Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning to Infer Generative Template Programs for Visual Concepts (2403.15476v2)

Published 20 Mar 2024 in cs.CV, cs.AI, cs.GR, and cs.LG

Abstract: People grasp flexible visual concepts from a few examples. We explore a neurosymbolic system that learns how to infer programs that capture visual concepts in a domain-general fashion. We introduce Template Programs: programmatic expressions from a domain-specific language that specify structural and parametric patterns common to an input concept. Our framework supports multiple concept-related tasks, including few-shot generation and co-segmentation through parsing. We develop a learning paradigm that allows us to train networks that infer Template Programs directly from visual datasets that contain concept groupings. We run experiments across multiple visual domains: 2D layouts, Omniglot characters, and 3D shapes. We find that our method outperforms task-specific alternatives, and performs competitively against domain-specific approaches for the limited domains where they exist.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. R. Kenny Jones (16 papers)
  2. Siddhartha Chaudhuri (40 papers)
  3. Daniel Ritchie (50 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.