Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Towards modular and programmable architecture search (1909.13404v1)

Published 30 Sep 2019 in cs.LG, cs.AI, and stat.ML

Abstract: Neural architecture search methods are able to find high performance deep learning architectures with minimal effort from an expert. However, current systems focus on specific use-cases (e.g. convolutional image classifiers and recurrent LLMs), making them unsuitable for general use-cases that an expert might wish to write. Hyperparameter optimization systems are general-purpose but lack the constructs needed for easy application to architecture search. In this work, we propose a formal language for encoding search spaces over general computational graphs. The language constructs allow us to write modular, composable, and reusable search space encodings and to reason about search space design. We use our language to encode search spaces from the architecture search literature. The language allows us to decouple the implementations of the search space and the search algorithm, allowing us to expose search spaces to search algorithms through a consistent interface. Our experiments show the ease with which we can experiment with different combinations of search spaces and search algorithms without having to implement each combination from scratch. We release an implementation of our language with this paper.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Renato Negrinho (8 papers)
  2. Darshan Patil (5 papers)
  3. Nghia Le (3 papers)
  4. Daniel Ferreira (6 papers)
  5. Matthew Gormley (3 papers)
  6. Geoffrey Gordon (15 papers)
Citations (26)

Summary

We haven't generated a summary for this paper yet.