Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Recombination of Artificial Neural Networks (1901.03900v1)

Published 12 Jan 2019 in cs.NE and cs.LG

Abstract: We propose a genetic algorithm (GA) for hyperparameter optimization of artificial neural networks which includes chromosomal crossover as well as a decoupling of parameters (i.e., weights and biases) from hyperparameters (e.g., learning rate, weight decay, and dropout) during sexual reproduction. Children are produced from three parents; two contributing hyperparameters and one contributing the parameters. Our version of population-based training (PBT) combines traditional gradient-based approaches such as stochastic gradient descent (SGD) with our GA to optimize both parameters and hyperparameters across SGD epochs. Our improvements over traditional PBT provide an increased speed of adaptation and a greater ability to shed deleterious genes from the population. Our methods improve final accuracy as well as time to fixed accuracy on a wide range of deep neural network architectures including convolutional neural networks, recurrent neural networks, dense neural networks, and capsule networks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Aaron Vose (2 papers)
  2. Jacob Balma (3 papers)
  3. Alex Heye (1 paper)
  4. Alessandro Rigazzi (6 papers)
  5. Charles Siegel (16 papers)
  6. Diana Moise (5 papers)
  7. Benjamin Robbins (2 papers)
  8. Rangan Sukumar (1 paper)
Citations (2)

Summary

We haven't generated a summary for this paper yet.