Publications

Learning Morphisms with Gauss-Newton Approximation for Growing Networks

Abstract

A popular method for Neural Architecture Search (NAS) is based on growing networks via small local changes to the network's architecture called network morphisms. These methods start with a small seed network and progressively grow the network by adding new neurons in an automated way. However, it remains a challenge to efficiently determine which parts of the network are best to grow. Here we propose a NAS method for growing a network by using a Gauss-Newton approximation of the loss function to efficiently learn and evaluate candidate network morphisms. We compare our method with state of the art NAS methods for CIFAR-10 and CIFAR-100 classification tasks, and conclude our method learns similar quality or better architectures at a smaller computational cost.

Metadata

publication
arXiv preprint arXiv:2411.05855, 2024
year
2024
publication date
2024/11/7
authors
Neal Lawton, Aram Galstyan, Greg Ver Steeg
link
https://arxiv.org/abs/2411.05855
resource_link
https://arxiv.org/pdf/2411.05855
journal
arXiv preprint arXiv:2411.05855