An algorithmic framework for the optimization of deep neural networks architectures and hyperparameters
Keisler, Julie, Talbi, El-Ghazali, Claudel, Sandra, Cabriel, Gilles
–arXiv.org Artificial Intelligence
While each new learning task requires the handcrafted design of a new DNN, automated deep learning facilitates the creation of powerful DNNs. Interests are to give access to deep learning to less experienced people, to reduce the tedious tasks of managing many parameters to reach the optimal DNN, and finally, to go beyond what humans can design by creating non-intuitive DNNs that can ultimately prove to be more efficient. Optimizing a DNN means automatically finding an optimal architecture for a given learning task: choosing the operations and the connections between those operations and the associated hyperparameters. The first task is also known as Neural Architecture Search [Elsken et al., 2019], also named NAS, and the second, as HyperParameters Optimization (HPO). Most works from the literature try to tackle only one of these two optimization problems. Many papers related to NAS [White et al., 2021, Loni et al., 2020b, Wang et al., 2019b, Sun et al., 2018b, Zhong, 2020] focus on designing optimal architectures for computer vision tasks with a lot of stacked convolution and pooling layers. Because each DNN training is time-consuming, researchers tried to reduce the search space by adding many constraints preventing from finding irrelevant architectures. It affects the flexibility of the designed search spaces and limits the hyperparameters optimization.
arXiv.org Artificial Intelligence
Feb-27-2023
- Country:
- Europe > France (0.04)
- North America > Trinidad and Tobago
- Genre:
- Research Report (1.00)
- Industry:
- Energy > Power Industry > Utilities (0.46)
- Technology: