Goto

Collaborating Authors

 concat range


Noisy Differentiable Architecture Search

Chu, Xiangxiang, Zhang, Bo, Li, Xudong

arXiv.org Machine Learning

Simplicity is the ultimate sophistication. Differentiable Architecture Search (DARTS) has now become one of the mainstream paradigms of neural architecture search. However, it largely suffers from several disturbing factors of the optimization process whose results are unstable to reproduce. Chu et al. [2019a] point out that skip connections natively have an unfair advantage in exclusive competition which primarily leads to dramatic performance collapse. While Fair-DARTS (Chu et al. [2019a]) turns the unfair competition into a collaborative one, we instead impede such unfair advantage by injecting unbiased random noise into skip operations' output. In effect, the optimizer should perceive this difficulty at each training step and refrain from overshooting on skip connections, but in a long run, it still converges to the right solution area since no bias is added to the gradient in terms of expectation. We name this novel approach as NoisyDARTS. Our experiments on CIFAR-10 and ImageNet attest that it can effectively break the skip connections' unfair advantage and yield better performance. It generates a series of models that achieve state-of-the-art results on both datasets.


Fair DARTS: Eliminating Unfair Advantages in Differentiable Architecture Search

Chu, Xiangxiang, Zhou, Tianbao, Zhang, Bo, Li, Jixiang

arXiv.org Artificial Intelligence

Differential Architecture Search (DARTS) is now a widely disseminated weight-sharing neural architecture search method. However, there are two fundamental weaknesses remain untackled. First, we observe that the well-known aggregation of skip connections during optimization is caused by an unfair advantage in an exclusive competition. Second, there is a non-negligible incongruence when discretizing continuous architectural weights to a one-hot representation. Because of these two reasons, DARTS delivers a biased solution that might not even be suboptimal. In this paper, we present a novel approach to curing both frailties. Specifically, as unfair advantages in a pure exclusive competition easily induce a monopoly, we relax the choice of operations to be collaborative, where we let each operation have an equal opportunity to develop its strength. We thus call our method Fair DARTS. Moreover, we propose a zero-one loss to directly reduce the discretization gap. Experiments are performed on two mainstream search spaces, in which we achieve new state-of-the-art networks on ImageNet. Our code is available on https://github.com/xiaomi-automl/fairdarts.