reduction cell
- Asia > China (0.05)
- Oceania > Australia > New South Wales > Sydney (0.04)
- North America > Canada > Ontario > Toronto (0.04)
Supplementary Material of IST A-NAS: Efficient and Consistent Neural Architecture Search by Sparse Coding Yibo Y ang
We perform our experiments on both CIFAR-10 and ImageNet. The images are normalized by mean and standard deviation. The images are normalized by mean and standard deviation. Concretely, the super-net for search is composed of 6 normal cells and 2 reduction cells, and has an initial number of channels of 16. Each cell has 6 nodes.
Theory-Inspired Path-Regularized Differential Network Architecture Search (Supplementary File)
Then Appendix C gives the proofs of the main results in Sec. 3, namely Theorem 1, by first introducing auxiliary theories Due to space limitation, we defer more experimental results and details to this appendix. Due to the high training cost, we fix two regularization parameters and then investigate the third one. This testifies the robustness of PR-DARTS to regularization parameters.Figure 3: Effects of regularization parameters Here we first display the selected reduction cell on CIRAR10 in Figure 1 (a). Next, we also report the average gate activate probability in the normal and reduction cells in Figure 1 (b). At the beginning of the search, we initialize the activation probability of each gate to be one.
- Asia > China (0.05)
- Oceania > Australia > New South Wales > Sydney (0.04)
- North America > Canada > Ontario > Toronto (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- Europe > Germany > Baden-Württemberg > Freiburg (0.05)
- North America > Canada (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- (2 more...)
Towards Accurate and Robust Architectures via Neural Architecture Search
Ou, Yuwei, Feng, Yuqi, Sun, Yanan
To defend deep neural networks from adversarial attacks, adversarial training has been drawing increasing attention for its effectiveness. However, the accuracy and robustness resulting from the adversarial training are limited by the architecture, because adversarial training improves accuracy and robustness by adjusting the weight connection affiliated to the architecture. In this work, we propose ARNAS to search for accurate and robust architectures for adversarial training. First we design an accurate and robust search space, in which the placement of the cells and the proportional relationship of the filter numbers are carefully determined. With the design, the architectures can obtain both accuracy and robustness by deploying accurate and robust structures to their sensitive positions, respectively. Then we propose a differentiable multi-objective search strategy, performing gradient descent towards directions that are beneficial for both natural loss and adversarial loss, thus the accuracy and robustness can be guaranteed at the same time. We conduct comprehensive experiments in terms of white-box attacks, black-box attacks, and transferability. Experimental results show that the searched architecture has the strongest robustness with the competitive accuracy, and breaks the traditional idea that NAS-based architectures cannot transfer well to complex tasks in robustness scenarios. By analyzing outstanding architectures searched, we also conclude that accurate and robust neural architectures tend to deploy different structures near the input and output, which has great practical significance on both hand-crafting and automatically designing of accurate and robust architectures.
ApproxDARTS: Differentiable Neural Architecture Search with Approximate Multipliers
Pinos, Michal, Sekanina, Lukas, Mrazek, Vojtech
Integrating the principles of approximate computing into the design of hardware-aware deep neural networks (DNN) has led to DNNs implementations showing good output quality and highly optimized hardware parameters such as low latency or inference energy. In this work, we present ApproxDARTS, a neural architecture search (NAS) method enabling the popular differentiable neural architecture search method called DARTS to exploit approximate multipliers and thus reduce the power consumption of generated neural networks. We showed on the CIFAR-10 data set that the ApproxDARTS is able to perform a complete architecture search within less than $10$ GPU hours and produce competitive convolutional neural networks (CNN) containing approximate multipliers in convolutional layers. For example, ApproxDARTS created a CNN showing an energy consumption reduction of (a) $53.84\%$ in the arithmetic operations of the inference phase compared to the CNN utilizing the native $32$-bit floating-point multipliers and (b) $5.97\%$ compared to the CNN utilizing the exact $8$-bit fixed-point multipliers, in both cases with a negligible accuracy drop. Moreover, the ApproxDARTS is $2.3\times$ faster than a similar but evolutionary algorithm-based method called EvoApproxNAS.
- Europe > Czechia > South Moravian Region > Brno (0.05)
- North America > United States > Virginia > Arlington County > Arlington (0.04)
- North America > United States > California > Los Angeles County > Long Beach (0.04)
- (3 more...)