Tsai, Yun-Chan
Optimizing Data Augmentation Policy Through Random Unidimensional Search
Dong, Xiaomeng, Potter, Michael, Kumar, Gaurav, Tsai, Yun-Chan, Saripalli, V. Ratna, Trafalis, Theodore
It is no secret amongst deep learning researchers that finding the optimal data augmentation strategy during training can mean the difference between state-of-the-art performance and a run-of-the-mill result. To that end, the community has seen many efforts to automate the process of finding the perfect augmentation procedure for any task at hand. Unfortunately, even recent cutting-edge methods bring massive computational overhead, requiring as many as 100 full model trainings to settle on an ideal configuration. We show how to achieve equivalent performance using just 6 trainings with Random Unidimensional Augmentation. Source code is available at https://github.com/fastestimator/RUA/tree/v1.0
To Raise or Not To Raise: The Autonomous Learning Rate Question
Dong, Xiaomeng, Tan, Tao, Potter, Michael, Tsai, Yun-Chan, Kumar, Gaurav, Saripalli, V. Ratna, Trafalis, Theodore
There is a parameter ubiquitous throughout the deep learning world: learning rate. There is likewise a ubiquitous question: what should that learning rate be? The true answer to this question is often tedious and time consuming to obtain, and a great deal of arcane knowledge has accumulated in recent years over how to pick and modify learning rates to achieve optimal training performance. Moreover, the long hours spent carefully crafting the perfect learning rate can come to nothing the moment your network architecture, optimizer, dataset, or initial conditions change ever so slightly. But it need not be this way. We propose a new answer to the great learning rate question: the Autonomous Learning Rate Controller. Find it at https://github.com/fastestimator/ARC/tree/v2.0