Appendix of Joint Data-T ask Generation for Auxiliary Learning Hong Chen
–Neural Information Processing Systems
We provide the derivation of the upper implicit gradient in eq. We summarize the whole DTG-AuxL algorithm in Algorithm 1, where the lower and upper optimization updates are conducted alternatingly. We use the batch stochastic gradient optimization for both the lower and upper update. STL: It is a natural baseline where we only train on the primary task. Equal: It is a multi-task learning method, where we assign an equal weight of 1.0 to the loss of each MAXL can be only applied to the classification problem.
Neural Information Processing Systems
Nov-14-2025, 13:30:10 GMT
- Country:
- Asia > China
- Europe
- North America
- Canada > British Columbia
- Vancouver (0.04)
- United States
- Maryland > Baltimore (0.04)
- Utah > Salt Lake County
- Salt Lake City (0.04)
- Canada > British Columbia
- Technology: