Generating Training Datasets Using Energy Based Models that Actually Scale

#artificialintelligence 

Energy-Based Models(EBM) is one of the most promising areas of deep learning that hasn't seen a tremendous level of adoption yet. Conceptually, EBMs are a form of generative modeling that learns the key characteristics of a target dataset and tries to generate similar datasets. While EBMs results appealing because of its simplicity they have experienced many challenges when applied in real world applications. Recently, AI-powerhouse OpenAI published a new research paper that explores a new technique to create EBM model that can scale across complex deep learning topologies. EBMs are typically used in one of the most complex problems of real world deep learning solutions: generating quality training datasets.