Import AI: #90: Training massive networks via 'codistillation', talking to books via a new Google AI experiment, and why the ACM thinks researchers should consider the downsides of research

#artificialintelligence 

Training unprecedentedly large networks with'codistillation': …New technique makes it easier to train very large, distributed AI systems, without adding too much complexity… When it comes to applied AI, bigger can frequently be better; access to more data, more compute, and (occasionally) more complex infrastructures can frequently allow people to obtain better performance at lower cost. One limit is in the ability for people to parallelize the computation of a single neural network during training. To deal with that, researchers at places like Google have introduced techniques like'ensemble distillation' which let you train multiple networks in parallel and use these to train a single'student' network that benefits from the aggregated learnings of its many parents. Though this technique has shown to be effective it is also quite fiddly and introduces additional complexity which can make people less keen to use it. New research from Google simplifies this idea via a technique they call'codistillaiton'.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found