Import AI: #90: Training massive networks via 'codistillation', talking to books via a new Google AI experiment, and why the ACM thinks researchers should consider the downsides of research
Training unprecedentedly large networks with'codistillation': …New technique makes it easier to train very large, distributed AI systems, without adding too much complexity… When it comes to applied AI, bigger can frequently be better; access to more data, more compute, and (occasionally) more complex infrastructures can frequently allow people to obtain better performance at lower cost. One limit is in the ability for people to parallelize the computation of a single neural network during training. To deal with that, researchers at places like Google have introduced techniques like'ensemble distillation' which let you train multiple networks in parallel and use these to train a single'student' network that benefits from the aggregated learnings of its many parents. Though this technique has shown to be effective it is also quite fiddly and introduces additional complexity which can make people less keen to use it. New research from Google simplifies this idea via a technique they call'codistillaiton'.
Apr-17-2018, 05:46:02 GMT
- Country:
- North America > United States > California (0.04)
- Genre:
- Research Report > New Finding (0.34)
- Technology: