Goto

Collaborating Authors

 programmable resistor


MIT Researchers, Working On Analog Deep Learning, Introduce A New Hardware Powered By Ultra-Fast Protonics And With Much Less Energy

#artificialintelligence

The amount of time, effort, and resources needed to train increasingly complicated neural network models is soaring as more machine learning experiments are being done. In order to combat this, a brand-new branch of artificial intelligence called "analog deep learning" is on the rise. It promises faster processing with far less energy consumption. Like transistors are the essential components of digital computers, programmable resistors are the fundamental building blocks of analog deep learning. Researchers have developed a network of analog artificial "neurons" and "synapses" that can do calculations similarly to a digital neural network by repeatedly repeating arrays of programmable resistors in intricate layers.


New hardware offers faster computation for artificial intelligence, with much less energy

#artificialintelligence

As scientists push the boundaries of machine learning, the amount of time, energy, and money required to train increasingly complex neural network models is skyrocketing. A new area of artificial intelligence called analog deep learning promises faster computation with a fraction of the energy usage. Programmable resistors are the key building blocks in analog deep learning, just like transistors are the core elements for digital processors. By repeating arrays of programmable resistors in complex layers, researchers can create a network of analog artificial "neurons" and "synapses" that execute computations just like a digital neural network. This network can then be trained to achieve complex AI tasks like image recognition and natural language processing. A multidisciplinary team of MIT researchers set out to push the speed limits of a type of human-made analog synapse that they had previously developed.


New hardware offers faster computation for artificial intelligence, with much less energy

#artificialintelligence

As scientists push the boundaries of machine learning, the amount of time, energy, and money required to train increasingly complex neural network models is skyrocketing. A new area of artificial intelligence called analog deep learning promises faster computation with a fraction of the energy usage. Programmable resistors are the key building blocks in analog deep learning, just like transistors are the core elements for digital processors. By repeating arrays of programmable resistors in complex layers, researchers can create a network of analog artificial "neurons" and "synapses" that execute computations just like a digital neural network. This network can then be trained to achieve complex AI tasks like image recognition and natural language processing.