Goto

Collaborating Authors

 pavlov


Why we should thank pigeons for our AI breakthroughs

MIT Technology Review

People looking for precursors to artificial intelligence often point to science fiction by authors like Isaac Asimov or thought experiments like the Turing test. But an equally important, if surprising and less appreciated, forerunner is Skinner's research with pigeons in the middle of the 20th century. Skinner believed that association--learning, through trial and error, to link an action with a punishment or reward--was the building block of every behavior, not just in pigeons but in all living organisms, including human beings. His "behaviorist" theories fell out of favor with psychologists and animal researchers in the 1960s but were taken up by computer scientists who eventually provided the foundation for many of the artificial-intelligence tools from leading firms like Google and OpenAI. These companies' programs are increasingly incorporating a kind of machine learning whose core concept--reinforcement--is taken directly from Skinner's school of psychology and whose main architects, the computer scientists Richard Sutton and Andrew Barto, won the 2024 Turing Award, an honor widely considered to be the Nobel Prize of computer science.


Artificial intelligence and the transformation of higher education institutions

Katsamakas, Evangelos, Pavlov, Oleg V., Saklad, Ryan

arXiv.org Artificial Intelligence

Artificial intelligence (AI) advances and the rapid adoption of generative AI tools like ChatGPT present new opportunities and challenges for higher education. While substantial literature discusses AI in higher education, there is a lack of a systemic approach that captures a holistic view of the AI transformation of higher education institutions (HEIs). To fill this gap, this article, taking a complex systems approach, develops a causal loop diagram (CLD) to map the causal feedback mechanisms of AI transformation in a typical HEI. Our model accounts for the forces that drive the AI transformation and the consequences of the AI transformation on value creation in a typical HEI. The article identifies and analyzes several reinforcing and balancing feedback loops, showing how, motivated by AI technology advances, the HEI invests in AI to improve student learning, research, and administration. The HEI must take measures to deal with academic integrity problems and adapt to changes in available jobs due to AI, emphasizing AI-complementary skills for its students. However, HEIs face a competitive threat and several policy traps that may lead to decline. HEI leaders need to become systems thinkers to manage the complexity of the AI transformation and benefit from the AI feedback loops while avoiding the associated pitfalls. We also discuss long-term scenarios, the notion of HEIs influencing the direction of AI, and directions for future research on AI transformation.


A new memristor-based neural network inspired by the notion of associative memory

#artificialintelligence

Classical conditioning is a psychological process through which animals or humans pair desired or unpleasant stimuli (e.g., food or a painful experiences) with a seemingly neutral stimulus (e.g., the sound of a bell, the flash of a light, etc.) after these two stimuli are repeatedly presented together. Russian psychologist Ivan Pavlov studied classical conditioning in great depth and introduced the idea of "associative memory," which entails building strong associations between the pleasant/unpleasant and neutral stimuli. Pavlov is renowned for his studies on dogs, in which he gave the animals food after they heard a specific sound for several trials. Interestingly, he observed that the dogs would eventually start salivating (i.e., anticipating the food) after hearing the sound, even if the food had not yet been presented to them. This suggests that they had learned to associate the sound with the arrival of food.


Dopamine and temporal difference learning: A fruitful relationship between neuroscience and AI

#artificialintelligence

Learning and motivation are driven by internal and external rewards. Many of our day-to-day behaviours are guided by predicting, or anticipating, whether a given action will result in a positive (that is, rewarding) outcome. The study of how organisms learn from experience to correctly anticipate rewards has been a productive research field for well over a century, since Ivan Pavlov's seminal psychological work. In his most famous experiment, dogs were trained to expect food some time after a buzzer sounded. These dogs began salivating as soon as they heard the sound, before the food had arrived, indicating they'd learned to predict the reward.


Demystifying artificial intelligence

#artificialintelligence

Natalie Lao was set on becoming an electrical engineer, like her parents, until she stumbled on course 6.S192 (Making Mobile Apps), taught by Professor Hal Abelson. Here was a blueprint for turning a smartphone into a tool for finding clean drinking water, or sorting pictures of faces, or doing just about anything. "I thought, I wish people knew building tech could be like this," she said on a recent afternoon, taking a break from writing her dissertation. After shifting her focus as an MIT undergraduate to computer science, Lao joined Abelson's lab, which was busy spreading its App Inventor platform and do-it-yourself philosophy to high school students around the world. App Inventor set Lao on her path to making it easy for anyone, from farmers to factory workers, to understand AI, and use it to improve their lives.


Neuroscience and Artificial Intelligence Are More Linked Than You'd Expect

#artificialintelligence

Artificial Intelligence (AI) is more linked to dopamine-reinforced learning than you may think. That's a mouthful, so for now just think of Pavlov's dog study. DeepMind AI published a blog post on their discovery that the human brain and AI learning methods are closely linked when it comes to learning through reward. Their findings were also published in the journal Nature on Wednesday. It's been a well-known fact for a while now that we humans, and many animals, learn through reward. We are motivated by external and internal factors to learn more.


When Free Is not Free: Pavlov's Humans and Behavior Modification Empires

#artificialintelligence

A Behavior Modification Empire is an organization that establishes the means of human interaction via highly addictive technologies that elicit Pavlovian reactions to social rewards and punishments and then sells the right to manipulate participants to third-parties. Jaron Lanier, a Virtual Reality pioneer and important Internet luminary, begins a 2018 TED talk by referring to an early computer scientist named Norbert Wiener. Weiner wrote a book called "The Human Use of Human Beings," in which he envisioned a dystopian future governed by a computer system that would gather "data from people and [provide] feedback to those people in real time in order to put them . . . in a Skinner box, in a behaviorist system." One could imagine a global computer system where everybody has devices on them all the time, and the devices are giving them feedback based on what they did, and the whole population is subject to a degree of behavior modification. And such a society would be insane, could not survive, could not face its problems. Of course, Weiner's notion proved to be eerily prophetic, and now that it's happened, Lanier believes we have to figure out how to survive it.


Tuna Fish School Human Engineers in Hydraulics

WIRED

Underwater robots do a lot of neat things--take photos of underwater volcanoes, track leopard sharks, and explore shipwrecks--but they could still learn a few things from fish. Tuna are built to cruise across oceans, usually at around 2 mph. But they can crank up to 45 mph at the drop of a snack (Michael Phelps races at around 5 or 6 mph, for comparison). And tuna are agile, too, able to whip after fast-turning squids or sardines. They owe their agility, in part, to a newfound hydraulic system that allows them to raise and lower some specialized fins.


Identifying and Tracking Switching, Non-Stationary Opponents: A Bayesian Approach

Hernandez-Leal, Pablo (Instituto Nacional de Astrofisica, Optica y Electronica (INAOE)) | Taylor, Matthew E. (Washington State University) | Rosman, Benjamin (University of the Witwatersrand) | Sucar, L. Enrique (Instituto Nacional de Astrofisica, Optica y Electronica (INAOE)) | Cote, Enrique Munoz de (Instituto Nacional de Astrofisica, Optica y Electronica (INAOE))

AAAI Conferences

In many situations, agents are required to use a set of strategies (behaviors) and switch among them during the course of an interaction. This work focuses on the problem of recognizing the strategy used by an agent within a small number of interactions. We propose using a Bayesian framework to address this problem. Bayesian policy reuse (BPR) has been empirically shown to be efficient at correctly detecting the best policy to use from a library in sequential decision tasks. In this paper we extend BPR to adversarial settings, in particular, to opponents that switch from one stationary strategy to another. Our proposed extension enables learning new models in an online fashion when the learning agent detects that the current policies are not performing optimally. Experiments presented in repeated games show that our approach is capable of efficiently detecting opponent strategies and reacting quickly to behavior switches, thereby yielding better performance than state-of-the-art approaches in terms of average rewards.


A Computational Study on Emotions and Temperament in Multi-Agent Systems

Reis, Luis Paulo, Barteneva, Daria, Lau, Nuno

arXiv.org Artificial Intelligence

Recent advances in neurosciences and psychology have provided evidence that affective phenomena pervade intelligence at many levels, being inseparable from the cognitionaction loop. Perception, attention, memory, learning, decisionmaking, adaptation, communication and social interaction are some of the aspects influenced by them. This work draws its inspirations from neurobiology, psychophysics and sociology to approach the problem of building autonomous robots capable of interacting with each other and building strategies based on temperamental decision mechanism. Modelling emotions is a relatively recent focus in artificial intelligence and cognitive modelling. Such models can ideally inform our understanding of human behavior. We may see the development of computational models of emotion as a core research focus that will facilitate advances in the large array of computational systems that model, interpret or influence human behavior. We propose a model based on a scalable, flexible and modular approach to emotion which allows runtime evaluation between emotional quality and performance. The results achieved showed that the strategies based on temperamental decision mechanism strongly influence the system performance and there are evident dependency between emotional state of the agents and their temperamental type, as well as the dependency between the team performance and the temperamental configuration of the team members, and this enable us to conclude that the modular approach to emotional programming based on temperamental theory is the good choice to develop computational mind models for emotional behavioral Multi-Agent systems.