Reward is not enough: can we liberate AI from the reinforcement learning paradigm?
–arXiv.org Artificial Intelligence
I present arguments against the hypothesis put forward by Silver, Singh, Precup, and Sutton ( https://www.sciencedirect.com/science/article/pii/S0004370221000862 ) : reward maximization is not enough to explain many activities associated with natural and artificial intelligence including knowledge, learning, perception, social intelligence, evolution, language, generalisation and imitation. I show such reductio ad lucrum has its intellectual origins in the political economy of Homo economicus and substantially overlaps with the radical version of behaviourism. I show why the reinforcement learning paradigm, despite its demonstrable usefulness in some practical application, is an incomplete framework for intelligence -- natural and artificial. Complexities of intelligent behaviour are not simply second-order complications on top of reward maximisation. This fact has profound implications for the development of practically usable, smart, safe and robust artificially intelligent agents.
arXiv.org Artificial Intelligence
Feb-8-2022
- Country:
- Europe > United Kingdom
- England > Cambridgeshire > Cambridge (0.04)
- North America
- Canada > Ontario
- Toronto (0.28)
- United States
- California
- San Francisco County > San Francisco (0.04)
- San Mateo County > San Mateo (0.04)
- Santa Clara County > Palo Alto (0.04)
- New York (0.04)
- California
- Canada > Ontario
- Europe > United Kingdom
- Genre:
- Research Report (1.00)
- Industry:
- Banking & Finance > Trading (0.67)
- Government > Military (0.46)
- Health & Medicine > Pharmaceuticals & Biotechnology (0.46)
- Law Enforcement & Public Safety > Crime Prevention & Enforcement (0.46)
- Leisure & Entertainment > Games (0.46)
- Technology: