This book presents a methodology and philosophy of empirical science based on large scale lossless data compression. In this view a theory is scientific if it can be used to build a data compression program, and it is valuable if it can compress a standard benchmark database to a small size, taking into account the length of the compressor itself. This methodology therefore includes an Occam principle as well as a solution to the problem of demarcation. Because of the fundamental difficulty of lossless compression, this type of research must be empirical in nature: compression can only be achieved by discovering and characterizing empirical regularities in the data. Because of this, the philosophy provides a way to reformulate fields such as computer vision and computational linguistics as empirical sciences: the former by attempting to compress databases of natural images, the latter by attempting to compress large text databases. The book argues that the rigor and objectivity of the compression principle should set the stage for systematic progress in these fields. The argument is especially strong in the context of computer vision, which is plagued by chronic problems of evaluation. The book also considers the field of machine learning. Here the traditional approach requires that the models proposed to solve learning problems be extremely simple, in order to avoid overfitting. However, the world may contain intrinsically complex phenomena, which would require complex models to understand. The compression philosophy can justify complex models because of the large quantity of data being modeled (if the target database is 100 Gb, it is easy to justify a 10 Mb model). The complex models and abstractions learned on the basis of the raw data (images, language, etc) can then be reused to solve any specific learning problem, such as face recognition or machine translation.
Albrecht, Stefano V. (University of Edinburgh) | Barreto, André M. S. (Brazilian National Laboratory for Scientific Computing) | Braziunas, Darius (Kobo Inc.) | Buckeridge, David L. (McGill University) | Cuayáhuitl, Heriberto (Heriot-Watt University) | Dethlefs, Nina (Heriot-Watt University) | Endres, Markus (University of Augsburg) | Farahmand, Amir-massoud (Carnegie Mellon University) | Fox, Mark (University of Toronto) | Frommberger, Lutz (University of Bremen) | Ganzfried, Sam (Carnegie Mellon University) | Gil, Yolanda (University of Southern California) | Guillet, Sébastien (Université du Québec à Chicoutimi) | Hunter, Lawrence E. (University of Colorado School of Medicine) | Jhala, Arnav (University of California Santa Cruz) | Kersting, Kristian (Technical University of Dortmund) | Konidaris, George (Massachusetts Institute of Technology) | Lecue, Freddy (IBM Research) | McIlraith, Sheila (University of Toronto) | Natarajan, Sriraam (Indiana University) | Noorian, Zeinab (University of Saskatchewan) | Poole, David (University of British Columbia) | Ronfard, Rémi (University of Grenoble) | Saffiotti, Alessandro (Orebro University) | Shaban-Nejad, Arash (McGill University) | Srivastava, Biplav (IBM Research) | Tesauro, Gerald (IBM Research) | Uceda-Sosa, Rosario (IBM Research) | Broeck, Guy Van den (Katholieke Universiteit Leuven) | Otterlo, Martijn van (Radboud University Nijmegen) | Wallace, Byron C. (University of Texas) | Weng, Paul (Pierre and Marie Curie University) | Wiens, Jenna (University of Michigan) | Zhang, Jie (Nanyang Technological University)
The AAAI-14 Workshop program was held Sunday and Monday, July 27–28, 2012, at the Québec City Convention Centre in Québec, Canada. Canada. The AAAI-14 workshop program included fifteen workshops covering a wide range of topics in artificial intelligence. The titles of the workshops were AI and Robotics; Artificial Intelligence Applied to Assistive Technologies and Smart Environments; Cognitive Computing for Augmented Human Intelligence; Computer Poker and Imperfect Information; Discovery Informatics; Incentives and Trust in Electronic Communities; Intelligent Cinematography and Editing; Machine Learning for Interactive Systems: Bridging the Gap between Perception, Action and Communication; Modern Artificial Intelligence for Health Analytics; Multiagent Interaction without Prior Coordination; Multidisciplinary Workshop on Advances in Preference Handling; Semantic Cities — Beyond Open Data to Models, Standards and Reasoning; Sequential Decision Making with Big Data; Statistical Relational AI; and The World Wide Web and Public Health Intelligence. This article presents short summaries of those events.
Decades of research in artificial intelligence (AI) have produced formidable technologies that are providing immense benefit to industry, government, and society. AI systems can now translate across multiple languages, identify objects in images and video, streamline manufacturing processes, and control cars. The deployment of AI systems has not only created a trillion-dollar industry that is projected to quadruple in three years, but has also exposed the need to make AI systems fair, explainable, trustworthy, and secure. Future AI systems will rightfully be expected to reason effectively about the world in which they (and people) operate, handling complex tasks and responsibilities effectively and ethically, engaging in meaningful communication, and improving their awareness through experience. Achieving the full potential of AI technologies poses research challenges that require a radical transformation of the AI research enterprise, facilitated by significant and sustained investment. These are the major recommendations of a recent community effort coordinated by the Computing Community Consortium and the Association for the Advancement of Artificial Intelligence to formulate a Roadmap for AI research and development over the next two decades.
In 1950, Claude Shannon published his seminal work on how to program a computer to play chess. Since then, developing game-playing programs that can compete with (and even exceed) the abilities of the human world champions has been a long-sought-after goal of the AI research community. In Shannon's time, it would have seemed unlikely that only a scant 50 years would be needed to develop programs that play world-class backgammon, checkers, chess, Othello, and Scrabble. These remarkable achievements are the result of a better understanding of the problems being solved, major algorithmic insights, and tremendous advances in hardware technology. Computer games research is one of the important success stories of AI. This article reviews the past successes, current projects, and future research directions for AI using computer games as a research test bed.
Apple on Thursday announced a handful of updates coming to its Apple TV set-top box. The biggest change will be the introduction of a new TV app, which essentially functions as a TV guide for your numerous streaming apps. Here's a closer look at what Apple announced during a media event: Apple's new TV app is designed to surface content from various streaming apps, like HBO Go and Hulu, and house them in one easily browsable place. The "Up Next" section, for example, will act as a hub for content you've already started watching across all of these services. If you started a movie but didn't finish it, the "Up Next" tab will let you pick up where you left off.