rulebase
Transformers as Soft Reasoners over Language
Clark, Peter, Tafjord, Oyvind, Richardson, Kyle
AI has long pursued the goal of having systems reason over *explicitly provided* knowledge, but building suitable representations has proved challenging. Here we explore whether transformers can similarly learn to reason (or emulate reasoning), but using rules expressed in language, thus bypassing a formal representation. We provide the first demonstration that this is possible, and characterize the extent of this capability. To do this, we use a collection of synthetic datasets that test increasing levels of reasoning complexity (number of rules, presence of negation, and depth of chaining). We find transformers appear to learn rule-based reasoning with high (99%) accuracy on these datasets, and in a way that generalizes to test data requiring substantially deeper chaining than in the training data (95%+ scores). We also demonstrate that the models transfer well to two hand-authored rulebases, and to rulebases paraphrased into more natural language. These findings are significant as it suggests a new role for transformers, namely as a limited "soft theorem prover" operating over explicit theories in language. This in turn suggests new possibilities for explainability, correctability, and counterfactual reasoning in question-answering. All datasets and a live demo are available at http://rule-reasoning.apps.allenai.org/
- Europe > Germany > Brandenburg > Potsdam (0.04)
- North America > United States > Washington > King County > Seattle (0.04)
- Europe > Slovenia > Drava > Municipality of Benedikt > Benedikt (0.04)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Rule-Based Reasoning (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Logic & Formal Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Machine Learning (1.00)
7 Inverting the Resolution Principle S. H. Muggleton
Duce uses a set of transformations of propositional Horn clauses which successively compress the example material on the basis of generalizations and the addition of new terms. In the following descriptions of three of the six Duce operators, lower-case Greek letters stand for conjunctions of propositional symbols.
Towards The Inductive Acquisition of Temporal Knowledge
The ability to predict the future in a given domain can be acquired by discovering empirically from experience certain temporal patterns that tend to repeat unerringly. Previous works in time series analysis allow one to make quantitative predictions on the likely values of certain linear variables. Since certain types of knowledge are better expressed in symbolic forms, making qualitative predictions based on symbolic representations require a different approach. A domain independent methodology called TIM (Time based Inductive Machine) for discovering potentially uncertain temporal patterns from real time observations using the technique of inductive inference is described here.
- North America > United States > Illinois > Champaign County > Urbana (0.14)
- North America > United States > California > Santa Clara County > Palo Alto (0.04)
- North America > United States > California > San Mateo County > San Mateo (0.05)
- North America > United States > Wisconsin > Milwaukee County > Milwaukee (0.04)
- North America > United States > California > Los Angeles County > Pasadena (0.04)
- Europe > United Kingdom > England > Greater London > London > Hackney (0.04)
- Media > Music (1.00)
- Leisure & Entertainment (1.00)
- North America > United States > California > San Mateo County > San Mateo (0.05)
- North America > United States > Wisconsin > Milwaukee County > Milwaukee (0.04)
- North America > United States > California > Los Angeles County > Pasadena (0.04)
- Europe > United Kingdom > England > Greater London > London > Hackney (0.04)
- Media > Music (1.00)
- Leisure & Entertainment (1.00)
Bach in a Box - Real-Time Harmony
Spangler, Randall R., Goodman, Rodney M., Hawkins, Jim
The learning and inferencing algorithms presented here speak an extended form of the classical figured bass representation common in Bach's time. Paired with a melody, figured bass provides a sufficient amount of information to reconstruct the harmonic content of a piece of music. Figured bass has several characteristics which make it well-disposed to learning rules. It is a symbolic format which uses a relatively small alphabet of symbols. It is also hierarchical - it specifies first the chord function that is to be played at the current note/timestep, then the scale step to be played by the bass voice, then additional information as needed to specify the alto and tenor scale steps. This allows our algorithm to fire sets of rules sequentially, to first determine the chord function which should be associated with a new melody note, and then to use that chord function as an input attribute to subsequent rulebases which determine the bass, alto, and tenor scale steps. In this way we can build up the final chord from simpler pieces, each governed by a specialized rulebase.
- North America > United States > California > San Mateo County > San Mateo (0.05)
- North America > United States > Wisconsin > Milwaukee County > Milwaukee (0.04)
- North America > United States > California > Los Angeles County > Pasadena (0.04)
- Europe > United Kingdom > England > Greater London > London > Hackney (0.04)
- Media > Music (1.00)
- Leisure & Entertainment (1.00)