Goto

Collaborating Authors

 construal


Estimating cognitive biases with attention-aware inverse planning

Banerjee, Sounak, Cornelisse, Daphne, Gopinath, Deepak, Sumner, Emily, DeCastro, Jonathan, Rosman, Guy, Vinitsky, Eugene, Ho, Mark K.

arXiv.org Artificial Intelligence

People's goal-directed behaviors are influenced by their cognitive biases, and autonomous systems that interact with people should be aware of this. For example, people's attention to objects in their environment will be biased in a way that systematically affects how they perform everyday tasks such as driving to work. Here, building on recent work in computational cognitive science, we formally articulate the attention-aware inverse planning problem, in which the goal is to estimate a person's attentional biases from their actions. We demonstrate how attention-aware inverse planning systematically differs from standard inverse reinforcement learning and how cognitive biases can be inferred from behavior. Finally, we present an approach to attention-aware inverse planning that combines deep reinforcement learning with computational cognitive modeling. We use this approach to infer the attentional strategies of RL agents in real-life driving scenarios selected from the Waymo Open Dataset, demonstrating the scalability of estimating cognitive biases with attention-aware inverse planning.


Analogy making as amortised model construction

Nagy, David G., Shen, Tingke, Zhou, Hanqi, Wu, Charley M., Dayan, Peter

arXiv.org Artificial Intelligence

Humans flexibly construct internal models to navigate novel situations. To be useful, these internal models must be sufficiently faithful to the environment that resource-limited planning leads to adequate outcomes; equally, they must be tractable to construct in the first place. We argue that analogy plays a central role in these processes, enabling agents to reuse solution-relevant structure from past experiences and amortise the computational costs of both model construction (construal) and planning. Formalis-ing analogies as partial homomorphisms between Markov decision processes, we sketch a framework in which abstract modules, derived from previous construals, serve as com-posable building blocks for new ones. This modular reuse allows for flexible adaptation of policies and representations across domains with shared structural essence.


Modelling Language

Grindrod, Jumbly

arXiv.org Artificial Intelligence

This paper argues that large language models have a valuable scientific role to play in serving as scientific models of a language. Linguistic study should not only be concerned with the cognitive processes behind linguistic competence, but also with language understood as an external, social entity. Once this is recognized, the value of large language models as scientific models becomes clear. This paper defends this position against a number of arguments to the effect that language models provide no linguistic insight. It also draws upon recent work in philosophy of science to show how large language models could serve as scientific models.


Concept Alignment as a Prerequisite for Value Alignment

Rane, Sunayana, Ho, Mark, Sucholutsky, Ilia, Griffiths, Thomas L.

arXiv.org Artificial Intelligence

Value alignment is essential for building AI systems that can safely and reliably interact with people. However, what a person values--and is even capable of valuing--depends on the concepts that they are currently using to understand and evaluate what happens in the world. The dependence of values on concepts means that concept alignment is a prerequisite for value alignment--agents need to align their representation of a situation with that of humans in order to successfully align their values. Here, we formally analyze the concept alignment problem in the inverse reinforcement learning setting, show how neglecting concept alignment can lead to systematic value mis-alignment, and describe an approach that helps minimize such failure modes by jointly reasoning about a person's concepts and values. Additionally, we report experimental results with human participants showing that humans reason about the concepts used by an agent when acting intentionally, in line with our joint reasoning model. People's thoughts and actions are fundamentally shaped by the concepts they use to represent the world and formulate their goals.


Control of mental representations in human planning

Ho, Mark K., Abel, David, Correa, Carlos G., Littman, Michael L., Cohen, Jonathan D., Griffiths, Thomas L.

arXiv.org Artificial Intelligence

One of the most striking features of human cognition is the capacity to plan. Two aspects of human planning stand out: its efficiency, even in complex environments, and its flexibility, even in changing environments. Efficiency is especially impressive because directly computing an optimal plan is intractable, even for modestly complex tasks, and yet people successfully solve myriad everyday problems despite limited cognitive resources. Standard accounts in psychology, economics, and artificial intelligence have suggested this is because people have a mental representation of a task and then use heuristics to plan in that representation. However, this approach generally assumes that mental representations are fixed. Here, we propose that mental representations can be controlled and that this provides opportunities to adaptively simplify problems so they can be more easily reasoned about -- a process we refer to as construal. We construct a formal model of this process and, in a series of large, pre-registered behavioral experiments, show both that construal is subject to online cognitive control and that people form value-guided construals that optimally balance the complexity of a representation and its utility for planning and acting. These results demonstrate how strategically perceiving and conceiving problems facilitates the effective use of limited cognitive resources.


Towards Better Forecasting by Fusing Near and Distant Future Visions

Cheng, Jiezhu, Huang, Kaizhu, Zheng, Zibin

arXiv.org Machine Learning

Multivariate time series forecasting is an important yet challenging problem in machine learning. Most existing approaches only forecast the series value of one future moment, ignoring the interactions between predictions of future moments with different temporal distance. Such a deficiency probably prevents the model from getting enough information about the future, thus limiting the forecasting accuracy. To address this problem, we propose Multi-Level Construal Neural Network (MLCNN), a novel multi-task deep learning framework. Inspired by the Construal Level Theory of psychology, this model aims to improve the predictive performance by fusing forecasting information (i.e., future visions) of different future time. We first use the Convolution Neural Network to extract multi-level abstract representations of the raw data for near and distant future predictions. We then model the interplay between multiple predictive tasks and fuse their future visions through a modified Encoder-Decoder architecture. Finally, we combine traditional Autoregression model with the neural network to solve the scale insensitive problem. Experiments on three real-world datasets show that our method achieves statistically significant improvements compared to the most state-of-the-art baseline methods, with average 4.59% reduction on RMSE metric and average 6.87% reduction on MAE metric.


The Language of Stories: A Conceptual Integration Approach

Dancygier, Barbara (University of British Columbia)

AAAI Conferences

Processing the language of a narrative text, be it a novel, a extended flashbacks). These subsequent levels of blending film, or a play, is a crucial component of narrative of narrative spaces eventually yield the emergent space, comprehension. The research reported here shows how traditionally described as'the story'. The final product of processes driven by general linguistic and conceptual narrative comprehension is thus a mental construct, a patterns of meaning construction prompt the reader's or mega-blend, which emerges through multiple levels of viewer's response to the narrative artifact.