Plotting

 taxnodes:Technology: Instructional Materials



The ToMCAT Dataset

Neural Information Processing Systems

We present a rich, multimodal dataset consisting of data from 40 teams of three humans conducting simulated urban search-and-rescue (SAR) missions in a Minecraftbased testbed, collected for the Theory of Mind-based Cognitive Architecture for Teams (ToMCAT) project. Modalities include two kinds of brain scan data-- functional near-infrared spectroscopy (fNIRS) and electroencephalography (EEG), as well as skin conductance, heart rate, eye tracking, face images, spoken dialog audio data with automatic speech recognition (ASR) transcriptions, game screenshots, gameplay data, game performance data, demographic data, and self-report questionnaires.


Appendix to: Predictive Querying for Autoregressive Neural Sequence Models 2

Neural Information Processing Systems

It is helpful to show both the exact summation form as well as the expected value representation as both will be useful in Section 4. Q3 The "hitting time" or the next occurrence of a specific event type a V is defined as ฯ„(a). The value a V can be easily replaced with a set of values A V in these representations. Interestingly, we can see that Q3 is a generalization of Q2 by noting that they are identical when A = {}. In practice, computing this exactly is intractable due to it being an infinite sum. There are two potential approaches one could take to subvert this. The other option is to produce a lower bound on this expression by evaluating the sum in Eq. (11) for the first K terms. As such, if we evaluate Eq. (11) up to K terms for both p Similar to Q3, we can also ask this query with sets A B V instead of values a, b.


No-Regret Online Reinforcement Learning with Adversarial Losses and Transitions

Neural Information Processing Systems

Existing online learning algorithms for adversarial Markov Decision Processes achieve O( T) regret after T rounds of interactions even if the loss functions are chosen arbitrarily by an adversary, with the caveat that the transition function has to be fixed. This is because it has been shown that adversarial transition functions make no-regret learning impossible. Despite such impossibility results, in this work, we develop algorithms that can handle both adversarial losses and adversarial transitions, with regret increasing smoothly in the degree of maliciousness of the adversary.


Lean Workbook: A large-scale Lean problem set formalized from natural language math problems

Neural Information Processing Systems

Large language models have demonstrated impressive capabilities across various natural language processing tasks, especially in solving mathematical problems. However, large language models are not good at math theorem proving using formal languages like Lean. A significant challenge in this area is the scarcity of training data available in these formal languages. To address this issue, we propose a novel pipeline that iteratively generates and filters synthetic data to translate natural language mathematical problems into Lean 4 statements, and vice versa. Our results indicate that the synthetic data pipeline can provide useful training data and improve the performance of LLMs in translating and understanding complex mathematical problems and proofs. Our final dataset contains about 57K formal-informal question pairs along with searched proof from the math contest forum and 21 new IMO questions.


Generative Visual Prompt: Unifying Distributional Control of Pre-Trained Generative Models

Neural Information Processing Systems

Generative models (e.g., GANs, diffusion models) learn the underlying data distribution in an unsupervised manner. However, many applications of interest require sampling from a particular region of the output space or sampling evenly over a range of characteristics. For efficient sampling in these scenarios, we propose Generative Visual Prompt (PromptGen), a framework for distributional control over pre-trained generative models by incorporating knowledge of other off-the-shelf models. PromptGen defines control as energy-based models (EBMs) and samples images in a feed-forward manner by approximating the EBM with invertible neural networks, avoiding optimization at inference. Our experiments demonstrate how PromptGen can efficiently sample from several unconditional generative models (e.g., StyleGAN2, StyleNeRF, diffusion autoencoder, NVAE) in a controlled or/and de-biased manner using various off-the-shelf models: (1) with the CLIP model as control, PromptGen can sample images guided by text, (2) with image classifiers as control, PromptGen can de-bias generative models across a set of attributes or attribute combinations, and (3) with inverse graphics models as control, PromptGen can sample images of the same identity in different poses.



Want to learn piano? AI can teach you faster than private lessons

Mashable

TL;DR: Learn to play piano with Skoove Premium Piano Lessons, an AI-powered teacher, now offering lifetime subscriptions for 116.41 (reg. Always wanted to learn piano? Nowadays, you don't need a teacher, and you certainly don't need to sit through boring classes. All it takes is a tablet, a keyboard, and your love of music. Skoove is an AI-powered piano tutoring app that listens while you play and gives you curated feedback and useful resources to improve your skills.


HOUDINI: Lifelong Learning as Program Synthesis

Neural Information Processing Systems

We present a neurosymbolic framework for the lifelong learning of algorithmic tasks that mix perception and procedural reasoning. Reusing high-level concepts across domains and learning complex procedures are key challenges in lifelong learning. We show that a program synthesis approach that combines gradient descent with combinatorial search over programs can be a more effective response to these challenges than purely neural methods.


Neural Attribution for Semantic Bug-Localization in Student Programs

Neural Information Processing Systems

Providing feedback is an integral part of teaching. Most open online courses on programming make use of automated grading systems to support programming assignments and give real-time feedback. These systems usually rely on test results to quantify the programs' functional correctness. They return failing tests to the students as feedback. However, students may find it difficult to debug their programs if they receive no hints about where the bug is and how to fix it. In this work, we present NeuralBugLocator, a deep learning based technique, that can localize the bugs in a faulty program with respect to a failing test, without even running the program. At the heart of our technique is a novel tree convolutional neural network which is trained to predict whether a program passes or fails a given test. To localize the bugs, we analyze the trained network using a state-of-the-art neural prediction attribution technique and see which lines of the programs make it predict the test outcomes. Our experiments show that NeuralBugLocator is generally more accurate than two state-of-the-art program-spectrum based and one syntactic difference based bug-localization baselines.