anthem
Relating Answer Set Programming and Many-sorted Logics for Formal Verification
Answer Set Programming (ASP) is an important logic programming paradigm within the field of Knowledge Representation and Reasoning. As a concise, human-readable, declarative language, ASP is an excellent tool for developing trustworthy (especially, artificially intelligent) software systems. However, formally verifying ASP programs offers some unique challenges, such as 1. a lack of modularity (the meanings of rules are difficult to define in isolation from the enclosing program), 2. the ground-and-solve semantics (the meanings of rules are dependent on the input data with which the program is grounded), and 3. limitations of existing tools. My research agenda has been focused on addressing these three issues with the intention of making ASP verification an accessible, routine task that is regularly performed alongside program development. In this vein, I have investigated alternative semantics for ASP based on translations into the logic of here-and-there and many-sorted first-order logic. These semantics promote a modular understanding of logic programs, bypass grounding, and enable us to use automated theorem provers to automatically verify properties of programs.
- North America > United States > Nebraska > Douglas County > Omaha (0.28)
- North America > United States > California > San Francisco County > San Francisco (0.14)
- South America > Paraguay > Asunción > Asunción (0.04)
- (10 more...)
RECALL: Library-Like Behavior In Language Models is Enhanced by Self-Referencing Causal Cycles
Nwadike, Munachiso, Iklassov, Zangir, Aremu, Toluwani, Hiraoka, Tatsuya, Bojkovic, Velibor, Heinzerling, Benjamin, Alqaubeh, Hilal, Takáč, Martin, Inui, Kentaro
We introduce the concept of the self-referencing causal cycle (abbreviated RECALL) - a mechanism that enables large language models (LLMs) to bypass the limitations of unidirectional causality, which underlies a phenomenon known as the reversal curse. When an LLM is prompted with sequential data, it often fails to recall preceding context. For example, when we ask an LLM to recall the line preceding "O say does that star-spangled banner yet wave" in the U.S. National Anthem, it often fails to correctly return "Gave proof through the night that our flag was still there" - this is due to the reversal curse. It occurs because language models such as ChatGPT and Llama generate text based on preceding tokens, requiring facts to be learned and reproduced in a consistent token order. While the reversal curse is often viewed as a limitation, we offer evidence of an alternative view: it is not always an obstacle in practice. We find that RECALL is driven by what we designate as cycle tokens - sequences that connect different parts of the training data, enabling recall of preceding tokens from succeeding ones. Through rigorous probabilistic formalization and controlled experiments, we demonstrate how the cycles they induce influence a model's ability to reproduce information. To facilitate reproducibility, we provide our code and experimental details at https://anonymous.4open.science/r/remember-B0B8/.
- North America > United States (0.48)
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.14)
- Asia > Japan > Honshū > Tōhoku (0.04)
Reverse Training to Nurse the Reversal Curse
Golovneva, Olga, Allen-Zhu, Zeyuan, Weston, Jason, Sukhbaatar, Sainbayar
Large language models (LLMs) have a surprising failure: when trained on "A has a feature B", they do not generalize to "B is a feature of A", which is termed the Reversal Curse. Even when training with trillions of tokens this issue still appears due to Zipf's law - hence even if we train on the entire internet. This work proposes an alternative training scheme, called reverse training, whereby all words are used twice, doubling the amount of available tokens. The LLM is trained in both forward and reverse directions by reversing the training strings while preserving (i.e., not reversing) chosen substrings, such as entities. We show that data-matched reverse-trained models provide superior performance to standard models on standard tasks, and compute-matched reverse-trained models provide far superior performance on reversal tasks, helping resolve the reversal curse issue.
- Europe > France (0.05)
- North America > United States > Massachusetts (0.04)
- North America > United States > New Jersey > Mercer County > Princeton (0.04)
- (2 more...)
Do Large Language Models Latently Perform Multi-Hop Reasoning?
Yang, Sohee, Gribovskaya, Elena, Kassner, Nora, Geva, Mor, Riedel, Sebastian
We study whether Large Language Models (LLMs) latently perform multi-hop reasoning with complex prompts such as "The mother of the singer of 'Superstition' is". We look for evidence of a latent reasoning pathway where an LLM (1) latently identifies "the singer of 'Superstition'" as Stevie Wonder, the bridge entity, and (2) uses its knowledge of Stevie Wonder's mother to complete the prompt. We analyze these two hops individually and consider their co-occurrence as indicative of latent multi-hop reasoning. For the first hop, we test if changing the prompt to indirectly mention the bridge entity instead of any other entity increases the LLM's internal recall of the bridge entity. For the second hop, we test if increasing this recall causes the LLM to better utilize what it knows about the bridge entity. We find strong evidence of latent multi-hop reasoning for the prompts of certain relation types, with the reasoning pathway used in more than 80% of the prompts. However, the utilization is highly contextual, varying across different types of prompts. Also, on average, the evidence for the second hop and the full multi-hop traversal is rather moderate and only substantial for the first hop. Moreover, we find a clear scaling trend with increasing model size for the first hop of reasoning but not for the second hop. Our experimental findings suggest potential challenges and opportunities for future development and applications of LLMs.
- Africa > Middle East > Somalia (0.14)
- North America > Nicaragua (0.14)
- Europe > Finland (0.14)
- (30 more...)
- Research Report > New Finding (0.66)
- Research Report > Experimental Study (0.46)
- Media (1.00)
- Leisure & Entertainment (1.00)
- Government > Regional Government > North America Government (0.46)
Automated Verification of Equivalence Properties in Advanced Logic Programs -- Bachelor Thesis
With the increase in industrial applications using Answer Set Programming, the need for formal verification tools, particularly for critical applications, has also increased. During the program optimisation process, it would be desirable to have a tool which can automatically verify whether an optimised subprogram can replace the original subprogram. Formally this corresponds to the problem of verifying the strong equivalence of two programs. In order to do so, the translation tool anthem was developed. It can be used in conjunction with an automated theorem prover for classical logic to verify that two programs are strongly equivalent. With the current version of anthem, only the strong equivalence of positive programs with a restricted input language can be verified. This is a result of the translation $\tau^*$ implemented in anthem that produces formulas in the logic of here-and-there, which coincides with classical logic only for positive programs. This thesis extends anthem in order to overcome these limitations. First, the transformation $\sigma^*$ is presented, which transforms formulas from the logic of here-and-there to classical logic. A theorem formalises how $\sigma^*$ can be used to express equivalence in the logic of here-and-there in classical logic. Second, the translation $\tau^*$ is extended to programs containing pools. Another theorem shows how $\sigma^*$ can be combined with $\tau^*$ to express the strong equivalence of two programs in classical logic. With $\sigma^*$ and the extended $\tau^*$, it is possible to express the strong equivalence of logic programs containing negation, simple choices, and pools. Both the extended $\tau^*$ and $\sigma^*$ are implemented in a new version of anthem. Several examples of logic programs containing pools, negation, and simple choice rules, which the new version of anthem can translate to classical logic, are presented. Some a...
'We got bored waiting for Oasis to re-form': AIsis, the band fronted by an AI Liam Gallagher
Before you do anything else with your day, you need to listen to this. A new "lost" Oasis album has been released, from the period between their third album, 1997's Be Here Now, and their fourth, 2000's Standing on the Shoulder of Giants. It was created by AI – or at least, it's an AI Liam Gallagher doing its best "hellooooos" and "sun-shiiiines" over a real band. But the eight songs, including Out of My Mind, Coming of Age and Forever, are practically indistinguishable from the real thing, with some seriously catchy melodies that give every post-What's the Story album – not to mention the whole of Liam and Noel's solo catalogues – a run for their money. How do you get a computer to sing like Liam?
- Media > Music (1.00)
- Leisure & Entertainment (1.00)
Tools and Methodologies for Verifying Answer Set Programs
Answer Set Programming (ASP) is a powerful declarative programming paradigm commonly used for solving challenging search and optimization problems. The modeling languages of ASP are supported by sophisticated solving algorithms (solvers) that make the solution search efficient while enabling the programmer to model the problem at a high level of abstraction. As an approach to Knowledge Representation and Reasoning, ASP benefits from its simplicity, conciseness and rigorously defined semantics. These characteristics make ASP a straightforward way to develop formally verifiable programs. In the context of artificial intelligence (AI), the clarity of ASP programs lends itself to the construction of explainable, trustworthy AI. In support of these goals, my research is concerned with extending the theory and tools supporting the verification of ASP progams.
- North America > United States > Nebraska > Douglas County > Omaha (0.14)
- North America > United States > Massachusetts > Suffolk County > Boston (0.04)
- Europe > Russia > Northwestern Federal District > Leningrad Oblast > Saint Petersburg (0.04)
- (2 more...)
Deepfake attacks can easily trick facial recognition
In brief Miscreants can easily steal someone else's identity by tricking live facial recognition software using deepfakes, according to a new report. Sensity AI, a startup focused on tackling identity fraud, carried out a series of pretend attacks. Engineers scanned the image of someone from an ID card, and mapped their likeness onto another person's face. Sensity then tested whether they could breach live facial recognition systems by tricking them into believing the pretend attacker is a real user. So-called "liveness tests" try to authenticate identities in real-time, relying on images or video streams from cameras like face recognition used to unlock mobile phones, for example.
Anthem Looks to Fuel AI Efforts With Petabytes of Synthetic Data
The ultimate goal, he said, is to validate and train AI algorithms on large amounts of data, while reducing privacy issues surrounding personal medical information. "More and more…synthetic data is going to overtake and be the way people do AI in the future," Mr. Bhatt said. Anthem, which has been using Amazon.com Inc.'s Amazon Web Services as a cloud provider since 2017, tapped Google Cloud last year for its data analytics and AI capabilities as part of an ongoing effort to become more customer-centric and focus on members' entire healthcare journeys, Mr. Bhatt said. It's a continuing effort that includes Anthem's work with synthetic data. This week, Anthem's shareholders are voting on a proposed rebranding of the company to Elevance Health as part of that same effort.
- Information Technology > Artificial Intelligence (1.00)
- Information Technology > Communications > Web (0.74)
- Information Technology > Data Science > Data Mining (0.74)
My Health Insurance Company Tries to Keep me Healthy
I am grateful to have the health insurance I have, and grateful for the payments they've made to resolve problems I've had. Nonetheless, I can't help but be astounded at the never-ending flow of expensive, incompetent, annoying and utterly useless interaction I have had with the company's computer systems. Why can't they (and others like them) get it right? The answer is simple: the company's leaders, like most enterprise companies, want to be leaders in technology. Today, that means funding big, publicized initiatives in AI and ML. Initiatives that will, of course, transform healthcare.
- Health & Medicine (1.00)
- Banking & Finance > Insurance (1.00)