Goto

Collaborating Authors

 houston


CORGI: Efficient Pattern Matching With Quadratic Guarantees

Weitekamp, Daniel

arXiv.org Artificial Intelligence

Rule-based systems must solve complex matching problems within tight time constraints to be effective in real-time applications, such as planning and reactive control for AI agents, as well as low-latency relational database querying. Pattern-matching systems can encounter issues where exponential time and space are required to find matches for rules with many underconstrained variables, or which produce combinatorial intermediate partial matches (but are otherwise well-constrained). When online AI systems automatically generate rules from example-driven induction or code synthesis, they can easily produce worst-case matching patterns that slow or halt program execution by exceeding available memory. In our own work with cognitive systems that learn from example, we've found that aggressive forms of anti-unification-based generalization can easily produce these circumstances. To make these systems practical without hand-engineering constraints or succumbing to unpredictable failure modes, we introduce a new matching algorithm called CORGI (Collection-Oriented Relational Graph Iteration). Unlike RETE-based approaches, CORGI offers quadratic time and space guarantees for finding single satisficing matches, and the ability to iteratively stream subsequent matches without committing entire conflict sets to memory. CORGI differs from RETE in that it does not have a traditional $β$-memory for collecting partial matches. Instead, CORGI takes a two-step approach: a graph of grounded relations is built/maintained in a forward pass, and an iterator generates matches as needed by working backward through the graph. This approach eliminates the high-latency delays and memory overflows that can result from populating full conflict sets. In a performance evaluation, we demonstrate that CORGI significantly outperforms RETE implementations from SOAR and OPS5 on a simple combinatorial matching task.


Texas's Water Wars

The New Yorker

As industrial operations move to the state, residents find that their drinking water has been promised to companies. In 2019, Corpus Christi, Texas's eighth-largest city, moved forward with plans to build a desalination plant. The facility, which was expected to be completed by 2023, at a cost of a hundred and forty million dollars, would convert seawater into fresh water to be used by the area's many refineries and chemical plants. The former mayor called it "a pretty significant day in the life of our city." In anticipation of the plant's opening, the city committed to provide tens of millions of gallons of water per day to new industrial operations, including a plastics plant co-owned by ExxonMobil and the Saudi Basic Industries Corporation, a lithium refinery for Tesla batteries, and a "specialty chemicals" plant operated by Chemours.


Addressing the alignment problem in transportation policy making: an LLM approach

Yan, Xiaoyu, Dai, Tianxing, Nie, Yu Marco

arXiv.org Artificial Intelligence

A key challenge in transportation planning is that the collective preferences of heterogeneous travelers often diverge from the policies produced by model-driven decision tools. This misalignment frequently results in implementation delays or failures. Here, we investigate whether large language models (LLMs)--noted for their capabilities in reasoning and simulating human decision-making--can help inform and address this alignment problem. We develop a multi-agent simulation in which LLMs, acting as agents representing residents from different communities in a city, participate in a referendum on a set of transit policy proposals. Using chain-of-thought reasoning, LLM agents provide Ranked-Choice or approval-based preferences, which are aggregated using instant-runoff voting (IRV) to model democratic consensus. We implement this simulation framework with both GPT-4o and Claude-3.5, and apply it for Chicago and Houston. Our findings suggest that LLM agents are capable of approximating plausible collective preferences and responding to local context, while also displaying model-specific behavioral biases and modest divergences from optimization-based benchmarks. These capabilities underscore both promise and limitations of LLMs as tools for solving the alignment problem in transportation decision-making. Introduction Urban transportation policy plays a central role in shaping regional development. Designing effective policy requires access to multidimensional data and a deep understanding of individual preferences across heterogeneous communities. Conventional approaches typically rely on structured mathematical models that identify an optimal policy under specified objectives and constraints. However, these models often rest on rigid assumptions and oversimplified behavioral representations. As a result, they may produce solutions that are analytically tractable yet poorly aligned with public sentiment or the complex realities of policy implementation. This misalignment frequently contributes to delays--or even failures--in policy approval and execution. Trained on vast corpora of text encompassing news, facts, and human discourse, LLMs possess a rich contextual understanding that could potentially help policymakers infer public preferences and explore trade-offs before implementation. Their ability to interpret unstructured information, reason about competing objectives in natural language, and adapt to specific contexts suggests a new form of decision support that complements the traditional paradigm. In this study, we implement a multi-agent voting framework to examine the potential of LLMs in supporting transportation policy design.


All Claims Are Equal, but Some Claims Are More Equal Than Others: Importance-Sensitive Factuality Evaluation of LLM Generations

Wanner, Miriam, Azzopardi, Leif, Thomas, Paul, Dan, Soham, Van Durme, Benjamin, Craswell, Nick

arXiv.org Artificial Intelligence

Existing methods for evaluating the factuality of large language model (LLM) responses treat all claims as equally important. This results in misleading evaluations when vital information is missing or incorrect as it receives the same weight as peripheral details, raising the question: how can we reliably detect such differences when there are errors in key information? Current approaches that measure factuality tend to be insensitive to omitted or false key information. To investigate this lack of sensitivity, we construct VITALERRORS, a benchmark of 6,733 queries with minimally altered LLM responses designed to omit or falsify key information. Using this dataset, we demonstrate the insensitivities of existing evaluation metrics to key information errors. To address this gap, we introduce VITAL, a set of metrics that provide greater sensitivity in measuring the factuality of responses by incorporating the relevance and importance of claims with respect to the query. Our analysis demonstrates that VITAL metrics more reliably detect errors in key information than previous methods. Our dataset, metrics, and analysis provide a foundation for more accurate and robust assessment of LLM factuality.


mmExpert: Integrating Large Language Models for Comprehensive mmWave Data Synthesis and Understanding

Yan, Yifan, Yang, Shuai, Guo, Xiuzhen, Wang, Xiangguang, Chow, Wei, Shu, Yuanchao, He, Shibo

arXiv.org Artificial Intelligence

Millimeter-wave (mmWave) sensing technology holds significant value in human-centric applications, yet the high costs associated with data acquisition and annotation limit its widespread adoption in our daily lives. Concurrently, the rapid evolution of large language models (LLMs) has opened up opportunities for addressing complex human needs. This paper presents mmExpert, an innovative mmWave understanding framework consisting of a data generation flywheel that leverages LLMs to automate the generation of synthetic mmWave radar datasets for specific application scenarios, thereby training models capable of zero-shot generalization in real-world environments. Extensive experiments demonstrate that the data synthesized by mmExpert significantly enhances the performance of downstream models and facilitates the successful deployment of large models for mmWave understanding.


To Theoretically Understand Transformer-Based In-Context Learning for Optimizing CSMA

Hao, Shugang, Li, Hongbo, Duan, Lingjie

arXiv.org Artificial Intelligence

The binary exponential backoff scheme is widely used in WiFi 7 and still incurs poor throughput performance under dynamic channel environments. Recent model-based approaches (e.g., non-persistent and $p$-persistent CSMA) simply optimize backoff strategies under a known and fixed node density, still leading to a large throughput loss due to inaccurate node density estimation. This paper is the first to propose LLM transformer-based in-context learning (ICL) theory for optimizing channel access. We design a transformer-based ICL optimizer to pre-collect collision-threshold data examples and a query collision case. They are constructed as a prompt as the input for the transformer to learn the pattern, which then generates a predicted contention window threshold (CWT). To train the transformer for effective ICL, we develop an efficient algorithm and guarantee a near-optimal CWT prediction within limited training steps. As it may be hard to gather perfect data examples for ICL in practice, we further extend to allow erroneous data input in the prompt. We prove that our optimizer maintains minimal prediction and throughput deviations from the optimal values. Experimental results on NS-3 further demonstrate our approach's fast convergence and near-optimal throughput over existing model-based and DRL-based approaches under unknown node densities.


159-year-old company embraces driverless trucks

FOX News

Driverless semitrucks raise questions about safety, reliability and the future of the trucking industry. A bold new pilot program is bringing autonomous trucking to the heart of Texas. That means robots are about to hit some of the country's busiest shipping lanes, with doors in tow. Sign up for my FREE CyberGuy Report Get my best tech tips, urgent security alerts and exclusive deals delivered straight to your inbox. Plus, you'll get instant access to my Ultimate Scam Survival Guide -- free when you join my CYBERGUY.COM/NEWSLETTER


Enhancing Privacy in Decentralized Min-Max Optimization: A Differentially Private Approach

Quan, Yueyang, Wang, Chang, Zhai, Shengjie, Fang, Minghong, Liu, Zhuqing

arXiv.org Artificial Intelligence

Decentralized min-max optimization allows multi-agent systems to collaboratively solve global min-max optimization problems by facilitating the exchange of model updates among neighboring agents, eliminating the need for a central server. However, sharing model updates in such systems carry a risk of exposing sensitive data to inference attacks, raising significant privacy concerns. To mitigate these privacy risks, differential privacy (DP) has become a widely adopted technique for safeguarding individual data. Despite its advantages, implementing DP in decentralized min-max optimization poses challenges, as the added noise can hinder convergence, particularly in non-convex scenarios with complex agent interactions in min-max optimization problems. In this work, we propose an algorithm called DPMixSGD (Differential Private Minmax Hybrid Stochastic Gradient Descent), a novel privacy-preserving algorithm specifically designed for non-convex decentralized min-max optimization. Our method builds on the state-of-the-art STORM-based algorithm, one of the fastest decentralized min-max solutions. We rigorously prove that the noise added to local gradients does not significantly compromise convergence performance, and we provide theoretical bounds to ensure privacy guarantees. To validate our theoretical findings, we conduct extensive experiments across various tasks and models, demonstrating the effectiveness of our approach.


emoji-development-face-tears-joy-book-keith-houston.html?via=rss

Slate

A couple of years ago, I frequently found myself driving past a roadside ice cream stand under construction. For weeks, the roof of this stand, a gigantic white swirl of fiberglass soft serve, sat on the ground next to the structure, waiting to be lowered onto the finished, cone-shaped building with a crane. I know what it was supposed to represent, but every time I glimpsed it, my instinctive first thought was There's a giant poop emoji. Keith Houston's history of emoji, Face With Tears of Joy, argues that emoji have "become so ubiquitous in our writing, so quotidian, that we should be talking about them in the same breath as grammar or punctuation." I don't know about grammar, which seems as fundamental to language, spoken and written, as words themselves.

  Country: Asia > Japan (0.15)
  Industry: Government (0.48)

From Prompts to Propositions: A Logic-Based Lens on Student-LLM Interactions

Alfageeh, Ali, Zarkouei, Sadegh AlMahdi Kazemi, Nam, Daye, Prol, Daniel, Amoozadeh, Matin, Chattopadhyay, Souti, Prather, James, Denny, Paul, Leinonen, Juho, Hilton, Michael, Ragavan, Sruti Srinivasa, Alipour, Mohammad Amin

arXiv.org Artificial Intelligence

Background and Context. The increasing integration of large language models (LLMs) in computing education presents an emerging challenge in understanding how students use LLMs and craft prompts to solve computational tasks. Prior research has used both qualitative and quantitative methods to analyze prompting behavior, but these approaches lack scalability or fail to effectively capture the semantic evolution of prompts. Objective. In this paper, we investigate whether students prompts can be systematically analyzed using propositional logic constraints. We examine whether this approach can identify patterns in prompt evolution, detect struggling students, and provide insights into effective and ineffective strategies. Method. We introduce Prompt2Constraints, a novel method that translates students prompts into logical constraints. The constraints are able to represent the intent of the prompts in succinct and quantifiable ways. We used this approach to analyze a dataset of 1,872 prompts from 203 students solving introductory programming tasks. Findings. We find that while successful and unsuccessful attempts tend to use a similar number of constraints overall, when students fail, they often modify their prompts more significantly, shifting problem-solving strategies midway. We also identify points where specific interventions could be most helpful to students for refining their prompts. Implications. This work offers a new and scalable way to detect students who struggle in solving natural language programming tasks. This work could be extended to investigate more complex tasks and integrated into programming tools to provide real-time support.