Goto

Collaborating Authors

 adopter


Attention to Non-Adopters

Zhou, Kaitlyn, Gligorić, Kristina, Cheng, Myra, Lam, Michelle S., Raman, Vyoma, Aminu, Boluwatife, Woo, Caeley, Brockman, Michael, Cha, Hannah, Jurafsky, Dan

arXiv.org Artificial Intelligence

Although language model-based chat systems are increasingly used in daily life, most Americans remain non-adopters of chat-based LLMs -- as of June 2025, 66% had never used ChatGPT. At the same time, LLM development and evaluation rely mainly on data from adopters (e.g., logs, preference data), focusing on the needs and tasks for a limited demographic group of adopters in terms of geographic location, education, and gender. In this position paper, we argue that incorporating non-adopter perspectives is essential for developing broadly useful and capable LLMs. We contend that relying on methods that focus primarily on adopters will risk missing a range of tasks and needs prioritized by non-adopters, entrenching inequalities in who benefits from LLMs, and creating oversights in model development and evaluation. To illustrate this claim, we conduct case studies with non-adopters and show: how non-adopter needs diverge from those of current users, how non-adopter needs point us towards novel reasoning tasks, and how to systematically integrate non-adopter needs via human-centered methods.


Selecting Critical Scenarios of DER Adoption in Distribution Grids Using Bayesian Optimization

Mulkin, Olivier, Heleno, Miguel, Ludkovski, Mike

arXiv.org Machine Learning

We develop a new methodology to select scenarios of DER adoption most critical for distribution grids. Anticipating risks of future voltage and line flow violations due to additional PV adopters is central for utility investment planning but continues to rely on deterministic or ad hoc scenario selection. We propose a highly efficient search framework based on multi-objective Bayesian Optimization. We treat underlying grid stress metrics as computationally expensive black-box functions, approximated via Gaussian Process surrogates and design an acquisition function based on probability of scenarios being Pareto-critical across a collection of line- and bus-based violation objectives. Our approach provides a statistical guarantee and offers an order of magnitude speed-up relative to a conservative exhaustive search. Case studies on realistic feeders with 200-400 buses demonstrate the effectiveness and accuracy of our approach.


Leakage-Robust Bayesian Persuasion

Haghtalab, Nika, Qiao, Mingda, Yang, Kunhe

arXiv.org Artificial Intelligence

We introduce the concept of leakage-robust Bayesian persuasion. Situated between public persuasion [KG11, CCG23, Xu20] and private persuasion [AB19], leakage-robust persuasion considers a setting where one or more signals privately sent by a sender to the receivers may be leaked. We study the design of leakage-robust persuasion schemes and quantify the price of robustness using two formalisms: - The first notion, $k$-worst-case persuasiveness, requires a scheme to remain persuasive as long as each receiver observes at most $k$ leaked signals. We quantify the Price of Worst-case Robustness (PoWR$_k$) -- i.e., the gap in sender's utility as compared to the optimal private scheme -- as $\Theta(\min\{2^k,n\})$ for supermodular sender utilities and $\Theta(k)$ for submodular or XOS utilities, where $n$ is the number of receivers. This result also establishes that in some instances, $\Theta(\log k)$ leakages are sufficient for the utility of the optimal leakage-robust persuasion to degenerate to that of public persuasion. - The second notion, expected downstream utility robustness, relaxes the persuasiveness and considers the impact on sender's utility when receivers best respond to their observations. By quantifying the Price of Downstream Robustness (PoDR) as the gap between the sender's expected utility over random leakage patterns as compared to private persuasion, we show that over several natural and structured distributions of leakage patterns, PoDR improves PoWR to $\Theta(k)$ or even $\Theta(1)$, where $k$ is the maximum number of leaked signals observable to each receiver across leakage patterns in the distribution. En route to these results, we show that subsampling and masking are general-purpose algorithmic paradigms for transforming private persuasion signaling schemes to leakage-robust ones, with minmax optimal loss in the sender's utility.


A Generative AI Technique for Synthesizing a Digital Twin for U.S. Residential Solar Adoption and Generation

Kishore, Aparna, Thorve, Swapna, Marathe, Madhav

arXiv.org Artificial Intelligence

Residential rooftop solar adoption is considered crucial for reducing carbon emissions. The lack of photovoltaic (PV) data at a finer resolution (e.g., household, hourly levels) poses a significant roadblock to informed decision-making. We discuss a novel methodology to generate a highly granular, residential-scale realistic dataset for rooftop solar adoption across the contiguous United States. The data-driven methodology consists of: (i) integrated machine learning models to identify PV adopters, (ii) methods to augment the data using explainable AI techniques to glean insights about key features and their interactions, and (iii) methods to generate household-level hourly solar energy output using an analytical model. The resulting synthetic datasets are validated using real-world data and can serve as a digital twin for modeling downstream tasks. Finally, a policy-based case study utilizing the digital twin for Virginia demonstrated increased rooftop solar adoption with the 30\% Federal Solar Investment Tax Credit, especially in Low-to-Moderate-Income communities.


Enhance Modality Robustness in Text-Centric Multimodal Alignment with Adversarial Prompting

Tsai, Yun-Da, Yen, Ting-Yu, Liao, Keng-Te, Lin, Shou-De

arXiv.org Artificial Intelligence

Converting different modalities into generalized text, which then serves as input prompts for large language models (LLMs), is a common approach for aligning multimodal models, particularly when pairwise data is limited. Text-centric alignment method leverages the unique properties of text as a modality space, transforming diverse inputs into a unified textual representation, thereby enabling downstream models to effectively interpret various modal inputs. This study evaluates the quality and robustness of multimodal representations in the face of noise imperfections, dynamic input order permutations, and missing modalities, revealing that current text-centric alignment methods can compromise downstream robustness. To address this issue, we propose a new text-centric adversarial training approach that significantly enhances robustness compared to traditional robust training methods and pre-trained multimodal foundation models. Our findings underscore the potential of this approach to improve the robustness and adaptability of multimodal representations, offering a promising solution for dynamic and real-world applications.


The Role of Network and Identity in the Diffusion of Hashtags

Ananthasubramaniam, Aparna, Zhu, Yufei, Jurgens, David, Romero, Daniel

arXiv.org Artificial Intelligence

Although the spread of behaviors is influenced by many social factors, existing literature tends to study the effects of single factors -- most often, properties of the social network -- on the final cascade. In order to move towards a more integrated view of cascades, this paper offers the first comprehensive investigation into the role of two social factors in the diffusion of 1,337 popular hashtags representing the production of novel culture on Twitter: 1) the topology of the Twitter social network and 2) performance of each user's probable demographic identity. Here, we show that cascades are best modeled using a combination of network and identity, rather than either factor alone. This combined model best reproduces a composite index of ten cascade properties across all 1,337 hashtags. However, there is important heterogeneity in what social factors are required to reproduce different properties of hashtag cascades. For instance, while a combined network+identity model best predicts the popularity of cascades, a network-only model has better performance in predicting cascade growth and an identity-only model in adopter composition. We are able to predict what type of hashtag is best modeled by each combination of features and use this to further improve performance. Additionally, consistent with prior literature on the combined network+identity model most outperforms the single-factor counterfactuals among hashtags used for expressing racial or regional identity, stance-taking, talking about sports, or variants of existing cultural trends with very slow- or fast-growing communicative need. In sum, our results imply the utility of multi-factor models in predicting cascades, in order to account for the varied ways in which network, identity, and other social factors play a role in the diffusion of hashtags on Twitter.


Enhance the Robustness of Text-Centric Multimodal Alignments

Yen, Ting-Yu, Tsai, Yun-Da, Liao, Keng-Te, Lin, Shou-De

arXiv.org Artificial Intelligence

Converting different modalities into general text, serving as input prompts for large language models (LLMs), is a common method to align multimodal models when there is limited pairwise data. This text-centric approach leverages the unique properties of text as a modality space, transforming diverse inputs into a unified textual representation. This enables downstream models to effectively interpret various modal inputs. This study assesses the quality and robustness of multimodal representations in the presence of missing entries, noise, or absent modalities, revealing that current text-centric alignment methods compromise downstream robustness. To address this issue, we propose a new text-centric approach that achieves superior robustness compared to previous methods across various modalities in different settings. Our findings highlight the potential of this approach to enhance the robustness and adaptability of multimodal representations, offering a promising solution for dynamic and real-world applications.


Google's latest AI tackles long and costly drug discovery

Engadget

It can cost billions of dollars to develop drugs and a large percentage fail at the trial stage, so a number of companies are deploying AI to help in that area. Google's Cloud division is the latest to join that race with two new suites aimed at addressing drug discovery while advancing precision medicine, it announced. The Target and Lead Identification Suite aims to help drug companies better understand proteins and amino acids that are key to drug development. Specifically, it's designed to help scientists identify biological targets that researchers can develop treatments around. This could effectively speed up drug discovery and lower costs. Early adopters for the suite "include multinational pharmaceutical companies like Pfizer and industry-leading biotech companies including Cereval," Google Cloud wrote in a press release.


A new and faster machine learning flywheel for enterprises

#artificialintelligence

This post is a commentary on the MLCommons article "Perspective: Unlocking ML requires an ecosystem approach" by Peter Mattson, Aarush Selvan, David Kanter, Vijay Janapa Reddi, Roger Roberts, and Jacomo Corbo. The world of artificial intelligence (AI) and machine learning (ML) is undergoing a sea change from science to engineering at scale. Over the past decade, the volume of AI research has skyrocketed as the cost to train and deploy commercial models has decreased. Between 2015 and 2021, the cost to train an image classification system fell by 64 percent, while training times improved by 94 percent in the same period.1 The emergence of foundation models--large-scale, deep learning models trained on massive, broad, unstructured data sets--has enabled entrepreneurs and business executives to see the possibility of true scale.


Council Post: 2023 Will Be A Defining Year For AI And The Future Of Work

#artificialintelligence

Cenk Sidar is the cofounder and CEO of Enquire AI, combining AI, data science, and human intelligence to deliver real-time insights. In recent years, tech-celeration has changed the way humans interact in and beyond the workplace. While rapid tech adoption is considered good, it also fuels the emergence of new risks and "unknown unknowns" in an ever-changing macro landscape. As we enter 2023 on the brink of economic strife, something must balance the scales and help business leaders tackle their biggest problems. One answer lies in another tech breakthrough: Artificial intelligence is ready to perform at scale. Its full implementation cannot be predicted at this point, but it promises real-time actionable insights and offers newfound agility in an uncertain world.