Goto

Collaborating Authors

 exposure



Hair samples reveal the benefits of lead regulation

Popular Science

Before the EPA, Utah saw 100 times more lead exposure. Breakthroughs, discoveries, and DIY tips sent six days a week. The evidence is clear--and in your hair. Americans were exposed to as much as 100 times more lead in their daily lives than they are today before the Environmental Protection Agency was established in 1970. In an effort to examine the dramatic reduction in toxic heavy metal exposure, researchers turned to human hair samples dating back a century.


Toxic 'forever chemicals' linked to cancer now associated with major pregnancy complication

Daily Mail - Science & tech

Senator accused of steamy affair with her bodyguard in bombshell lawsuit from his WIFE: 'Bring MDMA so I can guide you' Socialite who accused playboy twins of sex attack at Hamptons'castle' is found dead in unexplained circumstances Amy Schumer's friends reveal true meaning of thin bikini pictures and why they're'monitoring her'... as depth of ex Chris Fischer's heartbreak is laid bare Hunter Biden's stripper baby mama asks for him to be ARRESTED over claims he is still failing to pay her child support Ellen Greenberg's fiancé Sam Goldberg breaks cover as feds reopen probe into her'suicide'... and late teacher's mother shares incredible sign sent from beyond the grave Nicole Richie addresses her daughter's new identity after unveiling transformation on her 18th birthday '90s Vogue model Niki Taylor looks amazing as she sizzles at age 50 for new campaign Karoline Leavitt reveals the thinking behind Trump's call to cancel elections Family of Tyler Robinson's transgender lover speaks ...


Exploring Cumulative Effects in Survival Data Using Deep Learning Networks

Yang, Kang-Chung, Yuan, Shinsheng

arXiv.org Machine Learning

In epidemiological research, modeling the cumulative effects of time-dependent exposures on survival outcomes presents a challenge due to their intricate temporal dynamics. Conventional spline-based statistical methods, though effective, require repeated data transformation for each spline parameter tuning, with survival analysis computations relying on the entire dataset, posing difficulties for large datasets. Meanwhile, existing neural network-based survival analysis methods focus on accuracy but often overlook the interpretability of cumulative exposure patterns. To bridge this gap, we introduce CENNSurv, a novel deep learning approach that captures dynamic risk relationships from time-dependent data. Evaluated on two diverse real-world datasets, CENNSurv revealed a multi-year lagged association between chronic environmental exposure and a critical survival outcome, as well as a critical short-term behavioral shift prior to subscription lapse. This demonstrates CENNSurv's ability to model complex temporal patterns with improved scalability. CENNSurv provides researchers studying cumulative effects a practical tool with interpretable insights.


Estimating Propensity for Causality-based Recommendation without Exposure Data

Neural Information Processing Systems

Causality-based recommendation systems focus on the causal effects of user-item interactions resulting from item exposure (i.e., which items are recommended or exposed to the user), as opposed to conventional correlation-based recommendation. They are gaining popularity due to their multi-sided benefits to users, sellers and platforms alike. However, existing causality-based recommendation methods require additional input in the form of exposure data and/or propensity scores (i.e., the probability of exposure) for training. Such data, crucial for modeling causality in recommendation, are often not available in real-world situations due to technical or privacy constraints. In this paper, we bridge the gap by proposing a new framework, called Propensity Estimation for Causality-based Recommendation (PropCare). It can estimate the propensity and exposure from a more practical setup, where only interaction data are available any ground truth on exposure or propensity in training and inference. We demonstrate that, by relating the pairwise characteristics between propensity and item popularity, PropCare enables competitive causality-based recommendation given only the conventional interaction data. We further present a theoretical analysis on the bias of the causal effect under our model estimation. Finally, we empirically evaluate PropCare through both quantitative and qualitative experiments.


Indoor Air Quality Dataset with Activities of Daily Living in Low to Middle-income Communities

Neural Information Processing Systems

In recent years, indoor air pollution has posed a significant threat to our society, claiming over 3.2 million lives annually. Developing nations, such as India, are most affected since lack of knowledge, inadequate regulation, and outdoor air pollution lead to severe daily exposure to pollutants. However, only a limited number of studies have attempted to understand how indoor air pollution affects developing countries like India. To address this gap, we present spatiotemporal measurements of air quality from 30 indoor sites over six months during summer and winter seasons. The sites are geographically located across four regions of type: rural, suburban, and urban, covering the typical low to middle-income population in India. The dataset contains various types of indoor environments (e.g., studio apartments, classrooms, research laboratories, food canteens, and residential households), and can provide the basis for data-driven learning model research aimed at coping with unique pollution patterns in developing countries. This unique dataset demands advanced data cleaning and imputation techniques for handling missing data due to power failure or network outages during data collection. Furthermore, through a simple speech-to-text application, we provide real-time indoor activity labels annotated by occupants. Therefore, environmentalists and ML enthusiasts can utilize this dataset to understand the complex patterns of the pollutants under different indoor activities, identify recurring sources of pollution, forecast exposure, improve floor plans and room structures of modern indoor designs, develop pollution-aware recommender systems, etc.


DeepExposure: Learning to Expose Photos with Asynchronously Reinforced Adversarial Learning

Neural Information Processing Systems

The accurate exposure is the key of capturing high-quality photos in computational photography, especially for mobile phones that are limited by sizes of camera modules. Inspired by luminosity masks usually applied by professional photographers, in this paper, we develop a novel algorithm for learning local exposures with deep reinforcement adversarial learning. To be specific, we segment an image into sub-images that can reflect variations of dynamic range exposures according to raw low-level features. Based on these sub-images, a local exposure for each sub-image is automatically learned by virtue of policy network sequentially while the reward of learning is globally designed for striking a balance of overall exposures. The aesthetic evaluation function is approximated by discriminator in generative adversarial networks. The reinforcement learning and the adversarial learning are trained collaboratively by asynchronous deterministic policy gradient and generative loss approximation. To further simply the algorithmic architecture, we also prove the feasibility of leveraging the discriminator as the value function. Further more, we employ each local exposure to retouch the raw input image respectively, thus delivering multiple retouched images under different exposures which are fused with exposure blending. The extensive experiments verify that our algorithms are superior to state-of-the-art methods in terms of quantitative accuracy and visual illustration.


Scientists Thought Parkinson's Was in Our Genes. It Might Be in the Water

WIRED

Scientists Thought Parkinson's Was in Our Genes. New ideas about chronic illness could revolutionize treatment, if we take the research seriously. Amy Lindberg spent 26 years in the Navy and she still walked like it--with intention, like her chin had someplace to be. But around 2017, her right foot stopped following orders. Lindberg and her husband Brad were five years into their retirement. After moving 10 times for Uncle Sam, they'd bought their dream house near the North Carolina coast. They had a backyard that spilled out onto wetlands. From the kitchen, you could see cranes hunting. They kept bees and played pickleball and watched their children grow. But now Lindberg's right foot was out of rhythm. She worked hard to ignore it, but she couldn't disregard the tremors.


Overcoming the Generalization Limits of SLM Finetuning for Shape-Based Extraction of Datatype and Object Properties

Ringwald, Célian, Gandon, Fabien, Faron, Catherine, Michel, Franck, Akl, Hanna Abi

arXiv.org Artificial Intelligence

Small language models (SLMs) have shown promises for relation extraction (RE) when extracting RDF triples guided by SHACL shapes focused on common datatype properties. This paper investigates how SLMs handle both datatype and object properties for a complete RDF graph extraction. We show that the key bottleneck is related to long-tail distribution of rare properties. To solve this issue, we evaluate several strategies: stratified sampling, weighted loss, dataset scaling, and template-based synthetic data augmentation. We show that the best strategy to perform equally well over unbalanced target properties is to build a training set where the number of occurrences of each property exceeds a given threshold. To enable reproducibility, we publicly released our datasets, experimental results and code. Our findings offer practical guidance for training shape-aware SLMs and highlight promising directions for future work in semantic RE.


On the Interplay of Pre-Training, Mid-Training, and RL on Reasoning Language Models

Zhang, Charlie, Neubig, Graham, Yue, Xiang

arXiv.org Artificial Intelligence

Recent reinforcement learning (RL) techniques have yielded impressive reasoning improvements in language models, yet it remains unclear whether post-training truly extends a model's reasoning ability beyond what it acquires during pre-training. A central challenge is the lack of control in modern training pipelines: large-scale pre-training corpora are opaque, mid-training is often underexamined, and RL objectives interact with unknown prior knowledge in complex ways. To resolve this ambiguity, we develop a fully controlled experimental framework that isolates the causal contributions of pre-training, mid-training, and RL-based post-training. Our approach employs synthetic reasoning tasks with explicit atomic operations, parseable step-by-step reasoning traces, and systematic manipulation of training distributions. We evaluate models along two axes: extrapolative generalization to more complex compositions and contextual generalization across surface contexts. Using this framework, we reconcile competing views on RL's effectiveness. We show that: 1) RL produces true capability gains (pass@128) only when pre-training leaves sufficient headroom and when RL data target the model's edge of competence, tasks at the boundary that are difficult but not yet out of reach. 2) Contextual generalization requires minimal yet sufficient pre-training exposure, after which RL can reliably transfer. 3) Mid-training significantly enhances performance under fixed compute compared with RL only, demonstrating its central but underexplored role in training pipelines. 4) Process-level rewards reduce reward hacking and improve reasoning fidelity. Together, these results clarify the interplay between pre-training, mid-training, and RL, offering a foundation for understanding and improving reasoning LM training strategies.