Goto

Collaborating Authors

 aspirin


Supplementary Material for " Multi-task Causal Learning with Gaussian Processes "

Neural Information Processing Systems

Eq. (4) gives the causal operator.1.2 The set C represents the smallest set for which Eq. (2) holds. The conditions in Theorem 3.1 allow for full transfer across all intervention functions in This is equivalent to sampling from the mutilated graph. We compute the integrals in Eqs. Finally, we fix the variance in the likelihood of Eq.


This Is the First Time Scientists Have Seen Decisionmaking in a Brain

WIRED

Twelve laboratories around the world have joined forces to map neuronal activity in a mouse's brain as it makes decisions. All products featured on WIRED are independently selected by our editors. However, we may receive compensation from retailers and/or from purchases of products through these links. Neuroscientists from around the world have worked in parallel to map, for the first time, the entire brain activity of mice while they were making decisions. This achievement involved using electrodes inserted inside the brain to simultaneously record the activity of more than half a million neurons distributed across 95 percent of the rodents' brain volume.


Supplementary Material for " Multi-task Causal Learning with Gaussian Processes "

Neural Information Processing Systems

Eq. (4) gives the causal operator.1.2 The set C represents the smallest set for which Eq. (2) holds. The conditions in Theorem 3.1 allow for full transfer across all intervention functions in This is equivalent to sampling from the mutilated graph. We compute the integrals in Eqs. Finally, we fix the variance in the likelihood of Eq.


Clinical trial cohort selection using Large Language Models on n2c2 Challenges

Tai, Chi-en Amy, Tannier, Xavier

arXiv.org Artificial Intelligence

Clinical trials are a critical process in the medical field for introducing new treatments and innovations. However, cohort selection for clinical trials is a time-consuming process that often requires manual review of patient text records for specific keywords. Though there have been studies on standardizing the information across the various platforms, Natural Language Processing (NLP) tools remain crucial for spotting eligibility criteria in textual reports. Recently, pre-trained large language models (LLMs) have gained popularity for various NLP tasks due to their ability to acquire a nuanced understanding of text. In this paper, we study the performance of large language models on clinical trial cohort selection and leverage the n2c2 challenges to benchmark their performance. Our results are promising with regard to the incorporation of LLMs for simple cohort selection tasks, but also highlight the difficulties encountered by these models as soon as fine-grained knowledge and reasoning are required.


Medical Foundation Models are Susceptible to Targeted Misinformation Attacks

Han, Tianyu, Nebelung, Sven, Khader, Firas, Wang, Tianci, Mueller-Franzes, Gustav, Kuhl, Christiane, Försch, Sebastian, Kleesiek, Jens, Haarburger, Christoph, Bressem, Keno K., Kather, Jakob Nikolas, Truhn, Daniel

arXiv.org Artificial Intelligence

Large language models (LLMs) have broad medical knowledge and can reason about medical information across many domains, holding promising potential for diverse medical applications in the near future. In this study, we demonstrate a concerning vulnerability of LLMs in medicine. Through targeted manipulation of just 1.1% of the model's weights, we can deliberately inject an incorrect biomedical fact. The erroneous information is then propagated in the model's output, whilst its performance on other biomedical tasks remains intact. We validate our findings in a set of 1,038 incorrect biomedical facts. This peculiar susceptibility raises serious security and trustworthiness concerns for the application of LLMs in healthcare settings. It accentuates the need for robust protective measures, thorough verification mechanisms, and stringent management of access to these models, ensuring their reliable and safe use in medical practice.


Functional Causal Bayesian Optimization

Gultchin, Limor, Aglietti, Virginia, Bellot, Alexis, Chiappa, Silvia

arXiv.org Artificial Intelligence

We propose functional causal Bayesian optimization (fCBO), a method for finding interventions that optimize a target variable in a known causal graph. fCBO extends the CBO family of methods to enable functional interventions, which set a variable to be a deterministic function of other variables in the graph. fCBO models the unknown objectives with Gaussian processes whose inputs are defined in a reproducing kernel Hilbert space, thus allowing to compute distances among vector-valued functions. In turn, this enables to sequentially select functions to explore by maximizing an expected improvement acquisition functional while keeping the typical computational tractability of standard BO settings. We introduce graphical criteria that establish when considering functional interventions allows attaining better target effects, and conditions under which selected interventions are also optimal for conditional target effects. We demonstrate the benefits of the method in a synthetic and in a real-world causal graph.


Constrained Causal Bayesian Optimization

Aglietti, Virginia, Malek, Alan, Ktena, Ira, Chiappa, Silvia

arXiv.org Artificial Intelligence

We propose constrained causal Bayesian optimization (cCBO), an approach for finding interventions in a known causal graph that optimize a target variable under some constraints. cCBO first reduces the search space by exploiting the graph structure and, if available, an observational dataset; and then solves the restricted optimization problem by modelling target and constraint quantities using Gaussian processes and by sequentially selecting interventions via a constrained expected improvement acquisition function. We propose different surrogate models that enable to integrate observational and interventional data while capturing correlation among effects with increasing levels of sophistication. We evaluate cCBO on artificial and real-world causal graphs showing successful trade off between fast convergence and percentage of feasible interventions.


GWO-FI: A novel machine learning framework by combining Gray Wolf Optimizer and Frequent Itemsets to diagnose and investigate effective factors on In-Hospital Mortality and Length of Stay among Kermanshahian Cardiovascular Disease patients

Yavari, Ali, Janjani, Parisa, Motavaseli, Sayeh, Weysi, Seyran, Siabani, Soraya, Rouzbahani, Mohammad

arXiv.org Artificial Intelligence

Investigation and analysis of patient outcomes, including in-hospital mortality and length of stay, are crucial for assisting clinicians in determining a patient's result at the outset of their hospitalization and for assisting hospitals in allocating their resources. This paper proposes an approach based on combining the well-known gray wolf algorithm with frequent items extracted by association rule mining algorithms. First, original features are combined with the discriminative extracted frequent items. The best subset of these features is then chosen, and the parameters of the used classification algorithms are also adjusted, using the gray wolf algorithm. This framework was evaluated using a real dataset made up of 2816 patients from the Imam Ali Kermanshah Hospital in Iran. The study's findings indicate that low Ejection Fraction, old age, high CPK values, and high Creatinine levels are the main contributors to patients' mortality. Several significant and interesting rules related to mortality in hospitals and length of stay have also been extracted and presented. Additionally, the accuracy, sensitivity, specificity, and auroc of the proposed framework for the diagnosis of mortality in the hospital using the SVM classifier were 0.9961, 0.9477, 0.9992, and 0.9734, respectively. According to the framework's findings, adding frequent items as features considerably improves classification accuracy.


Multi-task Causal Learning with Gaussian Processes

Aglietti, Virginia, Damoulas, Theodoros, Álvarez, Mauricio, González, Javier

arXiv.org Machine Learning

This paper studies the problem of learning the correlation structure of a set of intervention functions defined on the directed acyclic graph (DAG) of a causal model. This is useful when we are interested in jointly learning the causal effects of interventions on different subsets of variables in a DAG, which is common in field such as healthcare or operations research. We propose the first multi-task causal Gaussian process (GP) model, which we call DAG-GP, that allows for information sharing across continuous interventions and across experiments on different variables. DAG-GP accommodates different assumptions in terms of data availability and captures the correlation between functions lying in input spaces of different dimensionality via a well-defined integral operator. We give theoretical results detailing when and how the DAG-GP model can be formulated depending on the DAG. We test both the quality of its predictions and its calibrated uncertainties. Compared to single-task models, DAG-GP achieves the best fitting performance in a variety of real and synthetic settings. In addition, it helps to select optimal interventions faster than competing approaches when used within sequential decision making frameworks, like active learning or Bayesian optimization.


Incremental Intervention Effects in Studies with Many Timepoints, Repeated Outcomes, and Dropout

Kim, Kwangho, Kennedy, Edward H., Naimi, Ashley I.

arXiv.org Machine Learning

Modern longitudinal studies feature data collected at many timepoints, often of the same order of sample size. Such studies are typically affected by dropout and positivity violations. We tackle these problems by generalizing effects of recent incremental interventions (which shift propensity scores rather than set treatment values deterministically) to accommodate multiple outcomes and subject dropout. We give an identifying expression for incremental effects when dropout is conditionally ignorable (without requiring treatment positivity), and derive the nonparametric efficiency bound for estimating such effects. Then we present efficient nonparametric estimators, showing that they converge at fast parametric rates and yield uniform inferential guarantees, even when nuisance functions are estimated flexibly at slower rates. We also study the efficiency of incremental effects relative to more conventional deterministic effects in a novel infinite time horizon setting, where the number of timepoints grows with sample size, and show that incremental effects yield near-exponential gains in this setup. Finally we conclude with simulations and apply our methods in a study of the effect of low-dose aspirin on pregnancy outcomes.