Goto

Collaborating Authors

 reference database


RadiomicsRetrieval: A Customizable Framework for Medical Image Retrieval Using Radiomics Features

Na, Inye, Rue, Nejung, Chung, Jiwon, Park, Hyunjin

arXiv.org Artificial Intelligence

Medical image retrieval is a valuable field for supporting clinical decision-making, yet current methods primarily support 2D images and require fully annotated queries, limiting clinical flexibility. To address this, we propose RadiomicsRetrieval, a 3D content-based retrieval framework bridging handcrafted radiomics descriptors with deep learning-based embeddings at the tumor level . Unlike existing 2D approaches, RadiomicsRetrieval fully exploits volumetric data to leverage richer spatial context in medical images. We employ a promptable segmentation model (e.g., SAM) to derive tumor-specific image embeddings, which are aligned with radiomics features extracted from the same tumor via contrastive learning. These representations are further enriched by anatomical positional embedding (APE). As a result, RadiomicsRe-trieval enables flexible querying based on shape, location, or partial feature sets. Extensive experiments on both lung CT and brain MRI public datasets demonstrate that radiomics features significantly enhance retrieval specificity, while APE provides global anatomical context essential for location-based searches. Notably, our framework requires only minimal user prompts (e.g., a single point), minimizing segmentation overhead and supporting diverse clinical scenarios. The capability to query using either image embeddings or selected radiomics attributes highlights its adaptability, potentially benefiting diagnosis, treatment planning, and research on large-scale medical imaging repositories.


Nearest Neighbor Normalization Improves Multimodal Retrieval

Chowdhury, Neil, Wang, Franklin, Shenoy, Sumedh, Kiela, Douwe, Schwettmann, Sarah, Thrush, Tristan

arXiv.org Artificial Intelligence

Multimodal models leverage large-scale pre-training to achieve strong but still imperfect performance on tasks such as image captioning, visual question answering, and cross-modal retrieval. In this paper, we present a simple and efficient method for correcting errors in trained contrastive image-text retrieval models with no additional training, called Nearest Neighbor Normalization (NNN). We show an improvement on retrieval metrics in both text retrieval and image retrieval for all of the contrastive models that we tested (CLIP, BLIP, ALBEF, SigLIP, BEiT) and for both of the datasets that we used (MS-COCO and Flickr30k). NNN requires a reference database, but does not require any training on this database, and can even increase the retrieval accuracy of a model after finetuning.


SoK: Anti-Facial Recognition Technology

Wenger, Emily, Shan, Shawn, Zheng, Haitao, Zhao, Ben Y.

arXiv.org Artificial Intelligence

The rapid adoption of facial recognition (FR) technology by both government and commercial entities in recent years has raised concerns about civil liberties and privacy. In response, a broad suite of so-called "anti-facial recognition" (AFR) tools has been developed to help users avoid unwanted facial recognition. The set of AFR tools proposed in the last few years is wide-ranging and rapidly evolving, necessitating a step back to consider the broader design space of AFR systems and long-term challenges. This paper aims to fill that gap and provides the first comprehensive analysis of the AFR research landscape. Using the operational stages of FR systems as a starting point, we create a systematic framework for analyzing the benefits and tradeoffs of different AFR approaches. We then consider both technical and social challenges facing AFR tools and propose directions for future research in this field.


Location retrieval using visible landmarks based qualitative place signatures

Wei, Lijun, Gouet-Brunet, Valerie, Cohn, Anthony

arXiv.org Artificial Intelligence

Location retrieval based on visual information is to retrieve the location of an agent (e.g. human, robot) or the area they see by comparing the observations with a certain form of representation of the environment. Existing methods generally require precise measurement and storage of the observed environment features, which may not always be robust due to the change of season, viewpoint, occlusion, etc. They are also challenging to scale up and may not be applicable for humans due to the lack of measuring/imaging devices. Considering that humans often use less precise but easily produced qualitative spatial language and high-level semantic landmarks when describing an environment, a qualitative location retrieval method is proposed in this work by describing locations/places using qualitative place signatures (QPS), defined as the perceived spatial relations between ordered pairs of co-visible landmarks from viewers' perspective. After dividing the space into place cells each with individual signatures attached, a coarse-to-fine location retrieval method is proposed to efficiently identify the possible location(s) of viewers based on their qualitative observations. The usability and effectiveness of the proposed method were evaluated using openly available landmark datasets, together with simulated observations by considering the possible perception error.


Artificial Intelligence to accelerate the fight against cancer - Actu IA

#artificialintelligence

Health Data Club (HDH) and Unicancer signed a partnership agreement in July 2021 to jointly build the Unibase program, with the goal of creating a working environment for processing health data through innovative analytical approaches. At the end of November, Unicancer and the HDH unveiled this program and launched the first Unibase call for expressions of interest (AMI) to accelerate research, based on real-life data, in oncology. In France, cancer is the leading cause of death in men and the second leading cause of death in women, with more than 380,000 cases reported each year. Medical and scientific progress has led to huge advances in the fight against this disease. Unicancer and the Cancer Research Centers (CLCC) are among the major players in recent developments in cancer research.


Studying Complex Phosphorus Systems with Machine Learning

#artificialintelligence

Machine learning and other artificial intelligence (AI) algorithms are becoming more commonplace in modern-day society. They are starting to become a very valuable tool for chemical research--at both the fundamental research and industrial-scale optimisation levels. This is primarily due to the rise in computational chemistry methods which use simulations and advanced numerical algorithms to predict best how molecules will behave (and how they will look structurally in the case of complex systems). While research is going into a lot of different chemicals, the various allotropes of elemental phosphorus (i.e. Still, it is relatively hard to simulate using conventional computational methods compared to other elements and molecules.


rOpenSci Introducing the 2018 rOpenSci Research Fellows!

@machinelearnbot

Since our inception, one of the mechanisms through which we have supported the community is by developing high-quality open source tools that lower barriers to working with scientific data. Equally important to our mission is to build capacity and promote researchers who are engaged in such practices within their disciplinary communities. This fellowship program is a unique opportunity for us to enable such individuals to have a bigger voice in their communities. This year, a diverse committee (comprised of Di Cook, Mine Cetinkaya-Rundel, Matt Jones, Ken Benoit and myself) reviewed 64 applications from researchers in various disciplines to select four winners. Emerging from this impressive pool of applicants are four outstanding researchers who constitute the 2018 Fellows.


Large-scale Machine Learning for Metagenomics Sequence Classification

Vervier, Kévin, Mahé, Pierre, Tournoud, Maud, Veyrieras, Jean-Baptiste, Vert, Jean-Philippe

arXiv.org Machine Learning

Metagenomics characterizes the taxonomic diversity of microbial communities by sequencing DNA directly from an environmental sample. One of the main challenges in metagenomics data analysis is the binning step, where each sequenced read is assigned to a taxonomic clade. Due to the large volume of metagenomics datasets, binning methods need fast and accurate algorithms that can operate with reasonable computing requirements. While standard alignment-based methods provide state-of-the-art performance, compositional approaches that assign a taxonomic class to a DNA read based on the k-mers it contains have the potential to provide faster solutions. In this work, we investigate the potential of modern, large-scale machine learning implementations for taxonomic affectation of next-generation sequencing reads based on their k-mers profile. We show that machine learning-based compositional approaches benefit from increasing the number of fragments sampled from reference genome to tune their parameters, up to a coverage of about 10, and from increasing the k-mer size to about 12. Tuning these models involves training a machine learning model on about 10 8 samples in 10 7 dimensions, which is out of reach of standard soft-wares but can be done efficiently with modern implementations for large-scale machine learning. The resulting models are competitive in terms of accuracy with well-established alignment tools for problems involving a small to moderate number of candidate species, and for reasonable amounts of sequencing errors. We show, however, that compositional approaches are still limited in their ability to deal with problems involving a greater number of species, and more sensitive to sequencing errors. We finally confirm that compositional approach achieve faster prediction times, with a gain of 3 to 15 times with respect to the BWA-MEM short read mapper, depending on the number of candidate species and the level of sequencing noise.


Pattern Matching for Self- Tuning of MapReduce Jobs

Rizvandi, Nikzad Babaii, Taheri, Javid, Zomaya, Albert Y.

arXiv.org Artificial Intelligence

In this paper, we study CPU utilization time patterns of several MapReduce applications. After extracting running patterns of several applications, they are saved in a reference database to be later used to tweak system parameters to efficiently execute unknown applications in future. To achieve this goal, CPU utilization patterns of new applications are compared with the already known ones in the reference database to find/predict their most probable execution patterns. Because of different patterns lengths, the Dynamic Time Warping (DTW) is utilized for such comparison; a correlation analysis is then applied to DTWs outcomes to produce feasible similarity patterns. Three real applications (WordCount, Exim Mainlog parsing and Terasort) are used to evaluate our hypothesis in tweaking system parameters in executing similar applications. Results were very promising and showed effectiveness of our approach on pseudo-distributed MapReduce platforms.


Improving accuracy and power with transfer learning using a meta-analytic database

Schwartz, Yannick, Varoquaux, Gaël, Pallier, Christophe, Pinel, Philippe, Poline, Jean-Baptiste, Thirion, Bertrand

arXiv.org Machine Learning

Typical cohorts in brain imaging studies are not large enough for systematic testing of all the information contained in the images. To build testable working hypotheses, investigators thus rely on analysis of previous work, sometimes formalized in a so-called meta-analysis. In brain imaging, this approach underlies the specification of regions of interest (ROIs) that are usually selected on the basis of the coordinates of previously detected effects. In this paper, we propose to use a database of images, rather than coordinates, and frame the problem as transfer learning: learning a discriminant model on a reference task to apply it to a different but related new task. To facilitate statistical analysis of small cohorts, we use a sparse discriminant model that selects predictive voxels on the reference task and thus provides a principled procedure to define ROIs. The benefits of our approach are twofold. First it uses the reference database for prediction, i.e. to provide potential biomarkers in a clinical setting. Second it increases statistical power on the new task. We demonstrate on a set of 18 pairs of functional MRI experimental conditions that our approach gives good prediction. In addition, on a specific transfer situation involving different scanners at different locations, we show that voxel selection based on transfer learning leads to higher detection power on small cohorts.