Goto

Collaborating Authors

 Montserrado


Weakly Supervised Object Segmentation by Background Conditional Divergence

Baker, Hassan, Emigh, Matthew S., Brockmeier, Austin J.

arXiv.org Artificial Intelligence

As a computer vision task, automatic object segmentation remains challenging in specialized image domains without massive labeled data, such as synthetic aperture sonar images, remote sensing, biomedical imaging, etc. In any domain, obtaining pixel-wise segmentation masks is expensive. In this work, we propose a method for training a masking network to perform binary object segmentation using weak supervision in the form of image-wise presence or absence of an object of interest, which provides less information but may be obtained more quickly from manual or automatic labeling. A key step in our method is that the segmented objects can be placed into background-only images to create realistic images of the objects with counterfactual backgrounds. To create a contrast between the original and counterfactual background images, we propose to first cluster the background-only images and then, during learning, create counterfactual images that blend objects segmented from their original source backgrounds to backgrounds chosen from a targeted cluster. One term in the training loss is the divergence between these counterfactual images and the real object images with backgrounds of the target cluster. The other term is a supervised loss for background-only images. While an adversarial critic could provide the divergence, we use sample-based divergences. We conduct experiments on side-scan and synthetic aperture sonar in which our approach succeeds compared to previous unsupervised segmentation baselines that were only tested on natural images. Furthermore, to show generality we extend our experiments to natural images, obtaining reasonable performance with our method that avoids pretrained networks, generative networks, and adversarial critics. The code for this work can be found at \href{GitHub}{https://github.com/bakerhassan/WSOS}.


MAGIC: A Multi-Hop and Graph-Based Benchmark for Inter-Context Conflicts in Retrieval-Augmented Generation

Lee, Jungyeon, Lee, Kangmin, Kim, Taeuk

arXiv.org Artificial Intelligence

Knowledge conflict often arises in retrieval-augmented generation (RAG) systems, where retrieved documents may be inconsistent with one another or contradict the model's parametric knowledge. Existing benchmarks for investigating the phenomenon have notable limitations, including a narrow focus on the question answering setup, heavy reliance on entity substitution techniques, and a restricted range of conflict types. To address these issues, we propose a knowledge graph (KG)-based framework that generates varied and subtle conflicts between two similar yet distinct contexts, while ensuring interpretability through the explicit relational structure of KGs. Experimental results on our benchmark, MAGIC, provide intriguing insights into the inner workings of LLMs regarding knowledge conflict: both open-source and proprietary models struggle with conflict detection -- especially when multi-hop reasoning is required -- and often fail to pinpoint the exact source of contradictions. Finally, we present in-depth analyses that serve as a foundation for improving LLMs in integrating diverse, sometimes even conflicting, information.


ShipwreckFinder: A QGIS Tool for Shipwreck Detection in Multibeam Sonar Data

Sheppard, Anja, Smithline, Tyler, Scheffer, Andrew, Smith, David, Sethuraman, Advaith V., Bird, Ryan, Lin, Sabrina, Skinner, Katherine A.

arXiv.org Artificial Intelligence

Abstract--In this paper, we introduce ShipwreckFinder, an open-source QGIS plugin that detects shipwrecks from multi-beam sonar data. Shipwrecks are an important historical marker of maritime history, and can be discovered through manual inspection of bathymetric data. However, this is a time-consuming process and often requires expert analysis. Our proposed tool allows users to automatically preprocess bathymetry data, perform deep learning inference, threshold model outputs, and produce either pixel-wise segmentation masks or bounding boxes of predicted shipwrecks. The backbone of this open-source tool is a deep learning model, which is trained on a variety of shipwreck data from the Great Lakes and the coasts of Ireland. Additionally, we employ synthetic data generation in order to increase the size and diversity of our dataset. We demonstrate superior segmentation performance with our open-source tool and training pipeline as compared to a deep learning-based ArcGIS toolkit and a more classical inverse sinkhole detection method.


Language Models That Walk the Talk: A Framework for Formal Fairness Certificates

Chen, Danqing, Ladner, Tobias, Mhadhbi, Ahmed Rayen, Althoff, Matthias

arXiv.org Artificial Intelligence

As large language models become integral to high-stakes applications, ensuring their robustness and fairness is critical. Despite their success, large language models remain vulnerable to adversarial attacks, where small perturbations, such as synonym substitutions, can alter model predictions, posing risks in fairness-critical areas, such as gender bias mitigation, and safety-critical areas, such as toxicity detection. While formal verification has been explored for neural networks, its application to large language models remains limited. This work presents a holistic verification framework to certify the robustness of transformer-based language models, with a focus on ensuring gender fairness and consistent outputs across different gender-related terms. Furthermore, we extend this methodology to toxicity detection, offering formal guarantees that adversarially manipulated toxic inputs are consistently detected and appropriately censored, thereby ensuring the reliability of moderation systems. By formalizing robustness within the embedding space, this work strengthens the reliability of language models in ethical AI deployment and content moderation.


Learning to Trust Your Feelings: Leveraging Self-awareness in LLMs for Hallucination Mitigation

Liang, Yuxin, Song, Zhuoyang, Wang, Hao, Zhang, Jiaxing

arXiv.org Artificial Intelligence

We evaluate the ability of Large Language Models (LLMs) to discern and express their internal knowledge state, a key factor in countering factual hallucination and ensuring reliable application of LLMs. We observe a robust self-awareness of internal knowledge state in LLMs, evidenced by over 85% accuracy in knowledge probing. However, LLMs often fail to express their internal knowledge during generation, leading to factual hallucinations. We develop an automated hallucination annotation tool, Dreamcatcher, which merges knowledge probing and consistency checking methods to rank factual preference data. Using knowledge preference as reward, We propose a Reinforcement Learning from Knowledge Feedback (RLKF) training framework, leveraging reinforcement learning to enhance the factuality and honesty of LLMs. Our experiments across multiple models show that RLKF training effectively enhances the ability of models to utilize their internal knowledge state, boosting performance in a variety of knowledge-based and honesty-related tasks.


Flickr Africa: Examining Geo-Diversity in Large-Scale, Human-Centric Visual Data

Naggita, Keziah, LaChance, Julienne, Xiang, Alice

arXiv.org Artificial Intelligence

Biases in large-scale image datasets are known to influence the performance of computer vision models as a function of geographic context. To investigate the limitations of standard Internet data collection methods in low- and middle-income countries, we analyze human-centric image geo-diversity on a massive scale using geotagged Flickr images associated with each nation in Africa. We report the quantity and content of available data with comparisons to population-matched nations in Europe as well as the distribution of data according to fine-grained intra-national wealth estimates. Temporal analyses are performed at two-year intervals to expose emerging data trends. Furthermore, we present findings for an ``othering'' phenomenon as evidenced by a substantial number of images from Africa being taken by non-local photographers. The results of our study suggest that further work is required to capture image data representative of African people and their environments and, ultimately, to improve the applicability of computer vision models in a global context.


Decadal Temperature Prediction via Chaotic Behavior Tracking

Ren, Jinfu, Liu, Yang, Liu, Jiming

arXiv.org Artificial Intelligence

Decadal temperature prediction provides crucial information for quantifying the expected effects of future climate changes and thus informs strategic planning and decision-making in various domains. However, such long-term predictions are extremely challenging, due to the chaotic nature of temperature variations. Moreover, the usefulness of existing simulation-based and machine learning-based methods for this task is limited because initial simulation or prediction errors increase exponentially over time. To address this challenging task, we devise a novel prediction method involving an information tracking mechanism that aims to track and adapt to changes in temperature dynamics during the prediction phase by providing probabilistic feedback on the prediction error of the next step based on the current prediction. We integrate this information tracking mechanism, which can be considered as a model calibrator, into the objective function of our method to obtain the corrections needed to avoid error accumulation. Our results show the ability of our method to accurately predict global land-surface temperatures over a decadal range. Furthermore, we demonstrate that our results are meaningful in a real-world context: the temperatures predicted using our method are consistent with and can be used to explain the well-known teleconnections within and between different continents.


Can robots impact our health? One study says so

#artificialintelligence

A growing number of Americans are seeing their job security erode in the face of automation and it's undermining their health, according to a new study. The report, conducted by three Ball State University researchers with the school's Center for Business and Economic Research, shows that a 10 percentage point increase in automation risk increases average per-county costs related to medical expenses and lost productivity. "People who live and work in areas where automation is taking place are sickened by the thought of losing their jobs and having no way of providing for themselves or their families," said Michael Hicks, the center's director, who helped conduct the research. Costs associated with an increase in poor or fair health rise by $24 million to $174 million, costs related to increased physical distress rise by $6 million to $40 million and costs linked to mental distress increase by $7 million to $47 million. "This should give us pause about thinking through the benefits and costs of these technologies," Hicks said.


OpenEarthMap: A Benchmark Dataset for Global High-Resolution Land Cover Mapping

Xia, Junshi, Yokoya, Naoto, Adriano, Bruno, Broni-Bediako, Clifford

arXiv.org Artificial Intelligence

We introduce OpenEarthMap, a benchmark dataset, for global high-resolution land cover mapping. OpenEarthMap consists of 2.2 million segments of 5000 aerial and satellite images covering 97 regions from 44 countries across 6 continents, with manually annotated 8-class land cover labels at a 0.25--0.5m ground sampling distance. Semantic segmentation models trained on the OpenEarthMap generalize worldwide and can be used as off-the-shelf models in a variety of applications. We evaluate the performance of state-of-the-art methods for unsupervised domain adaptation and present challenging problem settings suitable for further technical development. We also investigate lightweight models using automated neural architecture search for limited computational resources and fast mapping. The dataset is available at https://open-earth-map.org.


Apple computer built by Wozniak and Jobs fetches $500,000 at Southern California auction

Los Angeles Times

A piece of computer history and coveted collector's item with ties to Southern California fetched six figures at auction this week. An Apple-1 computer, hand-built by Steve Wozniak and Steve Jobs in the 1970s, sold for $500,000 at auction Tuesday in Monrovia. The final bid for the unit was $400,000, with the buyer -- who wishes to remain anonymous -- paying an additional $100,000 premium, or commission, to John Moran Auctioneers. The Southern California-based auction house estimated that the unit, dubbed the "Chaffey College Apple-1" after its original owner was identified as a Chaffey professor, would sell for between $400,000 to $600,000. In 2014, Bonhams auction house sold an Apple-1 for more than $900,000.