Goto

Collaborating Authors

 parcel


Monitoring digestate application on agricultural crops using Sentinel-2 Satellite imagery

Kalogeras, Andreas, Bormpoudakis, Dimitrios, Tsardanidis, Iason, Loka, Dimitra A., Kontoes, Charalampos

arXiv.org Artificial Intelligence

Abstract--The widespread use of Exogenous Organic Matter in agriculture necessitates monitoring to assess its effects on soil and crop health. This study evaluates optical Sentinel-2 satellite imagery for detecting digestate application, a practice that enhances soil fertility but poses environmental risks like mi-croplastic contamination and nitrogen losses. In the first instance, Sentinel-2 satellite image time series (SITS) analysis of specific indices (EOMI, NDVI, EVI) was used to characterize EOM's spectral behavior after application on the soils of four different crop types in Thessaly, Greece. Furthermore, Machine Learning (ML) models (namely Random Forest, k-NN, Gradient Boosting and a Feed-Forward Neural Network), were used to investigate digestate presence detection, achieving F1-scores up to 0.85. Agricultural systems can benefit from the application of Exogenous Organic Matter (EOM), which not only enhances soil fertility but also supports waste recycling and promotes circular economies [1], [2].



Will AI mean the end of call centres?

BBC News

Will AI mean the end of call centres? Ask ChatGPT whether AI will replace humans in the customer service industry, and it will offer a diplomatic answer, the summary of which is they will work side by side. Humans though, are not so optimistic. Last year, the chief executive of Indian technology firm Tata Consultancy Services, K Krithivasan, told the Financial Times that AI may soon mean that there is minimal need for call centres in Asia. Meanwhile, AI will autonomously resolve 80% of common customer service issues by 2029, predicts business and technology research firm Gartner.


Energy Approach from $\varepsilon$-Graph to Continuum Diffusion Model with Connectivity Functional

Yang, Yahong, Lee, Sun, Calder, Jeff, Hao, Wenrui

arXiv.org Machine Learning

We derive an energy-based continuum limit for $\varepsilon$-graphs endowed with a general connectivity functional. We prove that the discrete energy and its continuum counterpart differ by at most $O(\varepsilon)$; the prefactor involves only the $W^{1,1}$-norm of the connectivity density as $\varepsilon\to0$, so the error bound remains valid even when that density has strong local fluctuations. As an application, we introduce a neural-network procedure that reconstructs the connectivity density from edge-weight data and then embeds the resulting continuum model into a brain-dynamics framework. In this setting, the usual constant diffusion coefficient is replaced by the spatially varying coefficient produced by the learned density, yielding dynamics that differ significantly from those obtained with conventional constant-diffusion models.



Stacked Regression using Off-the-shelf, Stimulus-tuned and Fine-tuned Neural Networks for Predicting fMRI Brain Responses to Movies (Algonauts 2025 Report)

Scholz, Robert, Bagga, Kunal, Ahrends, Christine, Barbano, Carlo Alberto

arXiv.org Artificial Intelligence

Encoding models predict brain responses to a set of given stimuli. Recently, deep neural networks have been used as encoding models to predict brain activity as recorded by functional MRI (fMRI) [1, 2, 3, 4, 5, 6]. These studies investigate whether representations in deep neural networks correspond to those in the human brain. This relationship is often assessed using linear models, with successful prediction taken as evidence of shared representational structure. Studies have investigated representations from both unimodal and multimodal deep neural networks, including large language models (LLMs) [2, 4, 7, 8], vision models [9, 10], audio models [1, 11], and video-language models (VLMs) [12], to predict brain activity. However, existing studies face challenges in generalizability and comparability. Differences in stimulus modality, quantity, and content, as well as in preprocessing and scoring, make cross-study comparisons difficult. The Algonauts 2025 Challenge [13] provides a framework to address these issues, offering an openly available, preprocessed dataset with a large amount of data per subject and aligned stimuli across modalities, including video, audio, and transcripts, along with a standardized evaluation procedure. The challenge places particular emphasis on generalizability, including both in-distribution and out-of-distribution test sets to rigorously evaluate how well models transfer to new stimuli. 1


Coordinated Multi-Drone Last-mile Delivery: Learning Strategies for Energy-aware and Timely Operations

Qin, Chuhao, Narayanan, Arun, Pournaras, Evangelos

arXiv.org Artificial Intelligence

Abstract--Drones have recently emerged as a faster, safer, and cost-efficient way for last-mile deliveries of parcels, particularly for urgent medical deliveries highlighted during the pandemic. This paper addresses a new challenge of multi-parcel delivery with a swarm of energy-aware drones, accounting for time-sensitive customer requirements. Each drone plans an optimal multi-parcel route within its battery-restricted flight range to minimize delivery delays and reduce energy consumption. The problem is tackled by decomposing it into three sub-problems: (1) optimizing depot locations and service areas using K-means clustering; (2) determining the optimal flight range for drones through reinforcement learning; and (3) planning and selecting multi-parcel delivery routes via a new optimized plan selection approach. T o integrate these solutions and enhance long-term efficiency, we propose a novel algorithm leveraging actor-critic-based multi-agent deep reinforcement learning. Extensive experimentation using realistic delivery datasets demonstrate an exceptional performance of the proposed algorithm. We provide new insights into economic efficiency (minimize energy consumption), rapid operations (reduce delivery delays and overall execution time), and strategic guidance on depot deployment for practical logistics applications. Unmanned aerial vehicles (UA Vs), commonly known as drones, have gained significant attention as a solution for last-mile delivery, especially in recent years [1]. For instance, the COVID-19 pandemic has highlighted the vulnerabilities of traditional delivery methods, as deliverymen risk spreading the virus. This was particularly problematic in quarantine zones, where customers faced difficulties in accessing logistics services [2], [3]. In contrast, drones offer a safer and more flexible alternative. Due to their high mobility, carrying capacity, and accurate GPS navigation, drones are able to deliver parcels directly to small places such as doorways and balconies, avoiding human contact and traffic congestion.


Your next parcel could be delivered by a robot DOG: Major UK courier service starts using four-legged bots for deliveries

Daily Mail - Science & tech

It might not be able to fetch the paper for you, but a robot dog might soon bring you your parcels. Milo, the four-legged delivery bot, has started taking to the streets of Yorkshire as part of a new trial for delivery firm Evri. The robot dog has been trained to jump in and out of the van, navigate to customers' doors, and drop off packages without any assistance. Milo will be joining Evri's regular drivers over the next fortnight as they make their rounds in Morley, Leeds. Evri hopes that these robot co-pilots will take the strain off their human counterparts, freeing up more time for complex jobs like parking or navigating.


TRIBE: TRImodal Brain Encoder for whole-brain fMRI response prediction

d'Ascoli, Stéphane, Rapin, Jérémy, Benchetrit, Yohann, Banville, Hubert, King, Jean-Rémi

arXiv.org Artificial Intelligence

Historically, neuroscience has progressed by fragmenting into specialized domains, each focusing on isolated modalities, tasks, or brain regions. While fruitful, this approach hinders the development of a unified model of cognition. Here, we introduce TRIBE, the first deep neural network trained to predict brain responses to stimuli across multiple modalities, cortical areas and individuals. By combining the pretrained representations of text, audio and video foundational models and handling their time-evolving nature with a transformer, our model can precisely model the spatial and temporal fMRI responses to videos, achieving the first place in the Algonauts 2025 brain encoding competition with a significant margin over competitors. Ablations show that while unimodal models can reliably predict their corresponding cortical networks (e.g. visual or auditory networks), they are systematically outperformed by our multimodal model in high-level associative cortices. Currently applied to perception and comprehension, our approach paves the way towards building an integrative model of representations in the human brain. Our code is available at https://github.com/facebookresearch/algonauts-2025.


VIBE: Video-Input Brain Encoder for fMRI Response Modeling

Schad, Daniel Carlström, Dixit, Shrey, Keck, Janis, Studenyak, Viktor, Shpilevoi, Aleksandr, Bicanski, Andrej

arXiv.org Artificial Intelligence

We present VIBE, a two-stage Transformer that fuses multi-modal video, audio, and text features to predict fMRI activity. Representations from open-source models (Qwen2.5, BEATs, Whisper, SlowFast, V-JEPA) are merged by a modality-fusion transformer and temporally decoded by a prediction transformer with rotary embeddings. Trained on 65 hours of movie data from the CNeuroMod dataset and ensembled across 20 seeds, VIBE attains mean parcel-wise Pearson correlations of 0.3225 on in-distribution Friends S07 and 0.2125 on six out-of-distribution films. An earlier iteration of the same architecture obtained 0.3198 and 0.2096, respectively, winning Phase-1 and placing second overall in the Algonauts 2025 Challenge.