reid
- Africa > Nigeria (0.05)
- North America > United States > New York (0.04)
- North America > United States > Florida > Palm Beach County > Palm Beach (0.04)
- (2 more...)
- Media > News (1.00)
- Government > Regional Government > North America Government > United States Government (1.00)
Toward Re-Identifying Any Animal
The current state of re-identification (ReID) models poses limitations to their applicability in the open world, as they are primarily designed and trained for specific categories like person or vehicle. In light of the importance of ReID technology for tracking wildlife populations and migration patterns, we propose a new task called ``Re-identify Any Animal in the Wild'' (ReID-AW). This task aims to develop a ReID model capable of handling any unseen wildlife category it encounters. To address this challenge, we have created a comprehensive dataset called Wildlife-71, which includes ReID data from 71 different wildlife categories. This dataset is the first of its kind to encompass multiple object categories in the realm of ReID.
Cross-Modality Perturbation Synergy Attack for Person Re-identification
In recent years, there has been significant research focusing on addressing security concerns in single-modal person re-identification (ReID) systems that are based on RGB images. However, the safety of cross-modality scenarios, which are more commonly encountered in practical applications involving images captured by infrared cameras, has not received adequate attention. The main challenge in cross-modality ReID lies in effectively dealing with visual differences between different modalities. For instance, infrared images are typically grayscale, unlike visible images that contain color information. Existing attack methods have primarily focused on the characteristics of the visible image modality, overlooking the features of other modalities and the variations in data distribution among different modalities. This oversight can potentially undermine the effectiveness of these methods in image retrieval across diverse modalities. This study represents the first exploration into the security of cross-modality ReID models and proposes a universal perturbation attack specifically designed for cross-modality ReID.
Hierarchical Prompt Learning for Image- and Text-Based Person Re-Identification
Zhou, Linhan, Li, Shuang, Dong, Neng, Tai, Yonghang, Zhang, Yafei, Li, Huafeng
Person re-identification (ReID) aims to retrieve target pedestrian images given either visual queries (image-to-image, I2I) or textual descriptions (text-to-image, T2I). Although both tasks share a common retrieval objective, they pose distinct challenges: I2I emphasizes discriminative identity learning, while T2I requires accurate cross-modal semantic alignment. Existing methods often treat these tasks separately, which may lead to representation entanglement and suboptimal performance. To address this, we propose a unified framework named Hierarchical Prompt Learning (HPL), which leverages task-aware prompt modeling to jointly optimize both tasks. Specifically, we first introduce a Task-Routed Transformer, which incorporates dual classification tokens into a shared visual encoder to route features for I2I and T2I branches respectively. On top of this, we develop a hierarchical prompt generation scheme that integrates identity-level learnable tokens with instance-level pseudo-text tokens. These pseudo-tokens are derived from image or text features via modality-specific inversion networks, injecting fine-grained, instance-specific semantics into the prompts. Furthermore, we propose a Cross-Modal Prompt Regularization strategy to enforce semantic alignment in the prompt token space, ensuring that pseudo-prompts preserve source-modality characteristics while enhancing cross-modal transferability.
- Asia > China > Yunnan Province > Kunming (0.04)
- Asia > China > Jiangsu Province > Nanjing (0.04)
- Asia > China > Chongqing Province > Chongqing (0.04)
- Media > News (1.00)
- Government > Regional Government > North America Government > United States Government (1.00)
Meet the history-making Nasa astronauts headed for the Moon next year
The commander of Nasa's next mission to the Moon said that he and his crew would see things that no human has ever seen. Reid Wiseman told a news conference that it was likely that his spacecraft would fly over large areas of the Moon that previous Apollo missions had never mapped. Yesterday, Nasa announced it hoped it would be able to launch the first crewed Moon mission in 50 years as early as February 2026 . Mission specialist Christina Koch explained that the astronauts would be able to study the lunar surface in exquisite detail for a full three hours. Believe it or not, human eyes are one of the best scientific instruments that we have, she said.
- North America > Canada (0.15)
- South America (0.15)
- North America > Central America (0.15)
- (15 more...)
- Government > Space Agency (1.00)
- Government > Regional Government > North America Government > United States Government (1.00)
Generalizable Object Re-Identification via Visual In-Context Prompting
Huang, Zhizhong, Liu, Xiaoming
Current object re-identification (ReID) methods train domain-specific models (e.g., for persons or vehicles), which lack generalization and demand costly labeled data for new categories. While self-supervised learning reduces annotation needs by learning instance-wise invariance, it struggles to capture \textit{identity-sensitive} features critical for ReID. This paper proposes Visual In-Context Prompting~(VICP), a novel framework where models trained on seen categories can directly generalize to unseen novel categories using only \textit{in-context examples} as prompts, without requiring parameter adaptation. VICP synergizes LLMs and vision foundation models~(VFM): LLMs infer semantic identity rules from few-shot positive/negative pairs through task-specific prompting, which then guides a VFM (\eg, DINO) to extract ID-discriminative features via \textit{dynamic visual prompts}. By aligning LLM-derived semantic concepts with the VFM's pre-trained prior, VICP enables generalization to novel categories, eliminating the need for dataset-specific retraining. To support evaluation, we introduce ShopID10K, a dataset of 10K object instances from e-commerce platforms, featuring multi-view images and cross-domain testing. Experiments on ShopID10K and diverse ReID benchmarks demonstrate that VICP outperforms baselines by a clear margin on unseen categories. Code is available at https://github.com/Hzzone/VICP.
- North America > United States > Michigan > Ingham County > Lansing (0.40)
- North America > United States > Michigan > Ingham County > East Lansing (0.40)
- North America > United States > Tennessee > Davidson County > Nashville (0.04)
- (5 more...)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.67)
- Information Technology > Artificial Intelligence > Natural Language > Text Processing (0.66)
A CLIP-based Uncertainty Modal Modeling (UMM) Framework for Pedestrian Re-Identification in Autonomous Driving
Li, Jialin, Wu, Shuqi, Wang, Ning
Re-Identification (ReID) is a critical technology in intelligent perception systems, especially within autonomous driving, where onboard cameras must identify pedestrians across views and time in real-time to support safe navigation and trajectory prediction. However, the presence of uncertain or missing input modalities--such as RGB, infrared, sketches, or textual descriptions--poses significant challenges to conventional ReID approaches. While large-scale pre-trained models offer strong multimodal semantic modeling capabilities, their computational overhead limits practical deployment in resource-constrained environments. To address these challenges, we propose a lightweight Uncertainty Modal Modeling (UMM) framework, which integrates a multimodal token mapper, synthetic modality augmentation strategy, and cross-modal cue interactive learner. Together, these components enable unified feature representation, mitigate the impact of missing modalities, and extract complementary information across different data types. Additionally, UMM leverages CLIP's vision-language alignment ability to fuse multimodal inputs efficiently without extensive finetuning. Experimental results demonstrate that UMM achieves strong robustness, generalization, and computational efficiency under uncertain modality conditions, offering a scalable and practical solution for pedestrian re-identification in autonomous driving scenarios.
- Transportation > Ground > Road (0.83)
- Information Technology > Robotics & Automation (0.83)
- Automobiles & Trucks (0.83)
CORE-ReID V2: Advancing the Domain Adaptation for Object Re-Identification with Optimized Training and Ensemble Fusion
Nguyen, Trinh Quoc, Prima, Oky Dicky Ardiansyah, Irfan, Syahid Al, Purnomo, Hindriyanto Dwi, Tanone, Radius
This study presents CORE-ReID V2, an enhanced framework building upon CORE-ReID. The new framework extends its predecessor by addressing Unsupervised Domain Adaptation (UDA) challenges in Person ReID and Vehicle ReID, with further applicability to Object ReID. During pre-training, CycleGAN is employed to synthesize diverse data, bridging image characteristic gaps across different domains. In the fine-tuning, an advanced ensemble fusion mechanism, consisting of the Efficient Channel Attention Block (ECAB) and the Simplified Efficient Channel Attention Block (SECAB), enhances both local and global feature representations while reducing ambiguity in pseudo-labels for target samples. Experimental results on widely used UDA Person ReID and Vehicle ReID datasets demonstrate that the proposed framework outperforms state-of-the-art methods, achieving top performance in Mean Average Precision (mAP) and Rank-k Accuracy (Top-1, Top-5, Top-10). Moreover, the framework supports lightweight backbones such as ResNet18 and ResNet34, ensuring both scalability and efficiency. Our work not only pushes the boundaries of UDA-based Object ReID but also provides a solid foundation for further research and advancements in this domain. Our codes and models are available at https://github.com/TrinhQuocNguyen/CORE-ReID-V2.
- North America > United States > New York > New York County > New York City (0.14)
- North America > United States > California > Los Angeles County > Long Beach (0.14)
- North America > United States > Utah > Salt Lake County > Salt Lake City (0.04)
- (34 more...)
- Education (1.00)
- Information Technology > Security & Privacy (0.45)
- Information Technology > Artificial Intelligence > Machine Learning > Unsupervised or Indirectly Supervised Learning (0.87)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning > Clustering (0.68)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.67)
Inside the race to find GPS alternatives
"Just because of this shorter distance, we will put down signals that will be approximately a hundred times stronger than the GPS signal," says Tyler Reid, chief technology officer and cofounder of Xona. "That means the reach of jammers will be much smaller against our system, but we will also be able to reach deeper into indoor locations, penetrating through multiple walls." The first GPS system went live in 1993. In the decades since, it has become one of the foundational technologies that the world depends on. The precise positioning, navigation, and timing (PNT) signals beamed by its satellites underpin much more than Google Maps in your phone. They guide drill heads at offshore oil rigs, time-stamp financial transactions, and help sync power grids all over the world.
- Information Technology (0.96)
- Energy > Oil & Gas > Upstream (0.93)