coot
CO-Optimal Transport
Optimal transport (OT) is a powerful geometric and probabilistic tool for finding correspondences and measuring similarity between two distributions. Yet, its original formulation relies on the existence of a cost function between the samples of the two distributions, which makes it impractical when they are supported on different spaces. To circumvent this limitation, we propose a novel OT problem, named COOT for CO-Optimal Transport, that simultaneously optimizes two transport maps between both samples and features, contrary to other approaches that either discard the individual features by focusing on pairwise distances between samples or need to model explicitly the relations between them. We provide a thorough theoretical analysis of our problem, establish its rich connections with other OT-based distances and demonstrate its versatility with two machine learning applications in heterogeneous domain adaptation and co-clustering/data summarization, where COOT leads to performance improvements over the state-of-the-art methods.
All reviewers
Thank you for the constructive comments and suggestions. This indicates success of our model in capturing long-range semantics, which is the main theme of our paper. We report the results of video captioning in TabA 1-Left. VideoBERT uses more sophisticated transformer based method. VideoBERT if we use the same captioning method.
Supplementary materials for paper: CO-Optimal Transport
We recall the notations of the paper. The rest of the supplementary is organized as follows. Figure 1: Comparison between the coupling matrices obtained via GW and COOT on MNIST -USPS. Finally, as the cost is symmetric w.r .t This proof follows the proof of Theorem 2.2 in [ Note also that this result holds when we add a constant term to the cost function.2.2 Proofs of Propositions 2 and 3 We now prove all the theorems from Section 3 from the main paper.
- North America > United States (0.35)
- North America > Canada (0.04)
- North America > United States (0.15)
- North America > Canada (0.04)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.04)
- Africa > Middle East > Morocco > Casablanca-Settat Region > Casablanca (0.04)
Cognition-of-Thought Elicits Social-Aligned Reasoning in Large Language Models
Zhang, Xuanming, Chen, Yuxuan, Yeh, Samuel, Li, Sharon
Large language models (LLMs) excel at complex reasoning but can still exhibit harmful behaviors. Current alignment strategies typically embed safety into model weights, making these controls implicit, static, and difficult to modify. This paper introduces Cognition-of-Thought (CooT), a novel decoding-time framework that equips LLMs with an explicit cognitive self-monitoring loop. CooT couples a standard text Generator with a cognitive Perceiver that continuously monitors the unfolding sequence. The Perceiver uses a structured, precedence-based hierarchy of principles (e.g., safety over obedience) to detect potential misalignments as they arise. When violations are flagged, CooT intervenes by rolling back the generation to the point of error and regenerating under injected guidance that combines universal social priors with context-specific warnings. CooT thus transforms alignment from a fixed property into an explicit, dynamic, and auditable process active during inference, allowing for flexible policy updates without retraining the model. Extensive experiments across multiple benchmarks and model families confirm that CooT consistently improves safety and social reasoning performance.
- North America > United States > Wisconsin > Dane County > Madison (0.04)
- North America > United States > New York (0.04)
- Europe > United Kingdom > England > Greater London > London > Wimbledon (0.04)
- (3 more...)
- Research Report > New Finding (0.45)
- Research Report > Experimental Study (0.45)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Cognitive Science (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.94)
All reviewers
Thank you for the constructive comments and suggestions. This indicates success of our model in capturing long-range semantics, which is the main theme of our paper. We report the results of video captioning in TabA 1-Left. VideoBERT uses more sophisticated transformer based method. VideoBERT if we use the same captioning method.
- North America > United States (0.16)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- North America > United States (0.15)
- North America > Canada (0.04)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.04)
- Africa > Middle East > Morocco > Casablanca-Settat Region > Casablanca (0.04)
CO-Optimal Transport
Optimal transport (OT) is a powerful geometric and probabilistic tool for finding correspondences and measuring similarity between two distributions. Yet, its original formulation relies on the existence of a cost function between the samples of the two distributions, which makes it impractical when they are supported on different spaces. To circumvent this limitation, we propose a novel OT problem, named COOT for CO-Optimal Transport, that simultaneously optimizes two transport maps between both samples and features, contrary to other approaches that either discard the individual features by focusing on pairwise distances between samples or need to model explicitly the relations between them. We provide a thorough theoretical analysis of our problem, establish its rich connections with other OT-based distances and demonstrate its versatility with two machine learning applications in heterogeneous domain adaptation and co-clustering/data summarization, where COOT leads to performance improvements over the state-of-the-art methods.
Revisiting invariances and introducing priors in Gromov-Wasserstein distances
Demetci, Pinar, Tran, Quang Huy, Redko, Ievgen, Singh, Ritambhara
Gromov-Wasserstein distance has found many applications in machine learning due to its ability to compare measures across metric spaces and its invariance to isometric transformations. However, in certain applications, this invariance property can be too flexible, thus undesirable. Moreover, the Gromov-Wasserstein distance solely considers pairwise sample similarities in input datasets, disregarding the raw feature representations. We propose a new optimal transport-based distance, called Augmented Gromov-Wasserstein, that allows for some control over the level of rigidity to transformations. It also incorporates feature alignments, enabling us to better leverage prior knowledge on the input data for improved performance. We present theoretical insights into the proposed metric. We then demonstrate its usefulness for single-cell multi-omic alignment tasks and a transfer learning scenario in machine learning.
- Europe > France (0.04)
- North America > United States > Rhode Island > Providence County > Providence (0.04)
- Africa > Togo (0.04)