Goto

Collaborating Authors

 Phillips, Jeff


Quantized Wasserstein Procrustes Alignment of Word Embedding Spaces

arXiv.org Artificial Intelligence

In natural language processing (NLP), the problem of aligning monolingual embedding spaces to induce a shared cross-lingual vector space has been shown not only to be useful in a variety of tasks such as bilingual lexicon induction (BLI) (Mikolov et al., 2013; Barone, 2016; Artetxe et al., 2017; Aboagye et al., 2022), machine translation (Artetxe et al., 2018b), cross-lingual information retrieval (Vuliฤ‡ & Moens, 2015), but it plays a crucial role in facilitating the cross-lingual transfer of language technologies from high resource languages to low resource languages. Cross-lingual word embeddings (CLWEs) represent words from two or more languages in a shared cross-lingual vector space in which words with similar meanings obtain similar vectors regardless of their language. There has been a flurry of work dominated by the so-called projection-based CLWE models (Mikolov et al., 2013; Artetxe et al., 2016, 2017, 2018a; Smith et al., 2017; Ruder et al., 2019), which aim to improve CLWE model performance significantly. Projection-based CLWE models learn a transfer function or mapper between two independently trained monolingual word vector spaces with limited or no cross-lingual supervision. Famous among projection-based CLWE models are the unsupervised projection-based CLWE models (Artetxe et al., 2017; Lample et al., 2018; Alvarez-Melis & Jaakkola, 2018;


Learning In Practice: Reasoning About Quantization

arXiv.org Machine Learning

There is a mismatch between the standard theoretical analyses of statistical machine learning and how learning is used in practice. The foundational assumption supporting the theory is that we can represent features and models using real-valued parameters. In practice, however, we do not use real numbers at any point during training or deployment. Instead, we rely on discrete and finite quantizations of the reals, typically floating points. In this paper, we propose a framework for reasoning about learning under arbitrary quantizations. Using this formalization, we prove the convergence of quantization-aware versions of the Perceptron and Frank-Wolfe algorithms. Finally, we report the results of an extensive empirical study of the impact of quantization using a broad spectrum of datasets.


High Definition Fiber Tracking Exposes Circuit Diagram for Brain Showing Triarchic Representation, Domain General Control, and Metacognitive Subsystems

AAAI Conferences

Dramatic advances in the last six months in High Definition Fiber Tracking (HDFT) make it possible to image the fiber connectivity from source to destination mapping hundreds of thousands fiber tracks with sufficient resolution to identify the cable level circuit diagram of the human brain. Brain activity imaging studies using functional Magnetic Resonance Imagining (fMRI) identify differential activation patterns as a function of task and level of practice. These data show subnetworks with communication of high bandwidth vector associations, scalar priority and control signals, and interactions with control and meta cognition. The connectivity and activity data support a triarchic cognitive architecture. Processing is the synergistic interaction of three interlinked cognitive computational systems with differential computation role and evolutionary history. These data provided a detailed diagram to guide reverse engineering of the systems levels of the human brain.