Goto

Collaborating Authors

 Data Science


Efficient Kernel Machines Using the Improved Fast Gauss Transform

Neural Information Processing Systems

Such a complexity is significant even for moderate size problems and is prohibitive for large datasets. We present an approximation technique based on the improved fast Gauss transform to reduce the computation to O(N). We also give an error bound for the approximation, and provide experimental results on the UCI datasets.


Detecting Significant Multidimensional Spatial Clusters

Neural Information Processing Systems

Each of these problems can be solved using a spatial scan statistic (Kulldorff, 1997), where we compute the maximum of a likelihood ratio statistic over all spatial regions, and find the significance of this region by randomization. However, computing the scan statistic for all spatial regions is generally computationally infeasible, so we introduce a novel fast spatial scan algorithm, generalizing the 2D scan algorithm of (Neill and Moore, 2004) to arbitrary dimensions. Our new multidimensional multiresolution algorithm allows us to find spatial clusters up to 1400x faster than the naive spatial scan, without any loss of accuracy.


Learning Efficient Auditory Codes Using Spikes Predicts Cochlear Filters

Neural Information Processing Systems

The representation of acoustic signals at the cochlear nerve must serve a wide range of auditory tasks that require exquisite sensitivity in both time and frequency. Lewicki (2002) demonstrated that many of the filtering properties of the cochlea could be explained in terms of efficient coding of natural sounds. This model, however, did not account for properties such as phase-locking or how sound could be encoded in terms of action potentials. Here, we extend this theoretical approach with algorithm for learning efficient auditory codes using a spiking population code. Here, we propose an algorithm for learning efficient auditory codes using a theoretical model for coding sound in terms of spikes.


Euclidean Embedding of Co-Occurrence Data

Neural Information Processing Systems

Embedding algorithms search for low dimensional structure in complex data, but most algorithms only handle objects of a single type for which pairwise distances are specified. This paper describes a method for embedding objectsof different types, such as images and text, into a single common Euclidean space based on their co-occurrence statistics. The joint distributions are modeled as exponentials of Euclidean distances in the low-dimensional embedding space, which links the problem to convex optimizationover positive semidefinite matrices.


The Workshops at the Twentieth National Conference on Artificial Intelligence

AI Magazine

The AAAI-05 workshops were held on Saturday and Sunday, July 9-10, in Pittsburgh, Pennsylvania. The thirteen workshops were Contexts and Ontologies: Theory, Practice and Applications, Educational Data Mining, Exploring Planning and Scheduling for Web Services, Grid and Autonomic Computing, Human Comprehensible Machine Learning, Inference for Textual Question Answering, Integrating Planning into Scheduling, Learning in Computer Vision, Link Analysis, Mobile Robot Workshop, Modular Construction of Humanlike Intelligence, Multiagent Learning, Question Answering in Restricted Domains, and Spoken Language Understanding.


An Opinionated History of AAAI

AI Magazine

AAAI has seen great ups and downs, based largely on the perceived success of AI in business applications. Great early success allowed AAAI to weather the "AI winter" to enjoy the current "thaw." Other challenges to AAAI have resulted from its success in spinning out international conferences, thereby effectively removing several key AI areas from the AAAI National Conference. AAAI leadership continues to look for ways to deal with these challenges. AAI began life intending to be completely societies (such as ACM).


The Workshops at the Twentieth National Conference on Artificial Intelligence

AI Magazine

The AAAI-05 workshops were held on Saturday and Sunday, July 9-10, in Pittsburgh, Pennsylvania. The thirteen workshops were Contexts and Ontologies: Theory, Practice and Applications, Educational Data Mining, Exploring Planning and Scheduling for Web Services, Grid and Autonomic Computing, Human Comprehensible Machine Learning, Inference for Textual Question Answering, Integrating Planning into Scheduling, Learning in Computer Vision, Link Analysis, Mobile Robot Workshop, Modular Construction of Humanlike Intelligence, Multiagent Learning, Question Answering in Restricted Domains, and Spoken Language Understanding.


The General-Motors Variation-Reduction Adviser

AI Magazine

Additional initial ontologies include: search was used, queries were expanded to include (4) single part issues--relate to only one more words to search for, and thus, relevant vehicle component, such as a ding in a fender; documents could be found. Since the documents (5) multiple part issues--relate to two or more being searched were in a limited parts, especially misalignments, unsatisfactory domain, there were few problems with multiple gaps, malformations of joints between parts; senses of words introducing problems that (6) data analysis--results of analysis of measurement hurt precision. In our database, case entries are data generated by optical and mechanical similar--the textual fields do not contain long gages; and (7) plant locations--zones descriptions, and the content is limited to and stations organized topologically or functionally.


Learning Concept Hierarchies from Text Corpora using Formal Concept Analysis

Journal of Artificial Intelligence Research

We present a novel approach to the automatic acquisition of taxonomies or concept hierarchies from a text corpus. The approach is based on Formal Concept Analysis (FCA), a method mainly used for the analysis of data, i.e. for investigating and processing explicitly given information. We follow Harris' distributional hypothesis and model the context of a certain term as a vector representing syntactic dependencies which are automatically acquired from the text corpus with a linguistic parser. On the basis of this context information, FCA produces a lattice that we convert into a special kind of partial order constituting a concept hierarchy. The approach is evaluated by comparing the resulting concept hierarchies with hand-crafted taxonomies for two domains: tourism and finance. We also directly compare our approach with hierarchical agglomerative clustering as well as with Bi-Section-KMeans as an instance of a divisive clustering algorithm. Furthermore, we investigate the impact of using different measures weighting the contribution of each attribute as well as of applying a particular smoothing technique to cope with data sparseness.


Semantic Integration Research in the Database Community: A Brief Survey

AI Magazine

Semantic integration has been a long-standing challenge for the database community. It has received steady attention over the past two decades, and has now become a prominent area of database research. In this article, we first review database applications that require semantic integration and discuss the difficulties underlying the integration process. We then describe recent progress and identify open research issues. We focus in particular on schema matching, a topic that has received much attention in the database community, but also discuss data matching (for example, tuple deduplication) and open issues beyond the match discovery context (for example, reasoning with matches, match verification and repair, and reconciling inconsistent data values). For previous surveys of database research on semantic integration, see Rahm and Bernstein (2001); Ouksel and Seth (1999); and Batini, Lenzerini, and Navathe (1986).