Goto

Collaborating Authors

 Country


Exploratory Feature Extraction in Speech Signals

Neural Information Processing Systems

A novel unsupervised neural network for dimensionality reduction which seeks directions emphasizing multimodality is presented, and its connection toexploratory projection pursuit methods is discussed. This leads to a new statistical insight to the synaptic modification equations governing learning in Bienenstock, Cooper, and Munro (BCM) neurons (1982). The importance of a dimensionality reduction principle based solely on distinguishing features, is demonstrated using a linguistically motivated phoneme recognition experiment, and compared with feature extraction using back-propagation network. 1 Introduction Due to the curse of dimensionality (Bellman, 1961) it is desirable to extract features froma high dimensional data space before attempting a classification. How to perform this feature extraction/dimensionality reduction is not that clear. A first simplification is to consider only features defined by linear (or semi-linear) projections ofhigh dimensional data.


Direct memory access using two cues: Finding the intersection of sets in a connectionist model

Neural Information Processing Systems

For lack of alternative models, search and decision processes have provided the dominant paradigm for human memory access using two or more cues, despite evidence against search as an access process (Humphreys, Wiles & Bain, 1990). We present an alternative process to search, based on calculating the intersection of sets of targets activated by two or more cues. Two methods of computing the intersection are presented, one using information about the possible targets, the other constraining the cue-target strengths in the memory matrix. Analysis using orthogonal vectors to represent the cues and targets demonstrates the competence of both processes, and simulations using sparse distributed representations demonstrate the performance of the latter process for tasks involving 2 and 3 cues.


Phonetic Classification and Recognition Using the Multi-Layer Perceptron

Neural Information Processing Systems

In this paper, we will describe several extensions to our earlier work, utilizing asegment-based approach. We will formulate our segmental framework and report our study on the use of multi-layer perceptrons for detection and classification of phonemes. We will also examine the outputs of the network, and compare the network performance with other classifiers. Our investigation is performed within a set of experiments that attempts to recognize 38 vowels and consonants in American English independent of speaker.


The Devil and the Network: What Sparsity Implies to Robustness and Memory

Neural Information Processing Systems

Robustness is a commonly bruited property of neural networks; in particular, afolk theorem in neural computation asserts that neural networks-in contexts with large interconnectivity-continue to function efficiently, albeit withsome degradation, in the presence of component damage or loss. A second folk theorem in such contexts asserts that dense interconnectivity betweenneural elements is a sine qua non for the efficient usage of resources. These premises are formally examined in this communication in a setting that invokes the notion of the "devil"



Flight Control in the Dragonfly: A Neurobiological Simulation

Neural Information Processing Systems

Neural network simulations of the dragonfly flight neurocontrol system have been developed to understand how this insect uses complex, unsteady aerodynamics. The simulation networks account for the ganglionic spatial distribution of cells as well as the physiologic operating range and the stochastic cellular fIring history of each neuron. In addition the motor neuron firing patterns, "flight command sequences", were utilized. Simulation training was targeted against both the cellular and flight motor neuron firing patterns. The trained networks accurately resynthesized the intraganglionic cellular firing patterns. These in tum controlled the motor neuron fIring patterns that drive wing musculature during flight. Such networks provide both neurobiological analysis tools and fIrst generation controls for the use of "unsteady" aerodynamics.


Using Genetic Algorithms to Improve Pattern Classification Performance

Neural Information Processing Systems

Feature selection and creation are two of the most important and difficult tasks in the field of pattern classification. Good features improve the performance of both conventional and neural network pattern classifiers. Exemplar selection is another task that can reduce the memory and computation requirements of a KNN classifier. These three tasks require a search through a space which is typically so large that 797 798 Chang and Lippmann exhaustive search is impractical. The purpose of this research was to explore the usefulness of Genetic search algorithms for these tasks.



A Novel Approach to Prediction of the 3-Dimensional Structures of Protein Backbones by Neural Networks

Neural Information Processing Systems

One current aim of molecular biology is determination of the (3D) tertiary structures ofproteins in their folded native state from their sequences of amino acid 523 524 Fredholm, Bohr, Bohr, Brunak, Cotterill, Lautrup, and Thtersen residues. Since Kendrew & Perutz solved the first protein structures, myoglobin and hemoglobin, and explained from the discovered structures how these proteins perform their function, it has been widely recognized that protein function is intimately linkedwith protein structure[l]. Within the last two decades X-ray crystallographers have solved the 3-dimensional (3D) structures of a steadily increasing number of proteins in the crystalline state, and recently 2D-NMR spectroscopy has emerged as an alternative method for small proteins in solution. Today approximately three hundred 3D structures have been solved by these methods, although only about half of them can be considered as truly different, and only around a hundred of them are solved at high resolution (that is, less than 2A). The number of protein sequences known today is well over 20,000, and this number seems to be growing at least one order of magnitude faster than the number of known 3D protein structures. Obviously, it is of great importance to develop tools that can predict structural aspects of proteins on the basis of knowledge acquired from known 3D structures.