Not enough data to create a plot.
Try a different view from the menu above.
Roychowdhury, Vwani
Optimal Depth Neural Networks for Multiplication and Related Problems
Siu, Kai-Yeung, Roychowdhury, Vwani
An artificial neural network (ANN) is commonly modeled by a threshold circuit, a network of interconnected processing units called linear threshold gates. The depth of a network represents the number of unit delays or the time for parallel computation. The SIze of a circuit is the number of gates and measures the amount of hardware. It was known that traditional logic circuits consisting of only unbounded fan-in AND, OR, NOT gates would require at least O(log n/log log n) depth to compute common arithmetic functions such as the product or the quotient of two n-bit numbers, unless we allow the size (and fan-in) to increase exponentially (in n). We show in this paper that ANNs can be much more powerful than traditional logic circuits.
Computing with Almost Optimal Size Neural Networks
Siu, Kai-Yeung, Roychowdhury, Vwani, Kailath, Thomas
Artificial neural networks are comprised of an interconnected collection of certain nonlinear devices; examples of commonly used devices include linear threshold elements, sigmoidal elements and radial-basis elements. We employ results from harmonic analysis and the theory of rational approximation to obtain almost tight lower bounds on the size (i.e.
Computing with Almost Optimal Size Neural Networks
Siu, Kai-Yeung, Roychowdhury, Vwani, Kailath, Thomas
Artificial neural networks are comprised of an interconnected collection of certain nonlinear devices; examples of commonly used devices include linear threshold elements, sigmoidal elements and radial-basis elements. We employ results from harmonic analysis and the theory of rational approximation toobtain almost tight lower bounds on the size (i.e.