Goto

Collaborating Authors

 flint





A Framework to Learn with Interpretation

Neural Information Processing Systems

This is achieved by a dedicated architecture and well chosen regularization penalties. We seek for a small-size dictionary of high level attribute functions that take as inputs the outputs of selected hidden layers and whose outputs feed a linear classifier. We impose strong conciseness on the activation of attributes with an entropy-based criterion while enforcing fidelity to both inputs and outputs of the predictive model. A detailed pipeline to visualize the learnt features is also developed.


Flint water crisis led to spike in children with special needs and drop in school grades a decade later, according to research that likens fallout from disaster to Chernobyl

Daily Mail - Science & tech

The Flint water crisis has resulted in all-time high numbers of children with special needs and poor performance in school. More than 12,000 children to were exposed to toxic levels of lead in 2014 when the city switched it's public water source to the Flint River, where the water is considerably more acidic. This led to corrosion in lead pipes, which imbued the city's tap water with lead, and then introduced it into the drinking supply. Lead exposure has been linked to behavioral and cognitive problems, mental illness, and an underdeveloped brain. Now, researchers from Michigan and New Jersey experts have reported the rate of young children diagnosed with special needs increased by eight percent after 2014 while performance in math class dropped.


FLINT: A Platform for Federated Learning Integration

Wang, Ewen, Kannan, Ajay, Liang, Yuefeng, Chen, Boyi, Chowdhury, Mosharaf

arXiv.org Artificial Intelligence

Cross-device federated learning (FL) has been well-studied from algorithmic, system scalability, and training speed perspectives. Nonetheless, moving from centralized training to cross-device FL for millions or billions of devices presents many risks, including performance loss, developer inertia, poor user experience, and unexpected application failures. In addition, the corresponding infrastructure, development costs, and return on investment are difficult to estimate. In this paper, we present a device-cloud collaborative FL platform that integrates with an existing machine learning platform, providing tools to measure real-world constraints, assess infrastructure capabilities, evaluate model training performance, and estimate system resource requirements to responsibly bring FL into production. We also present a decision workflow that leverages the FL-integrated platform to comprehensively evaluate the trade-offs of cross-device FL and share our empirical evaluations of business-critical machine learning applications that impact hundreds of millions of users.


FLInt: Exploiting Floating Point Enabled Integer Arithmetic for Efficient Random Forest Inference

Hakert, Christian, Chen, Kuan-Hsun, Chen, Jian-Jia

arXiv.org Artificial Intelligence

In many machine learning applications, e.g., tree-based ensembles, floating point numbers are extensively utilized due to their expressiveness. Nowadays performing data analysis on embedded devices from dynamic data masses becomes available, but such systems often lack hardware capabilities to process floating point numbers, introducing large overheads for their processing. Even if such hardware is present in general computing systems, using integer operations instead of floating point operations promises to reduce operation overheads and improve the performance. In this paper, we provide \mdname, a full precision floating point comparison for random forests, by only using integer and logic operations. To ensure the same functionality preserves, we formally prove the correctness of this comparison. Since random forests only require comparison of floating point numbers during inference, we implement \mdname~in low level realizations and therefore eliminate the need for floating point hardware entirely, by keeping the model accuracy unchanged. The usage of \mdname~basically boils down to a one-by-one replacement of conditions: For instance, a comparison statement in C: if(pX[3]<=(float)10.074347) becomes if((*(((int*)(pX))+3))<=((int)(0x41213087))). Experimental evaluation on X86 and ARMv8 desktop and server class systems shows that the execution time can be reduced by up to $\approx 30\%$ with our novel approach.


ANT: Exploiting Adaptive Numerical Data Type for Low-bit Deep Neural Network Quantization

Guo, Cong, Zhang, Chen, Leng, Jingwen, Liu, Zihan, Yang, Fan, Liu, Yunxin, Guo, Minyi, Zhu, Yuhao

arXiv.org Artificial Intelligence

Quantization is a technique to reduce the computation and memory cost of DNN models, which are getting increasingly large. Existing quantization solutions use fixed-point integer or floating-point types, which have limited benefits, as both require more bits to maintain the accuracy of original models. On the other hand, variable-length quantization uses low-bit quantization for normal values and high-precision for a fraction of outlier values. Even though this line of work brings algorithmic benefits, it also introduces significant hardware overheads due to variable-length encoding and decoding. In this work, we propose a fixed-length adaptive numerical data type called ANT to achieve low-bit quantization with tiny hardware overheads. Our data type ANT leverages two key innovations to exploit the intra-tensor and inter-tensor adaptive opportunities in DNN models. First, we propose a particular data type, flint, that combines the advantages of float and int for adapting to the importance of different values within a tensor. Second, we propose an adaptive framework that selects the best type for each tensor according to its distribution characteristics. We design a unified processing element architecture for ANT and show its ease of integration with existing DNN accelerators. Our design results in 2.8$\times$ speedup and 2.5$\times$ energy efficiency improvement over the state-of-the-art quantization accelerators.


AI finds hidden evidence of ancient human fires 1 million years ago

New Scientist

An artificial intelligence tool has revealed hidden evidence of ancient fire at a 1-million-year-old archaeological site in Israel. Applying the technology at other sites could revolutionise our understanding of when and where humans first began controlling fire, which is widely considered to be one of the most significant innovations of all time. Archaeologists already have a few techniques for identifying whether ancient humans used fire. For instance, you can look for signs that prehistoric bones are discoloured – or that stone tools are warped – in a way that is consistent with exposure to temperatures of 450 C or more. But this sort of evidence is rarely found at sites that are more than 500,000 years old.


Using Spatial Information to Detect Lead Pipes

#artificialintelligence

For centuries, cities in the United States used an inexpensive, malleable, and leak-resistant material for constructing their water pipes: lead. Today, the health risks posed by lead pipes are well-known. Drinking lead-contaminated water can stunt children's development and cause heart and kidney problems among adults.¹ The Environmental Protection Agency (EPA) banned the use of lead pipes for new construction in 1986. Yet, today, lead services lines (the pipes that take water from city lines into individual homes) are still prevalent across the country.