Goto

Collaborating Authors

 edge tpus


Deep Learning on Edge TPUs

Sun, Yipeng, Kist, Andreas M

arXiv.org Artificial Intelligence

Computing at the edge is important in remote settings, however, conventional hardware is not optimized for utilizing deep neural networks. The Google Edge TPU is an emerging hardware accelerator that is cost, power and speed efficient, and is available for prototyping and production purposes. Here, I review the Edge TPU platform, the tasks that have been accomplished using the Edge TPU, and which steps are necessary to deploy a model to the Edge TPU hardware. The Edge TPU is not only capable of tackling common computer vision tasks, but also surpasses other hardware accelerators, especially when the entire model can be deployed to the Edge TPU. Co-embedding the Edge TPU in cameras allows a seamless analysis of primary data. In summary, the Edge TPU is a maturing system that has proven its usability across multiple tasks.


An Evaluation of Edge TPU Accelerators for Convolutional Neural Networks

Seshadri, Kiran, Akin, Berkin, Laudon, James, Narayanaswami, Ravi, Yazdanbakhsh, Amir

arXiv.org Artificial Intelligence

Edge TPUs are a domain of accelerators for low-power, edge devices and are widely used in various Google products such as Coral and Pixel devices. In this paper, we first discuss the major microarchitectural details of Edge TPUs. Then, we extensively evaluate three classes of Edge TPUs, covering different computing ecosystems, that are either currently deployed in Google products or are the product pipeline, across 423K unique convolutional neural networks. Building upon this extensive study, we discuss critical and interpretable microarchitectural insights about the studied classes of Edge TPUs. Mainly, we discuss how Edge TPU accelerators perform across convolutional neural networks with different structures. Finally, we present our ongoing efforts in developing high-accuracy learned machine learning models to estimate the major performance metrics of accelerators such as latency and energy consumption. These learned models enable significantly faster (in the order of milliseconds) evaluations of accelerators as an alternative to time-consuming cycle-accurate simulators and establish an exciting opportunity for rapid hard-ware/software co-design.


A Study on the Use of Edge TPUs for Eye Fundus Image Segmentation

#artificialintelligence

Medical image segmentation can be implemented using Deep Learning methods with fast and efficient segmentation networks. Single-board computers (SBCs) are difficult to use to train deep networks due to their memory and processing limitations. Specific hardware such as Google's Edge TPU makes them suitable for real time predictions using complex pre-trained networks. In this work, we study the performance of two SBCs, with and without hardware acceleration for fundus image segmentation, though the conclusions of this study can be applied to the segmentation by deep neural networks of other types of medical images. To test the benefits of hardware acceleration, we use networks and datasets from a previous published work and generalize them by testing with a dataset with ultrasound thyroid images.


New Coral APIs and tools for AI at the edge

#artificialintelligence

Fall has finally arrived and with it a new release of Coral's C and Python APIs and tools, along with new models optimized for the Edge TPU and further support for TensorFlow 2.0-based workflows. Coral is a complete toolkit to build products with local AI. Our on-device inferencing capabilities allow you to build products that are efficient, private, fast and offline with the help of TensorFlow Lite and the Edge TPU. From the beginning, we've provided APIs in Python and C that enable developers to take advantage of the Edge TPU's local inference speed. Offline processing for machine learning models allows for considerable savings on bandwidth and cloud compute costs, it keeps data local, and it preserves user privacy.


Google Forays Into Edge Computing With Cloud IoT Edge And TPU

Forbes - Tech

Tensor Processing Unit (TPU), an application specific integrated circuit, designed by Google for accelerating machine learning workloads, is going to be available at the edge. These tailor-made chips complement Cloud TPUs by inferencing machine learning models deployed at the edge. Google has also announced Cloud IoT Edge, an edge computing platform that extends Google Cloud's data processing and machine learning to edge devices. Google is the latest entrant into the edge computing market. The key competitors of Google Cloud – Amazon and Microsoft – have a comprehensive edge computing strategy.


Google's cloud team is building A.I. chips for web-connected devices

#artificialintelligence

Google is moving beyond developing artificial intelligence chips for its own data centers and is now designing them to work inside products made by other companies. After unveiling the Tensor Processing Unit two years ago, Google announced on Wednesday the Edge TPU, which will enable sensors and other gadgets to process data more quickly. The chips could be used in a wide variety of scenarios, but one initial use is in industrial manufacturing: Consumer electronics maker LG is testing them in a system that detects manufacturing defects in glass for displays. Google's jump into custom silicon is one way it's trying to expand its cloud market share against Amazon and Microsoft. Since 2015, Google has been using TPUs to accelerate certain workloads in its own data centers, rather than relying on commercially available hardware from vendors like Nvidia.


Google Unveils Tiny New AI Chips for On-device Machine Learning

#artificialintelligence

Two years ago, Google unveiled its Tensor Processing Units or TPUs -- specialized chips that live in the company's data centers and make light work of AI tasks. Now, the company is moving its AI expertise down from the cloud, and has taken the wraps off its new Edge TPU; a tiny AI accelerator that will carry out machine learning jobs in IoT devices. The Edge TPU is designed to do what's known as "inference." This is the part of machine learning where an algorithm actually carries out the task it was trained to do; like, for example, recognizing an object in a picture. Google's server-based TPUs are optimized for the training part of this process, while these new Edge TPUs will do the inference.


Google unveils tiny new AI chips for on-device machine learning

#artificialintelligence

Two years ago, Google unveiled its Tensor Processing Units or TPUs -- specialized chips that live in the company's data centers and make light work of AI tasks. Now, the company is moving its AI expertise down from the cloud, and has taken the wraps off its new Edge TPU; a tiny AI accelerator that will carry out machine learning jobs in IoT devices. The Edge TPU is designed to do what's known as "inference." This is the part of machine learning where an algorithm actually carries out the task it was trained to do; like, for example, recognizing an object in a picture. Google's server-based TPUs are optimized for the training part of this process, while these new Edge TPUs will do the inference. These new chips are destined to be used in enterprise jobs, not your next smartphone.