TensorFlow Lite: Model Optimization for On-Device Machine Learning
The recent trend in the development of larger and larger Deep Learning models for a slight increase in accuracy raises the concern about their computational efficiency and wide scaled usability. We can not use such huge models on resource-constrained devices like mobiles and embedded devices. Does it mean that such devices have to sacrifice accuracy at the cost of a smaller model? Is it possible at all to deploy these models on devices such as smartphones or a Raspberry Pi or even on Microcontrollers? Optimizing the models using TensorFlow Lite is the answer to these questions.
May-3-2022, 14:27:21 GMT
- Technology: