New Major Release for Nebullvm Speeds Up AI Inference by 2-30x

#artificialintelligence 

Nebuly is very excited to announce the new major release nebullvm 0.3.0, Nebullvm is an open-source library that generates an optimized version of your deep learning model that runs 2 to 10 times faster in inference without performance loss by leveraging multiple deep learning compilers (OpenVINO, TensorRT, ONNX Runtime, TVM, etc.). This additional acceleration is achieved by exploiting optimization techniques that slightly modify the model graph to make it lighter, such as quantization, half precision, distillation, sparsity, etc. Find tutorials and examples on how to use nebullvm, as well as installation instructions in the main readme of nebullvm library. It takes a few lines of code to install the library and optimize your models. The library now works on most CPU and GPU and will soon support TPU and other deep learning-specific ASIC.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found