In this blog post, we will learn to create a TensorFlow Lite model using the TF Lite Model Maker Library. We will fine-tune a pre-trained image classification model on the custom dataset and ...
TensorFlow Lite: Model Optimization for
On-Device Machine Learning
The recent trend in the development of larger and larger Deep Learning models for a slight increase in accuracy raises the concern about their computational efficiency and wide scaled usability. We ...
Building Industrial embedded deep learning inference pipelines with TensorRT
You can scarcely find a good article on deploying computer vision systems in industrial scenarios. So, we decided to write a blog post series on the topic. The topics we will cover in this ...
Model Selection and Benchmarking with Modelplace.AI
In this post, we will learn how to select the right model using Modelplace.AI. Selecting the right model will make your application faster, help you scale it to millions of requests, and save a ton ...
Introduction to OpenVINO Deep Learning Workbench
The Intel-OpenVINO Toolkit provides many great functionalities for Deep-Learning model optimization, inference and deployment. Perhaps the most interesting and practical tool among them is the ...
Running OpenVINO Models on Intel Integrated GPU
Traditionally, Deep-Learning models are trained on high-end GPUs. But for inference, Intel CPUs and edge devices like NVidia’s Jetson and Intel-Movidius VPUs are preferred. Most of these Intel CPUs ...