Contents
This article will review the Intel Neural Compute Stick 2 (NCS 2) based on the Intel Movidius Myriad X chip. Learn about the advantages of using the NCS 2 for Edge AI, computing Artificial Intelligence tasks on the edge.
Specifically, we will cover:
What is Intel Neural Compute Stick 2?
The edge is becoming an increasingly popular destination for deploying computer vision or deep learning models.
Edge Computing provides advantages such as local data processing, filtered data transfer to the cloud, or faster decision-making.
The Intel Neural Compute Stick 2 is powered by the Intel Movidius X VPU to power on-device AI applications at high performance with ultra-low power consumption. With new performance enhancements, the Intel Movidius Myriad X VPU is a power-efficient solution revolving around edge computing that brings computer vision and AI applications to edge devices such as connected computers, drones, smart cameras, smart home, security, VR/AR headsets, and 360-degree cameras.
The NCS 2 is a small, fan-less neural network training and deployment device that can be used for AI programming at the edge.
AI accelerators like the Intel Stick 2 VPU are useful for accelerating data-intensive deep learning inference on edge devices in a very cost-effective way. These accelerators work by assisting the edge device computer processing unit (CPU) by taking over the mathematical burden needed for running deep learning models.
Edge accelerators allow deep learning models to be run at low costs, low power consumption, and faster speeds. The benefits are primarily measured using throughput, value, latency, and efficiency.
Movidius Myriad X – High-Performance Computer Vision Inferencing
The Intel Movidius Myriad X Vision Processing Unit (VPU) is Intel’s first VPU to feature the Neural Compute Engine — a dedicated hardware accelerator for deep neural network inference. The Movidius Myriad X VPU is sold as a chip embedded in a Neural Compute Stick (similar to a USB drive) specifically built for processing images and video inputs. The USB containment of the chip allows it to be easily compatible with a Raspberry Pi or Intel NUC device (any popular computing architecture such as x86 PCs).
The Intel Movidius Neural Compute Stick engine is a chip-form hardware piece designed to run deep neural networks at high speed and low power without compromising accuracy, enabling computer vision processes in real time.
This engine is implemented within the USB casing along with Movidius Myriad X. When used concurrently, the Neural Compute Engine, 16 powerful SHAVE cores, and high throughput memory fabric make the Intel Movidius Myriad X neural compute stick ideal for training and deploying deep neural networks and computer vision products.
The benchmarks show the relative success rates of various machine learning algorithms on different inference engine samples. For the popular ssd300-CF (Caffe backend), the throughput rate is directly proportional to the accuracy, while the latency rate is inversely proportional to quality.
Accordingly, the Movidius accelerator achieves a higher throughput for the ssd300-CF than any of the core i3 to i9’s (except for the Intel Core i9-10920X). In addition, it has lower latency than any inferencing engine, which makes it a relatively robust platform for running artificial intelligence computer vision models.
Computer vision inferencing with object detection in real-time
OpenVINO with Movidius Neural Compute Stick 2
The Intel Distribution of OpenVINO (Open Visual Inference and Neural Network) allows computer vision models trained in the cloud to be run at the edge. The OpenVINO developer’s toolkit contains a full suite of development and deployment tools best used in conjunction with Movidius Myriad X.
The toolkit facilitates faster inference of deep learning models by creating cost-effective and robust computer vision applications. If you want to learn more about the toolkit, I recommend reading our full overview of OpenVINO. You can easily get started with “install OpenVINO.”
Supporting numerous deep learning models out of the box, the toolkit expedites the computer vision application production process by cutting down on raw creation and setup time. The pre-trained models, which range from deep learning frameworks such as YOLO (You Only Look Once) to R-CNN and ResNet, allow developers to create models that carry out complex computer vision applications such as face detection, person detection, vehicle detection, and people counting.
The Myriad Development Kit (MDK) further includes necessary development tools, frameworks, and APIs to implement custom vision, imaging, and deep neural network workloads on the chip. For example, existing Convolutional Neural Network (CNN) models can be converted into OpenVINO Intermediate Representation (IR) which drastically reduces the model’s size while simultaneously optimizing it for inferencing.
Intel Neural Compute Stick 2 (Intel NCS2) Review
Pro’s – Optimized for AI at the edge
Con’s – Things You Need to Know
How to Use the NCS 2
The Traditional Way – Learn it Yourself/Prototyping
The OpenVINO toolkit offers development tools to deploy applications and solutions for computer vision across Intel hardware, using the Neural Computer Stick 2. Read more about the distribution of the OpenVINO toolkit and how it can be used to perform CNN-based deep learning inference on the edge, with one or multiple NCS devices to distribute the workload. Intel offers extensive documentation for setting up OpenVINO on a device to which you can connect one or multiple NCS devices.
Computer Vision Platforms – Ready to Use/Production
Most probably, the easiest way to use the Neural Compute Stick and other VPU devices for Computer Vision are AI Low-Code Platforms. You can benefit from OpenVINO install integrations and built-in support for the latest algorithms and frameworks (YOLO, TensorFlow, PyTorch). Platforms make it easier to develop complete AI applications using the Neural Compute Stick. Also, you don’t need to start from scratch to perform video pre-processing tasks and create workflows to integrate processed output with third-party systems.
We at ProX PC power the End-to-End AI platform for Computer Vision. It’s probably the fastest way to build a computer vision prototype and scale it into production on scalable infrastructure. Hence, we do take care of the cloud and deployment infrastructure, container management, DevOps tasks, and security.
Related Products
Share this: