Deep Learning

Once an AI model has been trained, the resulting streamlined neural network can then be presented with new data for it to identify - whether it be to recognise and diagnose medical images, identify spoken words or predict habits – these results, dependant on the model’s training, is called inference.

The first approach when inferring looks at parts of the neural network that don’t get activated after it’s trained. These sections just aren’t needed and can be “pruned” away. The second approach looks for ways to fuse multiple layers of the neural network into a single computational step. It’s akin to the compression that happens to a digital image. Designers might work on these huge, beautiful, million pixel-wide and tall images, but when they go to put it online, they’ll turn into a jpeg. It’ll be almost exactly the same, indistinguishable to the human eye, but at a smaller resolution. Similarly, with inference you’ll get almost the same accuracy of the prediction, but simplified, compressed and optimized for runtime performance.

As the needs for inference are very different form training, a dedicated set pf hardware is recommended, often required if the place where inferring is to be done – such as remote cameras, drones or autonomous mobile vehicles. To cover all these options Scan has a number of options:

Intel Inferencing Solutions

The Intel Arria and Stratix range of FPGA cards alongside mobile-ready solutions such as the Movidius compute stick offer a comprehensive platform for the development of flexible and scalable inferencing solutions.

Find out more

NVIDIA Tesla T4

The NVIDIA Tesla T4 GPU is the world’s most advanced inference accelerator. Powered by Turing Tensor Cores, and packaged in an energy-efficient 70-watt, small PCIe form factor, T4 is optimized for scale-out servers and is purpose-built to deliver state-of-the-art inference in real time.

Find out more

Jetson AGX Xavier Developer Kit

With the NVIDIA Jetson AGX Xavier developer kit, you can easily create and deploy end-to-end AI robotics applications. In addition the NVIDIA JetPack and DeepStream SDKs, as well as CUDA, cuDNN, and TensorRT software libraries, so the kit provides all the tools you need to get started right away.

Find out more