This site uses cookies to improve customer service and for other purposes.
For more information, please click here.

Technology

  1. HOME
  2. Technology
  3. Fast AI Inference Engine

Fast AI Inference Engine

SoftNeuro®

“SoftNeuro®” is one of the world’s fastest deep learning inference engines. It operates in multiple environments, utilizing learning results that have been obtained through a variety of deep learning frameworks.

Use Case

  • Visual inspection
  • Self-driving / On-board camera
  • Structure deterioration diagnosis
  • Factory automation
  • Medical testing and diagnostic support
  • Smartphone / tablet
  • Surveillance camera
  • Ship / Train / Airplane
  • Autonomous vehicle

Function or Purpose

  • Accelerate AI inference

Features

  • Realize inference processing at the world's fastest processing speed in CPU-only processing.
  • Reasoning using learning results of major deep learning frameworks is feasible.
  • Multi-platform support allows learning results to be deployed on a wide range of operating platforms.
  • Since it includes encryption capability, the risk of leakage of know-how, and results of learning, is reduced.

Product Details

“SoftNeuro®” is one of the world’s fastest (*1) deep learning inference engines. It operates in multiple environments, utilizing learning results that have been obtained through a variety of deep learning frameworks. (Fig.1) Moreover, “SoftNeuro®” is not limited to image recognition. It is a general-purpose inference engine that can be used for text analysis, voice recognition, image recognition, etc.

Morpho not only licenses “SoftNeuro®” as a standalone inference engine, but also brings about a large increase in the processing speed of their existing products by embedding “SoftNeuro®” into them. An example is “Morpho Deep Recognizer™”, which is Morpho’s image recognition engine.

Features

1.One of the fastest in the world (According to Morpho’s research)
“SoftNeuro®” is faster than or as equally fast as the mainstream inference engines – when run on CPUs – while providing a set of key features as described later. This high-speed performance has been achieved through a variety of optimizations (neural network, memory usage and others) for each platform. Please refer to *1 for details.

 

2.Supports multiple frameworks
“SoftNeuro®” achieves fast processing by utilizing the learning results of these major frameworks (Fig.1). It is possible to achieve fast inference processing and multi-platform compatibility without discarding the learning assets that users have built up so far. Further, the compatibility of “SoftNeuro” with frameworks and layers will be expanded sequentially.

 

3.Multi-platform compatibility
“SoftNeuro®” is scheduled to be applied to a variety of platforms, and appropriate optimization (CPU speed-up instructions, use of GPU, DSP and other specialized hardware) for each platform will be performed.
Multi-platform compatibility facilitates deploying the learning results on a wide range of platforms. It also provides flexibility to quickly respond to changes in platforms and hardware trends.

 

4.Compatible with secure file formats
“SoftNeuro®” is capable of encrypting the trained networks, minimizing the risk of leaking the machine learning know-how and the learned model parameters.

 

*1:One of the Fastest in the world (According to Morpho’s research)
We compared the main inference engines that are publicly available. The measurement of the inference processing speed is performed on the CPU environment, comparing the main inference engines that are available to us.
The inference processing speed measurement results are shown in Fig. 2.

 

●Figure 1. Integrating “SoftNeuro” into Machine Learning Systems

 

●Figure 2. Comparison of Inference Speeds (CPU: ARM)
Measured on: Qualcomm Snapdragon 855 (Image size used for inference : 224×224 pixels)

 

●Relations Among AI Products

Awards

Vision Product of the Year Awards 2018

The Format to Provide the Product

Please contact us for more information.