Displaying 1 to 3 from 3 results

Tengine - Tengine is a lite, high performance, modular inference engine for embedded device

  •    C++

Tengine, developed by OPEN AI LAB, is a lite, high-performance, and modular inference engine for embedded device. Tengine is composed of six modules: core/operator/serializer/executor/driver/wrapper.

CHaiDNN - HLS based Deep Neural Network Accelerator Library for Xilinx Ultrascale+ MPSoCs

  •    C++

CHaiDNN is a Xilinx Deep Neural Network library for acceleration of deep neural networks on Xilinx UltraScale MPSoCs. It is designed for maximum compute efficiency at 6-bit integer data type. It also supports 8-bit integer data type. The design goal of CHaiDNN is to achieve best accuracy with maximum performance. The inference on CHaiDNN works in fixed point domain for better performance. All the feature maps and trained parameters are converted from single precision to fixed point based on the precision parameters specified by the user. The precision parameters can vary a lot depending upon the network, datasets, or even across layers in the same network. Accuracy of a network depends on the precision parameters used to represent the feature maps and trained parameters. Well-crafted precision parameters are expected to give accuracy similar to accuracy obtained from a single precision model.

MXNet-HRT - Heterogeneous Run Time version of MXNet

  •    C++

MXNet-HRT is a project that is maintained by OPEN AI LAB, it uses Arm Compute Library (NEON+GPU) to speed up MXNet and provide utilities to debug, profile and tune application performance. The MXNet based version is 26b1cb9ad0bcde9206863a6f847455ff3ec3c266.