Displaying 1 to 7 from 7 results

ncnn - ncnn is a high-performance neural network inference framework optimized for the mobile platform

  •    C

ncnn is a high-performance neural network inference computing framework optimized for mobile platforms. ncnn is deeply considerate about deployment and uses on mobile phones from the beginning of design. ncnn does not have third party dependencies. it is cross-platform, and runs faster than all known open source frameworks on mobile phone cpu. Developers can easily deploy deep learning algorithm models to the mobile platform by using efficient ncnn implementation, create intelligent APPs, and bring the artificial intelligence to your fingertips. ncnn is currently being used in many Tencent applications, such as QQ, Qzone, WeChat, Pitu and so on.

FeatherCNN - FeatherCNN is a high performance inference engine for convolutional neural networks.

  •    C++

FeatherCNN, developed by Tencent TEG AI Platform, is a high-performance lightweight CNN inference library. FeatherCNN is currently targeting at ARM CPUs, and is capable to extend to other devices in the future. Highly Performant FeatherCNN delivers state-of-the-art inference computing performance on a wide range of devices, including mobile phones (iOS/Android), embedded devices (Linux) as well as ARM-based servers (Linux).

Caffe-HRT - Heterogeneous Run Time version of Caffe

  •    C++

Caffe-HRT is a project that is maintained by OPEN AI LAB, it uses heterogeneous computing infrastructure framework to speed up Caffe and provide utilities to debug, profile and tune application performance. The Caffe based version is 793bd96351749cb8df16f1581baf3e7d8036ac37.

Tengine - Tengine is a lite, high performance, modular inference engine for embedded device

  •    C++

Tengine, developed by OPEN AI LAB, is a lite, high-performance, and modular inference engine for embedded device. Tengine is composed of six modules: core/operator/serializer/executor/driver/wrapper.




sse-popcount - SIMD (SSE) population count --- http://0x80.pl/articles/sse-popcount.html

  •    C++

Daniel Lemire, Nathan Kurz and I published an article Faster Population Counts using AVX2 Instructions. Subdirectory original contains code from 2008 --- it is 32-bit and GCC-centric. The root directory contains fresh C++11 code, written with intrinsics and tested on 64-bit machines.

MXNet-HRT - Heterogeneous Run Time version of MXNet

  •    C++

MXNet-HRT is a project that is maintained by OPEN AI LAB, it uses Arm Compute Library (NEON+GPU) to speed up MXNet and provide utilities to debug, profile and tune application performance. The MXNet based version is 26b1cb9ad0bcde9206863a6f847455ff3ec3c266.

TensorFlow-HRT - Heterogeneous Run Time version of TensorFlow

  •    C++

TensorFlow-HRT is a project that is maintained by OPEN AI LAB, it uses heterogeneous computing infrastructure framework to speed up Tensorflow and provide utilities to debug, profile and tune application performance. There are some compatibility issues between ACL and Tensorflow ops.