Go to file
jing.tang a85fe89cf6 add docs for ops 2021-05-21 18:39:59 +08:00
cmake Support local sdk for cmake (#41) 2021-05-13 22:28:36 +08:00
docs add docs for ops 2021-05-21 18:39:59 +08:00
include/tim add docs for ops 2021-05-21 18:39:59 +08:00
prebuilt-sdk Update simulator SDK to 6.4.6.2 2021-05-20 02:05:09 +08:00
samples Update tim lite api (#48) 2021-05-17 22:08:10 +08:00
src/tim add float unit_test for conv2d 2021-05-21 18:13:02 +08:00
toolchains support build for tensorflow A311D 2021-02-07 10:33:04 +08:00
.bazelrc Add prebuild support for VIPLite 2021-05-14 18:31:08 +08:00
.bazelversion Support build for A311D 2021-01-29 00:11:41 -08:00
.clang-format Add .clang-format 2021-01-19 09:54:50 +08:00
.gitignore Update gitignore 2021-05-11 12:53:31 +08:00
Android.mk Minor cleanup 2021-05-06 19:48:36 +08:00
BUILD Add prebuild support for VIPLite 2021-05-14 18:31:08 +08:00
CMakeLists.txt Refine cmake build: add gtest (#47) 2021-05-17 13:04:45 +08:00
LICENSE Initial Commit for VERSION 1.1.28 2021-01-11 18:27:48 +08:00
README.md Minor Cleanup 2021-05-08 12:13:02 +08:00
VERSION v1.1.30 2021-02-26 17:20:36 +08:00
WORKSPACE Add prebuild support for VIPLite 2021-05-14 18:31:08 +08:00

README.md

TIM-VX - Tensor Interface Module for OpenVX

TIM-VX is a software integration module provided by VeriSilicon to facilitate deployment of Neural-Networks on OpenVX enabled ML accelerators. It serves as the backend binding for runtime frameworks such as Android NN, Tensorflow-Lite, MLIR, TVM and more.

Main Features

  • Over 130 internal operators with rich format support for both quantized and floating point
  • Simplified binding API calls to create Tensors and Operations
  • Dynamic graph construction and supports shape inferencing
  • Built-in custom layer extensions
  • A set of utility functions for debugging

Framework Support

Roadmap

Roadmap of TIM-VX will be updated here in the future.

Get started

Build and Run

TIM-VX uses bazel build system by default. Install bazel first to get started.

TIM-VX needs to be compiled and linked against VeriSilicon OpenVX SDK which provides related header files and pre-compiled libraries. A default linux-x86_64 SDK is provided which contains the simulation environment on PC. Platform specific SDKs can be obtained from respective SoC vendors.

To build TIM-VX

bazel build libtim-vx.so

To run sample LeNet

# set VIVANTE_SDK_DIR for runtime compilation environment
export VIVANTE_SDK_DIR=`pwd`/prebuilt-sdk/x86_64_linux

bazel build //samples/lenet:lenet_asymu8_cc
bazel run //samples/lenet:lenet_asymu8_cc

To build and run Tensorflow-Lite delegate on A311D platform

# clone and cross build VeriSilicon tensorflow fork with TFlite delegate support
git clone --single-branch --branch vx-delegate.v2.4.1 git@github.com:VeriSilicon/tensorflow.git vx-delegate; cd vx-delegate
bazel build --config A311D //tensorflow/lite/tools/benchmark:benchmark_model

# push benchmark_model onto device and run
./benchmark_model --graph=mobilenet_v1_1.0_224_quant.tflite --use_vxdelegate=true