Update README

Signed-off-by: Kainan Cha <kainan.zha@verisilicon.com>
This commit is contained in:
Kainan Cha 2021-08-26 16:33:33 +08:00
parent eb28f8b3ed
commit ff22b3e34c
1 changed files with 6 additions and 5 deletions

View File

@ -1,13 +1,13 @@
# TIM-VX - Tensor Interface Module for OpenVX # TIM-VX - Tensor Interface Module
![VSim.X86.UnitTest](https://github.com/VeriSilicon/TIM-VX/actions/workflows/x86_vsim_unit_test.yml/badge.svg) ![VSim.X86.UnitTest](https://github.com/VeriSilicon/TIM-VX/actions/workflows/x86_vsim_unit_test.yml/badge.svg)
TIM-VX is a software integration module provided by VeriSilicon to facilitate deployment of Neural-Networks on OpenVX enabled ML accelerators. It serves as the backend binding for runtime frameworks such as Android NN, Tensorflow-Lite, MLIR, TVM and more. TIM-VX is a software integration module provided by VeriSilicon to facilitate deployment of Neural-Networks on Verisilicon ML accelerators. It serves as the backend binding for runtime frameworks such as Android NN, Tensorflow-Lite, MLIR, TVM and more.
Main Features Main Features
- Over [130 operators](https://github.com/VeriSilicon/TIM-VX/blob/main/src/tim/vx/ops/README.md) with rich format support for both quantized and floating point - Over [150 operators](https://github.com/VeriSilicon/TIM-VX/blob/main/src/tim/vx/ops/README.md) with rich format support for both quantized and floating point
- Simplified C++ binding API calls to create Tensors and Operations - Simplified C++ binding API calls to create Tensors and Operations [Guide](https://github.com/VeriSilicon/TIM-VX/blob/main/docs/Programming_Guide.md)
- Dynamic graph construction with support for shape inference and layout inference - Dynamic graph construction with support for shape inference and layout inference
- Built-in custom layer extensions - Built-in custom layer extensions
- A set of utility functions for debugging - A set of utility functions for debugging
@ -43,4 +43,5 @@ bazel run //samples/lenet:lenet_asymu8_cc
``` ```
To build and run Tensorflow-Lite with TIM-VX, please see [README](https://github.com/VeriSilicon/tflite-vx-delegate#readme) To build and run Tensorflow-Lite with TIM-VX, please see [README](https://github.com/VeriSilicon/tflite-vx-delegate#readme)
To build and run TVM with TIM-VX, please see [TVM](https://github.com/VeriSilicon/tvm)
To build and run TVM with TIM-VX, please see [TVM README](https://github.com/VeriSilicon/tvm/blob/vsi_npu/README.VSI.md)