diff --git a/docs/Operators.md b/docs/Operators.md new file mode 100644 index 0000000..4f272b7 --- /dev/null +++ b/docs/Operators.md @@ -0,0 +1,770 @@ + + +# Table of Contents +- [Operators](#operators) + - [Activation](#activation) + - [AddN](#addn) + - [ArgMin/ArgMax](#argminargmax) + - [Batch2Space](#batch2space) + - [BatchNorm](#batchnorm) + - [Clip](#clip) + - [Concat](#concat) + - [Conv2d](#conv2d) + - [DeConv2d](#deconv2d) + - [DeConv1d](#deconv1d) + - [DepthToSpace](#depthtospace) + - [Dropout](#dropout) + - [Add](#add) + - [Sub](#sub) + - [Multiply](#multiply) + - [Div](#div) + - [Pow](#pow) + - [Minimum](#minimum) + - [Maximum](#maximum) + - [FloorDiv](#floordiv) + - [FullyConnected](#fullyconnected) + - [Gather](#gather) + - [GatherNd](#gathernd) + - [GroupedConv2d](#groupedconv2d) + - [L2Normalization](#l2normalization) + - [LocalResponseNormalization](#localresponsenormalization) + - [And](#and) + - [Or](#or) + - [LogSoftmax](#logsoftmax) + - [Matmul](#matmul) + - [MaxpoolWithArgmax](#maxpoolwithargmax) + - [MaxUnpool2d](#maxunpool2d) + - [Moments](#moments) + - [NBG](#nbg) + - [Pad](#pad) + - [Pool2d](#pool2d) + - [ReduceMin](#reducemin) + - [ReduceMax](#reducemax) + - [ReduceAny](#reduceany) + - [ReduceAll](#reduceall) + - [ReduceProd](#reduceprod) + - [ReduceMean](#reducemean) + - [ReduceSum](#reducesum) + - [Greater](#greater) + - [GreaterOrEqual](#greaterorequal) + - [Less](#less) + - [LessOrEqual](#lessorequal) + - [NotEqual](#notequal) + - [Equal](#equal) + - [Reorg](#reorg) + - [Reshape](#reshape) + - [Resize](#resize) + - [Resize1d](#resize1d) + - [Reverse](#reverse) + - [ScatterND](#scatternd) + - [Select](#select) + - [DataConvert](#dataconvert) + - [Neg](#neg) + - [Abs](#abs) + - [Sin](#sin) + - [Exp](#exp) + - [Log](#log) + - [Sqrt](#sqrt) + - [Rsqrt](#rsqrt) + - [Square](#square) + - [LogicalNot](#logicalnot) + - [Floor](#floor) + - [Cast](#cast) + - [Slice](#slice) + - [Softmax](#softmax) + - [Space2Batch](#space2batch) + - [SpaceToDepth](#spacetodepth) + - [Split](#split) + - [Squeeze](#squeeze) + - [Stack](#stack) + - [StridedSlice](#stridedslice) + - [Tile](#tile) + - [Transpose](#transpose) + - [Unidirectional sequence lstm](#unidirectional-sequence-lstm) + - [Unstack](#unstack) + + +# Operators + + +## Activation + +Activation functions: + +``` +Relu(x) : max(0, x) + +Relu1(x) : -1 if x <= -1; x if -1 < x < 1; 1 if x >= 1 + +Relu6(x) : 0 if x <= 0; x if 0 < x < 6; 6 if x >= 6 + +Elu(x) : x if x >= 0 else alpha*(e^x - 1) + +Tanh(x) : (1 - e^{-2x})/(1 + e^{-2x}) + +Sigmoid(x) : 1/(1 + e^{-x}) + +Swish(x) : x * sigmoid(x) + +HardSwish(x) : 0 if x <= -3; x(x + 3)/6 if -3 < x < 3; x if x >= 3 + +Mish(x) : x if x >= 0 else alpha * x + +HardSigmoid(x) : min(max(alpha*x + beta, 0), 1) + +SoftRelu(x) : log(1 + e^x). Also known as SoftPlus. + +LeakyRelu(x) : alpha * x if x <= 0; x if x > 0. alpha is a scalar. + +Prelu(x) : alpha * x if x <= 0; x if x > 0. alpha is a tensor. +- axis : describes the axis of the inputs when coerced to 2D. + +Linear(x, a, b) : a*x + b. + +Gelu(x) : x * P(X <= x), where P(x) ~ N(0, 1). https://tensorflow.google.cn/api_docs/python/tf/nn/gelu +``` + + +## AddN + +``` +AddN(x) : Input0 + Input1 + ... + InputN +``` + + +## ArgMin/ArgMax + +Computes the indices of the **min/max** elements of the input tensor's element +along the provided **axis**. The type of the output tensor is integer. + + +## Batch2Space + +This operation reshapes the batch dimension (dimension 0) into M + 1 dimensions +of shape **block_size** + [batch], interleaves these blocks back into the grid +defined by the spatial dimensions [1, ..., M], to obtain a result with the same +rank as the input. This is the reverse transformation of Space2Batch. + +- crop : corp the output tensor for ROI usage. + + +## BatchNorm + +Carries out batch normalization as described in the paper +https://arxiv.org/abs/1502.03167. + +$$\hat x_i\leftarrow \frac{x_i-\mu_\mathcal{B}}{\sqrt{\sigma_\mathcal{B}^2+\epsilon}}$$ + +$$y_i=\gamma\hat x_i+\beta\equiv BN_{\gamma,\beta}(x_i)$$ + + +## Clip + +Clip(x) : min if x <= min; x if min < x < max; max if x >= max + + +## Concat + +Concatenate a list of tensors into a single tensor. + +- axis : Which axis to concat on. + + +## Conv2d + +Performs a 2-D convolution operation, include classic Conv2D / +Depthwise Conv2D / Group Conv2D / Dilation Conv2D. + +Input: +- input [WHCN or CWHN]. +- kernel [ WHIcOc ] (Ic: Input Channels. Oc: Output Channels). +- bias [ O ]. Optional. + +Attribute: +- weights : the output channel number for weight tensor. +- ksize : the height and width for weight tensor. +- padding : AUTO, VALID or SAME. +- pad : pad value for each spatial axis. +- stride : stride along each spatial axis. +- dilation : dilation value along each spatial axis of the filter. +- multiplier: function similar to group attribute on other framework, +but the value is different. multiplier = weights / group. +- layout : WHCN or CWHN. + + +## DeConv2d + +Performs the transpose of 2-D convolution operation. + +This operation is sometimes called "deconvolution" after Deconvolutional Networks, +but is actually the transpose (gradient) of Conv2D rather than an actual deconvolution. + +- oc_count_ : the out channel count for weight tensor. +- pad_type : SAME, VALID or AUTO. +- ksize : the height and width for weight tensor. +- padding : AUTO, VALID or SAME. +- pad : pad value for each spatial axis. +- stride : stride along each spatial axis. +- output_padding : specifying the amount of padding along the height and width of +the output tensor. +- group : the feature count of each group. +- input_layout : Layout for input, WHCN by default. +- kernel_layout: Layout for kernel, WHIO by default. + + +## DeConv1d + +Performs the transpose of 1-D convolution operation. + +This operation is sometimes called "deconvolution1d" after Deconvolutional Networks, +but is actually the transpose (gradient) of Conv2D rather than an actual deconvolution. + +- weights : the channel number for weight tensor. +- ksize : the length for weight tensor. +- padding : AUTO, VALID or SAME. +- pad : pad value for each spatial axis. +- stride : stride along each spatial axis. +- output_padding : specifying the amount of padding along the height and width of +the output tensor. + + +## DepthToSpace + +DepthToSpace rearranges (permutes) data from depth into blocks of spatial data. +This is the reverse transformation of SpaceToDepth. + +Chunks of data of size block_size * block_size from depth are rearranged into +non-overlapping blocks of size block_size x block_size. + +The width of the output tensor is input_depth * block_size, whereas the height +is input_height * block_size. The depth of the input tensor must be divisible +by block_size * block_size + +- crop : corp the output tensor for ROI usage. + + +## Dropout + +The Dropout layer randomly sets input units to 0 with a frequency of rate at +each step during training time, which helps prevent overfitting. + +TIM-VX only focus on inference time, and just scaling input tensor by **ratio** +for Dropout operator. + + +## Add + +Add(x, y) : x + y. This operation supports broadcasting. + + +## Sub + +Sub(x, y) : x - y. This operation supports broadcasting. + + +## Multiply + +Multiply(x, y) : Multiplies two tensors, element-wise, also known as Hadamard +product. This operation supports broadcasting. + +- scale: scaling the product. + + +## Div + +Div(x, y) : x / y. This operation supports broadcasting. + + +## Pow + +Pow(x, y) : x ^ y. This operation supports broadcasting. + + +## Minimum + +Minimum(x, y) : min(x, y). This operation supports broadcasting. + + +## Maximum + +Maximum(x, y) : max(x, y). This operation supports broadcasting. + + +## FloorDiv + +FloorDiv(x, y): floor( x / y ). This operation supports broadcasting. + + +## FullyConnected + +Denotes a fully (densely) connected layer, which connects all elements in the +input tensor with each element in the output tensor. + +- axis: Describes the axis of the inputs when coerced to 2D. +- weights: the output channel number for weight tensor. + + +## Gather + +Gather slices from input, **axis** according to **indices**. + + +## GatherNd + +An operation similar to Gather but gathers across multiple axis at once. + + +## GroupedConv2d + +Performs a grouped 2-D convolution operation. + +Input: +- input [WHCN or CWHN]. +- kernel [ WHIcOc ] (Ic: Input Channels. Oc: Output Channels). +- bias [ O ]. Optional. + +Attribute: +- weights : the output channel number for weight tensor. +- ksize : the height and width for weight tensor. +- padding : AUTO, VALID or SAME. +- pad : pad value for each spatial axis. +- stride : stride along each spatial axis. +- dilation : dilation value along each spatial axis of the filter. +- group_number: Split conv to n group. +- layout : WHCN or CWHN. + + +## L2Normalization + +Applies L2 normalization along the axis dimension: + +``` +output[batch, row, col, channel] = +input[batch, row, col, channel] / +sqrt(sum_{c} pow(input[batch, row, col, c], 2)) +``` + + +## LocalResponseNormalization + +Applies Local Response Normalization along the depth dimension: + +``` +sqr_sum[a, b, c, d] = sum( +pow(input[a, b, c, d - depth_radius : d + depth_radius + 1], 2)) +output = input / pow((bias + alpha * sqr_sum), beta) +``` + + +## And + +Returns the truth value of x AND y element-wise. This operation supports broadcasting. + + +## Or + +Returns the truth value of x OR y element-wise. This operation supports broadcasting. + + +## LogSoftmax + +Computes the log softmax activation on the input tensor element-wise, per batch. + +``` +logsoftmax = logits - log(reduce_sum(exp(logits), axis)) +``` + + +## Matmul + +Multiplies matrix a by matrix b, producing a * b. + +- transpose_a: If True, a is transposed before multiplication. +- transpose_b: If True, b is transposed before multiplication. +- adjoint_a: If True, a is conjugated and transposed before multiplication. +- adjoint_b: If True, b is conjugated and transposed before multiplication. + + +## MaxpoolWithArgmax + +Performs an 2-D Max pooling operation and return indices + +- padding : AUTO, VALID or SAME. +- ksize : filter size. +- stride : stride along each spatial axis. +- round_type : CEILING or FLOOR. + + +## MaxUnpool2d + +Performs an 2-D Max pooling operation upsample + +- stride : stride along each spatial axis. +- ksize : filter size. + + +## Moments + +The mean and variance are calculated by aggregating the contents of x across axes. +If x is 1-D and axes = [0] this is just the mean and variance of a vector. + +- axes : Axes along which to compute mean and variance. +- keep_dims : Produce moments with the same dimensionality as input. + + +## NBG + +Network Binary Graph is a precompile technology, which can compile a fuse graph into +a bianry file. + + +## Pad + +Pads a tensor. + +- const_val : the value to pad. + + +## Pool2d + +Performs an 2-D pooling operation. + +- type : MAX, AVG, L2 or AVG_ANDROID. +- padding : AUTO, VALID or SAME. +- ksize : filter size. +- stride : stride along each spatial axis. +- round_type : CEILING or FLOOR. + + +## ReduceMin + +Reduces a tensor by computing the minimum of elements along given dimensions. + +- axis : the dimensions to reduce. +- keep_dims : If keep_dims is true, the reduced dimensions are retained with +length 1. Otherwise, the rank of the tensor is reduced by 1 for each entry +in dimensions + + +## ReduceMax + +Reduces a tensor by computing the maximum of elements along given dimensions. + +- axis : the dimensions to reduce. +- keep_dims : If keep_dims is true, the reduced dimensions are retained with +length 1. Otherwise, the rank of the tensor is reduced by 1 for each entry +in dimensions + + +## ReduceAny + +Reduces a tensor by computing the "logical or" of elements along given dimensions. + +- axis : the dimensions to reduce. +- keep_dims : If keep_dims is true, the reduced dimensions are retained with +length 1. Otherwise, the rank of the tensor is reduced by 1 for each entry +in dimensions + + +## ReduceAll + +Reduces a tensor by computing the "logical and" of elements along given dimensions. + +- axis : the dimensions to reduce. +- keep_dims : If keep_dims is true, the reduced dimensions are retained with +length 1. Otherwise, the rank of the tensor is reduced by 1 for each entry +in dimensions + + +## ReduceProd + +Reduces a tensor by computing the multiplying of elements along given dimensions. + +- axis : the dimensions to reduce. +- keep_dims : If keep_dims is true, the reduced dimensions are retained with +length 1. Otherwise, the rank of the tensor is reduced by 1 for each entry +in dimensions + + +## ReduceMean + +Reduces a tensor by computing the mean of elements along given dimensions. + +- axis : the dimensions to reduce. +- keep_dims : If keep_dims is true, the reduced dimensions are retained with +length 1. Otherwise, the rank of the tensor is reduced by 1 for each entry +in dimensions + + +## ReduceSum + +Reduces a tensor by computing the summing of elements along given dimensions. + +- axis : the dimensions to reduce. +- keep_dims : If keep_dims is true, the reduced dimensions are retained with +length 1. Otherwise, the rank of the tensor is reduced by 1 for each entry +in dimensions + + +## Greater + +For input tensors x and y, computes x > y elementwise. + + +## GreaterOrEqual + +For input tensors x and y, computes x >= y elementwise. + + +## Less + +For input tensors x and y, computes x < y elementwise. + + +## LessOrEqual + +For input tensors x and y, computes x <= y elementwise. + + +## NotEqual + +For input tensors x and y, computes x != y elementwise. + + +## Equal + +For input tensors x and y, computes x == y elementwise. + + +## Reorg + +The layer used in YOLOv2. See also https://github.com/pjreddie/darknet/blob/master/src/reorg_layer.c + + +## Reshape + +Given tensor, this operation returns a tensor that has the same values as tensor, but with a newly specified shape. + +- size : defining the shape of the output tensor. + + +## Resize + +Resizes images to given size. + +- type : NEAREST_NEIGHBOR, BILINEAR or AREA. +- factor : scale the input size. DO NOT use it with target_height / target_width together. +- align_corners : If True, the centers of the 4 corner pixels of the input and output +tensors are aligned, preserving the values at the corner pixels. +- half_pixel_centers : If True, the pixel centers are assumed to be at (0.5, 0.5). +This is the default behavior of image.resize in TF 2.0. If this parameter is True, +then align_corners parameter must be False. +- target_height / target_width : output height / width. DO NOT use it with factor together. + + +## Resize1d + +Resize1ds 1D tensors to given size. + +- type : NEAREST_NEIGHBOR, BILINEAR or AREA. +- factor : scale the input size. DO NOT use it with target_height / target_width together. +- align_corners : If True, the centers of the 4 corner pixels of the input and output +tensors are aligned, preserving the values at the corner pixels. +- half_pixel_centers : If True, the pixel centers are assumed to be at (0.5, 0.5). +This is the default behavior of image.resize in TF 2.0. If this parameter is True, +then align_corners parameter must be False. +- target_height / target_width : output height / width. DO NOT use it with factor together. + + +## Reverse + +Reverses specific dimensions of a tensor. + +- axis : The indices of the dimensions to reverse. + + +## ScatterND + +Scatter updates into a new tensor according to indices. + +- shape : The shape of the resulting tensor. + + +## Select + +Using a tensor of booleans c and input tensors x and y select values elementwise +from both input tensors: O[i] = C[i] ? x[i] : y[i]. + + +## DataConvert + +Change the format from input tensor to output tensor. + + +## Neg + +Neg(x) : -x + + +## Abs + +Abs(x) : x if x >= 0; -x if x < 0. + + +## Sin + +Sin(x) : sin(x) + + +## Exp + +Exp(x) : e^x + + +## Log + +Log(x) : ln(x) + + +## Sqrt + +Sqrt(x) : $$\sqrt{x}$$ + + +## Rsqrt + +Rsqrt(x) : $$\frac{1}{\sqrt{x}}$$ + + +## Square + +Square : x^2 + + +## LogicalNot + +LogicalNot(x) : NOT x + + +## Floor + +returns the largest integer less than or equal to a given number. + + +## Cast + +Change the format from input tensor to output tensor. This operation ignores +the scale and zeroPoint of quanized tensors. + + +## Slice + +Extracts a slice of specified size from the input tensor starting at a specified location. + +- start : the beginning indices of the slice in each dimension. +- length : the size of the slice in each dimension. + + +## Softmax + +Computes the softmax activation on the input tensor element-wise, per batch, +by normalizing the input vector so the maximum coefficient is zero: + +``` +output[batch, i] = +exp((input[batch, i] - max(input[batch, :])) * beta) / +sum_{k}{exp((input[batch, k] - max(input[batch, :])) * beta)} +``` + + +## Space2Batch + +This operation divides "spatial" dimensions [1, ..., M] of the input into a grid +of blocks of shape **block_size**, and interleaves these blocks with the "batch" +dimension (0) such that in the output, the spatial dimensions [1, ..., M] correspond +to the position within the grid, and the batch dimension combines both the position +within a spatial block and the original batch position. Prior to division into blocks, +the spatial dimensions of the input are optionally zero padded according to paddings. +This is the reverse transformation of Batch2Space. + +- pad : the paddings for each spatial dimension of the input tensor. + + +## SpaceToDepth + +SpaceToDepth rearranges blocks of spatial data into depth. More specifically, +this op outputs a copy of the input tensor where values from the height and +width dimensions are moved to the depth dimension. This is the reverse +transformation of DepthToSpace. + + +## Split + +Splits a tensor along a given axis into num_splits subtensors. + +- axis : the axis along which to split. +- slices : indicating the number of splits along given axis. + + +## Squeeze + +Removes dimensions of size 1 from the shape of a tensor. + +- axis : the dimensions to squeeze. + + +## Stack + +Packs the list of tensors in inputs into a tensor with rank one higher than +each tensor in values, by packing them along the **axis** dimension. + + +## StridedSlice + +Extracts a strided slice of a tensor. + +Roughly speaking, this op extracts a slice of size (end - begin) / stride from +the given input tensor. Starting at the location specified by begin the slice +continues by adding stride to the index until all dimensions are not less than end. +Note that a stride can be negative, which causes a reverse slice. + +- begin_dims : the starts of the dimensions of the input tensor to be sliced. +- end_dims : the ends of the dimensions of the input tensor to be sliced. +- stride_dims : the strides of the dimensions of the input tensor to be sliced. +- begin_mask : if the ith bit of begin_mask is set, begin[i] is ignored and +the fullest possible range in that dimension is used instead. +- end_mask : if the ith bit of end_mask is set, end[i] is ignored and the fullest +possible range in that dimension is used instead. +- shrink_axis_mask : if the ith bit of shrink_axis_mask is set, the ith dimension +specification shrinks the dimensionality by 1, taking on the value at index begin[i]. +In this case, the ith specification must define a slice of size 1, +e.g. begin[i] = x, end[i] = x + 1. + + +## Tile + +Constructs a tensor by tiling a given tensor. +- multiples : Must be one of the following types: int32, int64. +Length must be the same as the number of dimensions in input. + + +## Transpose + +Transposes the input tensor, permuting the dimensions according to the +**perm** tensor. + +The returned tensor's dimension i corresponds to the input dimension perm[i]. +If perm is not given, it is set to (n-1...0), where n is the rank of the input +tensor. Hence by default, this operation performs a regular matrix transpose on +2-D input Tensors. + + +## Unidirectional sequence lstm +how to bind input/output: take unidirectional_sequence_lstm_test.cc + + +## Unstack + +Unpacks the given dimension of a rank-R tensor into rank-(R-1) tensors. +- axis : An int. The axis to unstack along. Defaults to the first dimension. +Negative values wrap around, so the valid range is [-R, R). \ No newline at end of file diff --git a/docs/Operators.md.template b/docs/Operators.md.template index e99ba27..8b12ab2 100644 --- a/docs/Operators.md.template +++ b/docs/Operators.md.template @@ -1,23 +1,3 @@ -# Memory Layout - -There are two different memory layout mode: - -- Row-Major. Also called C contiguous, which is firstly introduced in C by AT&T Bell Lab(1960s). Caffe/TensorFlow/Pytorch follow the Row-Major layout mode. - -- Column-Major. Also called Fortran contiguous, which is firstly introduced in Fortran by IBM(1958). Matlab/OpenGL/OpenCL/OpenVX follow the Column-Major layout mode. - -See also : - -http://eigen.tuxfamily.org/dox/group__TopicStorageOrders.html - -On the other hand, a memory layout can be described by both Row-Major and Column-Major mode. - -It easily tranlate the layout mode from one to another: just reverse the dimesions of the tensor. - -For example, TensorFlow usually describe the tensor as NHWC format. This is equal the CWHN format in OpenVX. - -Likewise, the WHCN format in OpenVX is equal the NCHW format in Caffe. - # Operators {DOCS} diff --git a/docs/Programming_Guide.md b/docs/Programming_Guide.md new file mode 100644 index 0000000..c209ae7 --- /dev/null +++ b/docs/Programming_Guide.md @@ -0,0 +1,83 @@ +# Basic Concepts + +## Context + +The OpenVX context is the object domain for all OpenVX objects. All data objects live in the context as well as all framework objects. The OpenVX context keeps reference counts on all objects and must do garbage collection during its deconstruction to free lost references. + +## Node + +A node is an instance of a kernel that will be paired with a specific set of references (the parameters). Nodes are created from and associated with a single graph only. + +**Node is also known as Operation.** + +## Graph + +A set of nodes connected in a directed (only goes one-way) acyclic (does not loop back) fashion. A Graph may have sets of Nodes that are unconnected to other sets of Nodes within the same Graph. + +**One context may have several graphs.** + +## Tensor + +An multidimensional data object. + +TIM-VX support the following tensor type: + +- Normal Tensor. The tensor which is allocated in host memory, which is managed by OpenVX driver. Normal Tensor is usually used for input/output tensors of the graph. +- Viratul Tensor. The tensor which is allocated in target memory. Viratul Tensor is usually used for intermediate tensors of the graph. Viratul Tensor is faster than the Normal Tensor due to reduce the memory copy between host and target. +- Handle Tensor. The tensor which is allocated in host memory, which is managed by user application. App usually use mmap() to allocate continuous memory for tensor usage. Handle Tensor is also usually used for input/output tensors of the graph. Sometimes the memory of Normal Tensor will be released/relocated by OpenVX driver, and this may be inefficient. You can use Handle Tensor to avoid release/relocate memory. + +## Memory Layout + +There are two different memory layout mode: + +- Row-Major. Also called C contiguous, which is firstly introduced in C by AT&T Bell Lab(1960s). Caffe/TensorFlow/Pytorch follow the Row-Major layout mode. + +- Column-Major. Also called Fortran contiguous, which is firstly introduced in Fortran by IBM(1958). Matlab/OpenGL/OpenCL/OpenVX follow the Column-Major layout mode. + +See also : + +http://eigen.tuxfamily.org/dox/group__TopicStorageOrders.html + +On the other hand, a memory layout can be described by both Row-Major and Column-Major mode. + +It easily tranlate the layout mode from one to another: just reverse the dimesions of the tensor. + +For example, TensorFlow usually describe the tensor as NHWC format. This is equal the CWHN format in OpenVX. + +Likewise, the WHCN format in OpenVX is equal the NCHW format in Caffe. + +## Data Types + +TIM-VX support the following data type: + +- INT8 +- INT16 +- INT32 +- INT64 +- UINT8 +- UINT16 +- UINT32 +- UINT64 +- FLOAT16 +- FLOAT32 +- FLOAT64 + +# Graph lifecycle + +![Graph lifecycle](/docs/image/graph_lifecycle.png) + +- Graph Construction: Create Node/Tensor/Graph. + +- Graph Verification: The graphs are checked for consistency, correctness, and other conditions.Memory allocation may occur. + +- Graph Execution: Between executions data may be updated by the client or some other external mechanism. + +- Graph Deconstruction: Release Node/Tensor/Graph. + +# Software architecture + +![Software architecture](/docs/image/architecture.png) + +- Ovxlib: the C wrapper for OpenVX driver. + +- TIM-VX: the C++ wrapper for OpenVX driver. diff --git a/docs/image/architecture.png b/docs/image/architecture.png new file mode 100644 index 0000000..dcd1d62 Binary files /dev/null and b/docs/image/architecture.png differ diff --git a/docs/image/graph_lifecycle.png b/docs/image/graph_lifecycle.png new file mode 100644 index 0000000..d12799f Binary files /dev/null and b/docs/image/graph_lifecycle.png differ diff --git a/include/tim/vx/ops/activations.h b/include/tim/vx/ops/activations.h index 2eadede..8b0366a 100644 --- a/include/tim/vx/ops/activations.h +++ b/include/tim/vx/ops/activations.h @@ -65,7 +65,6 @@ namespace ops { * Linear(x, a, b) : a*x + b. * * Gelu(x) : x * P(X <= x), where P(x) ~ N(0, 1). https://tensorflow.google.cn/api_docs/python/tf/nn/gelu - * ``` */