Support layout inference for transpose (#250)
Added interface GetProdeucerOp(tensor) in Graph Signed-off-by: yuenan.li <yuenan.li@verisilicon.com>
This commit is contained in:
parent
aed3a48248
commit
75d39e2cfd
|
|
@ -69,9 +69,16 @@ class Graph {
|
||||||
const std::shared_ptr<Tensor>& tensor,
|
const std::shared_ptr<Tensor>& tensor,
|
||||||
const Operation* op) = 0;
|
const Operation* op) = 0;
|
||||||
|
|
||||||
|
virtual void UpdateTensorProducerMap(
|
||||||
|
const std::shared_ptr<Tensor>& tensor,
|
||||||
|
const Operation* op) = 0;
|
||||||
|
|
||||||
virtual const std::vector<std::shared_ptr<Operation>> GetConsumersOp(
|
virtual const std::vector<std::shared_ptr<Operation>> GetConsumersOp(
|
||||||
std::shared_ptr<Tensor> tensor) const = 0;
|
std::shared_ptr<Tensor> tensor) const = 0;
|
||||||
|
|
||||||
|
virtual std::vector<std::shared_ptr<Operation>> GetProducerOp(
|
||||||
|
std::shared_ptr<Tensor> tensor) = 0;
|
||||||
|
|
||||||
virtual void PrintGraph() const = 0;
|
virtual void PrintGraph() const = 0;
|
||||||
|
|
||||||
protected:
|
protected:
|
||||||
|
|
|
||||||
|
|
@ -59,6 +59,7 @@
|
||||||
#include "ops/deconv2d_layout_inference.h"
|
#include "ops/deconv2d_layout_inference.h"
|
||||||
#include "ops/batchnorm_layout_inference.h"
|
#include "ops/batchnorm_layout_inference.h"
|
||||||
#include "ops/default_layout_inference.h"
|
#include "ops/default_layout_inference.h"
|
||||||
|
#include "ops/transpose_layout_inference.h"
|
||||||
|
|
||||||
#include <algorithm>
|
#include <algorithm>
|
||||||
#include <deque>
|
#include <deque>
|
||||||
|
|
@ -257,6 +258,7 @@ std::vector<std::shared_ptr<vx::Tensor>> HandleLayoutInfer(
|
||||||
REGIST_LAYOUT_INFERENCE(VSI_NN_OP_ARGMIN, Arg);
|
REGIST_LAYOUT_INFERENCE(VSI_NN_OP_ARGMIN, Arg);
|
||||||
REGIST_LAYOUT_INFERENCE(VSI_NN_OP_DECONVOLUTION, DeConv2d);
|
REGIST_LAYOUT_INFERENCE(VSI_NN_OP_DECONVOLUTION, DeConv2d);
|
||||||
REGIST_LAYOUT_INFERENCE(VSI_NN_OP_BATCH_NORM, BatchNorm);
|
REGIST_LAYOUT_INFERENCE(VSI_NN_OP_BATCH_NORM, BatchNorm);
|
||||||
|
REGIST_LAYOUT_INFERENCE(VSI_NN_OP_PERMUTE, Transpose);
|
||||||
REGIST_LOGICAL_LAYOUT_INFERENCE(VSI_NN_OP_LOGICAL_OPS);
|
REGIST_LOGICAL_LAYOUT_INFERENCE(VSI_NN_OP_LOGICAL_OPS);
|
||||||
REGIST_REDUCE_LAYOUT_INFERENCE(VSI_NN_OP_REDUCE);
|
REGIST_REDUCE_LAYOUT_INFERENCE(VSI_NN_OP_REDUCE);
|
||||||
// use default layout inference
|
// use default layout inference
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,76 @@
|
||||||
|
/****************************************************************************
|
||||||
|
*
|
||||||
|
* Copyright (c) 2020 Vivante Corporation
|
||||||
|
*
|
||||||
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||||
|
* copy of this software and associated documentation files (the "Software"),
|
||||||
|
* to deal in the Software without restriction, including without limitation
|
||||||
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
||||||
|
* and/or sell copies of the Software, and to permit persons to whom the
|
||||||
|
* Software is furnished to do so, subject to the following conditions:
|
||||||
|
*
|
||||||
|
* The above copyright notice and this permission notice shall be included in
|
||||||
|
* all copies or substantial portions of the Software.
|
||||||
|
*
|
||||||
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||||
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||||
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
||||||
|
* AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||||
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
||||||
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
||||||
|
* DEALINGS IN THE SOFTWARE.
|
||||||
|
*
|
||||||
|
*****************************************************************************/
|
||||||
|
#ifndef TIM_LAYOUT_INFER_TRANSPOSE_LAYOUT_INFERENCE_H_
|
||||||
|
#define TIM_LAYOUT_INFER_TRANSPOSE_LAYOUT_INFERENCE_H_
|
||||||
|
|
||||||
|
#include "tim/vx/ops/transpose.h"
|
||||||
|
|
||||||
|
#include "ops/op_layout_inference.h"
|
||||||
|
#include "permute_vector.h"
|
||||||
|
#include "operation_private.h"
|
||||||
|
|
||||||
|
namespace tim {
|
||||||
|
namespace transform {
|
||||||
|
class TransposeLayoutInfer : public OpLayoutInfer {
|
||||||
|
public:
|
||||||
|
TransposeLayoutInfer(
|
||||||
|
const std::shared_ptr<vx::Operation> op,
|
||||||
|
std::shared_ptr<layout_inference_impl::LayoutInferContext>& context)
|
||||||
|
: OpLayoutInfer(op, context) {}
|
||||||
|
|
||||||
|
void OnInputs(
|
||||||
|
std::vector<std::shared_ptr<vx::Tensor>>& next_tensors) override {
|
||||||
|
auto src_input = op_->impl()->InputsTensor()[0];
|
||||||
|
auto infer_input = context_->GetMapedTensor(src_input);
|
||||||
|
auto input_pv = context_->GetPermuteVector(src_input);
|
||||||
|
|
||||||
|
std::vector<uint32_t> perm(op_->impl()->node()->nn_param.permute.dim_num);
|
||||||
|
memcpy(perm.data(), op_->impl()->node()->nn_param.permute.perm,
|
||||||
|
op_->impl()->node()->nn_param.permute.dim_num * sizeof(uint32_t));
|
||||||
|
IPermuteVectorPtr perm_pv = MakeShared(perm.size());
|
||||||
|
for (uint32_t i = 0; i < perm.size(); i++) {
|
||||||
|
perm_pv->At(i) = perm[i];
|
||||||
|
}
|
||||||
|
|
||||||
|
IPermuteVectorPtr final_pv = input_pv->Reverse()->Add(perm_pv);
|
||||||
|
|
||||||
|
if (final_pv->IsAligned()) {
|
||||||
|
//skip transpose op by treating its input as its output.
|
||||||
|
context_->UpdateTensorMap(op_->impl()->OutputsTensor()[0], infer_input);
|
||||||
|
} else {
|
||||||
|
auto transpose_op =
|
||||||
|
context_->infer_graph_->CreateOperation<tim::vx::ops::Transpose>(
|
||||||
|
final_pv->AsStdVec());
|
||||||
|
transpose_op->BindInput(infer_input);
|
||||||
|
auto infer_out = CreateOutputsTensor(final_pv);
|
||||||
|
transpose_op->BindOutput(infer_out[0]);
|
||||||
|
}
|
||||||
|
context_->SetPermuteVector(op_->impl()->OutputsTensor()[0], MakeShared(perm.size()));
|
||||||
|
next_tensors.push_back(op_->impl()->OutputsTensor()[0]);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
} // namespace transform
|
||||||
|
} // namespace tim
|
||||||
|
#endif
|
||||||
|
|
@ -87,6 +87,15 @@ void GraphImpl::UpdateTensorConsumersMap(const std::shared_ptr<Tensor>& tensor,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void GraphImpl::UpdateTensorProducerMap(const std::shared_ptr<Tensor>& tensor,
|
||||||
|
const Operation* op) {
|
||||||
|
for (const auto& added_op : op_vector_) {
|
||||||
|
if (added_op.get() == op) {
|
||||||
|
tensor_producer_[tensor].push_back(added_op);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
const std::vector<std::shared_ptr<Operation>> GraphImpl::GetConsumersOp(
|
const std::vector<std::shared_ptr<Operation>> GraphImpl::GetConsumersOp(
|
||||||
std::shared_ptr<Tensor> tensor) const {
|
std::shared_ptr<Tensor> tensor) const {
|
||||||
auto consumers = tensor_consumers_.find(tensor);
|
auto consumers = tensor_consumers_.find(tensor);
|
||||||
|
|
@ -98,6 +107,17 @@ const std::vector<std::shared_ptr<Operation>> GraphImpl::GetConsumersOp(
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
std::vector<std::shared_ptr<Operation>> GraphImpl::GetProducerOp(
|
||||||
|
std::shared_ptr<Tensor> tensor) {
|
||||||
|
auto producer = tensor_producer_.find(tensor);
|
||||||
|
if (tensor_producer_.end() != producer) {
|
||||||
|
return producer->second;
|
||||||
|
} else {
|
||||||
|
VSILOGD("Tensor has no producer, may be graph input.");
|
||||||
|
return {};
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
void GraphImpl::PrintGraph() const { vsi_nn_PrintGraph(this->graph_); }
|
void GraphImpl::PrintGraph() const { vsi_nn_PrintGraph(this->graph_); }
|
||||||
|
|
||||||
std::shared_ptr<Tensor> GraphImpl::CreateTensor(const TensorSpec& spec,
|
std::shared_ptr<Tensor> GraphImpl::CreateTensor(const TensorSpec& spec,
|
||||||
|
|
|
||||||
|
|
@ -56,19 +56,24 @@ class GraphImpl : public Graph {
|
||||||
|
|
||||||
void UpdateTensorConsumersMap(const std::shared_ptr<Tensor>& tensor,
|
void UpdateTensorConsumersMap(const std::shared_ptr<Tensor>& tensor,
|
||||||
const Operation* op) override;
|
const Operation* op) override;
|
||||||
|
void UpdateTensorProducerMap(const std::shared_ptr<Tensor>& tensor,
|
||||||
|
const Operation* op) override;
|
||||||
const std::vector<std::shared_ptr<Operation>> GetConsumersOp(
|
const std::vector<std::shared_ptr<Operation>> GetConsumersOp(
|
||||||
std::shared_ptr<Tensor> tensor) const override;
|
std::shared_ptr<Tensor> tensor) const override;
|
||||||
void PrintGraph() const override;
|
std::vector<std::shared_ptr<Operation>> GetProducerOp(
|
||||||
/// Implement parents' virtual functions
|
std::shared_ptr<Tensor> tensor) override;
|
||||||
std::shared_ptr<Tensor> CreateTensor(const TensorSpec& spec,
|
|
||||||
const void* data = nullptr) override;
|
|
||||||
std::shared_ptr<Tensor> CreateTensor(const TensorSpec& spec,
|
|
||||||
const DmaBufferDesc& dmafd) override;
|
|
||||||
std::shared_ptr<Tensor> CreateTensorPlaceHolder() override;
|
|
||||||
bool Compile() override;
|
|
||||||
|
|
||||||
bool CompileToBinary(void* buf, size_t* size) override;
|
void PrintGraph() const override;
|
||||||
bool Run() override;
|
|
||||||
|
std::shared_ptr<Tensor> CreateTensor(const TensorSpec& spec,
|
||||||
|
const void* data = nullptr) override;
|
||||||
|
std::shared_ptr<Tensor> CreateTensor(const TensorSpec& spec,
|
||||||
|
const DmaBufferDesc& dmafd) override;
|
||||||
|
std::shared_ptr<Tensor> CreateTensorPlaceHolder() override;
|
||||||
|
|
||||||
|
bool Compile() override;
|
||||||
|
bool CompileToBinary(void* buf, size_t* size) override;
|
||||||
|
bool Run() override;
|
||||||
|
|
||||||
protected:
|
protected:
|
||||||
ContextImpl* context_;
|
ContextImpl* context_;
|
||||||
|
|
@ -82,6 +87,7 @@ class GraphImpl : public Graph {
|
||||||
std::vector<std::shared_ptr<Tensor>> inputs_tensor_;
|
std::vector<std::shared_ptr<Tensor>> inputs_tensor_;
|
||||||
std::vector<std::shared_ptr<Tensor>> outputs_tensor_;
|
std::vector<std::shared_ptr<Tensor>> outputs_tensor_;
|
||||||
std::map<std::shared_ptr<Tensor>, std::vector<std::shared_ptr<Operation>>> tensor_consumers_;
|
std::map<std::shared_ptr<Tensor>, std::vector<std::shared_ptr<Operation>>> tensor_consumers_;
|
||||||
|
std::map<std::shared_ptr<Tensor>, std::vector<std::shared_ptr<Operation>>> tensor_producer_;
|
||||||
};
|
};
|
||||||
|
|
||||||
} // namespace vx
|
} // namespace vx
|
||||||
|
|
|
||||||
|
|
@ -100,6 +100,7 @@ Operation& Operation::BindInput(const std::shared_ptr<Tensor>& tensor) {
|
||||||
|
|
||||||
Operation& Operation::BindOutput(const std::shared_ptr<Tensor>& tensor) {
|
Operation& Operation::BindOutput(const std::shared_ptr<Tensor>& tensor) {
|
||||||
impl_->BindOutput(tensor);
|
impl_->BindOutput(tensor);
|
||||||
|
impl_->graph_->UpdateTensorProducerMap(tensor, this);
|
||||||
return *this;
|
return *this;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue