From 2ef0f6785954681624d8eda5385601279812b067 Mon Sep 17 00:00:00 2001 From: Tian Jin Date: Tue, 24 Dec 2019 03:43:47 -0500 Subject: [PATCH] revert unnecessary changes --- test-tanh.mlir | 4 ---- test/mlir/onnx/onnx_lowering.mlir | 38 +++++++++++++++---------------- 2 files changed, 19 insertions(+), 23 deletions(-) delete mode 100644 test-tanh.mlir diff --git a/test-tanh.mlir b/test-tanh.mlir deleted file mode 100644 index 53395ed..0000000 --- a/test-tanh.mlir +++ /dev/null @@ -1,4 +0,0 @@ -func @test_tanh(%arg0 : tensor) -> tensor<*xf32> { - %0 = "onnx.Tanh"(%arg0) : (tensor) -> tensor<*xf32> - "std.return"(%0) : (tensor<*xf32>) -> () -} \ No newline at end of file diff --git a/test/mlir/onnx/onnx_lowering.mlir b/test/mlir/onnx/onnx_lowering.mlir index 7b19d72..92d4a0f 100644 --- a/test/mlir/onnx/onnx_lowering.mlir +++ b/test/mlir/onnx/onnx_lowering.mlir @@ -149,25 +149,25 @@ func @test_tanh(%arg0 : tensor) -> tensor<*xf32> { %0 = "onnx.Tanh"(%arg0) : (tensor) -> tensor<*xf32> "std.return"(%0) : (tensor<*xf32>) -> () -// // CHECK-LABEL: test_tanh -// // CHECK: [[DIM_0:%.+]] = dim %arg0, 0 : memref -// // CHECK: [[RES:%.+]] = alloc([[DIM_0]]) : memref -// // CHECK: [[DEF_LOOPS:%.+]]:2 = krnl.define_loops 2 -// // CHECK: [[OPT_LOOPS:%.+]]:2 = krnl.optimize_loops { -// // CHECK: krnl.return_loops [[DEF_LOOPS]]#0, [[DEF_LOOPS]]#1 -// // CHECK: } : () -> (!krnl.loop, !krnl.loop) -// // CHECK: [[DIM_2:%.+]] = dim %arg0, 0 : memref -// // CHECK: krnl.iterate([[OPT_LOOPS]]#0, [[OPT_LOOPS]]#1) with ([[DEF_LOOPS]]#0 -> %arg1 = 0 to [[DIM_2]], [[DEF_LOOPS]]#1 -> %arg2 = 0 to 10) { -// // CHECK: [[LOAD:%.+]] = load %arg0[%arg1, %arg2] : memref -// // CHECK: [[ZERO:%.+]] = constant {{0.+}} : f32 -// // CHECK: [[NLOAD:%.+]] = subf [[ZERO]], [[LOAD]] : f32 -// // CHECK: [[EXP:%.+]] = exp [[LOAD]] : f32 -// // CHECK: [[NEXP:%.+]] = exp [[NLOAD]] : f32 -// // CHECK: [[DIVIDEND:%.+]] = subf [[EXP]], [[NEXP]] : f32 -// // CHECK: [[DIVISOR:%.+]] = addf [[EXP]], [[NEXP]] : f32 -// // CHECK: [[TANH_RES:%.+]] = divf [[DIVIDEND]], [[DIVISOR]] : f32 -// // CHECK: store [[TANH_RES]], [[RES]][%arg1, %arg2] : memref -// // CHECK: return [[RES]] : memref + // CHECK-LABEL: test_tanh + // CHECK: [[DIM_0:%.+]] = dim %arg0, 0 : memref + // CHECK: [[RES:%.+]] = alloc([[DIM_0]]) : memref + // CHECK: [[DEF_LOOPS:%.+]]:2 = krnl.define_loops 2 + // CHECK: [[OPT_LOOPS:%.+]]:2 = krnl.optimize_loops { + // CHECK: krnl.return_loops [[DEF_LOOPS]]#0, [[DEF_LOOPS]]#1 + // CHECK: } : () -> (!krnl.loop, !krnl.loop) + // CHECK: [[DIM_2:%.+]] = dim %arg0, 0 : memref + // CHECK: krnl.iterate([[OPT_LOOPS]]#0, [[OPT_LOOPS]]#1) with ([[DEF_LOOPS]]#0 -> %arg1 = 0 to [[DIM_2]], [[DEF_LOOPS]]#1 -> %arg2 = 0 to 10) { + // CHECK: [[LOAD:%.+]] = load %arg0[%arg1, %arg2] : memref + // CHECK: [[ZERO:%.+]] = constant {{0.+}} : f32 + // CHECK: [[NLOAD:%.+]] = subf [[ZERO]], [[LOAD]] : f32 + // CHECK: [[EXP:%.+]] = exp [[LOAD]] : f32 + // CHECK: [[NEXP:%.+]] = exp [[NLOAD]] : f32 + // CHECK: [[DIVIDEND:%.+]] = subf [[EXP]], [[NEXP]] : f32 + // CHECK: [[DIVISOR:%.+]] = addf [[EXP]], [[NEXP]] : f32 + // CHECK: [[TANH_RES:%.+]] = divf [[DIVIDEND]], [[DIVISOR]] : f32 + // CHECK: store [[TANH_RES]], [[RES]][%arg1, %arg2] : memref + // CHECK: return [[RES]] : memref } func @test_sinh(%arg0 : tensor) -> tensor<*xf32> {