onnx-mlir/test/mlir/onnx
Tung D. Le 7c1e67898d
Fuse convolution and batch normalization (#253)
* Rewriting rule

* Fix formulas

* Reuse op results

* Const propagation for Div and Sqrt

* Explicitly use ONNXConstantOp

* Minor revise

* Const propagation for unsqueeze

* Do const propagationnce all tensors have inferred shapes

* LIT tests for fusion

* Add LIT tests for constant propagation on Div, Sqrt, and Unsqueeze

* Missing dash

Co-authored-by: Tian Jin <tjingrant@gmail.com>
2020-08-18 16:41:40 +08:00
..
onnx_bundle_memory_pool.mlir Assorted fixes for memory pool related passes (#244) 2020-08-11 17:34:59 -04:00
onnx_canonicalization.mlir Fuse convolution and batch normalization (#253) 2020-08-18 16:41:40 +08:00
onnx_constprop.mlir Fuse convolution and batch normalization (#253) 2020-08-18 16:41:40 +08:00
onnx_decompose.mlir ScalerOp support non-float input using CastOp (#231) 2020-07-24 10:57:52 -04:00
onnx_elide_constants.mlir Rework output to improve readability of intermediate MLIR code. (#87) 2020-04-24 16:15:36 -04:00
onnx_enable_memory_pool.mlir [NFC] Rename passes for stylistic consistency. (#232) 2020-07-31 21:37:35 +08:00
onnx_krnl_global_elision.mlir Remove optimize_loops/return_loops op. (#200) 2020-07-08 12:49:15 +08:00
onnx_lowering.mlir Add CastOp lowering (#259) 2020-08-11 16:07:13 -04:00
onnx_lowering_with_dealloc.mlir [NFC] Rename passes for stylistic consistency. (#232) 2020-07-31 21:37:35 +08:00
onnx_shape_inference.mlir OneHotEncoder Shape Inference (#265) 2020-08-14 16:13:31 -04:00
onnx_shape_inference_maxpool.mlir Add the split marker to test files (#90) 2020-04-16 15:17:27 +08:00
onnx_structure.mlir Seq type (#199) 2020-07-31 20:05:59 +08:00