What is it about?

This is a work to improve deep learning inference performance by compilation techniques. The main contribution is to enable flexible data layout optimization without re-implementing operators. With this, we can jointly tune data layouts, and existing loops easily.

Featured Image

Why is it important?

It shows that the bidirectional tuning flow is more beneficial than simply putting different optimization techniques into different system layers and no interaction is involved. In order to do this, a new infra for versatile data layout transformation is developed where some primitive functions help to transform data layouts without re-implementing operators like that in existing works. And, a bidirectional auto-tuning framework is built atop the transformation, which enables interaction between two important optimization techniques.

Read the Original

This page is a summary of: ALT: Breaking the Wall between Data Layout and Loop Optimizations for Deep Learning Compilation, May 2023, ACM (Association for Computing Machinery),
DOI: 10.1145/3552326.3587440.
You can read the full text:

Read

Contributors

The following have contributed to this page