TC-CIM: Empowering Tensor Comprehensions for Computing-In-Memory

Andi Drebes, Lorenzo Chelini, Oleksandr Zinenko, Albert Cohen, Henk Corporaal, Tobias Grosser, Kanishkan Vadivel, Nicolas Vasilache

Research output: Contribution to conferencePaperpeer-review

Abstract / Description of output

Memristor-based, non-von-Neumann architectures performing tensor operations directly in memory are a promising approach to address the ever-increasing demand for energyefficient, high-throughput hardware accelerators for Machine Learning (ML) inference. A major challenge for the programmability and exploitation of such Computing-InMemory (CIM) architectures consists in the efficient mapping of tensor operations from high-level ML frameworks to fixed-function hardware blocks implementing in-memory computations. We demonstrate the programmability of memristor-based accelerators with TC-CIM, a fully-automatic, end-to-end compilation flow from Tensor Comprehensions, a mathematical notation for tensor operations, to fixed-function memristor-based hardware blocks. Operations suitable for acceleration are identified using Loop Tactics, a declarative framework to describe computational patterns in a polyhedral representation. We evaluate our compilation flow on a system-level simulator based on Gem5, incorporating crossbar arrays of memristive devices. Our results show that TC-CIM reliably recognizes tensor operations commonly used in ML workloads across multiple benchmarks in order to offload these operations to the accelerator. 
Original languageEnglish
Number of pages12
Publication statusPublished - 22 Jan 2020
Event10th International Workshop on Polyhedral Compilation Techniques - Bologna, Italy
Duration: 22 Jan 202022 Jan 2020


Workshop10th International Workshop on Polyhedral Compilation Techniques
Abbreviated titleIMAPCT 2020
Internet address


Dive into the research topics of 'TC-CIM: Empowering Tensor Comprehensions for Computing-In-Memory'. Together they form a unique fingerprint.

Cite this