{"id":555282,"date":"2018-12-04T18:10:52","date_gmt":"2018-12-05T02:10:52","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?post_type=msr-project&p=555282"},"modified":"2023-07-10T03:41:13","modified_gmt":"2023-07-10T10:41:13","slug":"deep-learning-compiler-and-optimizer","status":"publish","type":"msr-project","link":"https:\/\/www.microsoft.com\/en-us\/research\/project\/deep-learning-compiler-and-optimizer\/","title":{"rendered":"Deep Learning Compiler and Optimizer"},"content":{"rendered":"

Project Overview<\/h3>\n

This project aims to build a deep learning compiler and optimizer infrastructure that can provide automatic scalability and efficiency optimization for distributed and local execution.\u00a0 Overall, this stack covers two types of general optimizations: fast distributed training over large-scale servers and efficient local execution on various hardware devices.\u00a0 Currently, our optimizations focus on many different parts of the system stack, such as fast distributed training over RDMA, automatic computation placement across devices, automatic operator batching and kernel fusion, tensor algebra compiler, sparse and quantization optimizations, and so on.<\/span><\/p>\n

\"graphical<\/p>\n

Open-source Release<\/h3>\n

Some of our projects have been open-sourced, and welcome to try, contribute and collaborate with us.<\/p>\n