News & features
Research Focus: Week of August 12, 2024
In this issue: Research Forum Ep. 4 explores multimodal AI. Registration is now open; Surveying developers’ AI needs; SuperBench improves cloud AI infrastructure reliability; Virtual Voices: Exploring factors influencing participation in virtual meetings.
Awards | USENIX ATC 2024
Best Paper Award at USENIX ATC 2024
Our paper titled "SuperBench: Improving Cloud AI Infrastructure Reliability with Proactive Validation" received the Best Paper Award at the 2024 USENIX Annual Technical Conference (USENIX ATC '24).
Tutel: An efficient mixture-of-experts implementation for large DNN model training
| Wei Cui, Yifan Xiong, Peng Cheng, and Rafael Salas
Mixture of experts (MoE) is a deep learning model architecture in which computational cost is sublinear to the number of parameters, making scaling easier. Nowadays, MoE is the only approach demonstrated to scale deep learning models to trillion-plus parameters, paving…