Efficient AI Computing,
Transforming the Future.

Team

Principle Investigator

Song Han is an associate professor at MIT EECS and distinguished scientist at NVIDIA. He received his PhD degree from Stanford University. He proposed the “Deep Compression” technique including pruning and quantization that is widely used for efficient AI computing, and “Efficient Inference Engine” that first brought weight sparsity to modern AI chips. He pioneered the TinyML research that brings deep learning to IoT devices, enabling learning on the edge (appeared on MIT home page). His team’s work on hardware-aware neural architecture search (once-for-all network) enables users to design, optimize, shrink and deploy AI models to resource-constrained hardware devices, receiving the first place in many low-power computer vision contests in flagship AI conferences.  His team’s recent work on large language model quantization/acceleration (SmoothQuant, AWQ, StreamingLLM) has effectively improved the efficiency of LLM inference, adopted by NVIDIA TensorRT-LLM. Song received best paper awards at ICLR and FPGA, faculty awards from Amazon, Facebook, NVIDIA, Samsung and SONY. Song was named “35 Innovators Under 35” by MIT Technology Review for his contribution on “deep compression” technique that “lets powerful artificial intelligence (AI) programs run more efficiently on low-power mobile devices.” Song received the NSF CAREER Award for “efficient algorithms and hardware for accelerated machine learning”, IEEE “AIs 10 to Watch: The Future of AI” award, and Sloan Research Fellowship. Song’s research in efficient AI computing has witnessed successful commercialization and influenced the industry. He was the cofounder of DeePhi (now part of AMD), and cofounder of OmniML (now part of NVIDIA). Song developed the EfficientML.ai course to disseminate this line of research.

Team Members

Postdoctoral

His research focuses on building efficient machine learning systems, particularly in the areas of training, serving, and scheduling for foundation models. His work received the Distinguished Paper Award at ASPLOS. He is also a recipient of the Google PhD Fellowship and has been recognized as one of the ML and Systems Rising Stars by MLCommons.

Ph.D

His research interests focus on the development of efficient algorithms and systems for deep learning, specifically large foundation models. His work has received over 8000 stars on GitHub. His work has a real-world impact: SmoothQuant has been integrated into NVIDIA's TensorRT-LLM, FasterTransformer and Intel's NeuralCompressor and is utilized in the LLMs of industry companies like Amazon, Meta, and Huggingface. StreamingLLM has been integrated into NVIDIA's TensorRT-LLM, Huggingface's transformers, and Intels' Extension for Transformers.

His research interests lie at the intersection of computer systems and machine learning, specifically in the area of full-stack efficient 3D deep learning for autonomous driving. Haotian has authored multiple papers with over 1,300 citations and over 4,200 GitHub stars in this field. Haotian has successfully advised several undergraduate students, and the intern students he mentored have continued as PhD students at MIT and UC Berkeley.

Jiaming Tang is a first-year Ph.D. student at MIT, advised by Prof. Song Han. He was a member of ACM Honors Class, Shanghai Jiao Tong University. His research interests lie in efficient systems and algorithms for large language models. His work AWQ receives the Best Paper Award at MLSys 2024 and has been integrated into Transformers, vLLM, FastChat, TensorRT-LLM, and TGI.

Muyang Li is a first-year Ph.D. student at MIT, advised by Prof. Song Han. He obtained his master’s degree at Robotics Institute, CMU, advised by Prof. Jun-Yan Zhu, and his Bachelor’s degree from Zhiyuan College (ACM Class), Shanghai Jiao Tong University. His research interest is in the intersection of machine learning, system, and computer graphics. He is currently working on building efficient and hardware-friendly generative models with its applications in computer vision and graphics.

Shang Yang is a first-year Ph.D. student at MIT, advised by Prof. Song Han. He received his B.Eng. degree from Tsinghua University. His research interests lie in efficient systems for machine learning applications, such as LLMs and Point Clouds.

His research focuses on the intersection of computer architecture and machine learning, particularly the co-design of software and hardware for deep learning and its applications. Yujun was awarded the 2021 Qualcomm Innovation Fellowship, and he is the founding member of the new course on TinyML and efficient deep learning computing (MIT 6.S965) teaching crew, which received 12k views on YouTube.

His research focuses on the development of high-performance and efficient hardware architectures for sparse linear algebra and deep learning. Zhekai has published several papers in the field of computer architecture, which have received over 250 citations.

Zhuoyang Zhang is a first-year Ph.D. student at MIT, advised by Prof. Song Han. He received his Bachelor’s degree from Yao class, Tsinghua University. His research interests lie in machine learning and computer vision. He aims to accelerate large foundation models for modern AI computing.

Master

She completed her undergrad in CS at MIT in 2023, and is currently an MEng student interested in performant systems and machine learning.

Undergraduate

Maggie Liu is a first-year undergraduate student who loves competitive programming and hackathons and is interested in machine learning.

Shreya Chaudhury is a second year undergraduate student at MIT EECS department. Her research interest is quantum computing, especially on variational quantum algorithms and quantum machine learning. She is the core developer of the TorchQuantum library.

Graduated

Ph.D

His research focuses on efficient machine learning and systems. His work has been featured in oral and spotlight presentations at conferences such as NeurIPS, ICLR, and CVPR. He received the Qualcomm Innovation Fellowship and was recognized as a Rising Star in ML and Systems by MLCommons and a Rising Star in Data Science by UChicago and UCSD. He received his B.Eng. degree from Shanghai Jiao Tong University. After graduation, Zhijian joined NVIDIA as a research scientist and will join UCSD as an assistant professor.

Ph.D

Han Cai is a final-year Ph.D. student at MIT EECS, advised by Professor Song Han. His research focuses on algorithms and acceleration of efficient deep learning computing. Han has made significant contributions to the field, including his work on hardware-aware neural architecture search (ProxylessNAS, Once-for-All), which has been integrated into PytorchHub@Meta, AutoGluon@Amazon, NNI@Microsoft, SONY Neural Architecture Search Library, SONY Model Compression Toolkit, and ADI Model Training and Synthesis Tool. His research has received 6.9K+ citations on Google Scholar and 5.2K+ stars on GitHub. He joined NVIDIA as a research scientist.

Ph.D

After graduation, Hanrui will join UCLA CS as an Assistant Professor. His research focuses on efficient AI and emerging hardware (e.g. quantum architecture). His research has been recognized by ACM student research competition 1st place award, best poster award at NSF AI Institute, Best Presentation Award as a DAC Young Fellow and appears in top conferences such as NeurIPS, ISCA, MICRO, HPCA, and DAC. His co-authored papers received ICML RL4RL Best Paper Award and QCE Best Paper Award. He is the recipient of Qualcomm Fellowship, Unitary Fund, and Nvidia Fellowship Finalist. He is the creator of TorchQuantum library which has been adopted by IBM and PyTorch Ecosystems. He is also the co-founder of QuCS lecture series for quantum education. Hanrui received his B. Eng. degree from Fudan University and S.M. degree from MIT.

Postdoctoral

His research focuses on efficient deep learning, TinyML, embedded systems, and memory/storage systems. Wei-Chen has received several accolades for his work, including the ACM/IEEE CODES+ISSS Best Paper Award, the IEEE NVMSA Best Paper Award, and the Best Poster Award at the NSF Athena AI Institute. In addition, he received first place (among 150 teams) in the flash consumption track of the ACM/IEEE TinyML Design Contest at ICCAD 2022. His research has received over 1,300 stars on GitHub, and his work "On-device training under 256KB memory" (MCUNetV3) was highlighted by the MIT homepage. He joined Amazon as a research scientist.

Ph.D

His research focuses on efficient deep learning computing, systems for ML and recently, accelerating large language models (LLMs). Ji is pioneering the research in the field of TinyML. His research has received over 6,500 citations on Google Scholar and over 7,000 stars on GitHub. Ji is an NVIDIA Graduate Fellowship Finalist in 2020, and Qualcomm Innovation Fellowship recipient in 2022. He joined OpenAI as a research staff.

Postdoctoral

Wei-Ming Chen is a Postdoctoral Associate at MIT EECS advised by Professor Song Han. His research focuses on TinyML, embedded systems, and real-time systems, with a particular emphasis on enabling efficient deep learning on Internet of Things (IoT) devices, such as microcontrollers. Chen's recent work on the MCUNet series (MCUNet, MCUNetv2, and MCUNetv3) has enabled efficient inference and training on devices with limited memory through the co-design of systems and algorithms. He is also a key contributor and maintainer of TinyEngine, an open-source library for high-performance and memory-efficient deep learning on microcontrollers. His work "On-device training under 256KB memory" (MCUNetV3) is highlighted by the MIT homepage in fall 2022. He received first place (among 150 teams) in the flash consumption track of the ACM/IEEE TinyML Design Contest at ICCAD 2022. He developed TinyChatEngine that enables LLM inference on the edge (laptop, Paspberry PI). His research has received more than 1,000 stars on GitHub. After graduation, he joined NVIDIA as a senior deep learning engineer working on large language model acceleration.

Undergraduate

Jessica Zheng was an M.Eng student at MIT Han Lab. Her research focused on efficient deep learning, anomaly detection and machine learning for healthcare. She received the 1st place award at the ACM TinyML Design Contest memory size track in 2022.

Master

Kevin Shao was an M.Eng student at MIT HAN Lab, working on autonomous driving and efficient 3D deep learning. After graduation, he joined Two Sigma as a Quantitative Researcher.

Master

Driss Hafdi was an M.Eng student at MIT EECS, working on specialized hardware for mixed-precision quantization. After graduation, he joined Hudson River Trading as an FPGA developer.

Openings

If you work on efficient LLM, VLM, GenAI and are interested in joining us, please fill in the recruiting form. Inquiry emails will not be replied if the recruiting form is incomplete. PhD applicants: select "ML+System" track in the MIT PhD application system.

Sponsors

We actively collaborate with industry partners on efficient AI, model compression and acceleration. Our research has influenced and landed in many industrial products: Intel OpenVino, Intel Neural Network Distiller, Intel Neural Compressor, Apple Neural Engine, NVIDIA Sparse Tensor Core, NVIDIA TensorRT LLM, AMD-Xilinx Vitis AI, Qualcomm AI Model Efficiency Toolkit (AIMET), Amazon AutoGluon, Facebook PyTorch, Microsoft NNI, SONY Neural Architecture Search Library, SONY Model Compression Toolkit,  ADI MAX78000/MAX78002 Model Training and Synthesis Tool.