Efficient AI Computing,
Transforming the Future.

Projects

To choose projects, simply check the boxes of the categories, topics and techniques.
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.

ProxylessNAS: Direct Neural Architecture Search on Target Task and Hardware

ICLR 2019
 (
)

ProxylessNAS is an efficient hardware-aware neural architecture search method, which can directly search on large-scale datasets. It can design specialized neural network architecture for different hardware platforms. With >74.5% top-1 accuracy, the latency of ProxylessNAS is 1.8x faster than MobileNetV2.

Deep Gradient Compression: Reducing the Communication Bandwidth in Distributed Training

ICLR 2018
 (
)

Deep Gradient Compression (DGC) reduces the communication bandwidth in the large-scale distributed training via four techniques: momentum correction, local gradient clipping, momentum factor masking, and warm-up training.

Learning to Design Circuits

NIPS 2019 MLSys Workshop
 (
oral
)

We develop a reinforcement learning framework for analog circuit design.

AMC: AutoML for Model Compression and Acceleration on Mobile Devices

ECCV 2018
 (
)

AutoML for Model Compression (AMC) leverages reinforcement learning to provide the model compression policy. This learning-based compression policy outperforms conventional rule-based compression policy by having higher compression ratio, better preserving the accuracy and freeing human labor.