Efficient AI Computing,
Transforming the Future.

Projects

To choose projects, simply check the boxes of the categories, topics and techniques.
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.

SpAtten: Efficient Sparse Attention Architecture with Cascade Token and Head Pruning

HPCA 2021
 (
oral
)

Pruning and Quantization for Transformer models such as BERT and GPT

Differentiable Augmentation for Data-Efficient GAN Training

NeurIPS 2020
 (
)

Differentiable augmentation to improve the data efficiency of GAN training.

TinyTL: Reduce Activations, Not Trainable Parameters for Efficient On-Device Learning

NeurIPS 2020
 (
)

Tiny-Transfer-Learning (TinyTL) provides memory-efficient on-device learning by freezing the weights while only learns the bias modules to get rid of the intermediate activations, and introducing the lite residual module to maintain the adaptation capacity.

Searching Efficient 3D Architectures with Sparse Point-Voxel Convolution

ECCV 2020
 (
)

SPVNAS enhances Point-Voxel Convolution in large-scale outdoor scenes with sparse convolutions. With 3D Neural Architecture Search (3D-NAS), it efficiently and effectively searches the optimal 3D neural network architecture under a given resource constraint.