Alert button
Picture for Qun Gao

Qun Gao

Alert button

Efficient Post-training Quantization with FP8 Formats

Add code
Bookmark button
Alert button
Sep 26, 2023
Haihao Shen, Naveen Mellempudi, Xin He, Qun Gao, Chang Wang, Mengni Wang

Figure 1 for Efficient Post-training Quantization with FP8 Formats
Figure 2 for Efficient Post-training Quantization with FP8 Formats
Figure 3 for Efficient Post-training Quantization with FP8 Formats
Figure 4 for Efficient Post-training Quantization with FP8 Formats
Viaarxiv icon

An Efficient Sparse Inference Software Accelerator for Transformer-based Language Models on CPUs

Add code
Bookmark button
Alert button
Jun 28, 2023
Haihao Shen, Hengyu Meng, Bo Dong, Zhe Wang, Ofir Zafrir, Yi Ding, Yu Luo, Hanwen Chang, Qun Gao, Ziheng Wang, Guy Boudoukh, Moshe Wasserblat

Figure 1 for An Efficient Sparse Inference Software Accelerator for Transformer-based Language Models on CPUs
Figure 2 for An Efficient Sparse Inference Software Accelerator for Transformer-based Language Models on CPUs
Figure 3 for An Efficient Sparse Inference Software Accelerator for Transformer-based Language Models on CPUs
Figure 4 for An Efficient Sparse Inference Software Accelerator for Transformer-based Language Models on CPUs
Viaarxiv icon