Alert button
Picture for Alexandre Marques

Alexandre Marques

Alert button

Enabling High-Sparsity Foundational Llama Models with Efficient Pretraining and Deployment

Add code
Bookmark button
Alert button
May 06, 2024
Abhinav Agarwalla, Abhay Gupta, Alexandre Marques, Shubhra Pandit, Michael Goin, Eldar Kurtic, Kevin Leong, Tuan Nguyen, Mahmoud Salem, Dan Alistarh, Sean Lie, Mark Kurtz

Viaarxiv icon

oBERTa: Improving Sparse Transfer Learning via improved initialization, distillation, and pruning regimes

Add code
Bookmark button
Alert button
Apr 04, 2023
Daniel Campos, Alexandre Marques, Mark Kurtz, ChengXiang Zhai

Figure 1 for oBERTa: Improving Sparse Transfer Learning via improved initialization, distillation, and pruning regimes
Figure 2 for oBERTa: Improving Sparse Transfer Learning via improved initialization, distillation, and pruning regimes
Figure 3 for oBERTa: Improving Sparse Transfer Learning via improved initialization, distillation, and pruning regimes
Figure 4 for oBERTa: Improving Sparse Transfer Learning via improved initialization, distillation, and pruning regimes
Viaarxiv icon

Sparse*BERT: Sparse Models are Robust

Add code
Bookmark button
Alert button
May 25, 2022
Daniel Campos, Alexandre Marques, Tuan Nguyen, Mark Kurtz, ChengXiang Zhai

Figure 1 for Sparse*BERT: Sparse Models are Robust
Figure 2 for Sparse*BERT: Sparse Models are Robust
Figure 3 for Sparse*BERT: Sparse Models are Robust
Figure 4 for Sparse*BERT: Sparse Models are Robust
Viaarxiv icon