Picture for Clara Na

Clara Na

Scalable Data Ablation Approximations for Language Models through Modular Training and Merging

Add code
Oct 21, 2024
Figure 1 for Scalable Data Ablation Approximations for Language Models through Modular Training and Merging
Figure 2 for Scalable Data Ablation Approximations for Language Models through Modular Training and Merging
Figure 3 for Scalable Data Ablation Approximations for Language Models through Modular Training and Merging
Figure 4 for Scalable Data Ablation Approximations for Language Models through Modular Training and Merging
Viaarxiv icon

Energy and Carbon Considerations of Fine-Tuning BERT

Add code
Nov 17, 2023
Viaarxiv icon

To Build Our Future, We Must Know Our Past: Contextualizing Paradigm Shifts in Natural Language Processing

Add code
Oct 11, 2023
Viaarxiv icon

The Framework Tax: Disparities Between Inference Efficiency in Research and Deployment

Add code
Feb 13, 2023
Viaarxiv icon

Train Flat, Then Compress: Sharpness-Aware Minimization Learns More Compressible Models

Add code
May 25, 2022
Figure 1 for Train Flat, Then Compress: Sharpness-Aware Minimization Learns More Compressible Models
Figure 2 for Train Flat, Then Compress: Sharpness-Aware Minimization Learns More Compressible Models
Figure 3 for Train Flat, Then Compress: Sharpness-Aware Minimization Learns More Compressible Models
Figure 4 for Train Flat, Then Compress: Sharpness-Aware Minimization Learns More Compressible Models
Viaarxiv icon