Picture for Konstantin Mishchenko

Konstantin Mishchenko

SIERRA, PSL

Hardware-Aware Parallel Prompt Decoding for Memory-Efficient Acceleration of LLM Inference

Add code
May 28, 2024
Figure 1 for Hardware-Aware Parallel Prompt Decoding for Memory-Efficient Acceleration of LLM Inference
Figure 2 for Hardware-Aware Parallel Prompt Decoding for Memory-Efficient Acceleration of LLM Inference
Figure 3 for Hardware-Aware Parallel Prompt Decoding for Memory-Efficient Acceleration of LLM Inference
Figure 4 for Hardware-Aware Parallel Prompt Decoding for Memory-Efficient Acceleration of LLM Inference
Viaarxiv icon

The Road Less Scheduled

Add code
May 24, 2024
Viaarxiv icon

When, Why and How Much? Adaptive Learning Rate Scheduling by Refinement

Add code
Oct 11, 2023
Viaarxiv icon

Adaptive Proximal Gradient Method for Convex Optimization

Add code
Aug 04, 2023
Viaarxiv icon

Prodigy: An Expeditiously Adaptive Parameter-Free Learner

Add code
Jun 09, 2023
Viaarxiv icon

Partially Personalized Federated Learning: Breaking the Curse of Data Heterogeneity

Add code
May 29, 2023
Viaarxiv icon

DoWG Unleashed: An Efficient Universal Parameter-Free Gradient Descent Method

Add code
May 25, 2023
Viaarxiv icon

Two Losses Are Better Than One: Faster Optimization Using a Cheaper Proxy

Add code
Feb 07, 2023
Viaarxiv icon

Learning-Rate-Free Learning by D-Adaptation

Add code
Jan 20, 2023
Viaarxiv icon

Convergence of First-Order Algorithms for Meta-Learning with Moreau Envelopes

Add code
Jan 17, 2023
Viaarxiv icon