Picture for Piotr Miłoś

Piotr Miłoś

Joint MoE Scaling Laws: Mixture of Experts Can Be Memory Efficient

Add code
Feb 07, 2025
Figure 1 for Joint MoE Scaling Laws: Mixture of Experts Can Be Memory Efficient
Figure 2 for Joint MoE Scaling Laws: Mixture of Experts Can Be Memory Efficient
Figure 3 for Joint MoE Scaling Laws: Mixture of Experts Can Be Memory Efficient
Figure 4 for Joint MoE Scaling Laws: Mixture of Experts Can Be Memory Efficient
Viaarxiv icon

Repurposing Language Models into Embedding Models: Finding the Compute-Optimal Recipe

Add code
Jun 06, 2024
Figure 1 for Repurposing Language Models into Embedding Models: Finding the Compute-Optimal Recipe
Figure 2 for Repurposing Language Models into Embedding Models: Finding the Compute-Optimal Recipe
Figure 3 for Repurposing Language Models into Embedding Models: Finding the Compute-Optimal Recipe
Figure 4 for Repurposing Language Models into Embedding Models: Finding the Compute-Optimal Recipe
Viaarxiv icon

What Matters in Hierarchical Search for Combinatorial Reasoning Problems?

Add code
Jun 05, 2024
Viaarxiv icon

Bigger, Regularized, Optimistic: scaling for compute and sample-efficient continuous control

Add code
May 25, 2024
Figure 1 for Bigger, Regularized, Optimistic: scaling for compute and sample-efficient continuous control
Figure 2 for Bigger, Regularized, Optimistic: scaling for compute and sample-efficient continuous control
Figure 3 for Bigger, Regularized, Optimistic: scaling for compute and sample-efficient continuous control
Figure 4 for Bigger, Regularized, Optimistic: scaling for compute and sample-efficient continuous control
Viaarxiv icon

tsGT: Stochastic Time Series Modeling With Transformer

Add code
Mar 15, 2024
Viaarxiv icon

Overestimation, Overfitting, and Plasticity in Actor-Critic: the Bitter Lesson of Reinforcement Learning

Add code
Mar 01, 2024
Figure 1 for Overestimation, Overfitting, and Plasticity in Actor-Critic: the Bitter Lesson of Reinforcement Learning
Figure 2 for Overestimation, Overfitting, and Plasticity in Actor-Critic: the Bitter Lesson of Reinforcement Learning
Figure 3 for Overestimation, Overfitting, and Plasticity in Actor-Critic: the Bitter Lesson of Reinforcement Learning
Figure 4 for Overestimation, Overfitting, and Plasticity in Actor-Critic: the Bitter Lesson of Reinforcement Learning
Viaarxiv icon

Analysing The Impact of Sequence Composition on Language Model Pre-Training

Add code
Feb 21, 2024
Figure 1 for Analysing The Impact of Sequence Composition on Language Model Pre-Training
Figure 2 for Analysing The Impact of Sequence Composition on Language Model Pre-Training
Figure 3 for Analysing The Impact of Sequence Composition on Language Model Pre-Training
Figure 4 for Analysing The Impact of Sequence Composition on Language Model Pre-Training
Viaarxiv icon

Fine-tuning Reinforcement Learning Models is Secretly a Forgetting Mitigation Problem

Add code
Feb 05, 2024
Figure 1 for Fine-tuning Reinforcement Learning Models is Secretly a Forgetting Mitigation Problem
Figure 2 for Fine-tuning Reinforcement Learning Models is Secretly a Forgetting Mitigation Problem
Figure 3 for Fine-tuning Reinforcement Learning Models is Secretly a Forgetting Mitigation Problem
Figure 4 for Fine-tuning Reinforcement Learning Models is Secretly a Forgetting Mitigation Problem
Viaarxiv icon

Structured Packing in LLM Training Improves Long Context Utilization

Add code
Jan 02, 2024
Viaarxiv icon

Focused Transformer: Contrastive Training for Context Scaling

Add code
Jul 06, 2023
Viaarxiv icon