Picture for Ahmet Üstün

Ahmet Üstün

Nexus: Specialization meets Adaptability for Efficiently Training Mixture of Experts

Add code
Aug 28, 2024
Viaarxiv icon

To Code, or Not To Code? Exploring Impact of Code in Pre-training

Add code
Aug 20, 2024
Figure 1 for To Code, or Not To Code? Exploring Impact of Code in Pre-training
Figure 2 for To Code, or Not To Code? Exploring Impact of Code in Pre-training
Figure 3 for To Code, or Not To Code? Exploring Impact of Code in Pre-training
Figure 4 for To Code, or Not To Code? Exploring Impact of Code in Pre-training
Viaarxiv icon

How Does Quantization Affect Multilingual LLMs?

Add code
Jul 03, 2024
Viaarxiv icon

RLHF Can Speak Many Languages: Unlocking Multilingual Preference Optimization for LLMs

Add code
Jul 02, 2024
Viaarxiv icon

Aya 23: Open Weight Releases to Further Multilingual Progress

Add code
May 23, 2024
Figure 1 for Aya 23: Open Weight Releases to Further Multilingual Progress
Figure 2 for Aya 23: Open Weight Releases to Further Multilingual Progress
Figure 3 for Aya 23: Open Weight Releases to Further Multilingual Progress
Figure 4 for Aya 23: Open Weight Releases to Further Multilingual Progress
Viaarxiv icon

Back to Basics: Revisiting REINFORCE Style Optimization for Learning from Human Feedback in LLMs

Add code
Feb 26, 2024
Viaarxiv icon

Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model

Add code
Feb 12, 2024
Viaarxiv icon

Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning

Add code
Feb 09, 2024
Figure 1 for Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning
Figure 2 for Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning
Figure 3 for Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning
Figure 4 for Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning
Viaarxiv icon

Pushing Mixture of Experts to the Limit: Extremely Parameter Efficient MoE for Instruction Tuning

Add code
Sep 11, 2023
Viaarxiv icon

When Less is More: Investigating Data Pruning for Pretraining LLMs at Scale

Add code
Sep 08, 2023
Viaarxiv icon