Picture for Weizhu Chen

Weizhu Chen

StreamAdapter: Efficient Test Time Adaptation from Contextual Streams

Add code
Nov 14, 2024
Viaarxiv icon

MTL-LoRA: Low-Rank Adaptation for Multi-Task Learning

Add code
Oct 15, 2024
Figure 1 for MTL-LoRA: Low-Rank Adaptation for Multi-Task Learning
Figure 2 for MTL-LoRA: Low-Rank Adaptation for Multi-Task Learning
Figure 3 for MTL-LoRA: Low-Rank Adaptation for Multi-Task Learning
Figure 4 for MTL-LoRA: Low-Rank Adaptation for Multi-Task Learning
Viaarxiv icon

GRIN: GRadient-INformed MoE

Add code
Sep 18, 2024
Figure 1 for GRIN: GRadient-INformed MoE
Figure 2 for GRIN: GRadient-INformed MoE
Figure 3 for GRIN: GRadient-INformed MoE
Figure 4 for GRIN: GRadient-INformed MoE
Viaarxiv icon

Arena Learning: Build Data Flywheel for LLMs Post-training via Simulated Chatbot Arena

Add code
Jul 15, 2024
Viaarxiv icon

Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling

Add code
Jun 11, 2024
Figure 1 for Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling
Figure 2 for Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling
Figure 3 for Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling
Figure 4 for Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling
Viaarxiv icon

Automatic Instruction Evolving for Large Language Models

Add code
Jun 02, 2024
Figure 1 for Automatic Instruction Evolving for Large Language Models
Figure 2 for Automatic Instruction Evolving for Large Language Models
Figure 3 for Automatic Instruction Evolving for Large Language Models
Figure 4 for Automatic Instruction Evolving for Large Language Models
Viaarxiv icon

Self-Augmented Preference Optimization: Off-Policy Paradigms for Language Model Alignment

Add code
May 31, 2024
Figure 1 for Self-Augmented Preference Optimization: Off-Policy Paradigms for Language Model Alignment
Figure 2 for Self-Augmented Preference Optimization: Off-Policy Paradigms for Language Model Alignment
Figure 3 for Self-Augmented Preference Optimization: Off-Policy Paradigms for Language Model Alignment
Figure 4 for Self-Augmented Preference Optimization: Off-Policy Paradigms for Language Model Alignment
Viaarxiv icon

Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone

Add code
Apr 23, 2024
Figure 1 for Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone
Figure 2 for Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone
Figure 3 for Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone
Figure 4 for Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone
Viaarxiv icon

Rho-1: Not All Tokens Are What You Need

Add code
Apr 11, 2024
Figure 1 for Rho-1: Not All Tokens Are What You Need
Figure 2 for Rho-1: Not All Tokens Are What You Need
Figure 3 for Rho-1: Not All Tokens Are What You Need
Figure 4 for Rho-1: Not All Tokens Are What You Need
Viaarxiv icon

A Note on LoRA

Add code
Apr 07, 2024
Viaarxiv icon