Picture for Yuxian Gu

Yuxian Gu

MiniPLM: Knowledge Distillation for Pre-Training Language Models

Add code
Oct 22, 2024
Viaarxiv icon

Data Selection via Optimal Control for Language Models

Add code
Oct 09, 2024
Viaarxiv icon

Direct Preference Knowledge Distillation for Large Language Models

Add code
Jun 28, 2024
Viaarxiv icon

Instruction Pre-Training: Language Models are Supervised Multitask Learners

Add code
Jun 20, 2024
Viaarxiv icon

Towards Optimal Learning of Language Models

Add code
Mar 03, 2024
Viaarxiv icon

Synthetic Data (Almost) from Scratch: Generalized Instruction Tuning for Language Models

Add code
Feb 20, 2024
Viaarxiv icon

Knowledge Distillation of Large Language Models

Add code
Jun 14, 2023
Viaarxiv icon

Pre-Training to Learn in Context

Add code
May 16, 2023
Viaarxiv icon

Structured Prompting: Scaling In-Context Learning to 1,000 Examples

Add code
Dec 13, 2022
Viaarxiv icon

Learning Instructions with Unlabeled Data for Zero-Shot Cross-Task Generalization

Add code
Oct 17, 2022
Figure 1 for Learning Instructions with Unlabeled Data for Zero-Shot Cross-Task Generalization
Figure 2 for Learning Instructions with Unlabeled Data for Zero-Shot Cross-Task Generalization
Figure 3 for Learning Instructions with Unlabeled Data for Zero-Shot Cross-Task Generalization
Figure 4 for Learning Instructions with Unlabeled Data for Zero-Shot Cross-Task Generalization
Viaarxiv icon