Picture for Shaoxiong Ji

Shaoxiong Ji

EMMA-500: Enhancing Massively Multilingual Adaptation of Large Language Models

Add code
Sep 26, 2024
Viaarxiv icon

Two Stacks Are Better Than One: A Comparison of Language Modeling and Translation as Multilingual Pretraining Objectives

Add code
Jul 22, 2024
Viaarxiv icon

Lucky 52: How Many Languages Are Needed to Instruction Fine-Tune Large Language Models?

Add code
Apr 07, 2024
Viaarxiv icon

Can Machine Translation Bridge Multilingual Pretraining and Cross-lingual Transfer Learning?

Add code
Mar 25, 2024
Viaarxiv icon

A New Massive Multilingual Dataset for High-Performance Language Technologies

Add code
Mar 20, 2024
Viaarxiv icon

MAMMOTH: Massively Multilingual Modular Open Translation @ Helsinki

Add code
Mar 12, 2024
Figure 1 for MAMMOTH: Massively Multilingual Modular Open Translation @ Helsinki
Figure 2 for MAMMOTH: Massively Multilingual Modular Open Translation @ Helsinki
Figure 3 for MAMMOTH: Massively Multilingual Modular Open Translation @ Helsinki
Figure 4 for MAMMOTH: Massively Multilingual Modular Open Translation @ Helsinki
Viaarxiv icon

MaLA-500: Massive Language Adaptation of Large Language Models

Add code
Jan 24, 2024
Viaarxiv icon

Rethinking Large Language Models in Mental Health Applications

Add code
Nov 19, 2023
Viaarxiv icon

Monolingual or Multilingual Instruction Tuning: Which Makes a Better Alpaca

Add code
Sep 16, 2023
Viaarxiv icon

Content Reduction, Surprisal and Information Density Estimation for Long Documents

Add code
Sep 12, 2023
Viaarxiv icon