Picture for Rui Hou

Rui Hou

Jack

Self-Generated Critiques Boost Reward Modeling for Language Models

Add code
Nov 25, 2024
Figure 1 for Self-Generated Critiques Boost Reward Modeling for Language Models
Figure 2 for Self-Generated Critiques Boost Reward Modeling for Language Models
Figure 3 for Self-Generated Critiques Boost Reward Modeling for Language Models
Figure 4 for Self-Generated Critiques Boost Reward Modeling for Language Models
Viaarxiv icon

Layer Swapping for Zero-Shot Cross-Lingual Transfer in Large Language Models

Add code
Oct 02, 2024
Viaarxiv icon

Law of the Weakest Link: Cross Capabilities of Large Language Models

Add code
Sep 30, 2024
Figure 1 for Law of the Weakest Link: Cross Capabilities of Large Language Models
Figure 2 for Law of the Weakest Link: Cross Capabilities of Large Language Models
Figure 3 for Law of the Weakest Link: Cross Capabilities of Large Language Models
Figure 4 for Law of the Weakest Link: Cross Capabilities of Large Language Models
Viaarxiv icon

The Llama 3 Herd of Models

Add code
Jul 31, 2024
Viaarxiv icon

Synergistic Anchored Contrastive Pre-training for Few-Shot Relation Extraction

Add code
Dec 24, 2023
Viaarxiv icon

Jack of All Tasks, Master of Many: Designing General-purpose Coarse-to-Fine Vision-Language Model

Add code
Dec 19, 2023
Viaarxiv icon

RoAST: Robustifying Language Models via Adversarial Perturbation with Selective Training

Add code
Dec 07, 2023
Viaarxiv icon

MART: Improving LLM Safety with Multi-round Automatic Red-Teaming

Add code
Nov 13, 2023
Figure 1 for MART: Improving LLM Safety with Multi-round Automatic Red-Teaming
Figure 2 for MART: Improving LLM Safety with Multi-round Automatic Red-Teaming
Figure 3 for MART: Improving LLM Safety with Multi-round Automatic Red-Teaming
Figure 4 for MART: Improving LLM Safety with Multi-round Automatic Red-Teaming
Viaarxiv icon

Co-training and Co-distillation for Quality Improvement and Compression of Language Models

Add code
Nov 07, 2023
Figure 1 for Co-training and Co-distillation for Quality Improvement and Compression of Language Models
Figure 2 for Co-training and Co-distillation for Quality Improvement and Compression of Language Models
Figure 3 for Co-training and Co-distillation for Quality Improvement and Compression of Language Models
Figure 4 for Co-training and Co-distillation for Quality Improvement and Compression of Language Models
Viaarxiv icon

Effective Long-Context Scaling of Foundation Models

Add code
Sep 27, 2023
Viaarxiv icon