Picture for James T. Kwok

James T. Kwok

Corrupted but Not Broken: Rethinking the Impact of Corrupted Data in Visual Instruction Tuning

Add code
Feb 18, 2025
Viaarxiv icon

Gradient-Based Multi-Objective Deep Learning: Algorithms, Theories, Applications, and Beyond

Add code
Jan 19, 2025
Viaarxiv icon

RouterDC: Query-Based Router by Dual Contrastive Learning for Assembling Large Language Models

Add code
Sep 30, 2024
Figure 1 for RouterDC: Query-Based Router by Dual Contrastive Learning for Assembling Large Language Models
Figure 2 for RouterDC: Query-Based Router by Dual Contrastive Learning for Assembling Large Language Models
Figure 3 for RouterDC: Query-Based Router by Dual Contrastive Learning for Assembling Large Language Models
Figure 4 for RouterDC: Query-Based Router by Dual Contrastive Learning for Assembling Large Language Models
Viaarxiv icon

EMOVA: Empowering Language Models to See, Hear and Speak with Vivid Emotions

Add code
Sep 26, 2024
Figure 1 for EMOVA: Empowering Language Models to See, Hear and Speak with Vivid Emotions
Figure 2 for EMOVA: Empowering Language Models to See, Hear and Speak with Vivid Emotions
Figure 3 for EMOVA: Empowering Language Models to See, Hear and Speak with Vivid Emotions
Figure 4 for EMOVA: Empowering Language Models to See, Hear and Speak with Vivid Emotions
Viaarxiv icon

Efficient Pareto Manifold Learning with Low-Rank Structure

Add code
Jul 30, 2024
Viaarxiv icon

Scalable Learned Model Soup on a Single GPU: An Efficient Subspace Training Strategy

Add code
Jul 04, 2024
Figure 1 for Scalable Learned Model Soup on a Single GPU: An Efficient Subspace Training Strategy
Figure 2 for Scalable Learned Model Soup on a Single GPU: An Efficient Subspace Training Strategy
Figure 3 for Scalable Learned Model Soup on a Single GPU: An Efficient Subspace Training Strategy
Figure 4 for Scalable Learned Model Soup on a Single GPU: An Efficient Subspace Training Strategy
Viaarxiv icon

Mixup Augmentation with Multiple Interpolations

Add code
Jun 03, 2024
Figure 1 for Mixup Augmentation with Multiple Interpolations
Figure 2 for Mixup Augmentation with Multiple Interpolations
Figure 3 for Mixup Augmentation with Multiple Interpolations
Figure 4 for Mixup Augmentation with Multiple Interpolations
Viaarxiv icon

Direct Alignment of Language Models via Quality-Aware Self-Refinement

Add code
May 31, 2024
Viaarxiv icon

Mixture of insighTful Experts : The Synergy of Thought Chains and Expert Mixtures in Self-Alignment

Add code
May 01, 2024
Figure 1 for Mixture of insighTful Experts : The Synergy of Thought Chains and Expert Mixtures in Self-Alignment
Figure 2 for Mixture of insighTful Experts : The Synergy of Thought Chains and Expert Mixtures in Self-Alignment
Figure 3 for Mixture of insighTful Experts : The Synergy of Thought Chains and Expert Mixtures in Self-Alignment
Figure 4 for Mixture of insighTful Experts : The Synergy of Thought Chains and Expert Mixtures in Self-Alignment
Viaarxiv icon

Eyes Closed, Safety On: Protecting Multimodal LLMs via Image-to-Text Transformation

Add code
Mar 22, 2024
Figure 1 for Eyes Closed, Safety On: Protecting Multimodal LLMs via Image-to-Text Transformation
Figure 2 for Eyes Closed, Safety On: Protecting Multimodal LLMs via Image-to-Text Transformation
Figure 3 for Eyes Closed, Safety On: Protecting Multimodal LLMs via Image-to-Text Transformation
Figure 4 for Eyes Closed, Safety On: Protecting Multimodal LLMs via Image-to-Text Transformation
Viaarxiv icon