Picture for James T. Kwok

James T. Kwok

RouterDC: Query-Based Router by Dual Contrastive Learning for Assembling Large Language Models

Add code
Sep 30, 2024
Viaarxiv icon

EMOVA: Empowering Language Models to See, Hear and Speak with Vivid Emotions

Add code
Sep 26, 2024
Figure 1 for EMOVA: Empowering Language Models to See, Hear and Speak with Vivid Emotions
Figure 2 for EMOVA: Empowering Language Models to See, Hear and Speak with Vivid Emotions
Figure 3 for EMOVA: Empowering Language Models to See, Hear and Speak with Vivid Emotions
Figure 4 for EMOVA: Empowering Language Models to See, Hear and Speak with Vivid Emotions
Viaarxiv icon

Efficient Pareto Manifold Learning with Low-Rank Structure

Add code
Jul 30, 2024
Viaarxiv icon

Scalable Learned Model Soup on a Single GPU: An Efficient Subspace Training Strategy

Add code
Jul 04, 2024
Viaarxiv icon

Mixup Augmentation with Multiple Interpolations

Add code
Jun 03, 2024
Figure 1 for Mixup Augmentation with Multiple Interpolations
Figure 2 for Mixup Augmentation with Multiple Interpolations
Figure 3 for Mixup Augmentation with Multiple Interpolations
Figure 4 for Mixup Augmentation with Multiple Interpolations
Viaarxiv icon

Direct Alignment of Language Models via Quality-Aware Self-Refinement

Add code
May 31, 2024
Viaarxiv icon

Mixture of insighTful Experts : The Synergy of Thought Chains and Expert Mixtures in Self-Alignment

Add code
May 01, 2024
Figure 1 for Mixture of insighTful Experts : The Synergy of Thought Chains and Expert Mixtures in Self-Alignment
Figure 2 for Mixture of insighTful Experts : The Synergy of Thought Chains and Expert Mixtures in Self-Alignment
Figure 3 for Mixture of insighTful Experts : The Synergy of Thought Chains and Expert Mixtures in Self-Alignment
Figure 4 for Mixture of insighTful Experts : The Synergy of Thought Chains and Expert Mixtures in Self-Alignment
Viaarxiv icon

Eyes Closed, Safety On: Protecting Multimodal LLMs via Image-to-Text Transformation

Add code
Mar 22, 2024
Figure 1 for Eyes Closed, Safety On: Protecting Multimodal LLMs via Image-to-Text Transformation
Figure 2 for Eyes Closed, Safety On: Protecting Multimodal LLMs via Image-to-Text Transformation
Figure 3 for Eyes Closed, Safety On: Protecting Multimodal LLMs via Image-to-Text Transformation
Figure 4 for Eyes Closed, Safety On: Protecting Multimodal LLMs via Image-to-Text Transformation
Viaarxiv icon

Task-customized Masked AutoEncoder via Mixture of Cluster-conditional Experts

Add code
Feb 08, 2024
Viaarxiv icon

KICGPT: Large Language Model with Knowledge in Context for Knowledge Graph Completion

Add code
Feb 04, 2024
Viaarxiv icon