Picture for Haoyuan Li

Haoyuan Li

Coverage-based Fairness in Multi-document Summarization

Add code
Dec 11, 2024
Viaarxiv icon

T2I-FactualBench: Benchmarking the Factuality of Text-to-Image Models with Knowledge-Intensive Concepts

Add code
Dec 05, 2024
Figure 1 for T2I-FactualBench: Benchmarking the Factuality of Text-to-Image Models with Knowledge-Intensive Concepts
Figure 2 for T2I-FactualBench: Benchmarking the Factuality of Text-to-Image Models with Knowledge-Intensive Concepts
Figure 3 for T2I-FactualBench: Benchmarking the Factuality of Text-to-Image Models with Knowledge-Intensive Concepts
Figure 4 for T2I-FactualBench: Benchmarking the Factuality of Text-to-Image Models with Knowledge-Intensive Concepts
Viaarxiv icon

Unsupervised Multi-view UAV Image Geo-localization via Iterative Rendering

Add code
Nov 22, 2024
Viaarxiv icon

Align$^2$LLaVA: Cascaded Human and Large Language Model Preference Alignment for Multi-modal Instruction Curation

Add code
Sep 27, 2024
Figure 1 for Align$^2$LLaVA: Cascaded Human and Large Language Model Preference Alignment for Multi-modal Instruction Curation
Figure 2 for Align$^2$LLaVA: Cascaded Human and Large Language Model Preference Alignment for Multi-modal Instruction Curation
Figure 3 for Align$^2$LLaVA: Cascaded Human and Large Language Model Preference Alignment for Multi-modal Instruction Curation
Figure 4 for Align$^2$LLaVA: Cascaded Human and Large Language Model Preference Alignment for Multi-modal Instruction Curation
Viaarxiv icon

EMOVA: Empowering Language Models to See, Hear and Speak with Vivid Emotions

Add code
Sep 26, 2024
Figure 1 for EMOVA: Empowering Language Models to See, Hear and Speak with Vivid Emotions
Figure 2 for EMOVA: Empowering Language Models to See, Hear and Speak with Vivid Emotions
Figure 3 for EMOVA: Empowering Language Models to See, Hear and Speak with Vivid Emotions
Figure 4 for EMOVA: Empowering Language Models to See, Hear and Speak with Vivid Emotions
Viaarxiv icon

Collaboratively Learning Federated Models from Noisy Decentralized Data

Add code
Sep 03, 2024
Viaarxiv icon

LLaVA-MoD: Making LLaVA Tiny via MoE Knowledge Distillation

Add code
Aug 28, 2024
Viaarxiv icon

NAS-Cap: Deep-Learning Driven 3-D Capacitance Extraction with Neural Architecture Search and Data Augmentation

Add code
Aug 23, 2024
Figure 1 for NAS-Cap: Deep-Learning Driven 3-D Capacitance Extraction with Neural Architecture Search and Data Augmentation
Figure 2 for NAS-Cap: Deep-Learning Driven 3-D Capacitance Extraction with Neural Architecture Search and Data Augmentation
Figure 3 for NAS-Cap: Deep-Learning Driven 3-D Capacitance Extraction with Neural Architecture Search and Data Augmentation
Figure 4 for NAS-Cap: Deep-Learning Driven 3-D Capacitance Extraction with Neural Architecture Search and Data Augmentation
Viaarxiv icon

TeamLoRA: Boosting Low-Rank Adaptation with Expert Collaboration and Competition

Add code
Aug 19, 2024
Figure 1 for TeamLoRA: Boosting Low-Rank Adaptation with Expert Collaboration and Competition
Figure 2 for TeamLoRA: Boosting Low-Rank Adaptation with Expert Collaboration and Competition
Figure 3 for TeamLoRA: Boosting Low-Rank Adaptation with Expert Collaboration and Competition
Figure 4 for TeamLoRA: Boosting Low-Rank Adaptation with Expert Collaboration and Competition
Viaarxiv icon

MARS: Mixture of Auto-Regressive Models for Fine-grained Text-to-image Synthesis

Add code
Jul 11, 2024
Figure 1 for MARS: Mixture of Auto-Regressive Models for Fine-grained Text-to-image Synthesis
Figure 2 for MARS: Mixture of Auto-Regressive Models for Fine-grained Text-to-image Synthesis
Figure 3 for MARS: Mixture of Auto-Regressive Models for Fine-grained Text-to-image Synthesis
Figure 4 for MARS: Mixture of Auto-Regressive Models for Fine-grained Text-to-image Synthesis
Viaarxiv icon