Picture for Huaxiu Yao

Huaxiu Yao

FactTest: Factuality Testing in Large Language Models with Statistical Guarantees

Add code
Nov 04, 2024
Viaarxiv icon

Unveiling Context-Aware Criteria in Self-Assessing LLMs

Add code
Oct 28, 2024
Figure 1 for Unveiling Context-Aware Criteria in Self-Assessing LLMs
Figure 2 for Unveiling Context-Aware Criteria in Self-Assessing LLMs
Figure 3 for Unveiling Context-Aware Criteria in Self-Assessing LLMs
Figure 4 for Unveiling Context-Aware Criteria in Self-Assessing LLMs
Viaarxiv icon

Fine-Grained Verifiers: Preference Modeling as Next-token Prediction in Vision-Language Alignment

Add code
Oct 18, 2024
Figure 1 for Fine-Grained Verifiers: Preference Modeling as Next-token Prediction in Vision-Language Alignment
Figure 2 for Fine-Grained Verifiers: Preference Modeling as Next-token Prediction in Vision-Language Alignment
Figure 3 for Fine-Grained Verifiers: Preference Modeling as Next-token Prediction in Vision-Language Alignment
Figure 4 for Fine-Grained Verifiers: Preference Modeling as Next-token Prediction in Vision-Language Alignment
Viaarxiv icon

CREAM: Consistency Regularized Self-Rewarding Language Models

Add code
Oct 17, 2024
Figure 1 for CREAM: Consistency Regularized Self-Rewarding Language Models
Figure 2 for CREAM: Consistency Regularized Self-Rewarding Language Models
Figure 3 for CREAM: Consistency Regularized Self-Rewarding Language Models
Figure 4 for CREAM: Consistency Regularized Self-Rewarding Language Models
Viaarxiv icon

MMed-RAG: Versatile Multimodal RAG System for Medical Vision Language Models

Add code
Oct 16, 2024
Viaarxiv icon

SAFREE: Training-Free and Adaptive Guard for Safe Text-to-Image And Video Generation

Add code
Oct 16, 2024
Figure 1 for SAFREE: Training-Free and Adaptive Guard for Safe Text-to-Image And Video Generation
Figure 2 for SAFREE: Training-Free and Adaptive Guard for Safe Text-to-Image And Video Generation
Figure 3 for SAFREE: Training-Free and Adaptive Guard for Safe Text-to-Image And Video Generation
Figure 4 for SAFREE: Training-Free and Adaptive Guard for Safe Text-to-Image And Video Generation
Viaarxiv icon

MMIE: Massive Multimodal Interleaved Comprehension Benchmark for Large Vision-Language Models

Add code
Oct 14, 2024
Figure 1 for MMIE: Massive Multimodal Interleaved Comprehension Benchmark for Large Vision-Language Models
Figure 2 for MMIE: Massive Multimodal Interleaved Comprehension Benchmark for Large Vision-Language Models
Figure 3 for MMIE: Massive Multimodal Interleaved Comprehension Benchmark for Large Vision-Language Models
Figure 4 for MMIE: Massive Multimodal Interleaved Comprehension Benchmark for Large Vision-Language Models
Viaarxiv icon

VHELM: A Holistic Evaluation of Vision Language Models

Add code
Oct 09, 2024
Figure 1 for VHELM: A Holistic Evaluation of Vision Language Models
Figure 2 for VHELM: A Holistic Evaluation of Vision Language Models
Figure 3 for VHELM: A Holistic Evaluation of Vision Language Models
Figure 4 for VHELM: A Holistic Evaluation of Vision Language Models
Viaarxiv icon

On Unsupervised Prompt Learning for Classification with Black-box Language Models

Add code
Oct 04, 2024
Figure 1 for On Unsupervised Prompt Learning for Classification with Black-box Language Models
Figure 2 for On Unsupervised Prompt Learning for Classification with Black-box Language Models
Figure 3 for On Unsupervised Prompt Learning for Classification with Black-box Language Models
Figure 4 for On Unsupervised Prompt Learning for Classification with Black-box Language Models
Viaarxiv icon

NEAT: Nonlinear Parameter-efficient Adaptation of Pre-trained Models

Add code
Oct 02, 2024
Viaarxiv icon