Picture for Cheng-Yu Hsieh

Cheng-Yu Hsieh

NVILA: Efficient Frontier Visual Language Models

Add code
Dec 05, 2024
Figure 1 for NVILA: Efficient Frontier Visual Language Models
Figure 2 for NVILA: Efficient Frontier Visual Language Models
Figure 3 for NVILA: Efficient Frontier Visual Language Models
Figure 4 for NVILA: Efficient Frontier Visual Language Models
Viaarxiv icon

Perception Tokens Enhance Visual Reasoning in Multimodal Language Models

Add code
Dec 04, 2024
Viaarxiv icon

Is C4 Dataset Optimal for Pruning? An Investigation of Calibration Data for LLM Pruning

Add code
Oct 09, 2024
Figure 1 for Is C4 Dataset Optimal for Pruning? An Investigation of Calibration Data for LLM Pruning
Figure 2 for Is C4 Dataset Optimal for Pruning? An Investigation of Calibration Data for LLM Pruning
Figure 3 for Is C4 Dataset Optimal for Pruning? An Investigation of Calibration Data for LLM Pruning
Figure 4 for Is C4 Dataset Optimal for Pruning? An Investigation of Calibration Data for LLM Pruning
Viaarxiv icon

The Hard Positive Truth about Vision-Language Compositionality

Add code
Sep 26, 2024
Viaarxiv icon

Graph-Based Captioning: Enhancing Visual Descriptions by Interconnecting Region Captions

Add code
Jul 09, 2024
Figure 1 for Graph-Based Captioning: Enhancing Visual Descriptions by Interconnecting Region Captions
Figure 2 for Graph-Based Captioning: Enhancing Visual Descriptions by Interconnecting Region Captions
Figure 3 for Graph-Based Captioning: Enhancing Visual Descriptions by Interconnecting Region Captions
Figure 4 for Graph-Based Captioning: Enhancing Visual Descriptions by Interconnecting Region Captions
Viaarxiv icon

Lookback Lens: Detecting and Mitigating Contextual Hallucinations in Large Language Models Using Only Attention Maps

Add code
Jul 09, 2024
Viaarxiv icon

Found in the Middle: Calibrating Positional Attention Bias Improves Long Context Utilization

Add code
Jun 23, 2024
Figure 1 for Found in the Middle: Calibrating Positional Attention Bias Improves Long Context Utilization
Figure 2 for Found in the Middle: Calibrating Positional Attention Bias Improves Long Context Utilization
Figure 3 for Found in the Middle: Calibrating Positional Attention Bias Improves Long Context Utilization
Figure 4 for Found in the Middle: Calibrating Positional Attention Bias Improves Long Context Utilization
Viaarxiv icon

DataComp-LM: In search of the next generation of training sets for language models

Add code
Jun 18, 2024
Figure 1 for DataComp-LM: In search of the next generation of training sets for language models
Figure 2 for DataComp-LM: In search of the next generation of training sets for language models
Figure 3 for DataComp-LM: In search of the next generation of training sets for language models
Figure 4 for DataComp-LM: In search of the next generation of training sets for language models
Viaarxiv icon

The Unmet Promise of Synthetic Training Images: Using Retrieved Real Images Performs Better

Add code
Jun 07, 2024
Viaarxiv icon

Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity

Add code
Oct 08, 2023
Viaarxiv icon