Picture for Xinsong Zhang

Xinsong Zhang

Power Line Aerial Image Restoration under dverse Weather: Datasets and Baselines

Add code
Sep 07, 2024
Figure 1 for Power Line Aerial Image Restoration under dverse Weather: Datasets and Baselines
Figure 2 for Power Line Aerial Image Restoration under dverse Weather: Datasets and Baselines
Figure 3 for Power Line Aerial Image Restoration under dverse Weather: Datasets and Baselines
Figure 4 for Power Line Aerial Image Restoration under dverse Weather: Datasets and Baselines
Viaarxiv icon

Toward Building General Foundation Models for Language, Vision, and Vision-Language Understanding Tasks

Add code
Jan 12, 2023
Viaarxiv icon

X$^2$-VLM: All-In-One Pre-trained Model For Vision-Language Tasks

Add code
Nov 22, 2022
Viaarxiv icon

EfficientVLM: Fast and Accurate Vision-Language Models via Knowledge Distillation and Modal-adaptive Pruning

Add code
Oct 14, 2022
Figure 1 for EfficientVLM: Fast and Accurate Vision-Language Models via Knowledge Distillation and Modal-adaptive Pruning
Figure 2 for EfficientVLM: Fast and Accurate Vision-Language Models via Knowledge Distillation and Modal-adaptive Pruning
Figure 3 for EfficientVLM: Fast and Accurate Vision-Language Models via Knowledge Distillation and Modal-adaptive Pruning
Figure 4 for EfficientVLM: Fast and Accurate Vision-Language Models via Knowledge Distillation and Modal-adaptive Pruning
Viaarxiv icon

Prefix Language Models are Unified Modal Learners

Add code
Jun 15, 2022
Figure 1 for Prefix Language Models are Unified Modal Learners
Figure 2 for Prefix Language Models are Unified Modal Learners
Figure 3 for Prefix Language Models are Unified Modal Learners
Figure 4 for Prefix Language Models are Unified Modal Learners
Viaarxiv icon

Cross-View Language Modeling: Towards Unified Cross-Lingual Cross-Modal Pre-training

Add code
Jun 01, 2022
Figure 1 for Cross-View Language Modeling: Towards Unified Cross-Lingual Cross-Modal Pre-training
Figure 2 for Cross-View Language Modeling: Towards Unified Cross-Lingual Cross-Modal Pre-training
Figure 3 for Cross-View Language Modeling: Towards Unified Cross-Lingual Cross-Modal Pre-training
Figure 4 for Cross-View Language Modeling: Towards Unified Cross-Lingual Cross-Modal Pre-training
Viaarxiv icon

VLUE: A Multi-Task Benchmark for Evaluating Vision-Language Models

Add code
May 30, 2022
Figure 1 for VLUE: A Multi-Task Benchmark for Evaluating Vision-Language Models
Figure 2 for VLUE: A Multi-Task Benchmark for Evaluating Vision-Language Models
Figure 3 for VLUE: A Multi-Task Benchmark for Evaluating Vision-Language Models
Figure 4 for VLUE: A Multi-Task Benchmark for Evaluating Vision-Language Models
Viaarxiv icon

Multi-Grained Vision Language Pre-Training: Aligning Texts with Visual Concepts

Add code
Nov 16, 2021
Figure 1 for Multi-Grained Vision Language Pre-Training: Aligning Texts with Visual Concepts
Figure 2 for Multi-Grained Vision Language Pre-Training: Aligning Texts with Visual Concepts
Figure 3 for Multi-Grained Vision Language Pre-Training: Aligning Texts with Visual Concepts
Figure 4 for Multi-Grained Vision Language Pre-Training: Aligning Texts with Visual Concepts
Viaarxiv icon

Active Testing: An Unbiased Evaluation Method for Distantly Supervised Relation Extraction

Add code
Oct 17, 2020
Figure 1 for Active Testing: An Unbiased Evaluation Method for Distantly Supervised Relation Extraction
Figure 2 for Active Testing: An Unbiased Evaluation Method for Distantly Supervised Relation Extraction
Figure 3 for Active Testing: An Unbiased Evaluation Method for Distantly Supervised Relation Extraction
Figure 4 for Active Testing: An Unbiased Evaluation Method for Distantly Supervised Relation Extraction
Viaarxiv icon

AMBERT: A Pre-trained Language Model with Multi-Grained Tokenization

Add code
Sep 01, 2020
Figure 1 for AMBERT: A Pre-trained Language Model with Multi-Grained Tokenization
Figure 2 for AMBERT: A Pre-trained Language Model with Multi-Grained Tokenization
Figure 3 for AMBERT: A Pre-trained Language Model with Multi-Grained Tokenization
Figure 4 for AMBERT: A Pre-trained Language Model with Multi-Grained Tokenization
Viaarxiv icon