Picture for Guanhong Wang

Guanhong Wang

User-Aware Prefix-Tuning is a Good Learner for Personalized Image Captioning

Add code
Dec 08, 2023
Viaarxiv icon

Sam-Guided Enhanced Fine-Grained Encoding with Mixed Semantic Learning for Medical Image Captioning

Add code
Nov 02, 2023
Viaarxiv icon

FrameRS: A Video Frame Compression Model Composed by Self supervised Video Frame Reconstructor and Key Frame Selector

Add code
Sep 16, 2023
Viaarxiv icon

MovieChat: From Dense Token to Sparse Memory for Long Video Understanding

Add code
Jul 31, 2023
Figure 1 for MovieChat: From Dense Token to Sparse Memory for Long Video Understanding
Figure 2 for MovieChat: From Dense Token to Sparse Memory for Long Video Understanding
Figure 3 for MovieChat: From Dense Token to Sparse Memory for Long Video Understanding
Figure 4 for MovieChat: From Dense Token to Sparse Memory for Long Video Understanding
Viaarxiv icon

A Survey of Deep Learning in Sports Applications: Perception, Comprehension, and Decision

Add code
Jul 07, 2023
Viaarxiv icon

Answering Private Linear Queries Adaptively using the Common Mechanism

Add code
Nov 30, 2022
Viaarxiv icon

Missing Modality meets Meta Sampling (M3S): An Efficient Universal Approach for Multimodal Sentiment Analysis with Missing Modality

Add code
Oct 07, 2022
Figure 1 for Missing Modality meets Meta Sampling (M3S): An Efficient Universal Approach for Multimodal Sentiment Analysis with Missing Modality
Figure 2 for Missing Modality meets Meta Sampling (M3S): An Efficient Universal Approach for Multimodal Sentiment Analysis with Missing Modality
Figure 3 for Missing Modality meets Meta Sampling (M3S): An Efficient Universal Approach for Multimodal Sentiment Analysis with Missing Modality
Figure 4 for Missing Modality meets Meta Sampling (M3S): An Efficient Universal Approach for Multimodal Sentiment Analysis with Missing Modality
Viaarxiv icon

Preserve Pre-trained Knowledge: Transfer Learning With Self-Distillation For Action Recognition

Add code
May 01, 2022
Figure 1 for Preserve Pre-trained Knowledge: Transfer Learning With Self-Distillation For Action Recognition
Figure 2 for Preserve Pre-trained Knowledge: Transfer Learning With Self-Distillation For Action Recognition
Figure 3 for Preserve Pre-trained Knowledge: Transfer Learning With Self-Distillation For Action Recognition
Viaarxiv icon

Human-Centered Prior-Guided and Task-Dependent Multi-Task Representation Learning for Action Recognition Pre-Training

Add code
Apr 27, 2022
Figure 1 for Human-Centered Prior-Guided and Task-Dependent Multi-Task Representation Learning for Action Recognition Pre-Training
Figure 2 for Human-Centered Prior-Guided and Task-Dependent Multi-Task Representation Learning for Action Recognition Pre-Training
Figure 3 for Human-Centered Prior-Guided and Task-Dependent Multi-Task Representation Learning for Action Recognition Pre-Training
Figure 4 for Human-Centered Prior-Guided and Task-Dependent Multi-Task Representation Learning for Action Recognition Pre-Training
Viaarxiv icon