Picture for Longxiang Tang

Longxiang Tang

Hybrid-Level Instruction Injection for Video Token Compression in Multi-modal Large Language Models

Add code
Mar 20, 2025
Viaarxiv icon

Does Your Vision-Language Model Get Lost in the Long Video Sampling Dilemma?

Add code
Mar 16, 2025
Viaarxiv icon

DreamRelation: Relation-Centric Video Customization

Add code
Mar 10, 2025
Viaarxiv icon

Gamma: Toward Generic Image Assessment with Mixture of Assessment Experts

Add code
Mar 09, 2025
Viaarxiv icon

Integrating Extra Modality Helps Segmentor Find Camouflaged Objects Well

Add code
Feb 20, 2025
Viaarxiv icon

RUN: Reversible Unfolding Network for Concealed Object Segmentation

Add code
Jan 30, 2025
Figure 1 for RUN: Reversible Unfolding Network for Concealed Object Segmentation
Figure 2 for RUN: Reversible Unfolding Network for Concealed Object Segmentation
Figure 3 for RUN: Reversible Unfolding Network for Concealed Object Segmentation
Figure 4 for RUN: Reversible Unfolding Network for Concealed Object Segmentation
Viaarxiv icon

Lyra: An Efficient and Speech-Centric Framework for Omni-Cognition

Add code
Dec 12, 2024
Figure 1 for Lyra: An Efficient and Speech-Centric Framework for Omni-Cognition
Figure 2 for Lyra: An Efficient and Speech-Centric Framework for Omni-Cognition
Figure 3 for Lyra: An Efficient and Speech-Centric Framework for Omni-Cognition
Figure 4 for Lyra: An Efficient and Speech-Centric Framework for Omni-Cognition
Viaarxiv icon

InstantSwap: Fast Customized Concept Swapping across Sharp Shape Differences

Add code
Dec 03, 2024
Figure 1 for InstantSwap: Fast Customized Concept Swapping across Sharp Shape Differences
Figure 2 for InstantSwap: Fast Customized Concept Swapping across Sharp Shape Differences
Figure 3 for InstantSwap: Fast Customized Concept Swapping across Sharp Shape Differences
Figure 4 for InstantSwap: Fast Customized Concept Swapping across Sharp Shape Differences
Viaarxiv icon

A Survey of Camouflaged Object Detection and Beyond

Add code
Aug 26, 2024
Viaarxiv icon

Mind the Interference: Retaining Pre-trained Knowledge in Parameter Efficient Continual Learning of Vision-Language Models

Add code
Jul 07, 2024
Figure 1 for Mind the Interference: Retaining Pre-trained Knowledge in Parameter Efficient Continual Learning of Vision-Language Models
Viaarxiv icon