Picture for Haokui Zhang

Haokui Zhang

Efficient Adaptation of Pre-trained Vision Transformer via Householder Transformation

Add code
Oct 30, 2024
Viaarxiv icon

Advancements in Visual Language Models for Remote Sensing: Datasets, Capabilities, and Enhancement Techniques

Add code
Oct 15, 2024
Figure 1 for Advancements in Visual Language Models for Remote Sensing: Datasets, Capabilities, and Enhancement Techniques
Figure 2 for Advancements in Visual Language Models for Remote Sensing: Datasets, Capabilities, and Enhancement Techniques
Figure 3 for Advancements in Visual Language Models for Remote Sensing: Datasets, Capabilities, and Enhancement Techniques
Figure 4 for Advancements in Visual Language Models for Remote Sensing: Datasets, Capabilities, and Enhancement Techniques
Viaarxiv icon

TG-LLaVA: Text Guided LLaVA via Learnable Latent Embeddings

Add code
Sep 15, 2024
Viaarxiv icon

3D-RCNet: Learning from Transformer to Build a 3D Relational ConvNet for Hyperspectral Image Classification

Add code
Aug 25, 2024
Viaarxiv icon

OVA-DETR: Open Vocabulary Aerial Object Detection Using Image-Text Alignment and Fusion

Add code
Aug 22, 2024
Figure 1 for OVA-DETR: Open Vocabulary Aerial Object Detection Using Image-Text Alignment and Fusion
Figure 2 for OVA-DETR: Open Vocabulary Aerial Object Detection Using Image-Text Alignment and Fusion
Figure 3 for OVA-DETR: Open Vocabulary Aerial Object Detection Using Image-Text Alignment and Fusion
Figure 4 for OVA-DETR: Open Vocabulary Aerial Object Detection Using Image-Text Alignment and Fusion
Viaarxiv icon

Bridging Sensor Gaps via Single-Direction Tuning for Hyperspectral Image Classification

Add code
Sep 22, 2023
Viaarxiv icon

NAR-Former V2: Rethinking Transformer for Universal Neural Network Representation Learning

Add code
Jun 19, 2023
Figure 1 for NAR-Former V2: Rethinking Transformer for Universal Neural Network Representation Learning
Figure 2 for NAR-Former V2: Rethinking Transformer for Universal Neural Network Representation Learning
Figure 3 for NAR-Former V2: Rethinking Transformer for Universal Neural Network Representation Learning
Figure 4 for NAR-Former V2: Rethinking Transformer for Universal Neural Network Representation Learning
Viaarxiv icon

Teacher Agent: A Non-Knowledge Distillation Method for Rehearsal-based Video Incremental Learning

Add code
Jun 01, 2023
Viaarxiv icon

NAR-Former: Neural Architecture Representation Learning towards Holistic Attributes Prediction

Add code
Nov 15, 2022
Viaarxiv icon

CabViT: Cross Attention among Blocks for Vision Transformer

Add code
Nov 14, 2022
Viaarxiv icon