Picture for Peng Wang

Peng Wang

Understanding Representation Dynamics of Diffusion Models via Low-Dimensional Modeling

Add code
Feb 09, 2025
Viaarxiv icon

Demystifying Catastrophic Forgetting in Two-Stage Incremental Object Detector

Add code
Feb 08, 2025
Viaarxiv icon

GP-GS: Gaussian Processes for Enhanced Gaussian Splatting

Add code
Feb 05, 2025
Viaarxiv icon

SceneVTG++: Controllable Multilingual Visual Text Generation in the Wild

Add code
Jan 07, 2025
Figure 1 for SceneVTG++: Controllable Multilingual Visual Text Generation in the Wild
Figure 2 for SceneVTG++: Controllable Multilingual Visual Text Generation in the Wild
Figure 3 for SceneVTG++: Controllable Multilingual Visual Text Generation in the Wild
Figure 4 for SceneVTG++: Controllable Multilingual Visual Text Generation in the Wild
Viaarxiv icon

Understanding How Nonlinear Layers Create Linearly Separable Features for Low-Dimensional Data

Add code
Jan 04, 2025
Figure 1 for Understanding How Nonlinear Layers Create Linearly Separable Features for Low-Dimensional Data
Figure 2 for Understanding How Nonlinear Layers Create Linearly Separable Features for Low-Dimensional Data
Figure 3 for Understanding How Nonlinear Layers Create Linearly Separable Features for Low-Dimensional Data
Figure 4 for Understanding How Nonlinear Layers Create Linearly Separable Features for Low-Dimensional Data
Viaarxiv icon

Dual Diffusion for Unified Image Generation and Understanding

Add code
Dec 31, 2024
Viaarxiv icon

TL-Training: A Task-Feature-Based Framework for Training Large Language Models in Tool Use

Add code
Dec 20, 2024
Figure 1 for TL-Training: A Task-Feature-Based Framework for Training Large Language Models in Tool Use
Figure 2 for TL-Training: A Task-Feature-Based Framework for Training Large Language Models in Tool Use
Figure 3 for TL-Training: A Task-Feature-Based Framework for Training Large Language Models in Tool Use
Figure 4 for TL-Training: A Task-Feature-Based Framework for Training Large Language Models in Tool Use
Viaarxiv icon

Explaining and Mitigating the Modality Gap in Contrastive Multimodal Learning

Add code
Dec 10, 2024
Figure 1 for Explaining and Mitigating the Modality Gap in Contrastive Multimodal Learning
Figure 2 for Explaining and Mitigating the Modality Gap in Contrastive Multimodal Learning
Figure 3 for Explaining and Mitigating the Modality Gap in Contrastive Multimodal Learning
Figure 4 for Explaining and Mitigating the Modality Gap in Contrastive Multimodal Learning
Viaarxiv icon

Pruning All-Rounder: Rethinking and Improving Inference Efficiency for Large Vision Language Models

Add code
Dec 09, 2024
Figure 1 for Pruning All-Rounder: Rethinking and Improving Inference Efficiency for Large Vision Language Models
Figure 2 for Pruning All-Rounder: Rethinking and Improving Inference Efficiency for Large Vision Language Models
Figure 3 for Pruning All-Rounder: Rethinking and Improving Inference Efficiency for Large Vision Language Models
Figure 4 for Pruning All-Rounder: Rethinking and Improving Inference Efficiency for Large Vision Language Models
Viaarxiv icon

Building a Family of Data Augmentation Models for Low-cost LLM Fine-tuning on the Cloud

Add code
Dec 06, 2024
Figure 1 for Building a Family of Data Augmentation Models for Low-cost LLM Fine-tuning on the Cloud
Figure 2 for Building a Family of Data Augmentation Models for Low-cost LLM Fine-tuning on the Cloud
Figure 3 for Building a Family of Data Augmentation Models for Low-cost LLM Fine-tuning on the Cloud
Figure 4 for Building a Family of Data Augmentation Models for Low-cost LLM Fine-tuning on the Cloud
Viaarxiv icon