Picture for Ming Tang

Ming Tang

Zero Token-Driven Deep Thinking in LLMs: Unlocking the Full Potential of Existing Parameters via Cyclic Refinement

Add code
Feb 17, 2025
Viaarxiv icon

Systematic Outliers in Large Language Models

Add code
Feb 10, 2025
Viaarxiv icon

Label Anything: An Interpretable, High-Fidelity and Prompt-Free Annotator

Add code
Feb 05, 2025
Viaarxiv icon

FiLo++: Zero-/Few-Shot Anomaly Detection by Fused Fine-Grained Descriptions and Deformable Localization

Add code
Jan 17, 2025
Viaarxiv icon

Enhancing Large Vision Model in Street Scene Semantic Understanding through Leveraging Posterior Optimization Trajectory

Add code
Jan 03, 2025
Viaarxiv icon

Cracking the Code of Hallucination in LVLMs with Vision-aware Head Divergence

Add code
Dec 18, 2024
Figure 1 for Cracking the Code of Hallucination in LVLMs with Vision-aware Head Divergence
Figure 2 for Cracking the Code of Hallucination in LVLMs with Vision-aware Head Divergence
Figure 3 for Cracking the Code of Hallucination in LVLMs with Vision-aware Head Divergence
Figure 4 for Cracking the Code of Hallucination in LVLMs with Vision-aware Head Divergence
Viaarxiv icon

UniVAD: A Training-free Unified Model for Few-shot Visual Anomaly Detection

Add code
Dec 05, 2024
Figure 1 for UniVAD: A Training-free Unified Model for Few-shot Visual Anomaly Detection
Figure 2 for UniVAD: A Training-free Unified Model for Few-shot Visual Anomaly Detection
Figure 3 for UniVAD: A Training-free Unified Model for Few-shot Visual Anomaly Detection
Figure 4 for UniVAD: A Training-free Unified Model for Few-shot Visual Anomaly Detection
Viaarxiv icon

Friend or Foe? Harnessing Controllable Overfitting for Anomaly Detection

Add code
Nov 30, 2024
Viaarxiv icon

SEEKR: Selective Attention-Guided Knowledge Retention for Continual Learning of Large Language Models

Add code
Nov 09, 2024
Viaarxiv icon

Griffon-G: Bridging Vision-Language and Vision-Centric Tasks via Large Multimodal Models

Add code
Oct 21, 2024
Figure 1 for Griffon-G: Bridging Vision-Language and Vision-Centric Tasks via Large Multimodal Models
Figure 2 for Griffon-G: Bridging Vision-Language and Vision-Centric Tasks via Large Multimodal Models
Figure 3 for Griffon-G: Bridging Vision-Language and Vision-Centric Tasks via Large Multimodal Models
Figure 4 for Griffon-G: Bridging Vision-Language and Vision-Centric Tasks via Large Multimodal Models
Viaarxiv icon