Picture for Ji Qi

Ji Qi

p-MoD: Building Mixture-of-Depths MLLMs via Progressive Ratio Decay

Add code
Dec 05, 2024
Figure 1 for p-MoD: Building Mixture-of-Depths MLLMs via Progressive Ratio Decay
Figure 2 for p-MoD: Building Mixture-of-Depths MLLMs via Progressive Ratio Decay
Figure 3 for p-MoD: Building Mixture-of-Depths MLLMs via Progressive Ratio Decay
Figure 4 for p-MoD: Building Mixture-of-Depths MLLMs via Progressive Ratio Decay
Viaarxiv icon

Weakly Supervised Framework Considering Multi-temporal Information for Large-scale Cropland Mapping with Satellite Imagery

Add code
Nov 27, 2024
Viaarxiv icon

Class-RAG: Content Moderation with Retrieval Augmented Generation

Add code
Oct 18, 2024
Viaarxiv icon

ExpLLM: Towards Chain of Thought for Facial Expression Recognition

Add code
Sep 04, 2024
Viaarxiv icon

CogVLM2: Visual Language Models for Image and Video Understanding

Add code
Aug 29, 2024
Figure 1 for CogVLM2: Visual Language Models for Image and Video Understanding
Figure 2 for CogVLM2: Visual Language Models for Image and Video Understanding
Figure 3 for CogVLM2: Visual Language Models for Image and Video Understanding
Figure 4 for CogVLM2: Visual Language Models for Image and Video Understanding
Viaarxiv icon

Exploring The Neural Burden In Pruned Models: An Insight Inspired By Neuroscience

Add code
Jul 27, 2024
Viaarxiv icon

LVBench: An Extreme Long Video Understanding Benchmark

Add code
Jun 12, 2024
Figure 1 for LVBench: An Extreme Long Video Understanding Benchmark
Figure 2 for LVBench: An Extreme Long Video Understanding Benchmark
Figure 3 for LVBench: An Extreme Long Video Understanding Benchmark
Figure 4 for LVBench: An Extreme Long Video Understanding Benchmark
Viaarxiv icon

An Empirical Study of Data Ability Boundary in LLMs' Math Reasoning

Add code
Feb 23, 2024
Viaarxiv icon

CogCoM: Train Large Vision-Language Models Diving into Details through Chain of Manipulations

Add code
Feb 06, 2024
Figure 1 for CogCoM: Train Large Vision-Language Models Diving into Details through Chain of Manipulations
Figure 2 for CogCoM: Train Large Vision-Language Models Diving into Details through Chain of Manipulations
Figure 3 for CogCoM: Train Large Vision-Language Models Diving into Details through Chain of Manipulations
Figure 4 for CogCoM: Train Large Vision-Language Models Diving into Details through Chain of Manipulations
Viaarxiv icon

LongAlign: A Recipe for Long Context Alignment of Large Language Models

Add code
Jan 31, 2024
Figure 1 for LongAlign: A Recipe for Long Context Alignment of Large Language Models
Figure 2 for LongAlign: A Recipe for Long Context Alignment of Large Language Models
Figure 3 for LongAlign: A Recipe for Long Context Alignment of Large Language Models
Figure 4 for LongAlign: A Recipe for Long Context Alignment of Large Language Models
Viaarxiv icon