Picture for Xiong Zhou

Xiong Zhou

Train Less, Learn More: Adaptive Efficient Rollout Optimization for Group-Based Reinforcement Learning

Add code
Feb 15, 2026
Viaarxiv icon

When LLMs get significantly worse: A statistical approach to detect model degradations

Add code
Feb 09, 2026
Viaarxiv icon

Variation-Bounded Loss for Noise-Tolerant Learning

Add code
Nov 15, 2025
Viaarxiv icon

Joint Asymmetric Loss for Learning with Noisy Labels

Add code
Jul 23, 2025
Viaarxiv icon

Proposer-Agent-Evaluator(PAE): Autonomous Skill Discovery For Foundation Model Internet Agents

Add code
Dec 17, 2024
Figure 1 for Proposer-Agent-Evaluator(PAE): Autonomous Skill Discovery For Foundation Model Internet Agents
Figure 2 for Proposer-Agent-Evaluator(PAE): Autonomous Skill Discovery For Foundation Model Internet Agents
Figure 3 for Proposer-Agent-Evaluator(PAE): Autonomous Skill Discovery For Foundation Model Internet Agents
Figure 4 for Proposer-Agent-Evaluator(PAE): Autonomous Skill Discovery For Foundation Model Internet Agents
Viaarxiv icon

Neural Field Classifiers via Target Encoding and Classification Loss

Add code
Mar 02, 2024
Figure 1 for Neural Field Classifiers via Target Encoding and Classification Loss
Figure 2 for Neural Field Classifiers via Target Encoding and Classification Loss
Figure 3 for Neural Field Classifiers via Target Encoding and Classification Loss
Figure 4 for Neural Field Classifiers via Target Encoding and Classification Loss
Viaarxiv icon

ViGoR: Improving Visual Grounding of Large Vision Language Models with Fine-Grained Reward Modeling

Add code
Feb 09, 2024
Figure 1 for ViGoR: Improving Visual Grounding of Large Vision Language Models with Fine-Grained Reward Modeling
Figure 2 for ViGoR: Improving Visual Grounding of Large Vision Language Models with Fine-Grained Reward Modeling
Figure 3 for ViGoR: Improving Visual Grounding of Large Vision Language Models with Fine-Grained Reward Modeling
Figure 4 for ViGoR: Improving Visual Grounding of Large Vision Language Models with Fine-Grained Reward Modeling
Viaarxiv icon

AffordanceLLM: Grounding Affordance from Vision Language Models

Add code
Jan 12, 2024
Figure 1 for AffordanceLLM: Grounding Affordance from Vision Language Models
Figure 2 for AffordanceLLM: Grounding Affordance from Vision Language Models
Figure 3 for AffordanceLLM: Grounding Affordance from Vision Language Models
Figure 4 for AffordanceLLM: Grounding Affordance from Vision Language Models
Viaarxiv icon

On the Dynamics Under the Unhinged Loss and Beyond

Add code
Dec 13, 2023
Figure 1 for On the Dynamics Under the Unhinged Loss and Beyond
Figure 2 for On the Dynamics Under the Unhinged Loss and Beyond
Figure 3 for On the Dynamics Under the Unhinged Loss and Beyond
Figure 4 for On the Dynamics Under the Unhinged Loss and Beyond
Viaarxiv icon

Visual Prompt Tuning for Test-time Domain Adaptation

Add code
Oct 10, 2022
Figure 1 for Visual Prompt Tuning for Test-time Domain Adaptation
Figure 2 for Visual Prompt Tuning for Test-time Domain Adaptation
Figure 3 for Visual Prompt Tuning for Test-time Domain Adaptation
Figure 4 for Visual Prompt Tuning for Test-time Domain Adaptation
Viaarxiv icon