Picture for Fangyun Wei

Fangyun Wei

Animate Any Character in Any World

Add code
Dec 18, 2025
Viaarxiv icon

Spatia: Video Generation with Updatable Spatial Memory

Add code
Dec 17, 2025
Figure 1 for Spatia: Video Generation with Updatable Spatial Memory
Figure 2 for Spatia: Video Generation with Updatable Spatial Memory
Figure 3 for Spatia: Video Generation with Updatable Spatial Memory
Figure 4 for Spatia: Video Generation with Updatable Spatial Memory
Viaarxiv icon

From Virtual Games to Real-World Play

Add code
Jun 23, 2025
Figure 1 for From Virtual Games to Real-World Play
Figure 2 for From Virtual Games to Real-World Play
Figure 3 for From Virtual Games to Real-World Play
Figure 4 for From Virtual Games to Real-World Play
Viaarxiv icon

Fast Autoregressive Models for Continuous Latent Generation

Add code
Apr 24, 2025
Viaarxiv icon

EAGLE-3: Scaling up Inference Acceleration of Large Language Models via Training-Time Test

Add code
Mar 03, 2025
Viaarxiv icon

Modelling Multi-modal Cross-interaction for ML-FSIC Based on Local Feature Selection

Add code
Dec 18, 2024
Figure 1 for Modelling Multi-modal Cross-interaction for ML-FSIC Based on Local Feature Selection
Figure 2 for Modelling Multi-modal Cross-interaction for ML-FSIC Based on Local Feature Selection
Figure 3 for Modelling Multi-modal Cross-interaction for ML-FSIC Based on Local Feature Selection
Figure 4 for Modelling Multi-modal Cross-interaction for ML-FSIC Based on Local Feature Selection
Viaarxiv icon

UniGraspTransformer: Simplified Policy Distillation for Scalable Dexterous Robotic Grasping

Add code
Dec 03, 2024
Figure 1 for UniGraspTransformer: Simplified Policy Distillation for Scalable Dexterous Robotic Grasping
Figure 2 for UniGraspTransformer: Simplified Policy Distillation for Scalable Dexterous Robotic Grasping
Figure 3 for UniGraspTransformer: Simplified Policy Distillation for Scalable Dexterous Robotic Grasping
Figure 4 for UniGraspTransformer: Simplified Policy Distillation for Scalable Dexterous Robotic Grasping
Viaarxiv icon

CogACT: A Foundational Vision-Language-Action Model for Synergizing Cognition and Action in Robotic Manipulation

Add code
Nov 29, 2024
Figure 1 for CogACT: A Foundational Vision-Language-Action Model for Synergizing Cognition and Action in Robotic Manipulation
Figure 2 for CogACT: A Foundational Vision-Language-Action Model for Synergizing Cognition and Action in Robotic Manipulation
Figure 3 for CogACT: A Foundational Vision-Language-Action Model for Synergizing Cognition and Action in Robotic Manipulation
Figure 4 for CogACT: A Foundational Vision-Language-Action Model for Synergizing Cognition and Action in Robotic Manipulation
Viaarxiv icon

Expressive and Generalizable Low-rank Adaptation for Large Models via Slow Cascaded Learning

Add code
Jul 01, 2024
Figure 1 for Expressive and Generalizable Low-rank Adaptation for Large Models via Slow Cascaded Learning
Figure 2 for Expressive and Generalizable Low-rank Adaptation for Large Models via Slow Cascaded Learning
Figure 3 for Expressive and Generalizable Low-rank Adaptation for Large Models via Slow Cascaded Learning
Figure 4 for Expressive and Generalizable Low-rank Adaptation for Large Models via Slow Cascaded Learning
Viaarxiv icon

EAGLE-2: Faster Inference of Language Models with Dynamic Draft Trees

Add code
Jun 24, 2024
Viaarxiv icon