Picture for Quanlu Zhang

Quanlu Zhang

$π_\texttt{RL}$: Online RL Fine-tuning for Flow-based Vision-Language-Action Models

Add code
Oct 29, 2025
Viaarxiv icon

RLinf: Flexible and Efficient Large-scale Reinforcement Learning via Macro-to-Micro Flow Transformation

Add code
Sep 19, 2025
Figure 1 for RLinf: Flexible and Efficient Large-scale Reinforcement Learning via Macro-to-Micro Flow Transformation
Figure 2 for RLinf: Flexible and Efficient Large-scale Reinforcement Learning via Macro-to-Micro Flow Transformation
Figure 3 for RLinf: Flexible and Efficient Large-scale Reinforcement Learning via Macro-to-Micro Flow Transformation
Figure 4 for RLinf: Flexible and Efficient Large-scale Reinforcement Learning via Macro-to-Micro Flow Transformation
Viaarxiv icon

You Only Cache Once: Decoder-Decoder Architectures for Language Models

Add code
May 08, 2024
Viaarxiv icon

Efficient Large Language Models: A Survey

Add code
Dec 23, 2023
Figure 1 for Efficient Large Language Models: A Survey
Figure 2 for Efficient Large Language Models: A Survey
Figure 3 for Efficient Large Language Models: A Survey
Figure 4 for Efficient Large Language Models: A Survey
Viaarxiv icon

AutoTaskFormer: Searching Vision Transformers for Multi-task Learning

Add code
Apr 20, 2023
Viaarxiv icon

ElasticViT: Conflict-aware Supernet Training for Deploying Fast Vision Transformer on Diverse Mobile Devices

Add code
Mar 21, 2023
Figure 1 for ElasticViT: Conflict-aware Supernet Training for Deploying Fast Vision Transformer on Diverse Mobile Devices
Figure 2 for ElasticViT: Conflict-aware Supernet Training for Deploying Fast Vision Transformer on Diverse Mobile Devices
Figure 3 for ElasticViT: Conflict-aware Supernet Training for Deploying Fast Vision Transformer on Diverse Mobile Devices
Figure 4 for ElasticViT: Conflict-aware Supernet Training for Deploying Fast Vision Transformer on Diverse Mobile Devices
Viaarxiv icon

SpaceEvo: Hardware-Friendly Search Space Design for Efficient INT8 Inference

Add code
Mar 15, 2023
Viaarxiv icon

SparDA: Accelerating Dynamic Sparse Deep Neural Networks via Sparse-Dense Transformation

Add code
Jan 26, 2023
Figure 1 for SparDA: Accelerating Dynamic Sparse Deep Neural Networks via Sparse-Dense Transformation
Figure 2 for SparDA: Accelerating Dynamic Sparse Deep Neural Networks via Sparse-Dense Transformation
Figure 3 for SparDA: Accelerating Dynamic Sparse Deep Neural Networks via Sparse-Dense Transformation
Figure 4 for SparDA: Accelerating Dynamic Sparse Deep Neural Networks via Sparse-Dense Transformation
Viaarxiv icon

SuperScaler: Supporting Flexible DNN Parallelization via a Unified Abstraction

Add code
Jan 21, 2023
Figure 1 for SuperScaler: Supporting Flexible DNN Parallelization via a Unified Abstraction
Figure 2 for SuperScaler: Supporting Flexible DNN Parallelization via a Unified Abstraction
Figure 3 for SuperScaler: Supporting Flexible DNN Parallelization via a Unified Abstraction
Figure 4 for SuperScaler: Supporting Flexible DNN Parallelization via a Unified Abstraction
Viaarxiv icon

Nesting Forward Automatic Differentiation for Memory-Efficient Deep Neural Network Training

Add code
Sep 22, 2022
Figure 1 for Nesting Forward Automatic Differentiation for Memory-Efficient Deep Neural Network Training
Figure 2 for Nesting Forward Automatic Differentiation for Memory-Efficient Deep Neural Network Training
Figure 3 for Nesting Forward Automatic Differentiation for Memory-Efficient Deep Neural Network Training
Figure 4 for Nesting Forward Automatic Differentiation for Memory-Efficient Deep Neural Network Training
Viaarxiv icon