Picture for Xuan Shen

Xuan Shen

Fully Open Source Moxin-7B Technical Report

Add code
Dec 08, 2024
Viaarxiv icon

A Survey of Small Language Models

Add code
Oct 25, 2024
Figure 1 for A Survey of Small Language Models
Figure 2 for A Survey of Small Language Models
Figure 3 for A Survey of Small Language Models
Viaarxiv icon

Pruning Foundation Models for High Accuracy without Retraining

Add code
Oct 21, 2024
Figure 1 for Pruning Foundation Models for High Accuracy without Retraining
Figure 2 for Pruning Foundation Models for High Accuracy without Retraining
Figure 3 for Pruning Foundation Models for High Accuracy without Retraining
Viaarxiv icon

Rethinking Token Reduction for State Space Models

Add code
Oct 16, 2024
Figure 1 for Rethinking Token Reduction for State Space Models
Figure 2 for Rethinking Token Reduction for State Space Models
Figure 3 for Rethinking Token Reduction for State Space Models
Figure 4 for Rethinking Token Reduction for State Space Models
Viaarxiv icon

Exploring Token Pruning in Vision State Space Models

Add code
Sep 27, 2024
Figure 1 for Exploring Token Pruning in Vision State Space Models
Figure 2 for Exploring Token Pruning in Vision State Space Models
Figure 3 for Exploring Token Pruning in Vision State Space Models
Figure 4 for Exploring Token Pruning in Vision State Space Models
Viaarxiv icon

Search for Efficient Large Language Models

Add code
Sep 25, 2024
Figure 1 for Search for Efficient Large Language Models
Figure 2 for Search for Efficient Large Language Models
Figure 3 for Search for Efficient Large Language Models
Figure 4 for Search for Efficient Large Language Models
Viaarxiv icon

EdgeQAT: Entropy and Distribution Guided Quantization-Aware Training for the Acceleration of Lightweight LLMs on the Edge

Add code
Feb 16, 2024
Viaarxiv icon

Agile-Quant: Activation-Guided Quantization for Faster Inference of LLMs on the Edge

Add code
Dec 09, 2023
Viaarxiv icon

DeepMAD: Mathematical Architecture Design for Deep Convolutional Neural Network

Add code
Mar 05, 2023
Viaarxiv icon

Peeling the Onion: Hierarchical Reduction of Data Redundancy for Efficient Vision Transformer Training

Add code
Nov 19, 2022
Figure 1 for Peeling the Onion: Hierarchical Reduction of Data Redundancy for Efficient Vision Transformer Training
Figure 2 for Peeling the Onion: Hierarchical Reduction of Data Redundancy for Efficient Vision Transformer Training
Figure 3 for Peeling the Onion: Hierarchical Reduction of Data Redundancy for Efficient Vision Transformer Training
Figure 4 for Peeling the Onion: Hierarchical Reduction of Data Redundancy for Efficient Vision Transformer Training
Viaarxiv icon