Picture for Chenyang Song

Chenyang Song

Sparsing Law: Towards Large Language Models with Greater Activation Sparsity

Add code
Nov 04, 2024
Figure 1 for Sparsing Law: Towards Large Language Models with Greater Activation Sparsity
Figure 2 for Sparsing Law: Towards Large Language Models with Greater Activation Sparsity
Figure 3 for Sparsing Law: Towards Large Language Models with Greater Activation Sparsity
Figure 4 for Sparsing Law: Towards Large Language Models with Greater Activation Sparsity
Viaarxiv icon

Configurable Foundation Models: Building LLMs from a Modular Perspective

Add code
Sep 04, 2024
Viaarxiv icon

Multi-Modal Multi-Granularity Tokenizer for Chu Bamboo Slip Scripts

Add code
Sep 02, 2024
Figure 1 for Multi-Modal Multi-Granularity Tokenizer for Chu Bamboo Slip Scripts
Figure 2 for Multi-Modal Multi-Granularity Tokenizer for Chu Bamboo Slip Scripts
Figure 3 for Multi-Modal Multi-Granularity Tokenizer for Chu Bamboo Slip Scripts
Figure 4 for Multi-Modal Multi-Granularity Tokenizer for Chu Bamboo Slip Scripts
Viaarxiv icon

ProSparse: Introducing and Enhancing Intrinsic Activation Sparsity within Large Language Models

Add code
Feb 27, 2024
Viaarxiv icon

ReLU$^2$ Wins: Discovering Efficient Activation Functions for Sparse LLMs

Add code
Feb 06, 2024
Viaarxiv icon

ConPET: Continual Parameter-Efficient Tuning for Large Language Models

Add code
Sep 26, 2023
Viaarxiv icon