Picture for Tao Ge

Tao Ge

Low-Bit Quantization Favors Undertrained LLMs: Scaling Laws for Quantized LLMs with 100T Training Tokens

Add code
Nov 26, 2024
Viaarxiv icon

Router-Tuning: A Simple and Effective Approach for Enabling Dynamic-Depth in Transformers

Add code
Oct 17, 2024
Viaarxiv icon

ParallelSpec: Parallel Drafter for Efficient Speculative Decoding

Add code
Oct 08, 2024
Viaarxiv icon

Refining Corpora from a Model Calibration Perspective for Chinese Spelling Correction

Add code
Jul 22, 2024
Figure 1 for Refining Corpora from a Model Calibration Perspective for Chinese Spelling Correction
Figure 2 for Refining Corpora from a Model Calibration Perspective for Chinese Spelling Correction
Figure 3 for Refining Corpora from a Model Calibration Perspective for Chinese Spelling Correction
Figure 4 for Refining Corpora from a Model Calibration Perspective for Chinese Spelling Correction
Viaarxiv icon

Enhancing Language Model Rationality with Bi-Directional Deliberation Reasoning

Add code
Jul 08, 2024
Viaarxiv icon

xRAG: Extreme Context Compression for Retrieval-augmented Generation with One Token

Add code
May 22, 2024
Figure 1 for xRAG: Extreme Context Compression for Retrieval-augmented Generation with One Token
Figure 2 for xRAG: Extreme Context Compression for Retrieval-augmented Generation with One Token
Figure 3 for xRAG: Extreme Context Compression for Retrieval-augmented Generation with One Token
Figure 4 for xRAG: Extreme Context Compression for Retrieval-augmented Generation with One Token
Viaarxiv icon

LLM as a Mastermind: A Survey of Strategic Reasoning with Large Language Models

Add code
Apr 01, 2024
Viaarxiv icon

K-Level Reasoning with Large Language Models

Add code
Feb 02, 2024
Viaarxiv icon

Unlocking Efficiency in Large Language Model Inference: A Comprehensive Survey of Speculative Decoding

Add code
Jan 15, 2024
Viaarxiv icon

ALYMPICS: Language Agents Meet Game Theory

Add code
Nov 16, 2023
Viaarxiv icon