Picture for Zhisong Zhang

Zhisong Zhang

Structured Extraction of Process Structure Properties Relationships in Materials Science

Add code
Apr 04, 2025
Viaarxiv icon

A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression

Add code
Dec 23, 2024
Figure 1 for A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression
Figure 2 for A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression
Figure 3 for A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression
Figure 4 for A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression
Viaarxiv icon

Attention Entropy is a Key Factor: An Analysis of Parallel Context Encoding with Full-attention-based Pre-trained Language Models

Add code
Dec 21, 2024
Figure 1 for Attention Entropy is a Key Factor: An Analysis of Parallel Context Encoding with Full-attention-based Pre-trained Language Models
Figure 2 for Attention Entropy is a Key Factor: An Analysis of Parallel Context Encoding with Full-attention-based Pre-trained Language Models
Figure 3 for Attention Entropy is a Key Factor: An Analysis of Parallel Context Encoding with Full-attention-based Pre-trained Language Models
Figure 4 for Attention Entropy is a Key Factor: An Analysis of Parallel Context Encoding with Full-attention-based Pre-trained Language Models
Viaarxiv icon

Low-Bit Quantization Favors Undertrained LLMs: Scaling Laws for Quantized LLMs with 100T Training Tokens

Add code
Nov 26, 2024
Figure 1 for Low-Bit Quantization Favors Undertrained LLMs: Scaling Laws for Quantized LLMs with 100T Training Tokens
Figure 2 for Low-Bit Quantization Favors Undertrained LLMs: Scaling Laws for Quantized LLMs with 100T Training Tokens
Figure 3 for Low-Bit Quantization Favors Undertrained LLMs: Scaling Laws for Quantized LLMs with 100T Training Tokens
Figure 4 for Low-Bit Quantization Favors Undertrained LLMs: Scaling Laws for Quantized LLMs with 100T Training Tokens
Viaarxiv icon

LoGU: Long-form Generation with Uncertainty Expressions

Add code
Oct 18, 2024
Figure 1 for LoGU: Long-form Generation with Uncertainty Expressions
Figure 2 for LoGU: Long-form Generation with Uncertainty Expressions
Figure 3 for LoGU: Long-form Generation with Uncertainty Expressions
Figure 4 for LoGU: Long-form Generation with Uncertainty Expressions
Viaarxiv icon

Atomic Calibration of LLMs in Long-Form Generations

Add code
Oct 17, 2024
Viaarxiv icon

On the Transformations across Reward Model, Parameter Update, and In-Context Prompt

Add code
Jun 24, 2024
Viaarxiv icon

On the Worst Prompt Performance of Large Language Models

Add code
Jun 08, 2024
Figure 1 for On the Worst Prompt Performance of Large Language Models
Figure 2 for On the Worst Prompt Performance of Large Language Models
Figure 3 for On the Worst Prompt Performance of Large Language Models
Figure 4 for On the Worst Prompt Performance of Large Language Models
Viaarxiv icon

Self-playing Adversarial Language Game Enhances LLM Reasoning

Add code
Apr 16, 2024
Viaarxiv icon

A Thorough Examination of Decoding Methods in the Era of LLMs

Add code
Feb 10, 2024
Figure 1 for A Thorough Examination of Decoding Methods in the Era of LLMs
Figure 2 for A Thorough Examination of Decoding Methods in the Era of LLMs
Figure 3 for A Thorough Examination of Decoding Methods in the Era of LLMs
Figure 4 for A Thorough Examination of Decoding Methods in the Era of LLMs
Viaarxiv icon