Picture for Da Ma

Da Ma

PACER: Blockwise Pre-verification for Speculative Decoding with Adaptive Length

Add code
Feb 01, 2026
Viaarxiv icon

PaperGuide: Making Small Language-Model Paper-Reading Agents More Efficient

Add code
Jan 19, 2026
Viaarxiv icon

ChemDFM-R: An Chemical Reasoner LLM Enhanced with Atomized Chemical Knowledge

Add code
Jul 30, 2025
Viaarxiv icon

Neuronal Activation States as Sample Embeddings for Data Selection in Task-Specific Instruction Tuning

Add code
Mar 19, 2025
Viaarxiv icon

AdaEAGLE: Optimizing Speculative Decoding via Explicit Modeling of Adaptive Draft Structures

Add code
Dec 25, 2024
Figure 1 for AdaEAGLE: Optimizing Speculative Decoding via Explicit Modeling of Adaptive Draft Structures
Figure 2 for AdaEAGLE: Optimizing Speculative Decoding via Explicit Modeling of Adaptive Draft Structures
Figure 3 for AdaEAGLE: Optimizing Speculative Decoding via Explicit Modeling of Adaptive Draft Structures
Figure 4 for AdaEAGLE: Optimizing Speculative Decoding via Explicit Modeling of Adaptive Draft Structures
Viaarxiv icon

Reducing Tool Hallucination via Reliability Alignment

Add code
Dec 05, 2024
Figure 1 for Reducing Tool Hallucination via Reliability Alignment
Figure 2 for Reducing Tool Hallucination via Reliability Alignment
Figure 3 for Reducing Tool Hallucination via Reliability Alignment
Figure 4 for Reducing Tool Hallucination via Reliability Alignment
Viaarxiv icon

Compressing KV Cache for Long-Context LLM Inference with Inter-Layer Attention Similarity

Add code
Dec 03, 2024
Figure 1 for Compressing KV Cache for Long-Context LLM Inference with Inter-Layer Attention Similarity
Figure 2 for Compressing KV Cache for Long-Context LLM Inference with Inter-Layer Attention Similarity
Figure 3 for Compressing KV Cache for Long-Context LLM Inference with Inter-Layer Attention Similarity
Figure 4 for Compressing KV Cache for Long-Context LLM Inference with Inter-Layer Attention Similarity
Viaarxiv icon

SciDFM: A Large Language Model with Mixture-of-Experts for Science

Add code
Sep 27, 2024
Figure 1 for SciDFM: A Large Language Model with Mixture-of-Experts for Science
Figure 2 for SciDFM: A Large Language Model with Mixture-of-Experts for Science
Figure 3 for SciDFM: A Large Language Model with Mixture-of-Experts for Science
Figure 4 for SciDFM: A Large Language Model with Mixture-of-Experts for Science
Viaarxiv icon

Evolving Subnetwork Training for Large Language Models

Add code
Jun 11, 2024
Figure 1 for Evolving Subnetwork Training for Large Language Models
Figure 2 for Evolving Subnetwork Training for Large Language Models
Figure 3 for Evolving Subnetwork Training for Large Language Models
Figure 4 for Evolving Subnetwork Training for Large Language Models
Viaarxiv icon

Sparsity-Accelerated Training for Large Language Models

Add code
Jun 03, 2024
Figure 1 for Sparsity-Accelerated Training for Large Language Models
Figure 2 for Sparsity-Accelerated Training for Large Language Models
Figure 3 for Sparsity-Accelerated Training for Large Language Models
Figure 4 for Sparsity-Accelerated Training for Large Language Models
Viaarxiv icon