Picture for Wen-tau Yih

Wen-tau Yih

DRAMA: Diverse Augmentation from Large Language Models to Smaller Dense Retrievers

Add code
Feb 25, 2025
Viaarxiv icon

Data-Efficient Pretraining with Group-Level Data Influence Modeling

Add code
Feb 20, 2025
Viaarxiv icon

SelfCite: Self-Supervised Alignment for Context Attribution in Large Language Models

Add code
Feb 13, 2025
Figure 1 for SelfCite: Self-Supervised Alignment for Context Attribution in Large Language Models
Figure 2 for SelfCite: Self-Supervised Alignment for Context Attribution in Large Language Models
Figure 3 for SelfCite: Self-Supervised Alignment for Context Attribution in Large Language Models
Figure 4 for SelfCite: Self-Supervised Alignment for Context Attribution in Large Language Models
Viaarxiv icon

Improving Factuality with Explicit Working Memory

Add code
Dec 24, 2024
Figure 1 for Improving Factuality with Explicit Working Memory
Figure 2 for Improving Factuality with Explicit Working Memory
Figure 3 for Improving Factuality with Explicit Working Memory
Figure 4 for Improving Factuality with Explicit Working Memory
Viaarxiv icon

Memory Layers at Scale

Add code
Dec 12, 2024
Figure 1 for Memory Layers at Scale
Figure 2 for Memory Layers at Scale
Figure 3 for Memory Layers at Scale
Figure 4 for Memory Layers at Scale
Viaarxiv icon

OpenScholar: Synthesizing Scientific Literature with Retrieval-augmented LMs

Add code
Nov 21, 2024
Figure 1 for OpenScholar: Synthesizing Scientific Literature with Retrieval-augmented LMs
Figure 2 for OpenScholar: Synthesizing Scientific Literature with Retrieval-augmented LMs
Figure 3 for OpenScholar: Synthesizing Scientific Literature with Retrieval-augmented LMs
Figure 4 for OpenScholar: Synthesizing Scientific Literature with Retrieval-augmented LMs
Viaarxiv icon

Mixture-of-Transformers: A Sparse and Scalable Architecture for Multi-Modal Foundation Models

Add code
Nov 07, 2024
Figure 1 for Mixture-of-Transformers: A Sparse and Scalable Architecture for Multi-Modal Foundation Models
Figure 2 for Mixture-of-Transformers: A Sparse and Scalable Architecture for Multi-Modal Foundation Models
Figure 3 for Mixture-of-Transformers: A Sparse and Scalable Architecture for Multi-Modal Foundation Models
Figure 4 for Mixture-of-Transformers: A Sparse and Scalable Architecture for Multi-Modal Foundation Models
Viaarxiv icon

Altogether: Image Captioning via Re-aligning Alt-text

Add code
Oct 22, 2024
Figure 1 for Altogether: Image Captioning via Re-aligning Alt-text
Figure 2 for Altogether: Image Captioning via Re-aligning Alt-text
Figure 3 for Altogether: Image Captioning via Re-aligning Alt-text
Figure 4 for Altogether: Image Captioning via Re-aligning Alt-text
Viaarxiv icon

CRAG -- Comprehensive RAG Benchmark

Add code
Jun 07, 2024
Figure 1 for CRAG -- Comprehensive RAG Benchmark
Figure 2 for CRAG -- Comprehensive RAG Benchmark
Figure 3 for CRAG -- Comprehensive RAG Benchmark
Figure 4 for CRAG -- Comprehensive RAG Benchmark
Viaarxiv icon

Nearest Neighbor Speculative Decoding for LLM Generation and Attribution

Add code
May 29, 2024
Figure 1 for Nearest Neighbor Speculative Decoding for LLM Generation and Attribution
Figure 2 for Nearest Neighbor Speculative Decoding for LLM Generation and Attribution
Figure 3 for Nearest Neighbor Speculative Decoding for LLM Generation and Attribution
Figure 4 for Nearest Neighbor Speculative Decoding for LLM Generation and Attribution
Viaarxiv icon