Picture for William Yang Wang

William Yang Wang

Scaling LLM Inference with Optimized Sample Compute Allocation

Add code
Oct 29, 2024
Viaarxiv icon

CBT-Bench: Evaluating Large Language Models on Assisting Cognitive Behavior Therapy

Add code
Oct 17, 2024
Figure 1 for CBT-Bench: Evaluating Large Language Models on Assisting Cognitive Behavior Therapy
Figure 2 for CBT-Bench: Evaluating Large Language Models on Assisting Cognitive Behavior Therapy
Figure 3 for CBT-Bench: Evaluating Large Language Models on Assisting Cognitive Behavior Therapy
Figure 4 for CBT-Bench: Evaluating Large Language Models on Assisting Cognitive Behavior Therapy
Viaarxiv icon

Speculative Knowledge Distillation: Bridging the Teacher-Student Gap Through Interleaved Sampling

Add code
Oct 15, 2024
Figure 1 for Speculative Knowledge Distillation: Bridging the Teacher-Student Gap Through Interleaved Sampling
Figure 2 for Speculative Knowledge Distillation: Bridging the Teacher-Student Gap Through Interleaved Sampling
Figure 3 for Speculative Knowledge Distillation: Bridging the Teacher-Student Gap Through Interleaved Sampling
Figure 4 for Speculative Knowledge Distillation: Bridging the Teacher-Student Gap Through Interleaved Sampling
Viaarxiv icon

COrAL: Order-Agnostic Language Modeling for Efficient Iterative Refinement

Add code
Oct 12, 2024
Viaarxiv icon

Understanding the Interplay between Parametric and Contextual Knowledge for Large Language Models

Add code
Oct 10, 2024
Viaarxiv icon

Detecting Training Data of Large Language Models via Expectation Maximization

Add code
Oct 10, 2024
Figure 1 for Detecting Training Data of Large Language Models via Expectation Maximization
Figure 2 for Detecting Training Data of Large Language Models via Expectation Maximization
Figure 3 for Detecting Training Data of Large Language Models via Expectation Maximization
Figure 4 for Detecting Training Data of Large Language Models via Expectation Maximization
Viaarxiv icon

Uncovering Factor Level Preferences to Improve Human-Model Alignment

Add code
Oct 09, 2024
Figure 1 for Uncovering Factor Level Preferences to Improve Human-Model Alignment
Figure 2 for Uncovering Factor Level Preferences to Improve Human-Model Alignment
Figure 3 for Uncovering Factor Level Preferences to Improve Human-Model Alignment
Figure 4 for Uncovering Factor Level Preferences to Improve Human-Model Alignment
Viaarxiv icon

T2V-Turbo-v2: Enhancing Video Generation Model Post-Training through Data, Reward, and Conditional Guidance Design

Add code
Oct 08, 2024
Figure 1 for T2V-Turbo-v2: Enhancing Video Generation Model Post-Training through Data, Reward, and Conditional Guidance Design
Figure 2 for T2V-Turbo-v2: Enhancing Video Generation Model Post-Training through Data, Reward, and Conditional Guidance Design
Figure 3 for T2V-Turbo-v2: Enhancing Video Generation Model Post-Training through Data, Reward, and Conditional Guidance Design
Figure 4 for T2V-Turbo-v2: Enhancing Video Generation Model Post-Training through Data, Reward, and Conditional Guidance Design
Viaarxiv icon

Gödel Agent: A Self-Referential Agent Framework for Recursive Self-Improvement

Add code
Oct 06, 2024
Viaarxiv icon

A Gradient Analysis Framework for Rewarding Good and Penalizing Bad Examples in Language Models

Add code
Aug 29, 2024
Figure 1 for A Gradient Analysis Framework for Rewarding Good and Penalizing Bad Examples in Language Models
Figure 2 for A Gradient Analysis Framework for Rewarding Good and Penalizing Bad Examples in Language Models
Figure 3 for A Gradient Analysis Framework for Rewarding Good and Penalizing Bad Examples in Language Models
Figure 4 for A Gradient Analysis Framework for Rewarding Good and Penalizing Bad Examples in Language Models
Viaarxiv icon