Picture for Sam Ringer

Sam Ringer

The Capacity for Moral Self-Correction in Large Language Models

Add code
Feb 18, 2023
Figure 1 for The Capacity for Moral Self-Correction in Large Language Models
Figure 2 for The Capacity for Moral Self-Correction in Large Language Models
Figure 3 for The Capacity for Moral Self-Correction in Large Language Models
Figure 4 for The Capacity for Moral Self-Correction in Large Language Models
Viaarxiv icon

Discovering Language Model Behaviors with Model-Written Evaluations

Add code
Dec 19, 2022
Viaarxiv icon

Constitutional AI: Harmlessness from AI Feedback

Add code
Dec 15, 2022
Figure 1 for Constitutional AI: Harmlessness from AI Feedback
Figure 2 for Constitutional AI: Harmlessness from AI Feedback
Figure 3 for Constitutional AI: Harmlessness from AI Feedback
Figure 4 for Constitutional AI: Harmlessness from AI Feedback
Viaarxiv icon

Language Models (Mostly) Know What They Know

Add code
Jul 16, 2022
Figure 1 for Language Models (Mostly) Know What They Know
Figure 2 for Language Models (Mostly) Know What They Know
Figure 3 for Language Models (Mostly) Know What They Know
Figure 4 for Language Models (Mostly) Know What They Know
Viaarxiv icon

Hierarchical Quantized Autoencoders

Add code
Feb 19, 2020
Figure 1 for Hierarchical Quantized Autoencoders
Figure 2 for Hierarchical Quantized Autoencoders
Figure 3 for Hierarchical Quantized Autoencoders
Figure 4 for Hierarchical Quantized Autoencoders
Viaarxiv icon

Texture Bias Of CNNs Limits Few-Shot Classification Performance

Add code
Oct 18, 2019
Figure 1 for Texture Bias Of CNNs Limits Few-Shot Classification Performance
Figure 2 for Texture Bias Of CNNs Limits Few-Shot Classification Performance
Figure 3 for Texture Bias Of CNNs Limits Few-Shot Classification Performance
Figure 4 for Texture Bias Of CNNs Limits Few-Shot Classification Performance
Viaarxiv icon