Picture for Yanhong Li

Yanhong Li

What Happened in LLMs Layers when Trained for Fast vs. Slow Thinking: A Gradient Perspective

Add code
Oct 31, 2024
Figure 1 for What Happened in LLMs Layers when Trained for Fast vs. Slow Thinking: A Gradient Perspective
Figure 2 for What Happened in LLMs Layers when Trained for Fast vs. Slow Thinking: A Gradient Perspective
Figure 3 for What Happened in LLMs Layers when Trained for Fast vs. Slow Thinking: A Gradient Perspective
Figure 4 for What Happened in LLMs Layers when Trained for Fast vs. Slow Thinking: A Gradient Perspective
Viaarxiv icon

When Hindsight is Not 20/20: Testing Limits on Reflective Thinking in Large Language Models

Add code
Apr 14, 2024
Viaarxiv icon

Towards Training A Chinese Large Language Model for Anesthesiology

Add code
Mar 05, 2024
Viaarxiv icon

Learning from Polar Representation: An Extreme-Adaptive Model for Long-Term Time Series Forecasting

Add code
Dec 16, 2023
Viaarxiv icon

Self-Supervised Video Transformers for Isolated Sign Language Recognition

Add code
Sep 02, 2023
Viaarxiv icon

An Extreme-Adaptive Time Series Prediction Model Based on Probability-Enhanced LSTM Neural Networks

Add code
Nov 29, 2022
Viaarxiv icon