Picture for Xiangning Chen

Xiangning Chen

Why Does Sharpness-Aware Minimization Generalize Better Than SGD?

Add code
Oct 11, 2023
Figure 1 for Why Does Sharpness-Aware Minimization Generalize Better Than SGD?
Figure 2 for Why Does Sharpness-Aware Minimization Generalize Better Than SGD?
Figure 3 for Why Does Sharpness-Aware Minimization Generalize Better Than SGD?
Figure 4 for Why Does Sharpness-Aware Minimization Generalize Better Than SGD?
Viaarxiv icon

Red Teaming Language Model Detectors with Language Models

Add code
May 31, 2023
Viaarxiv icon

Symbol tuning improves in-context learning in language models

Add code
May 15, 2023
Figure 1 for Symbol tuning improves in-context learning in language models
Figure 2 for Symbol tuning improves in-context learning in language models
Figure 3 for Symbol tuning improves in-context learning in language models
Figure 4 for Symbol tuning improves in-context learning in language models
Viaarxiv icon

Symbolic Discovery of Optimization Algorithms

Add code
Feb 17, 2023
Figure 1 for Symbolic Discovery of Optimization Algorithms
Figure 2 for Symbolic Discovery of Optimization Algorithms
Figure 3 for Symbolic Discovery of Optimization Algorithms
Figure 4 for Symbolic Discovery of Optimization Algorithms
Viaarxiv icon

Towards Efficient and Scalable Sharpness-Aware Minimization

Add code
Mar 05, 2022
Figure 1 for Towards Efficient and Scalable Sharpness-Aware Minimization
Figure 2 for Towards Efficient and Scalable Sharpness-Aware Minimization
Figure 3 for Towards Efficient and Scalable Sharpness-Aware Minimization
Figure 4 for Towards Efficient and Scalable Sharpness-Aware Minimization
Viaarxiv icon

Can Vision Transformers Perform Convolution?

Add code
Nov 03, 2021
Figure 1 for Can Vision Transformers Perform Convolution?
Figure 2 for Can Vision Transformers Perform Convolution?
Figure 3 for Can Vision Transformers Perform Convolution?
Viaarxiv icon

RANK-NOSH: Efficient Predictor-Based Architecture Search via Non-Uniform Successive Halving

Add code
Aug 18, 2021
Figure 1 for RANK-NOSH: Efficient Predictor-Based Architecture Search via Non-Uniform Successive Halving
Figure 2 for RANK-NOSH: Efficient Predictor-Based Architecture Search via Non-Uniform Successive Halving
Figure 3 for RANK-NOSH: Efficient Predictor-Based Architecture Search via Non-Uniform Successive Halving
Figure 4 for RANK-NOSH: Efficient Predictor-Based Architecture Search via Non-Uniform Successive Halving
Viaarxiv icon

Rethinking Architecture Selection in Differentiable NAS

Add code
Aug 10, 2021
Figure 1 for Rethinking Architecture Selection in Differentiable NAS
Figure 2 for Rethinking Architecture Selection in Differentiable NAS
Figure 3 for Rethinking Architecture Selection in Differentiable NAS
Figure 4 for Rethinking Architecture Selection in Differentiable NAS
Viaarxiv icon

When Vision Transformers Outperform ResNets without Pretraining or Strong Data Augmentations

Add code
Jun 03, 2021
Figure 1 for When Vision Transformers Outperform ResNets without Pretraining or Strong Data Augmentations
Figure 2 for When Vision Transformers Outperform ResNets without Pretraining or Strong Data Augmentations
Figure 3 for When Vision Transformers Outperform ResNets without Pretraining or Strong Data Augmentations
Figure 4 for When Vision Transformers Outperform ResNets without Pretraining or Strong Data Augmentations
Viaarxiv icon

Concurrent Adversarial Learning for Large-Batch Training

Add code
Jun 01, 2021
Figure 1 for Concurrent Adversarial Learning for Large-Batch Training
Figure 2 for Concurrent Adversarial Learning for Large-Batch Training
Figure 3 for Concurrent Adversarial Learning for Large-Batch Training
Figure 4 for Concurrent Adversarial Learning for Large-Batch Training
Viaarxiv icon