Picture for Xiangning Chen

Xiangning Chen

Why Does Sharpness-Aware Minimization Generalize Better Than SGD?

Add code
Oct 11, 2023
Viaarxiv icon

Red Teaming Language Model Detectors with Language Models

Add code
May 31, 2023
Viaarxiv icon

Symbol tuning improves in-context learning in language models

Add code
May 15, 2023
Viaarxiv icon

Symbolic Discovery of Optimization Algorithms

Add code
Feb 17, 2023
Viaarxiv icon

Towards Efficient and Scalable Sharpness-Aware Minimization

Add code
Mar 05, 2022
Figure 1 for Towards Efficient and Scalable Sharpness-Aware Minimization
Figure 2 for Towards Efficient and Scalable Sharpness-Aware Minimization
Figure 3 for Towards Efficient and Scalable Sharpness-Aware Minimization
Figure 4 for Towards Efficient and Scalable Sharpness-Aware Minimization
Viaarxiv icon

Can Vision Transformers Perform Convolution?

Add code
Nov 03, 2021
Figure 1 for Can Vision Transformers Perform Convolution?
Figure 2 for Can Vision Transformers Perform Convolution?
Figure 3 for Can Vision Transformers Perform Convolution?
Viaarxiv icon

RANK-NOSH: Efficient Predictor-Based Architecture Search via Non-Uniform Successive Halving

Add code
Aug 18, 2021
Figure 1 for RANK-NOSH: Efficient Predictor-Based Architecture Search via Non-Uniform Successive Halving
Figure 2 for RANK-NOSH: Efficient Predictor-Based Architecture Search via Non-Uniform Successive Halving
Figure 3 for RANK-NOSH: Efficient Predictor-Based Architecture Search via Non-Uniform Successive Halving
Figure 4 for RANK-NOSH: Efficient Predictor-Based Architecture Search via Non-Uniform Successive Halving
Viaarxiv icon

Rethinking Architecture Selection in Differentiable NAS

Add code
Aug 10, 2021
Figure 1 for Rethinking Architecture Selection in Differentiable NAS
Figure 2 for Rethinking Architecture Selection in Differentiable NAS
Figure 3 for Rethinking Architecture Selection in Differentiable NAS
Figure 4 for Rethinking Architecture Selection in Differentiable NAS
Viaarxiv icon

When Vision Transformers Outperform ResNets without Pretraining or Strong Data Augmentations

Add code
Jun 03, 2021
Figure 1 for When Vision Transformers Outperform ResNets without Pretraining or Strong Data Augmentations
Figure 2 for When Vision Transformers Outperform ResNets without Pretraining or Strong Data Augmentations
Figure 3 for When Vision Transformers Outperform ResNets without Pretraining or Strong Data Augmentations
Figure 4 for When Vision Transformers Outperform ResNets without Pretraining or Strong Data Augmentations
Viaarxiv icon

Concurrent Adversarial Learning for Large-Batch Training

Add code
Jun 01, 2021
Figure 1 for Concurrent Adversarial Learning for Large-Batch Training
Figure 2 for Concurrent Adversarial Learning for Large-Batch Training
Figure 3 for Concurrent Adversarial Learning for Large-Batch Training
Figure 4 for Concurrent Adversarial Learning for Large-Batch Training
Viaarxiv icon