Picture for Payal Bajaj

Payal Bajaj

Model-Generated Pretraining Signals Improves Zero-Shot Generalization of Text-to-Text Transformers

Add code
May 21, 2023
Figure 1 for Model-Generated Pretraining Signals Improves Zero-Shot Generalization of Text-to-Text Transformers
Figure 2 for Model-Generated Pretraining Signals Improves Zero-Shot Generalization of Text-to-Text Transformers
Figure 3 for Model-Generated Pretraining Signals Improves Zero-Shot Generalization of Text-to-Text Transformers
Figure 4 for Model-Generated Pretraining Signals Improves Zero-Shot Generalization of Text-to-Text Transformers
Viaarxiv icon

Understanding the Effectiveness of Very Large Language Models on Dialog Evaluation

Add code
Jan 27, 2023
Viaarxiv icon

Foundation Transformers

Add code
Oct 19, 2022
Figure 1 for Foundation Transformers
Figure 2 for Foundation Transformers
Figure 3 for Foundation Transformers
Figure 4 for Foundation Transformers
Viaarxiv icon

On the Representation Collapse of Sparse Mixture of Experts

Add code
Apr 20, 2022
Figure 1 for On the Representation Collapse of Sparse Mixture of Experts
Figure 2 for On the Representation Collapse of Sparse Mixture of Experts
Figure 3 for On the Representation Collapse of Sparse Mixture of Experts
Figure 4 for On the Representation Collapse of Sparse Mixture of Experts
Viaarxiv icon

METRO: Efficient Denoising Pretraining of Large Scale Autoencoding Language Models with Model Generated Signals

Add code
Apr 16, 2022
Figure 1 for METRO: Efficient Denoising Pretraining of Large Scale Autoencoding Language Models with Model Generated Signals
Figure 2 for METRO: Efficient Denoising Pretraining of Large Scale Autoencoding Language Models with Model Generated Signals
Figure 3 for METRO: Efficient Denoising Pretraining of Large Scale Autoencoding Language Models with Model Generated Signals
Figure 4 for METRO: Efficient Denoising Pretraining of Large Scale Autoencoding Language Models with Model Generated Signals
Viaarxiv icon

Pretraining Text Encoders with Adversarial Mixture of Training Signal Generators

Add code
Apr 07, 2022
Figure 1 for Pretraining Text Encoders with Adversarial Mixture of Training Signal Generators
Figure 2 for Pretraining Text Encoders with Adversarial Mixture of Training Signal Generators
Figure 3 for Pretraining Text Encoders with Adversarial Mixture of Training Signal Generators
Figure 4 for Pretraining Text Encoders with Adversarial Mixture of Training Signal Generators
Viaarxiv icon

XLM-E: Cross-lingual Language Model Pre-training via ELECTRA

Add code
Jun 30, 2021
Figure 1 for XLM-E: Cross-lingual Language Model Pre-training via ELECTRA
Figure 2 for XLM-E: Cross-lingual Language Model Pre-training via ELECTRA
Figure 3 for XLM-E: Cross-lingual Language Model Pre-training via ELECTRA
Figure 4 for XLM-E: Cross-lingual Language Model Pre-training via ELECTRA
Viaarxiv icon

Language Scaling for Universal Suggested Replies Model

Add code
Jun 04, 2021
Figure 1 for Language Scaling for Universal Suggested Replies Model
Figure 2 for Language Scaling for Universal Suggested Replies Model
Figure 3 for Language Scaling for Universal Suggested Replies Model
Figure 4 for Language Scaling for Universal Suggested Replies Model
Viaarxiv icon

COCO-LM: Correcting and Contrasting Text Sequences for Language Model Pretraining

Add code
Feb 16, 2021
Figure 1 for COCO-LM: Correcting and Contrasting Text Sequences for Language Model Pretraining
Figure 2 for COCO-LM: Correcting and Contrasting Text Sequences for Language Model Pretraining
Figure 3 for COCO-LM: Correcting and Contrasting Text Sequences for Language Model Pretraining
Figure 4 for COCO-LM: Correcting and Contrasting Text Sequences for Language Model Pretraining
Viaarxiv icon

MS MARCO: A Human Generated MAchine Reading COmprehension Dataset

Add code
Oct 31, 2018
Figure 1 for MS MARCO: A Human Generated MAchine Reading COmprehension Dataset
Figure 2 for MS MARCO: A Human Generated MAchine Reading COmprehension Dataset
Figure 3 for MS MARCO: A Human Generated MAchine Reading COmprehension Dataset
Figure 4 for MS MARCO: A Human Generated MAchine Reading COmprehension Dataset
Viaarxiv icon