Picture for Angelos Katharopoulos

Angelos Katharopoulos

No Need to Talk: Asynchronous Mixture of Language Models

Add code
Oct 04, 2024
Figure 1 for No Need to Talk: Asynchronous Mixture of Language Models
Figure 2 for No Need to Talk: Asynchronous Mixture of Language Models
Figure 3 for No Need to Talk: Asynchronous Mixture of Language Models
Figure 4 for No Need to Talk: Asynchronous Mixture of Language Models
Viaarxiv icon

Specialized Language Models with Cheap Inference from Limited Domain Data

Add code
Feb 02, 2024
Viaarxiv icon

Controllable Music Production with Diffusion Models and Guidance Gradients

Add code
Nov 01, 2023
Viaarxiv icon

Self Supervision Does Not Help Natural Language Supervision at Scale

Add code
Jan 20, 2023
Viaarxiv icon

Neural Parts: Learning Expressive 3D Shape Abstractions with Invertible Neural Networks

Add code
Mar 18, 2021
Figure 1 for Neural Parts: Learning Expressive 3D Shape Abstractions with Invertible Neural Networks
Figure 2 for Neural Parts: Learning Expressive 3D Shape Abstractions with Invertible Neural Networks
Figure 3 for Neural Parts: Learning Expressive 3D Shape Abstractions with Invertible Neural Networks
Figure 4 for Neural Parts: Learning Expressive 3D Shape Abstractions with Invertible Neural Networks
Viaarxiv icon

Fast Transformers with Clustered Attention

Add code
Jul 09, 2020
Figure 1 for Fast Transformers with Clustered Attention
Figure 2 for Fast Transformers with Clustered Attention
Figure 3 for Fast Transformers with Clustered Attention
Figure 4 for Fast Transformers with Clustered Attention
Viaarxiv icon

Transformers are RNNs: Fast Autoregressive Transformers with Linear Attention

Add code
Jun 30, 2020
Figure 1 for Transformers are RNNs: Fast Autoregressive Transformers with Linear Attention
Figure 2 for Transformers are RNNs: Fast Autoregressive Transformers with Linear Attention
Figure 3 for Transformers are RNNs: Fast Autoregressive Transformers with Linear Attention
Figure 4 for Transformers are RNNs: Fast Autoregressive Transformers with Linear Attention
Viaarxiv icon

Processing Megapixel Images with Deep Attention-Sampling Models

Add code
May 03, 2019
Figure 1 for Processing Megapixel Images with Deep Attention-Sampling Models
Figure 2 for Processing Megapixel Images with Deep Attention-Sampling Models
Figure 3 for Processing Megapixel Images with Deep Attention-Sampling Models
Figure 4 for Processing Megapixel Images with Deep Attention-Sampling Models
Viaarxiv icon

Not All Samples Are Created Equal: Deep Learning with Importance Sampling

Add code
Jun 09, 2018
Figure 1 for Not All Samples Are Created Equal: Deep Learning with Importance Sampling
Figure 2 for Not All Samples Are Created Equal: Deep Learning with Importance Sampling
Figure 3 for Not All Samples Are Created Equal: Deep Learning with Importance Sampling
Figure 4 for Not All Samples Are Created Equal: Deep Learning with Importance Sampling
Viaarxiv icon

Biased Importance Sampling for Deep Neural Network Training

Add code
Sep 13, 2017
Figure 1 for Biased Importance Sampling for Deep Neural Network Training
Figure 2 for Biased Importance Sampling for Deep Neural Network Training
Figure 3 for Biased Importance Sampling for Deep Neural Network Training
Figure 4 for Biased Importance Sampling for Deep Neural Network Training
Viaarxiv icon