Picture for S. Umesh

S. Umesh

FusDom: Combining In-Domain and Out-of-Domain Knowledge for Continuous Self-Supervised Learning

Add code
Dec 20, 2023
Figure 1 for FusDom: Combining In-Domain and Out-of-Domain Knowledge for Continuous Self-Supervised Learning
Figure 2 for FusDom: Combining In-Domain and Out-of-Domain Knowledge for Continuous Self-Supervised Learning
Figure 3 for FusDom: Combining In-Domain and Out-of-Domain Knowledge for Continuous Self-Supervised Learning
Figure 4 for FusDom: Combining In-Domain and Out-of-Domain Knowledge for Continuous Self-Supervised Learning
Viaarxiv icon

Stable Distillation: Regularizing Continued Pre-training for Low-Resource Automatic Speech Recognition

Add code
Dec 20, 2023
Figure 1 for Stable Distillation: Regularizing Continued Pre-training for Low-Resource Automatic Speech Recognition
Figure 2 for Stable Distillation: Regularizing Continued Pre-training for Low-Resource Automatic Speech Recognition
Figure 3 for Stable Distillation: Regularizing Continued Pre-training for Low-Resource Automatic Speech Recognition
Figure 4 for Stable Distillation: Regularizing Continued Pre-training for Low-Resource Automatic Speech Recognition
Viaarxiv icon

The Tag-Team Approach: Leveraging CLS and Language Tagging for Enhancing Multilingual ASR

Add code
May 31, 2023
Viaarxiv icon

UNFUSED: UNsupervised Finetuning Using SElf supervised Distillation

Add code
Mar 10, 2023
Viaarxiv icon

Channel-Aware Pretraining of Joint Encoder-Decoder Self-Supervised Model for Telephonic-Speech ASR

Add code
Nov 03, 2022
Figure 1 for Channel-Aware Pretraining of Joint Encoder-Decoder Self-Supervised Model for Telephonic-Speech ASR
Figure 2 for Channel-Aware Pretraining of Joint Encoder-Decoder Self-Supervised Model for Telephonic-Speech ASR
Figure 3 for Channel-Aware Pretraining of Joint Encoder-Decoder Self-Supervised Model for Telephonic-Speech ASR
Figure 4 for Channel-Aware Pretraining of Joint Encoder-Decoder Self-Supervised Model for Telephonic-Speech ASR
Viaarxiv icon

MAST: Multiscale Audio Spectrogram Transformers

Add code
Nov 02, 2022
Viaarxiv icon

SLICER: Learning universal audio representations using low-resource self-supervised pre-training

Add code
Nov 02, 2022
Viaarxiv icon

data2vec-aqc: Search for the right Teaching Assistant in the Teacher-Student training setup

Add code
Nov 02, 2022
Figure 1 for data2vec-aqc: Search for the right Teaching Assistant in the Teacher-Student training setup
Figure 2 for data2vec-aqc: Search for the right Teaching Assistant in the Teacher-Student training setup
Figure 3 for data2vec-aqc: Search for the right Teaching Assistant in the Teacher-Student training setup
Viaarxiv icon

CCC-wav2vec 2.0: Clustering aided Cross Contrastive Self-supervised learning of speech representations

Add code
Oct 05, 2022
Figure 1 for CCC-wav2vec 2.0: Clustering aided Cross Contrastive Self-supervised learning of speech representations
Figure 2 for CCC-wav2vec 2.0: Clustering aided Cross Contrastive Self-supervised learning of speech representations
Figure 3 for CCC-wav2vec 2.0: Clustering aided Cross Contrastive Self-supervised learning of speech representations
Figure 4 for CCC-wav2vec 2.0: Clustering aided Cross Contrastive Self-supervised learning of speech representations
Viaarxiv icon

Investigation of Ensemble features of Self-Supervised Pretrained Models for Automatic Speech Recognition

Add code
Jun 11, 2022
Figure 1 for Investigation of Ensemble features of Self-Supervised Pretrained Models for Automatic Speech Recognition
Figure 2 for Investigation of Ensemble features of Self-Supervised Pretrained Models for Automatic Speech Recognition
Figure 3 for Investigation of Ensemble features of Self-Supervised Pretrained Models for Automatic Speech Recognition
Figure 4 for Investigation of Ensemble features of Self-Supervised Pretrained Models for Automatic Speech Recognition
Viaarxiv icon