Picture for Kwangyoun Kim

Kwangyoun Kim

Sample-Efficient Diffusion for Text-To-Speech Synthesis

Add code
Sep 01, 2024
Viaarxiv icon

DiscreteSLU: A Large Language Model with Self-Supervised Discrete Speech Units for Spoken Language Understanding

Add code
Jun 13, 2024
Figure 1 for DiscreteSLU: A Large Language Model with Self-Supervised Discrete Speech Units for Spoken Language Understanding
Figure 2 for DiscreteSLU: A Large Language Model with Self-Supervised Discrete Speech Units for Spoken Language Understanding
Figure 3 for DiscreteSLU: A Large Language Model with Self-Supervised Discrete Speech Units for Spoken Language Understanding
Figure 4 for DiscreteSLU: A Large Language Model with Self-Supervised Discrete Speech Units for Spoken Language Understanding
Viaarxiv icon

Improving ASR Contextual Biasing with Guided Attention

Add code
Jan 16, 2024
Viaarxiv icon

Generative Context-aware Fine-tuning of Self-supervised Speech Models

Add code
Dec 15, 2023
Figure 1 for Generative Context-aware Fine-tuning of Self-supervised Speech Models
Figure 2 for Generative Context-aware Fine-tuning of Self-supervised Speech Models
Figure 3 for Generative Context-aware Fine-tuning of Self-supervised Speech Models
Figure 4 for Generative Context-aware Fine-tuning of Self-supervised Speech Models
Viaarxiv icon

A Comparative Study on E-Branchformer vs Conformer in Speech Recognition, Translation, and Understanding Tasks

Add code
May 18, 2023
Viaarxiv icon

Structured Pruning of Self-Supervised Pre-trained Models for Speech Recognition and Understanding

Add code
Feb 27, 2023
Viaarxiv icon

Context-aware Fine-tuning of Self-supervised Speech Models

Add code
Dec 16, 2022
Viaarxiv icon

E-Branchformer: Branchformer with Enhanced merging for speech recognition

Add code
Sep 30, 2022
Figure 1 for E-Branchformer: Branchformer with Enhanced merging for speech recognition
Figure 2 for E-Branchformer: Branchformer with Enhanced merging for speech recognition
Figure 3 for E-Branchformer: Branchformer with Enhanced merging for speech recognition
Figure 4 for E-Branchformer: Branchformer with Enhanced merging for speech recognition
Viaarxiv icon

Wav2Seq: Pre-training Speech-to-Text Encoder-Decoder Models Using Pseudo Languages

Add code
May 02, 2022
Figure 1 for Wav2Seq: Pre-training Speech-to-Text Encoder-Decoder Models Using Pseudo Languages
Figure 2 for Wav2Seq: Pre-training Speech-to-Text Encoder-Decoder Models Using Pseudo Languages
Figure 3 for Wav2Seq: Pre-training Speech-to-Text Encoder-Decoder Models Using Pseudo Languages
Figure 4 for Wav2Seq: Pre-training Speech-to-Text Encoder-Decoder Models Using Pseudo Languages
Viaarxiv icon

SRU++: Pioneering Fast Recurrence with Attention for Speech Recognition

Add code
Oct 11, 2021
Figure 1 for SRU++: Pioneering Fast Recurrence with Attention for Speech Recognition
Figure 2 for SRU++: Pioneering Fast Recurrence with Attention for Speech Recognition
Figure 3 for SRU++: Pioneering Fast Recurrence with Attention for Speech Recognition
Figure 4 for SRU++: Pioneering Fast Recurrence with Attention for Speech Recognition
Viaarxiv icon