Picture for Hakan Erdogan

Hakan Erdogan

Binaural Angular Separation Network

Add code
Jan 16, 2024
Viaarxiv icon

TokenSplit: Using Discrete Speech Representations for Direct, Refined, and Transcript-Conditioned Speech Separation and Recognition

Add code
Aug 21, 2023
Figure 1 for TokenSplit: Using Discrete Speech Representations for Direct, Refined, and Transcript-Conditioned Speech Separation and Recognition
Figure 2 for TokenSplit: Using Discrete Speech Representations for Direct, Refined, and Transcript-Conditioned Speech Separation and Recognition
Figure 3 for TokenSplit: Using Discrete Speech Representations for Direct, Refined, and Transcript-Conditioned Speech Separation and Recognition
Viaarxiv icon

Guided Speech Enhancement Network

Add code
Mar 13, 2023
Viaarxiv icon

CycleGAN-Based Unpaired Speech Dereverberation

Add code
Mar 29, 2022
Figure 1 for CycleGAN-Based Unpaired Speech Dereverberation
Figure 2 for CycleGAN-Based Unpaired Speech Dereverberation
Figure 3 for CycleGAN-Based Unpaired Speech Dereverberation
Viaarxiv icon

Adapting Speech Separation to Real-World Meetings Using Mixture Invariant Training

Add code
Oct 20, 2021
Figure 1 for Adapting Speech Separation to Real-World Meetings Using Mixture Invariant Training
Figure 2 for Adapting Speech Separation to Real-World Meetings Using Mixture Invariant Training
Viaarxiv icon

DF-Conformer: Integrated architecture of Conv-TasNet and Conformer using linear complexity self-attention for speech enhancement

Add code
Jun 30, 2021
Figure 1 for DF-Conformer: Integrated architecture of Conv-TasNet and Conformer using linear complexity self-attention for speech enhancement
Figure 2 for DF-Conformer: Integrated architecture of Conv-TasNet and Conformer using linear complexity self-attention for speech enhancement
Figure 3 for DF-Conformer: Integrated architecture of Conv-TasNet and Conformer using linear complexity self-attention for speech enhancement
Figure 4 for DF-Conformer: Integrated architecture of Conv-TasNet and Conformer using linear complexity self-attention for speech enhancement
Viaarxiv icon

Sparse, Efficient, and Semantic Mixture Invariant Training: Taming In-the-Wild Unsupervised Sound Separation

Add code
Jun 01, 2021
Figure 1 for Sparse, Efficient, and Semantic Mixture Invariant Training: Taming In-the-Wild Unsupervised Sound Separation
Figure 2 for Sparse, Efficient, and Semantic Mixture Invariant Training: Taming In-the-Wild Unsupervised Sound Separation
Viaarxiv icon

End-to-End Diarization for Variable Number of Speakers with Local-Global Networks and Discriminative Speaker Embeddings

Add code
May 05, 2021
Figure 1 for End-to-End Diarization for Variable Number of Speakers with Local-Global Networks and Discriminative Speaker Embeddings
Figure 2 for End-to-End Diarization for Variable Number of Speakers with Local-Global Networks and Discriminative Speaker Embeddings
Figure 3 for End-to-End Diarization for Variable Number of Speakers with Local-Global Networks and Discriminative Speaker Embeddings
Figure 4 for End-to-End Diarization for Variable Number of Speakers with Local-Global Networks and Discriminative Speaker Embeddings
Viaarxiv icon

Unsupervised Sound Separation Using Mixtures of Mixtures

Add code
Jun 23, 2020
Figure 1 for Unsupervised Sound Separation Using Mixtures of Mixtures
Figure 2 for Unsupervised Sound Separation Using Mixtures of Mixtures
Figure 3 for Unsupervised Sound Separation Using Mixtures of Mixtures
Figure 4 for Unsupervised Sound Separation Using Mixtures of Mixtures
Viaarxiv icon

Universal Sound Separation

Add code
May 08, 2019
Figure 1 for Universal Sound Separation
Figure 2 for Universal Sound Separation
Figure 3 for Universal Sound Separation
Figure 4 for Universal Sound Separation
Viaarxiv icon