Picture for Zhengkun Tian

Zhengkun Tian

MSR-86K: An Evolving, Multilingual Corpus with 86,300 Hours of Transcribed Audio for Speech Recognition Research

Add code
Jun 26, 2024
Figure 1 for MSR-86K: An Evolving, Multilingual Corpus with 86,300 Hours of Transcribed Audio for Speech Recognition Research
Figure 2 for MSR-86K: An Evolving, Multilingual Corpus with 86,300 Hours of Transcribed Audio for Speech Recognition Research
Figure 3 for MSR-86K: An Evolving, Multilingual Corpus with 86,300 Hours of Transcribed Audio for Speech Recognition Research
Figure 4 for MSR-86K: An Evolving, Multilingual Corpus with 86,300 Hours of Transcribed Audio for Speech Recognition Research
Viaarxiv icon

CPPF: A contextual and post-processing-free model for automatic speech recognition

Add code
Sep 21, 2023
Figure 1 for CPPF: A contextual and post-processing-free model for automatic speech recognition
Figure 2 for CPPF: A contextual and post-processing-free model for automatic speech recognition
Figure 3 for CPPF: A contextual and post-processing-free model for automatic speech recognition
Viaarxiv icon

TST: Time-Sparse Transducer for Automatic Speech Recognition

Add code
Jul 17, 2023
Viaarxiv icon

SceneFake: An Initial Dataset and Benchmarks for Scene Fake Audio Detection

Add code
Nov 11, 2022
Viaarxiv icon

Peak-First CTC: Reducing the Peak Latency of CTC Models by Applying Peak-First Regularization

Add code
Nov 07, 2022
Viaarxiv icon

System Fingerprints Detection for DeepFake Audio: An Initial Dataset and Investigation

Add code
Aug 21, 2022
Figure 1 for System Fingerprints Detection for DeepFake Audio: An Initial Dataset and Investigation
Figure 2 for System Fingerprints Detection for DeepFake Audio: An Initial Dataset and Investigation
Figure 3 for System Fingerprints Detection for DeepFake Audio: An Initial Dataset and Investigation
Figure 4 for System Fingerprints Detection for DeepFake Audio: An Initial Dataset and Investigation
Viaarxiv icon

Fully Automated End-to-End Fake Audio Detection

Add code
Aug 20, 2022
Figure 1 for Fully Automated End-to-End Fake Audio Detection
Figure 2 for Fully Automated End-to-End Fake Audio Detection
Figure 3 for Fully Automated End-to-End Fake Audio Detection
Figure 4 for Fully Automated End-to-End Fake Audio Detection
Viaarxiv icon

ADD 2022: the First Audio Deep Synthesis Detection Challenge

Add code
Feb 26, 2022
Figure 1 for ADD 2022: the First Audio Deep Synthesis Detection Challenge
Figure 2 for ADD 2022: the First Audio Deep Synthesis Detection Challenge
Figure 3 for ADD 2022: the First Audio Deep Synthesis Detection Challenge
Figure 4 for ADD 2022: the First Audio Deep Synthesis Detection Challenge
Viaarxiv icon

Reducing language context confusion for end-to-end code-switching automatic speech recognition

Add code
Jan 28, 2022
Figure 1 for Reducing language context confusion for end-to-end code-switching automatic speech recognition
Figure 2 for Reducing language context confusion for end-to-end code-switching automatic speech recognition
Figure 3 for Reducing language context confusion for end-to-end code-switching automatic speech recognition
Figure 4 for Reducing language context confusion for end-to-end code-switching automatic speech recognition
Viaarxiv icon

Continual Learning for Fake Audio Detection

Add code
Apr 15, 2021
Figure 1 for Continual Learning for Fake Audio Detection
Figure 2 for Continual Learning for Fake Audio Detection
Figure 3 for Continual Learning for Fake Audio Detection
Figure 4 for Continual Learning for Fake Audio Detection
Viaarxiv icon