Picture for Yutian Wang

Yutian Wang

MeloTrans: A Text to Symbolic Music Generation Model Following Human Composition Habit

Add code
Oct 17, 2024
Figure 1 for MeloTrans: A Text to Symbolic Music Generation Model Following Human Composition Habit
Figure 2 for MeloTrans: A Text to Symbolic Music Generation Model Following Human Composition Habit
Figure 3 for MeloTrans: A Text to Symbolic Music Generation Model Following Human Composition Habit
Figure 4 for MeloTrans: A Text to Symbolic Music Generation Model Following Human Composition Habit
Viaarxiv icon

Mimicking the Mavens: Agent-based Opinion Synthesis and Emotion Prediction for Social Media Influencers

Add code
Jul 30, 2024
Viaarxiv icon

Clifford Algebra-Based Iterated Extended Kalman Filter with Application to Low-Cost INS/GNSS Navigation

Add code
Nov 15, 2023
Figure 1 for Clifford Algebra-Based Iterated Extended Kalman Filter with Application to Low-Cost INS/GNSS Navigation
Figure 2 for Clifford Algebra-Based Iterated Extended Kalman Filter with Application to Low-Cost INS/GNSS Navigation
Figure 3 for Clifford Algebra-Based Iterated Extended Kalman Filter with Application to Low-Cost INS/GNSS Navigation
Figure 4 for Clifford Algebra-Based Iterated Extended Kalman Filter with Application to Low-Cost INS/GNSS Navigation
Viaarxiv icon

An Efficient Temporary Deepfake Location Approach Based Embeddings for Partially Spoofed Audio Detection

Add code
Sep 06, 2023
Figure 1 for An Efficient Temporary Deepfake Location Approach Based Embeddings for Partially Spoofed Audio Detection
Figure 2 for An Efficient Temporary Deepfake Location Approach Based Embeddings for Partially Spoofed Audio Detection
Figure 3 for An Efficient Temporary Deepfake Location Approach Based Embeddings for Partially Spoofed Audio Detection
Figure 4 for An Efficient Temporary Deepfake Location Approach Based Embeddings for Partially Spoofed Audio Detection
Viaarxiv icon

Unsupervised Quantized Prosody Representation for Controllable Speech Synthesis

Add code
Apr 07, 2022
Figure 1 for Unsupervised Quantized Prosody Representation for Controllable Speech Synthesis
Figure 2 for Unsupervised Quantized Prosody Representation for Controllable Speech Synthesis
Figure 3 for Unsupervised Quantized Prosody Representation for Controllable Speech Synthesis
Figure 4 for Unsupervised Quantized Prosody Representation for Controllable Speech Synthesis
Viaarxiv icon

Optimizing Shoulder to Shoulder: A Coordinated Sub-Band Fusion Model for Real-Time Full-Band Speech Enhancement

Add code
Mar 30, 2022
Figure 1 for Optimizing Shoulder to Shoulder: A Coordinated Sub-Band Fusion Model for Real-Time Full-Band Speech Enhancement
Figure 2 for Optimizing Shoulder to Shoulder: A Coordinated Sub-Band Fusion Model for Real-Time Full-Band Speech Enhancement
Figure 3 for Optimizing Shoulder to Shoulder: A Coordinated Sub-Band Fusion Model for Real-Time Full-Band Speech Enhancement
Figure 4 for Optimizing Shoulder to Shoulder: A Coordinated Sub-Band Fusion Model for Real-Time Full-Band Speech Enhancement
Viaarxiv icon

DBT-Net: Dual-branch federative magnitude and phase estimation with attention-in-attention transformer for monaural speech enhancement

Add code
Feb 16, 2022
Figure 1 for DBT-Net: Dual-branch federative magnitude and phase estimation with attention-in-attention transformer for monaural speech enhancement
Figure 2 for DBT-Net: Dual-branch federative magnitude and phase estimation with attention-in-attention transformer for monaural speech enhancement
Figure 3 for DBT-Net: Dual-branch federative magnitude and phase estimation with attention-in-attention transformer for monaural speech enhancement
Figure 4 for DBT-Net: Dual-branch federative magnitude and phase estimation with attention-in-attention transformer for monaural speech enhancement
Viaarxiv icon

Dual-branch Attention-In-Attention Transformer for single-channel speech enhancement

Add code
Nov 05, 2021
Figure 1 for Dual-branch Attention-In-Attention Transformer for single-channel speech enhancement
Figure 2 for Dual-branch Attention-In-Attention Transformer for single-channel speech enhancement
Figure 3 for Dual-branch Attention-In-Attention Transformer for single-channel speech enhancement
Figure 4 for Dual-branch Attention-In-Attention Transformer for single-channel speech enhancement
Viaarxiv icon

Joint magnitude estimation and phase recovery using Cycle-in-Cycle GAN for non-parallel speech enhancement

Add code
Oct 13, 2021
Figure 1 for Joint magnitude estimation and phase recovery using Cycle-in-Cycle GAN for non-parallel speech enhancement
Figure 2 for Joint magnitude estimation and phase recovery using Cycle-in-Cycle GAN for non-parallel speech enhancement
Figure 3 for Joint magnitude estimation and phase recovery using Cycle-in-Cycle GAN for non-parallel speech enhancement
Figure 4 for Joint magnitude estimation and phase recovery using Cycle-in-Cycle GAN for non-parallel speech enhancement
Viaarxiv icon

A Two-stage Complex Network using Cycle-consistent Generative Adversarial Networks for Speech Enhancement

Add code
Sep 05, 2021
Figure 1 for A Two-stage Complex Network using Cycle-consistent Generative Adversarial Networks for Speech Enhancement
Figure 2 for A Two-stage Complex Network using Cycle-consistent Generative Adversarial Networks for Speech Enhancement
Figure 3 for A Two-stage Complex Network using Cycle-consistent Generative Adversarial Networks for Speech Enhancement
Figure 4 for A Two-stage Complex Network using Cycle-consistent Generative Adversarial Networks for Speech Enhancement
Viaarxiv icon