Picture for Ziyi Chen

Ziyi Chen

Rectified Robust Policy Optimization for Model-Uncertain Constrained Reinforcement Learning without Strong Duality

Add code
Aug 24, 2025
Viaarxiv icon

A Topic Modeling Analysis of Stigma Dimensions, Social, and Related Behavioral Circumstances in Clinical Notes Among Patients with HIV

Add code
Jun 10, 2025
Viaarxiv icon

Natural Language Generation in Healthcare: A Review of Methods and Applications

Add code
May 07, 2025
Viaarxiv icon

EasyGenNet: An Efficient Framework for Audio-Driven Gesture Video Generation Based on Diffusion Model

Add code
Apr 11, 2025
Viaarxiv icon

Data-free Knowledge Distillation with Diffusion Models

Add code
Apr 01, 2025
Viaarxiv icon

SyncDiff: Diffusion-based Talking Head Synthesis with Bottlenecked Temporal Visual Prior for Improved Synchronization

Add code
Mar 17, 2025
Viaarxiv icon

Towards Optimal Multi-draft Speculative Decoding

Add code
Feb 26, 2025
Viaarxiv icon

Transformer-Enhanced Variational Autoencoder for Crystal Structure Prediction

Add code
Feb 13, 2025
Viaarxiv icon

CrySPAI: A new Crystal Structure Prediction Software Based on Artificial Intelligence

Add code
Jan 27, 2025
Figure 1 for CrySPAI: A new Crystal Structure Prediction Software Based on Artificial Intelligence
Figure 2 for CrySPAI: A new Crystal Structure Prediction Software Based on Artificial Intelligence
Figure 3 for CrySPAI: A new Crystal Structure Prediction Software Based on Artificial Intelligence
Figure 4 for CrySPAI: A new Crystal Structure Prediction Software Based on Artificial Intelligence
Viaarxiv icon

V2PE: Improving Multimodal Long-Context Capability of Vision-Language Models with Variable Visual Position Encoding

Add code
Dec 12, 2024
Figure 1 for V2PE: Improving Multimodal Long-Context Capability of Vision-Language Models with Variable Visual Position Encoding
Figure 2 for V2PE: Improving Multimodal Long-Context Capability of Vision-Language Models with Variable Visual Position Encoding
Figure 3 for V2PE: Improving Multimodal Long-Context Capability of Vision-Language Models with Variable Visual Position Encoding
Figure 4 for V2PE: Improving Multimodal Long-Context Capability of Vision-Language Models with Variable Visual Position Encoding
Viaarxiv icon