Picture for Yulan Hu

Yulan Hu

SPPD: Self-training with Process Preference Learning Using Dynamic Value Margin

Add code
Feb 19, 2025
Viaarxiv icon

Coarse-to-Fine Process Reward Modeling for Enhanced Mathematical Reasoning

Add code
Jan 23, 2025
Viaarxiv icon

Video-Text Dataset Construction from Multi-AI Feedback: Promoting Weak-to-Strong Preference Learning for Video Large Language Models

Add code
Nov 25, 2024
Viaarxiv icon

GUNDAM: Aligning Large Language Models with Graph Understanding

Add code
Sep 30, 2024
Viaarxiv icon

TSO: Self-Training with Scaled Preference Optimization

Add code
Aug 31, 2024
Figure 1 for TSO: Self-Training with Scaled Preference Optimization
Figure 2 for TSO: Self-Training with Scaled Preference Optimization
Figure 3 for TSO: Self-Training with Scaled Preference Optimization
Figure 4 for TSO: Self-Training with Scaled Preference Optimization
Viaarxiv icon

Preserving Node Distinctness in Graph Autoencoders via Similarity Distillation

Add code
Jun 25, 2024
Viaarxiv icon

Towards Comprehensive Preference Data Collection for Reward Modeling

Add code
Jun 24, 2024
Viaarxiv icon

Exploring Task Unification in Graph Representation Learning via Generative Approach

Add code
Mar 21, 2024
Viaarxiv icon

VIGraph: Self-supervised Learning for Class-Imbalanced Node Classification

Add code
Nov 02, 2023
Viaarxiv icon

Do We Really Need Contrastive Learning for Graph Representation?

Add code
Oct 23, 2023
Viaarxiv icon