Picture for Kun-Yu Lin

Kun-Yu Lin

TechCoach: Towards Technical Keypoint-Aware Descriptive Action Coaching

Add code
Nov 26, 2024
Figure 1 for TechCoach: Towards Technical Keypoint-Aware Descriptive Action Coaching
Figure 2 for TechCoach: Towards Technical Keypoint-Aware Descriptive Action Coaching
Figure 3 for TechCoach: Towards Technical Keypoint-Aware Descriptive Action Coaching
Figure 4 for TechCoach: Towards Technical Keypoint-Aware Descriptive Action Coaching
Viaarxiv icon

Towards Completeness: A Generalizable Action Proposal Generator for Zero-Shot Temporal Action Localization

Add code
Aug 25, 2024
Figure 1 for Towards Completeness: A Generalizable Action Proposal Generator for Zero-Shot Temporal Action Localization
Figure 2 for Towards Completeness: A Generalizable Action Proposal Generator for Zero-Shot Temporal Action Localization
Figure 3 for Towards Completeness: A Generalizable Action Proposal Generator for Zero-Shot Temporal Action Localization
Figure 4 for Towards Completeness: A Generalizable Action Proposal Generator for Zero-Shot Temporal Action Localization
Viaarxiv icon

ParGo: Bridging Vision-Language with Partial and Global Views

Add code
Aug 23, 2024
Viaarxiv icon

Human-Centric Transformer for Domain Adaptive Action Recognition

Add code
Jul 15, 2024
Viaarxiv icon

Mitigating the Human-Robot Domain Discrepancy in Visual Pre-training for Robotic Manipulation

Add code
Jun 20, 2024
Figure 1 for Mitigating the Human-Robot Domain Discrepancy in Visual Pre-training for Robotic Manipulation
Figure 2 for Mitigating the Human-Robot Domain Discrepancy in Visual Pre-training for Robotic Manipulation
Figure 3 for Mitigating the Human-Robot Domain Discrepancy in Visual Pre-training for Robotic Manipulation
Figure 4 for Mitigating the Human-Robot Domain Discrepancy in Visual Pre-training for Robotic Manipulation
Viaarxiv icon

Rethinking CLIP-based Video Learners in Cross-Domain Open-Vocabulary Action Recognition

Add code
Mar 03, 2024
Figure 1 for Rethinking CLIP-based Video Learners in Cross-Domain Open-Vocabulary Action Recognition
Figure 2 for Rethinking CLIP-based Video Learners in Cross-Domain Open-Vocabulary Action Recognition
Figure 3 for Rethinking CLIP-based Video Learners in Cross-Domain Open-Vocabulary Action Recognition
Figure 4 for Rethinking CLIP-based Video Learners in Cross-Domain Open-Vocabulary Action Recognition
Viaarxiv icon

ActionHub: A Large-scale Action Video Description Dataset for Zero-shot Action Recognition

Add code
Jan 22, 2024
Viaarxiv icon

AdaFocus: Towards End-to-end Weakly Supervised Learning for Long-Video Action Understanding

Add code
Nov 28, 2023
Viaarxiv icon

Diversifying Spatial-Temporal Perception for Video Domain Generalization

Add code
Oct 27, 2023
Viaarxiv icon

Event-Guided Procedure Planning from Instructional Videos with Text Supervision

Add code
Aug 17, 2023
Viaarxiv icon