Picture for Mingjie Lin

Mingjie Lin

Cross-Embodiment Robotic Manipulation Synthesis via Guided Demonstrations through CycleVAE and Human Behavior Transformer

Add code
Mar 11, 2025
Viaarxiv icon

APEX: Ambidextrous Dual-Arm Robotic Manipulation Using Collision-Free Generative Diffusion Models

Add code
Apr 02, 2024
Figure 1 for APEX: Ambidextrous Dual-Arm Robotic Manipulation Using Collision-Free Generative Diffusion Models
Figure 2 for APEX: Ambidextrous Dual-Arm Robotic Manipulation Using Collision-Free Generative Diffusion Models
Figure 3 for APEX: Ambidextrous Dual-Arm Robotic Manipulation Using Collision-Free Generative Diffusion Models
Figure 4 for APEX: Ambidextrous Dual-Arm Robotic Manipulation Using Collision-Free Generative Diffusion Models
Viaarxiv icon

FPGA-QHAR: Throughput-Optimized for Quantized Human Action Recognition on The Edge

Add code
Nov 04, 2023
Viaarxiv icon

RETRO: Reactive Trajectory Optimization for Real-Time Robot Motion Planning in Dynamic Environments

Add code
Oct 03, 2023
Viaarxiv icon

Intercepting A Flying Target While Avoiding Moving Obstacles: A Unified Control Framework With Deep Manifold Learning

Add code
Sep 27, 2022
Figure 1 for Intercepting A Flying Target While Avoiding Moving Obstacles: A Unified Control Framework With Deep Manifold Learning
Figure 2 for Intercepting A Flying Target While Avoiding Moving Obstacles: A Unified Control Framework With Deep Manifold Learning
Figure 3 for Intercepting A Flying Target While Avoiding Moving Obstacles: A Unified Control Framework With Deep Manifold Learning
Figure 4 for Intercepting A Flying Target While Avoiding Moving Obstacles: A Unified Control Framework With Deep Manifold Learning
Viaarxiv icon

Non-Parametric Stochastic Policy Gradient with Strategic Retreat for Non-Stationary Environment

Add code
Mar 24, 2022
Figure 1 for Non-Parametric Stochastic Policy Gradient with Strategic Retreat for Non-Stationary Environment
Figure 2 for Non-Parametric Stochastic Policy Gradient with Strategic Retreat for Non-Stationary Environment
Figure 3 for Non-Parametric Stochastic Policy Gradient with Strategic Retreat for Non-Stationary Environment
Figure 4 for Non-Parametric Stochastic Policy Gradient with Strategic Retreat for Non-Stationary Environment
Viaarxiv icon

Reactive Whole-Body Obstacle Avoidance for Collision-Free Human-Robot Interaction with Topological Manifold Learning

Add code
Mar 24, 2022
Figure 1 for Reactive Whole-Body Obstacle Avoidance for Collision-Free Human-Robot Interaction with Topological Manifold Learning
Figure 2 for Reactive Whole-Body Obstacle Avoidance for Collision-Free Human-Robot Interaction with Topological Manifold Learning
Figure 3 for Reactive Whole-Body Obstacle Avoidance for Collision-Free Human-Robot Interaction with Topological Manifold Learning
Figure 4 for Reactive Whole-Body Obstacle Avoidance for Collision-Free Human-Robot Interaction with Topological Manifold Learning
Viaarxiv icon

Dynamically Avoiding Amorphous Obstacles with Topological Manifold Learning and Deep Autoencoding

Add code
Mar 24, 2022
Figure 1 for Dynamically Avoiding Amorphous Obstacles with Topological Manifold Learning and Deep Autoencoding
Figure 2 for Dynamically Avoiding Amorphous Obstacles with Topological Manifold Learning and Deep Autoencoding
Figure 3 for Dynamically Avoiding Amorphous Obstacles with Topological Manifold Learning and Deep Autoencoding
Figure 4 for Dynamically Avoiding Amorphous Obstacles with Topological Manifold Learning and Deep Autoencoding
Viaarxiv icon

Hardware-Efficient Deconvolution-Based GAN for Edge Computing

Add code
Jan 18, 2022
Figure 1 for Hardware-Efficient Deconvolution-Based GAN for Edge Computing
Figure 2 for Hardware-Efficient Deconvolution-Based GAN for Edge Computing
Figure 3 for Hardware-Efficient Deconvolution-Based GAN for Edge Computing
Figure 4 for Hardware-Efficient Deconvolution-Based GAN for Edge Computing
Viaarxiv icon

Survivable Robotic Control through Guided Bayesian Policy Search with Deep Reinforcement Learning

Add code
Jun 29, 2021
Figure 1 for Survivable Robotic Control through Guided Bayesian Policy Search with Deep Reinforcement Learning
Figure 2 for Survivable Robotic Control through Guided Bayesian Policy Search with Deep Reinforcement Learning
Figure 3 for Survivable Robotic Control through Guided Bayesian Policy Search with Deep Reinforcement Learning
Figure 4 for Survivable Robotic Control through Guided Bayesian Policy Search with Deep Reinforcement Learning
Viaarxiv icon