Picture for Yasuhiro Fujita

Yasuhiro Fujita

Preferred Elements, Inc.

Entropy Controllable Direct Preference Optimization

Add code
Nov 12, 2024
Figure 1 for Entropy Controllable Direct Preference Optimization
Figure 2 for Entropy Controllable Direct Preference Optimization
Figure 3 for Entropy Controllable Direct Preference Optimization
Figure 4 for Entropy Controllable Direct Preference Optimization
Viaarxiv icon

PLaMo-100B: A Ground-Up Language Model Designed for Japanese Proficiency

Add code
Oct 10, 2024
Figure 1 for PLaMo-100B: A Ground-Up Language Model Designed for Japanese Proficiency
Figure 2 for PLaMo-100B: A Ground-Up Language Model Designed for Japanese Proficiency
Figure 3 for PLaMo-100B: A Ground-Up Language Model Designed for Japanese Proficiency
Figure 4 for PLaMo-100B: A Ground-Up Language Model Designed for Japanese Proficiency
Viaarxiv icon

Surface-Aligned Neural Radiance Fields for Controllable 3D Human Synthesis

Add code
Jan 05, 2022
Figure 1 for Surface-Aligned Neural Radiance Fields for Controllable 3D Human Synthesis
Figure 2 for Surface-Aligned Neural Radiance Fields for Controllable 3D Human Synthesis
Figure 3 for Surface-Aligned Neural Radiance Fields for Controllable 3D Human Synthesis
Figure 4 for Surface-Aligned Neural Radiance Fields for Controllable 3D Human Synthesis
Viaarxiv icon

Distributed Reinforcement Learning of Targeted Grasping with Active Vision for Mobile Manipulators

Add code
Jul 16, 2020
Figure 1 for Distributed Reinforcement Learning of Targeted Grasping with Active Vision for Mobile Manipulators
Figure 2 for Distributed Reinforcement Learning of Targeted Grasping with Active Vision for Mobile Manipulators
Figure 3 for Distributed Reinforcement Learning of Targeted Grasping with Active Vision for Mobile Manipulators
Figure 4 for Distributed Reinforcement Learning of Targeted Grasping with Active Vision for Mobile Manipulators
Viaarxiv icon

Learning Latent State Spaces for Planning through Reward Prediction

Add code
Dec 09, 2019
Figure 1 for Learning Latent State Spaces for Planning through Reward Prediction
Figure 2 for Learning Latent State Spaces for Planning through Reward Prediction
Figure 3 for Learning Latent State Spaces for Planning through Reward Prediction
Figure 4 for Learning Latent State Spaces for Planning through Reward Prediction
Viaarxiv icon

ChainerRL: A Deep Reinforcement Learning Library

Add code
Dec 09, 2019
Figure 1 for ChainerRL: A Deep Reinforcement Learning Library
Figure 2 for ChainerRL: A Deep Reinforcement Learning Library
Figure 3 for ChainerRL: A Deep Reinforcement Learning Library
Figure 4 for ChainerRL: A Deep Reinforcement Learning Library
Viaarxiv icon

A Differentiable Gaussian-like Distribution on Hyperbolic Space for Gradient-Based Learning

Add code
Feb 08, 2019
Figure 1 for A Differentiable Gaussian-like Distribution on Hyperbolic Space for Gradient-Based Learning
Figure 2 for A Differentiable Gaussian-like Distribution on Hyperbolic Space for Gradient-Based Learning
Figure 3 for A Differentiable Gaussian-like Distribution on Hyperbolic Space for Gradient-Based Learning
Viaarxiv icon

Model-Based Reinforcement Learning via Meta-Policy Optimization

Add code
Sep 14, 2018
Figure 1 for Model-Based Reinforcement Learning via Meta-Policy Optimization
Figure 2 for Model-Based Reinforcement Learning via Meta-Policy Optimization
Figure 3 for Model-Based Reinforcement Learning via Meta-Policy Optimization
Figure 4 for Model-Based Reinforcement Learning via Meta-Policy Optimization
Viaarxiv icon

Clipped Action Policy Gradient

Add code
Jun 22, 2018
Figure 1 for Clipped Action Policy Gradient
Figure 2 for Clipped Action Policy Gradient
Figure 3 for Clipped Action Policy Gradient
Figure 4 for Clipped Action Policy Gradient
Viaarxiv icon