Alert button
Picture for Takashi Onishi

Takashi Onishi

Alert button

Which Experiences Are Influential for Your Agent? Policy Iteration with Turn-over Dropout

Add code
Bookmark button
Alert button
Jan 26, 2023
Takuya Hiraoka, Takashi Onishi, Yoshimasa Tsuruoka

Figure 1 for Which Experiences Are Influential for Your Agent? Policy Iteration with Turn-over Dropout
Figure 2 for Which Experiences Are Influential for Your Agent? Policy Iteration with Turn-over Dropout
Figure 3 for Which Experiences Are Influential for Your Agent? Policy Iteration with Turn-over Dropout
Figure 4 for Which Experiences Are Influential for Your Agent? Policy Iteration with Turn-over Dropout
Viaarxiv icon

Soft Sensors and Process Control using AI and Dynamic Simulation

Add code
Bookmark button
Alert button
Aug 08, 2022
Shumpei Kubosawa, Takashi Onishi, Yoshimasa Tsuruoka

Figure 1 for Soft Sensors and Process Control using AI and Dynamic Simulation
Figure 2 for Soft Sensors and Process Control using AI and Dynamic Simulation
Figure 3 for Soft Sensors and Process Control using AI and Dynamic Simulation
Figure 4 for Soft Sensors and Process Control using AI and Dynamic Simulation
Viaarxiv icon

Railway Operation Rescheduling System via Dynamic Simulation and Reinforcement Learning

Add code
Bookmark button
Alert button
Jan 17, 2022
Shumpei Kubosawa, Takashi Onishi, Makoto Sakahara, Yoshimasa Tsuruoka

Figure 1 for Railway Operation Rescheduling System via Dynamic Simulation and Reinforcement Learning
Figure 2 for Railway Operation Rescheduling System via Dynamic Simulation and Reinforcement Learning
Figure 3 for Railway Operation Rescheduling System via Dynamic Simulation and Reinforcement Learning
Figure 4 for Railway Operation Rescheduling System via Dynamic Simulation and Reinforcement Learning
Viaarxiv icon

Dropout Q-Functions for Doubly Efficient Reinforcement Learning

Add code
Bookmark button
Alert button
Oct 05, 2021
Takuya Hiraoka, Takahisa Imagawa, Taisei Hashimoto, Takashi Onishi, Yoshimasa Tsuruoka

Figure 1 for Dropout Q-Functions for Doubly Efficient Reinforcement Learning
Figure 2 for Dropout Q-Functions for Doubly Efficient Reinforcement Learning
Figure 3 for Dropout Q-Functions for Doubly Efficient Reinforcement Learning
Figure 4 for Dropout Q-Functions for Doubly Efficient Reinforcement Learning
Viaarxiv icon

Meta-Model-Based Meta-Policy Optimization

Add code
Bookmark button
Alert button
Jun 05, 2020
Takuya Hiraoka, Takahisa Imagawa, Voot Tangkaratt, Takayuki Osa, Takashi Onishi, Yoshimasa Tsuruoka

Figure 1 for Meta-Model-Based Meta-Policy Optimization
Figure 2 for Meta-Model-Based Meta-Policy Optimization
Figure 3 for Meta-Model-Based Meta-Policy Optimization
Figure 4 for Meta-Model-Based Meta-Policy Optimization
Viaarxiv icon

Learning Robust Options by Conditional Value at Risk Optimization

Add code
Bookmark button
Alert button
Jun 11, 2019
Takuya Hiraoka, Takahisa Imagawa, Tatsuya Mori, Takashi Onishi, Yoshimasa Tsuruoka

Figure 1 for Learning Robust Options by Conditional Value at Risk Optimization
Figure 2 for Learning Robust Options by Conditional Value at Risk Optimization
Figure 3 for Learning Robust Options by Conditional Value at Risk Optimization
Figure 4 for Learning Robust Options by Conditional Value at Risk Optimization
Viaarxiv icon

Synthesizing Chemical Plant Operation Procedures using Knowledge, Dynamic Simulation and Deep Reinforcement Learning

Add code
Bookmark button
Alert button
Mar 06, 2019
Shumpei Kubosawa, Takashi Onishi, Yoshimasa Tsuruoka

Figure 1 for Synthesizing Chemical Plant Operation Procedures using Knowledge, Dynamic Simulation and Deep Reinforcement Learning
Figure 2 for Synthesizing Chemical Plant Operation Procedures using Knowledge, Dynamic Simulation and Deep Reinforcement Learning
Figure 3 for Synthesizing Chemical Plant Operation Procedures using Knowledge, Dynamic Simulation and Deep Reinforcement Learning
Figure 4 for Synthesizing Chemical Plant Operation Procedures using Knowledge, Dynamic Simulation and Deep Reinforcement Learning
Viaarxiv icon

Refining Manually-Designed Symbol Grounding and High-Level Planning by Policy Gradients

Add code
Bookmark button
Alert button
Sep 29, 2018
Takuya Hiraoka, Takashi Onishi, Takahisa Imagawa, Yoshimasa Tsuruoka

Figure 1 for Refining Manually-Designed Symbol Grounding and High-Level Planning by Policy Gradients
Figure 2 for Refining Manually-Designed Symbol Grounding and High-Level Planning by Policy Gradients
Figure 3 for Refining Manually-Designed Symbol Grounding and High-Level Planning by Policy Gradients
Figure 4 for Refining Manually-Designed Symbol Grounding and High-Level Planning by Policy Gradients
Viaarxiv icon

Monte Carlo Tree Search with Scalable Simulation Periods for Continuously Running Tasks

Add code
Bookmark button
Alert button
Sep 07, 2018
Seydou Ba, Takuya Hiraoka, Takashi Onishi, Toru Nakata, Yoshimasa Tsuruoka

Figure 1 for Monte Carlo Tree Search with Scalable Simulation Periods for Continuously Running Tasks
Figure 2 for Monte Carlo Tree Search with Scalable Simulation Periods for Continuously Running Tasks
Figure 3 for Monte Carlo Tree Search with Scalable Simulation Periods for Continuously Running Tasks
Figure 4 for Monte Carlo Tree Search with Scalable Simulation Periods for Continuously Running Tasks
Viaarxiv icon

Hierarchical Reinforcement Learning with Abductive Planning

Add code
Bookmark button
Alert button
Jun 28, 2018
Kazeto Yamamoto, Takashi Onishi, Yoshimasa Tsuruoka

Figure 1 for Hierarchical Reinforcement Learning with Abductive Planning
Figure 2 for Hierarchical Reinforcement Learning with Abductive Planning
Figure 3 for Hierarchical Reinforcement Learning with Abductive Planning
Figure 4 for Hierarchical Reinforcement Learning with Abductive Planning
Viaarxiv icon