default search action
Takayuki Osa
Person information
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
2020 – today
- 2024
- [c25]Motoki Omura, Takayuki Osa, Yusuke Mukuta, Tatsuya Harada:
Symmetric Q-learning: Reducing Skewness of Bellman Error in Online Reinforcement Learning. AAAI 2024: 14474-14481 - [c24]Takayuki Osa, Tatsuya Harada:
Discovering Multiple Solutions from a Single Task in Offline Reinforcement Learning. ICML 2024 - [c23]Abby O'Neill, Abdul Rehman, Abhiram Maddukuri, Abhishek Gupta, Abhishek Padalkar, Abraham Lee, Acorn Pooley, Agrim Gupta, Ajay Mandlekar, Ajinkya Jain, Albert Tung, Alex Bewley, Alexander Herzog, Alex Irpan, Alexander Khazatsky, Anant Rai, Anchit Gupta, Andrew Wang, Anikait Singh, Animesh Garg, Aniruddha Kembhavi, Annie Xie, Anthony Brohan, Antonin Raffin, Archit Sharma, Arefeh Yavary, Arhan Jain, Ashwin Balakrishna, Ayzaan Wahid, Ben Burgess-Limerick, Beomjoon Kim, Bernhard Schölkopf, Blake Wulfe, Brian Ichter, Cewu Lu, Charles Xu, Charlotte Le, Chelsea Finn, Chen Wang, Chenfeng Xu, Cheng Chi, Chenguang Huang, Christine Chan, Christopher Agia, Chuer Pan, Chuyuan Fu, Coline Devin, Danfei Xu, Daniel Morton, Danny Driess, Daphne Chen, Deepak Pathak, Dhruv Shah, Dieter Büchler, Dinesh Jayaraman, Dmitry Kalashnikov, Dorsa Sadigh, Edward Johns, Ethan Paul Foster, Fangchen Liu, Federico Ceola, Fei Xia, Feiyu Zhao, Freek Stulp, Gaoyue Zhou, Gaurav S. Sukhatme, Gautam Salhotra, Ge Yan, Gilbert Feng, Giulio Schiavi, Glen Berseth, Gregory Kahn, Guanzhi Wang, Hao Su, Haoshu Fang, Haochen Shi, Henghui Bao, Heni Ben Amor, Henrik I. Christensen, Hiroki Furuta, Homer Walke, Hongjie Fang, Huy Ha, Igor Mordatch, Ilija Radosavovic, Isabel Leal, Jacky Liang, Jad Abou-Chakra, Jaehyung Kim, Jaimyn Drake, Jan Peters, Jan Schneider, Jasmine Hsu, Jeannette Bohg, Jeffrey Bingham, Jeffrey Wu, Jensen Gao, Jiaheng Hu, Jiajun Wu, Jialin Wu, Jiankai Sun, Jianlan Luo, Jiayuan Gu, Jie Tan, Jihoon Oh, Jimmy Wu, Jingpei Lu, Jingyun Yang, Jitendra Malik, João Silvério, Joey Hejna, Jonathan Booher, Jonathan Tompson, Jonathan Yang, Jordi Salvador, Joseph J. Lim, Junhyek Han, Kaiyuan Wang, Kanishka Rao, Karl Pertsch, Karol Hausman, Keegan Go, Keerthana Gopalakrishnan, Ken Goldberg, Kendra Byrne, Kenneth Oslund, Kento Kawaharazuka, Kevin Black, Kevin Lin, Kevin Zhang, Kiana Ehsani, Kiran Lekkala, Kirsty Ellis, Krishan Rana, Krishnan Srinivasan, Kuan Fang, Kunal Pratap Singh, Kuo-Hao Zeng, Kyle Hatch, Kyle Hsu, Laurent Itti, Lawrence Yunliang Chen, Lerrel Pinto, Li Fei-Fei, Liam Tan, Linxi Jim Fan, Lionel Ott, Lisa Lee, Luca Weihs, Magnum Chen, Marion Lepert, Marius Memmel, Masayoshi Tomizuka, Masha Itkina, Mateo Guaman Castro, Max Spero, Maximilian Du, Michael Ahn, Michael C. Yip, Mingtong Zhang, Mingyu Ding, Minho Heo, Mohan Kumar Srirama, Mohit Sharma, Moo Jin Kim, Naoaki Kanazawa, Nicklas Hansen, Nicolas Heess, Nikhil J. Joshi, Niko Sünderhauf, Ning Liu, Norman Di Palo, Nur Muhammad (Mahi) Shafiullah, Oier Mees, Oliver Kroemer, Osbert Bastani, Pannag R. Sanketi, Patrick Tree Miller, Patrick Yin, Paul Wohlhart, Peng Xu, Peter David Fagan, Peter Mitrano, Pierre Sermanet, Pieter Abbeel, Priya Sundaresan, Qiuyu Chen, Quan Vuong, Rafael Rafailov, Ran Tian, Ria Doshi, Roberto Martín-Martín, Rohan Baijal, Rosario Scalise, Rose Hendrix, Roy Lin, Runjia Qian, Ruohan Zhang, Russell Mendonca, Rutav Shah, Ryan Hoque, Ryan Julian, Samuel Bustamante, Sean Kirmani, Sergey Levine, Shan Lin, Sherry Moore, Shikhar Bahl, Shivin Dass, Shubham D. Sonawani, Shuran Song, Sichun Xu, Siddhant Haldar, Siddharth Karamcheti, Simeon Adebola, Simon Guist, Soroush Nasiriany, Stefan Schaal, Stefan Welker, Stephen Tian, Subramanian Ramamoorthy, Sudeep Dasari, Suneel Belkhale, Sungjae Park, Suraj Nair, Suvir Mirchandani, Takayuki Osa, Tanmay Gupta, Tatsuya Harada, Tatsuya Matsushima, Ted Xiao, Thomas Kollar, Tianhe Yu, Tianli Ding, Todor Davchev, Tony Z. Zhao, Travis Armstrong, Trevor Darrell, Trinity Chung, Vidhi Jain, Vincent Vanhoucke, Wei Zhan, Wenxuan Zhou, Wolfram Burgard, Xi Chen, Xiaolong Wang, Xinghao Zhu, Xinyang Geng, Xiyuan Liu, Liangwei Xu, Xuanlin Li, Yao Lu, Yecheng Jason Ma, Yejin Kim, Yevgen Chebotar, Yifan Zhou, Yifeng Zhu, Yilin Wu, Ying Xu, Yixuan Wang, Yonatan Bisk, Yoonyoung Cho, Youngwoon Lee, Yuchen Cui, Yue Cao, Yueh-Hua Wu, Yujin Tang, Yuke Zhu, Yunchu Zhang, Yunfan Jiang, Yunshuang Li, Yunzhu Li, Yusuke Iwasawa, Yutaka Matsuo, Zehan Ma, Zhuo Xu, Zichen Jeff Cui, Zichen Zhang, Zipeng Lin:
Open X-Embodiment: Robotic Learning Datasets and RT-X Models : Open X-Embodiment Collaboration. ICRA 2024: 6892-6903 - [c22]Naoki Morihira, Pranav Deo, Manoj Bhadu, Akinobu Hayashi, Tadaaki Hasegawa, Satoshi Otsubo, Takayuki Osa:
Touch-Based Manipulation with Multi-Fingered Robot using Off-policy RL and Temporal Contrastive Learning. ICRA 2024: 7501-7507 - [c21]Takayuki Osa, Tatsuya Harada:
Robustifying a Policy in Multi-Agent RL with Diverse Cooperative Behaviors and Adversarial Style Sampling for Assistive Tasks. ICRA 2024: 15158-15164 - [c20]Motoki Omura, Takayuki Osa, Yusuke Mukuta, Tatsuya Harada:
Stabilizing Extreme Q-learning by Maclaurin Expansion. RLC 2024: 1427-1440 - [c19]Johannes Ackermann, Takayuki Osa, Masashi Sugiyama:
Offline Reinforcement Learning from Datasets with Structured Non-Stationarity. RLC 2024: 2140-2161 - [i15]Takayuki Osa, Tatsuya Harada:
Robustifying a Policy in Multi-Agent RL with Diverse Cooperative Behaviors and Adversarial Style Sampling for Assistive Tasks. CoRR abs/2403.00344 (2024) - [i14]Motoki Omura, Takayuki Osa, Yusuke Mukuta, Tatsuya Harada:
Symmetric Q-learning: Reducing Skewness of Bellman Error in Online Reinforcement Learning. CoRR abs/2403.07704 (2024) - [i13]Johannes Ackermann, Takayuki Osa, Masashi Sugiyama:
Offline Reinforcement Learning from Datasets with Structured Non-Stationarity. CoRR abs/2405.14114 (2024) - [i12]Motoki Omura, Takayuki Osa, Yusuke Mukuta, Tatsuya Harada:
Stabilizing Extreme Q-learning by Maclaurin Expansion. CoRR abs/2406.04896 (2024) - [i11]Takayuki Osa, Tatsuya Harada:
Discovering Multiple Solutions from a Single Task in Offline Reinforcement Learning. CoRR abs/2406.05993 (2024) - 2023
- [j13]Takayuki Osa, Naoto Osajima, Masanori Aizawa, Tatsuya Harada:
Learning Adaptive Policies for Autonomous Excavation Under Various Soil Conditions by Adversarial Domain Sampling. IEEE Robotics Autom. Lett. 8(9): 5536-5543 (2023) - [j12]Takayuki Osa, Akinobu Hayashi, Pranav Deo, Naoki Morihira, Takahide Yoshiike:
Offline Reinforcement Learning with Mixture of Deterministic Policies. Trans. Mach. Learn. Res. 2023 (2023) - [c18]Yuhei Yoshimitsu, Takayuki Osa, Shuhei Ikemoto:
Forward/Inverse Kinematics Modeling for Tensegrity Manipulator Based on Goal-Conditioned Variational Autoencoder. IROS 2023: 6668-6673 - 2022
- [j11]Takayuki Osa, Masanori Aizawa:
Deep Reinforcement Learning With Adversarial Training for Automated Excavation Using Depth Images. IEEE Access 10: 4523-4535 (2022) - [j10]Takayuki Osa:
Motion planning by learning the solution manifold in trajectory optimization. Int. J. Robotics Res. 41(3): 281-311 (2022) - [j9]Takayuki Osa, Voot Tangkaratt, Masashi Sugiyama:
Discovering diverse solutions in deep reinforcement learning by maximizing state-action-based mutual information. Neural Networks 152: 90-104 (2022) - 2021
- [c17]Takuya Hiraoka, Takahisa Imagawa, Voot Tangkaratt, Takayuki Osa, Takashi Onishi, Yoshimasa Tsuruoka:
Meta-Model-Based Meta-Policy Optimization. ACML 2021: 129-144 - [i10]Takayuki Osa, Voot Tangkaratt, Masashi Sugiyama:
Discovering Diverse Solutions in Deep Reinforcement Learning. CoRR abs/2103.07084 (2021) - [i9]Takayuki Osa:
Motion Planning by Learning the Solution Manifold in Trajectory Optimization. CoRR abs/2107.05842 (2021) - 2020
- [j8]Takayuki Osa:
Multimodal trajectory optimization for motion planning. Int. J. Robotics Res. 39(8) (2020) - [j7]Hiroyuki Karasawa, Tomohiro Kanemaki, Kei Oomae, Rui Fukui, Masayuki Nakao, Takayuki Osa:
Hierarchical Stochastic Optimization With Application to Parameter Tuning for Electronically Controlled Transmissions. IEEE Robotics Autom. Lett. 5(2): 628-635 (2020) - [j6]Takayuki Osa, Shuehi Ikemoto:
Goal-Conditioned Variational Autoencoder Trajectory Primitives with Continuous and Discrete Latent Codes. SN Comput. Sci. 1(5): 303 (2020) - [i8]Takayuki Osa:
Multimodal Trajectory Optimization for Motion Planning. CoRR abs/2003.07054 (2020) - [i7]Takuya Hiraoka, Takahisa Imagawa, Voot Tangkaratt, Takayuki Osa, Takashi Onishi, Yoshimasa Tsuruoka:
Meta-Model-Based Meta-Policy Optimization. CoRR abs/2006.02608 (2020) - [i6]Takayuki Osa:
Learning the Solution Manifold in Optimization and Its Application in Motion Planning. CoRR abs/2007.12397 (2020)
2010 – 2019
- 2019
- [c16]Takayuki Osa, Voot Tangkaratt, Masashi Sugiyama:
Hierarchical Reinforcement Learning via Advantage-Weighted Information Maximization. ICLR (Poster) 2019 - [i5]Takayuki Osa, Voot Tangkaratt, Masashi Sugiyama:
Hierarchical Reinforcement Learning via Advantage-Weighted Information Maximization. CoRR abs/1901.01365 (2019) - [i4]Johannes Ackermann, Volker Gabler, Takayuki Osa, Masashi Sugiyama:
Reducing Overestimation Bias in Multi-Agent Domains Using Double Centralized Critics. CoRR abs/1910.01465 (2019) - [i3]Takayuki Osa, Shuhei Ikemoto:
Variational Autoencoder Trajectory Primitives with Continuous and Discrete Latent Codes. CoRR abs/1912.04063 (2019) - 2018
- [j5]Takayuki Osa, Jan Peters, Gerhard Neumann:
Hierarchical reinforcement learning of multiple grasping strategies with human instructions. Adv. Robotics 32(18): 955-968 (2018) - [j4]Takayuki Osa, Joni Pajarinen, Gerhard Neumann, J. Andrew Bagnell, Pieter Abbeel, Jan Peters:
An Algorithmic Perspective on Imitation Learning. Found. Trends Robotics 7(1-2): 1-179 (2018) - [j3]Takayuki Osa, Naohiko Sugita, Mamoru Mitsuishi:
Online Trajectory Planning and Force Control for Automation of Surgical Tasks. IEEE Trans Autom. Sci. Eng. 15(2): 675-691 (2018) - [c15]Takayuki Osa, Masashi Sugiyama:
Hierarchical Policy Search via Return-Weighted Density Estimation. AAAI 2018: 3860-3867 - [c14]Robert Pinsler, Riad Akrour, Takayuki Osa, Jan Peters, Gerhard Neumann:
Sample and Feedback Efficient Hierarchical Reinforcement Learning from Human Preferences. ICRA 2018: 596-601 - [i2]Takayuki Osa, Joni Pajarinen, Gerhard Neumann, J. Andrew Bagnell, Pieter Abbeel, Jan Peters:
An Algorithmic Perspective on Imitation Learning. CoRR abs/1811.06711 (2018) - 2017
- [j2]Takayuki Osa, Amir Masoud Ghalamzan Esfahani, Rustam Stolkin, Rudolf Lioutikov, Jan Peters, Gerhard Neumann:
Guiding Trajectory Optimization by Demonstrated Distributions. IEEE Robotics Autom. Lett. 2(2): 819-826 (2017) - [c13]Guilherme Maeda, Marco Ewerton, Takayuki Osa, Baptiste Busch, Jan Peters:
Active Incremental Learning of Robot Movement Primitives. CoRL 2017: 37-46 - [c12]Firas Abi-Farraj, Takayuki Osa, Nicolo Pedemonte, Jan Peters, Gerhard Neumann, Paolo Robuffo Giordano:
A learning-based shared control architecture for interactive task execution. ICRA 2017: 329-335 - [i1]Takayuki Osa, Masashi Sugiyama:
Hierarchical Policy Search via Return-Weighted Density Estimation. CoRR abs/1711.10173 (2017) - 2016
- [c11]Takayuki Osa, Jan Peters, Gerhard Neumann:
Experiments with Hierarchical Reinforcement Learning of Multiple Grasping Policies. ISER 2016: 160-172 - 2014
- [c10]Takayuki Osa, Christian Farid Abawi, Naohiko Sugita, Hirotaka Chikuda, Shurei Sugita, Hideya Ito, Toru Moro, Yoshio Takatori, Sakae Tanaka, Mamoru Mitsuishi:
Autonomous penetration detection for bone cutting tool using demonstration-based learning. ICRA 2014: 290-296 - [c9]Takayuki Osa, Kanako Harada, Naohiko Sugita, Mamoru Mitsuishi:
Trajectory planning under different initial conditions for surgical task automation by learning from demonstration. ICRA 2014: 6507-6513 - [c8]Takayuki Osa, Satoshi Uchida, Naohiko Sugita, Mamoru Mitsuishi:
Hybrid control of master-slave velocity control and admittance control for safe remote surgery. IROS 2014: 1328-1334 - [c7]Takayuki Osa, Naohiko Sugita, Mamoru Mitsuishi:
Online Trajectory Planning in Dynamic Environments for Surgical Task Automation. Robotics: Science and Systems 2014 - 2013
- [c6]Takayuki Osa, Takuto Haniu, Kanako Harada, Naohiko Sugita, Mamoru Mitsuishi:
Perforation risk detector using demonstration-based learning for teleoperated robotic surgery. IROS 2013: 2572-2577 - 2010
- [c5]Christoph Staub, Alois C. Knoll, Takayuki Osa, Robert Bauernschmitt:
Autonomous High Precision Positioning of Surgical Instruments in Robot-Assisted Minimally Invasive Surgery under Visual Guidance. ICAS 2010: 64-69 - [c4]Christoph Staub, Takayuki Osa, Alois C. Knoll, Robert Bauernschmitt:
Automation of tissue piercing using circular needles and vision guidance for computer aided laparoscopic surgery. ICRA 2010: 4585-4590 - [c3]Kazushi Onda, Takayuki Osa, Naohiko Sugita, Makoto Hashizume, Mamoru Mitsuishi:
Asynchronous force and visual feedback in teleoperative laparoscopic surgical system. IROS 2010: 844-849 - [c2]Takayuki Osa, Christoph Staub, Alois C. Knoll:
Framework of automatic robot surgery system using Visual servoing. IROS 2010: 1837-1842
2000 – 2009
- 2009
- [j1]Naohiko Sugita, Taiga Nakano, Takayuki Osa, Yoshikazu Nakajima, Kazuo Fujiwara, Nobuhiro Abe, Toshifumi Ozaki, Masahiko Suzuki, Mamoru Mitsuishi:
Cutting Tool Protects for Soft Tissues in Bone-Milling Machining. Int. J. Autom. Technol. 3(2): 185-192 (2009) - 2008
- [c1]Naohiko Sugita, Takayuki Osa, Yoshikazu Nakajima, Mamoru Mitsuishi:
Deformation analysis and active compensation of surgical milling robot based on system error evaluation. ICRA 2008: 3389-3394
Coauthor Index
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2024-11-14 00:51 CET by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint