|
[1] B. D. Argall, S. Chernova, M. Veloso and B. Browning., "A survey of robot learning from demonstration," Robotics and Autonomous Sustems, 2009. [2] Z. M. O. J. Tianhao Zhang, D. Lee, X. Chen, K. Goldberg and P. Abbeel, "Deep imitation learning for complex manipulation," International Conference on Robotics and Automation, 2018. [3] A. Y. Ng and S. Russell, "Algorithms for inverse reinforcement," International Conference on Machine Learning, 2000. [4] P. Abbeel and A. Y. Ng, "Apprenticeship Learning via Inverse Reinforcement Learning," International Conference on Machine Learning, 2004. [5] A. Gupta, V. Kumar, C. Lynch, S. Levine and K. Hausman, "Relay Policy Learning: Solving Long-Horizon Tasks via Imitation and Reinforcement Learning," Conference on Robot Learning, 2019. [6] M. Bojarski, D. D. Testa, D. Dworakowski, B. Firner, B. Flepp, P. Goyal, L. D. Jackel, M. Monfort, U. Muller, J. Zhang, X. Zhang, J. Zhao and K. Zieba, "End to End Learning for Self-Driving Cars," arXiv:1604.07316, 2016. [7] J. Merel, Y. Tassa, D. TB, S. Srinivasan, J. Lemmon, Z. Wang, G. Wayne and N. Heess, "Learning human behaviors from motion capture by adversarial imitation," arXiv:1707.02201, 2017. [8] Y. Zhu, Z. Wang, J. Merel, A. Rusu, T. Erez, S. Cabi, S. Tunyasuvunakool, J. Kramár, R. Hadsell, N. d. Freitas and N. Heess, "Reinforcement and Imitation Learning for Diverse Visuomotor Skills," arXiv:1802.09564, 2018. [9] K. Hausman, Y. Chebotar, S. Schaal, G. Sukhatme and J. Lim, "Multi-Modal Imitation Learning from Unstructured Demonstrations using Generative Adversarial Nets," NIPS, 2017. [10] R. Rahmatizadeh, P. Abolghasemi, L. Bölöni and S. Levine, "Vision-Based Multi-Task Manipulation for Inexpensive Robots Using End-To-End Learning from Demonstration," International Conference on Robotics and Automation, pp. 3758-3765, 2018. [11] C. Lynch, M. Khansari, T. Xiao, V. Kumar, J. Tompson, S. Levine and P. Sermanet, "Learning Latent Plans from Play," Conference on Robot Learning, 2019. [12] C. Devin, A. Gupta, T. Darrell, P. Abbeel and S. Levine, "Learning Modular Neural Network Policies for Multi-Task and Multi-Robot Transfer," International Conference on Robotics and Automation, 2017. [13] C. Devin, P. Abbeel, T. Darrell and S. Levine, "Deep Object-Centric Representations for Generalizable Robot Learning," International Conference on Robotics and Automation, 2018. [14] C. Finn, T. Yu, T. Zhang, P. Abbeel and S. Levine, "One-Shot Visual Imitation Learning via Meta-Learning," Conference on Robot Learning, 2017. [15] S. James, M. Bloesch and A. J. Davison, "Task-Embedded Control Networks for Few-Shot Imitation Learning," Conference on Robot Learning, 2018. [16] A. Singh, E. Jang, A. Irpan, D. Kappler, M. Dalal, S. Levine, M. Khansari and C. Finn, "Scalable Multi-Task Imitation Learning with Autonomous Improvement," International Conference on Robotics and Automation, p. 2167–2173, 2020. [17] Y. Duan, M. Andrychowicz, B. C. Stadie, J. Ho, J. Schneider, I. Sutskever, P. Abbeel and W. Zaremba, "One-Shot Imitation Learning," NIPS, pp. 1087-1098, 2017. [18] Z. Wang, J. Merel, S. Reed, G. Wayne, N. d. Freitas and N. Heess, "Robust Imitation of Diverse Behaviors," NIPS, 2017. [19] S. Levine, C. Finn, T. Darrell and P. Abbeel, "End-to-End Training of Deep Visuomotor Policies," Journal of Machine Learning Research, 2016. [20] L. Yen-Chen, M. Bauza and P. Isola, "Experience-Embedded Visual Foresight," Conference on Robot Learning, 2019. [21] O. Vinyals, C. Blundell, T. Lillicrap, K. Kavukcuoglu and D. Wierstra, "Matching networks for one shot," Advances in neural information processing systems, p. 3630–3638, 2016. [22] W. Liu, D. Anguelov, D. Erhan, C. Szegedy, S. Reed, C.-Y. Fu and A. C. Berg, "SSD: Single Shot MultiBox Detector," Proceedings of the European Conference on Computer Vision, pp. 21-37, 2016.
|