LIVIVO - The Search Portal for Life Sciences

zur deutschen Oberfläche wechseln
Advanced search

Search results

Result 1 - 9 of total 9

Search options

  1. Book ; Online: Learning Time-Invariant Reward Functions through Model-Based Inverse Reinforcement Learning

    Davchev, Todor / Bechtle, Sarah / Ramamoorthy, Subramanian / Meier, Franziska

    2021  

    Abstract: Inverse reinforcement learning is a paradigm motivated by the goal of learning general reward functions from demonstrated behaviours. Yet the notion of generality for learnt costs is often evaluated in terms of robustness to various spatial perturbations ...

    Abstract Inverse reinforcement learning is a paradigm motivated by the goal of learning general reward functions from demonstrated behaviours. Yet the notion of generality for learnt costs is often evaluated in terms of robustness to various spatial perturbations only, assuming deployment at fixed speeds of execution. However, this is impractical in the context of robotics and building, time-invariant solutions is of crucial importance. In this work, we propose a formulation that allows us to 1) vary the length of execution by learning time-invariant costs, and 2) relax the temporal alignment requirements for learning from demonstration. We apply our method to two different types of cost formulations and evaluate their performance in the context of learning reward functions for simulated placement and peg in hole tasks executed on a 7DoF Kuka IIWA arm. Our results show that our approach enables learning temporally invariant rewards from misaligned demonstration that can also generalise spatially to out of distribution tasks.
    Keywords Computer Science - Robotics ; Statistics - Machine Learning
    Subject code 006
    Publishing date 2021-07-07
    Publishing country us
    Document type Book ; Online
    Database BASE - Bielefeld Academic Search Engine (life sciences selection)

    More links

    Kategorien

  2. Book ; Online: RoboTAP

    Vecerik, Mel / Doersch, Carl / Yang, Yi / Davchev, Todor / Aytar, Yusuf / Zhou, Guangyao / Hadsell, Raia / Agapito, Lourdes / Scholz, Jon

    Tracking Arbitrary Points for Few-Shot Visual Imitation

    2023  

    Abstract: For robots to be useful outside labs and specialized factories we need a way to teach them new useful behaviors quickly. Current approaches lack either the generality to onboard new tasks without task-specific engineering, or else lack the data- ... ...

    Abstract For robots to be useful outside labs and specialized factories we need a way to teach them new useful behaviors quickly. Current approaches lack either the generality to onboard new tasks without task-specific engineering, or else lack the data-efficiency to do so in an amount of time that enables practical use. In this work we explore dense tracking as a representational vehicle to allow faster and more general learning from demonstration. Our approach utilizes Track-Any-Point (TAP) models to isolate the relevant motion in a demonstration, and parameterize a low-level controller to reproduce this motion across changes in the scene configuration. We show this results in robust robot policies that can solve complex object-arrangement tasks such as shape-matching, stacking, and even full path-following tasks such as applying glue and sticking objects together, all from demonstrations that can be collected in minutes.

    Comment: Project website: https://robotap.github.io
    Keywords Computer Science - Robotics ; Computer Science - Artificial Intelligence ; Computer Science - Computer Vision and Pattern Recognition
    Subject code 629 ; 004
    Publishing date 2023-08-30
    Publishing country us
    Document type Book ; Online
    Database BASE - Bielefeld Academic Search Engine (life sciences selection)

    More links

    Kategorien

  3. Book ; Online: Learning Structured Representations of Spatial and Interactive Dynamics for Trajectory Prediction in Crowded Scenes

    Davchev, Todor / Burke, Michael / Ramamoorthy, Subramanian

    2019  

    Abstract: Context plays a significant role in the generation of motion for dynamic agents in interactive environments. This work proposes a modular method that utilises a learned model of the environment for motion prediction. This modularity explicitly allows for ...

    Abstract Context plays a significant role in the generation of motion for dynamic agents in interactive environments. This work proposes a modular method that utilises a learned model of the environment for motion prediction. This modularity explicitly allows for unsupervised adaptation of trajectory prediction models to unseen environments and new tasks by relying on unlabelled image data only. We model both the spatial and dynamic aspects of a given environment alongside the per agent motions. This results in more informed motion prediction and allows for performance comparable to the state-of-the-art. We highlight the model's prediction capability using a benchmark pedestrian prediction problem and a robot manipulation task and show that we can transfer the predictor across these tasks in a completely unsupervised way. The proposed approach allows for robust and label efficient forward modelling, and relaxes the need for full model re-training in new environments.
    Keywords Computer Science - Machine Learning ; Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Multiagent Systems ; Computer Science - Robotics ; Statistics - Machine Learning
    Subject code 629
    Publishing date 2019-11-29
    Publishing country us
    Document type Book ; Online
    Database BASE - Bielefeld Academic Search Engine (life sciences selection)

    More links

    Kategorien

  4. Book ; Online: An Empirical Evaluation of Adversarial Robustness under Transfer Learning

    Davchev, Todor / Korres, Timos / Fotiadis, Stathi / Antonopoulos, Nick / Ramamoorthy, Subramanian

    2019  

    Abstract: In this work, we evaluate adversarial robustness in the context of transfer learning from a source trained on CIFAR 100 to a target network trained on CIFAR 10. Specifically, we study the effects of using robust optimisation in the source and target ... ...

    Abstract In this work, we evaluate adversarial robustness in the context of transfer learning from a source trained on CIFAR 100 to a target network trained on CIFAR 10. Specifically, we study the effects of using robust optimisation in the source and target networks. This allows us to identify transfer learning strategies under which adversarial defences are successfully retained, in addition to revealing potential vulnerabilities. We study the extent to which features learnt by a fast gradient sign method (FGSM) and its iterative alternative (PGD) can preserve their defence properties against black and white-box attacks under three different transfer learning strategies. We find that using PGD examples during training on the source task leads to more general robust features that are easier to transfer. Furthermore, under successful transfer, it achieves 5.2% more accuracy against white-box PGD attacks than suitable baselines. Overall, our empirical evaluations give insights on how well adversarial robustness under transfer learning can generalise.
    Keywords Statistics - Machine Learning ; Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Machine Learning
    Subject code 006
    Publishing date 2019-05-07
    Publishing country us
    Document type Book ; Online
    Database BASE - Bielefeld Academic Search Engine (life sciences selection)

    More links

    Kategorien

  5. Book ; Online: Model-Based Inverse Reinforcement Learning from Visual Demonstrations

    Das, Neha / Bechtle, Sarah / Davchev, Todor / Jayaraman, Dinesh / Rai, Akshara / Meier, Franziska

    2020  

    Abstract: Scaling model-based inverse reinforcement learning (IRL) to real robotic manipulation tasks with unknown dynamics remains an open problem. The key challenges lie in learning good dynamics models, developing algorithms that scale to high-dimensional state- ...

    Abstract Scaling model-based inverse reinforcement learning (IRL) to real robotic manipulation tasks with unknown dynamics remains an open problem. The key challenges lie in learning good dynamics models, developing algorithms that scale to high-dimensional state-spaces and being able to learn from both visual and proprioceptive demonstrations. In this work, we present a gradient-based inverse reinforcement learning framework that utilizes a pre-trained visual dynamics model to learn cost functions when given only visual human demonstrations. The learned cost functions are then used to reproduce the demonstrated behavior via visual model predictive control. We evaluate our framework on hardware on two basic object manipulation tasks.

    Comment: Accepted at the 4th Conference on Robotic Learning (CoRL 2020), Cambridge MA, USA
    Keywords Computer Science - Robotics ; Computer Science - Machine Learning
    Publishing date 2020-10-18
    Publishing country us
    Document type Book ; Online
    Database BASE - Bielefeld Academic Search Engine (life sciences selection)

    More links

    Kategorien

  6. Book ; Online: Residual Learning from Demonstration

    Davchev, Todor / Luck, Kevin Sebastian / Burke, Michael / Meier, Franziska / Schaal, Stefan / Ramamoorthy, Subramanian

    Adapting DMPs for Contact-rich Manipulation

    2020  

    Abstract: Manipulation skills involving contact and friction are inherent to many robotics tasks. Using the class of motor primitives for peg-in-hole like insertions, we study how robots can learn such skills. Dynamic Movement Primitives (DMP) are a popular way of ...

    Abstract Manipulation skills involving contact and friction are inherent to many robotics tasks. Using the class of motor primitives for peg-in-hole like insertions, we study how robots can learn such skills. Dynamic Movement Primitives (DMP) are a popular way of extracting such policies through behaviour cloning (BC), but can struggle in the context of insertion. Policy adaptation strategies such as residual learning can help improve the overall performance of policies in the context of contact-rich manipulation. However, it is not clear how to best do this with DMPs. As result, we consider a number of possible ways for adapting a DMP formulation and propose ``residual Learning from Demonstration`` (rLfD), a framework that combines DMPs with Reinforcement Learning (RL) to learn a residual correction policy. Our evaluations suggest that applying residual learning directly in task space and operating on the full pose of the robot can significantly improve the overall performance of DMPs. We show that rLfD offers a gentle to the joints solution that improves the task success and generalisation of DMPs. The proposed framework is evaluated on a set of tasks in which a simulated robot and a real physical robot arm have to successfully insert pegs, gears and plugs into their respective sockets. Further material and videos accompanying this paper are provided at https://sites.google.com/view/rlfd/.
    Keywords Computer Science - Robotics ; Computer Science - Artificial Intelligence
    Subject code 629
    Publishing date 2020-08-17
    Publishing country us
    Document type Book ; Online
    Database BASE - Bielefeld Academic Search Engine (life sciences selection)

    More links

    Kategorien

  7. Book ; Online: Vid2Param

    Asenov, Martin / Burke, Michael / Angelov, Daniel / Davchev, Todor / Subr, Kartic / Ramamoorthy, Subramanian

    Modelling of Dynamics Parameters from Video

    2019  

    Abstract: Videos provide a rich source of information, but it is generally hard to extract dynamical parameters of interest. Inferring those parameters from a video stream would be beneficial for physical reasoning. Robots performing tasks in dynamic environments ... ...

    Abstract Videos provide a rich source of information, but it is generally hard to extract dynamical parameters of interest. Inferring those parameters from a video stream would be beneficial for physical reasoning. Robots performing tasks in dynamic environments would benefit greatly from understanding the underlying environment motion, in order to make future predictions and to synthesize effective control policies that use this inductive bias. Online physical reasoning is therefore a fundamental requirement for robust autonomous agents. When the dynamics involves multiple modes (due to contacts or interactions between objects) and sensing must proceed directly from a rich sensory stream such as video, then traditional methods for system identification may not be well suited. We propose an approach wherein fast parameter estimation can be achieved directly from video. We integrate a physically based dynamics model with a recurrent variational autoencoder, by introducing an additional loss to enforce desired constraints. The model, which we call Vid2Param, can be trained entirely in simulation, in an end-to-end manner with domain randomization, to perform online system identification, and make probabilistic forward predictions of parameters of interest. This enables the resulting model to encode parameters such as position, velocity, restitution, air drag and other physical properties of the system. We illustrate the utility of this in physical experiments wherein a PR2 robot with a velocity constrained arm must intercept an unknown bouncing ball with partly occluded vision, by estimating the physical parameters of this ball directly from the video trace after the ball is released.

    Comment: Accepted as a journal paper at IEEE Robotics and Automation Letters (RA-L)
    Keywords Computer Science - Robotics
    Subject code 629
    Publishing date 2019-07-15
    Publishing country us
    Document type Book ; Online
    Database BASE - Bielefeld Academic Search Engine (life sciences selection)

    More links

    Kategorien

  8. Book ; Online: RoboCat

    Bousmalis, Konstantinos / Vezzani, Giulia / Rao, Dushyant / Devin, Coline / Lee, Alex X. / Bauza, Maria / Davchev, Todor / Zhou, Yuxiang / Gupta, Agrim / Raju, Akhil / Laurens, Antoine / Fantacci, Claudio / Dalibard, Valentin / Zambelli, Martina / Martins, Murilo / Pevceviciute, Rugile / Blokzijl, Michiel / Denil, Misha / Batchelor, Nathan /
    Lampe, Thomas / Parisotto, Emilio / Żołna, Konrad / Reed, Scott / Colmenarejo, Sergio Gómez / Scholz, Jon / Abdolmaleki, Abbas / Groth, Oliver / Regli, Jean-Baptiste / Sushkov, Oleg / Rothörl, Tom / Chen, José Enrique / Aytar, Yusuf / Barker, Dave / Ortiz, Joy / Riedmiller, Martin / Springenberg, Jost Tobias / Hadsell, Raia / Nori, Francesco / Heess, Nicolas

    A Self-Improving Foundation Agent for Robotic Manipulation

    2023  

    Abstract: The ability to leverage heterogeneous robotic experience from different robots and tasks to quickly master novel skills and embodiments has the potential to transform robot learning. Inspired by recent advances in foundation models for vision and ... ...

    Abstract The ability to leverage heterogeneous robotic experience from different robots and tasks to quickly master novel skills and embodiments has the potential to transform robot learning. Inspired by recent advances in foundation models for vision and language, we propose a foundation agent for robotic manipulation. This agent, named RoboCat, is a visual goal-conditioned decision transformer capable of consuming multi-embodiment action-labelled visual experience. This data spans a large repertoire of motor control skills from simulated and real robotic arms with varying sets of observations and actions. With RoboCat, we demonstrate the ability to generalise to new tasks and robots, both zero-shot as well as through adaptation using only 100--1000 examples for the target task. We also show how a trained model itself can be used to generate data for subsequent training iterations, thus providing a basic building block for an autonomous improvement loop. We investigate the agent's capabilities, with large-scale evaluations both in simulation and on three different real robot embodiments. We find that as we grow and diversify its training data, RoboCat not only shows signs of cross-task transfer, but also becomes more efficient at adapting to new tasks.
    Keywords Computer Science - Robotics ; Computer Science - Machine Learning
    Subject code 629
    Publishing date 2023-06-20
    Publishing country us
    Document type Book ; Online
    Database BASE - Bielefeld Academic Search Engine (life sciences selection)

    More links

    Kategorien

  9. Book ; Online: Open X-Embodiment

    Collaboration, Open X-Embodiment / Padalkar, Abhishek / Pooley, Acorn / Mandlekar, Ajay / Jain, Ajinkya / Tung, Albert / Bewley, Alex / Herzog, Alex / Irpan, Alex / Khazatsky, Alexander / Rai, Anant / Singh, Anikait / Garg, Animesh / Brohan, Anthony / Raffin, Antonin / Wahid, Ayzaan / Burgess-Limerick, Ben / Kim, Beomjoon / Schölkopf, Bernhard /
    Ichter, Brian / Lu, Cewu / Xu, Charles / Finn, Chelsea / Xu, Chenfeng / Chi, Cheng / Huang, Chenguang / Chan, Christine / Pan, Chuer / Fu, Chuyuan / Devin, Coline / Driess, Danny / Pathak, Deepak / Shah, Dhruv / Büchler, Dieter / Kalashnikov, Dmitry / Sadigh, Dorsa / Johns, Edward / Ceola, Federico / Xia, Fei / Stulp, Freek / Zhou, Gaoyue / Sukhatme, Gaurav S. / Salhotra, Gautam / Yan, Ge / Schiavi, Giulio / Kahn, Gregory / Su, Hao / Fang, Hao-Shu / Shi, Haochen / Amor, Heni Ben / Christensen, Henrik I / Furuta, Hiroki / Walke, Homer / Fang, Hongjie / Mordatch, Igor / Radosavovic, Ilija / Leal, Isabel / Liang, Jacky / Abou-Chakra, Jad / Kim, Jaehyung / Peters, Jan / Schneider, Jan / Hsu, Jasmine / Bohg, Jeannette / Bingham, Jeffrey / Wu, Jiajun / Wu, Jialin / Luo, Jianlan / Gu, Jiayuan / Tan, Jie / Oh, Jihoon / Malik, Jitendra / Booher, Jonathan / Tompson, Jonathan / Yang, Jonathan / Lim, Joseph J. / Silvério, João / Han, Junhyek / Rao, Kanishka / Pertsch, Karl / Hausman, Karol / Go, Keegan / Gopalakrishnan, Keerthana / Goldberg, Ken / Byrne, Kendra / Oslund, Kenneth / Kawaharazuka, Kento / Zhang, Kevin / Rana, Krishan / Srinivasan, Krishnan / Chen, Lawrence Yunliang / Pinto, Lerrel / Fei-Fei, Li / Tan, Liam / Ott, Lionel / Lee, Lisa / Tomizuka, Masayoshi / Spero, Max / Du, Maximilian / Ahn, Michael / Zhang, Mingtong / Ding, Mingyu / Srirama, Mohan Kumar / Sharma, Mohit / Kim, Moo Jin / Kanazawa, Naoaki / Hansen, Nicklas / Heess, Nicolas / Joshi, Nikhil J / Suenderhauf, Niko / Di Palo, Norman / Shafiullah, Nur Muhammad Mahi / Mees, Oier / Kroemer, Oliver / Sanketi, Pannag R / Wohlhart, Paul / Xu, Peng / Sermanet, Pierre / Sundaresan, Priya / Vuong, Quan / Rafailov, Rafael / Tian, Ran / Doshi, Ria / Martín-Martín, Roberto / Mendonca, Russell / Shah, Rutav / Hoque, Ryan / Julian, Ryan / Bustamante, Samuel / Kirmani, Sean / Levine, Sergey / Moore, Sherry / Bahl, Shikhar / Dass, Shivin / Sonawani, Shubham / Song, Shuran / Xu, Sichun / Haldar, Siddhant / Adebola, Simeon / Guist, Simon / Nasiriany, Soroush / Schaal, Stefan / Welker, Stefan / Tian, Stephen / Dasari, Sudeep / Belkhale, Suneel / Osa, Takayuki / Harada, Tatsuya / Matsushima, Tatsuya / Xiao, Ted / Yu, Tianhe / Ding, Tianli / Davchev, Todor / Zhao, Tony Z. / Armstrong, Travis / Darrell, Trevor / Jain, Vidhi / Vanhoucke, Vincent / Zhan, Wei / Zhou, Wenxuan / Burgard, Wolfram / Chen, Xi / Wang, Xiaolong / Zhu, Xinghao / Li, Xuanlin / Lu, Yao / Chebotar, Yevgen / Zhou, Yifan / Zhu, Yifeng / Xu, Ying / Wang, Yixuan / Bisk, Yonatan / Cho, Yoonyoung / Lee, Youngwoon / Cui, Yuchen / Wu, Yueh-Hua / Tang, Yujin / Zhu, Yuke / Li, Yunzhu / Iwasawa, Yusuke / Matsuo, Yutaka / Xu, Zhuo / Cui, Zichen Jeff

    Robotic Learning Datasets and RT-X Models

    2023  

    Abstract: Large, high-capacity models trained on diverse datasets have shown remarkable successes on efficiently tackling downstream applications. In domains from NLP to Computer Vision, this has led to a consolidation of pretrained models, with general pretrained ...

    Abstract Large, high-capacity models trained on diverse datasets have shown remarkable successes on efficiently tackling downstream applications. In domains from NLP to Computer Vision, this has led to a consolidation of pretrained models, with general pretrained backbones serving as a starting point for many applications. Can such a consolidation happen in robotics? Conventionally, robotic learning methods train a separate model for every application, every robot, and even every environment. Can we instead train generalist X-robot policy that can be adapted efficiently to new robots, tasks, and environments? In this paper, we provide datasets in standardized data formats and models to make it possible to explore this possibility in the context of robotic manipulation, alongside experimental results that provide an example of effective X-robot policies. We assemble a dataset from 22 different robots collected through a collaboration between 21 institutions, demonstrating 527 skills (160266 tasks). We show that a high-capacity model trained on this data, which we call RT-X, exhibits positive transfer and improves the capabilities of multiple robots by leveraging experience from other platforms. More details can be found on the project website $\href{https://robotics-transformer-x.github.io}{\text{robotics-transformer-x.github.io}}$.
    Keywords Computer Science - Robotics
    Subject code 629
    Publishing date 2023-10-13
    Publishing country us
    Document type Book ; Online
    Database BASE - Bielefeld Academic Search Engine (life sciences selection)

    More links

    Kategorien

To top