LIVIVO - The Search Portal for Life Sciences

zur deutschen Oberfläche wechseln
Advanced search

Your last searches

  1. AU="Burgard, Wolfram"
  2. AU="Chrastil, Hillary"

Search results

Result 1 - 10 of total 105

Search options

  1. Book ; Online: Self-Supervised Moving Vehicle Detection from Audio-Visual Cues

    Zürn, Jannik / Burgard, Wolfram

    2022  

    Abstract: Robust detection of moving vehicles is a critical task for any autonomously operating outdoor robot or self-driving vehicle. Most modern approaches for solving this task rely on training image-based detectors using large-scale vehicle detection datasets ... ...

    Abstract Robust detection of moving vehicles is a critical task for any autonomously operating outdoor robot or self-driving vehicle. Most modern approaches for solving this task rely on training image-based detectors using large-scale vehicle detection datasets such as nuScenes or the Waymo Open Dataset. Providing manual annotations is an expensive and laborious exercise that does not scale well in practice. To tackle this problem, we propose a self-supervised approach that leverages audio-visual cues to detect moving vehicles in videos. Our approach employs contrastive learning for localizing vehicles in images from corresponding pairs of images and recorded audio. In extensive experiments carried out with a real-world dataset, we demonstrate that our approach provides accurate detections of moving vehicles and does not require manual annotations. We furthermore show that our model can be used as a teacher to supervise an audio-only detection model. This student model is invariant to illumination changes and thus effectively bridges the domain gap inherent to models leveraging exclusively vision as the predominant modality.

    Comment: 8 pages, 6 figures
    Keywords Computer Science - Computer Vision and Pattern Recognition
    Subject code 629
    Publishing date 2022-01-30
    Publishing country us
    Document type Book ; Online
    Database BASE - Bielefeld Academic Search Engine (life sciences selection)

    More links

    Kategorien

  2. Book ; Online: USegScene

    Vertens, Johan / Burgard, Wolfram

    Unsupervised Learning of Depth, Optical Flow and Ego-Motion with Semantic Guidance and Coupled Networks

    2022  

    Abstract: In this paper we propose USegScene, a framework for semantically guided unsupervised learning of depth, optical flow and ego-motion estimation for stereo camera images using convolutional neural networks. Our framework leverages semantic information for ... ...

    Abstract In this paper we propose USegScene, a framework for semantically guided unsupervised learning of depth, optical flow and ego-motion estimation for stereo camera images using convolutional neural networks. Our framework leverages semantic information for improved regularization of depth and optical flow maps, multimodal fusion and occlusion filling considering dynamic rigid object motions as independent SE(3) transformations. Furthermore, complementary to pure photo-metric matching, we propose matching of semantic features, pixel-wise classes and object instance borders between the consecutive images. In contrast to previous methods, we propose a network architecture that jointly predicts all outputs using shared encoders and allows passing information across the task-domains, e.g., the prediction of optical flow can benefit from the prediction of the depth. Furthermore, we explicitly learn the depth and optical flow occlusion maps inside the network, which are leveraged in order to improve the predictions in therespective regions. We present results on the popular KITTI dataset and show that our approach outperforms other methods by a large margin.
    Keywords Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Robotics
    Subject code 006
    Publishing date 2022-07-15
    Publishing country us
    Document type Book ; Online
    Database BASE - Bielefeld Academic Search Engine (life sciences selection)

    More links

    Kategorien

  3. Book ; Online ; Thesis: <dc:title>Techniques for localization and mapping in precision agriculture</dc:title>

    Winterhalter, Wera [Verfasser] / Burgard, Wolfram [Akademischer Betreuer] / Burgard, Wolfram / Pradalier, Cédric / Bödecker, Joschka

    2023  

    Keywords Landwirtschaft, Veterinärmedizin ; Agriculture, Veterinary Science
    Subject code sg630
    Language English
    Publisher Universität
    Publishing place Freiburg
    Document type Book ; Online ; Thesis
    Database Digital theses on the web

    More links

    Kategorien

  4. Book ; Online: AutoGraph

    Zürn, Jannik / Posner, Ingmar / Burgard, Wolfram

    Predicting Lane Graphs from Traffic Observations

    2023  

    Abstract: Lane graph estimation is a long-standing problem in the context of autonomous driving. Previous works aimed at solving this problem by relying on large-scale, hand-annotated lane graphs, introducing a data bottleneck for training models to solve this ... ...

    Abstract Lane graph estimation is a long-standing problem in the context of autonomous driving. Previous works aimed at solving this problem by relying on large-scale, hand-annotated lane graphs, introducing a data bottleneck for training models to solve this task. To overcome this limitation, we propose to use the motion patterns of traffic participants as lane graph annotations. In our AutoGraph approach, we employ a pre-trained object tracker to collect the tracklets of traffic participants such as vehicles and trucks. Based on the location of these tracklets, we predict the successor lane graph from an initial position using overhead RGB images only, not requiring any human supervision. In a subsequent stage, we show how the individual successor predictions can be aggregated into a consistent lane graph. We demonstrate the efficacy of our approach on the UrbanLaneGraph dataset and perform extensive quantitative and qualitative evaluations, indicating that AutoGraph is on par with models trained on hand-annotated graph data. Model and dataset will be made available at redacted-for-review.

    Comment: 8 pages, 6 figures
    Keywords Computer Science - Computer Vision and Pattern Recognition
    Subject code 006
    Publishing date 2023-06-27
    Publishing country us
    Document type Book ; Online
    Database BASE - Bielefeld Academic Search Engine (life sciences selection)

    More links

    Kategorien

  5. Book ; Online: FM-Loc

    Mirjalili, Reihaneh / Krawez, Michael / Burgard, Wolfram

    Using Foundation Models for Improved Vision-based Localization

    2023  

    Abstract: Visual place recognition is essential for vision-based robot localization and SLAM. Despite the tremendous progress made in recent years, place recognition in changing environments remains challenging. A promising approach to cope with appearance ... ...

    Abstract Visual place recognition is essential for vision-based robot localization and SLAM. Despite the tremendous progress made in recent years, place recognition in changing environments remains challenging. A promising approach to cope with appearance variations is to leverage high-level semantic features like objects or place categories. In this paper, we propose FM-Loc which is a novel image-based localization approach based on Foundation Models that uses the Large Language Model GPT-3 in combination with the Visual-Language Model CLIP to construct a semantic image descriptor that is robust to severe changes in scene geometry and camera viewpoint. We deploy CLIP to detect objects in an image, GPT-3 to suggest potential room labels based on the detected objects, and CLIP again to propose the most likely location label. The object labels and the scene label constitute an image descriptor that we use to calculate a similarity score between the query and database images. We validate our approach on real-world data that exhibit significant changes in camera viewpoints and object placement between the database and query trajectories. The experimental results demonstrate that our method is applicable to a wide range of indoor scenarios without the need for training or fine-tuning.
    Keywords Computer Science - Robotics
    Subject code 004
    Publishing date 2023-04-14
    Publishing country us
    Document type Book ; Online
    Database BASE - Bielefeld Academic Search Engine (life sciences selection)

    More links

    Kategorien

  6. Book ; Online: CoDEPS

    Vödisch, Niclas / Petek, Kürsat / Burgard, Wolfram / Valada, Abhinav

    Online Continual Learning for Depth Estimation and Panoptic Segmentation

    2023  

    Abstract: Operating a robot in the open world requires a high level of robustness with respect to previously unseen environments. Optimally, the robot is able to adapt by itself to new conditions without human supervision, e.g., automatically adjusting its ... ...

    Abstract Operating a robot in the open world requires a high level of robustness with respect to previously unseen environments. Optimally, the robot is able to adapt by itself to new conditions without human supervision, e.g., automatically adjusting its perception system to changing lighting conditions. In this work, we address the task of continual learning for deep learning-based monocular depth estimation and panoptic segmentation in new environments in an online manner. We introduce CoDEPS to perform continual learning involving multiple real-world domains while mitigating catastrophic forgetting by leveraging experience replay. In particular, we propose a novel domain-mixing strategy to generate pseudo-labels to adapt panoptic segmentation. Furthermore, we explicitly address the limited storage capacity of robotic systems by leveraging sampling strategies for constructing a fixed-size replay buffer based on rare semantic class sampling and image diversity. We perform extensive evaluations of CoDEPS on various real-world datasets demonstrating that it successfully adapts to unseen environments without sacrificing performance on previous domains while achieving state-of-the-art results. The code of our work is publicly available at http://codeps.cs.uni-freiburg.de.

    Comment: Accepted for "Robotics: Science and Systems (RSS) 2023"
    Keywords Computer Science - Robotics ; Computer Science - Computer Vision and Pattern Recognition
    Subject code 004 ; 629
    Publishing date 2023-03-17
    Publishing country us
    Document type Book ; Online
    Database BASE - Bielefeld Academic Search Engine (life sciences selection)

    More links

    Kategorien

  7. Book ; Online: Audio Visual Language Maps for Robot Navigation

    Huang, Chenguang / Mees, Oier / Zeng, Andy / Burgard, Wolfram

    2023  

    Abstract: While interacting in the world is a multi-sensory experience, many robots continue to predominantly rely on visual perception to map and navigate in their environments. In this work, we propose Audio-Visual-Language Maps (AVLMaps), a unified 3D spatial ... ...

    Abstract While interacting in the world is a multi-sensory experience, many robots continue to predominantly rely on visual perception to map and navigate in their environments. In this work, we propose Audio-Visual-Language Maps (AVLMaps), a unified 3D spatial map representation for storing cross-modal information from audio, visual, and language cues. AVLMaps integrate the open-vocabulary capabilities of multimodal foundation models pre-trained on Internet-scale data by fusing their features into a centralized 3D voxel grid. In the context of navigation, we show that AVLMaps enable robot systems to index goals in the map based on multimodal queries, e.g., textual descriptions, images, or audio snippets of landmarks. In particular, the addition of audio information enables robots to more reliably disambiguate goal locations. Extensive experiments in simulation show that AVLMaps enable zero-shot multimodal goal navigation from multimodal prompts and provide 50% better recall in ambiguous scenarios. These capabilities extend to mobile robots in the real world - navigating to landmarks referring to visual, audio, and spatial concepts. Videos and code are available at: https://avlmaps.github.io.

    Comment: Project page: https://avlmaps.github.io/
    Keywords Computer Science - Robotics ; Computer Science - Artificial Intelligence ; Computer Science - Computation and Language ; Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Machine Learning
    Subject code 629
    Publishing date 2023-03-13
    Publishing country us
    Document type Book ; Online
    Database BASE - Bielefeld Academic Search Engine (life sciences selection)

    More links

    Kategorien

  8. Article ; Online: Editorial: Language, affordance and physics in robot cognition and intelligent systems.

    Chen, Nutan / Mayol-Cuevas, Walterio W / Karl, Maximilian / Aljalbout, Elie / Zeng, Andy / Cortese, Aurelio / Burgard, Wolfram / van Hoof, Herke

    Frontiers in robotics and AI

    2024  Volume 10, Page(s) 1355576

    Language English
    Publishing date 2024-01-09
    Publishing country Switzerland
    Document type Editorial
    ZDB-ID 2781824-X
    ISSN 2296-9144 ; 2296-9144
    ISSN (online) 2296-9144
    ISSN 2296-9144
    DOI 10.3389/frobt.2023.1355576
    Database MEDical Literature Analysis and Retrieval System OnLINE

    More links

    Kategorien

  9. Book ; Online ; Thesis: <dc:title>Advanced techniques for autonomous navigation in precision agriculture</dc:title>

    Fleckenstein, Freya [Verfasser] / Burgard, Wolfram [Akademischer Betreuer] / Burgard, Wolfram [Sonstige] / Pradalier, Cédric Sonstige] / [Bödecker, Joschka [Sonstige]

    2022  

    Keywords Landwirtschaft, Veterinärmedizin ; Agriculture, Veterinary Science
    Subject code sg630
    Language English
    Publisher Universität
    Publishing place Freiburg
    Document type Book ; Online ; Thesis
    Database Digital theses on the web

    More links

    Kategorien

  10. Book ; Online: Few-Shot Panoptic Segmentation With Foundation Models

    Käppeler, Markus / Petek, Kürsat / Vödisch, Niclas / Burgard, Wolfram / Valada, Abhinav

    2023  

    Abstract: Current state-of-the-art methods for panoptic segmentation require an immense amount of annotated training data that is both arduous and expensive to obtain posing a significant challenge for their widespread adoption. Concurrently, recent breakthroughs ... ...

    Abstract Current state-of-the-art methods for panoptic segmentation require an immense amount of annotated training data that is both arduous and expensive to obtain posing a significant challenge for their widespread adoption. Concurrently, recent breakthroughs in visual representation learning have sparked a paradigm shift leading to the advent of large foundation models that can be trained with completely unlabeled images. In this work, we propose to leverage such task-agnostic image features to enable few-shot panoptic segmentation by presenting Segmenting Panoptic Information with Nearly 0 labels (SPINO). In detail, our method combines a DINOv2 backbone with lightweight network heads for semantic segmentation and boundary estimation. We show that our approach, albeit being trained with only ten annotated images, predicts high-quality pseudo-labels that can be used with any existing panoptic segmentation method. Notably, we demonstrate that SPINO achieves competitive results compared to fully supervised baselines while using less than 0.3% of the ground truth labels, paving the way for learning complex visual recognition tasks leveraging foundation models. To illustrate its general applicability, we further deploy SPINO on real-world robotic vision systems for both outdoor and indoor environments. To foster future research, we make the code and trained models publicly available at http://spino.cs.uni-freiburg.de.
    Keywords Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Robotics
    Subject code 004
    Publishing date 2023-09-19
    Publishing country us
    Document type Book ; Online
    Database BASE - Bielefeld Academic Search Engine (life sciences selection)

    More links

    Kategorien

To top