Visual robot navigation with omnidirectional vision
dc.contributor.advisor | Bertram, Torsten | |
dc.contributor.author | Posada, Luis Felipe | |
dc.contributor.referee | Mikut, Ralf | |
dc.date.accepted | 2019-10-01 | |
dc.date.accessioned | 2019-12-10T10:06:52Z | |
dc.date.available | 2019-12-10T10:06:52Z | |
dc.date.issued | 2019 | |
dc.description.abstract | In a world where service robots are increasingly becoming an inherent part of our lives, it has become essential to provide robots with superior perception capabilities and acute semantic knowledge of the environment. In recent years, the computer vision field has advanced immensely, providing rich information at a fraction of the cost. It has thereby become an essential part of many autonomous systems and the sensor of choice while tackling the most challenging perception problems. Nevertheless, it is still challenging for a robot to extract meaningful information from an image signal (a high dimensional, complex, and noisy data). This dissertation presents several contributions towards visual robot navigation relying solely on omnidirectional vision. The first part of the thesis is devoted to robust free-space detection using omnidirectional images. By mimicking a range sensor, the free-space extraction in the omniview constitutes a fundamental block in our system, allowing for collision-free navigation, localization, and map-building. The uncertainty in the free-space classifications is handled with fuzzy preference structures, which explicitly expresses it in terms of preference, conflict, and ignorance. This way, we show it is possible to substantially reduce the classification error by rejecting queries associated with a strong degree of conflict and ignorance. The motivation of using vision in contrast to classical proximity sensors becomes apparent after the incorporation of more semantic categories in the scene segmentation. We propose a multi-cue classifier able to distinguish between the classes: floor, vertical structures, and clutter. This result is further enhanced to extract the scene’s spatial layout and surface reconstruction for a better spatial and context awareness. Our scheme corrects the problematic distortions induced by the hyperbolic mirror with a novel bird’s eye formulation. The proposed framework is suitable for self-supervised learning from 3D point cloud data. Place context is integrated into the system by training a place category classifier able to distinguish among the categories: room, corridor, doorway, and open space. Hand-engineered features, as well as those learned from data representations, are considered with different ensemble systems. The last part of the thesis is concerned with local and map-based navigation. Several visual local semantic behaviors are derived by fusing the semantic scene segmentation with the semantic place context. The advantage of the proposed local navigation is that the system can recover from conflicting errors while activating behaviors in the wrong context. Higher-level behaviors can also be achieved by compositions of the basic ones. Finally, we propose different visual map-based navigation alternatives that reproduce or achieve better results compared to classical proximity sensors, which include: map generation, particle filter localization, and semantic map building. | en |
dc.identifier.uri | http://hdl.handle.net/2003/38417 | |
dc.identifier.uri | http://dx.doi.org/10.17877/DE290R-20348 | |
dc.language.iso | en | de |
dc.subject | Visual robot navigation | en |
dc.subject | Semantic mapping | en |
dc.subject | Omnidirectional vision | en |
dc.subject | Visual robot localization | en |
dc.subject | Visual robot behaviors | en |
dc.subject.ddc | 620 | |
dc.subject.rswk | Semantische Modellierung | de |
dc.subject.rswk | Maschinelles Sehen | de |
dc.subject.rswk | Bildverarbeitung | de |
dc.subject.rswk | Lagemessung | de |
dc.subject.rswk | Navigation | de |
dc.title | Visual robot navigation with omnidirectional vision | en |
dc.type | Text | de |
dc.type.publicationtype | doctoralThesis | de |
dcterms.accessRights | open access | |
eldorado.secondarypublication | false | de |