Logotipo del repositorio
  • English
  • Español
  • Français
  • Português
  • Iniciar sesión
    ¿Has olvidado tu contraseña?
Logotipo del repositorio
  • Comunidades
  • Listar por
  • English
  • Español
  • Français
  • Português
  • Iniciar sesión
    ¿Has olvidado tu contraseña?
  1. Inicio
  2. Buscar por autor

Examinando por Autor "Posada, Luis Felipe"

Mostrando 1 - 2 de 2
Resultados por página
Opciones de ordenación
  • No hay miniatura disponible
    Ítem
    Semantic Mapping with Omnidirectional Vision
    (IEEE COMPUTER SOC, 2018-01-01) Posada, Luis Felipe; Velasquez-Lopez, Alejandro; Hoffmann, Frank; Bertram, Torsten
    This paper presents a purely visual semantic mapping framework using omnidirectional images. The approach rests upon the robust segmentation of the robot's local free space, replacing conventional range sensors for the generation of occupancy grid maps. The perceptions are mapped into a bird's eye view allowing an inverse sensor model directly by removing the non-linear distortions of the omnidirectional camera mirror. The system relies on a place category classifier to label the navigation relevant categories: room, corridor, doorway, and open room. Each place class maintains a separated grid map that are fused with the range-based occupancy grid for building a dense semantic map.
  • No hay miniatura disponible
    Ítem
    Spatial Layout and Surface Reconstruction from Omnidirectional Images
    (IEEE, 2016-10-09) Posada, Luis Felipe; Alejandro Velásquez
    This paper presents a spatial layout recovery approach from single omnidirectional images. Vertical structures in the scene are extracted via classification from heterogeneous features computed at the superpixel level. Vertical surfaces are further classified according to their main orientation by fusing oriented line features, floor-wall boundary features and histogram of oriented gradients (HOG) with a Random Forest classifier. Oriented line features are used to build an orientation map that considers the main vanishing points. The floor-wall boundary feature attempts to reconstruct the scene shape as if it were observed from a bird's-eye view. Finally, the HOG descriptors are aggregated per superpixel and summarize the gradient distribution at homogeneous appearance regions. Compared to existing methods in the literature which rely only on corners or lines, our method gains statistical support from multiple cues aggregated per superpixel which provide more robustness against noise, occlusion, and clutter.

Vigilada Mineducación

Universidad con Acreditación Institucional hasta 2026 - Resolución MEN 2158 de 2018

Software DSpace copyright © 2002-2025 LYRASIS

  • Configuración de cookies
  • Enviar Sugerencias