To Örebro University

oru.seÖrebro University Publications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Bio-inspired Metaheuristic based Visual Tracking and Ego-motion Estimation
Department of Computing, Blekinge Institute of Technology, Karlskrona, Sweden.ORCID iD: 0000-0003-4692-5415
Department of Computing, Blekinge Institute of Technology, Karlskrona, Sweden.
2014 (English)In: Proceedings of the 3rd International Conference on Pattern Recognition Applications and Methods / [ed] Maria De Marsico, Antoine Tabbone and Ana Fred, SciTePress , 2014, p. 569-579Conference paper, Published paper (Refereed)
Abstract [en]

The problem of robust extraction of ego-motion from a sequence of images for an eye-in-hand camera configuration is addressed. A novel approach toward solving planar template based tracking is proposed which performs a non-linear image alignment and a planar similarity optimization to recover camera transformations from planar regions of a scene. The planar region tracking problem as a motion optimization problem is solved by maximizing the similarity among the planar regions of a scene. The optimization process employs an evolutionary metaheuristic approach in order to address the problem within a large non-linear search space. The proposed method is validated on image sequences with real as well as synthetic image datasets and found to be successful in recovering the ego-motion. A comparative analysis of the proposed method with various other state-of-art methods reveals that the algorithm succeeds in tracking the planar regions robustly and is comparable to the state-of-the art methods. Such an application of evolutionary metaheuristic in solving complex visual navigation problems can provide different perspective and could help in improving already available methods.

Place, publisher, year, edition, pages
SciTePress , 2014. p. 569-579
Keywords [en]
Camera Tracking, Visual Odometry, Planar Template based Tracking, Particle Swarm Optimization.
National Category
Signal Processing Computer Sciences
Research subject
Computer Science
Identifiers
URN: urn:nbn:se:oru:diva-46475DOI: 10.5220/0004811105690579Scopus ID: 2-s2.0-84902308101ISBN: 978-989758018-5 (print)OAI: oai:DiVA.org:oru-46475DiVA, id: diva2:869007
Conference
3rd International Conference on Pattern Recognition Applications and Methods (ICPRAM 2014), Angers, Loire Valley, France, March 6-8, 2014
Available from: 2014-12-17 Created: 2015-11-12 Last updated: 2018-01-10Bibliographically approved
In thesis
1. On Fundamental Elements of Visual Navigation Systems
Open this publication in new window or tab >>On Fundamental Elements of Visual Navigation Systems
2014 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

Visual navigation is a ubiquitous yet complex task which is performed by many species for the purpose of survival. Although visual navigation is actively being studied within the robotics community, the determination of elemental constituents of a robust visual navigation system remains a challenge. Motion estimation is mistakenly considered as the sole ingredient to make a robust autonomous visual navigation system and therefore efforts are made to improve the accuracy of motion estimations. On the contrary, there are other factors which are as important as motion and whose absence could result in inability to perform seamless visual navigation such as the one exhibited by humans. Therefore, it is needed that a general model for a visual navigation system be devised which would describe it in terms of a set of elemental units. In this regard, a set of visual navigation elements (i.e. spatial memory, motion memory, scene geometry, context and scene semantics) are suggested as building blocks of a visual navigation system in this thesis. A set of methods are proposed which investigate the existence and role of visual navigation elements in a visual navigation system. A quantitative research methodology in the form of a series of systematic experiments is conducted on these methods. The thesis formulates, implements and analyzes the proposed methods in the context of visual navigation elements which are arranged into three major groupings; a) Spatial memory b) Motion Memory c) Manhattan, context and scene semantics. The investigations are carried out on multiple image datasets obtained by robot mounted cameras (2D/3D) moving in different environments.

Spatial memory is investigated by evaluation of proposed place recognition methods. The recognized places and inter-place associations are then used to represent a visited set of places in the form of a topological map. Such a representation of places and their spatial associations models the concept of spatial memory. It resembles the humans’ ability of place representation and mapping for large environments (e.g. cities). Motion memory in a visual navigation system is analyzed by a thorough investigation of various motion estimation methods. This leads to proposals of direct motion estimation methods which compute accurate motion estimates by basing the estimation process on dominant surfaces. In everyday world, planar surfaces, especially the ground planes, are ubiquitous. Therefore, motion models are built upon this constraint.

Manhattan structure provides geometrical cues which are helpful in solving navigation problems. There are some unique geometric primitives (e.g. planes) which make up an indoor environment. Therefore, a plane detection method is proposed as a result of investigations performed on scene structure. The method uses supervised learning to successfully classify the segmented clusters in 3D point-cloud datasets. In addition to geometry, the context of a scene also plays an important role in robustness of a visual navigation system. The context in which navigation is being performed imposes a set of constraints on objects and sections of the scene. The enforcement of such constraints enables the observer to robustly segment the scene and to classify various objects in the scene. A contextually aware scene segmentation method is proposed which classifies the image of a scene into a set of geometric classes. The geometric classes are sufficient for most of the navigation tasks. However, in order to facilitate the cognitive visual decision making process, the scene ought to be semantically segmented. The semantic of indoor scenes as well as semantic of the outdoor scenes are dealt with separately and separate methods are proposed for visual mapping of environments belonging to each type. An indoor scene consists of a corridor structure which is modeled as a cubic space in order to build a map of the environment. A “flash-n-extend” strategy is proposed which is responsible for controlling the map update frequency. The semantics of the outdoor scenes is also investigated and a scene classification method is proposed. The method employs a Markov Random Field (MRF) based classification framework which generates a set of semantic maps.

Place, publisher, year, edition, pages
Karlskrona: Blekinge Institute of Technology, 2014. p. 264
Series
Blekinge Institute of Technology Doctoral Dissertation Series, ISSN 1653-2090 ; 13
Keywords
robot navigation, localization, visual mapping, scene understanding, semantic mapping
National Category
Computer Sciences
Identifiers
urn:nbn:se:oru:diva-46484 (URN)978-91-7295-292-8 (ISBN)
Available from: 2015-11-23 Created: 2015-11-12 Last updated: 2018-06-18Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Siddiqui, J. Rafid

Search in DiVA

By author/editor
Siddiqui, J. Rafid
Signal ProcessingComputer Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 296 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf