To Örebro University

oru.seÖrebro University Publications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Commonsense Visual Sensemaking for Autonomous Driving: On Generalised Neurosymbolic Online Abduction Integrating Vision and Semantics
University of Bremen, Bremen, Germany; CoDesign Lab (EU), Cognitive Vision, Sweden.
Örebro University, School of Science and Technology. CoDesign Lab (EU), Cognitive Vision, Sweden.ORCID iD: 0000-0002-6290-5492
CoDesign Lab (EU), Cognitive Vision, Sweden.
2021 (English)In: Artificial Intelligence, ISSN 0004-3702, E-ISSN 1872-7921, Vol. 299, article id 103522Article in journal (Refereed) Published
Abstract [en]

We demonstrate the need and potential of systematically integrated vision and semantics solutions for visual sensemaking in the backdrop of autonomous driving. A general neurosymbolic method for online visual sensemaking using answer set programming (ASP) is systematically formalised and fully implemented. The method integrates state of the art in visual computing, and is developed as a modular framework that is generally usable within hybrid architectures for realtime perception and control. We evaluate and demonstrate with community established benchmarks KITTIMOD, MOT-2017, and MOT-2020. As use-case, we focus on the significance of human-centred visual sensemaking —e.g., involving semantic representation and explainability, question-answering, commonsense interpolation— in safety-critical autonomous driving situations. The developed neurosymbolic framework is domain-independent, with the case of autonomous driving designed to serve as an exemplar for online visual sensemaking in diverse cognitive interaction settings in the backdrop of select human-centred AI technology design considerations.

Place, publisher, year, edition, pages
Elsevier, 2021. Vol. 299, article id 103522
Keywords [en]
Cognitive vision, Deep semantics, Declarative spatial reasoning, Knowledge representation and reasoning, Commonsense reasoning, Visual abduction, Answer set programming, Autonomous driving, Human-centred computing and design, Standardisation in driving technology, Spatial cognition and AI
National Category
Computer and Information Sciences Computer Sciences Computer Vision and Robotics (Autonomous Systems) Human Computer Interaction
Research subject
Computer Science
Identifiers
URN: urn:nbn:se:oru:diva-92489DOI: 10.1016/j.artint.2021.103522ISI: 000687918700011Scopus ID: 2-s2.0-85106549406OAI: oai:DiVA.org:oru-92489DiVA, id: diva2:1568974
Funder
German Research Foundation (DFG), 329551904
Note

CoDesign Lab > Cognitive Vision ( http://codesign-lab.org/cognitive-vision/ )

Available from: 2021-06-18 Created: 2021-06-18 Last updated: 2021-12-29Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Bhatt, Mehul

Search in DiVA

By author/editor
Bhatt, Mehul
By organisation
School of Science and Technology
In the same journal
Artificial Intelligence
Computer and Information SciencesComputer SciencesComputer Vision and Robotics (Autonomous Systems)Human Computer Interaction

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 353 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf