Commonsense visual sensemaking for autonomous driving – On generalised neurosymbolic online abduction integrating vision and semantics (bibtex)
by Jakob Suchan, Mehul Bhatt and Srikrishna Varadarajan
Abstract:
We demonstrate the need and potential of systematically integrated vision and semantics solutions for visual sensemaking in the backdrop of autonomous driving. A general neurosymbolic method for online visual sensemaking using answer set programming (ASP) is systematically formalised and fully implemented. The method integrates state of the art in visual computing, and is developed as a modular framework that is generally usable within hybrid architectures for realtime perception and control. We evaluate and demonstrate with community established benchmarks KITTIMOD, MOT-2017, and MOT-2020. As use-case, we focus on the significance of human-centred visual sensemaking —e.g., involving semantic representation and explainability, question-answering, commonsense interpolation— in safety-critical autonomous driving situations. The developed neurosymbolic framework is domain-independent, with the case of autonomous driving designed to serve as an exemplar for online visual sensemaking in diverse cognitive interaction settings in the backdrop of select human-centred AI technology design considerations.
Reference:
Commonsense visual sensemaking for autonomous driving – On generalised neurosymbolic online abduction integrating vision and semantics (Jakob Suchan, Mehul Bhatt and Srikrishna Varadarajan), In Artificial Intelligence, volume 299, 2021.
Bibtex Entry:
@article{SUCHAN2021103522,
title = {Commonsense visual sensemaking for autonomous driving – On generalised neurosymbolic online abduction integrating vision and semantics},
journal = {Artificial Intelligence},
volume = {299},
pages = {103522},
year = {2021},
issn = {0004-3702},
doi = {https://doi.org/10.1016/j.artint.2021.103522},
url = {https://www.sciencedirect.com/science/article/pii/S0004370221000734},
author = {Jakob Suchan and Mehul Bhatt and Srikrishna Varadarajan},
keywords = {Cognitive vision, Deep semantics, Declarative spatial reasoning, Knowledge representation and reasoning, Commonsense reasoning, Visual abduction, Answer set programming, Autonomous driving, Human-centred computing and design, Standardisation in driving technology, Spatial cognition and AI},
abstract = {We demonstrate the need and potential of systematically integrated vision and semantics solutions for visual sensemaking in the backdrop of autonomous driving. A general neurosymbolic method for online visual sensemaking using answer set programming (ASP) is systematically formalised and fully implemented. The method integrates state of the art in visual computing, and is developed as a modular framework that is generally usable within hybrid architectures for realtime perception and control. We evaluate and demonstrate with community established benchmarks KITTIMOD, MOT-2017, and MOT-2020. As use-case, we focus on the significance of human-centred visual sensemaking —e.g., involving semantic representation and explainability, question-answering, commonsense interpolation— in safety-critical autonomous driving situations. The developed neurosymbolic framework is domain-independent, with the case of autonomous driving designed to serve as an exemplar for online visual sensemaking in diverse cognitive interaction settings in the backdrop of select human-centred AI technology design considerations.}
}
Powered by bibtexbrowser