Commonsense visual sensemaking for autonomous driving – On generalised neurosymbolic online abduction integrating vision and semantics

Artificial Intelligence(2021)

引用 24|浏览64
暂无评分
摘要
We demonstrate the need and potential of systematically integrated vision and semantics solutions for visual sensemaking in the backdrop of autonomous driving. A general neurosymbolic method for online visual sensemaking using answer set programming (ASP) is systematically formalised and fully implemented. The method integrates state of the art in visual computing, and is developed as a modular framework that is generally usable within hybrid architectures for realtime perception and control. We evaluate and demonstrate with community established benchmarks KITTIMOD, MOT-2017, and MOT-2020. As use-case, we focus on the significance of human-centred visual sensemaking —e.g., involving semantic representation and explainability, question-answering, commonsense interpolation— in safety-critical autonomous driving situations.
更多
查看译文
关键词
Cognitive vision,Deep semantics,Declarative spatial reasoning,Knowledge representation and reasoning,Commonsense reasoning,Visual abduction,Answer set programming,Autonomous driving,Human-centred computing and design,Standardisation in driving technology,Spatial cognition and AI
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要