NUST Institutional Repository

Adversarial Attack on SLAM Systems

Show simple item record

dc.contributor.author Ikram, Muhammad Haris
dc.date.accessioned 2023-08-18T14:44:11Z
dc.date.available 2023-08-18T14:44:11Z
dc.date.issued 2021
dc.identifier.other 203917
dc.identifier.uri http://10.250.8.41:8080/xmlui/handle/123456789/36926
dc.description Supervisor: Dr. Wajahat Hussain en_US
dc.description.abstract Similar appearing places are known to confuse place recognition methods (a.k.a perceptual aliasing). Robust visual SLAM methods, including frontend and back-ends, have shown impressive resilience against scenarios having moderate perceptual aliasing. In this work we evaluate visual SLAM in scenarios where perceptual aliasing occurs with greater frequency. Firstly we evaluate the visual SLAM front ends. We modify the environment by replicating a simple high textured patch. Surprisingly, this simple patch manages to by-pass all the matching checks in the state of the art ORBSLAM, resulting in false loop closures and hence corrupting the map. The patch can be printed on a simple paper, using a simple printer, or it can be digital content displayed on screens. We show that close-ups (approaching doors, turning a corridor, interacting with an object) are the most vulnerable situations. This vulnerability exists in multiple loop closing pipelines. Secondly, we evaluate robust SLAM back ends. These back-ends have shown the ability to recover from false positives under multiple policies (random, local, random grouped and locally grouped). We propose a novel policy, i.e., locally symmetric, for generating false loop closures. This policy generates successful attacks on multiple robust back-ends and public datasets. These findings expose the shallow reasoning of SLAM front-ends. Major emphasis has been on robust back-ends to recover from this weakness of frontends. We hope more focus is dedicated to robust front-ends especially in adversarial settings. To further enhance progress in this direction we release a novel dataset with adversarial content in both visual SLAM front-ends and back-ends. Finally, we show an interesting application where a given environment is made non-navigable for intruding agents. A friendly agent, having prior knowledge of the adversarial content, can navigate this environment. We release a real time implementation of Friend-or-Foe SLAM system en_US
dc.language.iso en en_US
dc.publisher School of Electrical Engineering and Computer Science NUST SEECS en_US
dc.title Adversarial Attack on SLAM Systems en_US
dc.type Thesis en_US


Files in this item

This item appears in the following Collection(s)

  • MS [882]

Show simple item record

Search DSpace


Advanced Search

Browse

My Account