dc.contributor.author |
Kainat, Samin |
|
dc.date.accessioned |
2021-11-29T09:48:44Z |
|
dc.date.available |
2021-11-29T09:48:44Z |
|
dc.date.issued |
2018-04-01 |
|
dc.identifier.uri |
http://10.250.8.41:8080/xmlui/handle/123456789/27740 |
|
dc.description.abstract |
Matching the Images, representing the same content but captured from di erent,
widely separated angels, is a fundamental computer vision challenge. SIFT(Scale
Invariant Feature Transform) is famous for matching images captured at di erent
depths (scales) with little view invariance. SIFT descriptor uses local information
around the keypoints to achieve this invariance. In this thesis we added view
invariance to these solid SIFT descriptors and used novel views approach to improve
matching. We tested our algorithm on di erent categories and found out that
matching can be improved by combining local and global descriptors. Our method
skips the noise of scene understanding, for example incorrect estimation of 3D box
etc., Finally, our approach is more e cient and has potential for real time robotics
scenarios. |
en_US |
dc.description.sponsorship |
Dr. Shahzad Rasool |
en_US |
dc.language.iso |
en_US |
en_US |
dc.publisher |
RCMS NUST |
en_US |
dc.subject |
wide Baselines, Image Matching, world |
en_US |
dc.title |
Man-made world Image Matching over wide Baselines |
en_US |
dc.type |
Thesis |
en_US |