NUST Institutional Repository

Audio-Visual Person Recognition

Show simple item record

dc.contributor.author Ahmad Ali, supervised by Dr Hasan Sajid
dc.date.accessioned 2022-07-25T07:21:59Z
dc.date.available 2022-07-25T07:21:59Z
dc.date.issued 2022
dc.identifier.uri http://10.250.8.41:8080/xmlui/handle/123456789/29938
dc.description.abstract Person authentication is a primary element to consider wherever privacy is necessary. Deep learning based authentication algorithms have a number of applications in the said field. Adding multiple modalities makes the system more robust. In this research a joint multi-modal audio-visual deep learning based method has been devised to authenticate a person based on their voice as well as face. This two-step verification process works by learning face-feature based embeddings as well as voice-feature based embeddings to serve two purposes: 1) if the face presented matches with an identity in a reference database and 2) if the voice matches any voice in the reference database. This strategy can help prevent important systems from impostor attempts using modalities that are commonly present and available in consumer devices. en_US
dc.language.iso en en_US
dc.publisher SMME en_US
dc.subject Audio-Visual Person Recognition en_US
dc.title Audio-Visual Person Recognition en_US
dc.type Thesis en_US


Files in this item

This item appears in the following Collection(s)

  • MS [342]

Show simple item record

Search DSpace


Advanced Search

Browse

My Account