dc.contributor.author | Ahmad Ali, supervised by Dr Hasan Sajid | |
dc.date.accessioned | 2022-07-25T07:21:59Z | |
dc.date.available | 2022-07-25T07:21:59Z | |
dc.date.issued | 2022 | |
dc.identifier.uri | http://10.250.8.41:8080/xmlui/handle/123456789/29938 | |
dc.description.abstract | Person authentication is a primary element to consider wherever privacy is necessary. Deep learning based authentication algorithms have a number of applications in the said field. Adding multiple modalities makes the system more robust. In this research a joint multi-modal audio-visual deep learning based method has been devised to authenticate a person based on their voice as well as face. This two-step verification process works by learning face-feature based embeddings as well as voice-feature based embeddings to serve two purposes: 1) if the face presented matches with an identity in a reference database and 2) if the voice matches any voice in the reference database. This strategy can help prevent important systems from impostor attempts using modalities that are commonly present and available in consumer devices. | en_US |
dc.language.iso | en | en_US |
dc.publisher | SMME | en_US |
dc.subject | Audio-Visual Person Recognition | en_US |
dc.title | Audio-Visual Person Recognition | en_US |
dc.type | Thesis | en_US |