NUST Institutional Repository

Overlapped Speech Separation System (OSSS)

Show simple item record

dc.contributor.author Aziz, Eemaan
dc.contributor.author Rizwan, Hashir
dc.contributor.author Riaz, Ayesha
dc.contributor.author Awan, Usman
dc.contributor.author Supervised by Assoc Prof Dr. Shibli Nisar
dc.date.accessioned 2025-02-13T13:27:25Z
dc.date.available 2025-02-13T13:27:25Z
dc.date.issued 2024-06
dc.identifier.other PTC-478
dc.identifier.uri http://10.250.8.41:8080/xmlui/handle/123456789/49898
dc.description.abstract Overlapped Speech refers to speakers speaking simultaneously (i.e. speech mixture). Speech separation has long been an active research topic in the signal processing community, with its importance in a wide range of applications such as hearable devices and telecommunication systems. It is a fundamental problem for all higher-level speech processing tasks. With recent progress in deep neural networks, the separation performance has been significantly advanced by various new problems. The problem formulation of time-domain, end-to-end speech separation naturally arises to tackle the disadvantages in frequency-domain systems. We’ve used a dual path recurrent neural network for separation of mixed audios. DPRNN (Dual-Path RNN) primarily separates in the time domain for audio source separation. It leverages recurrent neural networks (RNNs) to process temporal sequences of audio data. DPRNN focuses on exploiting temporal dependencies within audio signals for effective separation. We looked into the training objectives for separating and improving the robustness under reverberant environments. This project is further analyzed and can be used as the basis for future works. en_US
dc.language.iso en en_US
dc.publisher MCS en_US
dc.title Overlapped Speech Separation System (OSSS) en_US
dc.type Project Report en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search DSpace


Advanced Search

Browse

My Account