dc.contributor.author |
Project Supervisor Dr. Usman Akram, Ns Zain Ul Abaidin Ns Muhammad Afif Ul Hasnain Ns Salman Ur Rehman |
|
dc.date.accessioned |
2025-03-13T06:11:43Z |
|
dc.date.available |
2025-03-13T06:11:43Z |
|
dc.date.issued |
2021 |
|
dc.identifier.other |
DE-COMP-39 |
|
dc.identifier.uri |
http://10.250.8.41:8080/xmlui/handle/123456789/50964 |
|
dc.description |
Project Supervisor Dr. Usman Akram |
en_US |
dc.description.abstract |
AI (Artificial Intelligence) must be able to grasp information holistically, like humans do, to
improve its effectiveness as a tool for identifying hate speech. When we look at a meme, for
example, we do not separate the words from the visuals; we get the meaning. Machines, on the
other hand, have a huge difficulty since they are unable to evaluate each word and image
separately. They must be able to combine these many modalities and understand how the meaning
shifts when they are provided together.
The goal is to create early in the categorization process tools that integrate the numerous modalities
present in a piece of material. The team created a Google Chrome Extension that scrapes posts
from a user's Facebook account and utilizes Machine Learning Models to determine whether the
messages are hostile. The Google Chrome Extension blurs posts containing hostile material if the
content is hateful. The text sentiment analysis model classifies the combined content from OCR
and picture captioning model as hateful and non-hateful, and the google extension blurs the hateful
post on the user's Facebook account. |
en_US |
dc.language.iso |
en |
en_US |
dc.publisher |
College of Electrical & Mechanical Engineering (CEME), NUST |
en_US |
dc.title |
Hateful Content Identification and Censorship on Facebook |
en_US |
dc.type |
Project Report |
en_US |