This Is AuburnElectronic Theses and Dissertations

Show simple item record

A Study of Adversarial Attacks on Machine Learning-based Fake News Detection Systems


Metadata FieldValueLanguage
dc.contributor.advisorDozier, Gerry
dc.contributor.authorBrown, Brandon
dc.date.accessioned2023-05-04T13:43:46Z
dc.date.available2023-05-04T13:43:46Z
dc.date.issued2023-05-04
dc.identifier.urihttps://etd.auburn.edu//handle/10415/8718
dc.description.abstractDue to the increased use and reliance on social media, fake news has become a significant problem that can cause great harm to individuals. Because of the dangerousness of fake news, techniques must be developed to detect and keep it from spreading. Currently, fact-checking is a damage control strategy that is essential to detecting and mitigating fake news. Websites such as Politifact, Snopes, and Factcheck.org use human verifiers to manually fact-check news articles. In cases where there are only a few articles to be fact-checked by human verifiers, relying on these websites to take the time to effectively research and debunk fake news is sufficient. However, in the case of social media, where news is generated at an extremely high volume (and with an extremely high velocity), automated approaches are needed for fake news detection. Recently, social media companies have begun to rely on automated systems in the form of machine learning-based fake news detection systems (ML-FNDSs). These ML-FNDSs are used to classify articles as either news or fake news. Although these ML-FNDSs effectively classify articles as either news or fake news, they are susceptible to adversarial attacks. Adversaries use these attacks to make a ML-FNDS misclassify its input. Adversarial attacks can be used to make a ML-FNDS accept fake news as news or flag news as being fake. In our research, we study two potential vulnerabilities of ML-FNDSs with respect to false positives (e.g., when news is erroneously classified as fake news) and false negatives (e.g., when fake news is erroneously classified as news). In this dissertation, we first introduce the concept of an Adversarial Universal False Positive (UFP) Attack and the Adversarial Universal False Negative (UFN) Attack. Next, we study the effectiveness of these two attacks on ML-FNDSs based on a single classifier, and finally, we study these attacks on ML-FNDSs based on a set of classifiers (ensemble machines).en_US
dc.rightsEMBARGO_GLOBALen_US
dc.subjectComputer Science and Software Engineeringen_US
dc.titleA Study of Adversarial Attacks on Machine Learning-based Fake News Detection Systemsen_US
dc.typePhD Dissertationen_US
dc.embargo.lengthMONTHS_WITHHELD:60en_US
dc.embargo.statusEMBARGOEDen_US
dc.embargo.enddate2028-05-04en_US

Files in this item

Show simple item record