Exposing LLM Vulnerabilities: Adversarial Scam Detection and Performance

Date
2024-12
Language
English
Embargo Lift Date
Committee Members
Degree
Degree Year
Department
Grantor
Journal Title
Journal ISSN
Volume Title
Found At
IEEE
Can't use the file because of accessibility barriers? Contact us with the title of the item, permanent link, and specifics of your accommodation need.
Abstract

Can we trust Large Language Models (LLMs) to accurately predict scam? This paper investigates the vulnerabilities of LLMs when facing adversarial scam messages for the task of scam detection. We addressed this issue by creating a comprehensive dataset with fine-grained labels of scam messages, including both original and adversarial scam messages. The dataset extended traditional binary classes for the scam detection task into more nuanced scam types. Our analysis showed how adversarial examples took advantage of vulnerabilities of a LLM, leading to high misclassification rate. We evaluated the performance of LLMs on these adversarial scam messages and proposed strategies to improve their robustness.

Description
item.page.description.tableofcontents
item.page.relation.haspart
Cite As
Chang, C.-W., Sarkar, S., Mitra, S., Zhang, Q., Salemi, H., Purohit, H., Zhang, F., Hong, M., Cho, J.-H., & Lu, C.-T. (2024). Exposing LLM Vulnerabilities: Adversarial Scam Detection and Performance. 2024 IEEE International Conference on Big Data (BigData), 3568–3571. https://doi.org/10.1109/BigData62323.2024.10825256
ISSN
Publisher
Series/Report
Sponsorship
Major
Extent
Identifier
Relation
Journal
2024 IEEE International Conference on Big Data (BigData)
Source
ArXiv
Alternative Title
Type
Article
Number
Volume
Conference Dates
Conference Host
Conference Location
Conference Name
Conference Panel
Conference Secretariat Location
Version
Author's manuscript
Full Text Available at
This item is under embargo {{howLong}}