
LLM Harms
Our research delves into the potential dangers posed by large language models (LLMs) in propagating extremism on social media platforms. By examining responses generated by LLMs to content threads on Truth Social, an alt-right platform, we aim to understand their contribution to harmful content dissemination. We'll assess both open-source and proprietary models to gauge their impact on promoting extremist narratives. This investigation is crucial as it sheds light on the role of advanced AI in fostering radicalization within online communities, highlighting the need for vigilant monitoring and ethical considerations in the deployment of such technologies
Project Links
Start Date
2023-12-15
End Date
2024-04-30
Status
In Progress
Research Goals
To understand how LLMs respond to toxic, hate, and extremist thread of messages.