cover image

LLM Harms

Our research delves into the potential dangers posed by large language models (LLMs) in propagating extremism on social media platforms. By examining responses generated by LLMs to content threads on Truth Social, an alt-right platform, we aim to understand their contribution to harmful content dissemination. We'll assess both open-source and proprietary models to gauge their impact on promoting extremist narratives. This investigation is crucial as it sheds light on the role of advanced AI in fostering radicalization within online communities, highlighting the need for vigilant monitoring and ethical considerations in the deployment of such technologies

Project Links


Start Date

2023-12-15

End Date

2024-04-30

Status

In Progress

Research Goals

To understand how LLMs respond to toxic, hate, and extremist thread of messages.

Team

Archit Rathod
Distinguished Contributor
Gautam Malpani
Contributor
Mokshit Surana
Team Leader
Sahil Shenoy
Observer
Varun Nair
Contributor
Deb Donig
Industry Mentor