Events
Ensuring Safety and Accountability in LLMs, Pre- and Post Training
Speaker: Swabha Swayamdipta (USC)
Location: 60 Fifth Avenue, Room 7th floor open space
Date: Thursday, November 7, 2024
As large language models have become ubiquitous, it has proven increasingly challenging to enforce their accountability and safe deployment. In this talk, I will discuss the importance of ensuring the safety, responsibility, and accountability of Large Language Models (LLMs) throughout all stages of their development: pre-training, post-training evaluation, and deployment. First, I will present the idea of a unique LLM signature that can identify the model to ensure accountability. Next, I will present our recent work on reliably evaluating LLMs through our novel formulation of generation separability, and how this could lead to more reliable generation. Finally, I will present some ongoing work that demonstrates LLMs’ ability to understand but not generate unsafe or untrustworthy content.