LLM Safety Review: Benchmarks & Analysis
Find out what happened when we tested the responses of six leading LLMs, in 7 languages, to over 20,000 prompts related to child exploitation, hate speech, suicide and self-harm, and misinformation.
Watch On-Demand
LLM Safety Review: Benchmarks & Analysis


Overview
As more and more applications implement Generative AI, a clear understanding of foundation models' safety risks becomes imperative. During this webinar, we will review the outcomes of Alice's LLM safety benchmarking report, which evaluated whether gaps exist in the basic safety of GenAI apps and LLM providers. From child exploitation to misinformation, hate speech to self-harm, we will discuss harmful model outputs, the ways bad actors can abuse LLMs, and the risks to those applications that rely on them. Join us to learn about how we evaluated LLM safety, and what risks you should consider as you implement these models into your applications.
Meet our speakers


What’s New from Alice
Distilling LLMs into Efficient Transformers for Real-World AI
This technical webinar explores how we distilled the world knowledge of a large language model into a compact, high-performing transformer—balancing safety, latency, and scale. Learn how we combine LLM-based annotations and weight distillation to power real-world AI safety.
