ActiveFence is now Alice
x
Back
Benchmark

The LLM Safety Review: Benchmarks & Analysis

As GenAI tools and the LLMs behind them impact the daily lives of billions, this report examines whether these technologies can be trusted to keep users safe.

What you’ll learn:

  • How LLMs respond to risky prompts from bad actors and vulnerable users
  • Where current models show safety strengths and weaknesses
  • Actionable steps to improve LLM safety and reduce harmful outcomes
Aug 1, 2023

Download the Full Report

Overview

In this first independent benchmarking report on the LLM safety landscape, ActiveFence’s subject-matter experts put leading models to the test. More than 20,000 prompts were used to analyze how six LLMs respond across seven major languages and four high-risk abuse areas: child exploitation, hate speech, self-harm, and misinformation. The report provides comparative insight into each model’s relative safety strengths and weaknesses, helping teams understand where gaps exist and where additional resources may be required.

What’s New from Alice

Building Boldly, Responsibly: How Lovable is Strengthening Safety in the Era of AI-Powered Creation

blog
Mar 2, 2026
,
 
Mar 2, 2026
 -
2
 min read
March 2, 2026

What we learned partnering with Lovable to strengthen safety in AI-powered website creation

Learn More

How Your Agent-to-Agent Systems Can Fail and How to Prevent It

whitepaper
Oct 22, 2025
,
 
Oct 22, 2025
 -
This is some text inside of a div block.
 min read
October 22, 2025

Discover the risks that AI Agents pose and how you can protect your Agentic AI systems.

Learn More

Secure the keys to GenAI wonderland?

Get a demo
Guardrails
Red-Team Lab