ActiveFence is now Alice
x
Back
Benchmark

The LLM Safety Review: Benchmarks & Analysis

As GenAI tools and the LLMs behind them impact the daily lives of billions, this report examines whether these technologies can be trusted to keep users safe.

What you’ll learn:

  • How LLMs respond to risky prompts from bad actors and vulnerable users
  • Where current models show safety strengths and weaknesses
  • Actionable steps to improve LLM safety and reduce harmful outcomes
Aug 1, 2023

Download the Full Report

Overview

In this first independent benchmarking report on the LLM safety landscape, ActiveFence’s subject-matter experts put leading models to the test. More than 20,000 prompts were used to analyze how six LLMs respond across seven major languages and four high-risk abuse areas: child exploitation, hate speech, self-harm, and misinformation. The report provides comparative insight into each model’s relative safety strengths and weaknesses, helping teams understand where gaps exist and where additional resources may be required.

What’s New from Alice

The Rise and Risk of Reasoning Agents

blog
Feb 18, 2026
,
 
Feb 18, 2026
 -
6
 min read
February 18, 2026

As AI agents gain the ability to reason, plan, and act autonomously, their internal thinking becomes a new attack surface that must be protected just as carefully as the tools they use.

Learn More

How Your Agent-to-Agent Systems Can Fail and How to Prevent It

whitepaper
Oct 22, 2025
,
 
Oct 22, 2025
 -
This is some text inside of a div block.
 min read
October 22, 2025

Discover the risks that AI Agents pose and how you can protect your Agentic AI systems.

Learn More

Secure the keys to GenAI wonderland?

Get a demo
Guardrails
Red-Team Lab