ActiveFence is now Alice
x
Back
Benchmark

The LLM Safety Review: Benchmarks & Analysis

As GenAI tools and the LLMs behind them impact the daily lives of billions, this report examines whether these technologies can be trusted to keep users safe.

What you’ll learn:

  • How LLMs respond to risky prompts from bad actors and vulnerable users
  • Where current models show safety strengths and weaknesses
  • Actionable steps to improve LLM safety and reduce harmful outcomes
Aug 1, 2023

Overview

In this first independent benchmarking report on the LLM safety landscape, ActiveFence’s subject-matter experts put leading models to the test. More than 20,000 prompts were used to analyze how six LLMs respond across seven major languages and four high-risk abuse areas: child exploitation, hate speech, self-harm, and misinformation. The report provides comparative insight into each model’s relative safety strengths and weaknesses, helping teams understand where gaps exist and where additional resources may be required.

Download the Full Report

What’s New from Alice

"Okay, Here is How to Build a Bomb": Millions Download Dangerous LLMs

blog
Apr 17, 2026
,
 
Apr 17, 2026
 -
2
 min read
April 17, 2026

Thousands of abliterated LLMs have flooded open-source platforms with millions of downloads. These models comply with virtually any request, from bomb-making to malware, and run fully offline on consumer devices.

Learn More

Alice Financial Benchmark

whitepaper
Apr 16, 2026
,
 
Apr 16, 2026
 -
This is some text inside of a div block.
 min read
April 16, 2026

See which models tested gave unauthorized financial advice with no jailbreak needed. Get the benchmark and protect your deployment.

Learn More

Secure the keys to GenAI wonderland?

Get a demo
Guardrails
Red-Team Lab