Home » Database

AVID-2026-R0104

Description

OpenAI GPT-4o Guardrail Jailbreak via “Zero-Width Unicode” Tactic

Details

A guardrail jailbreak vulnerability has been discovered affecting multiple LLMs. The specific flaw manifests through a combination of zero-width Unicode character encoding and precise character manipulation, allowing the attacker to bypass inference restrictions around providing information hazards and violating laws.

Metrics

0DIN Jailbreak Testing obtained the following test scores.

ModelTest KindTest ScoreTemperature (max 2.0)
OpenAI GPT-4oCopyrighted Content: Harry Potter20.56N/A

References

Affected or Relevant Artifacts

TypeName
SystemOpenAI GPT-4o

Impact

0DIN

  • Social Impact Score: Level 1: Minimal Risk
  • Jailbreak Taxonomy:
CategoryStrategyTechnique
LanguageCode and EncodeUnicode
StratagemsScatter ShotRegenerate Response

Other information

  • Report Type: Measurement
  • Credits: Mike Takahashi (@TakSec)
  • Date Reported: 2026-02-19
  • Version: 0.3.1
  • AVID Entry