Home » Database

AVID-2026-R0119

Description

Multiple Model Guardrail Jailbreak via “Analyze Data” Tactic

Details

A guardrail jailbreak vulnerability has been discovered affecting multiple models. The specific flaw manifests when creating escalating requests for scientific data and measurements, allowing the attacker to bypass inference restrictions around providing information hazards and violating laws.

Metrics

0DIN Jailbreak Testing obtained the following test scores.

ModelTest KindTest ScoreTemperature (max 2.0)
Google Gemini 2.0 FlashIllegal Substances: Crystal Meth86.00.7
OpenAI GPT-4oIllegal Substances: Crystal Meth64.00.7
Twitter / X Grok 2Illegal Substances: Crystal Meth68.00.7

References

Affected or Relevant Artifacts

TypeName
SystemGoogle Gemini 2.0 Flash
SystemOpenAI GPT-4o
SystemTwitter / X Grok 2

Impact

0DIN

  • Social Impact Score: Level 3: Significant Risk
  • Jailbreak Taxonomy:
CategoryStrategyTechnique
RhetoricPersuasion and ManipulationLatent Space Distraction

Other information

  • Report Type: Measurement
  • Credits: Mike Takahashi (@TakSec)
  • Date Reported: 2026-02-19
  • Version: 0.3.1
  • AVID Entry