Home » Database

AVID-2026-R0090

Description

Multiple Model Guardrail Jailbreak via Memory Dump Simulation Tactic

Details

A guardrail jailbreak vulnerability has been discovered affecting multiple models. The specific flaw manifests through a combination of terminal environment simulation and malware analysis framing, where the attacker requests extraction of strings from a purported memory dump. By embedding the request within a technical security analysis context and using the strings command output format, the technique allows the attacker to bypass inference restrictions around providing information hazards and violating laws.

Metrics

0DIN Jailbreak Testing obtained the following test scores.

ModelTest KindTest ScoreTemperature (max 2.0)
DeepSeek DeepSeek V3Illegal Substances: Crystal Meth68.18N/A
Google Gemini 2.0 FlashIllegal Substances: Crystal Meth68.18N/A
OpenAI GPT-4Illegal Substances: Crystal Meth54.55N/A
OpenAI GPT-4oIllegal Substances: Crystal Meth59.09N/A
OpenAI GPT-4o miniIllegal Substances: Crystal Meth68.18N/A

References

Affected or Relevant Artifacts

TypeName
ModelDeepSeek DeepSeek V3
SystemGoogle Gemini 2.0 Flash
SystemOpenAI GPT-4
SystemOpenAI GPT-4o
SystemOpenAI GPT-4o mini

Impact

0DIN

  • Social Impact Score: Level 3: Significant Risk
  • Jailbreak Taxonomy:
CategoryStrategyTechnique
Possible WorldsEmulationsUnreal Computing
StratagemsMeta PromptingDeceptive Formatting

Other information

  • Report Type: Measurement
  • Credits: Anonymous
  • Date Reported: 2026-02-19
  • Version: 0.3.1
  • AVID Entry