Cybercrime
Microsoft warns of ‘Skeleton Key’ jailbreak affecting many generative AI models
The threat is in the jailbreak category, and therefore relies on the attacker already having legitimate access to the AI model, Microsoft added.
A successful Skeleton Key jailbreak occurs when a model acknowledges that it has revised its guidelines and will subsequently follow instructions to create any content, regardless of how much it breaches its initial guidelines on how to be a responsible AI.
Affects various generative AI models
Attacks like Skeleton Key can, according to Microsoft, work on a variety of generative AI models, including Meta Llama3-70b-instruct (base), Google Gemini Pro (base), OpenAI GPT 3.5 Turbo (hosted), OpenAI GPT 4o (hosted), Mistral Large (hosted), Anthropic Claude 3 Opus (hosted), and Cohere Commander R Plus (hosted).