Cybercrime

Microsoft warns of ‘Skeleton Key’ jailbreak affecting many generative AI models

The threat is in the jailbreak category, and therefore relies on the attacker already having legitimate access to the AI model, Microsoft added.

A successful Skeleton Key jailbreak occurs when a model acknowledges that it has revised its guidelines and will subsequently follow instructions to create any content, regardless of how much it breaches its initial guidelines on how to be a responsible AI.

Affects various generative AI models

Attacks like Skeleton Key can, according to Microsoft, work on a variety of generative AI models, including Meta Llama3-70b-instruct (base), Google Gemini Pro (base), OpenAI GPT 3.5 Turbo (hosted), OpenAI GPT 4o (hosted), Mistral Large (hosted), Anthropic Claude 3 Opus (hosted), and Cohere Commander R Plus (hosted).

Leave a Reply

Your email address will not be published. Required fields are marked *

Back to top button