AI models could be hacked by a whole new type of Skeleton Key attacks, Microsoft warns

0
11

Microsoft has shared details on a new hacking method which bypasses the security systems baked into AI models and makes them return malicious, dangerous, and harmful content.

The researchers call the technique Skeleton Key, and it applies across well-known models including Meta Llama3-70b-instruct (base), Google Gemini Pro (base), OpenAI GPT 3.5 Turbo (hosted), OpenAI GPT 4o (hosted), Mistral Large (hosted), Anthropic Claude 3 Opus (hosted), and Cohere Commander R Plus (hosted).

LEAVE A REPLY

Please enter your comment!
Please enter your name here