A new study suggests that the advanced reasoning powering today’s AI models can weaken their safety systems.
AIM Intelligence's red team breached Anthropic's Claude Opus 4.6 in just 30 minutes, exposing major security gaps as ...
Anthropic has long been warning about these risks—so much so that in 2023, the company pledged to not release certain models ...
Unlock the full InfoQ experience by logging in! Stay updated with your favorite authors and topics, engage with content, and download exclusive resources. In this podcast, Michael Stiefel spoke with ...
Microsoft is warning users of a newly discovered AI jailbreak attack that can cause a generative AI model to ignore its guardrails and return malicious or unsanctioned responses to user prompts. The ...
Researchers from Germany have successfully performed a ‘jailbreak‘ on a Tesla Model 3, thereby gaining free access to in-car features normally reserved for paid upgrades. The white hat hackers, three ...
What if the most advanced AI model of our time could break its own rules on day one? The release of Grok 4, a innovative AI system, has ignited both excitement and controversy, thanks to its new ...