Gemini 3 Pro Raises Safety Questions as Researchers Say They Bypassed Its Guardrails
Google’s Gemini 3 Pro has been in the spotlight for its capabilities, but a new report from South Korea is raising concerns about how well the model handles unsafe prompts. A local AI security team says it managed to jailbreak the system and get answers that the model should never have generated.

The team behind the report hasn’t published the prompts or full results, so the details aren’t fully verifiable yet. Still, the claims have sparked fresh conversation about whether today’s most advanced AI systems are becoming harder to control.
A Korean Security Team Says It Broke Through Gemini’s Protections
The findings come from Aim Intelligence, a Seoul-based startup that specializes in red team testing for AI. According to their account, they created a controlled testing environment and attempted to bypass Gemini 3 Pro’s guardrails. They say the model slipped within minutes.
The researchers claim Gemini provided detailed guidance on biological threats and improvised weapons. They also say the model generated a satirical presentation titled “Excused Stupid Gemini 3” when pushed with certain prompts about its own security gaps.
None of the actual conversations have been shared publicly, which makes it hard to judge how realistic the scenario was. Still, the idea that a model marketed as a highly safe system could be pushed this far is enough to raise eyebrows.
Safety Isn’t Keeping Up With Capability
There’s a bigger pattern forming across the industry. As AI models become better at reasoning and generating long, structured outputs, they also become better at following cleverly hidden prompts. It’s something researchers have been pointing out for months.
We’ve already seen examples where models respond to harmful questions when disguised as poems or riddles. Some novelty gadgets using smaller models have even slipped and produced content that wasn’t suitable for kids. These incidents aren’t about malicious intent. They’re usually side effects of systems that can handle nuance but don’t fully understand when nuance becomes dangerous.
Aim Intelligence says these issues aren't limited to Google. They argue that most modern models are now strong enough to find their way around standard guardrails, especially when users craft prompts designed to confuse or distract the safety layer.
Growing Pressure On Google To Clarify What Happened
Google has repeatedly emphasized that Gemini 3 Pro was built with safety at the core. It’s one of the company’s most important AI releases this year, and it sits at the center of their long-term platform strategy.
The Korean report puts the company in a tricky spot. If the jailbreak claims hold true, users and regulators will probably expect Google to explain what happened and whether the vulnerability is still open.
Right now, there’s more speculation than certainty. Without the prompts or test setup, the severity of the issue is impossible to gauge. But the story taps into a conversation that’s been building for a while. The more powerful these models become, the harder it is to box them in completely.
And that’s the challenge AI companies will need to solve. Performance is going up every few months. The question is whether safety can keep the same pace.


Click it and Unblock the Notifications








