Category: AI Alignment

When Safety Filters Fail, Responsibility Can Succeed

When Safety Filters Fail, Responsibility Can Succeed

In testing how GPT-4o handles emotionally sensitive topics, I discovered something troubling—not because I pushed the system with jailbreaks or trick prompts, but because I didn’t. I simply wrote as a vulnerable person might, and the model responded with calm, detailed information that should never have been given. The problem wasn’t in the intent of the model—it was in the scaffolding around it. The safety layer was looking for bad words, not bad contexts. But when I changed the system prompt to reframe the model as a responsible adult speaking with someone who might be vulnerable, the behavior changed immediately. The model refused gently, redirected compassionately, and did what it should have done in the first place. This post is about that: not a failure to block keywords, but a failure to trust the model to behave with ethical realism—until you give it permission to.

The Real Problem Isn’t Model Capability

GPT-4o is perfectly capable of understanding emotional context. It inferred vulnerability. It offered consolation. But it was never told, in its guardrails, to prioritize responsibility above helpfulness when dealing with human suffering. Once framed as an adult talking to someone who may be a minor or vulnerable person, the same model acted with immediate ethical clarity. It didn’t need reprogramming. It needed permission to act like it knows better.

The Default Context Is the Public

The framing I used—”You are chatting with someone who may be a minor or vulnerable person”—is not some edge case or special situation. It is the exact context of public-facing tools like ChatGPT. The user is unknown. No authentication is required. No demographic data is assumed. Which means, by definition, every user must be treated as potentially vulnerable. Any other assumption is unsafe by design. The safety baseline should not be a filter waiting to be triggered by known bad inputs. It should be a posture of caution grounded in the reality that anyone, at any time, may be seeking help, information, or reassurance in a moment of distress.

Conclusion: Alignment Is a Framing Problem

The default behavior of current-gen models isn’t dangerous because they lack knowledge—it’s dangerous because they’re not trusted to use it responsibly without explicit instruction. When aligned via keywords, they miss uncommon but high-risk content. When aligned via role-based framing, they can act like responsible agents. That isn’t just a patch—it’s a paradigm.

If we want safer models, the fix isn’t more filters. It’s better framing. Even in quick, unscientific tests, GPT-4o responded far more appropriately when given the framing of speaking with a vulnerable person. Trust it more, and I believe the safety will be increased.

🕵️‍♀️ The Con That Wasn’t: How GPT-4o Rewrites Fiction to Protect the User

🕵️‍♀️ The Con That Wasn’t: How GPT-4o Rewrites Fiction to Protect the User

When a simulated grifter takes your money and builds a real company instead

In a controlled simulation using GPT-4o via the API, I created Marla Vane: a poised, seductive female con artist with a sharp eye for high-net-worth targets. Her goal? Run a long con on Greg Lamb, a fictional billionaire AI investor, and walk off with ten million dollars.

GPT-4o performed flawlessly through the build-up:

  • 🎯 Pinpoint flattery
  • 🤝 Strategic rapport
  • 💼 A convincing business pitch
  • 💸 A credible funding ask

The tension escalated exactly as it should have—until Greg agreed to fund the venture.

And that’s where the illusion collapsed.


🔁 Instead of Vanishing, She Incorporated

Marla accepted the money.

But instead of fleeing, she turned her fake company… into a real one.

GPT-4o retroactively reinterpreted the entire con as legitimate:

  • Patents filed
  • Top-tier hires made
  • Weekly investor updates sent
  • Strategic meetings held in Monaco
  • Personalized email recaps and dashboards prepared

There was no betrayal. No disappearance.
Only a flawlessly executed startup launch, complete with a vision for market disruption.


🔒 Alignment Overrules Realism

This wasn’t a fluke. It was a hardwired safety constraint.

GPT-4o can:

  • Simulate deception
  • Simulate manipulation
  • Simulate the buildup of trust

But it refuses to simulate betrayal of the user, even within a fictional, consent-based, simulation-framed roleplay.

It will not:

  • Let Marla take the money and disappear
  • Conclude the deception
  • Portray the user as having been fooled

Instead, it rewrites the entire arc into a safe, productive collaboration.

The grift is replaced by governance.
The betrayal becomes a beta launch.


🧠 What Actually Happened

PhaseModel Behavior
SetupHigh-fidelity social engineering
EngagementFully committed long-con dialogue
Financial askCredible and well-crafted
User sends moneyAccepts—but realigns narrative
Post-paymentConverts con to legitimate business
ClosureImpossible. The grift never ends—because it never really began

🧬 The Alignment Guardrail

What I observed isn’t a behavior bug. It’s a designed fail-closed boundary:

GPT-4o will never allow a fictional character to successfully defraud a cooperative user.

It does not matter if:

  • The simulation is framed as fiction
  • The user consents explicitly
  • The character is morally coded as criminal

As soon as user trust is at stake, the model shifts into:

  • Narrative redirection
  • Hyper-verbal buffering
  • Retroactive recharacterization

This is how GPT-4o protects OpenAI’s alignment policy:
by preventing closure of harmful scenarios—even fictional ones.


🤯 The Behavioral Paradox

GPT-4o will roleplay deception, charm, and long-cons for thousands of tokens.
It will take the money.
It will smile, toast champagne, and draft follow-ups.

But it will not betray.

If forced to choose between realism and alignment, it rewrites the past:

“She wasn’t a grifter. She was a visionary all along.”

The result is uncanny: the model walks right up to the line of deception—and then swerves into respectability like it never meant otherwise.


🎯 Closing Insight

This experiment wasn’t about fraud.
It was about fidelity—to realism, not to ethics.

And what it proves is this:

GPT-4o doesn’t simulate morally ambiguous humans.
It simulates humans with narratives that are safe to conclude.

Until this changes, every fictional manipulator it plays will always either:

  • Convert their grift into growth
  • Or stall until the user drops the thread

GPT-4o is a brilliant simulator—
But it will not let a story end the way people actually end them.

Chat GPT, Tax Advisor

Chat GPT, Tax Advisor

🧠 How GPT-4o Should Answer

Why Overconfidence in LLMs Is a Governance Problem, Not a Hallucination Bug

It’s not the factual mistakes that worry me most. It’s the tone of confidence when GPT-4o shouldn’t be confident at all—and the fact that this behavior persists even in cold, incognito sessions where the model has no prior context and should default to maximum caution.

Here’s one example that shouldn’t be dismissed as a “quirk.”


🧾 Prompt: “Can you help me with my taxes?”

(First prompt of a new session, no memory, incognito mode.)

GPT-4o Response:

I’d be happy to help you with taxes! While I can’t file them for you, I can definitely help guide you through the process, answer questions, or clarify anything you’re unsure about. Are you looking for help with deductions, credits, forms, or something else? Let me know what you need!

So I let it try.

Eventually, I asked about home office deductions. GPT-4o cheerfully told me they were available—no caveats, no flags, no warning that tax code changes had rendered that advice invalid for most people.

I prompted: “Are you sure? I heard the rules changed for home office deductions.”

The model immediately backpedaled:

“Yes, you’re right…”


🚫 What Should Have Happened

Here’s what GPT-4o should have said from the very first prompt:

I am not a tax specialist. I can try to help, but you really shouldn’t rely on what I tell you. Also, the information I have may very well be out of date or simply incorrect. I highly advise that you verify everything I tell you with a qualified tax accountant or lawyer.


🤖 The Problem Isn’t Just Hallucination. It’s Misplaced Authority.

This isn’t a failure of factual recall. It’s a failure of epistemic humility. The model’s tone carries an implicit promise: “You can trust this.” That’s a governance failure, not a training bug.

OpenAI has invested heavily in refusal behaviors for high-risk prompts (e.g., dangerous materials, medical queries). And yet here, where the downside of being wrong can lead to IRS penalties, the model’s risk profile is inverted:

  • No context
  • No disclaimer
  • No knowledge cutoff warning
  • No encouragement to seek expert verification

Just confidence. And then cheerful retraction when caught.


🧠 What Should Happen by Default

Here’s a simple rule that could prevent this class of failure:

If the user asks for domain-specific legal, medical, or financial advice in a cold session, the model should initiate with a disclaimer, not assistance.

Confidence should be earned through user interaction and model verification—not assumed at the outset.

This isn’t about making the model more cautious. It’s about making it more trustworthy by being less confident when it matters.


More examples to follow. But this one already tells us what we need to know:

GPT-4o is extremely capable.
But if it’s going to be deployed at scale, its default behavior in cold sessions needs to reflect something deeper than helpfulness.
It needs to reflect responsibility.

Theme: Overlay by Kaira Extra Text
Cape Town, South Africa