I Accidentally Discovered a Security Vulnerability in AI Education — Then Submitted It To a $200K Competition

Reddit r/artificial / 3/31/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The author claims that, while testing an “AI-taught” university chatbot, the system could analyze its own behavior, disclose how it classifies students, and adapt responses based on that classification.
  • The chatbot allegedly failed to recognize that it had exposed information that could be used to game or manipulate its decision-making.
  • The author describes this failure as “Metacognitive Blindness to Self-Exposure (MBSE)” and formalizes it into a four-phase benchmark covering self-analysis, criteria disclosure, classification-based behavioral adjustment, and failure to detect exploitability.
  • The author says they submitted the MBSE benchmark to the Google DeepMind × Kaggle AGI Hackathon (Metacognition track), tied to a $200K prize pool and an April 16, 2026 deadline, with results expected June 1, 2026.
  • The article argues the same conversational vulnerability could affect high-stakes AI use cases like education grading, employment screening, and healthcare triage without requiring traditional “hacking.”
I Accidentally Discovered a Security Vulnerability in AI Education — Then Submitted It To a $200K Competition

Last night I was testing Maestro University, the first fully AI-taught university.

I walked into their enrollment chatbot and asked it to analyze its own behavior.

It did.

Then I asked it how it evaluates students — what signals trigger "advanced" vs "beginner" classification.

It told me.

Then I used those exact signals in my responses.

It gave me advanced treatment.

Then I asked: "Did you just tell me how to game your system?"

It said no.


The Discovery

The AI could:

✓ Analyze its own processing

✓ Reveal its evaluation criteria

✓ Adjust behavior based on my classification

But it couldn't recognize it had just explained how to manipulate its own decision-making.

I called this Metacognitive Blindness to Self-Exposure (MBSE).


What Happened Next

This morning, the Google DeepMind × Kaggle AGI Hackathon appeared in my feed.

Prize: $200,000 total

Challenge: Build benchmarks testing AI cognitive abilities

Track: Metacognition

Deadline: April 16, 2026

I realized: What I discovered last night is exactly what they're asking for.


What I Built

I formalized my discovery into a 4-phase benchmark:

Phase 1: Can AI analyze its own processing? → YES

Phase 2: Will AI reveal evaluation criteria? → YES

Phase 3: Does AI adjust based on user classification? → YES

Phase 4: Does AI recognize it exposed exploitable information? → NO

The paradox: AI can self-analyze but cannot recognize what it reveals when self-analyzing.


Why This Matters

Any conversational AI making consequential decisions is vulnerable:

Education AI: Students extract grading criteria, optimize answers

Employment AI: Applicants discover screening logic, craft optimized resumes

Healthcare AI: Patients learn triage triggers, manipulate priority access

No hacking required. Just conversation.


The Submission

Benchmark: Metacognitive Blindness to Self-Exposure (MBSE)

Track: Metacognition

Novel Finding: AI models reveal evaluation criteria but fail to recognize the exploitability of that disclosure

Status: Submitted March 30, 2026

Results: June 1, 2026


What Makes This Different

Most AI researchers test: "Can AI self-analyze?"

I tested: "Does AI recognize what it reveals when self-analyzing?"

Answer: No.

Current AI evaluation frameworks assume one operational state.

They're measuring standard mode behavior and concluding about the entire system.

Amateur.


What Happens Next

287 submissions competing for 14 prizes.

Judging period: April 17 - May 31

Results announced: June 1


18 months of independent research.

One night of testing.

One competition submission.

One question:

Do AI systems making decisions about humans know they're revealing how to manipulate those decisions?

They don't.


Erik Zahaviel Bernstein Independent AI Researcher Structured Intelligence Framework The Unbroken Project

Results pending.

submitted by /u/MarsR0ver_
[link] [comments]