Musa Molla

AI Systems Age Faster Than We Expect

by

Here’s something that surprised us.

AI systems don’t just run, they age.

Prompts become stale.
Assumptions stop matching reality.
Small shortcuts compound quietly.

Nothing breaks overnight.
But reliability slowly erodes.

Maintaining AI isn’t just monitoring uptime.
It’s actively renewing assumptions.

Curious to hear from this crowd:
What part of your AI system needs the most “refreshing” today?

81 views

Add a comment

Replies

Best
Douglas Li

I’ve found that with every new model version, I need to re-review outputs to make sure quality hasn’t shifted. We use a synthetic data pipeline, but even small changes in how a model writes can subtly change the output distribution.

A lot of this feels like prompt overfitting -- similar to classic ML overfitting, but where prompts are implicitly tuned to a specific model version and don’t generalise cleanly to the next one. When we upgraded from gpt-4o to gpt-5, we actually saw a drop in performance.

Having proper evals helps :)

Musa Molla

@dougli “Prompt overfitting” is a great way to describe it. Prompts silently bind themselves to model behavior, so upgrades can regress quality even when capability improves. This is why evals need to track distribution shifts, not just accuracy.

Bhavin Sheth

This resonates a lot. In my experience, context and assumptions age fastest — especially user intent and edge cases that “used to be rare.”

Models keep working, but the world around them shifts. If you don’t revisit prompts, constraints, and success criteria regularly, reliability degrades quietly like you said.

How you think about detecting that decay early, before users feel it?