- AutoGPT Newsletter
- Posts
- ChatGPT Agent = Big Flop?
ChatGPT Agent = Big Flop?

Hey, Joey here.
I know everyone’s hyping GTP5 right now…
But I prefer to let the dust settle before writing about it so I’m not going to cover it this week.
That being said, I’ve been playing with ChatGPT Agent since launch, and let’s just say, we’re not there yet.
This week’s deep dive is all about that.
Plus a few solid picks:
📌 Video: I Built Real Upwork AI Projects For FREE
📌 Resource: Online courses are taking a bit hit, AI is to blame?
📌 Deep Dive: ChatGPT Agent sounded like a game-changer… but kinda flopped
Let’s get into it.
POWERED BY
OxyCon® 2025 Is Here!

Spent 3 weeks building the "perfect" web scraper that died after 4 minutes because of an anti-bot system.
OxyCon (the #1 web-scraping conference) just dropped a 2025 lineup that looks like the antidote:
↳ Fred de Villamil (CTO @ NielsenIQ) on scraping 10 BILLION products a day and scaling e-commerce data flows
↳ Zia Ahmad (Google/Turing) on the AI-scraper feedback loop—models that train themselves on the data they collect
↳ Oxylabs AI Studio live demo: build a price-comparison tool with nothing but natural-language prompts
↳ A no-fluff panel on Web Scraping & AI: Legal Touchpoints—because compliance questions never quit
I’m clearing 1 Oct 2025 (free, online) so I can stop patching bots every sprint while other teams are already living in 2030.
2,250+ people went last year. This year's agenda looks insane.
Who else is ready to stop playing whack-a-mole with anti-bot systems?

DEEP DIVE
ChatGPT Agent = Big Flop??
ChatGPT Agent has been out for a couple of weeks now.
And this was supposed to be a glimpse of AGI?
Well I have some thoughts then…
I actually planned to write a deep dive on “10 use cases for ChatGPT Agent.”
But honestly, I couldn’t even come up with more than 3. That says a lot.
The launch felt promising. Like a real step forward. A ChatGPT that actually does things, not just talks about them.
I mean the demo makes it look like something that will make a big difference. It gave me Manus AI vibes.
So I went ahead and gave it a real task.
Nothing crazy. I needed transparent PNG logos for a list of AI tools I was featuring in a YouTube video. The kind of thing that takes me 5–7 minutes usually, just enough to be annoying.
I gave ChatGPT Agent the list and asked it to find the transparent PNGs version.
Finding these logos is a bit of tedious work, you Google “Canva logo transparent png” and then pray.
And to its credit… it worked. Slowly. But it worked.
It took 17 minutes… And if I needed 100 logos, this might’ve saved me some time.
But it didn’t feel like a breakthrough.
There was no “wow” moment. Just a shrug. And based on how quickly the hype vanished after launch, I don’t think I was alone.
So what’s missing here?
There’s a tweet I really liked recently from Balaji:

That’s exactly my feeling and ChatGPT Agent is fitting in that category.
The demo promised something magical. A full loop. You say something, it gets done.
But what you actually get is:
A slightly smarter way to start something.
And a slightly messier way to end it.
We’re still stuck in the in-between. And that makes the experience feel… mid.

Now, prompting has become a lot less important over the past 6 months. Models are becoming much more efficient and if you have a lot of convo with your LLM, that means context, that means less prompting engineering needed.
It’s less of:
“You are a helpful assistant, here is all the context, now please…”
and more: “Do this…”
That part’s moving in the right direction.
But verification is still a real “bottleneck”.
To give you an example, I’ve been building with tools like Lovable, Replit, Cursor, and the biggest pain point is always the same:
You make a change, it half-works, then breaks something else.
You fix that, and something new breaks.
Even when you explicitly ask the AI to check for errors, to test its output, to validate before continuing, it skips, it assumes, it just kind of wings it and to be honest, gets lazy.
(Maybe to save on compute and inference cost?)
So you end up doing a lot of the work anyway.
Which is fine if you’re technical. But it defeats the whole purpose if you’re not.
Where this might be going
I do think a lot of this will change once AI agents are fully integrated into your browser.
When your workspace becomes the context window.
When you don’t have to explain what you’re doing, because the agent can already see it, that’s when prompt engineering really dies.
That’s when the “just do this” instruction actually lands.
And it also means more reliable results, because your agent will know more than what’s in the current message. It’ll know your project. Your tools. Your patterns.
In the meantime, we’re left in the middle.
And for now, you still need a “specialist/expert” that know what they’re doing and can check on the AI output.
Not quite automation.
Not quite delegation.
Just an eager assistant you still have to babysit.
As the article from Balaji mentions:
0% AI is slow
100% AI is slop.
Want to see more content like this? |

THAT‘S A WRAP
Before you go: Here’s how I can help
1) Sponsor Us — Reach 250,000+ AI enthusiasts, developers, and entrepreneurs monthly. Let’s collaborate →
2) The AI Content Machine Challenge — Join our 28-Day Generative AI Mastery Course. Master ChatGPT, Midjourney, and more with 5-minute daily challenges. Start creating 10x faster →
See you next week,
— Joey Mazars, Online Education & AI Expert 🥐
PS: Forward this to a friend who’s curious about AI. They’ll thank you (and so will I).
What'd you think of today's email? |
Reply