“It can take a PR that you would give to any other engineer and go build it. But not only does it build it, it does all the things that software engineers hate to do. It does its own QA, its own quality assurance, its own bug testing and bug bashing, and it does documentation, things you can never get software engineers to do so, suddenly, you can force, multiply your software engineering workforce.” Friar added.
OpenAI launched its first AI agent – Operator – in January, soon followed by Deep Research in February – both AI offerings are currently only available to ChatGPT’s paying customers.
Why you shouldn’t freak out right now?
OpenAI has a history of making tall claims about its products some of them never materialize. Take Deep Research for example, at the time of its launch, OpenAI said that the new tool will be able to replace a research assistant – a claim that Friar repeated on Friday.
While many of OpenAI’s peers, including xAI and Perplexity, have rolled out similar tools, it’s still unclear how much of a research assistant’s role these AI models can genuinely take over. The reason? They’re still prone to hallucinations—confidently generating information that simply isn’t true.
The problem isn’t that these large language models are inaccurate – after all, humans make mistakes too. What’s more concerning is how these models present false information with an air of absolute confidence, making it harder to separate fact from fabrication. And that hasn’t changed much since ChatGPT first rolled out to the public in late 2022.
So when OpenAI says that its upcoming AI agents will be able to essentially do all of what the current software engineers do and more, take those claims with a pinch of salt.
openai, chatgpt, ai, ai agent, chatgpt openai, ai chatgpt, chatgpt free
#OpenAIs #agent #selftesting #software #engineer #humans #wont