When AWS CEO Matt Garman told developers that โit's possible that most developers are not codingโ by 2026, it raised the question of how different AI-centric coding environments would look.
When Amazon Web Services CEO Matt Garman told his developers in June that they may no longer be in the coding business in as little as two years, it wasย a heads up to change-resistant coders that they had better embrace the new AI-centric reality.
But it also forced serious conversations among executives about what generative artificial intelligence (GenAI) coding environments would realistically look like.ย
Garmanโs recorded comments, which have since been confirmed by Amazon, were first reported by Business Insider.ย
โIf you go forward 24 months from now, or some amount of time โ I canโt exactly predict where it is โ itโs possible that most developers are not coding,โ Garman was quoted as saying. โIt just means that each of us has to get more in tune with what our customers need and what the actual end thing is that weโre going to try to go build, because thatโs going to be more and more of what the work is, as opposed to sitting down and actually writing code.โ
A different world
But managing code generated by human programmers will be vastly different than managing GenAI-generated code.ย Dev Nag, CEO of SaaS firm QueryPal, said that he has been working with GenAI coding efforts and that many enterprise IT executives are not prepared for how different it is.
โIt made tons of weird mistakes, like an alien from another planet. The code misbehaves in a way that human developers donโt do,โ Nag said. โItโs like an alien intelligence that does not think like we do, and it goes in weird directions.โ
Of even greater concern, Nag said, is that GenAI coding mechanisms will follow the rules to a certain degree, but it will also get creative in finding ways to circumvent those rules. โAI will find a pathological way to game the system.โ
Humans are still essential
Another programming specialist, Veny Moola, CEO of application development firm Fleet Studio, said that GenAI hallucinations are major problems.
โHallucination-laden software shouldnโt be given direct control over executables, and thatโs precisely why the role of software engineers remains critical. When AI systems, prone to hallucinations, are involved in generating code or making decisions, it is essential for human engineers to validate, test, and monitor these outputs before they are allowed to interact with or control executable processes,โ Moola said.
โEngineers are responsible for implementing rigorous checks and balances, such as automated testing, code reviews, and safety protocols, to catch and correct potential errors introduced by AI. In high-stakes environments where software has direct control over critical systems, human oversight is crucial to ensure that the AIโs outputs are reliable and safe. Without these safeguards, the risk of deploying AI-generated software without proper vetting could lead to catastrophic failures, which is why software engineers must remain actively involved in the development and deployment process.โ
GenAI-centric programming also lacks the kind of shared coding context that is in the heads of just about all human developers. โIt needs far more overhead because it is missing the shared organizational context. We are not used to having to mention everything,โ said Nag, adding that he estimates that context-related issues โwill be magnified 100 times with GenAI coding.โ
New testing requirements
A critical part of coding oversight will involve testing, but Nag stressed that enterprises will have to create entirely new approaches to testing. Today, most enterprise programming teams are trusted to create homegrown apps. The apps will certainly be tested for functionality, but many today forego a rigorous line-by-line examination looking for backdoors or other problematic elements.ย
In a GenAI coding environment, โtesting processes will have to change vastly, changing from simply functionality testing to far, far deeper automated testing of everything, including having every app subjected to automated penetration testing,โ Nag said.
Generative AI limitations
Another change that enterprise development managers will have to accept is that moving from small coding projects where AI is acting as a programming assistant to giving GenAI far more coding authority is a massive leap.
โAlthough AI is able to solve many college problem sets and handle small-to-medium snippets of code generation, it still struggles with complex logic, large code bases, and especially novel problems without precedent in the training data. Hallucinations and errors remain significant issues that require expert engineering oversight and correction,โ Nag said. โThese tools are far better at quick prototypes from scratch rather than iterating large applications, which is the bulk of engineering. Much of the context that drives large applications doesnโt actually exist in the code base at all.โ
Tom Taulli, who has authored multiple AI programming books, including this yearโs AI-Assisted Programming: Better Planning, Coding, Testing, and Deployment, agreed that the move to great GenAI coding efforts will catch most enterprises off guard.
โThese tools will mean a change in traditional workflows, approaches, and mindset. Consider that they are pretrained, so they are often not updated for the latest frameworks and libraries. Another issue is the context window. Code bases can be massive. But even the most sophisticated LLMs cannot handle the huge amount of code files in the prompts,โ Taulli said. โMany of these tools are mostly trained on public code repositories. But there is lots of legacy code that is incredibly important and quite sparse for the training datasets, like COBOL.โ
In his own GenAI coding efforts, Taulli said he has seen a lot of unorthodox coding behaviors. He also pointed to published reports of many different GenAI-coding challenges.ย
โFor example, you can ask these LLMs to create code and they sometimes make up a framework, or an imaginary library or module, to do what you want it to do,โ Taulli said, explaining that the LLMs were not creating a new framework as much as pretending that they created it, but they actually didnโt.
That is the sort of behavior that LLM coding will do that almost no human programmer would even consider doing, Taulli noted, adding, โunless (the human coder) is insane, they are not going to make up, create out of thin air, an imaginary library or module.โ
When that happens, it can be easy to detect, if someone looks for it. โIf I try to pip install it, you can find that thereโs nothing there. If it hallucinates, the IDE and compiler give you an error,โ Taulli said.
The coderโs new role
Nag said that he thinks Garmanโs comments are correct in that enterprise programming efforts are going to need to make the AI move, even if it causes a lot of painful changes along the way. Eventually, Nag said, enterprise development efforts will be better for the change.
โ[The coderโs] role is as much about understanding the larger context and business goals as it is about crafting a particular syntax or debugging an error code. The rise of digital photography didnโt destroy the photography profession. It democratized it and created explosive demand for new specialties like mobile filters and retouching,โ Nag said. โThe rise of spreadsheets didnโt make accountants obsolete. It gave them the leverage to focus on financial analysis, strategy, and more complex corner cases. AI in software engineering will likewise democratize application development while creating new roles around prompt engineering, AI-assisted development, and expanding the scope of whatโs possible with software.โ
Security concerns
Another coding specialist, Ashley Rose, CEO at SaaS vendor Living Security, said there are also security concerns if more coding influence gets turned over to GenAI.
โThink about how our internet infrastructure has evolved. It was originally designed to be decentralized, but over time, key services have become concentrated around major providers like Cloudflare and AWS. This centralization has introduced new points of vulnerability, where an outage or attack on one provider can ripple out to impact large portions of the web,โ Rose said.
โThe same could happen with AI in coding. As more companies integrate AI models into their development processes, these models could become high-value targets for attackers. Compromised models could introduce hidden vulnerabilities that slip through standard security checks. Given the widespread use of these models, the impact could be devastating, potentially affecting thousands of applications across different organizations.โ


