Software development
fromInfoWorld
17 hours agoHow agile practices ensure quality in GenAI-assisted development
Generative AI enhances coding speed but increases technical debt without Agile practices like pair programming and automated tests.
The most dangerous assumption in quality engineering right now is that you can validate an autonomous testing agent the same way you validated a deterministic application. When your systems can reason, adapt, and make decisions on their own, that linear validation model collapses.
This extends to the software development community, which is seeing a near-ubiquitous presence of AI-coding assistants as teams face pressures to generate more output in less time. While the huge spike in efficiencies greatly helps them, these teams too often fail to incorporate adequate safety controls and practices into AI deployments. The resulting risks leave their organizations exposed, and developers will struggle to backtrack in tracing and identifying where - and how - a security gap occurred.
"I've never felt this much behind as a programmer. The profession is being dramatically refactored as the bits contributed by the programmer are increasingly sparse and between. I have a sense that I could be 10X more powerful if I just properly string together what has become available over the last ~year and a failure to claim the boost feels decidedly like skill issue."
AI is no longer a research experiment or a novelty in the IDE: it is part of the software delivery pipeline. Teams are learning that integrating AI into production is less about model performance and more about architecture, process, and accountability. In this article series, we examine what happens after the proof of concept and how AI changes the way we build, test, and operate systems.
To find the typical example, just observe an average stand-up meeting. The ones who talk more get all the attention. In her article, software engineer Priyanka Jain tells the story of two colleagues assigned the same task. One posted updates, asked questions, and collaborated loudly. The other stayed silent and shipped clean code. Both delivered. Yet only one was praised as a "great team player."
A secure software development life cycle means baking security into plan, design, build, test, and maintenance, rather than sprinkling it on at the end, Sara Martinez said in her talk Ensuring Software Security at Online TestConf. Testers aren't bug finders but early defenders, building security and quality in from the first sprint. Culture first, automation second, continuous testing and monitoring all the way; that's how you make security a habit instead of a fire drill, she argued.
Olimpiu Pop: Hello everybody. I'm Olimpiu Pop, an InfoQ editor, and I have in front of me Erica Pisani, one of the track hosts of QCon London 2025, and a very important track in my opinion. One that is important in general, but even more important these days. And the name of the track was performance and sustainability, which seems to be two opposing words. So, Erica, please introduce yourself.
The recently updated SWEBOK Guide v4.0a represents a needful industry standard, following a thorough peer review and a consensus-based approach. With the rise of AI, a significant skills gap in IT and cybersecurity is emerging alongside changes in the global workforce. There has never been a greater need for a consensus-based framework. This guide, created and thoroughly reviewed by industry professionals, serves as a dynamic and evolving resource.
For years, reliability discussions have focused on uptime and whether a service met its internal SLO. However, as systems become more distributed, reliant on complex internet stacks, and integrated with AI, this binary perspective is no longer sufficient. Reliability now encompasses digital experience, speed, and business impact. For the second year in a row, The SRE Report highlights this shift.
Integrating databases into the CI/CD process or the DevOps pipeline is overlooked in the current DevOps landscape. Most organizations have adapted automated DevOps pipelines to handle application code, deployments, testing, and infrastructure configurations. However, database development and administration are left out of the DevOps process and handled separately. This can lead to unforeseen bugs, production issues, and delays in the software development life cycle.
Hast mentioned that they trust their unit tests and integration tests individually, and all of them together as a whole. They have no end-to-end tests: We achieved this by using good separation of concerns, modularity, abstraction, low coupling, and high cohesion. These mechanisms go hand in hand with TDD and pair programming. The result is a better domain-driven design with high code quality. Previously, they had more HTTP application integration tests that tested the whole app, but they have moved away from this (or just have some happy cases) to more focused tests that have shorter feedback loops, Hast mentioned.