Artificial intelligence
fromThe Verge
3 hours agoCanonical lays out a plan for AI in Ubuntu Linux
Ubuntu plans to integrate AI features to enhance OS functionality and provide new workflows over the next year.
The new feature via the Actual Result field enables you to record precise outcomes for each test step, improving traceability, audit readiness, and collaboration across your teams.
Dependabot sounded the alarm on a large scale. Thousands of repositories automatically received pull requests and warnings, including a high vulnerability score and signals about possible compatibility issues. According to Valsorda, this shows that the tool mainly checks whether a dependency is present, without analyzing whether the vulnerable code is actually accessible within a project.
JDK 26, the first non-LTS release since JDK 25, has reached its second release candidate as declared by Mark Reinhold, Chief Architect, Java Platform Group at Oracle. The main-line source repository, forked to the JDK stabilization repository in early-December 2025 (Rampdown Phase One), defines the feature set for JDK 26. Critical bugs, such as regressions or serious functionality issues, may be addressed, but must be approved via the Fix-Request process. As per the release schedule, JDK 26 will be formally released on March 17, 2026.
"I've never felt this much behind as a programmer. The profession is being dramatically refactored as the bits contributed by the programmer are increasingly sparse and between. I have a sense that I could be 10X more powerful if I just properly string together what has become available over the last ~year and a failure to claim the boost feels decidedly like skill issue."
The software industry is collectively hallucinating a familiar fantasy. We visited versions of it in the 2000s with offshoring and again in the 2010s with microservices. Each time, the dream was identical: a silver bullet for developer productivity, a lever managers can pull to make delivery faster, cheaper, and better. Today, that lever is generative AI, and the pitch is seductively simple: If shipping is bottlenecked by writing code, and large language models can write code instantly, then using an LLM means velocity should explode.
The real cost of poor observability isn't just downtime; it's lost trust, wasted engineering hours, and the strain of constant firefighting. But most teams are still working across fragmented monitoring tools, juggling endless alerts, dashboards, and escalation systems that barely talk to one another, which acts like chaos disguised as control. The result is alert storms without context, slow incident response times, and engineers burned out from reacting instead of improving.
AI made producing software cheap, but understanding it is still expensive. The Manifesto optimizes for the former. This addendum shifts the emphasis toward the latter. Four updated values, three refined principles, with reasoning for each.
One thing I always do when I prompt a coding agent is to tell it to ask me any questions that it might have about what I've asked it to do. (I need to add this to my default system prompt...) And, holy mackerel, if it doesn't ask good questions. It almost always asks me things that I should have thought of myself.
We build production platforms with AI every day, and we work with teams doing the same with their own stack -Cursor, Claude Code, Copilot. The difference shows up fast. By day two, some codebases are already harder to change than they were yesterday. Others keep getting easier. The difference is never the model. It's what the code lands in. The teams we work with that hit a wall? It's always the same story.