For decades in SAAS, products reduced ambiguity. Users supplied constrained inputs, and the system handled the output. It's never been Minority Report cinematic, but it was predictable. By providing predictable environments for manipulating data, users learned by moving things, adjusting variables - and the outcome emerged through interaction.
Instructions I created. Instructions I am continuing to hone - instructions that required me to study my own old essays, identifying what I do when I write. The sentence rhythms. The way I move between timescales. The zooming in and out from concept to detail. The instructions tell Claude how I would like ideas composed. I pull together concepts and experiences from my lived expertise to formulate a point of view - in this case, on this new AI technology.
Today we are at the cusp of revolutions in artificial intelligence, autonomous vehicles, renewable energy, and biotechnology. Each brings extraordinary promise, but each introduces more complexity, more interdependence, and more latent pathways to failure. This elevates prudence to be critical. Good design recognizes what cannot be foreseen. It acknowledges the limits of prediction and control. It builds not merely for performance, but for recovery.
The normative form for interacting with what we think of as "AI" is something like this: there's a chat you type a question you wait for a few seconds you start seeing an answer. you start reading it you read or scan some more tens of seconds longer, while the rest of the response appears you maybe study the response in more detail you respond the loop continues
When Sonos released its redesigned app in May 2024, the backlash was immediate and brutal. Users couldn't access basic features like volume control and alarms. Systems became unusable. The company's stock plummeted 25%. Eventually, the CEO was replaced, and lawsuits claimed over $5 million in damages from customers who'd lost functionality they'd paid for.
The question dropped into the Slack channel before the user research summary. Before the problem was clearly defined. Before anyone asked if users actually needed this feature. Your product manager already generated three interface options in ChatGPT. Now they're asking which one to build. Not whether to build. Not why to build. Which. And when you slow the conversation down to ask those questions, you're about to discover that strategic thinking now reads as bottleneck behavior.
In Andor, I got chills when Mon Mothma warns the senate of a chilling truth: When we let noise, conformity, or fear dominate, we lose sight of what matters. We risk allowing the loudest voices, often the safest, the most predictable, to drown out individuality, identity, and truth. To me, this line... This line echoes a growing tension I feel in content design.