AI: Weekly Summary (October 13-19, 2025)

Key trends, opinions and insights from personal blogs

I keep bumping into the same little knot in the week’s pile of AI posts. It’s this tug-of-war between a giddy, let‑the‑bots‑loose optimism and a quiet, stubborn caution that keeps popping up in different clothes. Some writers are shouting that agents, chips, and models will change everything. Others are saying—more softly, but with the same volume—that we still need plans, checks, and common sense. I would describe them as two playlists playing at once: one is techno on full blast, the other is an old acoustic record trying to be heard.

The agent moment: promise, friction, and a long list of caveats

There’s been an avalanche of takes about agents this week. A lot of posts look at agents as if they’re about to move into your house and do the chores. Dave Friedman writes that agents won’t save you. He points out something simple but easy to forget: the internet is not a neutral place. It’s a set of walled gardens. Agents need open plumbing to run. And the big platforms are basically saying, “nope.” To me, it feels like trying to teach a dog to mailbox‑run when all the doors are locked.

Then there’s the design and engineering side. Folks like Kaushik Gopal and Bart Wullems are doing the homework. ExecPlans, AGENTS.md and other process ideas pop up as a reaction to the chaos. They say: if you want reliable agents, give them a map before setting them loose. It’s not glamorous. It’s not a shiny demo. It’s planning. Peter Steinberger tells you to “just talk to it” — in practice that’s a kind of shorthand for interacting with models as tools, but he also stresses the need for oversight and context management.

And then there is trust. Pawel Brodzinski argues that the world is not ready to trust autonomous agent purchases. He lists transparency, alignment, and care as hard requirements. Which is a neat way to say: before handing your credit card to a bot, make sure it knows the difference between lunch and laundering.

If you’ve been swept up in the agent hype, read the skeptics and the planners back to back. They balance each other. Like tea and lemon. Don’t laugh—sometimes the right mix saves the day.

Developers: vibes, specs, and the slow rot of “vibe coding”

There’s a distinct thread about how engineers actually work with these tools. A few posts hit the same nerve. Onur Solmaz talks about redesigning his blog with AI help and jokes that default themes now scream ‘I gave up.’ That’s sideways commentary, but it pairs with more serious posts about coding practices.

Warning signs show up in the phrase people use: “vibe coding.” You’ll find it called out in English and Italian. The gist is: developers letting LLMs write code without understanding it is fast but risky. Lorin Hochstein wrote about the efficiency‑thoroughness trade-off — ETTO. Speed buys you output. It steals you careful review. After an incident, telling someone to “be careful” doesn’t fix the physics of that tradeoff.

Practical threads try to fix this. Spec-driven development is getting airtime. JP Posma and others explain how specs and AGENTS.md can guide agents. It’s like telling a sous chef the ingredients and the plating before asking them to cook. It changes the output. It adds safety. Tools like Haiku 4.5 and local sandboxes show we want both speed and guardrails. There’s a smell of pragmatism here: people want agents, but not at the cost of chaos.

A small, slightly annoying repeat appears across posts: more speed, more bugs, repeat. Developers are tired of running after their own shortcuts.

Models, chips, and the infrastructure arms race

If you’d read only press releases, you’d think the story is chips and data centers. That’s partially true. OpenAI partnering with Broadcom for 10‑gigawatt accelerators, Broadcom launching the Thor Ultra 800G NIC, and Apple’s new M5 chip—these are not mere specs pages. They set the stage for where compute goes and who pays for it. See notes from Brian Fagioli and Nick Heer for the industry playbook.

There’s also a parallel conversation about who builds the compute and where. JP Posma and others point to models like GLM‑4.6 and Qwen3 climbing fast in consumption. China’s stack without Nvidia chips is a quiet revolution. To me, it feels like a local bakery suddenly baking better loaves and selling them cheaper across town.

But the infrastructure story includes a darker B side: energy, water, and politics. Mike “Mish” Shedlock and others chronicle data centers building their own power plants. That’s not a metaphor. It’s literal. xAI building a wastewater plant in Memphis to cool servers is another of these hands‑on solutions (Stephen Hackett). And local communities are pushing back on bills and land use. Naked Capitalism and others describe a bipartisan rage about rising electricity costs. You can argue about whether the data‑center water story is a panic or not—Simon Willison thinks the water part is overblown—but the scene feels like building a house with no plumbers and hoping everything will drain fine.

There’s a lot of money in the middle of it. Posts about capex, circular investment, and Minsky moments suggest a market where financing and hardware deals can ripple faster than the technology itself. That’s worth reading if you like the smell of balance sheets more than model benchmarks.

Jobs, productivity, and the uneasy human bargain

You can’t avoid the labor question. Some pieces tilt optimistic. Richard Hanania claims a 33% bump in personal productivity thanks to AI. Fine—nice headline. But then you have the macro surveys and reports. Brian Fagioli shares data predicting which jobs will be hit first; [United Airlines] numbers discuss headcount reductions; Michael Spencer and others point out that despite the hype, generative AI hasn’t sparked a broad jobs boom.

There’s a recurring tension: AI can make many tasks easier, but often organizations use that as cover to compress labor costs. Daron Acemoglu’s voice appears across summaries: the real harm isn’t shiny AGI, it’s mediocre automation that sheds jobs without improving productivity. It’s the self‑checkout we all know: fewer cashiers, same store sales, more customer hassle.

At the same time, people are trying to figure out how to fit AI into careers. Nate is all over prompts for career planning, memory systems, and AI fluency. He suggests that AI fluency—knowing how to get useful things out of models—is becoming a star skill. Not just using the toy. Fluency. I’d say that’s similar to learning to read a map before you drive in a strange city.

Creativity, taste, and a culture weary of fakery

This week felt like the culture editors kept tapping their watches. A bunch of posts argue that people are tired of low‑grade, engagement‑first content. Tracy Durnell talks about “slop” as a symptom of capitalism; Adam Singer defends people who resist AI for taste’s sake; The Trichordist (on Universal Music) draws a firm line around artists’ voices.

There’s an emotional current here. Americans, apparently, are sick of fake AI content and crave something real (Brian Fagioli). That’s not a niche preference. It’s a feeling you see in restaurant lines and in how people talk about live music. When everything can be stitched together by an algorithm, people start to notice the seams. And they don’t like the look.

The pushback takes many forms. Legal fights around AI‑generated citations in court filings, musicians insisting on licensing for voice models, and radiologists pointing out that AI works in labs but struggles in messy hospitals (see A Learning a Day). All of it reads like a messy family reunion where some cousins insist the grill is broken and others argue they have the perfect marinade.

Safety, surveillance, and politics—where the slow burn lives

There were a lot of posts about politics and safety. Bruce Schneier teases chapters on AI and democracy, and others highlight real moves in surveillance and policy. The Trump administration’s increased use of social media surveillance and visa revocations tied to public speech is one example. There’s also a growing literature on how political campaigns are using AI to automate persuasion (Caroline Orr Bueno).

Artificial agents making purchases (Agentic Commerce Protocol) and the governance around advanced chips (GAIN Act) come up as real regulatory tensions. There’s a sense that policy and technology are sprinting on different tracks. The posts suggest we’ll be ironing out trust issues for years. It’s a boring fight, but the kind that changes how everything else works: customers, voters, and banks.

Memory, grounding, and the nitty‑gritty of intelligence

Some of the week’s best reading is quietly nerdy: memory systems, RAG/REFRAG tricks, episodic memory papers, and representation engineering. These aren’t sexy headlines. They are the parts inside the clock that make it tick.

A couple of posts stand out. Grigory Sapunov wrote about latent learning and episodic memory. The paper argues that models know more than they can flexibly reuse. That resonates with Anthropic’s new memory tool discussed by Shlok Khemani and with Nate who offers eight principles for memory fixes. These pieces, together, feel like plumbing work: boring while necessary, and making later features actually useful.

REFRAG from Meta gets a shout too. It’s a pragmatic rethink of the RAG pipeline. Instead of token‑expanding everything, expand only what’s needed. Speed up time‑to‑first‑token. Save cost. It’s the kind of optimization that makes an app usable on a budget.

If you like papers and slow progress, these posts are the ones to bookmark. They hint at what will make agents actually helpful rather than theatrical.

Edge, local models, and the return to privacy-ish workflows

A quieter movement is to run smaller models on devices. Local Browser AI, nanochat, Llama 3.1 locally, and EdgeXpert boxes from MSI are all attempts to get useful models close to users. Alex Ewerlöf and Simon Willison show how the Prompt API and small model tooling can move inference out of the cloud.

This matters to people who don’t want their worst email drafts feeding corporate corpuses. It’s like preferring to bake your bread at home rather than buying from a franchised bakery. You lose some scale. You gain a measure of privacy and control.

The recurring metaphors and the tone of the week

Several writers returned to similar metaphors. AI models as djinn or genies (Katherine Dee), agents as household helpers that need instructions, chips and data centers as power plants and plumbing. These metaphors matter because they anchor a complex debate in daily life. They make tech feel like dinner, not a research abstract.

I’d say the week’s mood is impatient, but in two different directions. Some people are impatient for more agentic automation. Others are impatient for the sober work: specs, tests, energy planning, legal rules, and memory systems. Both impatiences are valid. Both are annoying to the other side.

Little asides, small repeats, and why you should click through

You’ll see the same ideas reappear: trust, planning, and the limits of raw speed. The finance posts keep circling the same worry: lots of capex and a few big bets that could unsettle markets if they wobble. The infrastructure posts repeat a related worry: if data centres are going to suck down electricity, someone needs to pay for the pipes. The culture posts repeat the same note of fatigue: people want real, not slop.

If any of those lines tug at you, go read. Authors like Anup Jadhav and Nate give you practical playbooks for builders. Bert Hubert and Nick Heer give macro‑level skepticism that keeps the hype honest. The PyCoach and Grigory Sapunov will take you deep into research tweaks that actually stretch model capabilities.

Read the planners when you want to build something that doesn’t fall apart on day two. Read the critics when the headlines make you dizzy. Read the infrastructure posts if you like maps of how money and power will shape computing. Read the cultural pieces if you want to remember why humans still matter in storytelling and art.

A quick list of things that kept popping up

  • Agents: hype plus real engineering problems. (Friedman, Gopal, Wullems, Steinberger)
  • Dev workflow friction: vibe‑coding, spec‑driven development, AGENTS.md, ExecPlans. (Posma, Hochstein, Kaushik)
  • Compute and chips: Broadcom, OpenAI deals, Apple M5, Nvidia moves. (Fagioli, Heer, Tsai)
  • Energy and water: data centers building power plants and recycling water. (Shedlock, Hackett, Naked Capitalism)
  • Jobs and fluency: productivity claims vs structural labor impacts. (Hanania, Fagioli, Nate)
  • Memory + grounding: episodic memory, REFRAG, Anthropic memory. (Sapunov, Jadhav, Khemani)
  • Culture and trust: backlash to fake content and the defense of taste. (Durnell, Singer, The Trichordist)
  • Security and privacy: breaches, shadow AI sharing, phishing. (TCP, Mike McBride, Martin Brinkmann)

Yes, it is a lot. And the posts are not in agreement. Which is fine. They’re trying to map the territory at different scales.

There are more small gems in the stack: local LLM experiments, RAG tweaks that save money, design notes about making blogging feel nicer with AI help, reflections about learning starting at the end, and a parade of practical how‑tos. You’ll find recipes, code guides, and playbooks mixed with big‑picture essays on markets and society. It’s a weird market basket: the revelation and the instruction manual.

If I had to pick a mood for the next few months, it would be this: expect more tooling, expect more friction, and expect more regulation‑by‑default. Investors will keep placing big bets on compute. Engineers will keep inventing ways to make agents behave. Creators will keep asking for consent. And the rest of us will keep pretending we understand how the oven works.

If you want the sleepier technical reads, follow Grigory Sapunov, Anup Jadhav and Nate. If you like the policy and finance angle, the pieces by Nick Heer, Paul Kedrosky and Naked Capitalism are heavy hitters. And if you just want something readable and human, check out the cultural notes from Tracy Durnell and the practical notes from The PyCoach.

There’s a lot to chew on. Like a Sunday roast you didn’t mean to start, but now the whole neighborhood is sniffing smoke. Some pieces are smoky and overcooked, some are rare and worth savoring. Read a few. Argue with them. Then come back and read some more.

If you want to follow up on anything specific, I’d point you toward the planning‑and‑agents posts first. They feel like the place where the rubber will meet the road. The others add color and consequence. But, hey—pick what annoys you most. That’s often the signal worth paying attention to.