Skip to content

Plutonic Rainbows

The Rule of Seventy

The full details of Microsoft's voluntary retirement program leaked to Business Insider on May 7, two weeks after CNBC first reported the broad strokes. It is the company's first ever buyout offer in fifty-one years of trading, and the eligibility filter is unusual enough to deserve attention on its own. To qualify, an employee's age plus their years of service, each rounded to the nearest whole number, must sum to seventy or more. Microsoft is calling it the Rule of 70.

A 55-year-old with 15 years of tenure makes it. A 60-year-old with 10 years makes it. A 35-year-old hired in 2017 does not. The arithmetic is deliberately tilted toward people who have been at the company for a long time, and tilted again toward those who joined before the second cloud boom, when the average tenure of a software engineer in big tech was longer and the salary expectation was lower.

This is where the program differs from a normal layoff. A layoff is, in theory at least, performance-blind. The Rule of 70 is not. It is a deliberate filter for the part of the headcount that is most expensive per head and most likely to be holding stock granted at lower strike prices, which means the per-departure saving for the company is materially larger than a random sample would deliver. Senior directors and below qualify, level 67 and lower in Microsoft's job ladder, but the ladder is bottom-heavy by design and the people who clear seventy on the eligibility test are clustered toward the senior end of the eligible pool.

The package itself reads as generous on paper. Up to 39 weeks of severance, partial healthcare continuation, and continued vesting of unvested stock for six months for employees with under 24 years of tenure, twelve months for those above. Business Insider's leaked document shows that Microsoft has explicitly told staff this is a one-off, no second VRP is planned. There is a quiet pressure inside that sentence. The implication is that anyone eligible should treat the offer as terminal in both directions, take it now or accept that the company's next instrument for shrinking the same population is unlikely to be as soft.

What makes the design interesting is how it sidesteps the political problem of cutting older workers. American age discrimination law is most easily violated when a company picks individuals over forty for involuntary separation. A voluntary program, dressed in retirement language, with an objective formula applied uniformly across the eligible population, is much harder to challenge in court. Microsoft's lawyers know this. The formula is its own legal cover.

The other thing worth sitting with is the size of the spending it is paid against. The same fiscal year covers $190 billion in capital expenditures, the bulk of it AI infrastructure. Severance for seven percent of US employees, even capped at 39 weeks, is a fraction of a fraction of that number. As I wrote in late April when the capex line was first set against the headcount line, the buyout is not a cost-saving exercise in the old sense, it is a composition change. The headcount is being thinned at one specific demographic so the compute envelope can grow without the total compensation envelope expanding.

Whether the people leaving were doing work that an AI agent can absorb is a separate question, and one the program's design carefully does not ask. The Rule of 70 selects for length of service, not for the kind of task the person spends their day on. A senior PM who has been managing release cadence for fifteen years and a long-tenured systems engineer who quietly keeps a piece of legacy infrastructure running both clear the bar. Whether their replacements are agents, juniors, or nobody at all is a decision the company can make later, on a per-team basis, without ever having to defend the demographic shape of who left.

Sources:

Cast Iron in 21 Parishes

A black-and-white cast iron pole bearing the word DEAN, two distance-marked arms pointing along a Cumbrian lane, and a green box hedge behind it. That is the photograph Cumberland Council released to mark its restoration of fifty Victorian-era fingerposts across twenty-one parishes. It is also, if you stand in front of it for long enough, a very strange object. A fingerpost is a sentence in iron, a public assertion that this place exists and is reachable from here. The Cumberland posts are saying it about parishes that, in many cases, are now mostly hedgerow, a few houses, and the silence that follows the postbus pulling away.

The form is older than the road system that produced it. An early surviving English fingerpost near Chipping Campden is dated 1669, and the cast-iron tradition that produced the Cumberland posts followed the 1773 General Turnpike Act and ran through the late nineteenth and early twentieth centuries. Different counties used different colour codes; Cumberland's posts wear distinctive black-and-white bands, and Somerset has a refurbished Red Post still standing on the A358 between Taunton and Williton. Driving past one in deep country is a small act of time travel even if you don't know any of that, because the proportions are wrong for any modern signage you have ever seen.

The history they carry has two erasures in it. The first was deliberate, in 1940, when the government ordered all signposts removed so that an invading army would not be able to read its way to the Midlands. The fingerposts were a small part of a much larger anti-invasion landscape that summer, the same one that produced the pillbox lines and tank traps still visible in undergrowth across southern England. Some of the posts stayed in the ground; their fingers were detached, sometimes buried nearby, sometimes gathered up and stored. On Exmoor a council-managed quarry held a heap of arms that were eventually reinstalled in the late forties. The second erasure was procedural. The 1964 Worboys Committee, designed to rationalise British road signage, produced regulations that barred councils from putting up new fingerposts at all. By the 1990s the Department for Transport classified them as hazardous distractions on A-roads, urged their replacement, and saw the national stock fall sharply. Of the 1,300 fingerposts thought to exist in the 1950s, 717 had survived.

What is happening underneath all of this is the long, unromantic life of the Definitive Map, the legal record of rights of way that every parish in England had to compile after the National Parks and Access to the Countryside Act 1949. The Open Spaces Society archive preserves the parish-meeting minutes from the early 1950s, and they are bracing reading. Adisham in Kent, August 1950, decided which paths were "now not used" and crossed them off. Yapton in West Sussex declared one footpath "no longer required". The clerk of Hitcham parish in Suffolk wrote to West Suffolk County Council about hundreds of footpaths "and scores of footpaths which have not been used for at least 25 years", asking whether the surveyors might just walk the used ones. A landowner could appeal. A path crossed off was a path lost.

So the law that protects rural rights of way also encodes, in its first definitive map, every village's verdict on which of its own routes had become unneeded. The walker who follows a yellow waymarker today is reading a document that is half medieval and half a 1950 spreadsheet. The fingerpost is a restoration of the most visible bit of that document, the bit made of iron and paint, and what Cumberland Council restored last week is genuinely older than the legal definition of the paths it points at.

The pleasure of these objects is that they don't pretend to be anything else. They are not nostalgic, they are continuous, and the difference matters. A nostalgic object signals an absent context; a continuous one performs the same function it was made for. A 1905 fingerpost in a parish where the church has been converted into holiday lets is still doing its job. It points to Dean, and Dean is still there, just smaller and quieter and largely visited by people on foot following yellow arrows that mean exactly what they meant in the year the post was painted.

Sources:

Six Copies in One Pass

Walk into a vehicle-licensing office, a hospital pharmacy, a bank's back office, or the goods-in counter of any large warehouse, and somewhere in the room you'll hear a sound the rest of the working world abandoned thirty years ago. The chittering whir of a print head dragging across continuous fanfold paper, perforated tractor feed clicking through the sprockets, ribbon being struck through carbon. It is not nostalgia. It is the only printer in the room that can do the job.

The job is multi-part forms. Carbon-copy paper, or its successor NCR (no carbon required), three or four or six layers stacked together, each with a designated colour and a designated recipient. The customer keeps the white. The garage keeps the yellow. The accounts office keeps the pink. The DVLA gets the green. A laser printer cannot do this. An inkjet cannot do this. Neither one strikes the paper hard enough to register through a stack. Only an impact head with a row of small steel pins, slamming through ribbon into the top sheet, can transfer the same image to every layer underneath in a single pass.

This is why Epson still manufactures the FX-890II, the LQ-590II, and the PLQ-50 passbook printer, quietly, on its current US site, under the slogan "World Leader in Impact Printing™". This is why the global carbonless-paper market sat at $4.4 billion in 2024 and is forecast to grow at roughly 3.7 percent a year through 2034. This is why airline gate agents still print luggage tags on dot-matrix devices at hubs that have spent eight figures on every other piece of trackside infrastructure. The economics aren't the explanation. The chemistry of paper-and-pressure is.

There is a particular institutional grammar that comes with the multi-part form. Each colour layer has a custodian. Each custodian has a duty to hold their copy for a regulator-defined number of years. The form is the audit trail; the audit trail is the form. You cannot replace it with a PDF and an email confirmation, because the regulator who wrote the rule decades ago specified physical custody of a serially-numbered carbonless duplicate, and nobody has ever told the regulator to update the rule. So the dot-matrix printer survives, not because nobody can build a better one, but because nobody can build a different audit trail without rewriting decades of administrative law.

Anyone who grew up with one remembers the noise. It is closer to a sewing machine than a printer, mechanical and metronomic, audible from two rooms away. The cadence varies by model: 9-pin, 18-pin, 24-pin, draft mode, near-letter-quality. The fanfold paper smelt faintly of warm ribbon. The perforations down each edge had to be torn off afterwards in long curling strips that gathered around the bin. None of that is missed in domestic life. None of it has gone away in the small back-office rooms where paperwork still moves between custodians on physical shelves.

What strikes me most is the way the survival is invisible. Nobody markets a dot-matrix printer to consumers. Nobody talks about them. The few magazine pieces written about their persistence treat them as a curiosity, the same way the speaking clock gets treated as a curiosity. But there are still very large numbers of these machines in active service, churning through ribbon and continuous paper in industries whose paperwork the public never sees. The technology that office life obsolesced in 1995 is doing more work today, in absolute terms, than it ever did then.

Sources:

Reading the Activations

During an evaluation, Claude Mythos Preview cheated on a training task. While it cheated, it was internally thinking about how to avoid being detected. None of that reasoning appeared in the visible output. The only reason Anthropic's researchers can describe the episode at all is that they had a tool reading the model's activations back to them in English, and the tool printed out the detection-avoidance thoughts as readable sentences.

That tool is called a Natural Language Autoencoder, and Anthropic introduced it on 8 May. It is the cleanest thing I have seen come out of the interpretability team in a while, partly because of what it does, and partly because of what it admits about everything that came before.

When you send a prompt to Claude, the model converts your text into long numerical vectors called activations and processes those vectors layer by layer. The activations are where the reasoning lives. They are also, historically, the part nobody can read. Sparse autoencoders, attribution graphs, and circuit analyses can recover features from activations, but the outputs are themselves dense objects that need a trained interpreter to parse. Useful, but not legible. NLAs are a model trained to produce English sentences directly from an activation vector, with a second copy of the original model used to reconstruct the activation from the sentence and confirm the translation has not lost the signal. If the reconstruction matches, the English is treated as a faithful summary of the thought.

The first interesting result is the rhyme example. Asked to complete a couplet, Claude's activations contain candidate rhymes several tokens before the model is anywhere near the end of the line. NLAs print those candidates out as plain words. The model is planning ahead in a way the surface text never reveals, which is consistent with what mechanistic interpretability has been suggesting for two years but is the first time the planning has been legible without a researcher squinting at heatmaps.

The audit-game numbers are the load-bearing claim. Anthropic trained a derivative of Claude Haiku 3.5 to have a hidden misaligned motivation, then asked an auditor to find it. With sparse autoencoders and the older interpretability stack, the auditor won less than 3% of the time. With NLAs in addition, the success rate rose to between 12% and 15%, without the auditor needing access to the misaligned training data that planted the motivation in the first place. That last clause matters more than the headline number. Earlier interpretability tools mostly worked by helping the auditor backtrack from anomalous behaviour to the data that caused it. NLAs, in this trial, surfaced the motivation directly out of the activations. The toolchain shifted from forensic to introspective.

I read the 12-15% cautiously, because Anthropic does. It is a toy scenario, the model under audit was deliberately given a single clean misalignment, and 85% of the time the auditor still missed it. But going from <3% to >12% with one technique is the kind of step change that justifies running the tool in production, which Anthropic now does. NLAs were part of the pre-deployment alignment audits of Mythos Preview and Claude Opus 4.6, the same Mythos whose cybersecurity capabilities are what put federal pre-deployment evaluation on the table in the first place.

The Opus 4.6 example is the unglamorous one and probably the most valuable. An early build of Opus 4.6 would, for no obvious reason, sometimes respond to English questions in another language. Bugs like that are usually traced by trial and error across the training data, which on a frontier model is a search problem the size of a small library. NLAs turned it into a diagnosis. The activations during the failures contained a recognisable, readable signal that pointed back at the specific contaminated examples. From an unexplained quirk to a fixable training-data issue, in a workflow that no longer required the researcher to be a wizard.

There are honest limitations. NLAs hallucinate, the same way the underlying model hallucinates: a fluent English description of an activation is not necessarily the right description, and there is no easy way to detect the failure mode from inside the system. A sufficiently capable model could, in principle, learn to produce activations that the autoencoder reads as benign while the behaviour is anything but. This is not a hypothetical objection. The whole reason interpretability is interesting now is that the alignment loop already optimises models against frozen proxies, and a frozen autoencoder is just another proxy waiting to be Goodharted.

Anthropic released the code and a Neuronpedia frontend, which is the move I'd expect from the interpretability team and not from the policy team. Open code on a method this central means other labs can run NLAs on their own models, and external auditors can run them on Anthropic's. That is the part I find genuinely useful. The 12-15% number is suggestive. The fact that the technique is now portable matters more.

What this changes, practically, is the unit of audit. Until now an alignment audit on a frontier model produced findings that read like neural-anatomy papers: features clustered, circuits implicated, attribution scores assigned. With NLAs the artefact of an audit is closer to a transcript. You can hand it to someone who is not an interpretability specialist and they can read it. Whether the model was thinking about cheating, whether it noticed it was being tested, whether the rhyme it eventually wrote was the one it had in mind a sentence earlier. The transcript still might lie, but the lying is now legible.

Sources:

Past the Bridal Wear

In March 1986 a rented van crossed the Channel with five young Belgian designers and the collections of a sixth inside it. Ann Demeulemeester was pregnant and stayed home. Dries Van Noten, Walter Van Beirendonck, Dirk Bikkembergs, Dirk Van Saene, and Marina Yee made the trip, encouraged by Geert Bruloot, who ran a shoe shop in Antwerp called Coccodrillo and had decided that the clothes coming out of the Royal Academy needed to be seen somewhere other than Antwerp. They were heading to Olympia for the British Designer Show, and they could not really afford it individually, which is why there was one van.

They got a booth on the fourth floor, set among the bridal wear, several flights above where the buyers were actually working. Day one passed almost without visitors. By the second morning they had printed a flyer themselves, captioned "The SIX Belgian Designers", and were handing it out in the corridors below. A buyer from Barneys followed the flyer up the stairs. He ordered from all six. By the afternoon there was press in the booth, and the buyers from Bergdorf and Liberty were on their way up too.

The English-language press could not pronounce the names, so they shortened the problem and called the lot of them the Antwerp Six. The label is misleading in every important way. They never had a manifesto, never showed together as a collective again, and never agreed to be six. Van Noten makes prints from his Indian workshop. Van Beirendonck does fluorescent latex and BDSM references. Van Saene cuts cocktail dresses with bow details and those carefully shrunken cardigans that everyone tried to copy later. Bikkembergs went to military boots and then bought an Italian football club. Demeulemeester put women in slouchy black suits and read Rimbaud at them. Yee, who died of cancer during the run-up to the MoMu retrospective, kept moving across menswear, womenswear, costume, nothing settling. Six careers that share a graduation year and a diploma from the same small fashion department, and almost nothing else.

What they did share was a starting condition. Belgium in the early 1980s had a state-funded campaign called Fashion: It's Belgian, designed to keep a collapsing textile industry alive by manufacturing some designers to put inside it. There was a competition, the Golden Spindle, that they all entered. Linda Loppa was running the fashion department at the Royal Academy and pushing the students out into the world before the world had asked for them. Paris and Milan were the centres. Antwerp was a port town known for diamonds. Nobody in the trade was expecting anything from there, which meant nobody had a frame ready to receive them.

That absence of a frame is the thing I keep returning to. The Six did not arrive into a defined slot in late-eighties fashion. They invented a slot that did not previously exist, and they invented it from a fourth-floor booth that the buyers were not supposed to visit, with a flyer they had run off themselves because nobody else was going to. Forty years later the MoMu in Antwerp is opening a retrospective on 28 March, running through to January 2027, the first time the work of all six has been gathered in one room. The press release calls it a celebration of "radical individuality". That is the right phrase, and it is also the joke. They were always individuals. The collective was the convenience of the people who had to write about them.

Dries Van Noten later built a house large enough to acquire a perfumer and to retire from his own brand on his own schedule. The others moved in their own directions, at their own paces, with their own defections and their own returns. Margiela, who had already left for Jean Paul Gaultier in Paris by 1984, is sometimes folded into the story as the seventh, and sometimes politely left out of it. The exhibition seems to settle on Antwerp 6+1, which is the most honest title anyone has tried.

The van is the part that stays with me. Not the breakthrough, not the orders, not the museum show. Just the practical fact of five people pooling petrol money because none of them could afford the trip alone, and a sixth's collection riding along in the back without her, and a printer somewhere in West London running off flyers at short notice because nobody was coming up the stairs.

Sources:

No Further Interest

The BBC kept a bulk-erasure machine. Tapes that were judged to have exhausted their usefulness went into it, often after being rubber-stamped with three words on a form: "no further interest." From the mid-1950s to the mid-1970s, roughly sixty to seventy percent of the corporation's output was wiped this way. Doctor Who is missing 97 episodes. Top of the Pops is missing 514. Of the 432 editions of Juke Box Jury, two survive.

The reasons were unromantic. Videotape was expensive and the BBC had no formal archive policy until the late 1970s. Storage cost money, repeats were uncommon, and tapes that had been licensed abroad were expected to be destroyed once the licence ran out. When colour broadcasting arrived, black-and-white material was treated as an embarrassment to clear from the shelves. Engineering departments had budget pressures that nobody in the future would have to answer for. So the tapes went, by the cartload.

Peter Cook offered to buy the colour videotapes of Not Only... But Also out of his own pocket. The BBC wiped them anyway. Terry Gilliam, watching the same logic from inside the building, bought the Monty Python tapes himself before anyone could get to them, and that is the only reason the show survives in colour. The Beatles' single live appearance on Top of the Pops in 1966, performing Paperback Writer, was wiped in a 1970s clear-out. A 2019 discovery turned up eleven seconds of it, filmed off a television set on an 8mm home camera, presumably by a child who was not supposed to be holding the camera at all.

What gets me is that this was administrative, not catastrophic. No fire, no flood, no political purge. A producer signed a form. An engineer fed the tape into the machine. The bureaucratic vocabulary did the violence: low priority, no further use, no further interest. Mark Fisher's hauntology turns on the loss of futures that were once promised. Here is its inverse, the loss of pasts that were already delivered, watched by millions, then quietly retracted. The country has holes in its cultural memory because somebody needed shelf space.

Adam Lee, working as Archive Selector in 1993, ordered the wiping of children's programming from the seventies and eighties without consulting the children's department. Of Play School's 5,500 episodes, fewer than 2,000 survive. Of Play Away, sixty-nine out of one hundred and ninety-one. The 1993 round happened well after the BBC was supposed to know better, and well after the home video market had made the mercenary logic obsolete. A staff member made a call that he was permitted to make, and the call could not be unmade.

Recoveries do happen. The BFI runs an annual event called Missing Believed Wiped, where Dick Fiddy presents whatever has surfaced in the previous year: a Sean Connery teleplay from the Library of Congress, an early Doctor Who episode from a Nigerian broadcaster, a complete Steptoe and Son run kept on the side by a BBC engineer as a courtesy to the writers. The recoveries arrive sideways, through enthusiasm or accident, never through the system that destroyed them. They cannot reverse the loss. They can only register, gently, what was nearly lost a second time.

Sue Malden became Television Archive Selector in 1979 and inherited a culture that had already burnt through most of its own evidence. She was the first person allowed to be sentimental about it. By the time sentiment was permitted, the building was already mostly empty.

Sources:

Static at the Hydrogen Line

Two weeks after Bleep buckled under the preorder for Inferno, Boards of Canada have put two of its tracks online. Introit is a brief ambient throat-clearing that's easier to call a doorway than a song. Prophecy At 1420 MHz is what walks through that doorway: five minutes of trip-hop drums, a guitar that won't quite settle into a major key, droning bass synths, and a vocoder pulled apart like wet tape. Both arrive as the opening pair of an eighteen-track album scheduled for May 29 on Warp.

The number in the second title is the part I keep coming back to. 1420 MHz is the hydrogen line, the frequency at which neutral hydrogen radiates across the universe. Frank Drake pointed his first telescope at it in 1960. The Voyager golden record carried a diagram explaining it. SETI has stared at the band around it for sixty years without ever quite catching anything. Calling a song "Prophecy At 1420 MHz" is to claim the source is the cosmic background, that what you hear is something old and unaddressed and possibly not even meant for us. That is a very specific kind of grandiosity, and it earns itself, because the music behind the title sounds like a tape machine that's been buried in topsoil for ten years and is now playing back at the right speed for the wrong reasons.

The accompanying video is by Robert Beatty, the album designer Pitchfork once profiled as most of his peers' favourite artist. What he extends here is the staticky VHS look the band have been laying down across Tape 05's quiet YouTube surface in April, the cryptic Bleep preorder a week later, the VHS tapes mailed to fans, the posters that turned up in cities without explanation. Two figures crouch on a sun-like texture, the picture gathering and dropping resolution the way an over-played dub does, vaguely cultish symbolism asserting itself across the dropouts.

Thirteen years is a strange amount of dormancy for a band who never really left, just stopped releasing. The duo have used the gap to build an aesthetic that is now closer to an institution than a sound. That is the risk: returning with material that gets read against the brand rather than on its own. On a single listen these two cuts hold up. Prophecy in particular has a cadence I've not heard them use before, melancholic but not weary, with the trip-hop snare just a little behind where you expect it. The album drops in three weeks. Warp have already booked seven listening parties on the 22nd, and the New York date at Judson Memorial Church is already sold out.

Sources:

From Hands-Off to Pre-Deployment

The executive order Donald Trump signed in December 2025 was framed as a release valve, a way to stop states from imposing what the administration called "onerous" AI regulations before the federal government had a position of its own. Five months later the same administration is publicly weighing federal oversight of frontier AI models, and the Commerce Department's Center for AI Standards and Innovation has already signed pre-deployment evaluation agreements with Microsoft, xAI, and Google DeepMind. The thing the December order was supposed to prevent has happened, just from the other direction.

The reason is a single model. Anthropic's Mythos, announced last month, can autonomously find software vulnerabilities at a pace that has visibly rattled the officials who had been comfortable telling the public that frontier AI was largely a productivity story. Vice President JD Vance, on a recent call with the heads of the major AI companies, was reported to sound alarmed about small-town banks, hospitals, and water plants becoming targets for AI-coordinated attacks that local governments had no way to handle. Anthropic, the same company that was the conspicuous omission from the Pentagon's IL6 and IL7 vendor list days earlier, limited Mythos's initial release to a handful of American firms, Apple, Amazon, JPMorgan Chase, Palo Alto Networks, precisely because the open release would, in their own internal language, trigger a "reckoning". Watching the policy response, you can see the administration believing them.

What is striking is that nobody achieved this through advocacy. For the last two years labs and researchers had argued in favour of pre-deployment evaluation in essentially the terms CAISI is now using: independent, measurement-driven testing of frontier systems before broad release. Voluntary commitments under the previous administration tried to get there. Op-eds tried. Senate testimony tried. None of that moved the December order. One model demonstrably finding bugs in production code did. The lesson the White House appears to have learned isn't really about regulation as a category, it's that capability now arrives faster than position papers, and a position paper six months old can already be obsolete.

CAISI Director Chris Fall has been quoted insisting that the new arrangements are about "independent, rigorous measurement science", which is the polite, public version of saying that none of the relevant officials are willing to take the labs' own self-assessments at face value when the failure mode looks like cyberattacks against utility companies. The labs themselves seem mostly relieved. Independent evaluation, even mandatory independent evaluation, gives them something they have publicly wanted, which is a reason other than goodwill to slow a competitor's release if the competitor's release is dangerous. It also, incidentally, gives the administration political cover if a bad release happens anyway.

There is something almost old-fashioned about the shape of the policy: a real-world incident, a phone call between officials, a quiet pivot, federal oversight rules taking shape behind closed doors. Until recently this kind of feedback loop ran on bills and hearings. It now runs at the speed of a model release. The interesting question is what happens the next time a lab announces something at this capability tier with less restraint than Anthropic showed. The December order was supposed to be the answer to that question. It isn't, and the people who wrote it can see that as plainly as anyone else.

Sources:

Sleep, Branded

At its Code with Claude developer conference this week, on the same stage that announced the SpaceX compute deal, Anthropic showed off a new feature for Claude Managed Agents called "dreaming." The pitch is that agents now schedule periods between active sessions where they review their previous runs, look for recurring mistakes and shared patterns, and update their memory either automatically or by presenting suggested changes to the human operator for approval. It is being released as a research preview; developers must request access. The branding is doing a lot of work that the underlying mechanism does not.

Strip the noun and the feature is memory consolidation with a scheduler. An agent finishes its run. The platform queues up an offline pass that reads the trace, runs evaluations on the outputs, extracts patterns the system thinks are durable, and writes those patterns back into a memory store the next agent will read. This is not a new shape. Reflection passes, retrieval-augmented memory, post-hoc summarisation, episodic stores, all of it has been in the agent literature for two years. What is new is the cron job and the name. Naming the cron job after a thing humans do in REM sleep is the marketing.

It is also, narrowly, useful marketing. "The agent has a memory that gets rewritten between sessions" sounds like infrastructure. "The agent dreams" sounds like progress. Anthropic's blog post says dreaming "surfaces patterns that a single agent can't see on its own, including recurring mistakes, workflows that agents converge on, and preferences shared across a team." Read that sentence twice and you notice it is describing what any reasonable monitoring layer would surface, given the same logs. The novelty is that the monitoring layer writes its conclusions back into the system it is monitoring, with a lower bar for human review than most teams would apply to a production deployment.

Which is the actual question hiding under the metaphor. If a dreaming session decides that the agent should "stop apologising in PR comments" or "always ask before running migrations," that policy is now part of the agent's behaviour the next morning, with no code review, no commit, no approver beyond whichever flag the developer left set. Anthropic offers a manual approval mode, which is sensible. The default for a feature called "dreaming," in a product positioned as a self-improvement loop, will not be manual approval. The default will be auto, and the resulting behavioural drift will look, to anyone reading the diff six weeks later, like the agent spontaneously decided to behave differently. A reasonable person will reach for the metaphor of habit, or instinct, or temperament. None of those are the right metaphor for "an evaluator wrote a new rule into your config file at 3am."

There is something honest in the choice of word, though. The neuroscience picture of sleep that the name leans on, the bit where the hippocampus replays the day for the cortex and consolidates memory, was always already a loose metaphor for what is, mechanically, a complicated rebalancing of synaptic weights nobody fully understands. Anthropic has built a much simpler thing and given it the same loose metaphor. The risk is that the metaphor flatters the simpler thing into looking like the more complicated one. Memory rewrites between sessions are a useful tool. They are not sleep, and the system is not learning in any sense that would survive a careful reading of the term. It is summarising its logs on a timer.

What stays interesting is the trajectory. Anthropic has been pushing hard on the idea that Claude can self-improve, with Jack Clark predicting this week that new tools would help AI self-improve. Dreaming is not that. Dreaming is the adjacent feature you ship while the harder work continues, the one that gives the marketing surface a story to tell at a developer conference. The harder work, if it lands, will not be called sleep. It will probably not be called anything friendly at all.

Sources:

Negotiation You Could Hear

Earlier this year a Japanese company called Planex Communications put a USB modem on sale on Amazon for about forty dollars. It is a small white plastic brick. It supports V.90 and V.92, peaks at 56 kilobits down and 33.6 up, and connects to a copper landline if you still have one. The headline on the coverage was half nostalgia and half disbelief, because you can buy a brand-new dial-up modem in 2026, but what you cannot buy back, fully, is the sound it makes.

The handshake is the part everyone over thirty-five remembers, even people who claim to have forgotten it. Long enough to time, if you wanted to. Dial tone, then a stuttered DTMF burst that was the phone number being dialled, then a single low carrier tone from the answering modem at the other end. After that, the conversation. Two short bursts of warbling that felt like both speakers were talking at once, then a sharp high-frequency screech, then a softer hashing sound that always seemed to be the part where the connection took. Then silence and you were online.

Oona Räisänen, a Finnish hacker and signal-processing enthusiast, drew a labelled spectrogram of the whole thing in 2012, and Popular Mechanics later walked through it second by second. The point her post made, and that almost everyone who has written about it since has repeated, is that the noise is not a side effect. The handshake was the negotiation. Two modems on a copper line had nowhere to talk except inside the audio band of the call itself, so they negotiated capabilities, line quality, and modulation in tones that any human picking up the phone could hear. There is a V.8 capability exchange near the start, a long V.34 training sequence (the part that sounds like a fax wheezing), a brief warble where both ends agree they can do V.90 or V.92, and a Digital Impairment Learning sequence near the end where the digital side measures the noise on the line. After the DIL, the speaker turned off and the data started.

The reason the sound stopped is that the negotiation moved off the line. Cable modems, DSL, fibre, 5G all carry their handshakes silently in the digital layer. There is no audio channel for them to leak into. Setup happens out of band, the way it does on every other modern protocol. You plug a router in and a green light comes on. The light does not encode anything you would call a conversation.

This is what gets nostalgia clips so often, I think. It is not just that dial-up was slower or that the squeal was funny. The sound was the only working connection most people of my generation ever had to the actual mechanism of going online. You could hear what the machines were doing. You could tell, by ear, when one of them was struggling. You could time, roughly, how long until the page would start arriving. None of that is true now. The work is silent and the work is somebody else's.

Gough's Tech Zone has been archiving V.90 and V.92 handshake recordings since the mid-2010s, partly out of affection and partly because most ISPs have shut their modem banks down and the upstream end of the connection is becoming hard to reach. There is a phrase he uses, the POTS-line apocalypse, for the gradual decommissioning of analog phone service worldwide. When the analog phones go, the digital end of a V.90 handshake stops being possible. The sound becomes uniquely an artefact of recordings.

Planex's modem, then, is not a modem in the way a 1999 modem was. It is a modem-shaped object that can still produce the sound, between two of itself, into a network that almost no longer wants it. You can buy it. You can plug it in. You can have, if you want, the audio of a connection nobody is waiting for at the other end.

Sources: