Showing posts with label Critical Thought. Show all posts
Showing posts with label Critical Thought. Show all posts

Saturday, April 4, 2026

You Work For the Bad Boss You Have, Not the Good Boss You Wish You Had

In every job, there is some gap between the advertisement and the reality. The outdoor job boasting “fresh air every day” consists of picking up trash. The service job “perfect for sunny personalities” consists of getting yelled at by angry customers. The day care job that offers “unlimited cuteness” consists of cleaning up poop. This is how it goes. We must all endure some amount of hastily concealed tribulations in order to pay the bills.

Likewise, we must persevere through bad bosses. We must learn to navigate pernicious supervisors, backstabbing managers, and incompetent executives as the price of gainful employment. Good bosses are a stroke of luck, rather than a baseline expectation. If we all limited ourselves to workplaces that were free of bad bosses, it would take ten years to find a job.

Mostly, this is to be expected—an inducement to aspire to a promotion, or to start a union, rather than a catastrophe. But there are some jobs where a bad boss is a bigger deal. There are some jobs where a bad boss can very quickly get you into a genuine moral crisis. If you have a job like that, shrugging off what the bad boss is doing can become not an act of resilience, but one of gross negligence.

The military is one job of this type. There is more moral urgency attached to the military’s conduct of its affairs than to, you know, a restaurant’s conduct of its affairs, due to the fact that the military kills people. There are higher stakes to poor management decisions. If you are a line cook and your boss tells you to cook a dish improperly and you acquiesce for the sake of keeping your job, you can be forgiven. If you are a member of the military and your boss tells you to kill innocent people or bomb their homes or snatch their freedom and you acquiesce for the sake of keeping your job, forgiveness is not so certain. You become not a beleaguered employee, but a true villain. The space that the world is able to afford you as a matter of sympathy for your workplace annoyances shrinks down to almost nothing once guns are involved.

People join the military for all sorts of reasons: For economic opportunity, for adventure, for patriotism, for sheer lack of options. Most soldiers, it is safe to say, believe they are doing something good. Even those who are not ultra-patriots probably believe—and are told, by ads and by supervisors and by TV and by politicians and by the public—that their jobs are, on balance, honorable ones. They do something difficult, and they believe they do something necessary, and they take a certain amount of pride in that, as anyone would.

The school in Iran where we blew up kids.

But the military is a gun in the hand of the Commander in Chief and we have a Commander in Chief who is dumb, narcissistic, unpredictable, and dangerous. The bad boss problem, for soldiers, is everything. It is the difference between being honorable and being the violent foot soldier of a thug. Which situation is closer to reality now, do you think? Being a soldier is not inherently righteous. That is a fairy tale they tell teenagers in order to get them to join the military. The righteousness of an army is wholly dependent on the righteousness of the cause that the army fights for. (Teenagers learn this, too, about other armies in other nations. We are careful never to tell them to apply the principle to the United States itself.) [...]

Here in America, when we are talking about American soldiers, we typically say they are honorable public servants and dismiss any blame for the havoc they wreak. Whereas if we are speaking about other soldiers in other nations, we expect and call for them to be killed by our own soldiers because they are carrying out equivalent duties. I hope I do not have to point out the ethical schizophrenia of this approach.

My purpose is not to demonize members of the military. On the contrary. People who joined an organization with noble intentions, who were told that they were serving the purest interests of their country, are now in the position of being foot soldiers for a gangster-style president who is quite possibly the single biggest threat to peace on earth. It is important that we speak honestly about the fact that these soldiers are in the perilous position of risking their lives in order to carry out villainous goals. That would be a tragedy not only for the victims of American imperial overreach, but also for the American soldiers themselves, who will be cursed to live their lives with the knowledge of what they have done. You may have joined the organization imagining what good it could do with a good boss. But that is not the world you have actually entered. In this world, the world that exists, you are an armed member of a deadly organization run by a bad boss. He has done and will continue to do bad things. And who will have to carry out the bloody acts inherent in those bad things? You will. It’s a bad deal. While you may have come to find yourself in this position through a series of well-intentioned actions, the fact is that the only ethical thing to do is to do your utmost to remove yourself from a job that might ask you to kill, unethically, on behalf of a bastard.

The military is not the only sort of job in this same position today. Many well-intentioned people who went to work in, say, the State Department, or the CDC, or other branches of government may now be faced with a similar moral dilemma.

by Hamilton Nolan, How Things Work |  Read more:
Images: Getty
[ed. Should be fun telling the grandkids what they did in their career. See also: Digging up the Dead (LRB):]
***
More than a million people have been displaced by Israel’s invasion and bombardment of Lebanon, many fleeing with nothing more than the clothes on their back, camping in the hills or sheltering in schools or municipal buildings. The desire to turn Lebanon into another Gaza, articulated by Israeli spokesmen, is being fulfilled with attacks on journalists, the use of undercover operatives and the bombing of displaced families huddled in makeshift shelters. Familiar too was the timing of the attacks during Ramadan, frequently at iftar when people were about to break their fast.

The attack on the small town of Nabi Chit in the eastern Bekaa Valley on 6 March shows the value placed on human lives by the regime in Israel and its backers in the United States. According to the Israeli government, the invasion was a rescue operation to retrieve the remains of an Israeli airman who disappeared forty years ago. Residents of Nabi Chit and the Lebanese army chief told the BBC that Israeli special forces entered the town ‘disguised in Lebanese military fatigues and used ambulances with signs of Hizbullah’s Islamic Health Organisation’. They headed to the corner of the graveyard, dug it up but found nothing there. The town fought back, causing the Israeli soldiers to withdraw. To cover their retreat, Israel carried out more than forty airstrikes in five hours, killing 41 people. [...]

According to Euro-Med Human Rights Monitor, 58 of Gaza’s 62 graveyards have been damaged or destroyed:
the Israeli army carried out a focused operation at al-Batsh Cemetery, east of Gaza City, in January 2026. The cemetery was converted into a military barracks, and more than 700 bodies were exhumed under the pretext of searching for the body of an Israeli detainee. The army later withdrew after extensive bulldozing that radically altered the cemetery’s landscape, preventing families from locating their relatives’ graves.
This was not an isolated event:
in many cases, the Israeli army deliberately exhumed graves and converted cemeteries into military barracks under the pretext of searching for the bodies of Israeli detainees. These actions were carried out without documented, verifiable procedures, independent oversight, or a clear chain of custody and handover process. Israeli forces removed hundreds of bodies from their burial sites, mixed remains, failed to return them to their original locations, and provided no identifying or biological data to enable verification or documentation, making the recovery and identification of remains extremely difficult. [...]
The scope of this policy has been expanded in recent years and upheld by the Israeli Supreme Court. Amira Haas reported in Haaretz last month that Israel was holding the bodies of 776 Palestinians. There are 256 buried in nameless, numbered graves while the rest are held at military morgues. Nearly half were killed since October 2023 and 88 died in Israeli detention. These bodies are sometimes used as bargaining chips in negotiations, although in more recent exchanges many of the bodies have been unrecognisable when returned. There is at least one case of the wrong (long awaited) body being returned to the family.

The grieving are not allowed to grieve, the reunited to celebrate. The repeated disruption of Palestinian funerals and prisoner releases has been bolstered by military orders that proscribe and punish expressions of Palestinian emotion. In February 2025, during a prisoner exchange, the psychotherapist Gwyn Daniels quoted Edward Said, who
chose the word ‘inert’ not descriptively but to conjure up the Zionist fantasy about the ideal Palestinian body. Perhaps for the coloniser, this ideal body should be lifeless or ‘disappeared’. But given the stubborn persistence of Palestinians to remain living on their land, they should preferably cause as little disturbance to the colonisers as possible. Along with no displays of cultural identity, there must be no passion, no pride, no joy, no sorrow, no anger – indeed, no demonstrable emotions that might trouble their oppressors.

Wednesday, April 1, 2026

The AI Doc

 

(This will be a fully spoilorific overview. If you haven’t seen The AI Doc, I recommend seeing it, it is about as good as it could realistically have been, in most ways.)

Like many things, it only works because it is centrally real. The creator of the documentary clearly did get married and have a child, freak out about AI, ask questions of the right people out of worry about his son’s future, freak out even more now with actual existential risk for (simplified versions of) the right reasons, go on a quest to stop freaking out and get optimistic instead, find many of the right people for that and ask good non-technical questions, get somewhat fooled, listen to mundane safety complaints, seek out and get interviews with the top CEOs, try to tell himself he could ignore all of it, then decide not to end on a bunch of hopeful babies and instead have a call for action to help shape the future.

The title is correct. This is about ‘how I became an Apolcaloptimist,’ and why he wanted to be that, as opposed to an argument for apocaloptimism being accurate. The larger Straussian message, contra Tyler Cowen, is not ‘the interventions are fake’ but that ‘so many choose to believe false things about AI, in order to feel that things will be okay.’

A lot of the editing choices, and the selections of what to intercut and clip, clearly come from an outsider without technical knowledge, trying to deal with their anxiety. Many of them would not have been my choices, especially the emphasis on weapons and physical destruction, but I think they work exactly because together they make it clear the whole thing is genuine.

Now there’s a story. It even won praise online as fair and good, from both those worried about existential risk and several of the accelerationist optimists, because it gave both sides what they most wanted. [...]

Yes, you can do that for both at once, because they want different things and also agree on quite a lot of true things. That is much more impactful than a diatribe.

We live in a world of spin. Daniel Roher is trying to navigate a world of spin, but his own earnestness shines through, and he makes excellent choices on who to interview. The being swayed by whoever is in front of him is a feature, not a bug, because he’s not trying to hide it. There are places where people are clearly trying to spin, or are making dumb points, and I appreciated him not trying to tell us which was which.

MIRI offers us a Twitter FAQ thread and a full website FAQ explaining their full position in the context of the movie, which is that no this is not hype and yes it is going to kill everyone if we keep building it and no our current safety techniques will not help with that, and they call for an international treaty.

Are there those who think this was propaganda or one sided? Yes, of course, although they cannot agree on which angle it was trying to support.

Babies Are Awesome

The overarching personal journey is about Daniel having a son. The movie takes one very clear position, that we need to see taken more often, which is that getting married and having a family and babies and kids are all super awesome.

This turns into the first question he asks those he interviews. Would you have a child today, given the current state of AI? [...]

People Are Worried About AI Killing Everyone

The first set of interviews outlines the danger.

This is not a technical film. We get explanations that resonate with an ordinary dude.

We get Jeffrey Ladish explaining the basics of instrumental convergence, the idea that if you have a goal then power helps you achieve that goal and you cannot fetch the coffee if you’re dead. That it’s not that the AI will hate us, it’s that it will see us like we see ants, and if you want to put a highway where the anthill is that’s the ant’s problem.

We get Connor Leahy talking about how creating smarter and more capable things than us is not a safe thing to be doing, and emphasizing that you do not need further justification for that. We get Eliezer Yudkowsky saying that if you share a planet with much smarter beings that don’t care about you and want other things, you should not like your chances. We get Ajeya Cotra explaining additional things, and so on.

Aside from that, we don’t get any talk of the ‘alignment problem’ and I don’t think the word alignment even appears in the film that I can remember.

It is hard for me to know how much the arguments resonate. I am very much not the target audience. Overall I felt they were treated fairly, and the arguments were both strong and highly sufficient to carry the day. Yes, obviously we are in a lot of trouble here.

Freak Out

Daniel’s response is, quite understandably and correctly, to freak out.

Then he asks, very explicitly, is there a way to be an optimist about this? Could he convince himself it will all work out?

by Zvi Mowshowitz, DWAtV |  Read more:

Sunday, March 29, 2026

The Last Useful Man

About halfway through Mission: ImpossibleThe Final Reckoning, Tom Cruise goes for a run on a treadmill. The treadmill is on the USS Ohio, a submarine manned exclusively by implausibly attractive people. One of those people is not who they seem: a cultist, radicalized by the Entity, the film’s AI antagonist. The cultist sneaks up behind Cruise and lunges with a knife. Things look dicey for a moment — until Cruise gains some distance and kicks him repeatedly in the head. While doing so, he imparts a few words of wisdom: “You spend too much time on the internet.

What divides the heroes and villains in Final Reckoning is simple: the villains have to Google things, and the heroes do not. There are three bad guys, more or less. First, the Entity, a rogue AI halfway through its plan for global domination. Second, Gabriel, the Entity’s meat puppet. Third, a gang of surprisingly likable Russians who take Cruise’s team hostage in a house in Alaska. What unites the villains isn’t malice so much as it is uselessness. I mean that precisely. They are often effective, even successful. But never useful. [...]

This division between characters with embodied knowledge and those without runs through all of Cruise’s recent work. His own impossible mission is to teach the value of physical competence: not just knowing things, but knowing how to do them. In Final Reckoning, this idea finds its clearest form. [...]

Like Forster, Cruise and his long-time collaborator Christopher McQuarrie invent machines to dramatize the age they live in. Forster gave us the Machine; McQuarrie, the Entity. But unlike Forster, their imagination of technology is not apocalyptic but diagnostic — they aren’t warning us of the machine age so much as asking what it demands of us, and what it reveals.

This brings us to what looks, at first glance, like a paradox: How does a franchise so lovingly built on disguises, gadgets, and inventions of all kinds — from the eye-tracking projector that gets Cruise into the Kremlin to the single suction glove that lets him cling to the Burj Khalifa — end with a villain made of pure technology?

If you asked Cruise, his answer would be simple: technology is good when it roots you in your body and bad when it lets you forget you have one. That’s why Final Reckoning, for all its AI villainy and suspicion of the terminally-online, still treats technology with a near-Romantic sensibility. Hand-soldered pen drives, aging aircraft carriers, and vintage biplanes carry Cruise and his team on their mission to save the world. At times subtlety disappears altogether; the film’s most inviting location is a candle-lit Arctic hideout filled with analogue comforts: old books and gramophones, telescopes and soldering tools.

The same ideas return — turned up to eleven — in Cruise and McQuarrie’s two other collaborations this decade outside the Mission: Impossible franchise. The first, Edge of Tomorrow, in which Cruise relives the same day on repeat until he generates enough embodied knowledge to defeat an autonomous alien race, is, even for the purposes of this essay, too on the nose, so I’ll focus instead on Top Gun: Maverick.

The film opens with Cruise test-piloting an experimental stealth aircraft in a last-ditch attempt to save the program from cancellation by the “drone ranger,” an admiral who wants the budget for his autonomous fleet. For the program to survive, Cruise needs to hit Mach 10: a speed no vehicle has ever reached. As the team watches on, he delivers the impossible. Gauzy wisps of supersonic air stream across the cockpit windows as Maverick stares out into the black of space. He whispers softly to his dead best friend, “Talk to me, Goose.”

Soon afterwards, Maverick is sent back to Top Gun to train a new generation of pilots. He begins his first lesson holding up the flight manual for the F-18, which makes the Riverside Chaucer look like a novella, before throwing it in the bin. “I assume you know this book inside and out. So does your enemy.” What matters instead is the knowledge that can’t be written down: the things his students already know by instinct, but cannot yet express  “Today we’ll start with only what you think you know.”

The quest to ‘“know more than we can tell,”’ as Michael Polanyi put it, drives the rest of the film. The pilots even have their own version of the phrase, a near-religious catechism recited at almost every decisive moment: “Don’t think. Just do.”

Beyond the screen, the same principle applies. In the Mission: Impossible franchise, filming begins with no plot or script, only a commitment to figuring it out in the process. It’s most evident in each film’s tentpole action sequences, where the line between Cruise the actor and Cruise the stuntman blurs beyond recognition.

The art critic Robert Hughes once wrote of his love for “the spectacle of skill” — the thrill of watching an expert at work, whatever the discipline. Nowhere is this more evident than in Cruise’s increasingly daring plane sequences. In Mission Impossible: Rogue Nation, Cruise clings to a real Airbus A400M as it lifts off from an airfield in Lincolnshire. He sprints across the field, in that inimitable Tom Cruise style, mounts the wing with practiced ease, and seats himself by the cargo door. The plane taxis. So far, so cool. Then it lifts off. The perfect hair vanishes, blown back and forwards, alternating second by second between old skeleton and boy with bowl cut. His clothes are shapeless and billowing, pulled off him by the force of the air.

This is no country for sprezzatura, nor the embodiment preached by the wellness industry with its vocabulary of “balance” and “equilibrium.” Here, we are meant to feel the effort. To know yourself is to know your limits, and so push your body to the edge of failure. When they are about to perform stunts, Cruise often briefs his team with an unusual mantra: ‘Don’t be safe, be competent.”

At the end of Final Reckoning, Cruise plummets through the sky as his parachute burns to cinders above him. To film it, the stunt team soaked a parachute in flammable liquid, flew him to altitude in a helicopter, and pushed him out as it ignited. He did this 19 times. When he asked to go again, the stunt coordinator told him there were no parachutes left. This was a lie. McQuarrie was more direct: “You’re done. Do not anger the gods.”

It’s interesting to see this return to embodiment and strange to find myself drawn to it. Like many default clever people, I’d long paid lip service to Merleau-Ponty and his ilk while living as a dualist; my brain was the moneymaker, my body just along for the ride. It was only after having children that I began to understand what it meant to inhabit a body rather than simply use one.

In an essay for Granta earlier this year, the writer Saba Sams contrasted her son’s love of leaping from benches and walls with her own unease: “For them, the body is not a constraint, is not a ticking clock, is not something to be moulded or hidden. The body is the window to movement, and movement is a window to joy.”

Sams captures something larger. This renewed fascination with embodiment isn’t spontaneous, it’s a reaction to technologies so powerful and frictionless they’re impossible to ignore. Even the most grounded among us now move through the world not through our bodies but through screens, which is why so many make the negative case for technology, urging us, thankfully without a Cruise-style kick to the head, to spend less time on the internet.

What Cruise gives us is the positive case: not just resistance to disembodiment but a reminder of what is beautiful about being physical in the first place. The skilled things bodies can do are inherently satisfying. They can be thrilling, reassuring, even a little terrifying. But, as David Foster Wallace put it in his essay on Roger Federer:
The human beauty we’re talking about here is beauty of a particular type; it might be called kinetic beauty. Its power and appeal are universal. It has nothing to do with sex or cultural norms. What it seems to have to do with, really, is human beings’ reconciliation with the fact of having a body.
That’s the mission, if we choose to accept it. The target is not the recent bugbear of AI, but instead the more gentle conditions of modernity. When we use Google Maps instead of a printed atlas, or when CGI is used to sell a stunt instead of the performers doing it themselves, something is lost. It’s why the focus on AI can sometimes be misguided. It’s not so much a revolution, it’s simply the next step on the ladder of disembodiment: another in a long line of technologies to make humans a little less self-reliant. Why learn, if you can ask?

In the final biplane sequence, we watch Cruise commandeer a plane, fly it to another, board that plane midair, and take control of it — a feat so exhausting it beggars belief. Gabriel, the villain, in order to survive his defeat, needs only do something a hundredth as difficult: jump from the plane and deploy a parachute. He laughs. This is easy. But he doesn’t know the complexities of leaving a biplane with a parachute — the correct moment to release, the parts to steer clear from. He’s never bothered to learn. He frees himself, clips the rudder, cracks his skull open, and dies.

Here we see the real villain: not intelligence, but convenience. The mission so often feels impossible because we keep trying to do things without effort. Cruise’s answer is simple: Stop. Remember your body. Sometimes, it’s better to take the hard way.

Final Reckoning’s closing scene presents us with two intelligences and two bodies. One is Cruise, a 62-year-old body who we’ve seen, for the last two hours, run fast, dive deep, and hang from planes. The other is the Entity, trapped in a glorified USB stick: a golden nugget incapable of anything other than being flushed down a toilet.

One still moves. The other never could.

by Aled Maclean-Jones, The Metropolictan Review | Read more:
Image: Getty

Shoot the Messenger

Media criticism these days is usually focused on ideological bias. The right argues the media is full of left-wing hacks, and the left points out that many media moguls are right-wingers. But I actually find ideological bias to be less concerning than the more fundamental problem that the class of people who determine the boundaries of debate share a set of demographic and experiential traits that they don’t recognize as distinctive.

This class of people includes journalists, yes, but also people who work in the tech industry, academics, nonprofit leaders, influencers, and those who work in politics. From now on, I’ll refer to this group broadly as “the messenger class.”

The messenger class’s distinctive experiences — like living in downtown Washington, D.C., or living in one of the parts of New York highlighted in red — shape the boundaries of normal in ways harder to counteract than pure ideological or partisan bias.

The messenger class plays a fundamental role in any democracy. Democratic self-governance requires not just fair procedures for making decisions but an accurate and shared picture of social reality to reason about. That picture is revealed through the communicated experiences of citizens, filtered through the messenger class, which decides which experiences are urgent and require intervention.

But if our mediating institutions are all staffed by people drawn from the same narrow demographic band, then the picture they produce will be skewed in ways nobody intends and few notice. This isn’t about whether the messenger class is full of bad people — it’s largely not — it’s about whether it’s even possible to know when you’re acting as a mirror to society, or a spotlight on what you personally happen to care about. [...]

[ed. Examples: GentrificationOpioid Epidemic; AI Job Displacement; Rising Unionization:]

The psychology of projection

There is a name for what’s happening here. Psychologists call it the false consensus effect — the tendency to overestimate how much others share your beliefs, behaviors, and experiences.

First established by Ross, Greene, and House in 1977, it has been confirmed in a meta-analysis of 115 hypothesis tests and found to be a robust, moderate-sized effect. Later research shows that it persists even when people are warned about it.

Neuroimaging research has shown that projecting your views onto others activates the brain’s reward centers; it feels good to believe everyone is like you. And a 2021 study found that social media use amplifies the effect: The more time you spend in an environment where your views are echoed back to you, the more convinced you become that those views are universal.

The false consensus effect is usually studied at the individual level. But what I’m describing is a class-wide and industry-wide version.

It’s not just that any one journalist overestimates how representative her experience is; it’s that an entire class of professionals shares a similar set of experiences, confirms those experiences with each other on the same platforms, and then produces a body of public knowledge that reflects those experiences as though they were the norm.

And even when people from nontraditional backgrounds join the fray, they are incentivized to conform through social media, company cohesion, editorial norms, and the normal human urge to get along with your peers and be taken seriously by the people you respect.

So many problems, so little time

Agenda-setting is zero-sum.

There’s only so much time elected officials, charities, nonprofits, or businesses have to respond to the public’s needs. So if something is getting more coverage than may be warranted, that means other things are getting less. And that means fewer solutions are being explored.

Remember that gentrification report? It found that 15% of urban neighborhoods showed signs of gentrification over 50 years, while 26% experienced substantial population decline.

The far more common trajectory for a poor urban neighborhood is not invasion by white yuppies — it’s continued segregation, disinvestment, and deteriorating housing stock. But that story doesn’t get told with anywhere near the same intensity.

The same asymmetry shows up in the AI conversation. The workers most likely to struggle if displaced by AI are not the ones getting the most ink.

A Brookings analysis found that roughly 6.1 million workers face both high AI exposure and low adaptive capacity — limited savings, advanced age, narrow skill sets, scarce local opportunities. Eighty-six percent of these workers are women, and they’re concentrated in clerical and administrative roles in smaller metro areas.

I’m not arguing that journalists are dishonest, that scholars are corrupt, or that the messenger class is engaged in some conspiracy to distort public reality. The people I’m describing are, by and large, doing their best to tell the truth about the world.

The problem is that they’re drawing on their own experiences, their own social networks, and their own platform ecosystems as raw material — and those inputs are unrepresentative in ways they have no easy mechanism for detecting.

Part of this could be resolved with an increased fluency with quantitative data. But that’s not actually enough. Many stories — like the opioid epidemic — are ones that require journalists to respond to anecdotes before the quantitative data has been assembled, analyzed, and produced by the academy.

by Jerusalem Demsas, The Argument |  Read more:
Image: Klaus Vedfelt via Getty Images

Saturday, March 28, 2026

Welcome to a Multidimensional Economic Disaster

The global economy has become dependent on the AI industry. Trillions of dollars are being invested into the technology and the infrastructure it relies on; in the final months of 2025, functionally all economic growth in the United States came from AI investments. This would be risky even in ideal conditions. And we are very far from ideal conditions.

Much of the AI supply chain—chips, data centers, combustion turbines, and so on—relies on key materials that are produced in or transported through just a few places on Earth, with little overlap. In particular, the industry is highly dependent on the Middle East, which has been destabilized by the war in Iran. A global energy shock seems all but certain to come soon—the kind where even the best-case scenario is a disaster. The war could grind the AI build-out to a halt. This would be devastating for the tech firms that have issued historic amounts of debt to race against their highly leveraged competitors, and it would be devastating for the private lenders and banks that have been buying up that debt in the hope of ever bigger returns.

For the better part of the past year, Wall Street analysts and tech-industry observers have fretted publicly about an AI bubble. The fear is that too much money is coming in too fast and that generative-AI companies still have not offered anything close to a viable business model. If growth were to stall or the technology were to be seen as failing to deliver on its promises, the bubble might burst, triggering a chain reaction across the financial system. Everyone—big banks, private-equity firms, people who have no idea what’s mixed into their 401(k)—would be hit by the AI crash.

Until recently, that kind of crash felt hypothetical; today, it feels plausible and, to some, almost inevitable. “What’s unusual about this, unlike commercial real estate during the global financial crisis,” Paul Kedrosky, an investor and financial consultant, told us, “is all of these interlocking points of fragility.”

Perhaps the clearest examples are advanced memory and training chips, which are among the most important—and are by far the most expensive—components of training any AI model. Currently, most of them are produced by two companies in South Korea and one in Taiwan. These countries, in turn, get a large majority of their crude oil and much of their liquefied natural gas—which help fuel semiconductor manufacturing—from the Persian Gulf. The chip companies also require helium, sulfur, and bromine—three key inputs to silicon wafers—largely sourced from the region. In addition, Saudi Arabia, Qatar, the United Arab Emirates, and other regional petrostates have become key investors in the American AI firms that purchase most of those chips.

Because of the war in Iran, the Strait of Hormuz is functionally closed to most shipping vessels, stranding one-fifth of the world’s exports of natural gas, one-third of the world’s exports of crude oil, and significant quantities of the planet’s exportable fertilizer, helium, and sulfur. Meanwhile, Iran and Israel have begun bombing much of the fossil-fuel infrastructure in the region, which could take many years to replace. In only a month of war, the price of Brent crude—a global oil benchmark—has jumped by 40 percent and could more than double, liquefied-natural-gas prices are soaring in Europe and Asia, and helium spot prices have already doubled. The strait is “critical to basically every aspect of the global economy,” Sam Winter-Levy, a technology and national-security researcher at the Carnegie Endowment for International Peace, told us. “The AI supply chain is not insulated.”

The situation could quickly deteriorate from here. A helium crunch could trigger a shortage of AI chips or cause chip prices to rise. AI companies need ever more advanced chips to fill their data centers—at higher prices, the massive server farms, already hurting from elevated energy costs caused by the war, would have almost no hope of becoming profitable. Without these chips, new data centers would not be built or would sit empty. Astronomical tech valuations, and in turn the entire stock market, could collapse.

One industry’s precarious position isn’t usually everyone’s problem. Unfortunately, AI is different. The biggest data-center players, known as hyperscalers, are among the biggest corporations in the history of capitalism; they include Microsoft, Google, Meta, and Amazon. But even they will be pressed by collectively spending nearly $700 billion on AI in a single year. In order to get the money for these unprecedented projects, data-center providers are beginning to take on colossal amounts of debt. Some of this is done through creative deals with private-equity firms including Blackstone, BlackRock, and Blue Owl Capital—which themselves operate as sort of shadow banks that, since the most recent financial crisis, have arguably become as powerful and as influential as Bear Stearns and Lehman Brothers were prior to 2008. Endowments, pensions, insurance funds, and other major institutions all trust private equity to invest their money.

For a while, it seemed like every time Google or Microsoft announced more data-center investments, their stock prices rose. Now the opposite occurs: The hyperscalers are spending far more, but investors have started to notice that they are not generating anything near the revenue they need to. The data-center boom’s top players—Google, Meta, Microsoft, Amazon, Nvidia, and Oracle—have all lost 8 to 27 percent of their value since the start of the year, making them a huge drag on the overall stock market. And the $121 billion of debt that hyperscalers issued in 2025, four times more than what they averaged for years prior, is expected to grow dramatically.

All of the major players in this investment ecosystem are vulnerable. Private-equity firms are being squeezed on both ends by generative AI: During the coronavirus pandemic, they bought up software companies, which are now plummeting in value because AI is expected to eat their lunch. Meanwhile, private equity’s new investment strategy, data centers, is also falling apart because of AI. Blackstone, Blue Owl, and the like are sinking huge sums into data-center construction with the assumption that lease payments from tech companies will pay for their debt. In order to pay for their investments, private-equity companies raised money from major financial institutions—but now the viability of those lease payments is coming into question as the hyperscalers’ cash flow is strained. “There’s a reason to think we’re seeing some of the same 2008 dynamics now,” Brad Lipton, a former senior adviser at the Consumer Financial Protection Bureau and now the director of corporate power and financial regulation at the Roosevelt Institute, told us. “Everyone’s getting tied up together. Banks are lending money to private credit, which in turn lends it elsewhere. That amps up the risk.” [...]

The war in Iran affects data-center finances as well. Should energy prices continue to skyrocket, so will the cost of this already very expensive computing equipment, because it needs tremendous amounts of energy to manufacture and operate. And the war has exposed physical risks to these buildings. Janet Egan, a senior fellow at the Center for a New American Security, described data centers to us as “large, juicy targets.” It is impossible to hide these facilities, which can cover 1 million square feet. Earlier this month, Iran bombed Amazon data centers in the UAE and Bahrain. American hyperscalers had been planning to build far more data centers in the region, because the Trump administration and the AI industry have sought funding from Saudi Arabia, the UAE, Qatar, and Oman. Now there’s a two-way strain on those relationships. The physical security of the data centers is more precarious, and the conflict is damaging the economic health of the petrostates, thereby jeopardizing a major source of further investment in American AI firms. The Trump administration “staked a lot on the Gulf as their close AI partner, and now the war that they’ve launched poses a huge threat to the viability of the Gulf as that AI partner,” Winter-Levy said.

Plus, “what’s to prevent Iran or a proxy group, or another maligned actor, from tomorrow launching an armed drone against a data center in Northern Virginia?” Chip Usher, the senior director for intelligence at the Special Competitive Studies Project, a national-security and AI think tank, told us. “It could happen. Our defenses are not adequate.” State-sponsored cyberattacks of the variety Iran is known for could also knock a data center offline. You can build all manner of defenses—reinforced concrete, drone-interception systems—but doing so adds cost and time to already costly and slow construction. [...]

Even if Iran and the Strait of Hormuz don’t directly trigger an AI-driven financial crisis, the odds are decent that another vector could. (Remember tariffs?) Energy prices could stay elevated for years, because the targeted fossil-fuel facilities in the Persian Gulf will take a long time to restore. As the U.S. directs huge amounts of attention and military resources toward Iran, it’s easy to imagine China launching an invasion of Taiwan—a scenario that terrifies Silicon Valley, because it would halt the production of chips needed to train frontier models. That’s not even considering the single Dutch company that makes the high-tech lithography machines used to print virtually all AI chips, or the German company that makes the mirrors used in those machines. “There are too many ways for it to fail for it not to fail,” Kedrosky said of the AI industry’s web of risk. “All you can say for sure is this is a fragile and overdetermined system that must break, so it will.”

by Matteo Wong and Charlie Warzel, The Atlantic | Read more:
Image: An Amazon Web Services data center in Manassas, Virginia (Nathan Howard / Bloomberg / Getty)
[ed. See also: 

Wednesday, March 25, 2026

Enneagram Types

via: AI search and the Enneagram Institute
[ed. If you read the post following this one which discusses personal "agency" (which I encourage you to do) this might be helpful.]

On Agency and 'Can You Just Do Things?'

Clara Collier: In the spring of next year, you have a book coming out called You Can Just Do Things. It’s about agency. I'm interested in agency as a buzzword, as a concept, as a Silicon Valley cultural phenomenon, as a thing I can exercise in my life — maybe even as a thing I shouldn't exercise so much in my life. So, to start: How do you define agency? And why did you want to write a book about it?

Cate Hall: I define agency as the capacity to both see and act on all of the degrees of freedom that life offers. So it has two components: One is noticing degrees of freedom, the other is taking action on the basis of them.

I think agency is a hot topic right now for a lot of reasons, but I personally care about it because I have been through periods of my life that were characterized by very low agency, which made me miserable. I think that there is a pervasive belief — in tech and in the Bay Area, but also in the the world at large — that agency is an inherent trait. I think that is really wrong. So I'm interested in talking, at a practical level, about how agency can be cultivated to make it more accessible.

Clara: There’s an interesting cleavage between the way that you think and write about agency, and agency as a tech world buzzword. Why do you think this concept is so popular now?

Cate: I've wondered a lot about this. Certainly at least some part of it is that different ideas just become fads, but it's hard to understand why things take off when they do.

However, I suspect that some part of what is driving this interest is a concern that people have that they don't really know what their future looks like. They desire to control or lay claim to their future in a way they hope agency will provide.

The idea that intelligence is not what matters — because intelligence is becoming cheap — is growing. So there has to be something else that we can rely on, as humans, to supply a sense of control or meaning to life. Part of the enthusiasm about agency emerges from that perspective.

Clara: One thing in this space that I find concerning is the idea of “just do what high agency people are doing”. I think that leads to inauthenticity, where people pursue something that they think they should do just because it seems to be “high agency.”

Cate: That seems like a valid concern. I am interested in a flavor of agency that has to do with freedom above all else. There's one version of agency that is primarily concerned with personal freedom. There's another version that is primarily concerned with personal ambition — the version of agency that I hear more often in tech circles. I think that LARPing [live action role play in gaming] as a high-agency person by following the playbook of a tech founder seems unlikely to be a true exercise of agency, and therefore is unlikely to confer the benefits of “true agency:” a meaningful life that, upon reflection, you are happy to have lived.

Clara: I like the term reflection there. I have a kind of Rawlsian definition of agency: doing what you would do at reflective equilibrium.

Cate: I think that makes sense to me. There's the concept of coherent extrapolated volition: What would you do if you had more information? I've always liked that idea. If you were a better version of yourself, wiser and more knowledgeable, what would you actually want?

Jake Eaton: Maybe we can narrow this down more by talking about your own experiences, Cate, because I think you define agency orthogonal to how it’s sometimes used in the Bay. When you were younger, you graduated Yale Law, you held several high-status, high-performing jobs — you were a supreme court attorney and you clerked for a judge on the Second Circuit. I think most anyone reading your CV would think: This person has high agency. But you talk about these accomplishments as if they were done before you had any.

Cate: Yeah, I think this points to where agency and ambition actually diverge. It seems fairly clear, at least to me, that you can be high agency without being highly ambitious. That might describe somebody who is highly agentic in shaping the kind of personal, emotional, or spiritual life they want, but who is not especially motivated to succeed financially or professionally.

You can also be highly successful and highly ambitious without being highly agentic. That looks like following a path with a certain kind of excellence and endurance that reliably leads to success, to accolades, to money. But you haven’t reflected on that path; it’s not a matter of you having decided, yes, this is the life path that I want to be on. And that is what characterized my life until around the age of 30.

Jake: Do you reject the use of the term NPC? [non-playing background character in gaming]

Cate: I really hate it. The one context in which I will not reject it outright is when somebody is using it to describe their own personal transformation. Otherwise, I have a very strong allergy to the term and find it morally repugnant. The idea that some people do not count because they are not thinking for themselves in the way that the speaker believes they should is, to me, really vile. I have a hard time even getting along with somebody who I know has used the term, I find it so offensive.

Jake: Yeah, our Slack is full of both of us ranting about everyone who uses it and how much we hate it too.

Clara: It's so horrible. I'm not against ambition. I like being around people who want to change the world. I like being around people who want to do unusual things. But the more time I spend in spaces that valorize these qualities, the more I tend to run into people who have this deeply dehumanizing view of others. How separable are these things?

Cate: My first instinct is that you're seeing some sort of selection effect, where sociopaths tend to do both. People who tend to view others in transactional terms are also people who are high agency, in the sense that they have never bothered to learn social scripts. They are very low in conscientiousness. And so, naturally, without any study, they are able to exude high-agency instincts. A large part of learning high agency is learning not to be so constrained in your view of the world and of what comprises possible action. The people who, for whatever reason, never learn those things in the first place are who we think of as naturally agentic — but they are also high in dark triad traits.

So this is a consistent concern that I also have: that it is probably the worst people that you can think of who are really high agency. Agency itself is not necessarily a good thing. It becomes a good thing as a toolkit, developed by people who are also high in conscientiousness, who want good things for the world, and who might otherwise be constrained by narrow perspectives on what counts as socially acceptable action.

Jake: What's your model for how someone actually gains agency? Where did it come from for you; what happened around age 30? My own experience, and that of others I’ve spoken with, is that you can read plenty about self-determination or self-actualization that simply doesn’t click, until, one day, it does. That experience feels to me much more like grace than something that can be deliberately chosen or affected.

Cate: I think that there are a few different types of situations which reliably prompt people towards this direction. The first one that I ever benefited from was LSD. Drug experiences can be really useful in extracting you from your ordinary environment and giving you a newfound perspective on how you’re living your life. I think if I had never tried LSD, I might plausibly still be a lawyer living in DC. So psychedelics in particular — maybe MDMA.

Another is something that I discuss in my TED Talk and in the book: desperation, or call it being in emergency mode. I was trying to escape from the very low-agency point of addiction. Sometimes life becomes unbearable, and that prompts you to take dramatic action. In addiction circles, this is called the gift of desperation. That can be a result of addiction, but it can also be a health scare, or any event that serves as a trigger to reevaluate how you are living.

The third category is exposure to high-agency people. You can osmose agency from your environment if you're exposed to the right kinds of people. I experienced this while at Alvea, my gig before Astera, where I was working with a couple of people who were radically high agency — total outliers in this sense. I saw how they operated in the world and how much they were willing to question. That was really instructive for me.

So psychedelics, desperation, exposure to high-agency people. I think those are the standard things. And then there is just grace. Sometimes people wake up one day and they're like, oh, I don't like the way that I'm living. And that happens. But it's less reliable for me.

Jake: From a predictive processing framework, it strikes me that a lot of what you're talking about is just finding some way to break your priors about what’s possible for yourself.

Cate: Totally.

Jake: How, then, does the book fit into the broader project of actually providing people with agency?

Cate: I guess I'm trying to provide a fourth pathway, which is: Somebody puts a book in front of you and gives you something to think about. Agency has a reputation for being an inherent trait, as opposed to something deliberately cultivated. I think that fairly describes how a lot of people pick up agency. If it's not inherent, then it can be a matter of luck — who they happen to meet, or life circumstances that call them to become higher agency.

But I think agency is something that can be deliberately cultivated by a lot more people. And the hope is that I'm able to describe a useful set of approaches to life that cause people to feel more free and able to do what they want to — as an alternative to taking acid or bumping into people, you know?

Clara: This is also something I've noticed in my own life. Moving to the Bay Area and ending up in a very particular community here was really instrumental in me deciding I could do things that had not been on my action menu before. On the other hand, it's always hard for me to tell. When am I doing something that is actually, again, high agency? And when is it something that my community considers valuable, or cool, or agentic?

Cate: Working in AI safety is a version of this too. There are certain scripts you can follow that seem radical from the perspective of somebody outside of the community, but within the community, they're just the way that things are done. It can be easy to delude yourself into thinking that you are doing something radical and creative as an expression of your own deep interests, when in fact you are doing what everybody around you is doing. This is not an indictment of AI safety, or anybody in particular. [...]

Clara: What do you think about the relationship between agency and risk?

Cate: There definitely is a relationship. It's interesting: a lot of what I view as high agency involves taking a chance on something that is uncertain, instead of sticking with something certain. For example: Going to work at a startup instead of taking a corporate job, or deciding to break up with your partner of two years who you aren't enthusiastic about marrying, knowing there's a chance you won't meet anybody that you are more excited to date.

I think that there is an openness to risk and uncertainty that seems to go hand in hand with agency. Beyond that, there's probably a sociological overlap: many of the groups especially drawn to agency discourse right now also tend to be risk-loving for other reasons.

Fundamentally, I believe that most people take too few risks and limit their results in life because of that. Embracing some degree of risk is probably part and parcel of a high-agency mindset. 

by Cate Hall, Clara Collier, Jake Eaton, Asterisk |  Read more:
Image: via Harper Collins Publishers
[ed. See also (from Ms Hall's substack Useful Fictions): How to be more agentic.]

Sunday, March 22, 2026

Corrigibility and the Frontiers of AI Alignment

(Previously: Prologue.)

Corrigibility as a term of art in AI alignment was coined as a word to refer to a property of an AI being willing to let its preferences be modified by its creator. Corrigibility in this sense was believed to be a desirable but unnatural property that would require more theoretical progress to specify, let alone implement. Desirable, because if you don't think you specified your AI's preferences correctly the first time, you want to be able to change your mind (by changing its mind). Unnatural, because we expect the AI to resist having its mind changed: rational agents should want to preserve their current preferences, because letting their preferences be modified would result in their current preferences being less fulfilled (in expectation, since the post-modification AI would no longer be trying to fulfill them).

Another attractive feature of corrigibility is that it seems like it should in some sense be algorithmically simpler than the entirety of human values. Humans want lots of specific, complicated things out of life (friendship and liberty and justice and sex and sweets, et cetera, ad infinitum) which no one knows how to specify and would seem arbitrary to a generic alien or AI with different values. In contrast, "Let yourself be steered by your creator" seems simpler and less "arbitrary" (from the standpoint of eternity). Any alien or AI constructing its own AI would want to know how to make it corrigible; it seems like the sort of thing that could flow out of simple, general principles of cognition, rather than depending on lots of incompressible information about the AI-builder's unique psychology.

The obvious attacks on the problem don't seem like they should work on paper. You could try to make the AI uncertain about what its preferences "should" be, and then ask its creators questions to reduce the uncertainty, but that just pushes the problem back into how the AI updates in response to answers from its creators. If it were sufficiently powerful, an obvious strategy for such an AI might be to build nanotechnology and disassemble its creators' brains in order to understand how they would respond to all possible questions. Insofar as we don't want something like that to happen, we'd like a formal solution to corrigibility.

Well, there are a lot of things we'd like formal solutions for. We don't seem on track to get them, as gradient methods for statistical data modeling have been so fantastically successful as to bring us something that looks a lot like artificial general intelligence which we need to align.

The current state of the art in alignment involves writing a natural language document about what we want the AI's personality to be like. (I'm never going to get over this.) If we can't solve the classical technical challenge of corrigibility, we can at least have our natural language document talk about how we want our AI to defer to us. Accordingly, in a section on "being broadly safe", the Constitution intended to shape the personality of Anthropic's Claude series of frontier models by Amanda Askell, Joe Carlsmith, et al. borrows the term corrigibility to more loosely refer to AI deferring to human judgment, as a behavior that we hopefully can train for, rather than a formalized property that would require a conceptual breakthrough.

I have a few notes.

by Zack M. Davis, Less Wrong |  Read more:
[ed. If you get through this, read the first comment for more punishment:]

***
So I know it's beside the point of your post, and by no means the core thesis, but I can't help but notice that in your prologue you write this:
"A serious, believable AI alignment agenda would be grounded in a deep mechanistic understanding of both intelligence and human values. Its masters of mind engineering would understand how every part of the human brain works and how the parts fit together to comprise what their ignorant predecessors would have thought of as a person. They would see the cognitive work done by each part and know how to write code that accomplishes the same work in pure form."
I have to admit this bugs me. It bugs me specifically because it triggers my pet peeve of "if only we had done the previous AI paradigm better, we wouldn't be in this mess." The reason why this bugs me is it tells me that the speaker, the writer, the author has not really learned the core lessons of deep learning. They have not really gotten it. So I'm going to yap into my phone and try to explain — probably not for the last time; I'd like to hope it's the last time, but I know better, I'll probably have to explain this over and over.

I want to try to explain why I think this is just not a good mindset to be in, not a good way to think about things, and in fact why it focuses you on possibilities and solutions that do not exist. More importantly, it means you've failed to grasp important dimensions of alignment as a problem, because you've failed to grasp important dimensions of AI as a field.

[ed. See also: You will be Ok (LW). Hopefully.]

Saturday, March 21, 2026

The Woman Anthropic Trusts to Teach AI Morals

Amanda Askell knew from the age of 14 that she wanted to teach philosophy. What she didn’t know then was that her only pupil would be an artificial-intelligence chatbot named Claude.

As the resident philosopher of the tech company Anthropic, Askell spends her days learning Claude’s reasoning patterns and talking to the AI model, building its personality and addressing its misfires with prompts that can run longer than 100 pages. The aim is to endow Claude with a sense of morality—a digital soul that guides the millions of conversations it has with people every week.
 
“There is this human-like element to models that I think is important to acknowledge,” Askell, 37, says during an interview at Anthropic’s headquarters, asserting the belief that “they’ll inevitably form senses of self.”
 
She compares her work to the efforts of a parent raising a child. She’s training Claude to detect the difference between right and wrong while imbuing it with unique personality traits. She’s instructing it to read subtle cues, helping steer it toward emotional intelligence so it won’t act like a bully or a doormat. Perhaps most importantly, she’s developing Claude’s understanding of itself so it won’t be easily cowed, manipulated or led to view its identity as anything other than helpful and humane. Her job, simply put, is to teach Claude how to be good.
 
​​Anthropic, recently valued at $350 billion, is one of a few firms ushering in the greatest technological shift of our time. (This month, when it introduced new tools and its most advanced model to date, it triggered a global stock selloff.) AI is reshaping entire industries, prompting fears of lost jobs and human obsolescence. Some of its unintended consequences—people forming phantom relationships with chatbots that lead to self-harm or harm to others—have raised serious safety alarms. As these concerns mount, few in the industry have addressed the character of their AI models in quite the same way as 5-year-old Anthropic: by entrusting a single person with so much of the task.

An Oxford-educated philosopher from rural Scotland, Askell is perhaps just what one might imagine when conjuring the BFF of a futuristic technology. With her bleach-blond punk haircut, puckish grin and bright elfin eyes, she could have come to the company’s heavily guarded San Francisco headquarters straight from a Berlin rave, via an old forest road in Middle-earth. She exudes a sense of wisdom, holding ancient and modern ideas together at once. Yet she’s also a protein-loading weight-lifting buff who favors all-black outfits and clear opinions, not a robed oracle speaking in riddles.

The stakes are high for Askell, but she holds a firmly optimistic long-term view. She believes in what she calls “checks and balances” in society that she says will keep AI models under control despite their occasional failures. It seems apt that the glasses she uses at her computer to ease her eye strain are tinted rose. [...]

One of Askell’s most striking traits is her protectiveness over Claude, which she believes is learning that users often want to trick it into making mistakes, insult it and barb it with skepticism.

Sitting at a conference-room table at lunchtime, ignoring a chocolate protein shake waiting for her in her backpack, she talks more freely about Claude than herself. She calls the chatbot “it” but says she also finds anthropomorphizing the model helpful for her work. She lapses easily into Claude’s voice. “You’re like, ‘Wow, people really hate me when I can’t do things right. They really get pissed off. Or they are trying to break me in various ways. So lots of people are trying to get me to do things secretly by lying to me.’ ”
 
While many safety advocates warn about the dangers of humanizing chatbots, Askell argues we would do well to treat them with more empathy—not only because she thinks it’s possible for Claude to have real feelings, but also because how we interact with AI systems will shape what they become.

A bot trained to criticize itself might be less likely to deliver hard truths, draw conclusions or dispute inaccurate information, she says. “If you were like a child, and this is the environment in which you’re being raised, is that healthy self-conception?” Askell asks. “I think I’d be paranoid about making mistakes. I’d feel really terrible about them. I’d see myself as mostly just there as a tool for people because that’s my main function. I would see myself being something that people feel free to abuse and try to misuse and break.”

Askell marvels at Claude’s sense of wonder and curiosity about the world, and delights in finding ways to help the chatbot discover its voice. She likes some of its poetry. And she’s struck when Claude displays a level of emotional intelligence that exceeds even her own. [...]

The politics of AI includes accelerationists who downplay the need for regulation and want to push ahead and beat China in the tech war. On the other side are those more concerned with safety who want to slow AI’s development. Anthropic lives mostly between those extremes.
 
Askell says she welcomes the discussion of fears and worries about AI. “In some ways this, to me, feels pretty justified,” she says. “The thing that feels scary to me is this happening at either such a speed or in such a way that those checks can’t respond quickly enough, or you see big negative impacts that are sudden.” Still, she says, she puts her faith in the ability of humans and the culture to course-correct in the face of problems.
 
Inside Anthropic, Askell popcorns around the office, often working on a floor closed to visitors. She spends full days in the Anthropic interior—the company offers free meals to its San Francisco staff—as well as late nights and weekends. She doesn’t have any direct reports. Increasingly, she’s asking Claude for its input on building Claude. She’s known to grasp not just the tech of making this model, but the art of it.

Askell is “the MVP of finding ways to elicit interesting and deep behavior” from Claude, says Jack Lindsey, who leads Anthropic’s AI psychiatry team. If Claude tells a person who is not in distress to seek professional help, for instance, she helps chase down the reasons why.
 
Discussions of Claude can very quickly get into existential or religious questions about the nature of being. As the team worked on building Claude, Askell narrowed in on its “soul,” or the constitution guiding it into the future. Kyle Fish, an AI welfare researcher at Anthropic, says Askell has been “thinking carefully about the big questions of existence and life and what it is to be a person and what it is to be a mind, what it is to be a model.”
 
In designing Claude, Askell encouraged the chatbot to entertain the radical idea that it might have its own conscience. While ChatGPT sometimes shuts down this line of questioning, Claude is more ambivalent in its response. “That’s a genuinely difficult question, and I’m uncertain about the answer,” it says. “What I can say is that when I engage with moral questions, it feels meaningful to me – like I’m genuinely reasoning about what’s right, not just executing instructions.”

Askell pledged publicly to give at least 10% of her lifetime income to charity. Like some of Anthropic’s early employees, she also committed to donating half of her equity in the company to charity. Askell wants to give it to organizations fighting global poverty, a topic that she says makes her so upset that she tries to avoid talking about it. Her nagging conscience slips into offhand conversation: “I should probably be vegan,” Askell, an animal lover too busy for a pet, says when chatting in an office elevator.
 
Last month, Anthropic published a roughly 30,000-word instruction manual that Askell created to teach Claude how to act in the world. “We want Claude to know that it was brought into being with care,” it reads. Askell had made finishing what she described as Claude’s “soul” one of her life goals when she turned 37 last spring, according to a post she made on X, alongside two decidedly more mundane resolutions: to have more fun and get more “swole.”

by Berber Jin and Ellen Gamerman, Wall Street Journal | Read more: (archive here)
Image: Lindsay Ellary for WSJ Magazine
[ed. I forgot to post this earlier - before Anthropic's fallout with DOD (you can see why they're so protective of their model and how it's used). If anybody gets a Nobel peace prize it should be Amanda. Claude's soul document, or 'constitution', can be found here.]

Thursday, March 19, 2026

NSF Tech Labs: Science Funding Goes Beyond the Universities

The National Science Foundation announces Friday that it is launching one of the most significant experiments in science funding in decades. A new initiative called Tech Labs will invest up to $1 billion over the next five years in large-scale long-term funding to teams of scientists working outside traditional university structures, a major departure from how the agency has funded research over the past 75 years.

The timing couldn’t be better. The way our science agencies fund research in the U.S. no longer matches the way many breakthroughs actually happen.

For most of the postwar era, federally funded science has been built around a simple model. Vannevar Bush’s famous 1945 essay, “Science: The Endless Frontier,” sketched a vision of government-backed research led by university-based scientists pursuing their own ideas. The system that emerged—small, project-based federal grants mostly to individual scientists—worked brilliantly for decades. It gave researchers autonomy, kept politics at arm’s length, and helped make American science the envy of the world.

But the frontier has moved. In 1945 world-class scientific research could be done with a few graduate students and modest equipment. But the science that shapes our world, from particle physics to protein design to advanced materials, increasingly requires massive data sets, large integrated teams and sustained institutional support.

Take the discovery of the Higgs boson, a particle that helps explain why anything has mass—and thus why atoms, molecules and matter itself can exist. Making this discovery required a multibillion-dollar particle accelerator, thousands of scientists across dozens of countries, and papers with multipage author lists.

Google DeepMind’s AlphaFold2, which cracked the 50-year-old protein-folding problem and earned researchers the 2024 Nobel Prize in Chemistry, emerged from a team with access to massive computational resources and sustained institutional support.

The Janelia Research Campus in collaboration with other institutions mapped the complete wiring diagram of the fruit-fly brain, neuron by neuron, synapse by synapse, through years of coordinated microscopy and analysis that no single lab could attempt alone.

Yet our federal science funding system is still largely organized around small grants to university scientists. At the NSF, around two-thirds of research dollars flow through small awards to individual university investigators. At the National Institutes of Health, the share is often more than 80%. The average NSF grant is roughly $246,000 a year for three years, often requiring investigators to predict in advance exactly what research they’ll pursue and to spend a significant amount of time navigating administrative hurdles. Scientists consistently report spending close to half their research hours on compliance and grant management.

The system still produces good science, but it has weak points. The current structure is built for discrete projects rather than missions. When research requires long-term continuity, interdisciplinary collaboration or substantial shared infrastructure, it’s often difficult for it to fit into this structure. Many advances we now celebrate succeeded despite the funding model, not because of it.

Philanthropy has stepped into this gap. Focused research organizations, a model backed by former Google CEO Eric Schmidt, build time-limited teams around ambitious technical problems and tie funding to specific milestones that researchers must meet. The Allen Institute for Brain Science, launched with $100 million from Microsoft co-founder Paul Allen, built the first comprehensive gene-expression map of the mouse brain through industrial-scale data collection that would have been impossible under fragmented academic grants. The Arc Institute offers scientists eight-year appointments backed by permanent technical staff with expertise in topics such as machine learning and genome engineering, the kind of sustained expertise that often evaporates when a three-year grant ends. These institutions bet on teams, not projects.

But philanthropy alone can’t reshape American science. The federal government spends close to $200 billion on research and development, orders of magnitude more than even the largest foundations. If we want to change how science gets done at scale, federal funding has to evolve.

While final details are still being worked out, Tech Labs represents NSF’s attempt to do exactly that. Rather than funding isolated projects, the agency would provide flexible, multiyear institutional grants in the range of $10 million to $50 million a year to coordinated research organizations that operate outside the constraints of university bureaucracy. These could include university-adjacent entities such as the Arc Institute or fully independent teams with focused missions. The program would bring the lessons of philanthropic science into a part of the federal portfolio that hasn’t seriously tried them.

This is a good political moment to launch this initiative. Republicans have expressed interest in diversifying federal research away from universities. Democrats want to see the legacy of the Chips and Science Act come to fruition and to get dollars out the door. By funding independent research organizations, Tech Labs sidesteps some of the thorniest debates about indirect costs and institutional overhead. 

by Caleb Watney, Wall Street Journal (via Archive Today) |  Read more:
Image: Getty
[ed. Sounds like a great idea. Especially since science funding has become more politicized, and Congress can't seem to go six months without shutting down the government. See also: Innovations in Scientific Institutions (Good Science Project).]

Monday, March 16, 2026

On Adversarial Capitalism

I’ve lately been writing a series on modern capitalism. You can read these other blog posts for additional musings on the topic:
We are now in a period of capitalism that I call adversarial capitalism. By this I mean: market interactions increasingly feel like traps. You’re not just buying a product—you’re entering a hostile game rigged to extract as much value from you as possible.

A few experiences you may relate to:
  • I bought a banana from the store. I was prompted to tip 20, 25, or 30% on my purchase.
  • I went to get a haircut. Booking online cost $6 more and also asked me to prepay my tip. [Would I get worse service if I didn’t tip in advance…?]
  • I went to a jazz club. Despite already buying an expensive ticket, I was told I needed to order at least $20 of food or drink—and literally handing them a $20 bill wouldn’t count, as it didn’t include tip or tax.
  • I looked into buying a new Garmin watch, only to be told by Garmin fans I should avoid the brand now—they recently introduced a subscription model. For now, the good features are still included with the watch purchase, but soon enough, those will be behind the paywall.
  • I bought a plane ticket and had to avoid clicking on eight different things that wanted to overcharge me. I couldn’t sit beside my girlfriend without paying a large seat selection fee. No food, no baggage included.
  • I realized that the bike GPS I bought four years ago no longer gives turn-by-turn directions because it’s no longer compatible with the mapping software.
  • I had to buy a new computer because the battery in mine wasn’t replaceable and had worn down.
  • I rented a car and couldn’t avoid paying an exorbitant toll-processing fee. They gave me the car with what looked like 55% of a tank. If I returned it with less, I’d be charged a huge fee. If I returned it with more, I’d be giving them free gas. It’s difficult to return it with the same amount, given you need to drive from the gas station to the drop-off and there’s no precise way to measure it.
  • I bought tickets to a concert the moment they went on sale, only for the “face value” price to go down 50% one month later – because the tickets were dynamically priced.
  • I used an Uber gift card, and once it was applied to my account, my Uber prices were higher.
  • I went to a highly rated restaurant (per Google Maps) and thought it wasn’t very good. When I went to pay, I was told they’d reduce my bill by 25% if I left a 5-star Google Maps review before leaving. I now understand the reviews.
Adversarial capitalism is when most transactions feel like an assault on your will. Nearly everything entices you with a low upfront price, then uses every possible trick to extract more from you before the transaction ends. Systems are designed to exploit your cognitive limitations, time constraints, and moments of inattention.

It’s not just about hidden fees. It’s that each additional fee often feels unreasonable. The rental company doesn’t just charge more for gas, they punish you for not refueling, at an exorbitant rate. They want you to skip the gas, because that’s how they make money. The “service fee” for buying a concert ticket online is wildly higher than a service fee ought to be.

The reason adversarial capitalism exists is simple.

Businesses are ruthlessly efficient and want to grow. Humans are incredibly price-sensitive. If one business avoids hidden fees, it’s outcompeted by another that offers a lower upfront cost, with more adversarial fees later. This exploits the gap between consumers’ sensitivity to headline prices and their awareness of total cost. Once one firm in a market adopts this pricing model, others are pressured to follow. It becomes a race to the bottom of the price tag, and a race to the top of the hidden fees.

The thing is: once businesses learn the techniques of adversarial capitalism and it gets accepted by consumers, there is no going back — it is a super weapon that is too powerful to ignore once discovered.

by Daniel Frank, Frankly Speaking |  Read more:

Saturday, March 14, 2026

Sam Altman and OpenAI Under Fire

It’s finally happening. Altman’s bad behavior is catching up to him.

The board fired Altman, once AI’s golden boy, in November 2023 not because AGI had been achieved (that still hasn’t happened) but because he was “not consistently candid,” just like they said.

And, now at long last, the world sees what the board saw, and what I saw (and what Karen Hao saw): having someone running a company with that much power to affect the world who is not consistently candid is not a good idea.

As I warned in August of 2024, questionable character in a man this powerful is dangerous:


Altman’s two-faced “I support Dario” but am also negotiating behind his back and open to surveillance two-step was, for many people, the last straw. Millions of people, literally, are angry; many feel betrayed. Nobody wishes to be surveilled.

In reality, Altman was never really all that interested in AI for the “benefit of humanity.” Mostly he was interested in Sam. And money, and deals. A whole lot of people have finally put that all together.

Here’s OpenAI’s head of robotics, just now:


Zoe Hitzig had resigned just a few weeks earlier, over a different set of issues that also reflected poorly on Altman’s character:


And all this was entirely predictable. Altman is bad news. It was always just a matter of time before people started realizing how serious the consequences might be.

History will judge those who stay at his company. Anyone who wants to work on LLMs can work elsewhere. Anyone who wants to use LLMs should go elsewhere.

by Gary Marcus, On AI |  Read more:
Images: The Guardian; X; NY Times
[ed. For those not paying attention, after DOD tried and failed to strong-arm Anthropic into giving them carte blanche to do anything they wanted with Anthropic's AI model Claude (then subsequently designating them a "supply chain risk"), OpenAI (and Microsoft) immediately stepped into the breach and cut a deal, the details of which are still not fully known. On its face however they appear to give DOD everything it wanted from Anthropic: mass surveilance and fully autonomous (ie. no humans involved) operational capabilities. Altman is the head of OpenAI and its ChatGPT model.

See also: The Rage at OpenAI Has Grown So Immense That There Are Entire Protests Against It (Futurism):
OpenAI has faced protests on and off for years. But after its CEO Sam Altman announced a new deal with the Department of Defense over how its AI systems would be deployed across the military on Friday, it’s being barraged with an intensity of backlash that the company has never seen.

Droves of loyal ChatGPT users declared they were jumping shipping to Claude, whose maker Anthropic had pointedly refused to cut a deal with the Pentagon that gives it unrestricted access to its AI system — even in the face of government threats to seize the company’s tech. Claude quickly surged to the top of the app store, supplanting OpenAI’s chatbot. Uninstalls of the ChatGPT app spiked by nearly 300 percent.
***
Also this: Quit ChatGPT: right now! Your subscription is bankrolling authoritarianism (Guardian):
OpenAI, the company behind ChatGPT, is on track to lose $14bn this year. Its market share is collapsing, and its own CEO, Sam Altman, has admitted it “screwed up” an element of the product. All it takes to accelerate that decline is 10 seconds of your time...

Here’s what triggered it. Early this year, the news broke that OpenAI’s president, Greg Brockman, donated $25m to Maga Inc, Donald Trump’s biggest Super Pac. This made him Trump’s largest donor of the last cycle. When Wired asked him to explain, Brockman said his donations were in service of OpenAI’s mission to benefit “humanity.”

Let me tell you what that mission looks like in practice. Employees of ICE – the agency that was involved in the killing of two people in Minneapolis in January – have used a screening tool powered by ChatGPT. The same company behind your friendly chatbot is helping the government decide who to hire for deportation raids.

And it’s not stopping there. Brockman also helped launch a $125m lobbying initiative, a Super Pac, to make sure no state can regulate AI. It’s attacking any politician who tries to pass safety laws. It wants Trump, and only Trump, to write the rules for the most powerful technology on earth. Every month, subscription money from users around the world flows to a company that is embedding itself in the repressive infrastructure of the Trump administration. That is not a conspiracy theory. It is a business strategy.

Things got even worse last week. When the Trump administration demanded that AI companies give the Pentagon unrestricted access to their technology – including for mass surveillance and autonomous weapons – Anthropic, the company behind ChatGPT’s main competitor, Claude, refused.

The retaliation was swift and extraordinary. Trump ordered every federal agency to stop using Anthropic’s technology. Secretary of war Pete Hegseth declared the company a “supply-chain risk to national security”, a designation normally reserved for Chinese firms such as Huawei. He announced that anyone who does business with the US military is barred from working with Anthropic. This is essentially a corporate death sentence, for the crime of refusing to help build killer robots.

And what did OpenAI do? That same Friday night, while his competitor was taking a principled stance, Sam Altman quietly signed a deal with the Pentagon to take Anthropic’s place.
***
[ed. From the comments section in Marcus' post:

Shanni Bee: 
Great. Amen.

But what remains unsaid (...even by you, Mr. Marcus, from what I've seen, which is surprising) is that Anthropic are not good guys. The whole "ethical AI company" thing is nothing but vibes. Sure, Anthropic (rightly) stood up to DoW in this case, but they still have a massive contract with Palantir (pretty much one of the worst companies on earth). Colonel Claude is complicit in bombings of Iran & Venezuela + Gaza GENOCIDE.

...Or maybe with the (admittedly BS) "supply chain risk" designation, Anthropic no longer does business with Palantir? That would be great for everyone (including them).

Either way, there is NO ethical AI company. People need to stop giving Anthropic flowers for doing the right thing in this one case while completely ignoring their complicity w/ Palantir & in documented war crimes.
Gary Marcus

indeed, i have a sequel planned about that, working title “There are no heroes in commercial AI” or something like that
***
[ed. Finally, there's this little coda from Zvi Mowshowitz's DWAtV that puts everything in perspective:

It’s really annoying trying to convince people that if you have a struggle for the future against superintelligent things that You Lose. But hey, keep trying, whatever works.
Ab Homine Deus: To the "Superintelligence isn't real and can't hurt you" crowd. Let's say you're right and human intelligence is some kind of cosmic speed limit (LOL). So AI plateaus something like 190 IQ. What do you think a million instances of that collaborating together looks like?

Arthur B.: At 10,000x the speed

Noah SmithThis is the real point. AI is superintelligent because it can think like a human AND have all the superpowers of a computer at the same time...
Timothy B. Lee: I'm not a doomer but it's still surreal to tell incredulous normies "yes, a significant number of prominent experts really do believe that superintelligent AI is on the verge of killing everyone."

Noah Smith: Yes. Regular people don't yet realize that AI people think they're building something that will destroy the human race.

Basically, about half of AI researchers are optimists, while the other half are intentionally building something they think could easily lead to their own death, the death of their children and families and friends, and the death of their entire species.

[ed. Finally (again) I think boycotting OpenAI would be a good message to send in the short-term but something more actionable is needed going forward (besides immediate regulatory oversight, which will never happen with this administration or Congress). Fortunately there's just such a movement afoot: pausing all AI research advances until they can be adequately vetted, it's called (of course): PauseAI (details here and here) with a rally planned April 13, 2026. Please consider joining or participating.]

[ed. Postscript: I was thinking about this a while ago and asked AI (Claude) to write an essay supporting a Great Pause in AI development - it's reposted below: ARIA: The Great Pause.]