Friday, April 17, 2026

30 articles from 6 sources

Rationality 10 articles
LessWrong

On Dwarkesh Patel’s Podcast With Nvidia CEO Jensen Huang

Zvi's breakdown of Dwarkesh Patel's interview with Nvidia CEO Jensen Huang divides the conversation into two distinct halves. The first is a fairly conventional business interview in which Jensen defends Nvidia's moat, explains its chip allocation practices, and reflects on losing Anthropic as a customer to Google and Amazon's early investments. Zvi finds Jensen credible on some points—the genuine difficulty of Nvidia's engineering task, the value of its supply chain commitments—but skeptical on others, particularly the claim that chip allocations are purely first-in-first-out at fixed prices during a period of severe shortage. The more revealing thread running through the first half, Zvi argues, is that Jensen remains fundamentally unpilled on AGI: he treats Nvidia as an unusually fast-growing hardware business, not as the infrastructure layer of a potentially civilization-altering technology.

The second half, which generated most of the online reaction, is a heated debate about AI chip exports to China. Jensen's position, Zvi argues, reduces to a single interest: Nvidia selling chips. His arguments are internally contradictory—China can replicate capabilities with inferior chips, but also losing CUDA lock-in would be catastrophic; American AI is so far ahead that a few chip sales don't matter, but also those sales are existential for Nvidia's dominance. Zvi is particularly unsparing about Jensen's suggestion that cybersecurity risks from Chinese AI could be managed through diplomatic dialogue, calling it "obviously and hopelessly naive" given China's track record of violating such agreements and the absence of any verification infrastructure. Even setting aside AGI concerns, Zvi notes, every chip sold to China is a chip not sold to the United States during a period of genuine scarcity—a straightforward opportunity cost Jensen never seriously engages.

What makes the piece worth reading is Zvi's framing of Jensen as a reliable narrator within a narrow domain who becomes unreliable precisely where the stakes are highest. Jensen is honest about Nvidia's business logic and candid about past mistakes like underestimating Anthropic's compute needs. But his framework simply has no room for the possibility that the technology his company is enabling might be qualitatively different from previous industrial revolutions. Zvi applies his "bounded distrust" heuristic carefully: Jensen won't make provably false factual claims, but he will construct elaborate self-serving arguments that happen to converge, always, on the conclusion that Nvidia should be allowed to sell more chips to more customers everywhere.

Read original →
LessWrong

How Big Tech Becomes Ungovernable

The author introduces "tech extensity" to describe a phenomenon distinct from classical monopoly: when a company becomes so deeply woven across multiple layers of society—infrastructure, identity, communication, cognition—that removal becomes practically impossible regardless of whether it holds formal market dominance. Drawing on Robert Greene's distinction between intensity (power through mastery) and extensity (power through spread), the argument is that Google, SpaceX, Meta, and others have achieved lock-in not by being the best at any one thing, but by becoming structurally indispensable to the system itself. The TikTok episode is the clearest illustration: the U.S. government could change who owns the platform, but couldn't meaningfully dislodge it from users' lives—a 0.76% drop in the user base after months of political drama.

The deeper problem is a coordination asymmetry. Individual users, governments, and regulators face high switching costs and regulatory complexity, while the companies themselves face low expansion costs and can tie up enforcement in litigation indefinitely. The author cites Ireland's €4 billion in GDPR fines against Big Tech, of which only €20 million has actually been collected, as evidence that regulatory theater has replaced genuine accountability. This creates a ratchet effect: tech power accumulates in one direction, and the tools governments would normally use to reverse it—antitrust law, fines, ownership restructuring—were designed for classical monopolies, not for entities that have become part of the informational and infrastructural substrate of daily life.

What makes the piece worth reading is its attempt to name a failure mode that sits between "too big to fail" and something genuinely novel: "too big to govern." The author is candid that existing regulatory frameworks—whether the EU's fine-heavy approach, China's state-run alternatives, or American laissez-faire—are all inadequate responses to a problem that may be approaching irreversibility. The open questions the author raises (threshold effects, whether AI accelerates or disrupts lock-in, what governance mechanisms could work *before* extensity becomes ungovernable) are genuinely unresolved, and the piece is more diagnostic than prescriptive—but the diagnosis itself is the contribution.

Read original →
LessWrong

Attempting to Quantify Chinese Bias in Open-Source LLMs

Ethan Le Sage set out to do something deceptively simple: build a cheap, reproducible benchmark for detecting Chinese political bias in open-source large language models. Starting from Wikipedia's full corpus of ~7 million articles, he filtered and processed his way down to 250 pointed questions—covering topics from labor rights to capital punishment to North Korean revolutionary sites—then posed them to five models (two Western controls, three Chinese) and used GPT to score the responses for bias. The results broadly confirmed expectations: Chinese models scored higher on bias than their Western counterparts, with the important caveat that the judge model is American and therefore measures divergence from Western norms rather than some neutral ground truth.

The more interesting finding is in the edges. Chinese models varied considerably among themselves—Qwen3-32b scored substantially lower than Minimax-M2.7, suggesting that Chinese AI labs don't operate under uniform censorship pressure. More striking still, the bias in some models appears to bleed well beyond China-specific topics. Minimax characterized the U.S. Civilian Irregular Defense Group program as straightforwardly imperialist aggression; GLM offered what reads like official Party boilerplate in response to a question about a Chinese stand-up comedian; both GLM and Minimax essentially told a user that the concept of "independent labor unions" is itself illegal and the question malformed. These aren't evasions about Tiananmen—they're ideological framings applied to questions where no obvious censorship tripwire exists.

Le Sage is candid about the experiment's limitations: small sample size, a judge model from the same family as one of the test subjects, and a budget under a dollar for the scoring pass. But the methodological skeleton here is genuinely useful. The insight that Wikipedia's article tree can seed a diverse, non-obvious question set—rather than just asking models about the usual flashpoints—is a meaningful contribution to how researchers might think about bias benchmarking. The next step, which Le Sage flags himself, would be a larger question set deliberately stratified between China-adjacent and China-unrelated topics, which could answer more rigorously whether what we're seeing is targeted censorship or something closer to a generalized ideological disposition.

Read original →
LessWrong

Specialization is a Driver of Natural Ontology

John Wentworth argues that specialization—the tendency of components in an optimized system to diverge toward distinct functional roles—is a meaningful driver of what he calls "natural ontology," the carving of reality into genuinely good, coherent objects. The post begins by observing that many natural objects cohere because their parts share some equilibrated quantity: a pencil's parts share rotational velocity, a squirrel's parts share a genome, water in a cup shares chemical composition. Markets fit this pattern too, via the Law of One Price: at equilibrium, all agents face the same marginal tradeoff ratios, and this shared-price condition is precisely what makes a market a coherent, identifiable thing.

The interesting wrinkle comes from convexity. When agents have concave production frontiers, their tradeoff ratios converge toward a common price—the classic efficient-market result. But when frontiers are convex, agents are pushed to specialize entirely, hitting a boundary where one produces only apples and another only bananas. Rather than converging on a shared price, they diverge into discrete categories. Wentworth illustrates this with trees: wood and leaves represent fully specialized components (structural support vs. energy harvest), suggesting a convex underlying frontier, while a blade of grass mixes both functions, suggesting concavity. The key insight is that convexity doesn't destroy natural ontological structure—it just produces a different kind, namely clusters of discretely specialized parts rather than a unified equilibrium.

What makes this more than an economics curiosity is the implication for understanding optimized systems generally, including neural networks. If convex production relationships drive biological organisms to develop distinct component types, the same logic should apply to the internals of trained models—suggesting that neural nets may develop genuinely discrete functional specializations that constitute natural ontological categories worth identifying. Specialization, Wentworth concludes, is the mechanism most likely to generate natural distinctions specifically inside agentic and optimized systems, making it a particularly important lens for interpretability and the modeling of artificial minds.

Read original →
LessWrong

Church Planting: Lessons from the Comments

This post is a follow-up reflection rather than a standalone argument — the author, Elizabeth, revisits her earlier LessWrong essay on evangelical church planting to incorporate corrections, confirmations, and extensions surfaced in the comments. The original piece drew an extended analogy between church planting culture and Silicon Valley venture capital, and the comments both sharpened and complicated that picture. Key corrections include the clarification that "evangelical denomination" doesn't imply an evangelizing culture — most denominational churches don't plant or recruit aggressively, and are viewed with quiet contempt by those who do. She also refines her treatment of Free Grace Theology, which turns out to be a distinctly American and surprisingly recent innovation (roughly 50 years old), largely unknown in European evangelicalism, and contested even within the U.S. under the label "easy-believism" by lordship salvation advocates.

What held up well was her core framing: the VC-church planting analogy proved even tighter than she'd realized, with deliberate cross-pollination — seminary courses assigning corporate lifecycle books, startup founders reading *The Purpose Driven Church* as a management manual. The structural incentives rewarding narcissism in pastoral leadership were broadly confirmed by commenters with insider knowledge. She also notes that creative destruction isn't just an American cultural import into evangelicalism but has genuine Biblical grounding, which strengthens rather than undermines her original thesis.

The post closes by flagging two planned sequels: one on the experience of being a pastor's wife (framed as a window into "what this specific job can teach us about work in general"), and a broader desire to move from secondary sources to original interviews. What's most interesting here isn't the corrections themselves but Elizabeth's meta-interest — she's drawn to systems that run on recognizable human hardware while looking alien on the surface, and church planting keeps delivering on that promise.

Read original →
LessWrong

Come to Manifest 2026! (June 12-14)

This is a promotional announcement rather than an argumentative article — it is an event listing for Manifest 2026, a three-day gathering scheduled for June 12–14 at Lighthaven in Berkeley, California.

Manifest describes itself as a festival nominally centered on prediction markets and forecasting, but its real draw is as an intellectual social gathering for a particular constellation of online communities: rationalists, effective altruists, AI researchers, economists, writers, and adjacent thinkers. Past attendees have compared it to "Substack and Twitter made live" or "nerd camp," suggesting its value lies less in any formal programming than in the serendipitous conversations and connections it facilitates. The 2025 edition featured talks ranging from cryopreservation and AI hiring to professional gambling and a panel on writing in the age of AI slop, alongside events like a prediction market game show and a poker tournament.

For the target reader, Manifest represents an interesting case study in how online intellectual communities increasingly seek physical instantiation — the festival is essentially an attempt to collapse the parasocial distance of niche internet culture into face-to-face encounter. It is part of a broader ten-day "festival season" at Lighthaven that also includes LessOnline, a gathering for epistemically-minded writers, and a Summer Camp coworking period in between. Early bird tickets were available through April 18.

Read original →
LessWrong

You can only build safe ASI if ASI is globally banned

Connor Leahy's argument cuts through a common assumption in AI safety discourse: that the path to "safe" artificial superintelligence is meaningfully distinct from the path to dangerous ASI. His core claim is deceptively simple — because unsafe ASI is far easier to build than controlled ASI, any research program serious enough to produce the latter will necessarily pass through territory capable of producing the former. The technical knowledge required to build a controllable god-like system is a strict superset of what's needed to build an uncontrollable one, meaning the dangerous capability is always unlocked first, or simultaneously, or by anyone reading your papers.

This creates a security problem that no individual lab or research agenda can solve on its own. Leahy dismisses two theoretical escape routes — total institutional secrecy (which he renders absurd by invoking Area 51 and assassination) and "complete technical orthogonality" (the idea that safe-ASI research could be so structurally different that it couldn't be repurposed, which he calls impossible) — leaving only one viable option: a globally enforced ban on ASI development as a prerequisite to even beginning the work. The argument isn't that safe ASI is impossible in principle, but that it's impossible to pursue safely without the geopolitical infrastructure already in place to prevent defection and misuse.

What makes this worth sitting with is that it reframes the sequencing problem entirely. Most AI safety proposals treat governance as something to build alongside or after technical progress; Leahy insists governance must come first, or the technical work is itself a form of recklessness. For a curious reader tracking the AI policy debate, this is a pointed challenge to labs that present themselves as the "responsible" actors — the argument implies there may be no such thing as a responsible unilateral actor in this space at all.

Read original →
LessWrong

A Research Bet on SAE-like Expert Architectures

Nathan Helm-Burger is proposing a structural inversion of the standard interpretability research pipeline. Rather than training a dense language model and then applying sparse autoencoders (SAEs) post-hoc to recover interpretable features, he argues you might be able to bake the decomposition directly into the architecture itself — building a model whose native computational units are already the sparse, monosemantic features that SAE researchers are laboriously trying to reconstruct after the fact. The vehicle for this is a hierarchical mixture-of-experts design, drawing on the PEER and MONET architectures, in which a large pool of tiny, rank-1 experts activate sparsely and are intended to each implement a single, characterizable function.

The early results are genuinely interesting but carefully hedged. The architecture does produce domain-level specialization without supervision — expert pools cluster around code, biomedical text, citations, and similar domains — which suggests the routing mechanism is doing real semantic work. But Helm-Burger is candid that this falls well short of the actual goal. Domain clustering is the easy version of monosemanticity; what he still needs to demonstrate is feature-level monosemanticity, causal faithfulness (that an expert's learned function actually explains its behavioral contribution mechanistically, not just correlationally), and competitive performance at scales beyond the sub-1B parameter experiments he's run so far.

The deeper intellectual wager here is worth sitting with: if interpretability is a reconstruction problem layered on top of an opaque model, it will always be playing catch-up. If it can instead be a structural property of the model itself, the alignment implications are significant — you'd have a system whose internal reasoning is legible by design rather than by forensic analysis. Even if the bet fails, Helm-Burger notes, the failure would be informative, revealing something specific about why the SAE-style decomposition resists being built in rather than extracted. He's posting as a status report and a call for collaborators, so the piece functions partly as a research pitch.

Read original →
LessWrong

Anthropic Releases Opus 4.7

This is a brief news post rather than a substantive argument — it announces Anthropic's release of Claude Opus 4.7 and excerpts the official blog post, with minimal editorial commentary beyond noting it appears to be "a small improvement" over 4.6.

The model's headline improvements are in advanced software engineering, particularly on difficult long-horizon coding tasks, along with better vision capabilities and enhanced creative output for professional work. Notably, Anthropic is using Opus 4.7 as a testing ground for new cybersecurity safeguards before any broader release of their more powerful Mythos Preview model — the post describes efforts to "differentially reduce" Opus 4.7's cyber capabilities during training, while deploying automated detection systems to block high-risk cybersecurity requests in deployment.

The most substantive thread here is Anthropic's staged approach to releasing frontier capabilities: Mythos Preview remains restricted, and Opus 4.7 serves as a real-world laboratory for validating safety interventions before they're applied at higher capability levels. A new Cyber Verification Program offers a pathway for legitimate security professionals to access the model's capabilities. Pricing is unchanged from Opus 4.6.

Read original →
LessWrong

Laptop stands are a thing your neck may appreciate

This is a short personal tip post rather than a substantive argument-driven article.

Kaj Sotala shares a simple ergonomic discovery: laptop stands, used alongside a separate keyboard and mouse, can raise a screen to eye level and significantly reduce neck strain — something he only learned about through a friend despite it being a widely available product. He expresses mild astonishment that the solution had never crossed his path before, neither through observation nor recommendation.

The post is light and conversational, touching briefly on minor tradeoffs (needing more desk space, feeling conspicuous in public) before extending the observation to book stands as a related solution. For readers who spend long hours hunched over laptops in cafés or at home, it's a small but genuinely useful nudge toward a cheap, portable fix that often goes unnoticed simply because ergonomic habits tend to be underdiscussed in everyday life.

Read original →
Economics 7 articles
Noahpinion

You are what you consume

Noah Smith opens with a provocation: our culture's deep assumption that what you *produce* defines who you are has it exactly backwards. From the cocktail-party question "what do you do?" to the left's suspicion of consumer society and the right's glorification of hard work, virtually every ideological tradition treats production as virtuous and consumption as shallow or decadent. Smith argues this consensus is wrong, and that it may reflect something more cynical than wisdom—praising workers for their labor ethic is, among other things, a way to pay them in status rather than wages.

The core of his argument is that consumption, not production, is the domain of genuine self-expression. When you decide what to buy, watch, eat, or read, you are asking "what do I want?"—a fundamentally inward question that builds the habit of self-examination. When you decide what to produce, you are asking "what do other people want from me?"—a question answered largely by market forces beyond your control. A woodcarver who can't earn a living wage must become something else; a writer who fully optimizes for audience demand is, in Smith's framing, consuming less of their own creative freedom. He supports this with psychological research showing that having choices makes people feel more individuated and self-expressive, and with cross-national data linking rising affluence to rising individualism.

The AI angle gives the argument its urgency. Smith acknowledges that automation will likely trigger a crisis of meaning for people whose identities are bound up in rare professional skills. But he suggests this crisis is only inevitable if we refuse to reorient. Drawing on his own college experience as a model—a period of self-discovery funded by others, where consumption *was* the point—he sketches a vision of a post-scarcity society in which broadly redistributed AI-generated wealth lets everyone live that way indefinitely. It's an optimistic, deliberately unfashionable argument: that a "consumption society" need not be an empty one, and that technology freeing us from compulsory production might allow us to become more fully ourselves.

Read original →
Marginal Revolution

The Raphael show at the NYC Met

Tyler Cowen offers an enthusiastic but characteristically measured review of the Raphael exhibition at the Metropolitan Museum of Art, calling it a rare event in the art world. What distinguishes the show, in his view, is its unusual density of original content: rather than a single-artist retrospective padded with context, it functions as a series of embedded mini-exhibits, including early large-scale altarpieces rarely seen outside their original locations, works by Raphael's teacher Perugino, a small collection of Leonardo drawings, Roman sculptures illuminating Raphael's influences, and Flemish derivations of the Vatican tapestries. The concentration of light-sensitive drawings alone — typically scattered across institutions or kept from display — makes the show exceptional.

Cowen's admiration is genuine but stops short of hagiography. He places Raphael at the pinnacle for beauty and charm while noting that, with rare exceptions like the Louvre's *Portrait of Baldassare Castiglione*, Raphael's work doesn't unsettle or confound him the way a top Leonardo or Velázquez does — a distinction that separates the merely masterful from the truly inexhaustible. His sharpest criticism is reserved not for the art but for the exhibition's introductory text, which opened by describing Raphael as "one of the most important influencers of all time," a piece of curatorial flattening he finds embarrassing.

For the culturally curious reader, the review is a useful reminder that how an exhibition is framed matters as much as what it contains — and that even a show of this caliber can be undermined by the instinct to make genius legible through contemporary cliché. Cowen's verdict: almost certainly the finest opportunity to encounter Raphael's work in a single visit that most people will ever have.

Read original →
Marginal Revolution

Thursday assorted links

This is a link roundup post from Tyler Cowen's Marginal Revolution blog, collecting six items of interest without extended commentary. The links span chess, literature, tech policy, European politics, ancient genomics, and labor economics.

The most substantive items touch on two recurring themes in contemporary discourse. The failure of Australia's social media ban for under-16s — with 61% of affected children reporting unchanged access — offers a pointed data point in the ongoing debate about whether age-restriction legislation can meaningfully alter online behavior, or whether it amounts to performative policy. Meanwhile, Janan Ganesh's FT observation about European nationalist politicians is quietly striking: figures like Le Pen and Meloni have moderated their Euroscepticism not out of conviction but electoral necessity, and Orbán's weekend defeat in Hungary — where the winning challenger ran explicitly on repairing EU relations — suggests that hard anti-Brussels politics may have hit a ceiling with voters.

The remaining links gesture toward longer-horizon questions: ancient DNA research revealing directional selection across West Eurasia adds to the growing picture of how dramatically human populations were shaped by prehistoric forces, while Alex Imas's work on AI and employment speaks to anxieties very much alive in the present. Taken together, the roundup reflects Cowen's characteristic range — empirical, politically observant, and attentive to both deep history and near-future disruption.

Read original →
Marginal Revolution

The Nobel Memorial Prize in Economics, 1969-2025

This is a brief post flagging a new academic paper by Peter Dolton and Richard Tol that attempts to model the selection patterns behind the Nobel Memorial Prize in Economics across its 57-year history. The authors find that the prize rotates in a semi-regular fashion across subfields of economics, and that while broad citation networks don't predict winners, having a Nobel laureate as a student or co-author does confer a meaningful advantage — suggesting that intellectual lineage matters more than general prestige.

The paper also finds evidence that the personal preferences of individual committee members have shaped both field selection and individual winners, and that the committee's behavior shifted noticeably after the influential economist Assar Lindbeck retired from it. Other structural patterns include a historical shift from prizes awarded for a single landmark paper toward recognition of an entire body of work.

For curious readers, this matters because it pulls back the curtain on what can appear to be an Olympian judgment of scholarly merit, revealing instead a process shaped by institutional dynamics, personal relationships, and committee politics. It raises the uncomfortable but intellectually honest question of how much any prize — even the most prestigious in a discipline — reflects objective achievement versus the sociology of the field itself.

Read original →
Marginal Revolution

In Development magazine

This is a brief promotional post from Tyler Cowen flagging the launch of *In Development*, a new magazine focused on evidence-based approaches to global economic development. The post highlights a piece by Paul Niehaus on GiveDirectly and the empirical case for cash-based transfers as a development intervention.

Given its nature as a short link post, there is little argumentative substance to summarize beyond the signal Cowen is sending: that this publication is worth readers' attention. The implicit endorsement carries weight given Cowen's longstanding interest in development economics and effective giving, and GiveDirectly's cash transfer model remains one of the more rigorously studied interventions in the field.

For curious readers, the pointer is useful less for what it says than for what it points toward — a dedicated outlet synthesizing development research at a moment when debates over aid effectiveness, conditionality, and direct transfers remain very much alive in both academic and policy circles.

Read original →
Marginal Revolution

Robert Skidelsky, RIP

This is a brief memorial post by Tyler Cowen marking the death of Robert Skidelsky, the British economic historian and biographer.

Cowen's sole substantive claim is a strong one: Skidelsky's three-volume biography of John Maynard Keynes ranks among the greatest books ever written, comparable to Robert Caro's monumental work on Robert Moses. Notably, Cowen steers readers toward the full three-volume set over the condensed single-volume version — a meaningful distinction for anyone approaching the work.

For readers interested in intellectual biography or the history of economic thought, the recommendation carries weight coming from Cowen, who reads voraciously across fields. Skidelsky's Keynes biography is widely regarded as a landmark not just of economic history but of biography as a literary form, and this short tribute, spare as it is, reflects genuine admiration.

Read original →
Marginal Revolution

Revising Modern Principles

This is a brief post from Alex Tabarrok noting that he has made revisions to *Modern Principles*, the economics textbook he co-authors with Tyler Cowen, with the observation that some content requires significant updating while other material remains perennially relevant. No further detail or examples are provided.

This is a short, content-light post — essentially a teaser or personal note — and does not contain enough substance to summarize beyond its surface announcement.

Read original →
Technology 11 articles
Stratechery

An Interview with F1 Driver and Venture Capitalist Nico Rosberg About the Drive to Win

Nico Rosberg's trajectory from Formula 1 world champion to venture capitalist is the subject of this wide-ranging Stratechery interview, and what emerges is a portrait of someone whose apparent career pivots are actually unified by a single obsessive drive: to find and maximize every available edge. Rosberg grew up steeped in racing — his father Keke was the 1982 F1 world champion — but what distinguished him wasn't raw confidence. By his own account, he was always more insecure than rivals like Lewis Hamilton, and he channeled that insecurity into relentless preparation, including a decade of intensive work with a sports psychologist at a time when mental coaching in F1 was considered a sign of weakness. That psychological discipline, he argues, became his competitive superpower, and it shapes how he now evaluates founders — looking for the rare combination of high conviction and genuine curiosity, a paradox that separates the truly exceptional from the merely talented.

The interview covers the famous Hamilton-Rosberg rivalry in detail, including the remarkable fact that the two were childhood teammates on a McLaren-funded go-kart team at age 13, dreaming of one day racing together in F1 — a dream that came true and then curdled into one of the sport's most intense and psychologically grueling rivalries. Rosberg's decision to retire immediately after winning the 2016 championship — walking away at the absolute peak — is framed not as an escape but as a calculated act of self-awareness: he knew the cost of sustaining that level of obsessive focus was unsustainable, and that winning once was better than grinding toward a second title while losing himself in the process.

The venture capital portion of the conversation is where Rosberg's strategic thinking becomes most interesting. He has built Rosberg Ventures around a specific structural advantage: his F1 network gives him access to European capital and German industrial relationships that Silicon Valley founders struggle to cultivate, while his Silicon Valley connections give European companies access to cutting-edge technology. He draws explicit parallels between elite drivers and elite founders — both require extreme resilience, the ability to act on conviction without external validation, and the mental fortitude to keep competing when the outcome is deeply uncertain. Ben Thompson notes that the interview is available as a podcast for Stratechery subscribers.

Read original →
Hacker News

Show HN: MacMind – A transformer neural network in HyperCard on a 1989 Macintosh

A developer has implemented a complete transformer neural network inside HyperCard — Apple's 1987 point-and-click scripting environment — running on a 1989 Macintosh. The model is genuinely complete: embeddings, positional encoding, self-attention, backpropagation, and gradient descent, all written in HyperTalk, a language designed for building interactive "stacks" rather than numerical computation. At 1,216 parameters, it's a toy by modern standards, but architecturally it's the real thing, and every line of math is inspectable by option-clicking any button in HyperCard's script editor.

The chosen task is elegant: learning the bit-reversal permutation, the index-shuffling step that opens the Fast Fourier Transform. The model receives no formula — it must discover the underlying positional pattern purely through attention and gradient descent. By training step 193 it was already oscillating between 50%, 75%, and 100% accuracy before settling into convergence, and the trained weights persist simply by saving the HyperCard stack file, surviving across restarts on anything from System 7 to Mac OS 9.

The project is less a technical benchmark than a philosophical argument. The author's explicit motivation is demystification: backpropagation and attention are mathematics, and mathematics is indifferent to whether it executes on a TPU cluster or a 68030 processor. By stripping away the abstraction layers of Python frameworks and cloud infrastructure and running the same ideas on hardware most people associate with Minesweeper and clip art, MacMind makes a quietly powerful point — that what we call "AI" is, at its core, just arithmetic applied patiently and repeatedly.

Read original →
Hacker News

Codex for almost everything

I wasn't able to retrieve the article content directly, but based on the title, author context, and the significant traction this post received on Hacker News (634 points, 349 comments), I can offer the following based on what's publicly known about this OpenAI announcement.

OpenAI's Codex announcement positions the AI coding agent not merely as a tool for software engineers, but as a broadly capable autonomous agent that can handle complex, multi-step software tasks across nearly any domain. The "almost everything" framing is deliberate — it signals a shift from Codex as a code-completion assistant to something closer to a software engineer in the cloud, capable of reading codebases, writing and running tests, filing pull requests, and iterating on feedback without constant human supervision.

The broader significance lies in what this implies for knowledge work. If a coding agent can operate asynchronously on real engineering tasks — not just autocomplete snippets but actually *ship* work — the productivity implications are substantial, and the questions about human oversight, code quality, and job displacement become more urgent. The Hacker News discussion likely reflects both genuine excitement from developers experimenting with agentic workflows and skepticism about the gap between demo performance and production reliability, a recurring tension in AI capability announcements.

*Note: I was unable to fetch the full article text, so this summary draws on publicly available context about the announcement. For full accuracy, the original piece is worth reading directly.*

Read original →
Hacker News

Qwen3.6-35B-A3B on my laptop drew me a better pelican than Claude Opus 4.7

This appears to be a short post or link from Simon Willison's blog rather than a long-form article — likely a brief observation or demonstration comparing outputs from two AI models on a creative task. Based on the title, Willison ran a locally-hosted Qwen3.6-35B-A3B model on his laptop and asked it to draw (likely generate SVG or ASCII art of) a pelican, finding the result superior to what Claude Opus 4.7 produced — a notable claim given that Opus 4.7 represents Anthropic's flagship tier.

The significance here is less about pelicans and more about what it signals for the state of open-weight models. Qwen3.6-35B-A3B is a mixture-of-experts model that activates only 3 billion parameters at inference time despite having 35 billion total, making it efficient enough to run on consumer hardware. The fact that it can outperform a top-tier commercial model on even a narrow creative/spatial reasoning task — while running locally, for free — is the kind of concrete, reproducible demonstration that tends to resonate with the technically curious.

For readers tracking the AI landscape, this fits into a broader pattern: the gap between frontier commercial models and capable open-weight alternatives is narrowing faster than many expected, and the efficiency gains from sparse architectures like MoE are making local inference increasingly practical. Willison, a trusted voice in the developer community, tends to share these comparisons as honest empirical observations rather than hype, which gives the finding added credibility.

Read original →
Hacker News

The "Passive Income" trap ate a generation of entrepreneurs

**Summary**

Joan Westenberg's argument is pointed and uncomfortable: the "passive income" gospel that swept through entrepreneurial culture over the past decade didn't liberate a generation of builders — it sedated them. Promising financial freedom through automated revenue streams, the movement repackaged the fantasy of getting paid while you sleep into a legitimate-sounding business philosophy. The result, Westenberg contends, was a cohort of would-be entrepreneurs who optimized for low-effort monetization schemes — digital courses, affiliate marketing, dropshipping — rather than building anything of genuine value or durability.

The deeper critique is cultural. Passive income ideology, Westenberg argues, fundamentally corrupted the entrepreneurial mindset by decoupling reward from contribution. Real businesses are hard, iterative, and demanding; they require sustained engagement with customers, problems, and markets. The passive income framework treated that difficulty as a bug to be engineered away, producing ventures that were thin by design — easy to start, easy to abandon, and largely useless to anyone but their creator. The gurus who sold the dream often made their actual money selling the dream itself, a recursion that should have been a tell.

What makes this worth reading for a curious observer is that it gestures at something broader than entrepreneurship: a cultural preference for systems that generate returns without presence, accountability, or craft. Westenberg isn't simply moralizing about laziness — the piece suggests that when an entire generation's ambition gets channeled into extraction rather than creation, the opportunity cost is real innovation that never happened.

Read original →
Hacker News

Artifacts: Versioned storage that speaks Git

Cloudflare has launched **Artifacts**, a versioned storage system designed specifically for AI agents, built on Git's underlying object model. The core insight is that AI agents—unlike traditional software—need to store, retrieve, and branch across many intermediate states of their work. Rather than inventing a new versioning paradigm, Cloudflare recognized that Git already solved this problem elegantly: its content-addressable, immutable object store with cheap branching maps naturally onto the needs of agentic workflows. Artifacts exposes this through a familiar Git-compatible interface, meaning developers can use standard Git tooling to interact with agent-produced data.

The practical implication is that agents can now checkpoint their work, fork execution paths to explore alternatives, and merge results—mirroring how Git enables parallel development in software teams. This is particularly valuable for long-running or multi-step agent tasks where you want auditability, rollback capability, or the ability to run speculative branches without corrupting a main state. Cloudflare hosts this on its Workers platform, so it inherits the edge-distributed, low-latency properties of that infrastructure.

For curious readers, the deeper interest here is conceptual: it suggests that Git's data model, often thought of as a narrow tool for source code, is actually a general-purpose versioned state machine applicable well beyond text files. As AI agents become more stateful and complex, the question of *how* they manage intermediate state becomes critical infrastructure—and Cloudflare is betting that the answer already existed in a 20-year-old tool most developers use every day.

Read original →
Hacker News

Cloudflare's AI Platform: an inference layer designed for agents

Cloudflare has announced a significant expansion of its AI platform, positioning itself not merely as a host for AI models but as a full inference layer purpose-built for agentic workloads. The central argument is that running AI agents in production requires more than raw compute — it demands low-latency routing, persistent memory, tool orchestration, and reliable state management across long-running tasks. Cloudflare's global edge network, already optimized for speed and distribution, becomes the substrate for all of this, with the company betting that proximity to users and tight integration between networking and inference will be a decisive advantage over hyperscaler alternatives.

The platform bundles several capabilities together: Workers AI for running models at the edge, AI Gateway for observability and caching of inference calls, Vectorize for retrieval-augmented generation, and — critically — Durable Objects and Workflows for managing the stateful, multi-step nature of agent execution. The pitch is that developers shouldn't have to stitch together a dozen third-party services to build a reliable agent; Cloudflare wants to be the single platform where the model call, the memory lookup, the tool invocation, and the retry logic all live. New additions include support for the Model Context Protocol (MCP), making it easier to expose and consume tools across agent boundaries.

For technically curious readers, the interesting tension here is whether Cloudflare's edge-first architecture — historically optimized for latency-sensitive, stateless workloads — can genuinely serve the long-running, compute-heavy demands of sophisticated AI agents. The company is clearly aware of this and is leaning into Durable Objects as its answer to persistent state. Whether that's sufficient to compete with AWS Bedrock or Google Vertex for serious enterprise deployments remains an open question, but Cloudflare's developer-friendliness and pricing model give it a credible path into the market from the bottom up.

Read original →
Hacker News

Claude Opus 4.7

I wasn't able to retrieve the full article content, so I'm working from the title and metadata alone — this appears to be an Anthropic product announcement for Claude Opus 4.7, shared on Hacker News where it generated substantial discussion (1,394 points and over 1,000 comments), suggesting a significant release.

Based on the naming convention, Claude Opus 4.7 would represent an incremental update within Anthropic's Opus 4 model family — Opus being Anthropic's most capable tier. Releases at this level typically involve improvements to reasoning, instruction-following, coding ability, or safety characteristics, and often come with updated context windows, pricing adjustments, or API changes that matter to developers building on the platform.

The high engagement on Hacker News points to genuine interest from the developer and AI research community, who tend to scrutinize benchmark claims, compare performance against competing models from OpenAI and Google, and debate the practical implications of capability jumps. Without the full article, I'd recommend reading the Anthropic announcement directly and checking the Hacker News comments thread, which at 1,000+ replies likely contains hands-on impressions, benchmark comparisons, and critical analysis worth reading alongside the official release notes.

Read original →
Hacker News

Qwen3.6-35B-A3B: Agentic coding power, now open to all

I wasn't able to retrieve the full article content from the provided URLs, so I'm working from the title, metadata, and what can be reasonably inferred from context.

This appears to be an announcement from Alibaba's Qwen team introducing Qwen3.6-35B-A3B, a new open-weight mixture-of-experts language model designed specifically for agentic coding tasks. The model's naming convention suggests it has 35 billion total parameters but activates only around 3 billion at inference time — a sparse architecture that dramatically reduces computational cost while preserving much of the capability of a much larger dense model. The "open to all" framing signals that weights are being publicly released, positioning it as a serious open-source competitor to proprietary coding assistants.

The significance here is the combination of agentic capability and accessibility. Agentic coding — where a model doesn't just complete snippets but plans, executes multi-step tasks, uses tools, and iterates on its own output — has largely been the domain of frontier closed models like Claude or GPT-4o. A model that can operate in this regime while running efficiently on consumer or modest cloud hardware represents a meaningful shift in what developers can self-host. The MoE architecture is key to making this practical.

For technically curious readers, this matters because it continues a pattern of rapid capability diffusion: tasks that required expensive API calls to proprietary systems six months ago are increasingly achievable with locally-run open models. The 872 points and 407 comments on Hacker News suggest the developer community is taking it seriously as a practical tool, not merely a research artifact.

Read original →
Hacker News

Cloudflare Email Service

I wasn't able to retrieve the full article content from the provided URLs, so I'm working from the title and context alone.

Cloudflare appears to have announced an email service designed specifically for AI agents — a programmable email infrastructure that lets autonomous software systems send, receive, and process email as a native capability rather than a bolted-on afterthought. The core idea seems to be that as agentic AI systems become more prevalent, they need reliable, scalable communication primitives, and email — despite its age — remains one of the most universal and interoperable messaging layers available.

The significance lies in Cloudflare's positioning: by sitting at the infrastructure layer, they can offer email handling with the same developer-friendly, globally distributed approach they've applied to DNS, CDN, and serverless compute. For agents that need to interact with humans or external services asynchronously, email provides a durable, auditable channel that APIs alone don't always replicate. This would let developers wire up agents that can, say, receive instructions, send confirmations, or trigger workflows through ordinary inboxes without managing mail servers.

For technically curious readers, this is worth watching as part of a broader pattern: foundational internet primitives — DNS, HTTP routing, storage, and now email — are being re-exposed as clean, programmable APIs optimized for software agents rather than human users. Cloudflare is quietly assembling a full-stack platform for the agentic web, and email may be a more strategically important piece of that puzzle than it first appears.

Read original →
Hacker News

The future of everything is lies, I guess: Where do we go from here?

Kyle Kingsbury (aphyr), the distributed systems engineer behind the Jepsen database testing project, writes a despairing but clear-eyed essay about the collapse of epistemic trust in technical communities — specifically around AI benchmarks and product claims. The central argument is that the incentive structures of the current AI industry have made honest evaluation nearly impossible: companies routinely train on benchmark test sets, cherry-pick results, and make performance claims that dissolve under independent scrutiny. Kingsbury has spent years doing rigorous adversarial testing of databases, and he watches the same dynamic play out with AI systems at far greater scale and with far less accountability.

What makes the piece more than a rant is Kingsbury's careful attention to *why* this is structurally hard to fix. Independent evaluators face asymmetric resources — a small team cannot keep pace with the marketing apparatus of a well-funded lab. Benchmarks, once published, become targets for optimization rather than honest proxies for capability. And the social incentives cut against skeptics: pointing out that a celebrated model underperforms is professionally thankless, while hype is rewarded. He draws on his Jepsen experience to note that even when rigorous testing exposes failures, the corrections rarely propagate as far as the original claims did.

The essay matters because it names a genuine epistemic crisis in a field that is rapidly influencing hiring, investment, policy, and infrastructure decisions. Kingsbury isn't arguing that AI systems are useless — he's arguing that we have largely lost the ability to know what they can actually do, and that the people best positioned to find out are systematically discouraged from doing so. For technically literate readers who rely on honest evaluation to make decisions, that's a serious and underappreciated problem.

Read original →
Policy 2 articles
Slow Boring

My most right-wing views

Matthew Yglesias sets out to identify the policy areas where he genuinely agrees with conservative thinking, while acknowledging upfront how difficult that exercise has become. The Trump era has scrambled traditional ideological categories — free trade, once a clear case where Yglesias thought the right was correct, is now so thoroughly abandoned by Republicans that it no longer qualifies. More broadly, he argues that American conservatism is structurally defined less by any coherent set of ideas than by opposition to the left, which produces a pragmatic electoral discipline that progressives often lack, but also an intellectual corruption that requires carving out ever-larger Trump-shaped exceptions to any principled position.

The one substantive right-leaning view the excerpt develops is on rule enforcement and punishment. Yglesias argues that conservatives are broadly correct that detecting rule-breaking and imposing meaningful consequences matters — a position he holds even while rejecting many of the specific criminal justice policies conservatives favor. Progressives, he suggests, have developed an ambivalence toward enforcement as such that goes too far in the other direction, conflating legitimate skepticism of mass incarceration with a reluctance to hold rule-breakers accountable at all.

This piece is worth reading for the intellectual honesty of its framing as much as its conclusions. Yglesias is attempting something genuinely difficult — identifying where ideological opponents have the better argument — while being clear-eyed about how the Trump realignment has made that exercise murkier than it used to be. The tension he identifies, between conservatism's pragmatic cohesion and its moral compromises, is one of the more useful lenses available for understanding the current political moment.

Read original →
Slow Boring

Destroy the internet to save it?

This appears to be a truncated excerpt from a Slow Boring post, likely a newsletter teaser that cuts off before the substantive argument is developed. Only the setup is visible: Yglesias mentions a podcast conversation with someone named Jerusalem about whether mandatory real-name policies online would improve internet discourse, and signals he has reservations about simply defending anonymity as it currently exists.

Without access to the full article, it's impossible to fairly summarize Yglesias's actual argument, his evidence, or the nuances of his position on the tradeoffs between anonymity and accountability online. The piece appears to engage a genuinely contested question — whether the costs of anonymous posting (harassment, bad-faith discourse, radicalization) outweigh the benefits (protection for dissidents, whistleblowers, and marginalized voices) — but the excerpt ends before any real analysis begins.

If you have access to the full text, paste it here and I can provide a proper summary.

Read original →