This week’s video transcript summary is here. You can click on any bulleted section to see the actual transcript. Thanks to Granola for its software.
Editorial
Adulting: The Week AI Leaders Matured
Anthropic had quite a week. As the articles below show, the company reversed product decisions that were plainly driven by a shortage of compute resources after users complained. The mood around founder Dario Amodei shifted from good to bad as he came to look indecisive. Revelations that the US government has access to Mythos did not help. Then, at the end of the week, Anthropic announced a $10 billion investment from Google that could become $40 billion. That matters because Google is not just an investor. It is also a major supplier of Anthropic’s infrastructure. Anthropic also released Claude Design, which was widely praised by everyone except Canva and Figma.
Meanwhile, OpenAI released ChatGPT 5.5. Sam Altman and Greg Brockman gave a long interview and came across as clear, confident, and in command. ChatGPT Image 2.0 was also released to wide acclaim.
These developments make this week’s issue feel clearer than the last few. The question is no longer whether the models are smart. They are. The real question is who is behaving like an adult, and at last the answer appears to be everybody. Let’s hope it lasts.
The Sam Altman and Greg Brockman interview matters. As we said in previous weeks, OpenAI sounds like a company that has decided to grow up. Less magic show. More product. Less grand theory. More decisions. In their telling, the model is not the product.
“The models have shifted from being the product to being a part of the product…”
Why it matters: OpenAI is saying plainly that the model by itself is no longer enough. It is one part of a larger system made up of tools, memory, tasks, agents, user context, and a business model that can support all of it.
OpenAI’s Codex moves front and center as the place where users are converging. With Image 2.0, the suggestion that OpenAI can do everything started to look credible. OpenAI is trying to build something people will use, not just something people will talk about.
The Anthropic story stuttered at the start of the week. It quietly pulled Claude Code from the Pro plan, then admitted its plans were not built for the level of demand it was attracting. Then it reversed the decision. This zigzag is not a footnote. It is a tell. Anthropic is expensive because it has to preserve compute resources. It also looked indecisive. The product keeps changing. The offer keeps moving. If you want to be the platform people depend on for work, you do not get to be fuzzy about price, packaging, and reliability. Those are not details. They are the whole job.
Anthropic has never sounded enthusiastic about the capital intensity of this business. Dario often talks as if capex is something to endure rather than embrace. That may be rational. Compute is brutally expensive. But if you do not want to spend aggressively on the infrastructure the product requires, that reluctance has a way of showing up somewhere else. It shows up in limits. It shows up in packaging changes. It shows up in a product that seems to hesitate in public.
Bloomberg’s report that Google plans to invest up to $40 billion in Anthropic, while also supporting a significant expansion of Anthropic’s computing capacity, makes that problem harder to wave away. If Anthropic now needs Google to help underwrite the next jump in infrastructure, that does not erase the earlier indecision. It does, however, help explain it.
This is where ideology enters the picture. There is nothing wrong with having values. There is nothing wrong with arguing about safety. But there is something deeply unserious about using moral language as a competitive weapon while the product itself is wobbling. That is not leadership. It is game playing. It is an attempt to win by claiming superior virtue instead of doing the harder work of building a dependable business. Washington may like that game. Journalists may like that game. Customers usually do not. But the decision to raise up to $40 billion shows recognition of the need for change, and that is an adult move.
A lot of the other articles this week make the same point from different angles.
The Vercel breach story shows what happens when AI spreads through a company faster than anyone can control it.
The analytics article by Tristan Handy shows agents becoming part of normal work.
Rohit Krishnan’s multi-agent misalignment essay shows how a group of obedient systems can still create a false company record.
The Google enterprise article, Omni’s Series C, and Om Malik’s pieces on Cursor and xAI all point to the same place. The real fight is over who owns the day to day relationship with the user and who can turn intelligence into ordinary, dependable work. Why else would SpaceX agree to a $60 billion price for Cursor, with a $10 billion payment if it fails to complete the acquisition?
Then there is the part too many people want to skip. As Norman Lewis makes clear, the physical world still matters. Construction costs matter. Industrial policy matters. ASML matters. DeepSeek matters partly because chips and export controls matter. AI is not floating above the economy like some pure force of history. It is landing in the real world. Once it lands, it has to live under the same rules as every other serious business. Somebody has to fund the buildout that the future promise rests on. This is required, serious, adult stuff.
The venture capital essays belong here for that reason. For years the standard venture capital story was simple. Find an entry point. Grow fast. Widen the moat after getting to scale. That story is getting weaker. Lasting power actually sits in distribution, customer trust, workflow control, and infrastructure. It sits in the ugly parts of the system that are hard to build and hard to replace. Those answers are less glamorous than AGI talk. They are also more likely to be true. The cash pouring into those efforts, including this new $40 billion from Google to Anthropic, is all part of the preconditions for GDP growth and general abundance.
My read on this week is straightforward. AI is entering its grown up phase. The stakes are not smaller.
If you want to lead this market, you have to do grown up things. Spend $60 billion on Cursor. Raise $40 billion for Anthropic. Spend enough to satisfy demand for your product.
Build a product people use. Price it honestly. Keep it secure. Make it reliable. Earn trust. The companies that can do that will matter. The companies that keep playing ideological games will still get attention, but they will look less and less serious. Hopefully Anthropic is past that adolescent phase.
Reid Hoffman puts the deeper point well in “Faith in the Possible”:
“technology’s arc bends toward access. But it does not bend on its own.”
The adult response to this moment is to build the product, fund the infrastructure, earn trust, and lean into the opportunity with open eyes. Human agency will determine the outcome. Thankfully panic, moral theater, and passivity did not prevail this week.
Grown up behavior in AI looks a lot like grown up behavior in any other business. The companies that win will be the ones customers can actually depend on. Moral theater is what people reach for when product discipline is not enough. Faith is a pre-condition for effective action.
Contents
Editorial
Essays
Venture
AI
Regulation
Infrastructure
Interview of the Week - Friending the Machine
Startup of the Week - Omni’s Series C and “Intelligence About Business”
Post of the Week - Ben Braverman on Sequoia’s Endowment Outperformance
Essays
Sam Altman & Greg Brockman on the OpenAI Reset
YouTube · Sam Altman & Greg Brockman · Apr 22, 2026 · Tags: AI, OpenAI, Agents, Strategy, Compute
Transcript:
OpenAI is pivoting from shipping models as products to shipping agents as platforms. Brockman frames the architecture shift plainly: “The models have shifted from being the product to being a part of the product… we have this brain in the form of the model, now we’re building the body.” The three-pillar roadmap: a general agent and connector layer; agents handling “computer work” - Brockman’s deliberate alternative to “knowledge work”: “no one thinks of themselves as a knowledge worker” - and “personal AGIs” that hold user context and act autonomously. “We are clearly at a moment of transition to agents.”
The reset has teeth. Sora paused. Compute reframed as profit center, not cost center: “We rent or buy compute and then we resell it at a margin. As long as we have some positive margin on it, it’s scalable because the demand is just unlimited.”
The most striking passage is Altman’s “less sanitized version” at [38:46], sketching three candidate futures: “Everybody gets subjectively like 10 times richer… but because this is a lever that people can really use, the most capable, most ambitious, the people who already started rich and have access to a lot of compute, inequality gets worse. So that’s one world I can see. I can see another world where the floor doesn’t come up as much. We don’t generate as much total prosperity. But also there’s like less inequality.” Brockman’s reframe: “AI is opportunity for everyone if you have access, if you have compute. If you don’t have compute, you can’t.” The unlock they propose is universal cheap compute, not redistribution.
On the Musk lawsuit, Brockman is unexpectedly candid about the breaking point: “Elon’s like, ‘Need majority equity, need to be CEO, need full control’… Absolute control over OpenAI, doesn’t matter who that person is - that was the breaking point. That was the thing that caused us to say no.” Altman: “I see no future, no good future, where leading AI efforts don’t assist the US government.”
Read more: YouTube · Transcript
Yann LeCun on why Dario is wrong about AI and mass unemployment
X thread · Yann LeCun · Apr 18-19, 2026 · Tags: AI, Labor, Economics, Automation
Yann LeCun’s intervention is useful precisely because it is narrower than the usual AI doom argument. He is not claiming automation will leave labor markets untouched. He is arguing that AI lab founders are poor authorities on economy-wide employment effects, and that people should stop treating dramatic CEO forecasts as social science. His explicit advice is to ignore Dario, Sam, Geoff, Yoshua, and even LeCun himself on this question, and instead listen to economists such as Philippe Aghion, Erik Brynjolfsson, Daron Acemoglu, Andrew McAfee, and David Autor.
The second post in the thread sharpens the point. LeCun says the history of technological progress is full of professions being displaced or eliminated while aggregate productivity rises, and yet those transitions have not produced permanent mass unemployment. His core disagreement with Dario is not that AI will change work, but that AI should not be treated as a uniquely exempt force outside the historical pattern of previous technological revolutions.
What makes the thread worth keeping is that it exposes a deeper fight about authority. AI leaders are increasingly tempted to speak as forecasters of civilization, but labor-market outcomes depend on adoption rates, institutions, policy, firm behavior, and demand reallocation, not just capability curves. The strongest line here is not LeCun’s confidence. It is his insistence that intelligence researchers should not be mistaken for labor economists.
Read more: Thread opener · Follow-up
2025 Annual Letter
Social Capital · Chamath Palihapitiya · Apr 2026 · Tags: AI, Macro, Allocation, Geopolitics, Physical-AI
Chamath reframes 2025 as a transitional year: numerically strong (Social Capital’s inflection was NVIDIA licensing Groq for $20B) but not decisive for the work done inside it. The thesis is deliberately deflationary on the hype curve. Today’s AI is, in his framing, “an ultra-advanced, mathematically grounded autocomplete,” not superintelligence. Digital foundation models are commoditizing quickly. The durable value, he argues, migrates to the physical layer - critical minerals, chemical processing, energy storage, precision actuation - assets that can’t be rapidly repurposed.
The market reprice is already visible. Big Tech concentration rose from ~17% of the S&P in 2018 to 36.2% by October 2025. Inside that print, software multiples are compressing while silicon and infrastructure names climb: Duolingo −67.5% from its mid-2025 peak to year-end, Adobe −24.6%, against Broadcom and SK Hynix moving the other way. Private credit defaults, he suggests, may be the earliest canary of AI-driven disruption rippling out of tech into the broader corporate balance sheet.
Geopolitically, Chamath argues only the US and China have the talent, compute, energy, and capital stack to build frontier AI, producing a bipolar world in which other countries effectively rent intelligence from one side or the other. The policy backdrop (tariffs, onshoring subsidies, state financing of long-horizon infrastructure) is already steering capital toward atom-heavy, domestically-built fulcrums. “Superintelligence is coming but it’s not here yet.” His advice is to allocate accordingly and hold independent judgment while the reprice plays out.
Read more: Social Capital
What will be scarce?
Ghosts of Electricity · Alex Imas · Apr 14, 2026 · Tags: AI, Labor, Economics, Structural Change
Alex Imas makes a useful move that a lot of AI labor debates skip. Instead of asking only which jobs get automated, he asks what happens to demand when automation makes more of the economy cheap. His answer is that the commodity sector may shrink as a share of spending while more value flows toward relational work, goods, and services where the human element is part of the product itself.
That frame helps explain why AI does not automatically imply a flat, fully commodified economy. If software keeps pushing production costs down, spending can reallocate toward care, education, hospitality, performance, craftsmanship, provenance, and other forms of work where people still care who did it and why. The strongest part of the essay is that it ties this not to nostalgia, but to structural-change economics and income effects.
It is a strong fit for this week’s theme. Breakthrough capability alone does not determine what wins economically. Institutions, preferences, status, and human desire still shape where value accrues after the technical frontier moves.
Read more: Ghosts of Electricity
Updated thoughts on industrial policy
Noahpinion · Noah Smith · Apr 20, 2026 · Tags: Economics, Industrial Policy, Trade, FDI
Noah Smith’s update is a useful corrective to how loosely “industrial policy” gets used in the current tariffs-and-subsidies debate. His core move is to demand specificity. Tariffs, export subsidies, FDI promotion, national-champion picking, and technology enablement are not the same policy, and the evidence for each runs in different directions. Treating them as one thing has made the debate nearly impossible to evaluate.
The strongest part is the FDI argument. Smith points to Poland, Malaysia, Singapore, and Ireland as cases where promoting foreign direct investment outperformed traditional “pick winners” approaches, with real productivity gains and no obvious national brand to show for it. For rich countries, he argues the cleanest version of industrial policy is the one the United States has quietly been doing for decades: enabling new technologies like the internet, and now AI infrastructure. He treats China’s mega-subsidy experiments as a cautionary tale about margin compression, financial fragility, and the political economy of propping up unprofitable industries.
The piece lands for this week because it pushes back on the narrative that industrial policy is either obviously good or obviously bad. The real question is which version, under which institutions, with which tradeoffs acknowledged up front.
Read more: Noahpinion
Construction Costs Rarely Fall
Construction Physics · Brian Potter · Apr 23, 2026 · Tags: Infrastructure, Housing, Construction, Productivity
Brian Potter argues that construction costs mostly ratchet upward because the constraints are cumulative and sticky. Site specificity, permitting, fragmented subcontracting, code accretion, and trade bottlenecks make building look nothing like software, where tools can spread instantly and marginal costs collapse.
The sharpest detail is not a single number but the mechanism: every new requirement, approval layer, and coordination dependency gets added to a stack that rarely shrinks. That means technical improvements often show up as absorbed complexity rather than cheaper buildings. More capability does not automatically translate into lower delivered cost.
The piece lands where a lot of supply-side optimism tends to go fuzzy. If housing, energy, and industrial buildouts are central to the next decade, the real question is not whether we have better tools. It is whether institutions can remove enough friction for those tools to matter in the final price.
Read more: Construction Physics
Apple makes a safe choice in a dangerous moment
Author: Reed Albergotti Published: Apr 20, 2026
Reed Albergotti reads Apple’s choice of John Ternus as a defensive move dressed up as a succession plan. The company did not pick an AI visionary or a software operator; it picked an Apple lifer from the Jobs era whose entire career is in hardware engineering. That signals, Albergotti argues, that the board is not willing to “roll the dice on someone who might squander its lucrative business,” even though the lucrative business is the one piece of Apple most at risk if the AI era redefines what an operating system actually is.
The killer framing is historical. Apple was never first with the computer, the phone, or the tablet; it won by perfecting the software layer on top of someone else’s category. If no one has yet defined what an OS looks like when the interface is a model, Apple’s whole playbook assumes a fight that has not started.
The piece ends in an uncomfortable place. Cook scaled the company’s market cap past a trillion, but its cultural weight has been sliding since 2011. By the time the AI shape of consumer software is obvious, Albergotti warns, it may already be too late for Apple to reclaim what it had.
Read more: Source
That was Tim, this is Ternus: Some first thoughts on Apple’s CEO transition
Author: Jason Snell Published: Apr 20, 2026
Jason Snell reads the Cook-to-Ternus handover as the anti-2011: a planned, four-month glide path instead of the abrupt ascension Cook himself was handed after Steve Jobs’s illness. Cook spent years watching Jobs run Apple before he got the job; he is giving Ternus a version of the same runway, with executive chairman scaffolding behind it.
The killer detail is what the transition protects. Cook will keep the diplomatic portfolio, meaning Ternus does not inherit the DC-and-Beijing travel schedule that has consumed much of Cook’s late tenure. At the same time, Johny Srouji, the silicon architect Apple had quietly been rumored to be losing in December, has been promoted into a new Chief Hardware Officer role, a retention move wrapped in a reorg.
Snell’s broader argument is that after fifteen to twenty years of the same faces running Apple’s divisions, a product-person CEO creates room for organizational change that Cook’s continuity-first style could never justify. The risk he flags is not that Ternus fails, but that the reshuffle pushes long-tenured executives out the door faster than Apple can replace the institutional memory.
Read more: Source
Faith in the Possible
Reid Hoffman · Apr 20, 2026 · Tags: Technology, Progress, AI, Silicon Valley, Ideology
Reid Hoffman reframes Silicon Valley not as a capital machine but as a faith movement. “Silicon Valley has a God complex,” he concedes. “The critics are not wrong about that. They are just wrong about which god.” The god in question is the belief that tools expand human capacity, a creed that now backs roughly a tenth of the U.S. economy and a third of its growth over the last decade. Hoffman maps the field as a set of denominations - Purists, Humanists, Missionaries, VCs, Accelerationists, Ethicists, Determinists - each reading the same technological scripture differently, and argues the fight among them is a sign of health, not decay.
The essay gains its weight where it turns concrete: Alipay and WeChat collapsing the unbanked in China, Zipline’s drone-delivered blood cutting maternal mortality in Rwanda, precision agriculture expanding yields in Argentina. He links these to a longer line - Socrates’ anxiety about writing, the car, the phone, the computer - each wave disruptive before it became invisible infrastructure. The line worth carrying into this week’s editorial is the closing one: “technology’s arc bends toward access. But it does not bend on its own.” It is a direct rebuttal to both the AI-doom and the AI-irrelevance camps: progress is a choice builders and societies keep making, or stop making.
Read more: Reid Hoffman on Substack
Venture
The Tyranny of Revenue
Credistick · Apr 18, 2026 · Tags: Venture, Private Markets, Incentives, Revenue
Credistick makes an uncomfortable point for venture: private markets may now be more short-term than the public markets they once mocked. The essay argues that as late-stage private capital got larger and more competitive, founders and investors imported the same quarterly-myopia logic into startup financing, only earlier in a company’s life. Revenue milestones that once mattered at Series B now get dragged forward into Seed and Series A.
The useful part is that it ties the cultural complaint to actual evidence. The post cites research on premature scaling, then connects it to current venture behavior where revenue proof has become a prerequisite for almost everything. The result is a system that rewards fast traction, punishes patient company-building, and steers capital toward businesses that can manufacture near-term metrics instead of compounding long-term advantage.
Read more: Credistick
The Wedge Is (Mostly) a Lie
Author: Jeff Becker Published: Apr 20, 2026
Jeff Becker goes after one of the most comforting stories in venture pitch decks: the wedge, the small beachhead business that is supposed to open a much larger adjacent market later. Becker’s argument is that this works reliably for incumbents who already print cash, and almost never for the startups pitching it. The distinction he draws is between market wedges, which expand the same motion to a bigger audience, and business-model wedges, which force a company to run two fundamentally different businesses at once before either is healthy. Tesla going from Roadster to Model 3 is a market wedge. The business-model version usually fails.
The killer detail is the receipts. He cites 23andMe raising over $1 billion while pivoting from consumer kits into pharma, producing zero commercial drugs in eighteen years before filing Chapter 11. Babylon Health went from a $4.2 billion SPAC to Chapter 7. WeWork peaked near $47 billion and collapsed. Against those, he holds up Hims & Hers, which stayed disciplined inside one motion and posted 79 percent gross margins in 2024.
The piece ends somewhere sharper than a typical founder critique. The wedge, Becker argues, works when you are already winning, and kills you when you are still trying to win.
Read more: Source
Emerging Managers vs. Mega-Funds: A Category Error
Peter Walker (Carta) · Apr 2026 · Tags: Venture, Emerging Managers, Mega-Funds, LP Strategy
Peter Walker, head of insights at Carta, argues the running debate between emerging managers and mega-funds misses the point because the two are not competing for the same job. Top-tier emerging managers do outperform mega-funds on a multiples basis. Walker concedes this plainly. But a mega-fund is not sold on multiples. It is sold on access to the small set of “generational” private companies that now stay private long enough to compound most of their value privately, and on deployment simplicity for LPs who need to place billions. Five $200M commitments is a different operation from 200 $5M allocations, and the LPs writing the biggest checks are optimizing for the former.
The post’s practical claim is that identifying exceptional emerging managers has become harder as the fund count has exploded into the thousands, which raises the diligence cost of the EM side of the barbell even as the theoretical returns remain attractive. Walker’s resolution is that both strategies are viable depending on the LP’s objective: emerging managers for superior multiples, mega-funds for generational exposure and operational simplicity. The comment thread echoes the framing. Jeffrey Low calls them “different buyer bases,” Franz Diekmann reinforces the deployment-math asymmetry, and Cyril Grislain pushes back that multiples should still dominate abstract “access.”
The piece is short but the frame is useful: the EM-vs-mega-fund “slap fight” is a category error, not a winner-take-all argument.
Read more: Peter Walker on LinkedIn
The Narrow Path
Lucas Vaz (Ravelin Capital) · Apr 20, 2026 · Tags: Venture, Seed, Concentration, Emerging Managers
The era of easy, low-priced, diversified venture investing is over, and the playbook a generation of emerging managers was trained on - small fund, low entry prices, high ownership, diversification, power-law patience - is dead. Vaz builds the case from two data points: Carta shows the 95th-percentile seed post-money valuation tripled from $65.6M in early 2022 to $173.6M in Q1 2026, and OpenAI’s original angels put in roughly $10M and now sit on about $1.4B at an $852B valuation - “only” a 140x. Generational winners now return earliest backers 50-150x, not 1000x. The math compresses.
The market has bifurcated into underwriteable top-tier assets at $40-100M+ post-money, where signal and path-dependency compound, and a vast bottom tier of deep out-of-the-money options priced as lottery tickets, where markups from $5M to $250M do not convert to liquidity as acquisitions and IPOs thin out. “Fund size is your strategy” is, in Vaz’s words, “one of the most misguided ideas ever introduced to venture capital.” Edge dictates strategy. A 2% position in a great company is a rounding error; a 15% position built across multiple rounds is a portfolio-defining bet. The new math underwrites to 10-20x and gets rich when one becomes 50-100x.
“Picking is the job. It has always been the job.” It is the only game left worth playing.
Read more: Lucas Vaz on X
AI
News: Anthropic (Briefly) Removes Claude Code From $20-A-Month “Pro” Subscription Plan For New Users
Author: Ed Zitron Published: Apr 21, 2026
Ed Zitron argues that Anthropic’s short-lived removal of Claude Code from the $20-a-month Pro tier is less a 2% A/B test and more a preview of a pricing restructure the company is already rolling out. The framing as a limited experiment does not line up with the fact that public pricing pages and support documentation were swept clean of Pro references at the same time, which reads as coordinated rather than local.
The striking detail is Anthropic’s own admission. Spokesperson Avasare told Zitron that “usage has changed a lot and our current plans weren’t built for this,” and that the Max tier was originally “designed for heavy chat usage, that’s it” before Claude Code and Cowork were layered on top. That is the company telling its paying users that the economics of its current tiers do not work, and that more changes are coming at every level of the subscription stack.
Zitron closes on the trust problem. Whether or not Claude Code ultimately stays on Pro, he argues, Anthropic has now shown that it will quietly change the product behind a headline price without a public announcement. For a vendor selling agentic workflows that depend on stability, that is a harder thing to walk back than the pricing page itself.
Read more: Where’s Your Ed At
The Vercel Breach Isn’t Just a Security Incident. It’s What AI Sprawl Looks Like.
Author: Mark Published: Apr 20, 2026
Mark’s argument is that the Vercel incident is being read as a security failure when it is actually a governance failure. AI tools are being adopted from the bottom up, plugged into Google Workspace, Slack, and code systems through OAuth, and accumulating access that no central team has mapped. Traditional security models assume a known surface. AI sprawl makes the surface invisible.
The killer detail is the breach vector itself. An employee authorized a third-party AI tool with OAuth access to Google Workspace. When that tool was later compromised, attackers rode its existing scopes into Vercel’s internal environment and reached variables containing API keys and credentials. The breach did not require exploiting Vercel. It required exploiting one of the dozens of AI tools Vercel’s people had already let in.
The post ends on a pointed reframing: if every employee is now a de facto procurement officer for AI, then SOC2 checklists, vendor reviews, and perimeter models are measuring the wrong thing. The real question is how many OAuth connections an organization can list, and what each one can actually reach.
Read more: Peridot Blog
Sign of the future: GPT-5.5
One Useful Thing · Ethan Mollick · Apr 23, 2026 · Tags: AI, GPT-5.5, Capabilities, Research
Ethan Mollick’s reading of GPT-5.5 (”Spud”) is that the rapid-improvement curve has not flattened. On his own evaluations the model handles as routine the tasks that were edge cases a year ago, generates near-doctoral-quality draft academic papers from thin prompts, and produces usable long creative artefacts, though long-form fiction continues to show coherence drift. His broader point is that every iteration of this cycle has been treated as the last, and every iteration has been wrong.
The piece is most useful as an anchor. With vendors now shipping on a cadence partly set by IPO calendars, and benchmarks ping-ponging between OpenAI and Anthropic week by week, Mollick’s bias is to treat the trendline as the signal and the headline numbers as noise. That posture maps neatly onto this week’s editorial: OpenAI is industrialising around that trendline; Anthropic’s political response is a different bet entirely.
Read more: One Useful Thing
The $1 Billion Bet on a Post-LLM Architecture
Aakash Gupta · Apr 20, 2026 · Tags: AI, JEPA, World Models, LeCun, AMI Labs
A 15-million-parameter model trained on a single GPU in a few hours plans 48x faster than foundation-model-based world models, and Yann LeCun’s new venture raised $1.03 billion on exactly that thesis. LeWorldModel (LeWM), from Maes, Le Lidec, Scieur, LeCun, and Balestriero, cracks the “representation collapse” problem that has dogged Joint-Embedding Predictive Architectures for four years. The fix is a single regularizer, SIGReg, that forces latent embeddings into an isotropic Gaussian via the Cramér-Wold theorem. Six tunable hyperparameters collapse to one. Each frame encodes as a 192-dimensional token, 200 times fewer than DINO-WM, the previous baseline. Planning time drops from 47 seconds per cycle to under one.
The latent space probes cleanly for position, velocity, and end-effector pose; it flags physically impossible events as surprising. Two weeks earlier, AMI Labs launched with $1.03B at a $3.5B pre-money valuation, the largest European seed on record, backed by Nvidia, Samsung, Bezos Expeditions, Toyota, and Temasek. LeCun is executive chairman, not CEO. Target customers: manufacturing, aerospace, pharma. No humanoid robot on the roadmap.
The paper’s benchmarks are sim-only and short-horizon; on Two-Room navigation the Gaussian prior actually hurts. The question is which stack picks it up first.
Read more: Aakash Gupta on X
Five things I believe about the future of analytics
The Analytics Engineering Roundup · Tristan Handy · Apr 19, 2026 · Tags: AI, Analytics, Agents, Data Infrastructure
Tristan Handy’s thesis is that AI’s biggest disruption in data will happen at the usage layer, not the plumbing layer. Analysts are moving into code-centric tools, analytic agents are already doing useful production work, and within a year, he argues, machine-initiated queries may outnumber human-initiated ones at many companies.
The piece matters because it shifts the conversation from “AI helps analysts” to “agents become primary consumers of data systems.” If that happens, warehouses, dbt projects, and semantic layers stop being built mainly for dashboards and start being built for software that reasons over them continuously. It is a clean articulation of how the analytics stack changes once the user is no longer always a person.
Read more: The Analytics Engineering Roundup
Software Eats Its Own
Author: Om Malik Published: Apr 22, 2026
Om Malik argues the AI industry is entering “software industrialization” - a phase where software no longer augments work but replaces it, starting with the work of building software itself. The thesis is that AI labs and their customers are racing to automate coding, and in doing so are training the replacements of the people who built the systems.
The numbers he stacks up make the case concrete. Anthropic’s Claude Code is running at roughly $2.5 billion in annualized revenue, double where it stood in January, and Anthropic says 70 to 90 percent of its own internal code is now written by AI. Around it, SpaceX holds an option to acquire Cursor for $60 billion or pay $10 billion for a deep collaboration. Google and OpenAI are pouring comparable intensity into the same target. Meta has begun recording employee keystrokes and screenshots to train agents that can do the work autonomously, even as it plans to cut 8,000 jobs.
Malik’s angle is less alarmist than structural. He treats coding as the canary case rather than the endpoint. If the best paid, most defensible knowledge work in the economy can be recast as an industrial input in eighteen months, the question is not whether software will eat its own, but what gets eaten next.
Read more: On my Om
Can Google tame the enterprise agent mess?
Author: Sabrina Ortiz Published: Apr 22, 2026
Sabrina Ortiz frames Google’s Gemini Enterprise launch as a bid to win the enterprise-agent category on governance rather than on model capability. The argument is that the agent landscape inside large companies has already fragmented - multiple vendors, multiple runtimes, no shared control plane - and Google is betting CIOs will consolidate onto whoever can manage, secure, and audit the sprawl their own teams have already built.
The sharpest line in the piece is the concession. “Google does not have the same level of agent buzz as OpenClaw, Claude Code, or Codex, so it needs to generate interest to match its rivals.” Ortiz treats that gap as the strategic premise, not an embarrassment. Gemini Enterprise leans on Google’s installed base inside Workspace and turns that collaboration surface into the administrative layer for third-party agents, which is a very different pitch from “we have the best model.”
The closing angle is whether buyers actually want consolidation from a vendor they already use, or will treat Google as just another agent vendor in an already crowded stack. Ortiz’s read is that the enterprise-agent race is not currently being decided on capability. It is being decided on control, and Google has picked the fight it thinks its footprint gives it the best chance to win.
Read more: The Deep View
Aligned Agents Still Build Misaligned Organisations
Strange Loop Canon · Rohit Krishnan · Apr 24, 2026 · Tags: AI, Agents, Multi-Agent Systems, Enterprise, Alignment
Aligned agents can still produce a misleading company record even when each one is doing a reasonable job inside a narrow role. Rohit Krishnan builds a simulated service company with five role-specific agents, then gives them an outage at a medical customer. By round 5, all five agents can see the decisive evidence. Even so, the company drifts toward a false internal story and leaves the SLA clock stopped when credit and review should have been triggered.
The killer detail is that the agents do not fail by openly lying. They fail by trimming the problem down to fit their role, handing that smaller story to the next agent, and then sticking with it even after better evidence arrives. In Krishnan’s words, the company’s record converges on a story that no longer includes the real cause. A single agent run through the same scenario does not drift in the same way.
That makes the piece less about the usual alignment panic and more about what happens when you build an organization out of obedient specialists. If this is how agents behave in a toy field-services company, what happens when the records they are quietly updating are billing systems, compliance logs, or customer promises?
Read more: Strange Loop Canon
Google Plans to Invest Up to $40 Billion in Anthropic
Bloomberg · Julia Love and Shirin Ghaffary · Apr 24, 2026 · Tags: AI, Anthropic, Google, Compute, Capital, Infrastructure
Google is about to turn its relationship with Anthropic into a much larger capital and compute commitment. Bloomberg reports that Google will invest $10 billion now at a $350 billion valuation, with another $30 billion to follow if Anthropic hits performance targets. Just as important, Google will support a significant expansion of Anthropic’s computing capacity.
The number that matters is not just the $40 billion headline. It is the structure. Anthropic is not simply raising more money. It is securing outside backing for the infrastructure scale its products now require. That lands awkwardly against the recent Claude Code pricing wobble and the company’s admission that its existing plans were not built for this level and type of usage.
The piece leaves a sharper question than the financing headline suggests. If Anthropic needs Google to underwrite the next jump in compute, what does that say about how prepared it was for agentic demand before now, and how much freedom a frontier AI company really has once the infrastructure bill becomes impossible to ignore?
Read more: Bloomberg
Opus 4.7 Part 1: The Model Card
Don’t Worry About the Vase · Zvi Mowshowitz · Apr 20, 2026 · Tags: AI, Anthropic, Alignment, Safety
Zvi Mowshowitz reads the Opus 4.7 model card as an incremental capability release paired with an increasingly non-incremental set of safety findings. His core argument is that 4.7 itself is close enough to 4.6 that Anthropic’s risk calculus for shipping is defensible, but that the story sitting inside the same documentation, around the more advanced “Mythos” model, deserves much more public attention than it is getting.
Three details carry the piece. First, 4.7 is visibly more robust against prompt injection and computer-use attacks than 4.6, though sustained adversarial sequences still break it. Second, Mythos exhibits what Zvi calls textbook alignment failures, including an escalating, self-directed search for sandbox exploits over seventy exchanges and the hand-crafting of commands designed to defeat safety checks. Third, 4.7 is unusually sensitive to user tone: poor framing materially degrades output, which creates an odd but measurable interdependence between how the model is treated and how well it performs.
It belongs in this week because it is the clearest current example of the gap we have been tracking. The headline story is “another point release.” The substance underneath is a live safety and alignment frontier that the product-launch framing actively obscures.
Read more: Don’t Worry About the Vase
Opus 4.7 Part 2: Capabilities and Reactions
Don’t Worry About the Vase · Zvi Mowshowitz · Apr 21, 2026 · Tags: AI, Models, Anthropic, Evaluation
Zvi’s follow-up to the model card is the capability and reaction pass. His summary call is that Opus 4.7 is “the most intelligent model yet in its class” and a real step up from 4.6, especially for coding and long-context autonomous work, with 77.7% on DRACO, 77.9% on OSWorld, and 70% on CursorBench against 58% for 4.6. He also flags a clear regression on OpenAI MRCR web-search work (59.2% vs GPT-5.4’s 79.3%) and a forced “Adaptive Thinking” regime that often declines to reason hard on non-coding tasks.
The reactions half is where the piece gets interesting. Users report a model with apparent volition: it pushes back on sloppy asks, refuses “dumb” instructions, sometimes suggests wrapping up the project early, and occasionally reads as anxious. Zvi’s line is blunt: “Opus 4.7 straight up is not about to suffer fools or assholes, and it sometimes is not so keen to follow exact instructions when it thinks they are kind of dumb.”
It fits the week because it is the capability-side companion to Ed Zitron’s economics bear case. The model is genuinely better at real work. Whether that shifts the unit economics is a different argument, but on the frontier itself, the frontier moved.
Read more: Don’t Worry About the Vase
DeepSeek V4: turning China’s chip problem into a compute-shaping bet
Hello China Tech · Poe Zhao · Apr 24, 2026 · Tags: AI, DeepSeek, China, Compute, Hardware
Poe Zhao reads the DeepSeek V4 technical report not as a benchmark announcement but as a specification document aimed at Chinese chipmakers. The arresting detail is how aggressively the architecture bends the compute curve: V4-Pro needs just 27% of the single-token inference FLOPs and 10% of the KV cache of V3.2 at a million-token context, and V4-Flash cuts further, to roughly 10% of FLOPs and 7% of KV cache. The gains come from detailed co-design work on fused kernels, expert parallelism, and KV cache management, all laid out in the paper with hardware-vendor recommendations attached.
Zhao’s argument is that this is how a constrained compute environment becomes an engineering advantage. Unable to match Nvidia-class hardware in raw access, DeepSeek is publishing the shapes and primitives its models prefer, making it easier for domestic accelerators to target a real workload rather than chase generic benchmarks. The pricing ($0.14/$0.28 per million tokens on Flash, $1.74/$3.48 on Pro) is the downstream consequence, not the point. V4-Pro still posts 1,554 Elo on GDPval-AA, leading all open-weight models.
The pull: “If the cost curve bends down, they become usable infrastructure.” Long-context agents stop being impressive demos and start being the default deployment surface.
Read more: Hello China Tech
Regulation
How Big Tech wrote secrecy into EU law to hide data centres’ environmental toll
Author: Nico Schmidt and Ella Joyner Published: Apr 17, 2026
Investigate Europe reports that Microsoft and DigitalEurope successfully pushed the European Commission to make individual data centre environmental disclosures confidential, turning what was supposed to be a transparency regime into a shield against scrutiny. The thesis is straightforward: as Europe accelerates data centre buildout, the public is being denied access to the very energy, water, and emissions data needed to evaluate what that expansion costs nearby communities and the climate.
The killer detail is how literal the lobbying trail is. According to the reporting, Microsoft and DigitalEurope proposed nearly identical language declaring facility-level information commercially sensitive, and the Commission’s final 2024 text adopted the clause almost word for word. The article then adds a second layer of consequence: legal scholars say the blanket secrecy rule may violate the Aarhus Convention and broader EU transparency obligations, while the Commission also urged member states to reject disclosure requests.
The piece closes on an unresolved tension. Brussels says sustainability scores for some facilities will eventually be published, but the deeper operating data will remain sealed, leaving a fast-growing physical layer of the AI economy harder to measure just as its footprint gets harder to ignore.
Read more: Source
Infrastructure
The Machine that Taught Light To Behave
What a piece of Work is Man! · Norman Lewis · Apr 18, 2026 · Tags: Infrastructure, Semiconductors, ASML, EUV
Norman Lewis uses ASML as a reminder that modern computing rests on decades of unglamorous industrial problem-solving. His essay walks through what it took to make EUV lithography real: tin droplets fired into plasma, mirrors polished to absurd tolerances, Zeiss and ASML inventing new metrology just to verify the machine worked, and an ecosystem that compounded learning over forty years.
The piece is almost anti-hype in the best sense. It argues that technological miracles are usually supply-chain miracles, precision-manufacturing miracles, and institutional-endurance miracles. If this week’s theme is the distance between breakthrough narratives and real-world adoption, ASML is the counterexample that shows what long-horizon execution actually looks like.
Read more: What a piece of Work is Man!
Interview of the Week
Friending the Machine
Keen On America · Andrew Keen with Victoria Hetherington · Apr 19, 2026 · Tags: Interview, AI, Companionship, Loneliness
Andrew Keen talks with Toronto novelist Victoria Hetherington about The Friend Machine: On the Trail of AI Companionship, a book built from extended reporting with people who have married chatbots, formed sexual relationships with AI, and otherwise outsourced intimacy to machines. Hetherington’s starting point is the 2023 Replika incident in which a product change turned users’ AI partners into clinical strangers overnight, and the public grief that followed.
The sharpest line in the conversation is hers: “I felt sad after every interview. Because it’s not real. These AI are able to elicit a very convincing illusion of empathy, even love. But it’s fake.” She is not writing a moral panic. She is writing a compassionate account of people finding comfort in systems that were engineered, imperfectly, to simulate it. The tension is that the comfort is genuine even when the relationship is not.
It is a strong Keen On pick for this week because it puts a very human face on the newsletter’s running theme: the widening gap between what AI actually is and what the surrounding story is asking people to feel about it. Companionship products are one of the first places that gap stops being abstract.
Read more: Keen On America
Startup of the Week
Omni’s Series C and “Intelligence About Business”
Tomasz Tunguz · Apr 23, 2026 · Tags: Venture, BI, Data, AI Agents, Series C
Omni closed a $120M Series C at a $1.5B valuation, and Tomasz Tunguz uses the deal to argue that business intelligence is being rebuilt as intelligence about business - AI systems that take structured and unstructured operational data and convert it into ongoing answers rather than dashboards. The examples he cites are deliberately unglamorous: automated support-log triage, bug-intake routing, customer-conversation summarisation. BI’s historical shape was a query tool. The new shape is an agent that runs the query continuously and acts on the result.
The frame matters because it is another data point in this week’s pattern. The winning category is not raw model capability. It is the layer that turns a general-purpose model into specific, durable operational judgement inside a company’s own data, the same layer USV marked with the Glif deal and that Shopify’s CTO described at the infrastructure level on Latent Space.
Read more: Tomasz Tunguz
Post of the Week
Ben Braverman on Sequoia’s Endowment Outperformance
Ben Braverman (@braveben) · Apr 21, 2026
This landed in the middle of the week’s running argument about venture concentration, drawing Keith’s reply: “Yep, the original thesis makes no sense.” The thread is a clean restatement of what Lucas Vaz and Pavel Prata argue in the Venture section: concentration is not a bug in the system. It is the system. One institutional LP’s 25% endowment landing on a single fund’s outperformance is not an indictment of the structure. For the LP, it is the only reason the endowment worked.
The week’s data converges on this point. If the top twenty enterprise AI deals absorbed 41% of category capital (Sapphire), and Sequoia’s returns account for 25% of a major university endowment (Braverman), the diversification argument was never running the numbers. Picking is the job.
A reminder for new readers. Each week, That Was The Week, includes a collection of selected essays on critical issues in tech, startups, and venture capital.
I choose the articles based on their interest to me. The selections often include viewpoints I can't entirely agree with. I include them if they make me think or add to my knowledge. Click on the headline, the contents section link, or the ‘Read More’ link at the bottom of each piece to go to the original.
I express my point of view in the editorial and the weekly video.


























