Latest Posts (20 found)
Jim Nielsen -30 days ago

You Might Debate It — If You Could See It

Imagine I’m the design leader at your org and I present the following guidelines I want us to adopt as a team for doing design work: How do you think that conversation would go? I can easily imagine a spirited debate where some folks disagree with any or all of my points, arguing that they should be struck as guidelines from our collective ethos of craft. Perhaps some are boring, or too opinionated, or too reliant on trends. There are lots of valid, defensible reasons. I can easily see this discussion being an exercise in frustration, where we debate for hours and get nowhere — “I suppose we can all agree to disagree”. And yet — thanks to a link to Codex’s front-end tool guidelines in Simon Willison’s article about how coding agents work — I see that these are exactly the kind of guidelines that are tucked away inside an LLM that’s generating output for many teams. It’s like a Trojan Horse of craft: guidelines you might never agree to explicitly are guiding LLM outputs, which means you are agreeing to them implicitly. It’s a good reminder about the opacity of the instructions baked in to generative tools. We would debate an open set of guidelines for hours, but if there’re opaquely baked in to a tool without our knowledge does anybody even care? When you offload your thinking, you might be on-loading someone else’s you’d never agree to — personally or collectively. Reply via: Email · Mastodon · Bluesky Typography: Use expressive, purposeful fonts and avoid default stacks (Inter, Roboto, Arial, system). Motion: Use a few meaningful animations (page-load, staggered reveals) instead of generic micro-motions. Background: Don't rely on flat, single-color backgrounds; use gradients, shapes, or subtle patterns to build atmosphere. Overall: Avoid boilerplate layouts and interchangeable UI patterns. Vary themes, type families, and visual languages.

0 views

Jensen Huang and Andy Grove, Groq LPUs and Vera CPUs, Hotel California

GTC 2026 marked an important inflection point for Nvidia, as the company is selling multiple architectures, instead of focusing on just one GPU. The motivation is serve all needs and keep all customers.

0 views

Exploiting brain flaws

In my “ closing thoughts ” post about the phone usage experiment, I mentioned I had deeper thoughts I wanted to share. Here I am, sharing those thoughts. I ran various month-long life experiments over the years, many of which I chronicled here on this blog. For that reason, the outcome of this recent phone experiment wasn’t a surprise: if I make the conscious decision to pay attention to some specific aspect of my life, there’s a high likelihood I’ll manage to enact significant changes in that specific area. Or so I thought. You see, I am a flawed human being, like many—most?—of the people out there. If I were in therapy, there would be a plethora of issues I’d be discussing with my therapist, but in therapy I am not, and so I thought it would be fun—for me at least, not sure about you—to tackle one of them here, since it’s strictly related to this recent phone experiment. «Wait a second, if that’s the case, then why aren’t you in therapy, Manu?» Good question, I’m glad you asked. There are two main reasons. The first, and less important reason, is that I am a stubborn motherfucker, and the idea of asking someone else to help me fix my inner issues is something that doesn’t sit right with me. The second, and more important reason, is that I have a fundamental distrust of psychologists. Not in psychology as a field, I have no issues with that. I even considered going into psychology back when I was about to finish high school and was thinking about possible career paths. I also read plenty of psychology books, and the book that had more impact on me growing up was a psychology book written by a psychologist. The issue I have with psychologists is that all the ones I had the pleasure to meet in person were deeply flawed and fucked up individuals, and that left an impression on me. Now I carry this fundamental (and partly irrational) distrust in them, which is a bit problematic since it’s hard to go to therapy when you don’t trust the person on the other side. Maybe this will change at some point in the future, who knows. I'm open to that possibility. Anyway, to get back on track, the issue I wanted to discuss is related to disappointment. Specifically, my issue with the concept of disappointing others. This is something I had to deal with since I was a kid, and I’m not sure why that is. I don’t know if it was triggered by something specific that happened or if it’s just part of my character, but disappointing others and especially the thought of seeing them disappointed because of something I did or didn’t do, is something I have always struggled with. To this day, I still do. The reason why I think this is all related to my weird life experiments is that those experiments usually follow a pattern: I experiment with something, I blog about it, I get to enjoy the benefits of some positive change, the experiment ends, I stop blogging about it, and slowly but surely the old habits manage to creep back in. It happens every time, like clockwork. But this time around, I realised that the reason why it happens is that I, fundamentally, do not give much of a fuck about myself. That itself is a topic for another time, but in the context of this discussion, the thing that matters is that as long as I’m blogging and I’m sharing my experience, the irrational pressure of disappointing someone keeps me on track. At a rational level, I know that no one gives a fuck if I fail at these silly experiments, and yet, for some reason, that extra pressure is what keeps me in check. Now, is this a healthy way to exist in this world? Probably not. Do I care? Definitely not. But, having realised this, I’m now wondering how I can exploit this to my advantage. Because there are things I’d love to change in my life, and I’m starting to think leveraging the disappointment-lever to my advantage could be the way to go. My phone usage, for example, is still under control, and that’s because I know I’m gonna keep sharing those numbers. Not weekly, because that’s boring, but probably every couple of months. And this fact alone, the decision of doing this, is apparently enough to keep my brain on track. Brains are weird, what can I say? I’m still figuring out which changes I want to put in place in my life. The tricky part is that they need to be trackable and shareable somehow; otherwise, this will not work, but I’m sure I’ll manage to come up with a solution. Thank you for keeping RSS alive. You're awesome. Email me :: Sign my guestbook :: Support for 1$/month :: See my generous supporters :: Subscribe to People and Blogs

0 views

Here’s my list of reasons for using Opencode

Here’s my list of reasons for using Opencode . I’m often experimenting with the bleeding edge models as they come out. I actively switch between models for tasks and I use them all enough where I can tell the difference. Opencode lets me switch between models mid-task or mid-conversation. Fluidly. I wrote about this and agentic fluidity in more detail but tldr: Opencode has the client/server architecture baked in. So I can just start an opencode server on one machine, expose it through and start using it on my phone or other machines. I talked about this on my podcast in some detail but Opencode has the best implementation of subagents and modes. You can switch to a subagent definition as your primary mode, then operate other subagents from there. It makes orchestrator-type tasks super easy. I love that OpenCode is opinionated about their UX. They don’t try to be Claude Code or Codex. In the process they have some really nice UX patterns like a sidebar with ongoing file changes, context/cost, MCPs connected etc. It’s the first time I’ve not needed to worry about a custom statusline.sh or building one. The plugin ecosystem is highly customizable. To the point where you can add new features, integrate with external services or even modify OpenCode’s default behavior. The wonderful Jesse Vincent mentioned this to me when I was stupidly contemplating a fork. It’s not all rainbows and sunshine. Anomaly — the team behind OpenCode — is small . Which sometimes shows, because there’s definitely bug s and some features missing . But I will say… none that’s deterred me from using it for the last two months, exclusively . Go give it a shot . Many of the serious AI coders I know are really liking it and switching.

0 views

GPT-5.4 mini and GPT-5.4 nano, which can describe 76,000 photos for $52

OpenAI today: Introducing GPT‑5.4 mini and nano . These models join GPT-5.4 which was released two weeks ago . OpenAI's self-reported benchmarks show the new 5.4-nano out-performing their previous GPT-5 mini model when run at maximum reasoning effort. The new mini is also 2x faster than the previous mini. Here's how the pricing looks - all prices are per million tokens. is notably even cheaper than Google's Gemini 3.1 Flash-Lite: I used GPT-5.4 nano to generate a description of this photo I took at the John M. Mossman Lock Collection : Here's the output: The image shows the interior of a museum gallery with a long display wall. White-painted brick walls are covered with many framed portraits arranged in neat rows. Below the portraits, there are multiple glass display cases with dark wooden frames and glass tops/fronts, containing various old historical objects and equipment. The room has a polished wooden floor, hanging ceiling light fixtures/cords, and a few visible pipes near the top of the wall. In the foreground, glass cases run along the length of the room, reflecting items from other sections of the gallery. That took 2,751 input tokens and 112 output tokens, at a cost of 0.069 cents (less than a tenth of a cent). That means describing every single photo in my 76,000 photo collection would cost around $52.44. I released llm 0.29 with support for the new models. Then I had OpenAI Codex loop through all five reasoning effort levels and all three models and produce this combined SVG grid of pelicans riding bicycles ( generation transcripts here ). I do like the gpt-5.4 xhigh one the best, it has a good bicycle (with nice spokes) and the pelican has a fish in its beak! You are only seeing the long-form articles from my blog. Subscribe to /atom/everything/ to get all of my posts, or take a look at my other subscription options .

0 views
fLaMEd fury Yesterday

Damn, I Can Still Read

What’s going on, Internet? Last December I finally got off my ass and committed to reading Jared Savages books, Gangland , Gangster’s Paradise , and the recently released Underworld . These books had been on my radar since the release of Gangland, but I was waiting on an ebook version. Then I went all in on audiobooks and decided to wait until they were available in audio format. So, back to December. It was my birthday. My wife sorted me some kid-free time so I dug out my Kobo Libra, charged it up a bit, reconnected to libby, borrowed Gangland and got stuck in. After hundreds of audiobooks and not much ebook reading outside of comics I thought I was in for a bad time. Much to my amazement I found out rather quickly that I could still read books with words, not sound. I also went through a period where I’d get into bed and snuggle in with a book only to find myself asleep after maybe getting through a single page. This made finishing books an audacious task. When I did switch to audiobooks, they became almost the only way I read. Night time reading defaulted to comic books, which I enjoyed but these have taken a back seat so far this year. I’ve got three months of X-Men to catch up on. I’ve read 10 books so far this year, four audiobooks and six books on the Kobo, a big change from previous years since I started my audiobook journey. I’ve got at least three more books lined up after the one I’ve just finished. After finishing the amazing 1985 I started another audiobook that just didn’t click so I quickly abandoned it before falling into sunk cost territory. I’ve picked up a few more podcasts to listen to during the day and have been listening to more music recently. I’m not worried though, I’m sure I’ll pick up the audiobooks again, just waiting for the right ones to make their way into my orbit. The question is, will the backlog of X-Men comics continue to grow or will I be able to find some balance in my physical reading? I just need some more of that kid-free time, right? Hey, thanks for reading this post in your feed reader! Want to chat? Reply by email or add me on XMPP , or send a webmention . Check out the posts archive on the website.

0 views
ava's blog Yesterday

i got featured @ noyb!

Every now and then, Noyb (European Center for Digital Rights) highlights some of their volunteers for their GDPRhub project. Now I got my entry :) Check it out on Mastodon and LinkedIn . Fittingly, one of my translated and summarized decisions finally made it into the weekly newsletter last Thursday! Also, they give you some goodies when you reach some of the volunteer milestones they have. I received mine :) Reply via email Published 17 Mar, 2026

0 views
David Bushell Yesterday

SMTP on the edge

Disclaimer: this post includes my worst idea yet! Until now my contact form submissions were posted to a Cloudflare worker. The worker encrypted the details with PGP encryption . It then used the Amazon AWS “Simple Email Service” API to send an email to myself. PGP encryption meant that any middleman after the worker, like Amazon, could not snoop. (TLS only encrypts in transit.) The setup was okay but involved too many services. If you thought that was over-engineered, get a load of my next idea. My experiment with a self-hosted SMTP server was short-lived but I did learn to code SMTP protocol with server-side JavaScript. During that tinkering I had issue upgrading TLS on the SMTP server for receiving email. In my recent AT Protocol PDS adventure I learned that Proton Mail can generate restricted tokens for SMTP client auth. I’ve also been slowly migrating from Cloudflare to Bunny in my spare time. I was reminded that Bunny has Deno edge workers. Lightbulb moment: can I rawdog SMTP in a Bunny worker? This cuts out the AWS middleman. Neither Bunny nor Proton ever see the unencrypted data. True end-to-end encryption for my contact form! I threw together a proof-of-concept. My script opened a TCP connection to Proton using and sent the SMTP message. The connection was upgraded with to secure it. It then followed a very fragile sequence of SMTP messages to authenticate and send an email. If the unexpected happened it bailed immediately. Surprisingly this worked! I’m not sharing code because I don’t want to be responsible for any misuse. There is nothing in Bunny’s Terms of Service or Acceptable Use policy that explicitly prohibits sending email. Magic containers do block ports but edge scripting doesn’t. I asked Bunny support who replied: While Edge Scripting doesn’t expose the same explicit port limitation table as Magic Containers, it’s not intended to be used as a general-purpose SMTP client or email relay. Outbound traffic is still subject to internal network controls, abuse prevention systems, and our Acceptable Use Policy. Even if SMTP connections may technically work in some cases, sending email directly from Edge Scripts (especially at scale) can trigger automated abuse protections. We actively monitor for spam and unsolicited email patterns, and this type of usage can be restricted without a specific “port block” being publicly documented. If you need to send transactional emails from your application, we strongly recommend using a dedicated email service provider (via API) rather than direct SMTP from Edge Scripting. bunny.net support …that isn’t an outright “no” but it’s obviously a bad idea. To avoid risking an account ban I decided to use the Bunny edge worker to forward the encrypted data to a self-hosted API. That service handles the SMTP. In theory I could decrypt and log locally, but I’d prefer to let Proton Mail manage security. I’m more likely to check my email inbox than a custom GUI anyway. The OpenPGP JavaScript module is a big boy at 388 KB (minified) and 144 KB (compressed). I load this very lazily after an event on my contact form. Last year in a final attempt to save my contact form I added a Cloudflare CAPTCHA to thwart bots. I’ve removed that now because I believe there is sufficient obfuscation and “proof-of-work” to deter bad guys. Binning both Cloudflare and Amazon feels good. I deleted my entire AWS account. My new contact form seems to be working. Please let me know if you’ve tried to contact me in the last two weeks and it errored. If this setup fails, I really will remove the form forever! Thanks for reading! Follow me on Mastodon and Bluesky . Subscribe to my Blog and Notes or Combined feeds. PGP encryption in the browser to Bunny edge worker SMTP directly to Proton

0 views
Martin Fowler Yesterday

Context Anchoring

Conversations with AI are ephemeral, decisions made early lose attention as the conversation continues, and disappear entirely with a new session. Rahul Garg explains how Context Anchoring externalizes the decision context into a living document.

0 views

TiNA: Tiered Network Buffer Architecture for Fast Networking in Chiplet-based CPUs

TiNA: Tiered Network Buffer Architecture for Fast Networking in Chiplet-based CPUs Siddharth Agarwal, Tianchen Wang, Jinghan Huang, Saksham Agarwal, and Nam Sung Kim ASPLOS'26 Here we go again , another paper in a top-tier conference on the classic CS problem: how to DMA received packets from NIC to host. It would be interesting to understand why this is such a hot topic these days. This paper deals with the case where the host CPU comprises multiple chiplets. If you get nothing else from this, I hope you will learn something about SNC mode (I had not heard of it before). Recent Intel CPUs can be placed into Sub-NUMA Clustering mode (via a BIOS setting). This causes each chiplet to appear as a separate NUMA node. It is like a single socket CPU is transformed into a 4 socket CPU. The DRAM memory space is divided into four regions (one per chiplet), and the LLC slices within a chiplet only cache data from one memory space. This can be advantageous for some applications, because it can lower average LLC and DRAM access latency (by avoiding inter-chiplet communication). The downside is that the peak LLC capacity available to a single core is reduced. Fig. 3 illustrates these tradeoffs: Source: https://dl.acm.org/doi/10.1145/3760250.3762224 SNC and DDIO Recall that DDIO is a feature of Intel CPUs that allows a NIC to write received packets directly into the LLC, which the host CPU can then read. PCIe lanes are distributed among chiplets. This means that the NIC is directly connected to one chiplet. One way to support DDIO with SNC is to allocate buffers for received packets in the memory region associated with the chiplet that the NIC is connected to. This improves LLC bandwidth (for both the NIC and CPU cores) but decreases the LLC capacity available for network packets. In practice, this means that longer bursts of network packets degrade performance more when SNC is enabled (i.e., leaky DMA is a larger problem in SNC mode). Fig. 6 has data from a microbenchmark to back this up: Source: https://dl.acm.org/doi/10.1145/3760250.3762224 TiNA The solution proposed by this paper requires a change to the NIC/driver interface. Each ring buffer of received network packets is replaced by ring buffers (where is the number of chiplets). Ring buffer is placed in the memory region associated with chiplet . The NIC knows about all of these ring buffers and dynamically decides which one to use. The NIC prefers to use the ring buffer associated with the chiplet that it is directly connected to. However, if a burst of traffic causes high utilization of the LLC capacity of that chiplet, then the NIC will fall back to using the other ring buffers. The NIC estimates LLC utilization based on two competing rates: The rate that received network packets are produced by the NIC The rate that received network packets are consumed by the host The first rate is easy for the NIC to compute as it knows how fast it is sending bytes to the host. The second rate is computed by networking software running on the host, and periodically sent to the NIC. The overall approach reminds me of CEIO . The key difference is the set of memory segments available. CEIO uses NIC-local DRAM as the fallback path. One complication of splitting a single ring buffer into multiple is ensuring that the host processes received packets in order. This paper proposes using sequence numbers associated with each packet. Most protocols already use per-packet sequence numbers. For other protocols (e.g., UDP), the NIC adds a sequence number based on the order in which packets were received. When the host reads a packet from a logical ring buffer, it examines the sequence numbers from the packets at the head of each of the physical ring buffers and chooses the packet with the lowest sequence number. Fig. 9 has benchmark results: lower latency than SNC and non-SNC across a range of microbenchmarks. Source: https://dl.acm.org/doi/10.1145/3760250.3762224 Dangling Pointers It would be nice if SNC allowed more fine-grained configuration. For example, there may be applications where ideal performance is achieved if each CPU core only has access to the L3 slice that is directly connected to it. Subscribe now Source: https://dl.acm.org/doi/10.1145/3760250.3762224 SNC and DDIO Recall that DDIO is a feature of Intel CPUs that allows a NIC to write received packets directly into the LLC, which the host CPU can then read. PCIe lanes are distributed among chiplets. This means that the NIC is directly connected to one chiplet. One way to support DDIO with SNC is to allocate buffers for received packets in the memory region associated with the chiplet that the NIC is connected to. This improves LLC bandwidth (for both the NIC and CPU cores) but decreases the LLC capacity available for network packets. In practice, this means that longer bursts of network packets degrade performance more when SNC is enabled (i.e., leaky DMA is a larger problem in SNC mode). Fig. 6 has data from a microbenchmark to back this up: Source: https://dl.acm.org/doi/10.1145/3760250.3762224 TiNA The solution proposed by this paper requires a change to the NIC/driver interface. Each ring buffer of received network packets is replaced by ring buffers (where is the number of chiplets). Ring buffer is placed in the memory region associated with chiplet . The NIC knows about all of these ring buffers and dynamically decides which one to use. The NIC prefers to use the ring buffer associated with the chiplet that it is directly connected to. However, if a burst of traffic causes high utilization of the LLC capacity of that chiplet, then the NIC will fall back to using the other ring buffers. The NIC estimates LLC utilization based on two competing rates: The rate that received network packets are produced by the NIC The rate that received network packets are consumed by the host

0 views
ava's blog Yesterday

radically accepting my flawed and dysfunctional body

I remember years ago, especially early on in the pandemic (2020/2021), I was still not diagnosed with my illnesses ( Bechterew's disease and Crohn's disease ). For a decade at least, I had dealt with a variety of symptoms, most of it around joints, my spine, and my digestive tract, and separately from those, also hormonal issues. Food was unpredictable and made me feel sick and caused me a lot of pain, and all the inflammation showed on my skin too: Around that time, gut health information was really booming online (probably still is, but I keep away from that content now and I'm less online). The idea was that by cutting out certain stuff or mostly eating this or that diet or taking these supplements would regenerate your gut health and make all the symptoms go away - the joint pain, the sluggishness, the acid reflux, the rashes, the hormone imbalances, the allergies and intolerances, and so on. If you see your body as a naturally wholesome and healthy body that is just temporarily imbalanced by some exposure and can be brought back into balance, these products and lifestyle changes are basically the magic pill. Just do this and avoid expensive pharmaceutical drugs with side effects! I'm not trying to act like that can never happen; people have successfully reversed or lessened some illnesses and issues by eating differently, working out, losing weight or limiting their exposure to something. But for me, this approach just led to disordered eating habits and holding off on effective treatment in some things for a while. The thing is, lots of people online peddling this stuff are in the business of snake oil. Buy their classes, their book, their supplements to finally be free from all these issues that doctor's can't or won't diagnose or only have evil medicines for that have side effects! Your body is good as is, it just needs a nudge in the right direction! It puts so much responsibility on you. Yes, we should limit our exposure to pesticides, PFAS etc., but you go insane in the grocery store thinking: " I can't buy this, it's not organic, can't buy this, it's wrapped in plastic, can't buy this, it's canned, can't buy this, it's high inflammatory/against FODMAP diet, can't buy this because it's too processed, can't buy this because it has so much sugar... ". Back then, every grocery store trip had me on the verge of a mental breakdown or actually breaking down. Everything felt contaminated, unsafe, or something my body couldn't tolerate. It felt impossible to " treat my body naturally " or bring it " back into balance ". Even when you do manage for a while, it significantly inhibits your ability to socialize with people because so much of it is about food: going out to eat together, attending festivities, being invited to dinner, being gifted food, traveling. A very restrictive diet can also cause deficiencies or starve you. It's also a bottomless pit: If it doesn't work for you and you don't see results, they say you need to try harder, also cut out this and that, buy this other supplement, and now consider other areas of your life too. Aggressively filter all your water, move away from any kind of busy street to limit the exhaust fume exposure, have your home checked for mold, switch out all your synthetic dyed clothes for unbleached undyed linen, switch out all your cooking utensils and pans to the "non-toxic" varieties, check if you live near some kind of coal plant or electricity lines or so, and if you are in the really weird circles, you will hear about chemtrails and Electromagnetic Hypersensitivity and all that. Yes, mold exposure and harmful substances in water are a problem, but I'm just saying: Doing all this next to everything else in life is a huge undertaking, mentally taxing, making people extremely paranoid and isolated, and bleeding them dry when it's often not even the issue . It's taking advantage of vulnerable people who either have no access to healthcare or aren't taken seriously or cannot afford the testing or medication required. It's good when one simple change can genuinely help you - for example, I know what foods not to eat to avoid triggering acid reflux. I love it for you if you figured out that eating gluten was behind it all and are now happy and healthy. But my body was never a naturally healthy and balanced one that got out of whack by some behavior or exposure, and even if it happened because of exposure in utero, or as a child, or just living in our modern environment nowadays, I can't undo or change that. My body, in its natural state, is not normal or healthy, and all that helps is proper medication. It's not temporary, this is just how my body functions. The baseline I was born with isn't the norm, and as experience showed, no amount of gut health stuff or limiting exposure or other lifestyle changes were going to change that. All that helped was finally getting properly diagnosed and receiving treatment . It was easy for me to accept treatment for the above issues because life had become unlivable with my intense flare ups and affected by daily ability to function all the time, and any possible side effect was worth the risk. I still don't regret any of it, and it works fine for me. Where I struggled to seek and accept help was for my hormone issues, as they only affected me every other month or so and were easier to ignore otherwise. As I talked about in a different post, I received hormone therapy as early as 11 years old because my periods and hormone levels were not normal and I otherwise wouldn't have developed how I am expected to as a cis woman 1 . I needed T-blockers like cypro to have the puberty my body and mind needed 2 . I stopped at 19 or 20 because I had started having issues with pain and spotting for a while and thought I could try and see whether after puberty, my situation had resolved and I'd naturally have the hormone levels I needed. It hadn't. So since then, I either took nothing, or tried reigning in my PCOS and endometriosis with things like Maca root powder. It did bring down my cycle days from 60-70 down to 30, but other issues still persisted. Lots of menstrual pain, flareups of my other issues, PMDD , and so on. When I still went to therapy years ago, my therapist suggested getting antidepressants to take just for the phase between ovulation and period, so I'd stop feeling the awful effects of PMDD. I declined, because while I had been on antidepressants previously for a while and they helped, I also knew what it was like to start and stop them, and I didn't want to constantly put my body through that; plus, the scary side effects! The same happened with hormone treatment. Even though I had spent years of my life on artificial hormones, I was scared to go back on it because I couldn't rule out that they had played a part in my depression back then (or at least amplified it). I was also scared of thrombosis, meningioma and other issues 3 . I thought it would just naturally fade away, or I could make without until menopause, or later: My treatment for my Bechterew's and Crohn's will finally bring my body into natural alignment! At first, it looked like it; I suddenly experienced cycles like a normal person. On time, barely or no pain, very light bleeding. But it went back to how it was over months, even after switching from infliximab to adalimumab. So turns out, fighting the inflammation in my body didn't do anything to normalize my hormones. I wrote something about accepting my natural menstrual cycle that retroactively is just a huge cope. There it was again, the idea that there is a natural state a body can return to and that everyone's default state is automatically healthy, now warped into the idea that I was just naturally meant to have elevated androgens and all this, and that I should just accept how it is. The idea that natural is automatically good is such an easy fallacy to fall prey to, and natural also meant unmedicated to me. I tried to find so many reasons for why being so destroyed by my cycle every time was actually somehow a good thing or had any advantages. There's no shortage of supposedly empowering and encouraging content online about this as well: People who present having a cycle as something magical and romanticizing it as living with the moon tides or living in tune with nature. Just be proud of it and feel like those TikTok witches brewing your own herbal solution and gulping it down with some pumpkin seed oil. Ugh! Recently, I just grew tired of it all. The weeks of feeling sluggish, moody, forgetful and weak; my Crohn's and Bechterew's flaring up with it every time; feeling suicidal and calling in sick due to menstrual pain. 2-3 weeks until I felt normal again derailed good routines and fitness goals all the time, and it was hard to plan around such an irregular cycle. These times could fall on important dates at work or in my degree (exam season etc.) and jeopardize my reliability and skills. If I wanted to reach the goals I had set myself and would thrive in and feel the happiest in, I needed to address this. I owe myself that. No one will ever notice your avoidable suffering and pat you on the back for enduring it when there is another way. You aren't impressing anyone with choosing "natural" over comfortable and happy. All people will see and remember are the times you seemed unhappy, uncomfortable, snappy or missed out on being even being there. In that one post about accepting my cycle, I wasn't actually accepting it. I see now that to actually accept my sick body, it also means accepting treatment where possible . Everything else is not acceptance, it's just giving up and ignoring the issue. So recently, I had my yearly checkup at the gynecologist and finally got help. I am very lucky to have a very attentive and knowledgeable gynecologist 4 , and we went through all the options with pros and cons, also in connection with my Crohn's that can affect absorption, and we settled on dienogest daily and skipping my period altogether. Independently of that, I finally accepted that my hair needs additional help as I am prone to telogen effluvium and androgenic alopecia , and if I am regrowing it now since cutting it off in October 2024 due to losing like half my hair back then, I need to do something. So I am trying out minoxidil on top of going back to scalp massages and all. I know seeking medical help can be daunting, stressful, humiliating, costly, inaccessible, and scary. I almost cancelled that appointment about four times. But I hope it motivates you to seek help for the thing you put off or gave up on. You don't need to suffer, you don't need to self-sabotage or prove it to yourself, and you weren't " meant to be like this ". If " natural remedies " or snake oil and obsessive rules don't work for you, allow yourself to accept proper help. Reply via email Published 17 Mar, 2026 This is also why I have very small hands and feet, and remained at an average size. I was expected to become 1,80m tall, now I am just 1,66m, with a EU shoe size of 36/37. I didn't change that much from that age in terms of size. ↩ Yes, they do that for cis children, so stop clutching your pearls about trans children getting the same care! ↩ This is unfortunately what happens when you work with medical data, particularly side effects and adverse events; you know way too much about some meds. ↩ She's always been great, but it felt like in the year since we last saw each other, she went extra hard in researching how my illnesses can interact with my cycle before I showed up. ↩ This is also why I have very small hands and feet, and remained at an average size. I was expected to become 1,80m tall, now I am just 1,66m, with a EU shoe size of 36/37. I didn't change that much from that age in terms of size. ↩ Yes, they do that for cis children, so stop clutching your pearls about trans children getting the same care! ↩ This is unfortunately what happens when you work with medical data, particularly side effects and adverse events; you know way too much about some meds. ↩ She's always been great, but it felt like in the year since we last saw each other, she went extra hard in researching how my illnesses can interact with my cycle before I showed up. ↩

0 views
Stratechery Yesterday

An Interview with Nvidia CEO Jensen Huang About Accelerated Computing

Listen to this post: Good morning, This week’s Stratechery Interview is running early this week, as I had the chance to speak in person with Nvidia CEO Jensen Huang at the conclusion of his GTC 2026 keynote , which took place yesterday in San Jose . I have spoken to Huang four times previously, in May 2025 , March 2023 , September 2022 , and March 2022 . In this interview we talk about a keynote that came across like a bit of a history lesson, and what that says about a company that still feels small even as it’s the most valuable in the world, as well as what has changed in AI over the last year. Then we discuss a number of announcements that might feel like a change in approach (although Huang disagrees), including Nvidia’s burgeoning CPU business and the Groq acquisition. Finally we discuss scarcity in the AI stack and how that affects Nvidia, the China question, and Huang’s frustration with doomers and their influence in Washington. As a reminder, all Stratechery content, including interviews, is available as a podcast; click the link at the top of this email to add Stratechery to your podcast player. On to the Interview: This interview is lightly edited for clarity. Jensen Huang, welcome back to Stratechery. JH: It’s great to be with you. You literally just walked off the stage, went a little long, I think, but you spent a lot of this keynote , which I quite enjoyed, explaining what Nvidia is, starting with the history of the programmable shader, the launch of CUDA 20 years ago. We don’t need to spend too much time recounting this, you did a good job, and Stratechery readers are certainly familiar — sorry, this is a bit of a lead up here — Stratechery readers are familiar , and I remember this exactly, someone asked me to explain how is it that Nvidia can announce so many things at a single GTC, this is like six, seven years ago, maybe even longer than that, and I explained the whole thing with CUDA and all the libraries is it’s just sort of doing the same thing again and again , but for specific industries. That’s the story you told today, and it’s kind of a back-to-the-future moment after the last few GTC keynotes have kind of just been pretty AI-centered, CES was pretty AI-centered . Why did you feel the need tell that story now? To recast CUDA and why is it important? JH: Well, because we’re going into a whole lot of new new industries and because AI is going to use tools, and when AI uses tools, those are tools that we created for humans. AI is going to use Excel, AI is going to use Photoshop, AI is going to use logic synthesis tools, Synopsis tools, and Cadence tools. Those tools have to be super-accelerated, they’re going to use databases they have to be super-accelerated because AI’s are fast. And so I think in this era, we need to get all of the world’s software now as fast as possible accelerated, and then put them in front of AI so that AI could agentically use them. So is that a bit where we’ve already done this for a bunch of sectors and now we’re going to do it for a bunch more? JH: Yeah, a whole bunch more. For example, data processing. Well, that was sort of a surprise. I didn’t expect you to be opening with an IBM partnership . JH: Yeah, right, that kind of puts it in perspective. I mean, they really started it all. You wrote last week that AI is a five-layer cake : power, chips, infrastructure, models, and applications. Is there a concern that in the last four or five years, that you are worried about being squeezed into the chips box, so it’s important to both remind people and also yourselves about you being this vertically integrated company — not just in terms of building systems, but into the entire software stack, you’re not just a chip company. JH: I guess my mind doesn’t start with, “What I’m not”, it starts with, “What do we need to be?”. And back then, we realized that accelerated computing was a full stack problem, you have to understand the application to accelerate it. We realized that we had to understand the application, we had to have the developer ecosystem, we needed to have excellent expertise in algorithm development, because the old algorithms that were developed for CPUs don’t work well for GPUs, so we had to rewrite, refactor algorithms so that they could be accelerated by our GPUs. If we do that, though, you get 50 times speed up, 100 times speed up, 10 times speed up, and so it’s totally worth it. I think since the very beginning, we realized, “Ok, what do we want to do, and what does it take to achieve that?”. Now, today we’re building AI factories, we’re building AI infrastructure all over the world. That’s a lot more than building chips, and building chips is obviously important, it’s the foundation of it. Right, that’s like one full stack of doing the networking and doing the storage, and now you’re into CPUs. JH: Now you’ve got to put it all together into these giant systems — a gigawatt factory is probably $50, $60 billion. Out of that $50, $60 billion, probably about, call it $15, $17 or so, is infrastructure: land, power, and shell. The rest of it is compute and networking and storage and things like that, and so that level of investment, unless you’re helping customers achieve the level of confidence that they’re going to succeed in building it, you just have no hope, nobody’s going to risk $50 billion. So I think that that’s the big idea, that we need to help customers not just build chips, but build systems and then after we build systems, not just build systems, but build AI factories. AI Factories has a lot of software inside, it’s not just our software, it’s a ton of software for cooling management and electricals and things like that, and redundancies and a lot of it is over-designed, it’s over-designed because nobody talked to each other. When you have a lot of people who don’t talk to each other, integrate systems, you have to, by definition, over-design your part of it. But if we’re working together as one team, we’ll make sure that we can push the limits and get more throughput out of the power that we have or save money for whatever throughput you want to have. Just to go back to that software bit, you mentioned Excel wasn’t designed to be used by AI. You have things like Claude has this new functionality to use Excel , so when you talk about that, you want to invest in these libraries, is that to enable models like that to do better? Or is that something for Microsoft or for enterprises — you want to use this, you don’t want to be beholden to this sort of other player in the world? JH: Well, SQL’s a good example. SQL’s used by people, and we bang on the SQL systems like anybody else, and it is the ground truth of businesses. Well, it’s not just gonna be people banging on our SQL database now, it’s gonna be a whole bunch of agents banging on it. Right, they’re gonna do it way faster. JH: They’re gonna need to do it way faster. And so the first thing we have to do is accelerate SQL, that’s kind of the simple logic of it. That makes sense. In terms of models, you noted that language models are only one category. “Some of the most transformative work is happening in protein AI, chemical AI, physical simulation, robots, and autonomous systems”, and this is from the piece you wrote last week. You’ve previously made this point while noting in other keynotes, “Everything is a token”, I think, is a phrase that you’ve used before. Do you see transformers as being the key to everything, or do we need new fundamental breakthroughs to enable these applications? JH: We need all kinds of new models. For example, transformers, its ability to do attention scales quadratically, and so how do you have quite long memory? How can you have a conversation that lasts a very long time and not have the KV cache essentially become, over time, garbage? Or have entire racks of solid-state drives that are holding KV cache . JH: And of course, let’s say that you were able to record all of our conversation, when you go back and reference some conversation, which part of the reference is most important? There needs to be some new architecture that thinks about attention properly and be able to process that very quickly. We came up with a hybrid architecture of a transformer with an SSM, and that is what enables Nemotron 3 to be super intelligent and super efficient at the same time, that’s an example. Another example is coming up with models that are geometry aware, meaning a lot of things in life, in nature, are symmetrical. And so when you’re generating these models, you don’t want it to generate what is just statistically plausible, it has to also be physically based, and so it has to come out symmetrical. And so cuEquivariance , for example, allows you to do things like that. So we have all these different technologies that are designed — or, for example, when we’re generating tokens in words, it comes out in chunks at a time, little bits, tokens at a time, when you’re generating motion, you need it to be continuous. And so there’s discrete information that you generate and understand, and there’s continuous information that you want to generate and understand. Transformers is not ideal for both. Right, that makes sense. One more quote from the piece, you write, “In the past year, AI crossed an important threshold. Models became good enough to be useful at scale. Reasoning improved. Hallucinations dropped. Grounding improved dramatically. For the first time, applications built on AI began generating real economic value”. What specifically was that change? Because I think about the timing, I feel like this upcoming year is definitely about agents, I just wrote about it today — but for last year, was that the reasoning? Was that the big breakthrough? JH: Generative, of course, was a big breakthrough, but it hallucinated a lot and so we had to ground it, and the way to ground it is reasoning, reflection, retrieval, search, so we helped it ground. Without reasoning, you couldn’t do any of that, and so reasoning allowed us to ground the generative AI. And once you ground it, then you could use that system to reason through problems and decompose it, and decompose it into things that you could actually do something about, and so the next generation was tool use. Turns out it probably tells you something that search was a service that nobody paid for, and the reason for that is getting information is very important and very useful but it’s not something you pay for. The bar to reach to get somebody to pay you for something has to be higher than just information. “Where’s a good restaurant?” — information is just, I don’t think is worthy enough to get paid for. Some people pay for it, I pay for it. We now know that we’ve now crossed that threshold. Not only is it able to converse with us and generate information for us, it can now, of course, do things for us. Coding is just a perfect example for that. If you think about it for a second, you realize this, coding is not really the same modality as language, you have to teach it empty spaces and indentations and symbols, it’s almost like a new modality and you can’t generate code just one token at a time, you have to reflect on the chunk of code. That chunk of code has to be factored properly and has to be optimal and has to obviously compile, it has to be grounded not on probable truth, it has to be grounded on execution. Right, does it run or not? JH: It has to run or not. And so I think the code, learning that modality was a big deal. Once you’re able to now do — we pay engineers several hundred thousand dollars a year to code, and so now they have a coding assistant. They could think about architecture. Instead of describe programs in code, which is very laborious, they can now describe software in specification, which is much more abstract and allows them to be much more productive. And so they describe specification, architecture, they’re able to use their time to solve and innovate, and so our software engineers 100% use coding agents now. Many of them haven’t generated a line of code in a while, but they’re super productive and super busy. Do you think there is a temptation to over-extrapolate from coding, though, precisely because it’s verifiable? You have this agent idea where they can go — it’s not just that they will generate code, then they can actually verify it, see if it works, if it doesn’t, they can go back and do it again, and this can happen all without humans because there’s a clear, “Does it work or not?”. JH: Well, because you can reflect, you could have, let’s say, design a house. Designing a house or designing a kitchen used to be the work of architects, designers, but now you could have carpenters do that. So now you elevated the capability of a carpenter, now you use an agent for that carpenter to go design a house, design a kitchen, come up with some interesting styles. The agent doesn’t have some tool to execute. However, you could give an example. You say, “these are the styles I’m looking for, I want it to be aesthetic like that”. Because the agent is able to reflect, is able to compare its quality of code, its quality of result against some reference, it could say, “You know what, it didn’t turn out as well as I hope, I’m going to go back at it again”, and so it iterates. It doesn’t have to be fully executable, in fact, the more probabilistic, the more aesthetic, the more subjective, if you will, AI actually does better. Right, well that’s why you almost have two extremes. You have generating images where there’s no right answer and then you have coding where there is a right answer and AI seems to do good on those sides and the question is how much will it collapse into the middle there. JH: We’re fairly certain it could do architecture now, we’re fairly certain it could design kitchens and living rooms. Well, to this point, one of the big things with agents coming online is, you’ve talked a lot about accelerated computing, I think you’ve trash talked as it were, maybe the CPUs to the day, they’re all gonna be removed, like everything’s gonna be accelerated. Suddenly CPUs are hot again. It turns out they’re pretty useful and important to the extent you are selling CPUs now, how’s it feel to be a CPU salesman ? JH: There’s no question that Moore’s law is over. Accelerated computing is not parallel computing. Go back in time — 30 years ago, there were probably 10, 20, 30 parallel computing companies, only one survived, Nvidia and the reason why is because we had the good wisdom of recognizing the goal wasn’t to get rid of the CPU, the goal was to accelerate the application. So what I just falsely accused you of was actually true for everybody else. JH: We were never against CPUs, we don’t want to violate Amdahl’s Law . Accelerated computing, in fact, inside our systems, we choose the best CPUs, we buy the most expensive CPUs, and the reason for that is because that CPU, if not the best and not the most performant, holds back millions of dollars of chips. When it comes to branch prediction , you worried about wasting CPU time, now you’re worried about wasting GPU time. JH: That’s right, you just never can have GPUs be squandered, GPU time be idle. And so we always use the best CPUs to the point where we went and built Grace so that we could have the highest performance single-threaded CPU and move data around a lot faster. And so accelerated computing was never against CPUs, my basis is still true that Amdahl’s Law is over, the idea that you would use general purpose computing and just keep adding transistors, that is so dead, and so I think fundamentally we’re not against CPUs. However, these agents are now able to do tool use, and the tools that they want to use are tools created for humans and they’re basically two types. There’s the stuff that we run in data centers and most of it is SQL, most of it is database related, and the other type is personal computers. We’re now going to have AIs that are able to learn unstructured tool use, the first type of tool use is structured. CLIs are tool use, APIs, they’re all structured tool use, the commands are very explicit, the arguments are explicit, the way you talk to that application is very specific. However, there’s a whole bunch of applications that were never designed to have CLIs and APIs and those tools need AIs to learn multi-modality, unstructured, and it has to go and be able to go surf a website and it has to be able to recognize buttons and pull down menus and just kind of work its way through it like we do. That tool use are going to want to use PCs and we have both sides, we have incredibly great data processing systems, and as you know, Nvidia’s PCs are the most performant in the world. So what makes an agent-focused CPU different from other CPUs? So you’re going to have a rack of just Vera CPUs. JH: Oh, really good, excellent. So the way that CPUs were designed in the last decade, they were all designed for hyperscale cloud and the way that hyperscale cloud monetizes CPUs is by the CPU core. So you want to design CPUs that have as many cores as possible that are rentable, the performance of it is kind of secondary. You’re dealing with web latency by and large. JH: That’s exactly right, exactly. And so the number of CPU instances is what you’re optimizing for. That’s why you see these CPUs with a couple of hundred, 300, 400 cores coming. Well, they’re not performant and for tool use, where you have this GPU waiting for the tool use— And you’re going over NVLink. JH: That’s right, you want the fastest single-threaded computer you can possibly get. So is it just the speed? Or does the CPU itself need to be increasingly parallel so it doesn’t have misses and things like that? Or so it’s like just all the way down the pipeline is very different? JH: Yeah, the most important thing is single-threaded performance and the I/O has to be really great. Because it’s now in the data center, the number of single-threaded instances running is going to be quite high and therefore, it’s going to bang on the I/O system, it’s going to bang on the memory controller really hard. Vera’s bandwidth-per-CPU core, bandwidth-per-CPU, is three times higher than any CPU that’s ever been designed, and so it’s designed so that it has lots and lots of I/O bandwidth and lots and lots of memory bandwidth, so that it never throttles the CPU. If the CPU gets throttled, then we’re holding back a whole bunch of GPUs. Is this Vera rack, is it still, you talked about it being very tightly linked to the GPU rack, but is it still disaggregated so that the GPUs can be serving multiple different Vera cores? Whereas you have a Vera core on a board with- Okay, got it, that makes sense. How does your Intel partnership and the NVLink thing fit into this, if at all? JH: Excellent. Some of the world is happy with Arm, some of the world still needs, particularly, you know, enterprise computing, a whole bunch of stacks that people don’t want to move and so x86 is really important to that. Has the resiliency of x86 code been surprising to you? JH: No. Nvidia’s PC is still x86, all of our workstations are x86. I did want to congratulate you, as you talked about in the keynote today, you are the token king . So in your article, you also talked about that energy is the first principle of AI infrastructure and the constraint on how much intelligence the system can produce. If that’s the case, if it’s the amount of tokens you can produce and you’re constrained by how much energy is in the data center, why do companies even try to compete with the token king? JH: It’s going to be hard because it’s not reasonable to build a chip and somehow achieve results that are fairly dramatic. Even in the case of Groq , Groq couldn’t deliver the results unless we paired it with Vera Rubin . Well tell me about this, my next question was about Groq. JH: So if you look at the entire envelope of inference, on the one hand, you want to deliver as much throughput as possible, on the other hand, you want to deliver as many smart tokens as possible — the smarter the token, the higher price you could charge. These two balance, this tension of maximizing throughput on the one hand, maximizing intelligence on the other hand, is really, really tough to work out. I do have to say, last year you had a slide talking about this Pareto Curve , and you talked about, I think it was when you introduced Dynamo, how your GPUs could cover the whole thing, and so you didn’t have to think about it, just buy an Nvidia GPU, and Dynamo will do both. But now you’re here saying, “Well, it doesn’t quite cover the whole thing”. JH: We cover the whole thing still better than any system that can do it. Where we could extend that Pareto is particularly on the extremely high token rates and extremely low latency, but it also reduces the throughput. However, because of coding agents, because they’re now AI agents that are producing really, really great economics, and because the agents are being attached to humans that are actually making extremely, I mean, they’re extremely valuable. Right, they’re even more expensive than GPUs. JH: And so I want to give my software engineers the highest token rate service, and so if Anthropic has a tier of Anthropic Claude Code that increases coding rate by a factor of 10, I would pay for it, I would absolutely pay for it. So you’re building this product for yourself? JH: I think most great products are kind of because you see a pain point and you feel the pain point and you know that that’s where the market’s going to go. We would love for our coding agents to run 10 times faster, but in order to do that, it’s just very, very difficult to do that in a high throughput system and so we decided to add the Groq low latency system to it and then we basically co-run, co-process. Right. And is this just separating decode and prefill ? JH: We’re going to do even the high processing, high FLOPS part of decode, attention part of decode. So you’re disaggregating even down to the decode level. JH: That’s right, and that requires really tight coupling and really, really close integration of software. So how are you able to do that? You say you’re shipping later this year, this deal was just announced a couple of months ago. JH: Well, we started working on disaggregated inferencing, Dynamo really put Nvidia’s ideas on the table. The day that I announced Dynamo, everybody should have internalized that, I was already thinking about, “How do we disaggregate inference across a heterogeneous infrastructure more finely?”, and Groq’s architecture is such an extreme version of ours, they had a very hard time. Dynamo was a year ago, and Groq just happened sort of over Christmas. Was there an event that sort of made you think this needed to happen? JH: Well remember, I announced Dynamo a year ago, we’ve been working on Dynamo for two years, so we’ve been thinking about disaggregated inference thing for two, three years, and we started working with Groq maybe before we announced the deal, maybe six months earlier. So we’ve been thinking about working with them about unifying Grace Blackwell and Groq fairly early on. So the interaction with them, I really like the team and we don’t want their cloud service. They had another business that they really believe in and they still believe in, they’re doing really well with it and that wasn’t a part of the business that we wanted, so we decided to acquire the team and license the technology. Then we’ll take the fundamental architecture and we’ll evolve it from here. So it was just a happy coincidence or not a happy coincidence, maybe not a happy coincidence. JH: Strategic serendipity. Because OpenAI, you know, has an instance now with Cerebras that they announced in January . JH: That was done completely independent of us and frankly, I didn’t even know about it, but it wouldn’t have changed anything. I think the Groq architecture is the one I would have chosen anyways, it’s much more sensible to us. Was this the first time where there was sort of an ASIC approach that sort of made you raise your eyebrows like, “Oh, that’s actually fundamentally different”? JH: No, Mellanox . That’s a good example. JH: Yeah, Mellanox. We took a bunch of our computing stack and we put it into the Mellanox stack. NVLink wouldn’t be possible, you know, at the scale we’re talking about without the in-network fabric computing that we did with Mellanox. Taking the software stack, disaggregating it, and putting it where it needs to be, is a specialty of Nvidia. We’re not obsessed about where computing is done, we just want to accelerate the application. Remember, Nvidia is an accelerated computing company, not a GPU company. Right. So you talk about power being the constraint. When your customers are thinking about what to buy, we could buy all sort of traditional GPUs, or we could buy these LPU racks. Is that just, they should be thinking about it in terms of you’re just confident they can drive way more revenue? JH: It really depends on the kind of products they have. Suppose you really don’t have enterprise use cases at the moment, I don’t really think that adding Groq makes much sense, and the reason for that is because most of your customers are free tier customers, and they’re moving towards paying. So it might be two-thirds free tier, one-third paid, in that case, adding Groq to it, you’re adding a lot of expense. You’re taking some power, it’s not worth it. Complexity. And you’re taking away servers, the opportunity cost. JH: What you could actually be serving the free tier, yeah. However, if you have Anthropic-like business and you have OpenAI-like business where Codex is capturing really great economics, but you just wish you could generate more tokens, this is where adding that accelerator can really boost your revenues. Are we actually constrained by power right now in 2026 or by fab capacity or what? Everyone’s saying we don’t have enough supply. What’s the actual limiting factor? JH: I think it’s probably close on everything. You couldn’t double anything, really. Because you’ll hit some other constraints. It does feel like, though, the U.S. has I think done a pretty good job of scrounging up power , maybe more than people expected a couple years ago, it feels like chips are really much more of a limiter right now . JH: Our supply chain is fairly well planned. You know, we were planning for a very, very big year, and we’re planning for a very big year next year. We saw all the soju drinking and fried chickens. JH: (laughing) Yeah, right. We’re planning, we plan for, in our supply chain, we have got, you know, a couple of hundred partners in our supply chain and we’ve got long-term partnerships with them. So I feel pretty good about that part of it. I don’t think we have twice as much power as we need, I don’t think we have twice as much chip supply as we need, I don’t think we have twice of anything as we need. But I think everything is, everything that I see in the horizon, we will be able to support from a supply chain perspective and the thing that I wish probably more than anything is that all the land, power, and shell would just get stood up faster. Is it fair to say, is there a bit where Nvidia is actually the biggest beneficiary of scarcity, though, to the extent it exists? Like, if there’s a power scarcity, you’re the most efficient chip, so you’re going to be utilizing that power better. Or if there’s fab capacity, like you just said, you’ve been out there securing the supply chain, you got it sort of sorted, are you the big winners in that regard? JH: Well, we’re the largest company in this space, and we did a good job planning. And we plan upstream of the supply chain, we plan downstream of the supply chain and so I think we’ve done a really good job preparing everyone for growth. Right, but is this a bit where, at its core, why not having access to the Chinese market maybe is a threat? Like if China ends up with plenty of power and plenty of chips, even though those chips are only 7nm, they have the capacity to build up an ecosystem to potentially rival CUDA in the long run, is that the concern that you have? JH: There’s no question we need to have American tech stack in China, and I’ve been very consistent about that since the very beginning recognizing that open source software will come. No country contributes more to open source software than China does and we also know that 50% of the world’s AI researchers come from China, and we also know that they’re really inventive. DeepSeek is not a nominal piece of technology, it’s really, really good. And Kimi is really good, and Qwen is really good and they make unique contributions to architecture, and they make unique contributions to the AI stack so I think we have to take these companies seriously. To the extent that American tech stack is what the world builds on top of, then when that technology diffuses out of China, which it will, because it’s open source, and when it comes out of China, it goes into American industries, it goes into Southeast Asia, it goes into Europe, the American tech stack will be prepared to receive them. I’ve been really consistent that this is probably the single most geopolitical strategic issue for the American tech industry. Yeah, when we talked last time , the Trump administration had banned the H20. Were you surprised you were able to get the Trump administration to see your point of view? And then were you even more surprised that now you’re stymied by the Chinese government ? JH: I’m not surprised by us being stymied by them and the reason for that is because, of course, China would like to have their tech stack develop. In the time that we’ve left that market, you know how fast the Chinese industry moves, and Huawei achieved a record year for their company’s history. This is a very long-running company, and they had a record year. They had, what, five, six IPOs of chip companies that are addressing the AI industry. I think we need to be more strategic in how we think about American leadership and American geopolitical and technology leadership. AI is not just a model, and that’s a deep misunderstanding — AI, as I said and as you mentioned in the beginning, AI is a five-layer cake and we have to win the infrastructure layer, we have to win the chips layer, we have to win the platform layer, we have to win the model layer and we have to win the application layer. Some of the things that we do are jeopardizing our ability as a country to lead in each one of those five layers. I think it’s a terrible mistake to think that the way to win is to bundle all of it top-to-bottom and tie every company together into one holistic stack so that we can only win or win at the limits of what any one of the layers can win. We’ve got to let all the layers go out and try to win the market. Have those other layers maybe benefited from their longer experience in Washington and you sort of showed up a little late to the scene? JH: Yeah, maybe. What have you learned? What’s been the biggest thing you’ve learned about Washington? JH: Well, the thing that I was surprised by is how deep the doomers were integrated into Washington D.C. and how the messages of doomers affected the psychology of the policy makers. Everyone was scared instead of optimistic. JH: That’s right, and I think it has two fundamental problems. In this Industrial Revolution, if we don’t allow the technology to diffuse across the United States and we don’t take advantage of it ourselves, what will happen to us is what happened to Europe in the last Industrial Revolution — we left them behind. And they, in a lot of ways, they invented all the technologies of the last Industrial Revolution and we just took advantage of it. I hope that we have the historic wisdom, that we have the technological understanding and not get trapped in science fiction, doomerism, these incredible stories that are being invented to scare the living daylights out of policy makers who don’t understand technology very well and they give them these science fiction embodiments that are just not helpful. One of the situations that is most concerning to me is when you poll the United States, the population, the popularity of AI is decreasing, that’s a real problem. It’s no different than the popularity of electricity, the popularity of electric motors, the popularity of gasoline engines, in the last Industrial Revolution became less popular. The popularity of the Internet, could you just imagine? Other countries took advantage of it much more quickly than we did and then technology diffused into its industries and society much more quickly and so we just have to be much, much more concerned that we don’t give this technology some kind of a mystical science fiction embodiment that’s just not helpful and scaring people. And so I don’t like it when doomers are out scaring people, I think there’s a difference between genuinely being concerned and warning people versus is creating rhetoric that scares people. I think a characteristic you see all the time is people put on their big thinking hats and try to tease out all these nuances and forget the fact that actual popular communication is done in broad strokes. You don’t get to say, “Oh, you’re a little scared of this, but not this XYZ”— you’re just communicating fear as opposed to communicating optimism. JH: Yeah, and somehow it makes them sound smarter. People love to sound smart. JH: Sometimes it’s maybe, and we now know, it helps them with their fundraising and sometimes it helps them secure regulatory capture. So there’s a lot of different reasons why they do it, and these are incredibly smart people but I would just warn them that most of these things will likely backlash and will likely come back and they’ll be probably disappointed that they did it someday. I’m gonna tie a few questions together because I know we’re a little short on time. In the self-driving car space , you’re working with multiple automakers, you have your Alpamayo model , while still supplying chips to Tesla. You had a big bit about OpenClaw today in your presentation — meanwhile, a huge thing driving the Vera chips, for example, we talk about agents, is what’s happening with say, Claude Code and happening with Codex from OpenAI. Am I right to tease out a consistent element here, and your investment in your open source models goes with that, where you’re happy to supply the leading provider, or the inventor in a space with chips, but then you’re going to fast follow what they do for everyone else that is threatened by them? So you simultaneously broaden your customer base, you’re not just dependent on the leaders, but then also the leaders are helping you sell to everyone else because they’re worried about being left behind. JH: No, nothing like that. We’re at the frontier on so many different domains. In a lot of ways, we are the leader in many of these domains, but we never turn them into products. We’re a technology stack and so we have to be at the frontier, we have to be the world leader of the technology stack, but we’re not a solutions manufacturer, we’re not a service provider. And so that’s number one. Will that always be the case? JH: Yeah, always be the case. There’s no reason to, and we’re delighted not to. And so we create all this technology, we make it available to everybody. Well, it’s funny though, if you go back to like your boards, for example, like the products you ship, more and more of that, there’s what, 30,000 specific SKUs in a rack today or something like that. More and more of those are defined by you, “This is what it’s going to be”, in part to make it easier to assemble, all those sorts of pieces. Is there a bit where that’s gonna happen on the software side too, as you talk about those vertical bits and your open source model? JH: We create a thing vertically and then we open it horizontally and so everybody could use whatever piece they would like. As long as they’re running on Nvidia chips? JH: Whatever piece they would like, they don’t have to use all Nvidia chips, they don’t have to use all Nvidia software. We have to build it vertically, we have to integrate it vertically and optimize it vertically. But afterwards, we give them source, we give them — they just figure out how they want to do it. Do you think Nvidia can actually produce and keep up in terms of having a frontier model that can win that space or be a necessary provider of that space given that folks like Meta seem to have fallen off or the alternative is, seems to be by and large Chinese models. JH: Winning that space is not important to us. Right, well important not in terms of winning, but important in terms of there needs to be an open source frontier model, so if not you, then who? JH: That’s right, that’s right, somebody has to create open source models and Nvidia has a real capability in doing so. Whenever we create these open source models, we also learn a lot about the computation. Was that a bit of a problem with Blackwell? I’ve heard mutters that the training runs were maybe a little more difficult than they were sort of previously. JH: The challenge with Blackwell was 100% NVLink 72, NVLink 72 was backbreaking work. And it was the only time that I thanked the audience for working with us. Yeah, I noticed when you said that today, it came across as very sincere. JH: Yeah, because we tortured everybody, but everybody loves it now. This is the second time we’ve had a chance to talk in person, and my takeaway when I met you previously in Taipei was the extent that Nvidia still feels like a small company. Are you worried about getting stretched too thin, or do you still think you have sort of that CUDA-esque flywheel where, “It looks like we’re doing a lot, we’re just kind of doing the same thing over and over again?”. JH: The reason why Nvidia can move so fast is because we always have a unifying theory for the company, and that’s my job, I need to come up with a unifying theory for what’s important and why things connect together and how they connect together and then create an organization, an organism that’s really, really good at delivering on that unifying theory. And so the unifying theory for Nvidia is actually fairly simple. On the one hand, we have the computing platform, the software platform that’s related to CUDA-X . On the other hand, we’re a computing systems company, we optimize things vertically, we apply extreme co-design across the stack and all the different components of a computer and now that computer is a platform of ours and we integrate that platform into all the clouds and to all the OEMs and then we have another platform that’s now the data center platform, or the AI factory platform. So once you have a unifying theory about what Nvidia builds and how it goes about doing it — and I used the keynote to kind of tell that story even partly to our own employees. That’s what it felt like. That whole first hour of the keynote felt like you talking to your employees, reminding them of what you do. JH: It’s important that we’re always constantly reminded of what’s important to us and AI is important to us, but of course CUDA-X and all of the solvers and all of the applications that we can accelerate is really important to us. Thank you very much. JH: Thank you. It’s great to see you, Ben. Keep up the good work. This Daily Update Interview is also available as a podcast. To receive it in your podcast player, visit Stratechery . The Daily Update is intended for a single recipient, but occasional forwarding is totally fine! If you would like to order multiple subscriptions for your team with a group discount (minimum 5), please contact me directly. Thanks for being a supporter, and have a great day!

0 views
Lalit Maganti Yesterday

syntaqlite: high-fidelity devtools that SQLite deserves

Most SQL tools treat SQLite as a “flavor” of a generic SQL parser. They approximate the language, which means they break on SQLite-exclusive features like virtual tables , miss syntax like UPSERT , and ignore the 22 compile-time flags that change the syntax SQLite accepts. So I built syntaqlite : an open-source parser, formatter, validator, and LSP built directly on SQLite’s own Lemon-generated grammar. It sees SQL exactly how SQLite sees it, no matter which version of SQLite you’re using or which feature flags you compiled with. It ships as a CLI , VS Code extension , Claude Code LSP plugin , and C / Rust libraries. There’s also a web playground which you can try now: paste any SQLite SQL and see parsing, formatting, and validation live in the browser, no install needed. Full documentation is available here . Here’s syntaqlite in action: Formatting with the CLI Validation with the CLI

0 views
Michael Lynch Yesterday

Refactoring English: Month 15

Hi, I’m Michael. I’m a software developer and founder of small, indie tech businesses. I’m currently working on a book called Refactoring English: Effective Writing for Software Developers . Every month, I publish a retrospective like this one to share how things are going with my book and my professional life overall. At the start of each month, I declare what I’d like to accomplish. Here’s how I did against those goals: Visits and orders are down, but mainly because January was such an outlier due to “The Most Popular Blogs of Hacker News in 2025.” I got another lucky bump from the HN moderators putting “My Eighth Year as a Bootstrapped Founder” on the front page. I mentioned in January that I added regional pricing for my book. I wasn’t tracking data carefully, but just based on order notifications, it seemed like most of my orders were coming from countries outside the US, so I took a closer look at the data. The first question was: is it really true that the majority of orders use regional pricing now? It’s true. The majority of Refactoring English customers are now outside of the US. The US accounts for only 28% of orders by volume and 40% by revenue. I was also surprised to see how many customers purchase from countries like India and Brazil, where English is not the primary language, so I checked English vs. non-English primary countries: Surprisingly, the majority of orders for Refactoring English come from countries where English is not the primary language, though English-speaking countries are a small majority revenue-wise. Next question: Do readers from certain countries purchase at a higher rate than others relative to total website visitors? Wow! One out of every six readers in Kazakhstan purchases the book! I need to start advertising in Kazakhstan. Okay, the extreme Kazakhstan result is based on a single customer, so that’s probably an outlier. And I bet my website analytics undercount visitors from Kazakhstan. What if I focus on the top countries based on website visitors? The US is my top country for website visitors, but a relatively low share (0.5%) purchase the book. Indian readers purchase at the highest rate, with 2.5% of website visitors purchasing the book. Canadian readers purchase the most by revenue, with every Canadian reader giving me about $0.47 in additional book sales. Clearly, I need to start pandering more to India and Canada in the book. I could change all the Docker examples to cricket examples and look for more opportunities to praise Shopify. After the US, most website visitors come from China (5.9% of total), but I’ve had zero sales in China. At first, I thought buying ebooks was not so popular in China, but I just checked what regional discount I was offering in China and was surprised to find it was zero. I wasn’t offering a regional discount in China at all. I made two mistakes in my price generation scripts that excluded a huge number of countries: The local currency thing is silly in retrospect because I can still offer a discount and just accept payment in USD. And I’m not sure how I ended up missing so many Stripe-supported countries. I even missed Kazakhstan, my new favorite country! I was only offering regional discounts in about 39 countries. After my fixes, the list grew to 156. And within 12 hours, I got a new order from Kazakhstan. With the majority of Refactoring English readers coming from countries where English is a second language, should I adjust the book to better serve non-native speakers? A few readers have asked about English tips for non-native speakers. I’d like to tackle the subject, but I have no experience writing as a non-native speaker. I want everything in the book to be techniques I personally use rather than things I’ve heard secondhand . My best idea is to find editing clients who are non-native speakers and look for patterns in their writing to include in the book. But right now, I’d like to get the v1 finished. The beauty of an ebook is that you can keep iterating on it and find ways to improve it even after official release. I’ve been using AI for software development for about a year and a half, but there have been two major inflection points: Since December, I’ve been spending more and more time doing AI-assisted coding. It’s become an ever-increasing part of my workday and non-work time. I used to have a bad habit of checking email and social media excessively. During the past month, I’ve repeatedly had the experience of noticing that it’s 4pm, but I haven’t checked email or social media. Except it’s because I’ve fallen into an AI vortex and forgot everything else. Every month, I think, “Is this a problem?” And in the past few weeks, I’ve had to face the fact that, yes, it’s a problem. I generally start each workday by writing a schedule on a little notepad on my desk. I break the day into 30-minute blocks and write down how I’ll spend that block. Historically, I stick to the schedule when I’m disciplined. When I have less will power, I let fun tasks exceed their budgets by a block or two. With AI-assisted coding, I was getting to the point where I’d make a schedule and then completely ignore it and play with AI all day. I wouldn’t say that I have an “addiction” to AI in the way people develop addictions to drugs or alcohol, but I am letting AI-assisted coding distract me from work that I recognize is more important, like finishing my book. There are a few factors that make AI especially compelling and easy for me to get sucked into: I feel like I can integrate any technology, write in any programming language, install any tool. There used to be an annoying level of friction in using any new software, but now I can mostly just hand it to AI and ask it to figure out how to install it or debug it, and it just works. In the 90s, Bill Gates published a book called Business @ the Speed of Thought . I’ve never read it, but I keep thinking back to that book title as I use AI. It’s not literally at the speed of thought, but it’s closer than anything I ever imagined. I can have an idea for a feature, give a brief explanation to an AI agent, and see the feature materialize in minutes. Even before AI, I’d often intend to spend an hour coding and instead spent three. But there were natural limits to how long I could code. A few hours of intense dev work fries my brain, and work becomes unpleasant, unproductive, or both. With AI, you can build for hours without doing any deep thought. And even when something does require thought, AI makes it easier than ever to take on tech debt. When I’m coding myself, I don’t want to do something the ugly way because then I’m the one who has to maintain that hack. But if I’m making AI do everything, I don’t feel the pain of hacky, ugly code. One of the things that makes gambling addictive is variable rewards . Our brains are more captivated by a system that gives you $10 at random intervals than one that delivers you money on a fixed, predictable schedule. Whether intentional or not, my experience with AI agents varies wildly. Sometimes, I point it at a 2,000 line log file and it diagnoses the issue before I’ve even asked a question. Other times, I give it a simple task, and it spends the next 20 minutes aimlessly roaming my codebase. Because I don’t know if the wait will be 5 seconds or 20 minutes, I sit there staring at the agent for a minute, then compulsively check it every few minutes, then start some other AI task while I’m waiting. And then I’m cycling between multiple agents and don’t even remember what they’re all doing. One of the most maddening experiences I have with AI is when I’ve set up the AI agent to complete a long task, and I come back hours later to find the AI paused its work a few minutes after I left and asked, “Okay, the next step is to try a full build, but that will take 30-60 minutes. Would you like me to continue?” Yes! That’s why I left the task to you! It’s hard to predict exactly what effect AI will have on the software industry, but I feel confident that it will completely upend the ecosystem. We’re in the early stages of a massive shake-up. Depending on how things turn out, there are paths forward for me as a software developer, but I also think there’s at least a 20% chance that we’re in the last year or two of “software developer” being a job that requires any special knowledge or skill. It could be like what happened to elevator operators . Right now, there are a few factors that make AI-assisted development especially attractive for developers in my position: The current situation with AI can’t last. The AI bubble could burst, and I’ll have to start paying the non-subsidized, metered rate. Or AI will continue to improve to the point where I have no advantage over junior engineers or even people with no software experience. I’ve found a few techniques for getting my AI usage back to a manageable place: It turns out that most of Refactoring English ’s readers come from outside the US. I’m using AI-assisted coding too much. Result : Published “Why Improve Your Writing?” and “Improve Your Grammar Incrementally” Result : Scheduled a discussion about design reviews I only included countries where Stripe supports the local currency. Even with this filter, I accidentally omitted a lot of countries where Stripe supports the local currency. In February 2025, I started using an integrated AI agent in my code editor In December 2025, I started running AI agents with full permissions (within isolated environments) AI is helpful for junior engineers, but senior engineers are the ones who can use it best There are multiple AI companies competing heavily on price and using VC money to subsidize costs. I use flat-rate plans, but I consume the equivalent of about $4k/month in API costs, and even those rates are probably VC-subsidized. Don’t start the day with an AI project If I start with AI and then work on my book, then I’m switching from an exciting, easy task to a hard, unsexy task. If I instead start the day with an hour of writing , I’ve done my hard task for the day and don’t have to move uphill. This is challenging because I often set up long AI tasks overnight, and I’m always curious in the morning to see how they turned out. Reduce parallel AI-driven projects. Parallel work feels appealing because I can cycle between agents. In practice, I find it sucks me in too much because there’s a spinning plates mentality of some agent always needing attention. Published two new book chapters Published “Eversource EV Rebate Program Exposed Massachusetts Customer Data” and complained to the MA Department of Public Utilities Don’t start the day with an AI coding project. It’s too distracting and too hard to switch to something harder but more important. Finish Refactoring English It won’t be fully polished and edited, but I want to complete all the chapters.

0 views
Anton Sten Yesterday

Onboarding is a transaction

A design post that's not about AI. I know. Rare. I've been working with two different teams lately, both early-stage, both building something genuinely useful. And both had made the same decision before I arrived: keep onboarding as short as possible. Fewer screens. Fewer questions. Get users to the product fast. I understand the instinct. Churn during onboarding is the thing that keeps founders up at night. Every extra screen feels like a risk. So you cut, and cut, and cut until what's left is a signup flow so frictionless it almost feels rude — like meeting someone and immediately handing them a set of keys. But here's the thing they were both missing: onboarding is one of the few moments where you have a user's complete attention *and* their clear intent. They just decided they want what you're building. They're motivated. They're present. That is not a moment to rush through. ## Efficient doesn't mean fewer questions When people talk about efficient onboarding, they usually mean fewer steps. But that's not what efficient actually means. Efficient means getting as much value as possible — for both sides — while keeping the user willing and engaged. A user who flies through a four-screen signup and lands in a generic empty state is not a success story. You got them in the door, sure. But you know nothing about them, and they're already wondering what to do next. Compare that to a user who spends two more minutes during onboarding, answers a few specific questions, and arrives in an experience that already feels like it was made for them. That's efficient. Not because it was fast, but because it worked. ## The deal people are actually willing to make At Summer Health, we asked parents for their home address during onboarding. On paper, that sounds like exactly the kind of friction you'd want to cut. A home address? For a telehealth service treating your kids? That's not just friction — that's a trust test. But we didn't just ask for it — we explained why. If you give us your address, we can route prescriptions to your closest pharmacy. You give us something, we give something back. We also asked parents, early in the flow, whether they had an urgent question right now. If they said yes, we stopped onboarding entirely and connected them straight to a pediatrician. If they said no, we'd say great — and carry on. That's not a question designed to collect data. It's a signal to the user that we understand why they're here, and that we'll drop everything if they need us to. The onboarding can wait. Then we asked about medical history and allergies. Heavy questions. The kind that make people hesitate. But we were upfront about why: we're asking now so we already know when something urgent is happening. Nobody wants to answer questions about their child's penicillin allergy while they're panicking at midnight. We ask during onboarding so we never have to ask then. People don't mind sharing. The problem isn't the questions — it's when it feels like a company collecting data points rather than actually caring about the answer. The moment it feels like a form versus a conversation, people shut down. Onboarding is a transaction. You're asking for information, time, and trust. In return, you owe them a better experience. When that exchange is clear and honest, users lean in. When it's not, they abandon. ## What the questions you skip are telling you There's a useful test here that I keep coming back to: if you're not sure whether to include a question in onboarding, ask yourself whether you can justify *why* you're asking it — not to yourself, but out loud, to the user. "We're asking for your role so we can show you the features most relevant to how you work." "We're asking about your team size so we don't waste your time on things that don't apply." If you can say it plainly and it sounds reasonable, ask it. If you find yourself reaching for vague justifications, or worse, deciding you don't actually need the answer for anything specific — that's telling you something. Either the question shouldn't be there, or you haven't yet figured out what you'd do with the answer. Both are worth knowing. ## The moment won't come back The startup instinct to minimize onboarding comes from a real fear, and I'm not saying ignore it. Drop-off during signup is real, and a bloated onboarding flow with irrelevant questions is genuinely a problem. But so is the missed opportunity. You will never again have this user's attention the way you have it right now. They signed up. They're curious. They want to be here. Ask them something. Make it worth answering. Tell them why. That's not friction. That's just a conversation.

0 views
Justin Duke Yesterday

Stop Making Sense

A polite man is driven to murder. He becomes a prophet and screams manifestos on love, war, and the increasingly alarming impact of technology and progress. Driven to insanity by his own insights into the human condition, he travels to a river in an attempt to drown himself but instead is baptized and absolved of sin. He dies, crosseyed yet painless. This is the definitive fairytale of my generation, and the moral is "watch out, you might get what you're after". Jesus lives, and he's wearing a giant suit. A film that is so flatly and universally beloved by all who watch it, regardless of affiliation with the band itself. And truth be told, I don't really care much for the Talking Heads — not that I dislike them or their music, but to me they are one of many bands that I can recognize the artistic and aesthetic value in at an intellectual level more than a Dionysian level. (And I don't really prefer my listening to be pleasurable on the intellectual level.) What did I think about while watching this excellent film, a master of its genre? I thought about the greatest concerts in my life: Lost in the Trees playing in the tea house in Charlottesville, an equal number of band members and audience members; Blind Pilot playing in the Crystal Ballroom, an entirely acoustic set and an audience willing enough to go along with it; CHVRCHES at the Paramount in Seattle, sweaty and glowlit. What Demme captures here is that same indelible feel of the best live music, where you feel in the same breath and beat both completely alone and completely surrounded by the only people who matter: building, building, higher, higher. I have half-joked with friends over the past couple years that I'm done with concerts as a medium. The event no longer holds any sort of allure outside of special occasions (once-in-a-lifetimes, family). The highest praise I can give this film is that it made me reconsider that stance.

0 views
Xe Iaso Yesterday

My homelab will be down for at least 20 days

Quick post for y'all now that I can use my macbook while standing (long story, I can't sit due to surgical recovery, it SUCKS). My homelab went offline at about 13:00 UTC today likely because of a power outage. I'm going to just keep it offline and not fight it. I'll get home in early April and restore things then. An incomplete list of the services that are down: Guess it's just gonna be down, hope I didn't lose any data. I'll keep y'all updated as things change if they do. The vanity Go import server The preview site for this blog Various internal services including the one that announces new posts on social media My experimental OpenClaw bot Moss I was using to kill time in bed My DGX Spark for self hosted language models, mainly used with Moss

0 views

Questions about the future of MacOS in the age of the MacBook Neo

As far as I can see, the majority of MacBook Neo reviews are overwhelmingly positive . Other reviews are simply acknowledging that this new laptop will be a huge success, while also recommending other laptops, including the refurbished MacBook Air . These reviews share the same overall message: the Neo, especially after the August-September back-to-school season, will be an immense hit, potentially becoming the best-selling Mac computer of all time, maybe outselling the previous bestseller, I want to say three to four times (just speculating here). With this upcoming increased volume of sales in the traditional computer market, i.e. not phones or tablets, and with these millions of users new to the Mac platform, what can this mean for MacOS and the ecosystem? I have a lot of questions, and very few answers, as you can see below. Will the Neo become a second chance for the Mac App Store? Will the popularity of the Neo, on the contrary, make the Mac App Store experience even worse? Will it become flooded with crappy apps, trying to take advantage of trusting users new to the platform? Will this change the average app price or business model on the Mac? Looking at the Top Free Apps list on the Mac App Store as I write this line, the 6th most popular app is called “ AI Chatbot · Ask AI Anything 5.2 ”. * 1 It sits right after Microsoft Excel and CapCut, and before Microsoft PowerPoint. No, this app — unrelated to OpenAI — is not fishy at all (!) and the Mac App Store is very safe. The 12th most popular app on the list is “ HP: Print and Support ”. Great, great stuff. I wonder what will happen with millions of extra Mac users. Will the Neo help the Mac become a proper gaming platform? The Neo may not be equipped for “serious” gaming, due to its basic screen and “modest” GPU, but all the casual games and older games like Minecraft would be perfectly fine on this machine: there is definitely an opportunity for Apple and developers here, especially with the Mac being compatible with PlayStation, Xbox, and Switch controllers out of the box. Will the popularity of the MacBook Neo be an opportunity for Apple to mobilise more third-party developers to build apps for MacOS, now that the potential user base can be significantly larger? How many of these new apps will be truly native, and how many will be built on top of frameworks like Electron, since the majority of these new users probably won’t care? Is the Neo a new opportunity for the Swift language? Will the Neo push Apple to finally update the Stickies app? I guess we will have to wait until WWDC 2026 to have parts of these answers. Will this increased popularity of the Mac, arguably the first modern Mac for the masses, bring more heat to MacOS when it comes to viruses and security flaws? This is one of the first questions I asked myself when I started to read about how the MacBook Neo could sell millions, on top of the current Mac sales. I understand that MacOS itself is pretty secure, but if MacOS becomes more appealing to apps and games developers, it will also be more appealing to virus makers. How much of the iPad market will the Neo capture? How much of an impact will it have on the Safari vs. Chrome market share: will new Mac users just use Chrome on their new Macs or stick to Safari? Will the Neo push Apple to release more frequent updates for Safari? How many Safari extensions will be available by the end of the year? How many of the new Mac users, brought to the platform via the Neo, will eventually become MacOS enthusiasts? What does it mean for the direction of MacOS? If, by the end of 2026, 80 to 90% of active Macs are MacBooks Neo (again, just speculating), what does it mean for the future of Liquid Glass? * 2 Is an increased line of revenue for the Mac a reason for Apple to mobilise more people to work on MacOS ? I am a little worried that a never-seen-before popularity for the Mac would encourage Apple to make MacOS look and behave more like iOS. Will the increased popularity of the Mac make the Mac less cool in the eyes of others, less exclusive? Is the Mac ready to become more than the cooler alternative to Windows? I have a lot of questions, as you can see. I’m sure most of these questions have been asked hundreds of times already. Answers to these questions will appear obvious to some, less so to others. We don’t even know if the Neo will be as successful as most people predict. But I’m sure the Neo’s success is the one thing that raises the fewest questions. Note: App Store rankings vary by region (I think). My observations relate to the French store. ^ Yeah, sorry in advance, I never know how to write the plural of MacBooks, so in this post I will use the “MacBooks Neo” form. ^ Note: App Store rankings vary by region (I think). My observations relate to the French store. ^ Yeah, sorry in advance, I never know how to write the plural of MacBooks, so in this post I will use the “MacBooks Neo” form. ^

1 views
DHH 2 days ago

ONCE (Again)

The original concept for ONCE sought to sell self-hostable web apps for a one-time fee. That didn't work. Sure, we recouped the investment on Campfire, our chat app, but that was it. You gotta listen when the market tells you what it wants! And it didn't seem to want to pay for self-hosted web apps in a one-off way. So we set Campfire, Writebook, and now Fizzy free by releasing them all as open source with a permissive license. That worked! Tons of people have been running these apps on their own servers, contributing code back, and learning how we build real production applications at 37signals. Now we're doubling down on the gift and adding an integrated way to run all these apps, and your own vibe-coded adventures too, on a brand-new application server we're also calling ONCE. If you twist my arm, I can make that spell "Open Network Container Executor", but we don't even have to go there. Once is just a cool word, we already own the domain, and it's running all the original applications released under that banner as free and open-source installations. That's good enough! The pitch here is that installing a whole suite of applications on your own server should be dead easy. The original ONCE model wanted a dedicated box or VM per app, which was just cumbersome and costly to maintain. Now you can use a single machine — even your laptop! — to run everything all at once. ONCE gives you a beautiful terminal interface to track application metrics, like RAM + CPU usage, as well as basic visitor + request/second counts. It also gives you zero-downtime upgrades and scheduled backups. It's meant to be able to run all the infrastructure apps you'd need, like our full suite and all the ones your AI agents will soon be building for you. Give it a spin. It's just a single command to install. I can show you how with this YouTube video tour. Enjoy!

0 views
David Bushell 2 days ago

What is agentic engineering?

Below is a parody of Simon Willison’s What is agentic engineering? I use the term agentic engineering to describe the practice of casino gambling with the assistance of random superstitions. What are random superstitions ? They’re superstitions that can both write and execute entropy. Popular examples include blowing on dice, wearing lucky socks, and saying a prayer. What’s a superstition ? Clearly defining that term is a challenge that has frustrated gambling researchers since at least the 1990s BC but the definition I’ve come to accept, at least in the field of Random Number Generators (RNGs) like GPT-5 and Gemini and Claude, is this one: The “superstition” is a belief that calls upon God with your prompt and passes it a set of magic definitions, then calls any ritual that the deity requests and feeds the results back into the slot machine. For random superstitions, those rituals include one that can confirm bias. You prompt the random superstition to define a bias. The superstition then generates and executes random numbers in a loop until that bias has been confirmed. Dogmatic faith is the defining capability that makes agentic engineering possible. Without the ability to directly play a hand, anything output by an RNG is of limited value. With automated card shuffling, these superstitions can start iterating towards gambling that demonstrably “works”. Enough of that. If you want to experience agenetic engineering yourself, visit my homepage and play the one-armed code bandit! Thanks for reading! Follow me on Mastodon and Bluesky . Subscribe to my Blog and Notes or Combined feeds.

0 views