Posts in Career (20 found)
Martin Fowler Yesterday

Fragments: April 14

I attended the first Pragmatic Summit early this year, and while there host Gergely Orosz interviewed Kent Beck and myself on stage . The video runs for about half-an-hour. I always enjoy nattering with Kent like this, and Gergely pushed into some worthwhile topics. Given the timing, AI dominated the conversation - we compared it to earlier technology shifts, the experience of agile methods, the role of TDD, the danger of unhealthy performance metrics, and how to thrive in an AI-native industry. ❄                ❄                ❄                ❄                ❄ Perl is a language I used a little, but never loved. However the definitive book on it, by its designer Larry Wall, contains a wonderful gem. The three virtues of a programmer: hubris, impatience - and above all - laziness . Bryan Cantrill also loves this virtue : Of these virtues, I have always found laziness to be the most profound: packed within its tongue-in-cheek self-deprecation is a commentary on not just the need for abstraction, but the aesthetics of it. Laziness drives us to make the system as simple as possible (but no simpler!) — to develop the powerful abstractions that then allow us to do much more, much more easily. Of course, the implicit wink here is that it takes a lot of work to be lazy Understanding how to think about a problem domain by building abstractions (models) is my favorite part of programming. I love it because I think it’s what gives me a deeper understanding of a problem domain, and because once I find a good set of abstractions, I get a buzz from the way they make difficulties melt away, allowing me to achieve much more functionality with less lines of code. Cantrill worries that AI is so good at writing code, we risk losing that virtue, something that’s reinforced by brogrammers bragging about how they produce thirty-seven thousand lines of code a day. The problem is that LLMs inherently lack the virtue of laziness. Work costs nothing to an LLM. LLMs do not feel a need to optimize for their own (or anyone’s) future time, and will happily dump more and more onto a layercake of garbage. Left unchecked, LLMs will make systems larger, not better — appealing to perverse vanity metrics, perhaps, but at the cost of everything that matters. As such, LLMs highlight how essential our human laziness is: our finite time forces us to develop crisp abstractions in part because we don’t want to waste our (human!) time on the consequences of clunky ones. The best engineering is always borne of constraints, and the constraint of our time places limits on the cognitive load of the system that we’re willing to accept. This is what drives us to make the system simpler, despite its essential complexity. This reflection particularly struck me this Sunday evening. I’d spent a bit of time making a modification of how my music playlist generator worked. I needed a new capability, spent some time adding it, got frustrated at how long it was taking, and wondered about maybe throwing a coding agent at it. More thought led to realizing that I was doing it in a more complicated way than it needed to be. I was including a facility that I didn’t need, and by applying yagni , I could make the whole thing much easier, doing the task in just a couple of dozen lines of code. If I had used an LLM for this, it may well have done the task much more quickly, but would it have made a similar over-complication? If so would I just shrug and say LGTM? Would that complication cause me (or the LLM) problems in the future? ❄                ❄                ❄                ❄                ❄ Jessica Kerr (Jessitron) has a simple example of applying the principle of Test-Driven Development to prompting agents . She wants all updates to include updating the documentation. Instructions – We can change AGENTS.md to instruct our coding agent to look for documentation files and update them. Verification – We can add a reviewer agent to check each PR for missed documentation updates. This is two changes, so I can break this work into two parts. Which of these should we do first? Of course my initial comment about TDD answers that question ❄                ❄                ❄                ❄                ❄ Mark Little prodded an old memory of mine as he wondered about to work with AIs that are over-confident of their knowledge and thus prone to make up answers to questions, or to act when they should be more hesitant. He draws inspiration from an old, low-budget, but classic SciFi movie: Dark Star . I saw that movie once in my 20s (ie a long time ago), but I still remember the crisis scene where a crew member has to use philosophical argument to prevent a sentient bomb from detonating . Doolittle: You have no absolute proof that Sergeant Pinback ordered you to detonate. Bomb #20: I recall distinctly the detonation order. My memory is good on matters like these. Doolittle: Of course you remember it, but all you remember is merely a series of sensory impulses which you now realize have no real, definite connection with outside reality. Bomb #20: True. But since this is so, I have no real proof that you’re telling me all this. Doolittle: That’s all beside the point. I mean, the concept is valid no matter where it originates. Bomb #20: Hmmmm…. Doolittle: So, if you detonate… Bomb #20: In nine seconds…. Doolittle: …you could be doing so on the basis of false data. Bomb #20: I have no proof it was false data. Doolittle: You have no proof it was correct data! Bomb #20: I must think on this further. Doolittle has to expand the bomb’s consciousness, teaching it to doubt its sensors. As Little puts it: That’s a useful metaphor for where we are with AI today. Most AI systems are optimised for decisiveness. Given an input, produce an output. Given ambiguity, resolve it probabilistically. Given uncertainty, infer. This works well in bounded domains, but it breaks down in open systems where the cost of a wrong decision is asymmetric or irreversible. In those cases, the correct behaviour is often deferral, or even deliberate inaction. But inaction is not a natural outcome of most AI architectures. It has to be designed in. In my more human interactions, I’ve always valued doubt, and distrust people who operate under undue certainty. Doubt doesn’t necessarily lead to indecisiveness, but it does suggest that we include the risk of inaccurate information or faulty reasoning into decisions with profound consequences. If we want AI systems that can operate safely without constant human oversight, we need to teach them not just how to decide, but when not to. In a world of increasing autonomy, restraint isn’t a limitation, it’s a capability. And in many cases, it may be the most important one we build.

0 views
André Arko 2 days ago

Software developers have become their own joke

Creating software is complicated. It’s hard to figure out exactly what you need to build without a lot of trial and error. It almost always requires both exploring possible options and refining something until it works really well. But those things aren’t the same! Your research prototype is not a good product that people will happily pay for. Back in the olden days, when software literally came from BigCo R&D departments, we managed to invent Unix, and the mouse, and GUIs, and Ethernet, and TCP/IP, and a ton of other stuff we all use constantly today. Those research divisions didn’t ship viable consumer products, though. Doug Englebart demoed a mouse-driven GUI in 1968, but you couldn’t buy a home computer with a mouse-driven GUI until 1979, and they didn’t become commercially popular until the Macintosh in 1984. Even years or decades of research wasn’t enough, and years (or decades!) of development work also needed to be done before the results was ready for people to use. Early literature about creating software, written by Fred Brooks and his peers, seems to contain the internalized view that both R and D are required. That’s not surprising, since R&D departments created most software back then, but we seem to have lost track of that connection. Even though our jobs are descended from those R&D labs of yore, we somehow lost the industry job of “software researcher”, and only “software developer” remains. Instead, research happens in academia, where an argument and some pseudocode is all you need to publish a paper. In that world, development is effectively non-existent. (I admit the division isn’t perfectly clear-cut. Sometimes academics will start companies around their research that create a product, or more likely get acquired to add a feature to a product. And sometimes Linus Torvalds will just build a new operating system, without doing any academic research on it, and it will get so popular everyone uses it. The point is that industry and academia have each publicly claimed one half of R&D while disowning the other.) The broader separation of research and development into academia and industry is really unfortunate, because good software needs both research and development as inputs. If you don’t do any research, you can’t identify which parts will be hard (or impossible) until after it’s too late. You also won’t have a good idea of what parts are important until after you’ve put in most or all of the work to create the parts that don’t matter. If you don’t do development, you won’t ever have something robust enough that other people can use it successfully. Meanwhile, on the other side, it feels like developers work hard to convince themselves there are no research aspects involved in their jobs. We call anything research-ish by another name, like “design”, “user experience”, “prototyping”, “de-risking”, “a spike”, and a lot of other funny euphemisms that avoid referring to the work as research. It seems like we’re trying to convince ourselves that we don’t do Research any more, because we are just Developers. This cultural lack of clarity around research in software development spaces really hit hard for me this week, as I read yet another treatise on working with LLM-driven agents for development. The two most popular takes that I have seen are “these tools are a fundamental shift in the nature of software development” and “these tools change nothing about building software at all”. Then the two sides start screaming at each other about how the other side is delusional and time will prove them completely wrong, and I lose interest. If we instead start from the premise that all software work requires research (where the problem space must be explored) and development (where solutions must be implemented and refined), there’s something hiding in the sometimes messy overlap between those two ideas that I’m not seeing come up in any discussions. No one can take the output of software research and treat it like it’s the output of software development. Not Bell Labs, not Xerox PARC, not Microsoft middle managers, and not “solo founders managing a team of AI agents” today. Unfortunately, seeing a prototype and becoming convinced it’s complete is not a new problem. It’s been the bane of software development possibly since the very beginning, when (apocryphally) a manager would review a mockup and conclude the project was now complete and could be shipped to customers immediately. Today, instead of telling that story as a joke, software developers have have somehow turned themselves into the boss from the joke, shouting that it’s time to ship the research prototype because it “looks finished”. How did we do this to ourselves? It seems like, back when we always had to do all the work ourselves, it was harder for software developers to be confused this way. If a developer knows they skipped every validation and edge case, it’s much easier to realize it’s not finished. If an LLM agent says “here’s a comprehensive implementation”, without mentioning all the validations and edge cases it skipped, many (and possibly most) developers will not notice the parts that are missing. This phenomenon is bad for a lot of reasons, including one reason you have probably already thought of: we’re going to get a lot more software claiming to be “comprehensive” and “fully implemented” when it’s really a partially finished prototype that’s full of holes. In a world full of research prototypes being pitched as completed development work, life is about to get worse for everyone who uses software. The docs are even more wildly wrong than they were before, customer support is telling you that your problem is solved by a feature that doesn’t exist, and company leadership is so excited they are planning to fire as many humans as possible so they can have more of it. I don’t want worse software! The software we already have is mostly terrible. Not only much worse software, but also much more of it, is pretty much my worst case scenario. What I actually want is better software, even if that means less of it. Unfortunately, instead of making better software, software developers have decided to become the butt of their own joke, shipping software that doesn’t work, with a footnote that says they know it doesn’t work but they are still shipping it. I don’t see any way to stop it, but I hate it anyway.

0 views
iDiallo 5 days ago

Your friends are hiding their best ideas from you

Back in college, the final project in our JavaScript class was to build a website. We were a group of four, and we built the best website in class. It was for a restaurant called the Coral Reef. We found pictures online, created a menu, and settled on a solid theme. I was taking a digital art class in parallel, so I used my Photoshop skills to place our logo inside pictures of our fake restaurant. All of a sudden, something clicked. We were admiring our website on a CRT monitor when my classmate pulled me aside. She had an idea. A business idea. An idea so great that she couldn't share it with the rest of the team. She whispered, covering her mouth with one hand so a lip reader couldn't steal this fantastic idea: "what if we build websites for people?" This was the 2000s, of course it was a fantastic idea. The perfect time to spin up an online business after a market crash. But what she didn't know was that, while I was in class in the mornings, my afternoons were spent scouring Craigslist and building crappy websites for a hundred to two hundred dollars a piece. I wasn't going to share my measly spoils. If anything, this was the perfect time to build that kind of service. That's a great idea , I said. There is something satisfying about having an idea validated. A sort of satisfaction we get from the acknowledgment. We are smart, and our ideas are good. Whenever someone learned that I was a developer, they felt this urge to share their "someday" idea. It's an app, a website, or some technology I couldn't even make sense of. I used to try to dissect these ideas, get to the nitty-gritty details, scrutinize them. But that always ended in hostility. "Yeah, you don't get it. You probably don't have enough experience" was a common response when I didn't give a resounding yes. I don't get those questions anymore, at least not framed in the same way. I have worked for decades in the field, and I even have a few failed start-ups under my belt. I'm ready to hear your ideas. But that job has been taken, not by another eager developer with even more experience, or maybe a successful start-up on their résumé. No, not a person. AI took this job. Somewhere behind a chatbot interface, an AI is telling one of your friends that their idea is brilliant. Another AI is telling them to write out the full details in a prompt and it will build the app in a single stroke. That friend probably shared a localhost:3000 link with you, or a Lovable app, last year. That same friend was satisfied with the demo they saw then and has most likely moved on. In the days when I stood as a judge, validating an idea was rarely what sparked a business. The satisfaction was in the telling. And today, a prompt is rarely a spark either. In fact, the prompt is not enough. My friends share a link to their ChatGPT conversation as proof that their idea is brilliant. I can't deny it, the robot has already spoken. I'm not the authority on good or bad ideas. I've called ideas stupid that went on to make millions of dollars. (A ChatGPT wrapper for SMS, for instance.) A decade ago, I was in Y Combinator's Startup School. In my batch, there were two co-founders: one was the developer, and the other was the idea guy. In every meeting, the idea guy would come up with a brand new idea that had nothing to do with their start-up. The instructor tried to steer him toward being the salesman, but he wouldn't budge. "My talent is in coming up with ideas," he said. We love having great ideas. We're just not interested in starting a business, because that's what it actually takes. A friend will joke, "here's an idea" then proceeds to tell me their idea. "If you ever build it, send me my share." They are not expecting me to build it. They are happy to have shared a great idea. As for my classmate, she never spoke of the business again. But over the years, she must have sent me at least a dozen clients. It was a great idea after all.

0 views
Martin Fowler 6 days ago

Fragments: April 9

I mostly link to written material here, but I’ve recently listened to two excellent podcasts that I can recommend. Anyone who regularly reads these fragments knows that I’m a big fan of Simon Willison, his (also very fragmentary) posts have earned a regular spot in my RSS reader. But the problem with fragments, however valuable, is that they don’t provide a cohesive overview of the situation. So his podcast with Lenny Rachitsky is a welcome survey of that state of world as seen through a discerning pair of eyeballs. He paints a good picture of how programming has changed for him since the “November inflection point”, important patterns for this work, and his concern about the security bomb nestled inside the beast. My other great listening was on a regular podcast that I listen to, as Gergely Orosz interviewed Thuan Pham - the former CTO of Uber. As with so many of Gergely’s podcasts, they focused on Thuan Pham’s fascinating career direction, giving listeners an opportunity to learn from a successful professional. There’s also an informative insight into Uber’s use of microservices (they had 5000 of them), and the way high-growth software necessarily gets rewritten a lot (a phenomenon I dubbed Sacrificial Architecture ) ❄                ❄                ❄                ❄                ❄ Axios published their post-mortem on their recent supply chain compromise . It’s quite a story, the attackers spent a couple of weeks developing contact with the lead maintainer, leading to a video call where the meeting software indicated something on the maintainer’s system was out of date. That led to the maintainer installing the update, which in fact was a Remote Access Trojan (RAT). they tailored this process specifically to me by doing the following: Simon Willison has a summary and further links . ❄                ❄                ❄                ❄                ❄ I recently bumped into Diátaxis , a framework for organizing technical documentation. I only looked at it briefly, but there’s much to like. In particular I appreciated how it classified four forms of documentation: The distinction between tutorials and how-to guides is interesting A tutorial serves the needs of the user who is at study. Its obligation is to provide a successful learning experience. A how-to guide serves the needs of the user who is at work. Its obligation is to help the user accomplish a task. I also appreciated its point of pulling explanations out into separate areas. The idea is that other forms should contain only minimal explanations, linking to the explanation material for more depth. That way we keep the flow on the goal and allow the user to seek deeper explanations in their own way. The study/work distinction between explanation and reference mirrors that same distinction between tutorials and how-to guides. ❄                ❄                ❄                ❄                ❄ For eight years, Lalit Maganti wanted a set of tools for working with SQLite. But it would be hard and tedious work, “getting into the weeds of SQLite source code, a fiendishly difficult codebase to understand”. So he didn’t try it. But after the November inflection point , he decided to tackle this need. His account of this exercise is an excellent description of the benefits and perils of developing with AI agents. Through most of January, I iterated, acting as semi-technical manager and delegating almost all the design and all the implementation to Claude. Functionally, I ended up in a reasonable place: a parser in C extracted from SQLite sources using a bunch of Python scripts, a formatter built on top, support for both the SQLite language and the PerfettoSQL extensions, all exposed in a web playground. But when I reviewed the codebase in detail in late January, the downside was obvious: the codebase was complete spaghetti. I didn’t understand large parts of the Python source extraction pipeline, functions were scattered in random files without a clear shape, and a few files had grown to several thousand lines. It was extremely fragile; it solved the immediate problem but it was never going to cope with my larger vision, never mind integrating it into the Perfetto tools. The saving grace was that it had proved the approach was viable and generated more than 500 tests, many of which I felt I could reuse. He threw it all away and worked more closely with the AI on the second attempt, with lots of thinking about the design, reviewing all the code, and refactoring with every step In the rewrite, refactoring became the core of my workflow. After every large batch of generated code, I’d step back and ask “is this ugly?” Sometimes AI could clean it up. Other times there was a large-scale abstraction that AI couldn’t see but I could; I’d give it the direction and let it execute. If you have taste, the cost of a wrong approach drops dramatically because you can restructure quickly. He ended up with a working system, and the AI proved its value in allowing him to tackle something that he’d been leaving on the todo pile for years. But even with the rewrite, the AI had its potholes. His conclusion of the relative value of AI in different scenarios: When I was working on something I already understood deeply, AI was excellent…. When I was working on something I could describe but didn’t yet know, AI was good but required more care…. When I was working on something where I didn’t even know what I wanted, AI was somewhere between unhelpful and harmful… At the heart of this is that AI works at its best when there is an objectively checkable answer. If we want an implementation that can pass some tests, then AI does a good job. But when it came to the public API: I spent several days in early March doing nothing but API refactoring, manually fixing things any experienced engineer would have instinctively avoided but AI made a total mess of. There’s no test or objective metric for “is this API pleasant to use” and “will this API help users solve the problems they have” and that’s exactly why the coding agents did so badly at it. ❄                ❄                ❄                ❄                ❄ I became familiar with Ryan Avent’s writing when he wrote the Free Exchange column for The Economist. His recent post talks about how James Talarico and Zohran Mamdani have made their religion an important part of their electoral appeal, and their faith is centered on caring for others. He explains that a focus on care leads to an important perspective on economic growth. The first thing to understand is that we should not want growth for its own sake. What is good about growth is that it expands our collective capacities: we come to know more and we are able to do more. This, in turn, allows us to alleviate suffering, to discover more things about the universe, and to spend more time being complete people. they reached out masquerading as the founder of a company they had cloned the companys founders likeness as well as the company itself. they then invited me to a real slack workspace. this workspace was branded to the companies ci and named in a plausible manner. the slack was thought out very well, they had channels where they were sharing linked-in posts, the linked in posts i presume just went to the real companys account but it was super convincing etc. they even had what i presume were fake profiles of the team of the company but also number of other oss maintainers. they scheduled a meeting with me to connect. the meeting was on ms teams. the meeting had what seemed to be a group of people that were involved. the meeting said something on my system was out of date. i installed the missing item as i presumed it was something to do with teams, and this was the RAT. everything was extremely well co-ordinated looked legit and was done in a professional manner. Tutorials: to learn how to use the product How-to guides: for users to follow to achieve particular goals with the product Reference: to describe what the product does Explanations: background and context to educate the user on the product’s rationale

0 views
DYNOMIGHT 1 weeks ago

I quit drinking for a year

In early January 2025, a family friend was over for lunch. One of my many guilty midwit pleasures is a love of New Year’s resolutions, so I asked her if she had made any. She said no, but mentioned that she had some relatives that were doing “damp January”. In case you’re not aware, Dry January is a challenge many people do to quit drinking alcohol during the month of January. These folks were doing a variant in which, instead of not drinking, one simply drinks less. For some reason, this triggered me. I thought, “Are you kidding? You can’t even stop drinking for a single month? Do you know how pathetic that is?” And then, “Fuck you! Fuck you for doing damp January! You know what, I’m going to stop drinking for a year !” To be clear, these thoughts were directed at people I’ve never even met. In retrospect, I wonder what was going on with me emotionally. But I take resolutions seriously, so I felt committed. We are now 15 months down the timeline, so I’ll make my report. This will sound odd, but I swear it’s true. Not drinking was so easy that it was almost easier than my previous baseline of not-not-drinking. Before starting this resolution, I didn’t drink much—perhaps two or three drinks per week. But I often thought about drinking. Every time I saw friends or went to a restaurant, I thought, “Should I have a drink?” Usually I decided not to. But making that decision required effort. After a few weeks of not drinking, that question never even came up. Drinking was simply not a thing I did, so I never needed to negotiate with myself. Theoretically, you could allow yourself one drink a month instead of zero. Theoretically, that should be easier. But I’m pretty sure I’d find it harder, because alcohol would still be an option , a thing to consider. Early on, I sometimes wanted a drink. But gradually I noticed that I didn’t really want a drink, I just wanted a thing . I can’t find a precise name for this concept in psychology, but often, some deep part of my brain seems to scream, “I WANT A THING.” It could be alcohol, but I found dessert worked just as well. I suspect that a new shirt or meeting a new dog would also work. I was not able to stop my brain from doing this. When it demanded a thing, I gave it a thing. I just substituted a non-alcohol thing. So, over the year, I became interested in desserts and even-more interested in tea. The struggle was The Chocolates. Shortly after I made this resolution, my mother gave me a bag of chocolates that each contained a bit of whiskey. In general, I don’t keep chocolate at home. If anyone gives me chocolate, I immediately eat all of it and then text the giver, “Thanks for the chocolate, I ate it instead of dinner, it’s all gone, this is what will always happen if you give me chocolate.” But I couldn’t eat the Chocolates, because they contained alcohol. I managed to get guests to eat a few. A couple of times I came close to draining out the alcohol and eating the chocolate container. I even considered throwing them away, but that felt wrong. So instead I spent a year glaring at them and waiting for them to apologize for the anguish they were causing me. This represented half the difficulty of this resolution. I do not recommend it. Keep your things separate. Have you heard that alcohol is bad for sleep? Because alcohol is bad for sleep . I’ve always known that was true, abstractly. But sleep is variable. If I didn’t sleep well on an individual night, I was never sure: Was that because of the alcohol, or was it random variation? After a year without alcohol, I am very confident that yes indeed, alcohol is bad for sleep , because my sleep during 2025 was much better than in previous years. Sure, like anyone else, I still sometimes wake up and start thinking about oblivion rushing towards me, and how everything I love will vanish into time, and how all that was once future and hope inevitably becomes static and dust, and how the plague of bluetooth speakers continues to spread across the globe. But now: less! I wish there was a drug I could take that would give me energy and improve my mood and make me physically healthier and smarter, all without side-effects. I don’t think such a drug exists. But we do have the opposite! So, sadly, I’ve come to believe that alcohol is basically the perfect anti-nootropic. That’s not because it makes you dumb while you’re drunk. (True, but who cares?) Rather, that’s because it is bad for sleep , and therefore makes you worse across all dimensions the next day. I did find not drinking to have one clear downside: It’s just not that much fun to hang out with people who are drinking if you are not drinking yourself. To be clear, this is a limited effect. It’s only an issue at bars or certain parties where people are there to drink . I don’t go to many such gatherings, but when I did, I felt it was less fun. It’s not that I missed alcohol. Instead, my theory is that drinking parties are a sort of joint role-playing exercise: “Let’s all get together and collectively reduce our inhibitions and see what happens.” It’s fun not (just) because everyone is taking a recreational drug, but because it’s a joint social experience. If you don’t drink, then you aren’t fully participating. It seems like it should be possible to reproduce this effect without alcohol. You could imagine other ways to push the social equilibrium out of balance. Like… Masks? Or weird environments? Or mutual disclosure games? Should people get together and do a group cold plunge? Unfortunately, all these are complicated and/or carry some kind of social stigma. So until we figure something better out, this is a real cost of not drinking. It was minor for me, but it probably depends a lot on where you are in life. All other effects were minor. I guess I saved money at restaurants. I actually lost a bit of weight over the year, despite all the extra desserts, though I can’t say for sure if alcohol was the cause. Otherwise, once I stopped thinking of alcohol as an option, I rarely thought about the resolution at all, except when I saw those damn chocolates. Towards the end of the year, I started wondering if I should quit drinking forever. But I never came to a conclusion, because I rarely thought about alcohol. I considered having a drink at midnight on New Year’s eve, but I happened to be on a plane that crossed the international date line and thus skipped New Year’s eve. And then… for the first few months of 2026, I still didn’t drink. That wasn’t because of any decision. It just never seemed appealing because (a) sleep and (b) I’d broken the mental link between want thing and drink alcohol . Eventually, I ate the chocolates, and I had a glass of wine when visiting some friends. If I can continue rarely drinking while almost never thinking about drinking, I’ll probably do that. If I slowly slide back into always thinking of alcohol as a live option and always negotiating with myself, I might just resolve to quit forever. So that’s my story. Obviously, it’s heavily colored by my own idiosyncrasies, so it’s hard to say if it offers any general lesson. I do think people underrate the long-term health impact of drinking. The effect on heart disease is debated, but everyone agrees that any alcohol increases the risk of cancer. Still, the long-term effects from occasional light drinking probably aren’t huge. What’s really underrated is the short-term effects, via worse sleep. If I had to give advice, it would be this: If you drink, and you think you might be better off not drinking, why not try it? Maybe you’ll find that champagne is essential to your happiness and drink it every night, to hell with the costs. Maybe you’ll find a different baseline, or maybe you’ll quit forever. Whatever you decide, you’ll have full information.

1 views
Takuya Matsuyama 1 weeks ago

What a Japanese cooking principle taught me about overcoming AI fatigue

Hey, what's up? It's Takuya . I've been thinking about how to live well in this AI era as a developer, content creator, and artist. By “living well,” I mean enjoying the act of creating while maintaining good mental health. I imagine many of my readers are also wondering how to survive — and even thrive — amid the rapid changes brought by recent AI advancements. I don’t have all the answers. No one can predict the future precisely. But I believe it’s more a matter of direction than strategy — rather than trying to build some kind of moat around your life as a software-oriented artist. Where do you want to go? What do you want to see? That's what matters. Strategies/moats can be flexibly changed along with the situations. In this article, I’d like to explore a few life principles from Japanese culture. I recently read a book called "一汁一菜でよいという提案" (The Proposal for One Soup, One Dish) by traditional food expert Yoshiharu Doi (土井善晴), and found it very interesting for keeping the pace of your life healthy. 一汁一菜 pic.twitter.com/hJVYhJX4lE We are currently in the midst of "AI fatigue." New services emerge every day, and big company releases change workflows weekly. Chasing every hype doesn't make us more secure; it just fogs up the crucial skills we need to foster. It’s easy to fall into the trap of trying too hard to keep up, only to end up exhausted. Traditional Japanese culture offers a different perspective — one that helps us stay grounded and resilient in the face of uncertainty. Let's dive into it. Video version: Doi-san's book emphasizes that by stripping away the unnecessary, we find our "comfortable place." As he puts it: It's easy to get your wishlist or to-do list flooded if you don't have any clear rules, policy, or direction in your life. Let's think about it. For example, on social media, the algorithms try to grab and hold your attention as long as possible by displaying so much gossip and drama depending on your interests. But you have full control of not willingly seeing them. I'm always careful not to spend time tracking gossip or drama on the internet. It lets me keep calm and avoid comparing myself with others, causing me to feel miserable. To form a rhythm of life, you have to decide what NOT to do in your daily life. In terms of cooking, Doi-san proposes a system called " One-soup One-dish (一汁一菜)", which helped a lot of people who cook for their family every day. This is a style of meal centered around rice, with one soup and one side dish. Eating is an essential part of living. Yet, keeping it simple is surprisingly hard. There are endless food products, recipes, and health tips — and TV shows constantly push us to make beautiful, varied meals for our families every day. Oh, it looks very similar to today's tech industry, doesn't it? People are unconsciously exhausted by all of this, feeling as if they have to create something great every single day. Doi-san saw through this — and freed them by saying: Software developers can't live without software. It is literally an essential part of our lives, and keeping it simple is surprisingly hard. Let's learn from his philosophy behind his principle. Why does cooking matter so much? Because it is something you do every day, which makes you or someone you love feel really comfortable, as he puts it: I was deeply moved by this. I cook for my family every evening. It helps me shift from work to home—a transition I don’t get naturally, since I don’t have a commute. My 4-year-old daughter sometimes says, “I can smell something good.” It makes me happy, too. What’s important is having something you do every day that makes you feel safe, comfortable, and happy. It could be anything, such as playing an instrument, going for a walk in the morning, painting, singing, swimming, reading before bed, or meditating. It should be something you never get bored with. Something you’ve truly enjoyed in the past. Something that doesn’t make you compare yourself to others, but instead helps you be mindful. If you don't have it yet, step away from your computer and go outside to experience new things. I feel like it’s important that this habit doesn’t involve a screen, as he suggests: What matters is finding something you can return to every day — something you never tire of, like rice and miso soup, rather than something instantly stimulating but quickly exhausting, like Netflix or doom-scrolling on social media. The more I adapt to algorithms and AI, the more I value organic connections – both with people and with ideas. In his book, Doi-san explains that the things we never tire of are often the things humans didn't "engineer" to be perfect: This concept of "not a human feat" (or rather, not a calculated feat) is exactly what’s missing from our digital lives. Algorithms are the "processed seasoning" of the internet — designed to give you an instant hit of dopamine, but leaving you feeling empty and "tired of the flavor" an hour later. I’ve realized that my most resilient moments don't come from a perfectly optimized prompt or a viral post. They come from the "fermentation" of daily life — the slow, messy, unscripted interactions that haven't been optimized for engagement. For example, when I have a quick, casual chat with the barista at Starbucks, or when I’m swapping stories with other parents ( mama-tomo ) while dropping my daughter off at kindergarten, I feel like I’m participating in a natural rhythm, not an algorithm. Small moments like these give me a real sense that I belong to society and am truly living in it. My best ideas work the same way. They rarely strike when I’m glaring at a screen, trying to force a breakthrough. Instead, they "descend" upon me when I’m out for a walk or simply staring blankly at the scenery. It feels less like distillation — which focuses on seeking speed, purity, and efficiency — and more like fermentation . It’s about creating the right environment and then letting the subconscious work its magic over time. You can't rush miso, and you can't rush a truly original thought. Each season brings a variety of foods to enjoy, and Doi-san emphasizes the importance of appreciating them: In Japan, we celebrate the arrival of the first bonito or the last of the winter cabbage. It’s enjoyable to appreciate these changes, but notice one thing: people don’t "chase" them. You don't feel like a failure if you missed the peak of cherry blossom season; you simply look forward to the next cycle. Yet, in the tech industry, we treat trends like a race we are constantly losing. Instead of trying so hard to "stay in the loop," why not view new technologies as seasonal arrivals? You don't have to master every single one. You are living in "Technical Nature" just as much as you are living in real nature. If a new AI model drops, it’s like the first bamboo shoots of spring — interesting, worth trying, but not something to stress over. You can learn anything when it becomes necessary for your craft. By shifting from "chasing" to "appreciating," you replace FOMO with curiosity. When we stop being obsessed with "catching up" and start allowing ourselves the "leisure time" mentioned earlier, something vital happens: playfulness is born. True creativity doesn't come from a place of survival or anxiety. It comes from playing with the tools available to us, much like a chef plays with the ingredients of the season. For a developer, this might be the optimal way of life. Don't just be a user of technology; be someone who resonates with its constant birth and decay, using that rhythm to create something that feels truly alive. I experienced burnout last year . The philosophy of one soup, one dish has helped me step back from that. It reminds me to strip away the noise and return to what really matters. Relax! You’re not as bad as you think. Trust your instincts. Listen to your body. And let’s enjoy this rare moment of change we’re living through :)

1 views
Brain Baking 1 weeks ago

A Commentary On GenAI Inspected Through Different Lenses

The amount of concerning reports related to generative AI is rising at an alrming rate, yet all we do is make ourselves more dependent on the brand new technology. Why? It’s not just that we’re lazy—we are!—there are many more variables involved. As part of my quest to try and understand what the heck is going on and what is becoming of one of my prime professional fields: software engineering, I read and read and read. And then I read and read and read. And then I became disappointed and depressed. I see colleagues jumping the gun, others being more prudent. I see industry discovering there’s yet another buck to be made. I see students forgoing learning at all. I wanted to try to form my own judgement of genAI in its modern form by looking at it from four different viewpoints: that of the software engineer, that of the teacher, that of the creativity researcher, and that of the concerned civilian living in this capitalist world.. References can be found at the end of this article. Does anyone remember Dan North’s Programming is not a craft post from 2011? I do, and I often think about it. With the advent of genAI, North’s port might be even more polarising: Well, congrats to you, you’ve won the lottery: here’s a tool that immediately can add customer value. If you don’t care about the inner code quality, you can have genAI generate (slop) code faster than you can think. If you love the impact of software itself, you’ll love Claude Code et al. Are you perhaps an enterprise software engineer? In that case you’ll be able to scaffold and generate CRUD crap even faster, hooray! But wait a minute. You obviously won’t take true ownership of this code: you’ll want to impress your clients with the results, but keep the lid closed at all times. The less ownership and feeling of responsibility, the easier it comes to completely let go of all the breaks and just accept any future changes without code reviewing at all. People who are now claiming they will keep themselves in the loop as an architectural reviewer don’t need to lie to themselves. After the nth time pressing the green button, and as the technology further evolves, you’ll wind up eventually accepting the slop anyway. Verification burnout will pop up next: because it’s not your own code you’re attempting to so carefully review, it actually takes more instead of less effort, increasing your stress level instead of reducing it! Does the code quality really matter if all clients see is the end product? As a gamer, I just want the game to run smoothly, I don’t care about the spaghetti. Or do I? I do, implicitly—the more spaghetti, the less smoothly it’ll run. The more holes, the more soft locks and crashes. So programming might or might not be a craft, but as Cal Newport and Robert M. Pirsig say: the concept of Quality is important! Maybe it’s time to become a goose farmer instead. The only thing left for you to do is to move to a depressing quality control position instead of crafting something yourself. No more “I built this”, but “I managed its orchestration”. Depending on how you view this, It’s either a promotion or demotion. I tend to agree with the latter. Why? Because we humans are the Homo Faber , the ones who like to control their fate and environment with the use of tools. Yes, genAI certainly is a tool, but it’s a tool that takes away all other tools. Instead of kneading dough by hand, feeling it, knowing when to ferment and when to bake, we’re forced to oversee the industrial Wonder Bread production process. Instead of manipulating leather to create a pair of shoes, we’re being employed by Nike to watch shoes being made by machines. This somehow reminds me of David Graeber’s bullshit jobs where useless paper pushing is prevalent but also called a “revolution” when it comes to a professional purpose. I beg to differ. Humans want to make things. They want to be proud of the things they made. The fact that the open source community rejects this slop code is a telling sign: if you’re programming in the open, your peers who also think highly of software development will keep you in check. But when it’s “for (enterprise) work”, we don’t care, generate away, I’m not the true owner anyway. If programming is a craft, then the recently leaked Claude Code CLI source code will be a big joke to you, where constructs are endlessly repeated, and spaghetti is topped up with more spaghetti. Code that is being generated doesn’t even seem to be made to be (re)read: how then, are we expecting to maintain it, or guarantee its security? By letting the agent maintain it and guarantee its security, I can hear you say? What is there left to say? I’ve already asserted that genAI tools are worse than Stack Overflow . Sure, mindless copy-pasting has long existed before this AI storm, but not on this scale. GenAI is able to provide a working solution to an assignment faster than I can come op with the assignment itself. Suddenly, all our traditional evaluation systems and grading workflows became useless: scoring high on a checklist is just a matter of pasting the requirements into Claude. We try to adapt by requiring oral defences, having students explain what they did and why, and asking them to walk us through a small imaginative change. The result is a spectacular fall in grades from previous years: they are just not able (1) to explain the code they did not make but generated and (2) to make small adjustments as they skipped the hard part: the learning and understanding. Yet in the hallways, I hear lots of students bragging to each other about how they let ChatGPT do their homework. Congrats. We’ll see each other again in September for your second try. We often forget something else very important: peer pressure . About a year ago, on the train I overheard a few girls on their way to a university lecture chatting about their homework. One of them complained: “I put in all that hard work, but all the others are just using ChatGPT to do it. Next time, I’m not doing all that, I’m also just using AI, that’s not fair!”. I should have gotten up to congratulate her: the only one actively learning is the one putting in the hard work! There is no shortcut to becoming proficient. There is only hard work. Sure, the more you prompt your way through your curriculum, the more proficient you’ll become with the tool, but ask yourself: did you learn what you wanted to learn or did you learn to prompt? When I was an undergraduate, I used to fill A4 pages with summaries of courses to help me study. Just before the exams, I could quickly glance over these pages to remembers the core concepts. Some students sold their summaries to others. Now, genAI can generate summaries for you. But smart students will know this will only fool yourself: the purpose of the summaries is to make them : to study and gradually fill the pages. Not to acquire a summary. The journey is the destination. When my summaries were done, I could just as well throw them away: they were just a tool to help with the hard work. Yet it’s next to impossible to explain this to a student who only sees how easy it is to jump to an outcome by leveraging AI. Maybe legislation will help here? (Not really; see below) In case all this is not clear: students are becoming dumber yet the programming projects they hand in are becoming better than ever. As the inventor of the framework presented in The Creative Programmer , I thought it would be interesting to take a look at the seven domains and how genAI fits in these. In The Creative Programmer , I present seven distinct but heavily intertwined themes that define the way we are creative when we solve a programming problem: I might be overly focusing on the negative here and have to recognise the possible advantages of having genAI as a tool available in our creative toolbox—but only when we learn to yield it properly and with moderation, which is not exactly what we are doing lately, is it. In an interesting systematic literature review (2025) with lots of references to other academic material if that’s what you’re looking for, Holzner et al. conclude with: […] human-GenAI collaboration shows small but consistent gains in creative output across tasks and contexts. However, collaboration with GenAI reduces the diversity of ideas, indicating a risk of creative outputs that could become more homogeneous. More same-ness; exactly what we need when it comes to creativity, right? The more we use genAI, the more creatively we will be able to prompt, but the less creative we will be in actually applying a solution to the problem. We no longer create: we generate. We know that genAI will do everything in its power to keep you locked within that chat box. Its tendency to talk to your mouth, agree with your statements, and serve you whatever you want to hear creates biases and dependencies. It’s not unlike a drug that slowly but surely diminished your critical thinking, and thus, creativity. This is where the true nature of humans are unfolded: when it comes to earning something for themselves, ethics suddenly becomes a very malleable subject. On the morality, ethics, and privacy, everyone agrees that genAI is what Ron Gilbert calls a train wreck . This bears no further explanation from me: Microsoft slurped all GitHub repositories dry without taking any licenses into account, the book that I painstakingly produced in almost two years was ingested OpenAI’s systems in about two seconds, … Yet at the same time, everyone also consistently ignores all these topics in favour of their own self-interest. Why, I wonder? Everyone knows they should eat less meat. Yet almost nobody does. Everyone knows Microsoft (and probably other big tech companies) power genocide yet the adoption rate of Windows as an OS is still 95%. Why? Everyone knows the climate is going to shits yet we happily turn the other way and take the plane on a weekend trip to sip some wine and do some shopping in Italy. As Gretea Thunberg said: knowing is not enough . For GenAI, similar patterns emerge. We know it’s bad for us, yet we happily close our eyes and use it anyway. Why, I wonder? The power of a drug, the pull, the ease at which something can be done without breaking too much sweat? Here’s a possible answer I suggested before: because humans are inherently lazy. As long as Belgian supermarkets keep on stocking apples from New Zealand and Belgium, most people won’t care and just pick up whatever. As long as we keep handing out company cars and making infrastructure geared towards car drivers, most people will be driving to work instead of biking. A possible answer to the problem then might be governmental legislation to protect people living in a society from making the wrong choices. And I’m 100% sure that will work! Yet legislation is always (1) either happening way too late; or (2) minimised or manipulated by the people who wield the power because they have bought out key politicians to prevent laws like this from happening. Hence my depression. In the case of GenAI, a technology that evolves at lightning speed and is taking the world by storm, legislation will be way too late. To prove my point, in an attempt to modernise, many Belgian governmental instances already “embraced” the technology and made many blunders in doing so. The EU is currently evaluating the options. Meanwhile, the San Francisco bros are laughing. Prompt engineering is the most degenerative thing that ever happened to engineering . It’s a capitalist’s way to minimise the cost of the human. Yet I don’t see genAI disappearing any time soon. Companies and decision makers smelled the green and won’t let go. I don’t understand how capitalism works, but I know it’s been growing in power ever since we centralised cane sugar plantations with the help of slavery. GenAI is evidently yet another product of capitalism. The companies I’ve worked for wanted more and more profit each year: even though they were sometimes satisfied with last year’s profit, the target for the next year was always increased no matter what. GenAI is already responsible for thousands of layoffs in an attempt to even more aggressively push profit up. To what end, I wonder? Why? To our own detriment. It seems that our cognition is for sale, and the sale has already been made. You know what they say: no returns are accepted. Peer pressure to use genAI on the job is already prevalent as it “gets things done faster”, so quite logically also brings in money faster. Let’s worry about durability and maintenance later, shall we. Also, I’ve seen colleagues fall into the trap of obsessive agent babysitting. Whether at work, on the lunch break, or in the very late evenings: you’ve got to keep those agents spinning! Squeeze the maximum out of your tokens because they squeeze the maximum out of you. There goes our work-life balance, coming from the tools that are supposed to take over our work so we can focus more on the life part. So as long as I remain in a position to be able to choose whether I can put in the work myself for my (hobby) programming projects, I will. As long as I am in a position to bike instead of drive, to be a vegetarian instead of meat-eater, or in short, to be a concerned civilian, I will. And so should you. Even though that won’t stop this devolution from happening at all. Sure I will occasionally consult Gemini et al. to ask it a specific question regarding a broken config file that has me scratching my head. But I treat these queries as specialised internet searchers, not as a way to evade the hard work completely. I’ve become Albert Camus’s pessimist. I’m genuinely afraid of how our kids will turn out if we don’t act quickly to save our youth. Yet I won’t stop being an activist. Reading List I’d rather link to personal blog posts instead of academic publications here as we’re dealing with something that impacts us on a personal level and by the time the relevant 2026 studies are published, the landscape will have changed yet again. The following folks expressed their experience and opinion on genAI: Related topics: / genai / By Wouter Groeneveld on 8 April 2026.  Reply via email . Technical Knowledge—if we don’t have any knowledge, we won’t have the creative ability to combine them. Guess what; GenAI is actively deskilling us. The more you generate, the less you actively learn, harming your creative ability to solve problems. Creativity requires a rich mental toolbox to draw from. By prompting, you’re not exactly filling that toolbox. Communication–I see both a good and a bad thing here: if your colleagues aren’t immediately available, rubber ducking with an AI agent might help identifying that problem. On the other hand, it’s also awfully easy to stay locked inside that comfortable genAI chatbox. Why ask anyone when it talks to your mouth? Constraints—If you manage to constraint yourself (ha!) to only ask AI for 10 possible ways to approach a problem you don’t know how to approach without having it solve the problem for you , this might help you learn how to approach certain heavily constrained environments. Unfortunately, it’s very easy to just have it generate the solution as well, rendering a possible learning path useless. Critical Thinking—The more we use genAI, the less critical we are and the more likely we are accepting whatever comes out of it. Validating the the source material outside of that chatbox suddenly requires a lot of willpower. I’ve even heard people changing their entire preferred technology stack to something more popular because genAI is better at it. That’s very sad. Curiosity—Judge for yourself. What does reliance on genAI tell you about your curiosity to discover other things? Creative state of Mind—without Cal Newport’s “Deep Work”, there won’t be an “aha!” moment. The 90% transpiration, 10% inspiration is suddenly turned on its head: Claude is the one sweating for us, even at night, while all we do is press the green button and write “LGTM!”. Maybe we should take the time to read Newport’s new book Slow Productivity . Creative Techniques—GenAI itself as a technique might belong in this section; but the question is; are we the one yielding the tool or is the tool yielding us? Nolan Lawsom; How I use AI agents to write code . A clear conflicted state: it’s okay to generate away at work, but “I also don’t use AI for my open-source work, because it just feels… ick. The code is ‘mine’ in some sense, but ultimately, I don’t feel true ownership over it, because I didn’t write it”. John Allsopp: The Structure of Engineering Revolutions Dave Gauer; A programmer’s loss of social identity Cory Zue; Software got weird Doug Belshaw; Claude’s Constitution and the trap of corporate AI ethics Tom Hall; Towards a Slow Code Manifesto Rishi Baldawa; The Reviewer Isn’t the Bottleneck Information/superhighway.net; On The Need For Understanding Antoine Leblanc; Chatbot psychosis (Mastodon) “this is the main reason why i believe that chatbot addiction / chatbot psychosis is a LOT more widespread than we realise: people with a clear understanding of the ethical issues try claude once, it does a thing correctly enough, they get one-shot, and they start posting like if sephiroth was on linked-in, ethical concerns be damned. it keeps happening.” Exactly. Sean Boots; Generative AI vegetarianism Simon Willison; Perhaps not Boring Technology after all Sophie from Localghost; Stop Generating Start Thinking Micaheal Harley; AI Stance Lauren Woolsey; AI Sucks And You Shouldn’t Use It Ron Gilbert; My Dinner With AI Matthew Lamont; Generative AI is an Evil Technology Arne Brasseur; The AI Divide (Mastodon) Zach Manson; CoPilot Edited an Ad Into My PR Michael Taggart; I Used AI. It Worked. I Hated It. Bob Nystrom; The Value of Things . GenAI can have utility but not meaning. Jonny; Dismantling Claude Code source (Mastodon) . Another train wreck, as expected. Cal Newport; In Defense of Thining Hamilton Greene; Why I’m moving from F# to C# Senator Bernie Sanders vs. Claude (YouTube) Joel Chrono; Not having to work would be nice (but not like this)

0 views
Hugo 1 weeks ago

AI & Layoffs: What if Artificial Intelligence Is Just an Excuse?

Well, here we are—tech layoffs are exploding. According to RationalFX , the total number of departures is expected to reach 273,000 by the end of the year. And while this figure alone doesn't mean much, know this: it represents roughly 10 times the annual volume of pre-COVID layoffs. So, can we really say that humans are being progressively replaced by AI as so many claim ? In France, INSEE speaks of a contraction in the job market directly linked to the rise of AI . But correlation doesn't imply causation, so we're entitled to wonder if there's something else hiding behind all the hype. So I wanted to dig deeper and explore the root causes to understand this wave. And it turns out AI might not be our biggest concern. If you read the latest news, there's plenty to worry about: And I could've cited Meta, Amazon, Klarna, ASML, Ericsson, Salesforce—the list goes on. In most cases, AI is cited as one of the reasons. And this narrative has a major advantage because on paper, these companies say: we're automating, we're gaining productivity, and we're cutting fixed costs. Which tends to reassure shareholders. Block's stock price, for example, recovered a bit in February following the announcements. Same with Oracle's stock price (announcement made March 30th). Now doubts linger, and as one article put it : "Isn't this just layoffs with better marketing—AI washing?" Block is the new name for Square, a payments company you might know from its little payment terminal that's now fairly ubiquitous: But Block isn't just a payment terminal—it's also companies in crypto because its founder, Jack Dorsey, is a big believer in cryptocurrencies. Jack Dorsey is also the former founder of Twitter, which he sold to Elon Musk a few years ago. And Jack tends to think big. Twitter had 8,000 employees when he sold it—a company that now runs with 2,800 people. At Block, the company tripled its headcount post-COVID. We're talking about a 12,000-employee company that had just 4,000 pre-2020. Sure, you can understand it by looking at the COVID effect on Block's stock. Except the return to reality in 2022 hit hard. The company stagnated, and with an explosion in the payroll, things couldn't end well. So we started seeing performance improvement plans emerge. Because if you look at the economic fundamentals, as this article does , you realize that Block is far less profitable than its competitors, with gross margins that are half theirs. Today, AI is mostly a "pretty" way to hide management mistakes and reassure investors. Oracle's case is a bit different. Officially, it's not about cuts driven by productivity gains, but a reorientation of investments toward infrastructure to support AI. In their case too, the stock price is rather concerning, but it's not the main driver of changes. As one article puts it, it's primarily about investment : The job cuts at Oracle come as it has invested heavily in AI, spending both on its own infrastructure and on partnerships with other companies like OpenAI. It plans to spend at least $50bn on infrastructure this year, and it has also raised $50bn in debt in order to "meet demand" for even more AI infrastructure. Oracle is also part of the Stargate initiative, alongside OpenAI, SoftBank and MGX, an AI investment fund backed by US President Donald Trump. Here, it's really about reorienting capital from a traditional activity that's flagging to one that's supposed to replace it in a few years. In reality, I won't criticize it. It's a strategy, a bet. A huge bet, but one that falls into the same category as what Kodak should've done when digital arrived. And Oracle doesn't want to be the next Kodak. And that's really the issue—nobody wants to be the next Kodak. When a leader (like Block, Google, or Meta) lays off 10% of its workforce and its stock goes up the next day, every other company is tempted to do the same. Laying people off because you mismanaged your company would be an admission of failure. But laying people off because you're "transforming through AI" is a vision of the future. And this FOMO—fear of missing out—explains a lot of the current departure plans. Gartner calls it "RIFs before reality", the anticipation of unrealized gains : The employment deal is being rewritten in real time. CEOs are making bold moves based on AI's promise rather than its proven impact. Layoffs linked to AI dominated headlines last year, but Gartner data shows fewer than 1% were due to actual productivity gains. This anticipation drives investment reorientations. Oracle's case is representative here. Not everyone is investing in infrastructure, but many are reinvesting in engineering to automate other business functions and, most importantly, to be ready for the future. AI is no longer just a growth story; it's a cost-reduction tool, and firms are restructuring accordingly. What we're witnessing is a shift from headcount-driven expansion to automation-led productivity, a transition that will define the tech sector in the coming years. —Alan Cohen, analyst at RationalFX Now, this isn't new, and I notice a certain hypocrisy among some developers who are discovering today that their profession has always been about automating others' jobs. It's a shame to discover it when it touches us personally. Anyway, what's certain is that companies are anticipating cuts without yet having proof of the gains to come. It's not just a few layoffs—we're seeing signs, notably raised by levels.fyi's founder : we're witnessing a simplification of career paths. A layoff plan is temporary. But when you start eliminating rungs in career ladders, it signals you're anticipating a durable, global reduction in headcount. And yet, once again, the gains aren't that obvious so far. We all have our opinion on this. I consider myself more productive with AI. But not everyone agrees. But in any case, these are just opinions. There are studies on the topic of productivity, but there's no consensus. You can find studies showing we're less productive , but you can also find others saying the opposite . The causes are multiple. The first is what's called the productivity paradox : You can see the computer age everywhere but in the productivity statistics Yes, back then we wondered if computers really made us more productive. It was far from certain. This paradox is explained two ways. First, companies spend more time configuring tools, training people, and reorganizing workflows than actually producing more. Second, a new technology requires a learning period that can be quite long to master. And that's what we're seeing today—AI usage is totally new. Many are just faster at doing what they did badly before. And it's not like we know how to measure developer productivity anyway. I'll remind you that this question still hasn't found a universal answer since we started asking it. Now, I've also heard plenty of CTOs and IT directors privately say they have the means to prove it. But they don't want to . Because proving it would mean making decisions they don't want to make. And I can tell you that in this period, I'm glad I'm no longer a CTO. Still, as we've seen, productivity gains or not, can we really say all current layoffs are AI-related? Probably not. A recently cited study shows that 59% of HR leaders admit that AI was used as a "cover" to justify budget cuts that were actually driven by : But I think that would be overly reductive. It's mostly the nth demonstration that we've entered a new era post-COVID. Between rising inflation, ongoing trade wars, endless debates about tariffs, skyrocketing energy costs, various conflicts that paralyze parts of international commerce— we're really in a recession . AI is a facade to hide the rest. When Trump gets excited about his Stargate project (building datacenters), it's storytelling to hide the mess, even if it's true that AI is probably one of the drivers of the military sector in coming years and the US losing ground on it is probably making them nervous. Yes, because the worst part is that even on AI, it's not certain the people leading the dance will be American. Recent Chinese models like Ernie, DeepSeek, Qwen, and Kimi are largely on par with Gemini or ChatGPT, without necessarily costing the same. Kimi and DeepSeek reportedly cost 10% of their American counterparts during training phases. Which, incidentally, is encouraging but mostly logical—technology improves, and we've never seen tech stay this inefficient over time. The computer that sent a rocket to the moon was less powerful than our smartphone despite consuming far more energy. And for all these reasons, US companies are in full downsizing mode. Players in AI need to become more competitive. They're investing heavily while cutting payroll at the same time. Other tech companies are following suit, further constrained by hyper-unfavorable economic conditions and in a context where saying you're laying off to increase productivity is more sellable than admitting reality. And us in the middle of all this? Well... I'll be honest—I really wondered how to conclude this piece. I always try to end on a positive note, but the exercise is difficult here. I'll try anyway. Is this the end of an era? Probably the era of unreasonable hyper-growth, which isn't so bad. This forced downsizing might help us get back to basics instead of just chasing vanity metrics (like headcount). It's also a global economic shift and a US bloc that seems to be faltering. I want to see some positivity in thinking that Europe has cards to play. We're less affected than the US by the recent massive waves of layoffs. Probably because we have less insane payrolls than the US and more solid social models. While American giants painfully refocus, it's our moment in Europe to catch up. These new technologies, more accessible and efficient, let us move faster with fewer resources. Maybe it's finally time to create real European tech alternatives—more sober and pragmatic. On that note, you can go back to normal activities. Oracle just laid off 30,000 people (20% of its workforce) Block cut 40% of its headcount Over-hiring post-COVID Investor pressure to increase margins Internal strategy mistakes

0 views
A Room of My Own 1 weeks ago

Craving Quiet: Stepping Away for a While

Lately I've realised that even though I'm barely on social media, my life still feels 95% digital. I don't post on LinkedIn. My Instagram account mostly exists so I can open links people send me when I absolutely have to. I only keep a fake Facebook account for Marketplace and I use my real account (I've had it since the beginning of Facebook and all my friends live there so it stays) for Messenger only. But there is more than social media to occupy our time now. My days are still full of feeds, links, apps, messages (whatsapp groups and such), digital projects, and little things I feel like I should be keeping up with. And they are easy to keep up with, my phone is always in my hand anyway. RELATED: I Choose Living Over Documenting On the Compulsion to Record The Journal Project I Can’t Quit The Art of Organizing (Things That Don’t Need to Be Organized) At work we showcase our AI agents and I wonder (from my anecdotal experience) if we are creating more busy work for ourselves and replacing reflection and with it, the actual prouctivity and output and good old ““getting the job done.” Most of our work meetings now have extensive transcripts that turn into minutes, notes, action points and insights. I remember when the output of such a meeting would be 2-3 points that we actually remembered. AI Generated Workslop certainly is a thing now. I need a break from it all. And from all the self-imposed shoulds such as scanning my old journals into Day One. Backing up Day One, which hasn't been backed up in a while. An external hard drive backup that's probably a year overdue. A Trello board full of things I want to do but don't really want to or have to do, or maybe I want to do them but can't justify the time when I already feel so busy. After a full day of work and virtual meetings, I feel completely depleted. Those self-imposed obligations, things that used to be fun because they were few and far between, are no longer acceptable. I used to sneak in 15 minutes of personal things at work. Now when I have a break, I'd rather grab a coffee with someone or go for a walk. I crave analog. I crave nature. I crave quiet thinking time (not with a meditation app). I have made some changes already and they seem to be sticking. We have dinner at the table now, which has been good, at least we get some family time before everyone retreats to their own corners. We used to eat while watching a show together as a family, which is fine every now and then, but it was too much of it all. But still my phone is somewhere nearby, and I'm half-watching TV and half-checking a message or voice journaling into an app. None of it is thoughtful. It's just me blabbering. My brain feels like it's all over the place. I used to be able to sit with my own thoughts. I haven't been able to do that in a long time. My daughter broke her arm two weeks ago. She has a purple cast all her friends signed, and she was wondering whether to keep it when it comes off. I told her how I broke my arm as a kid, and she asked if I kept my cast. I said I would have liked to, but what we have now is better. I can take a clear photo of hers and she'll have that memory without keeping the physical thing. Then she asked if I had a photo of mine. I didn't. It never even occurred to me. Back then we took maybe 20 photos a year, if that, and they were all the more precious for it. Now I'm struggling to keep my monthly saves under 150 photos and screenshots, most of which I probably don't need. RELATED: My Photo Management and Memory Keeping Workflow I love my Day One journals , I really do. I just exported all of 2025 to PDF and JSON. But reading back through it, it's every tiny minutia of my life. I like to think it'll be interesting to me one day. Probably not to anyone else. And I wonder whether the time I spent on it was worth it. Yes, there are some insights there , but nothing that I didn’t already know. Had I allowed myself that thinking time instead of outsourcing it to AI. RELATED: Committing to the Thinking Life If my house burned down and I lost everything, the memories that matter are still in my head. I'm a cumulative experience of all of it. Do I need the artifact to know who I am? I still have journals from my 20s and 30s sitting back home in Bosnia. Thick ones, full of pasted tickets and stubs and mementos. I haven't looked at them in years but I can't let them go. My plan is to eventually scan them, maybe pay one of my kids to do it since they won't be able to read my handwriting anyway. RELATED: Letting Go of Old Journals and Mementos But anyway. The point is, I just need a break. From reading things online, from note-keeping, from digital journaling, blogging, saving notes and highlights (even my Readwise subscription feels intrusive now), from all of it. I've decided to do a 30-day digital detox. Within reason, because I still have to work. But I'm off until Tuesday, so I have a few days to ease into it. I'm lucky and privileged that I can do this. That I can shut down for a while and stop following things I can't influence and let go of expectations I put on myself. So that's what I'm doing. Simplifying my phone, deleting apps, putting the phone away when I get home. If we're watching something as a family, fine. One episode. But otherwise, even if I'm bored and restless, I'll go for a walk or play a board game, read a book. Journal (on paper). I'll do nothing, like I used to. Go to bed early. Meet a friend for coffee (and be more proactive about that). It's all become too hard because easy distractions that scratch the itch of everything are too easy. Calm my mind. Slow down. It's been too much. Time to reclaim myself. And if you've gotten this far, the world is reminding me once again of E.M. Forster's The Machine Stops , which I wrote about in 2020 . It feels eerily even more relevant now.

0 views
Manuel Moreale 1 weeks ago

Anthony Nelzin-Santos

This week on the People and Blogs series we have an interview with Anthony Nelzin-Santos, whose blog can be found at z1nz0l1n.com . Tired of RSS? Read this in your browser or sign up for the newsletter . People and Blogs is supported by the "One a Month" club members. If you enjoy P&B, consider becoming one for as little as 1 dollar a month. Bonjour ! I’m a militant wayfarer, budding typographer, pathological reader, slow cyclist, obsessive tinkerer, dangerous cook, amateur bookbinder, homicidal gardener, mediocre sewist, and fanatical melomaniac living in Lyon (France). I was a technology journalist and journalism teacher for sixteen years, but i now work in instructional design. In my spare time, i take photos of old storefronts to preserve a rapidly fading typographical tradition. One of these days, i’ll finally finish the typefaces i’ve been working on forever. And my novel. And the painting of the bathroom. (My wife is a saint.) I was born a few years before the web was invented and grew up at this fascinating time when everybody wanted to do something with it, but nobody knew quite what yet. We were still supposed to learn Logo and Pascal in technology class, but most of the teachers understood the importance of the web and taught us the basics of HTML and CSS. I built my first website in 2000… as a school assignment! By 2007, i was one of those insufferable tech bloggers who made enough money to feel entitled, but not enough to feel safe. (I moonlighted as a graphic designer.) When more established outlets came knocking at my door, i shut down my blog and became one of those insufferable tech journalists who make enough money to feel entitled, but not enough to feel safe. (I moonlighted as a journalism teacher.) I kept a personal blog under the “zinzolin” moniker. This shade of purple is my favourite colour, partly because it sounds a bit like my name. Over the years, it became more and more difficult to find the energy to write recreationally after having spent the day writing professionally. In 2025, feeling more than a little burnt out, i rebooted my blog and switched from French to English. Fortunately, the name is equally weird in both languages. I don’t have a process so much as a way of managing the incessant chatter in my head. I write to give myself the permission to forget, and i publish to gift myself the ability to remember. You’ll never catch me without some way to capture those little “brain itches” — a notebook, the Bloom app, a digital recorder, the back of my hand… (I wrote part of this interview as a long series of text messages to myself!) In the middle of the week, i start reviewing my notes to find a common theme or extract the strongest idea. When an incomplete thought keeps coming back, i don’t try to force it by staring at a blinking cursor. I take a long walk, and usually, i have to stop part way to write. Most of the actual blogging is done long before i sit down to properly draft my weekly note. I have this romantic notion that the more comfortable i am, the more i can edit, the worse my writing tends to get. If i could, i’d write everything longhand in a rickety train, stream-of-consciousness style, and publish the raw scans of my notebooks. You wouldn’t be able to read half of it, but i can assure you the illegible half would be Nobel-prize worthy. But then, some things only happen after a few hours of diligent editing. If i give myself enough time, i can stop transcribing my notes and start conversing with them. There’s always something worth exploring in the gap between our past and present selves – even if the past was two days ago – but that delicate work requires a conducive environment. Judging by my recent output, it looks like this environment comprises a good chair , a MacBook Air on one of those ugly lap desks, my custom international QWERTY layout , iA Writer for writing and Antidote for proofreading, cosy lighting, just the right amount of background noise, and most important of all, a pot of delicious coffee. I’ve tried pretty much every CMS and SSG under the sun, but i’ve always come back to WordPress, until Matt Mullenweg reminded us that a benevolent dictator still is a dictator . Z1NZ0L1N is now built on Ghost and hosted by Magic Pages . I used to use Tinylytics and Buttondown , but i’m now using Ghost’s integrated analytics and newsletter features. My other websites are hosted on a VPS with Infomaniak , which is also where i get my domain names, e-mail, and assorted cloud services. That’s a question i had to ask myself when i rebooted Z1NZ0L1N last year. I switched to English in a bid to better separate my professional output from my recreational output. I jettisoned most of my audience, but i found a new community around the IndieWeb Carnival and quickly rebuilt a readership on my own merits. I get excited each time i get an e-mail from someone i don’t know from a country on the other side of the globe. I wanted to find a way to publish regularly without turning Z1NZ0L1N into the umpteenth link blog. After a few experiments, i’ve settled on a weekly note that’s part “what i’m doing”, part “what the rest of the world is doing”. This is old-school blogging meets recommendation algorithms — and i love it. Some things haven’t changed, though, and will never change. I use an open-source CMS that i could host myself, not a proprietary platform that i can’t control. I designed my theme myself. I don’t play the SEO/GEO game. I pay a little less than €10/month for Magic Pages’ starter plan with the custom themes add-on. Considering that it saves me €15/month in third-party services, i’d say it’s a fair price. I pay €12/year for the domain, but i also registered a few variations, including , which was first registered in 1999! Blogging is my least expensive hobby — by far. As someone who’s worked a lot on the economics of independent publishing, i’m happily subscribed to a few news outlets and magazines. I like the idea of $1/month memberships for blogs, but in practice, i find it hard to track multiple micro-subscriptions on top of my existing (and frankly far too numerous) digital subscriptions. I wonder if we should create blogging collectives, almost like unions and coops, to collect and redistribute a single subscription in between members. In the meantime, i’ll continue not talking about my Ko-Fi page . The Forest and Ye Olde Blogroll are fantastic discovery tools. A lot of my favourite bloggers have already been featured in People and blogs : VH Belvadi, BSAG, Frank Chimero, Keenan, Piper Haywood, Nick Heer, Tom McWright, Riccardo Mori, Jim Nielsen, Kev Quirk, Arun Venkatesan, Zinzy… I’d love to see how Rob Weychert , Chris Glass , Josh Ginter or Melanie Richards would answer. Their approach to blogging couldn’t be more different, but they each informed mine in their own way. Since 2008, i’ve taken thousands of photos of old storefronts. It began as a way to inform my typographical practice, but it rapidly became an excuse to go out and pay attention – really pay attention – to the world around me. You wouldn’t believe the things i’ve discovered in side streets, the number of conversations i’ve struck after taking a picture of a once-beloved shop, and how my way of looking at the evolution of cities has entirely changed. If you’re up for a little challenge, find your own collection. It might be cool doors, weird postboxes, triangular things, every bookshop in Nova Scotia , sewer manholes, purple things, number signs… It’ll give you another perspective not only when travelling in foreign places, but also on your (not so) familiar surroundings. It doesn’t cost a penny, but it’ll pay off immensely. Now that you're done reading the interview, go check the blog and subscribe to the RSS feed . If you're looking for more content, go read one of the previous 135 interviews . People and Blogs is possible because kind people support it.

0 views
Kaushik Gopal 1 weeks ago

We are becoming Harness Engineers

The role of a software engineer is shifting. Not toward writing more code but toward building the environment that makes agents reliable. Think about what you actually do with Claude Code or Codex today: you configure AGENTS.md files, set up MCP servers, write skills and hooks, build feedback loops and tune sub-agents. You’re not writing as much of the software anymore. You’re engineering the harness around the thing that writes the software. Mitchell Hashimoto first coined the term harness engineering — the work of shaping the environment around an agent so it can act reliably. What the model sees, what tools it has, how it gets feedback, when humans step in. We keep hearing that agents will replace engineers. That shouldn’t be the focus of the change we’re seeing. What’s actually happening is product people shipping features directly. A well-harnessed agent lets someone with product instinct but little engineering background make meaningful changes — safely . The harness engineer makes that possible. Guardrails, design choices, blast radius controls, feedback loops. The scaffolding that turns “just prompt it” into something a team can trust. I say this from first-hand experience. If you want to go deeper, listen to the episode where my cohost and I dug into it. We landed on five pillars: Honestly one of the most important episodes we’ve recorded. agent legibility closed feedback loops persistent memory entropy control blast radius controls

0 views
Simon Willison 1 weeks ago

Highlights from my conversation about agentic engineering on Lenny's Podcast

I was a guest on Lenny Rachitsky's podcast, in a new episode titled An AI state of the union: We've passed the inflection point, dark factories are coming, and automation timelines . It's available on YouTube , Spotify , and Apple Podcasts . Here are my highlights from our conversation, with relevant links. 4:19 - The end result of these two labs throwing everything they had at making their models better at code is that in November we had what I call the inflection point where GPT 5.1 and Claude Opus 4.5 came along. They were both incrementally better than the previous models, but in a way that crossed a threshold where previously the code would mostly work, but you had to pay very close attention to it. And suddenly we went from that to... almost all of the time it does what you told it to do, which makes all of the difference in the world. Now you can spin up a coding agent and say, build me a Mac application that does this thing , and you'll get something back which won't just be a buggy pile of rubbish that doesn't do anything. 5:49 - I can churn out 10,000 lines of code in a day. And most of it works. Is that good? Like, how do we get from most of it works to all of it works? There are so many new questions that we're facing, which I think makes us a bellwether for other information workers. Code is easier than almost every other problem that you pose these agents because code is obviously right or wrong - either it works or it doesn't work. There might be a few subtle hidden bugs, but generally you can tell if the thing actually works. If it writes you an essay, if it prepares a lawsuit for you, it's so much harder to derive if it's actually done a good job, and to figure out if it got things right or wrong. But it's happening to us as software engineers. It came for us first. And we're figuring out, OK, what do our careers look like? How do we work as teams when part of what we did that used to take most of the time doesn't take most of the time anymore? What does that look like? And it's going to be very interesting seeing how this rolls out to other information work in the future. Lawyers are falling for this really badly. The AI hallucination cases database is up to 1,228 cases now! Plus this bit from the cold open at the start : It used to be you'd ask ChatGPT for some code, and it would spit out some code, and you'd have to run it and test it. The coding agents take that step for you now. And an open question for me is how many other knowledge work fields are actually prone to these agent loops? 8:19 - I write so much of my code on my phone. It's wild. I can get good work done walking the dog along the beach, which is delightful. I mainly use the Claude iPhone app for this, both with a regular Claude chat session (which can execute code now ) or using it to control Claude Code for web . 9:55 If you're vibe coding something for yourself, where the only person who gets hurt if it has bugs is you, go wild. That's completely fine. The moment you ship your vibe coding code for other people to use, where your bugs might actually harm somebody else, that's when you need to take a step back. See also When is it OK to vibe code? 12:49 The reason it's called the dark factory is there's this idea in factory automation that if your factory is so automated that you don't need any people there, you can turn the lights off. Like the machines can operate in complete darkness if you don't need people on the factory floor. What does that look like for software? [...] So there's this policy that nobody writes any code: you cannot type code into a computer. And honestly, six months ago, I thought that was crazy. And today, probably 95% of the code that I produce, I didn't type myself. That world is practical already because the latest models are good enough that you can tell them to rename that variable and refactor and add this line there... and they'll just do it - it's faster than you typing on the keyboard yourself. The next rule though, is nobody reads the code. And this is the thing which StrongDM started doing last year. I wrote a lot more about StrongDM's dark factory explorations back in February. 21:27 - It used to be, you'd come up with a spec and you hand it to your engineering team. And three weeks later, if you're lucky, they'd come back with an implementation. And now that maybe takes three hours, depending on how well the coding agents are established for that kind of thing. So now what, right? Now, where else are the bottlenecks? Anyone who's done any product work knows that your initial ideas are always wrong. What matters is proving them, and testing them. We can test things so much faster now because we can build workable prototypes so much quicker. So there's an interesting thing I've been doing in my own work where any feature that I want to design, I'll often prototype three different ways it could work because that takes very little time. I've always loved prototyping things, and prototyping is even more valuable now. 22:40 - A UI prototype is free now. ChatGPT and Claude will just build you a very convincing UI for anything that you describe. And that's how you should be working. I think anyone who's doing product design and isn't vibe coding little prototypes is missing out on the most powerful boost that we get in that step. But then what do you do? Given your three options that you have instead of one option, how do you prove to yourself which one of those is the best? I don't have a confident answer to that. I expect this is where the good old fashioned usability testing comes in. More on prototyping later on: 46:35 - Throughout my entire career, my superpower has been prototyping. I've been very quick at knocking out working prototypes of things. I'm the person who can show up at a meeting and say, look, here's how it could work. And that was kind of my unique selling point. And that's gone. Anyone can do what I could do. 26:25 - I'm finding that using coding agents well is taking every inch of my 25 years of experience as a software engineer, and it is mentally exhausting. I can fire up four agents in parallel and have them work on four different problems. And by like 11 AM, I am wiped out for the day. [...] There's a personal skill we have to learn in finding our new limits - what's a responsible way for us not to burn out. I've talked to a lot of people who are losing sleep because they're like, my coding agents could be doing work for me. I'm just going to stay up an extra half hour and set off a bunch of extra things... and then waking up at four in the morning. That's obviously unsustainable. [...] There's an element of sort of gambling and addiction to how we're using some of these tools. 45:16 - People talk about how important it is not to interrupt your coders. Your coders need to have solid two to four hour blocks of uninterrupted work so they can spin up their mental model and churn out the code. That's changed completely. My programming work, I need two minutes every now and then to prompt my agent about what to do next. And then I can do the other stuff and I can go back. I'm much more interruptible than I used to be. 28:19 - I've got 25 years of experience in how long it takes to build something. And that's all completely gone - it doesn't work anymore because I can look at a problem and say that this is going to take two weeks, so it's not worth it. And now it's like... maybe it's going to take 20 minutes because the reason it would have taken two weeks was all of the sort of crufty coding things that the AI is now covering for us. I constantly throw tasks at AI that I don't think it'll be able to do because every now and then it does it. And when it doesn't do it, you learn, right? But when it does do something, especially something that the previous models couldn't do, that's actually cutting edge AI research. And a related anecdote: 36:56 - A lot of my friends have been talking about how they have this backlog of side projects, right? For the last 10, 15 years, they've got projects they never quite finished. And some of them are like, well, I've done them all now. Last couple of months, I just went through and every evening I'm like, let's take that project and finish it. And they almost feel a sort of sense of loss at the end where they're like, well, okay, my backlog's gone. Now what am I going to build? 29:29 - So ThoughtWorks, the big IT consultancy, did an offsite about a month ago , and they got a whole bunch of engineering VPs in from different companies to talk about this stuff. And one of the interesting theories they came up with is they think this stuff is really good for experienced engineers, like it amplifies their skills. It's really good for new engineers because it solves so many of those onboarding problems. The problem is the people in the middle. If you're mid-career, if you haven't made it to sort of super senior engineer yet, but you're not sort of new either, that's the group which is probably in the most trouble right now. I mentioned Cloudflare hiring 1,000 interns , and Shopify too. Lenny asked for my advice for people stuck in that middle: 31:21 - That's a big responsibility you're putting on me there! I think the way forward is to lean into this stuff and figure out how do I help this make me better? A lot of people worry about skill atrophy: if the AI is doing it for you, you're not learning anything. I think if you're worried about that, you push back at it. You have to be mindful about how you're applying the technology and think, okay, I've been given this thing that can answer any question and often gets it right. How can I use this to amplify my own skills, to learn new things, to take on much more ambitious projects? [...] 33:05 - Everything is changing so fast right now. The only universal skill is being able to roll with the changes. That's the thing that we all need. The term that comes up most in these conversations about how you can be great with AI is agency . I think agents have no agency at all. I would argue that the one thing AI can never have is agency because it doesn't have human motivations. So I'd say that's the thing is to invest in your own agency and invest in how to use this technology to get better at what you do and to do new things. The fact that it's so easy to create software with detailed documentation and robust tests means it's harder to figure out what's a credible project. 37:47 Sometimes I'll have an idea for a piece of software, Python library or whatever, and I can knock it out in like an hour and get to a point where it's got documentation and tests and all of those things, and it looks like the kind of software that previously I'd have spent several weeks on - and I can stick it up on GitHub And yet... I don't believe in it. And the reason I don't believe in it is that I got to rush through all of those things... I think the quality is probably good, but I haven't spent enough time with it to feel confident in that quality. Most importantly, I haven't used it yet . It turns out when I'm using somebody else's software, the thing I care most about is I want them to have used it for months. I've got some very cool software that I built that I've never used . It was quicker to build it than to actually try and use it! 41:31 - Everyone's like, oh, it must be easy. It's just a chat bot. It's not easy. That's one of the great misconceptions in AI is that using these tools effectively is easy. It takes a lot of practice and it takes a lot of trying things that didn't work and trying things that did work. 19:04 - In the past sort of three to six months, they've started being credible as security researchers, which is sending shockwaves through the security research industry. See Thomas Ptacek: Vulnerability Research Is Cooked . At the same time, open source projects are being bombarded with junk security reports: 20:05 - There are these people who don't know what they're doing, who are asking ChatGPT to find a security hole and then reporting it to the maintainer. And the report looks good. ChatGPT can produce a very well formatted report of a vulnerability. It's a total waste of time. It's not actually verified as being a real problem. A good example of the right way to do this is Anthropic's collaboration with Firefox , where Anthropic's security team verified every security problem before passing them to Mozilla. Of course we had to talk about OpenClaw! Lenny had his running on a Mac Mini. 1:29:23 - OpenClaw demonstrates that people want a personal digital assistant so much that they are willing to not just overlook the security side of things, but also getting the thing running is not easy. You've got to create API keys and tokens and install stuff. It's not trivial to get set up and hundreds of thousands of people got it set up. [...] The first line of code for OpenClaw was written on November the 25th. And then in the Super Bowl, there was an ad for AI.com, which was effectively a vaporware white labeled OpenClaw hosting provider. So we went from first line of code in November to Super Bowl ad in what? Three and a half months. I continue to love Drew Breunig's description of OpenClaw as a digital pet: A friend of mine said that OpenClaw is basically a Tamagotchi. It's a digital pet and you buy the Mac Mini as an aquarium. In talking about my explorations of AI for data journalism through Datasette : 1:34:58 - You would have thought that AI is a very bad fit for journalism where the whole idea is to find the truth. But the flip side is journalists deal with untrustworthy sources all the time. The art of journalism is you talk to a bunch of people and some of them lie to you and you figure out what's true. So as long as the journalist treats the AI as yet another unreliable source, they're actually better equipped to work with AI than most other professions are. Obviously we talked about pelicans riding bicycles : 56:10 - There appears to be a very strong correlation between how good their drawing of a pelican riding a bicycle is and how good they are at everything else. And nobody can explain to me why that is. [...] People kept on asking me, what if labs cheat on the benchmark? And my answer has always been, really, all I want from life is a really good picture of a pelican riding a bicycle . And if I can trick every AI lab in the world into cheating on benchmarks to get it, then that just achieves my goal. 59:56 - I think something people often miss is that this space is inherently funny. The fact that we have these incredibly expensive, power hungry, supposedly the most advanced computers of all time. And if you ask them to draw a pelican on a bicycle, it looks like a five-year-old drew it. That's really funny to me. Lenny asked if I had anything else I wanted to leave listeners with to wrap up the show, so I went with the best piece of news in the world right now. 1:38:10 - There is a rare parrot in New Zealand called the Kākāpō. There are only 250 of these parrots left in the world. They are flightless nocturnal parrots - beautiful green dumpy looking things. And the good news is they're having a fantastic breeding season in 2026, They only breed when the Rimu trees in New Zealand have a mass fruiting season, and the Rimu trees haven't done that since 2022 - so there has not been a single baby kākāpō born in four years. This year, the Rimu trees are in fruit. The kākāpō are breeding. There have been dozens of new chicks born. It's a really, really good time. It's great news for rare New Zealand parrots and you should look them up because they're delightful. Everyone should watch the live stream of Rakiura on her nest with two chicks ! Here's the full list of chapters Lenny's team defined for the YouTube video: You are only seeing the long-form articles from my blog. Subscribe to /atom/everything/ to get all of my posts, or take a look at my other subscription options . The November inflection point Software engineers as bellwethers for other information workers Writing code on my phone Responsible vibe coding Dark Factories and StrongDM The bottleneck has moved to testing This stuff is exhausting Interruptions cost a lot less now My ability to estimate software is broken It's tough for people in the middle It's harder to evaluate software The misconception that AI tools are easy Coding agents are useful for security research now Journalists are good at dealing with unreliable sources The pelican benchmark And finally, some good news about parrots YouTube chapters 00:00 : Introduction to Simon Willison 02:40 : The November 2025 inflection point 08:01 : What's possible now with AI coding 10:42 : Vibe coding vs. agentic engineering 13:57 : The dark-factory pattern 20:41 : Where bottlenecks have shifted 23:36 : Where human brains will continue to be valuable 25:32 : Defending of software engineers 29:12 : Why experienced engineers get better results 30:48 : Advice for avoiding the permanent underclass 33:52 : Leaning into AI to amplify your skills 35:12 : Why Simon says he's working harder than ever 37:23 : The market for pre-2022 human-written code 40:01 : Prediction: 50% of engineers writing 95% AI code by the end of 2026 44:34 : The impact of cheap code 48:27 : Simon's AI stack 54:08 : Using AI for research 55:12 : The pelican-riding-a-bicycle benchmark 59:01 : The inherent ridiculousness of AI 1:00:52 : Hoarding things you know how to do 1:08:21 : Red/green TDD pattern for better AI code 1:14:43 : Starting projects with good templates 1:16:31 : The lethal trifecta and prompt injection 1:21:53 : Why 97% effectiveness is a failing grade 1:25:19 : The normalization of deviance 1:28:32 : OpenClaw: the security nightmare everyone is looking past 1:34:22 : What's next for Simon 1:36:47 : Zero-deliverable consulting 1:38:05 : Good news about Kakapo parrots

0 views
Stratechery 1 weeks ago

An Interview with Asymco’s Horace Dediu About Apple at 50

An interview with Asymco's Horace Dediu about his career in tech, Apple's first 50 years, and the prospects for the next 50, particularly in the face of AI

0 views
David Bushell 2 weeks ago

I quit. The clankers won.

… is what I’m reading far too often! Some of you are losing faith! A growing sentiment amongst my peers — those who haven’t already resigned to an NPC career path † — is that blogging is over. Coding is cooked. What’s the point of sharing insights and expertise when the Cognitive Dark Forest will feed on our humanity? Before I’m dismissed as an ill-informed hater please note: I’ve done my research. † To be fair it’s a valid choice in this economy. Clock in, slop around, clock out. Why not? It’s never been more important to blog. There has never been a better time to blog. I will tell you why. We’re being starved for human conversation and authentic voices. What’s more: everyone is trying to take your voice away. Do not opt-out of using it yourself. First let’s accept the realities. The giant plagiarism machines have already stolen everything. Copyright is dead. Licenses are washed away in clean rooms . Mass surveillance and tracking are a feature, privacy is a bug. Everything is an “algorithm” optimised to exploit. How can we possibly combat that? From a purely selfish perspective it’s never been easier to stand out and assert yourself as an authority. When everyone is deferring to the big bullshitter in the cloud your original thoughts are invaluable. Your brain is your biggest asset. Share it with others for mutual benefit. I find writing stuff down improves my memory and hardens my resolve. I bet that’s true for you too. It’s part rote learning part rubberducking † . Writing publicly in blog form forces me to question assumptions. Even when research fails me Cunningham’s Law saves me. † Some will claim writing into a predictive chat box helps too, and sure, they’re absolutely right! Blogging makes you a better professional. No matter how small your audience, someone will eventually stumble upon your blog and it will unblock their path. Don’t accept a fate being forced upon you. The AI industry is 99% hype; a billion dollar industrial complex to put a price tag on creation. At this point if you believe AI is ‘just a tool’ you’re wilfully ignoring the harm . (Regardless, why do I keep being told it’s an ‘extreme’ stance if I decide not to buy something?) The 1% utility AI has is overshadowed by the overwhelming mediocracy it regurgitates. We’re saying goodbye to Sora. To everyone who created with Sora, shared it, and built community around it: thank you. What you made with Sora mattered, and we know this news is disappointing. @soraofficialapp - XCancel Is there anything, in the entire recorded history of human creation, that could have possibly mattered less than the flatulence Sora produced? NFTs had more value. I’m not protective over the word “art”. Generative AI is art. It’s irredeemably shit art; end of conversation. A child’s crayon doodle is also lacking refined artistry but we hang it on our fridge because a human made it and that matters. We care and caring has a positive effect on our lives. When you pass human creativity through the slop wringer, or just prompt an incantation, the result is continvoucly morged ; a vapid mockery of the input. The garbage out no longer matters, nobody cares, nobody benefits. I forgot where I was going with this… oh right: don’t resign yourself to the deskilling of our craft . You should keep blogging! Take pride in your ability and unique voice. But please don’t desecrate yourself with slop. The only winning move is not to play. WarGames (1983) We’ve gotten too comfortable with the convenience of Big Tech . We do not have to continue playing their game. Don’t buy the narratives they’re selling. The AI industry is built on the predatory business model of casinos. Except they’ve forget the house is supposed to win. One upside of this looming economic and intellectual depression is that the media is beginning to recognise gate keepers are no longer the hand that feeds them. Big Tech is not the web. You don’t have to use it nor support it. Blog for the old web , the open web , the indie web — the web you want to see. And if you think I’m being dramatic and I’ve upset your new toys, you’re welcome to be left behind in the miasmatic dystopia these technofacists are racing to build. Thanks for reading! Follow me on Mastodon and Bluesky . Subscribe to my Blog and Notes or Combined feeds.

0 views
./techtipsy 2 weeks ago

Improving my focus by giving up my big monitor

Keeping my focus has been challenging. It’s not a new phenomenon, and I suspect that there are contributing factors that have lead to the unfocused state dominating. For example, I’ve been that guy who wants to be on top of things, to be in the loop, to respond to urgent issues. It feels fantastic to be in that firefighter role as it gives me the feeling of having an impact, but it results in me being drained at the end of the day and often over-caffeinated. One day I was doing work on my laptop on a couch because hitting 30 apparently means that sleeping slightly incorrectly results in debilitating back pain. During that session, I was working on a larger task and making tons of tiny little changes that needed to be done in order to release a new feature. I was finally in the zone again, and it felt fantastic! That’s when I decided to start an experiment: can I improve my focus by giving up my big monitor? I’ve done this type of “experiment” a few times in the past when the power has gone out and my super duper ergonomic setup has become useless. No power, no USB-C dock, no monitor. It wasn’t that fun and my eyes hated reading text off of a laptop screen. A few things have changed since then: Almost a month in, I’ve had a pleasant experience with this experiment. I feel more focused. Yeah, that’s it. Am I actually more focused is up for debate, as I’m not sure how to measure it objectively. 1 Working off of a single screen forces me to focus at what’s at hand. Alt-tabbing to a different app is quick, but just enough to deter me from doing it in meetings or other focused tasks. In my personal free time 2 , this has also resulted in computer use becoming more intentional. On a 34" ultrawide monitor, it was too easy to put YouTube running on the left side, and whatever else on the right. It was distracting and resulted in time being wasted doing nothing. Interestingly enough, making computer use more intentional was a trick that I tried when recovering from burnout, and it helped a lot. As a side effect, the power consumption of my whole home office setup is significantly smaller, as I don’t have to power my ultrawide monitor. That made up most of the power consumption, with peaks of up to 100W. I also don’t have to fight with my dock killing my whole network, because there is no dock. If you’re just cleaning up your desk and plopping your laptop on there, you will likely have a bad time. The posture will be off, and depending on your laptop, the keyboard and touchpad combination can prove to be an ergonomic nightmare. At the very least, you should put your laptop up somewhere higher. Ideally, it should be using a stand that allows you to use your favourite wireless keyboard and mouse below it. A simple laptop stand could get you most of the way there, but the ideal solution is a freely adjustable monitor arm combined with a VESA-mounted laptop holder. This gives you the freedom to place the laptop exactly as you’d like while leaving the desk free for your peripherals. Most monitor arm laptop holders have side arms that keep it in place, but I found them to be extremely annoying, so I removed them by disassembling the holder and yanking out the side arms and springs. You may still need them if you are using a very aggressive vertical angle, but I hated having to give up one USB-A port and blocking about 25% of the exhaust fan also didn’t seem like a good idea. Mounting the laptop with the springy side arms was also awkward. If you’re using a desktop and have a big display, then intentionally using a smaller and cheaper one for a while may prove to be just as effective. If you’re using a laptop with a horrible display with poor viewing angles, glare and crappy resolution (which a lot of older ThinkPads have), then you can still try this out, but I suspect that you’ll not have a very good experience with it due to this reason alone. I still prefer to do my gaming sessions on a big screen. It’s more immersive, and I can make out tiny details better, such as spotting a car in the distance while driving in the oncoming lane in Need for Speed Most Wanted. I’m happy with this setup. That’s all I ever needed. go ahead, try to measure developer productivity objectively. Good luck!  ↩︎ that’s what I call the time window between putting my son to sleep and midnight.  ↩︎ GNOME has working fractional scaling that you can simply enable in display settings ThinkPad displays have gotten better, with the picture being quite cromulent, and the 16:10 aspect ratio helps fit more on the screen the nature of my work has changed and will keep changing in the near future go ahead, try to measure developer productivity objectively. Good luck!  ↩︎ that’s what I call the time window between putting my son to sleep and midnight.  ↩︎

0 views
Allen Pike 2 weeks ago

The Rise of Transparency

Small companies are, by default, very transparent. When there are 4 people working in a room, you have a direct line of sight on what everybody else is doing, and why. Your docs, Slack channels, and repositories are open to everybody. When the CEO has an epiphany that changes everything, you all know right away – probably because you were at lunch together when it happened. Thus, startup founders will often get religion about transparency. “Our culture,” they’ll declare, “is to be radically transparent! Everything defaults to open. We hire adults, expect them to do great work, and give them the context they need.” Yay transparency! And this works pretty well. Transparent orgs tend to delegate more effectively, have higher accountability, less politics, faster trust, and just plain ship more. Transparency helps bigger orgs adapt more quickly to the ground truth, responding to customer signals that execs might not be directly exposed to. But, at a certain scale, radical transparency strains. Some idle musing by the CEO sends a team off on an unimportant side quest. A well-justified compensation anomaly upsets a group who is missing background information. A 450-message Slack thread about bike shed paint color choices devolves into factions, hashtags, and philosophical arguments about the morality of taupe. #nevertaupe And if you talk to people at a large yet highly transparent company, you’ll hear about the hazards of the relentless firehose . A thousand shared Slack channels, to start. But also a glut of docs – some critical, most unmaintained. Then there’s the meeting notes, meeting recordings, and meeting invites. Plus proposals, requests for comment, and requests to comment on your proposals’ comments’ resolutions. “So, you like information, eh? Well, have all the information in the world!” How do you make sense of all this? While some people are tenaciously able to find, within this chaos, the important info they need to do great work, a lot of otherwise-capable people get easily distracted by information that just might be urgent, provocative, or even just… shiny. 💫 Meanwhile, allowing everybody access to every historical doc is occasionally useful, but it also presents an ever-growing surface area for leaks and legal liability. Are you sure there isn’t something highly sensitive or disagreeable in those 99,999 unmaintained Notion docs? So, as companies grow, they tend to lock information down. Some – Netflix, Stripe, Shopify – do their best to keep as transparent as possible while still complying with necessary guardrails. Others – Apple, Palantir, Oracle – move toward a need-to-know basis, ensuring information flows top-down. With more control over information, it’s easier to ensure that leaks or internal distractions don’t derail your plans for surprising product launches and/or world domination. Of course, every company’s culture is forged by the market they operate in, but there’s always some tradeoff here. And as companies grow, they tend to regress to a boring middle ground. However. As with many tradeoffs, the balance has recently begun to shift. Recently, we’ve seen a revolution in tools that can make better use of the firehose. Slack can now summarize your unread messages, albeit with mixed effectiveness. Tools like Glean and Unblocked can consider a mountain of your company’s data and answer important questions about it, albeit limited to the data they can actually see. And large open companies like Shopify and Stripe have internal tools that let employees’ agents query, analyze, and act on the copious data any given employee has access to – albeit with some sharp edges and exfiltration risks. Just as LLMs are making the world’s data more useful to the world, they’re making companies’ internal data more useful to employees. Of course, this can be misused! In some companies we’ll see further secrecy – I’ve heard of AI search tools and MCPs letting employees find accidentally-visible compensation data and other spicy docs that hadn’t been audited. I’ve heard of support agents giving customers true-but-problematic information because they surfaced it with internal AI tooling without proper training. But as we evolve past early growing pains, and into teams and processes fully making use of this stuff, the anecdata points toward this new tooling becoming a superpower. Agents’ newfound ability to effectively query and reason about far more data than can fit into context is making the long tail of communications and docs much more useful for decision-making – but only when people have access to the relevant data. Given that, the maturation of AI tooling will motivate companies to become more transparent . In 2024, the cost of being internally secretive was meaningful but manageable. Although Apple keeping information need-to-know sometimes leads to waste, or important changes being slow to diffuse through layers of management, they’ve done, like, pretty well for themselves? With all the scrutiny from press, competitors, and regulators, you can see why they’ve kept it up. But as all companies increasingly have tools that can assess, consider, analyze, and make use of all the business’ communications and documents, what kinds of org are going to benefit most? Well, the ones that let their employees access more context. Extremely transparent orgs like Zapier, GitLab, and PostHog that might have struggled to cope with their firehoses – and who often had gaps in the data due to untranscribed meetings and decisions – will increasingly be able to leverage it. Sure, not all of it, certainly not at first. (Some of it is just junk.) But increasingly more of it. And critically, it won’t just be executives that will be able to attend to all this knowledge. The frontend dev working on your internal admin dashboard should be flagged that the React upgrade issue they’re battling right now was just solved by the customer-facing dev team. The intermediate developer who is incensed about a company-wide tech decision should be able to build their understanding of why it was made without booking a 1:1 with the responsible Principal Engineer. Your go-to-market team should be able to “see” through to the code, developers’ conversations, and the recent decisions around a given feature, letting them give customers correct and timely information about what to actually expect from the product today. And everybody in your company should, when it’s useful, have key company-wide strategy docs available to their agents as they make plans and decisions. And then, when a new revelation motivates the exec team to improve those docs, then bam. All the product engineers’ agents will take this new strategy into account right away. Anybody who’s worked at a large company and/or used CLAUDE.md knows this won’t be a silver bullet – deeply ingrained habits and momentum can not be simply prompted away. But as the tools and the data improve, the advantage will accumulate. When we launched a realtime meeting agent last month, we expected to get feedback about its defaults being too open – currently, Cedarloop defaults to sharing its collaborative notes and tools with all attendees live. But instead, we’ve seen two diverging kinds of feedback: many of our users want the tool to be less visible to external guests and customers, but more open internally within their companies. Which in retrospect makes a lot of sense: decisions and actions in your team’s work are increasingly useful across your company, but your customers shouldn’t need to worry about all that. So long story short, more internal transparency is coming. It will take some time. Apple isn’t doomed, and just because Zapier and Shopify are already working that way doesn’t mean they’re going to instantly be turbo-boosted. But it seems a new era is coming, where siloed knowledge, information hoarding, and secrecy-by-default will become less tenable. The firehose will evolve from a spicy distraction to a useful input to important work.

0 views
ava's blog 2 weeks ago

rose ▪ bud ▪ thorn - march 2026

Reply via email Published 31 Mar, 2026 I was featured as a Country Reporter on noyb's channels! My summaries made it into their newsletter 4 times this month. I reached Gold Status in my volunteering (20+ summarized and translated decisions for GDPRhub) now. Next up is the Magenta Status at 35+ :) I've written 4 exams this month; if I'd pass all of them, that's 30 ECTS! I think I'll pass 3. Switched away from Discord . I have no issue with being classified as a teen on the platform because it doesn't stop me from doing anything, but the move fit in with living my actual values like I do with other tech/media things (preferring open source, EU, etc.). I'm both on Matrix and Fluxer. Did some spring cleaning, like clearing out the fridge, wiping the inside, and rearranging the contents, together with throwing away expired toiletries, putting like 2 years of used batteries in the battery collection bin, decluttering a drawer, and vacuuming under and behind the sofa and bed. I've really felt like pouring extra energy into my looks lately. Got back into oil massages for my scalp, hair treatments, sheet masks, teeth bleaching, and got my nails done again (after going natural since December) and got a pedicure, too. I bought new dress pants that are so insanely comfortable, good looking and flattering, it's ridiculous! My yearly gyn checkup came back fine, and I finally caved and got proper treatment for my PCOS and endometriosis. I went out for some runs in the late evening :) haven't run outside in ages, I usually limit it to the treadmill. I went out to parks and forests , enjoying the weather and my free time after the exams. It was super healing and relaxing. Journaled more. Went to a vegan food fair. I applied to a job opening sent to me by a fellow blogger (James) and got an interview !!! I think I did well :) Upcoming: More decluttering and selling, tidying up the basement. Planning to go to two museum exhibitions soon before they close. Gonna go on vacation with two friends for 8 days next month! Booked tickets for an upcoming data protection event. Working on business cards (and maybe stickers?) for it. I've had some issues with my illnesses . :( The stress of intense studying most of February and March, weird weather changes, straining work stuff, eating a little too much sugar, the family situation, and starting two new medications this month sent my body over the edge. That made my fitness goals and studying a bit harder. I also unfortunately didn’t taper off a bigger dose of an anti-anxiety med I occasionally take as needed and accidentally caused agonizing withdrawal symptoms without realizing in time 🥴 I cut contact to last family member I was still talking to. It's stressful to withstand all the attempts to reach out to me, and to stick with the decision without guilt. My wardrobe is stressing me a little. I preferred not to own much. Unfortunately, the less you have, the more you wear the same things, the more they get washed and worn out. At some point, you want to replace a lot of it at the same time. That's not only financially hurtful, but also annoying when you have the goal to sew most of your clothes yourself, and you currently neither have the time nor the energy to buy fabric and sew the things you need. I am annoyed at walking into these fast fashion places, seeing nothing I like, then forcing myself to look at stuff more closely and everything is XS, feels like a trash bag, and costs too much for how flimsy and unethical it is. I'll have to try my luck with thrifting more, but even that has been overrun with Shein trash. If I make it to the second interview round, I might have to deny it. I like the company, they’re a great and respected employer, generous, and the interview was fun… but there are some dealbreakers for me, which hurts. I sat with it after, and slept over it now, and I just don’t think I’ll be happy in these circumstances. :( I wish it wasn’t so, because they were in the Top 3 of places I’d wanna work at, and I want a job in data protection badly. But it doesn’t feel right, and I can’t justify moving forward with it, all things considered. It feels like the wrong time for me. Maybe another open position in a couple years?

0 views

The Transposed Organization

One of the most common mistakes when building multi-agent systems is giving each sub-agent a specialized role: the “dev ops” agent, the “tester” agent, the “senior architect” agent. It mirrors how we organize humans but in my experience it doesn’t work well 1 . I’m starting to think the same thing is true for the humans who make up an organization. Most companies organize by what people know how to do but as AI pushes towards generalists (“Product Engineers”) and faster iteration cycles (“software factories”), I think they should instead organize by what problems people close. Cartoon via Nano Banana. This post proposes Company T (“company-transpose” 2 ), an operation that fundamentally flips how software companies organize talent and operate. Why I think this will happen, what will go wrong, and how an AI-native software company probably doesn’t even really look like a ‘software company’. The transposed organization organizes around what I’ll call “loops” rather than traditional specialities. A loop is the full chain of decisions between a problem and a deployed solution, owned by one person. It’s recurring, not one-shot, which is what makes it an org design unit rather than a project. Customer bugs → fixes in prod is a loop. Revenue pipeline → closed deals is a loop. The strongest loops close back to an external source (the customer who reported the bug gets the fix), but internal loops exist too. Here’s today’s org matrix. Rows are problems to solve, columns are specialist roles. Each cell is the person who handles that step. A single problem touches many people: A toy example of the org matrix, where rows are problems and columns are specialists. Traditional organizations organize around specialists. Read down any column: the same set of specialists handle that function across every problem. Bob writes the code, Carol does the design, Dave handles ops. A customer reports a bug and it touches five people across four handoffs. Frank, the person closest to the customer, has zero ability to fix the problem. Bob, the person who can fix it, has no direct relationship with the customer. Now transpose it. Same grid, but read across any row: the same generalist closes every step of that problem. A toy example of the “transposed” org matrix, where rows are loops and columns are specialties. Transposed organizations organize around loops. Bold means native skill; regular means AI-augmented. Alice was a PM; now she owns the full product loop, with agents handling the code, design, and ops she couldn’t do before. Eve was in sales; now she closes the full revenue loop. Frank was in support; now he diagnoses and fixes issues rather than just logging them. The loop owner’s role is closer to an architect than a traditional individual contributor. They’re not writing code the way an engineer writes code. They’re directing agent execution, making architectural and judgment calls across the full chain, deciding what to build and what “good” looks like before delivery. The primary output is decisions, not artifacts. Bob, Carol, and Dave didn’t disappear. Their roles shifted from direct execution to platform: encoding their specialist taste into the systems, guardrails, and context that every loop owner’s agents rely on. While this has always been the promise of platform teams, AI makes it more tractable because the encoding target is now a context layer that agents can interpret. It’s still hard, but it’s now the dedicated role rather than a side-effect. Bob doesn’t write the fix for Alice’s customer bug anymore, but he built the testing framework and code quality standards that Alice’s agents run against. Carol encoded the design system that ensures Eve’s agent-generated demos meet the brand bar. Dave built the deployment pipeline and observability stack that Frank’s agents use to ship and monitor fixes. Their expertise became infrastructure. They run their own loops now (platform reliability, design systems, etc.), and their output compounds across every other loop without requiring direct coordination with any of them. The relay chain between them collapsed into a shared context layer, made possible by agents that can absorb and apply specialist judgment at the point of execution rather than requiring the specialist to be in the chain. Today this works for some loops better than others. The gap between agent-assisted and specialist-quality output is real, and it narrows unevenly. A PM shipping production code via agents is not yet the same as a staff engineer shipping it. But the trajectory is directional, and the organizational question is whether you begin to restructure around where the capability is heading, wait until it arrives, or continue to assume things will stay where they are today. Thanks for reading Shrivu’s Substack! Subscribe for free to receive new posts and support my work. Every handoff in a relay chain has (at least) three costs: delay (waiting for the next person to pick it up), context loss (the next person only gets a summary of what the last person understood), and coordination overhead (syncs, updates, ticket management). These costs can be so deeply embedded in how companies operate that often we’ve stopped seeing them as costs. In a loop, there’s nobody to hand off to. The same person who heard the customer describe the problem is the one deploying the fix. There’s no “let me loop in engineering.” The context is native because the same human carried it the entire way. For external-facing loops, the output ships directly back to the source of the problem: customer reports a bug, customer gets a fix in prod. Prospect asks a question, prospect gets a live demo built for their use case. For internal loops, the closure is the same structure: system hits a scaling wall, the infra loop owner ships the fix; new hire joins, the people-development loop owner ramps them into a functioning loop owner. The topology is the same whether the source is a customer or the organization itself. This also changes accountability in ways that are hard to overstate. When six people touch a customer issue, responsibility diffuses. “X is working on it” is a phrase that defacto implies nobody specific is responsible for the outcome. When one person owns the full loop, they own the outcome. Innovation in a transposed org is encouraged and diffuses rapidly. It’s each loop owner getting better at their end-to-end vertical and being creative about how they close it. The product loop owner who finds a novel way to diagnose customer problems, the revenue loop owner who invents a new demo format that converts better: these emerge from depth within a loop, not from a separate R&D department. That doesn’t mean loop owners innovate in isolation. Shared skills and harnesses propagate ideas between loops through agents. Demo hours, brownbags, and on-the-loop peer review across loop owners propagate them between humans. The innovation surface is the combination both depth within a loop plus breadth across the shared context layer. The codified institutional knowledge (standards, conventions, judgment artifacts) compounds across every loop without requiring direct coordination between them. This doesn’t happen automatically. Skill and context governance, deciding which artifacts propagate, resolving conflicts between competing approaches, maintaining quality as the context layer grows, becomes a critical organizational capacity in an AI-native company. When you reorganize around loops, the company compresses. Not necessarily fewer people, but fewer coordination surfaces. Each loop operates with a degree of independence that looks a lot like a startup within a company 3 , shared infrastructure, shared values, shared context, but autonomous execution. The weekly meeting where twelve teams give status updates starts to feel unnecessary when each loop is self-contained enough to not need the other eleven to unblock it. This also changes what “management” means. In a function-based org, managers exist partly to coordinate across the relay chain, to make sure the handoff from Product to Engineering actually happens, that priorities align, that nothing falls through the cracks. In a loop-based org, that coordination job mostly evaporates. What remains is setting direction (which loops should exist, what problems they should target) and developing people (training the taste, agency, and judgment that make loops work). The claim is “one person closes a much wider loop,” not “one person does everything.” An Account Owner who closes the full revenue loop is not also the person who takes a customer bug to a fix in production. The loops are distinct; they share infrastructure and context, not responsibilities. The common binding constraint is cross-domain taste. A person can only hold good judgment across so many domains simultaneously. Every artifact on the loop that doesn’t meet the bar is slop, and there’s no ( in the loop ) downstream review to catch it. This is what determines how wide a loop can get: not execution capacity (agents handle that), but the breadth of judgment one person can maintain with quality. The pool of people who can close a full product loop, from customer problem to deployed fix, with taste at every step, is smaller than the pool of people who can do any one step well. Training taste is hard; you can teach someone a new framework in a week, but teaching them to recognize when an architecture decision will cause problems six months from now takes years of pattern-matching. When a loop is too wide for one person’s taste to cover, you pair loops. A revenue loop that requires both deep technical credibility and relationship selling splits into a technical loop co-owner and a commercial one working the same customer from complementary angles. This mirrors what happens in agent systems. A single agent with tools has zero internal coordination cost. But the moment you have two agents sharing a database or a file system, you need orchestration. You’ve reduced coordination from O(n) handoffs per problem to O(k) dependencies between loops, where k is much smaller than n. Two loops sharing a dependency (a demo environment, a roadmap, a brand promise, a production database) still create coordination costs between them. Between loops, you still need shared context, shared infrastructure, and someone who holds the picture of how the loops interact. That “someone” is probably the exec team or an uber-architect. Some steps in a loop are irreducibly human. Enterprise sales requires being in a room with a business leader. People management requires the kind of trust that doesn’t transfer through an API. These touchpoints are an additional hard floor on how much a loop can compress. A revenue loop can collapse AE, SE, and Deal Desk into one person, but that person still needs to show up to the dinner, still needs to have the relationship, still needs to read the room. The agent handles the demo, the quote, the technical deep-dive. It doesn’t handle the handshake. This also means that the loops with the most human touchpoints can be the ones that compress the least. Support for a self-serve product can compress dramatically. Enterprise sales to Fortune 500 accounts, less so. The transpose is not uniform across the matrix. A human can make a fixed number of high-quality decisions per day 4 . Agents handle execution, but every loop still requires judgment calls: what to prioritize, when to ship, whether the output meets the bar. The number of loops a person can own is also bounded by their decision capacity. And not all decisions are equal. Thirty deployment calls are different from five calls that each blend technical judgment, customer empathy, and business risk. The weight and variety of decisions matters as much as the count. A loop that requires constant high-stakes calls across multiple domains drains capacity faster than one with routine decisions in a familiar domain. This means you can’t just keep adding loops to a person until they’re doing everything. At some point, decision fatigue degrades the quality of every loop they touch. The right load is the number of loops where taste stays high, and that number is probably lower than most executives think. Even if you buy the model, most organizations are actively incentivizing against it. Performance frameworks often reward functional depth: you get promoted for being a better engineer, not for closing a wider loop. Compensation structures may assume specialization. Career ladders push for a specific function to climb. The latent motivation of what people think their job is, “I’m a designer,” “I’m in sales”, cements as identity-level, not just structural. Transposing the org requires transposing the reward functions. Performance need to measure loop ownership and outcomes, not functional output. Compensation needs to reward breadth of judgment, not depth of specialization. And the bar needs to move continuously: what counted as “full loop ownership” three months ago is the baseline today, because the agents keep getting better and the definition of what one person can drive execution on keeps expanding. Expectations ratchet every N months. Organizations that don’t explicitly reset this bar will find their people settling into a comfortable local maximum that’s already behind. If one person owns an end-to-end loop and they leave, you lose the entire capability. Specialist orgs have redundancy baked in: three engineers all know the billing system, so losing one is survivable. In a transposed org, the product loop owner carries the full context of their vertical. The mitigation is structural, and it introduces two roles that survive the transpose: loop managers and trainees. A loop manager owns the people-development loop: new hire → ramped loop owner is their recurring end-to-end responsibility. They set the targets for their loops (what problems to attack, what the bar looks like), develop the people running them, and step in when someone is out or ramping. They don’t coordinate handoffs, because there are no handoffs. They develop loop owners. Training someone into a loop is the harder problem. In a specialist org, onboarding is narrow: learn one tool, one codebase, one function. In a loop, the new person needs to develop judgment across the full chain. The ramp could look something like: shadow the current loop owner, run the loop with training wheels (the loop manager reviews output and flags where taste is off), then take full ownership as the manager steps back. The shared organizational context, the system guardrails, skill files, and encoded judgment that every loop inherits, means the trainee doesn’t start from zero. They start from the accumulated taste of the institution. The more the organization invests in making its context explicit and portable, the lower the risk on any individual loop and the faster new people ramp. If the transpose compresses loops and agents handle execution, why have a company at all? Why not a single “make money” loop run by one person with a swarm of agents? You can. And for many problems, a solo operator with agents will outperform a team. But companies still exist for the things that don’t fit inside a single loop: pooled trust, legal liability, shared infrastructure, the cross-org context layer that makes every loop better. A cluster of individuals with taste, sharing context and compounding each other’s judgment, outperforms the same individuals operating independently. The size of a transposed company is something like: Size = sum of all loops, where each loop is bounded by: The generalist bound. Can you hire someone capable of closing the full loop with taste? The wider the loop, the rarer the person. The human touchpoint floor. How many steps in the loop require a human talking to another human? These are the execution steps agents can’t absorb. The decision capacity ceiling. How many high-judgment calls does the loop require per day, and how heavy are they? Weight and variety matter as much as count. The volume threshold. A support loop for 10 customers and one for 10,000 are different loops entirely. At volume, loops split into sub-verticals, each still owned end-to-end. The authority surface. Can the person actually close the loop? Deploy access, customer access, spending authority. Without these, wider ownership is just more steps in the same relay chain. The shared surface. How many dependencies does this loop share with other loops? A production database, a brand promise, a deployment pipeline: each creates a coordination edge. More loops sharing more surfaces means more governance overhead, even with zero handoffs within any single loop. The company ends up feeling smaller than its pre-transpose version, not because it necessarily has fewer people, but because each person is more autonomous and the coordination overhead between them drops. The “startup within a startup” cliche becomes more structurally real rather than aspirationally fake. The hiring question follows directly from the formula. In the old model, you hire to fill columns: another engineer, another designer, another sales rep. In Company T , you hire to widen rows: people who can span more of the loop with judgment, not just execution. The value of a hire is roughly how many columns they can cover with taste, multiplied by how many loops they can carry at once. The profile that thrives in a transposed org is the person who has built things end-to-end before, someone who has shipped a product, closed a deal, debugged a system, and talked to the customer 5 . Builders, founders, people who’ve run their own thing. You’ll see companies increasingly marketing to exactly this profile , and it won’t just be for engineering roles. The revenue loop owner who can build their own demos, the support loop owner who can ship their own fixes. If you’re currently in a specialist role reading this, the question isn’t whether your function disappears; it’s whether you’re the person whose judgment becomes a loop, or the person whose execution becomes an agent. The path from specialist to loop owner is: widen. What infrastructure does Company T need? The loop only works if agents can actually provide the specialist capabilities at every step. That means the right agent tooling, the right context systems, and the right permissions model that gives loop owners the authority to actually close their loops. Just how generalist can a person be? The system creates incentives for maximal generalists, people who can close the widest possible loop with taste. But there’s presumably some bound on how many domains a person can hold enough judgment in simultaneously. Or maybe post-ai-neo-SaaS does just look like a group of micro-CEOs. Who owns organizational taste? A company of micro-CEOs each closing their own loops will develop their own judgment about what good looks like. Some of that divergence is healthy: the support loop owner knows what good support feels like better than the exec does. But the product a company sells still needs to be cohesive (at least I think so). The customer shouldn’t experience three different philosophies depending on which loop they touch. What gets decided at the loop level versus the org level, and how you maintain a shared thesis across independent loop owners without re-introducing the coordination overhead you just eliminated, is a core architectural challenge of the transposed org. Thanks for reading Shrivu’s Substack! Subscribe for free to receive new posts and support my work. I discuss this a bit in my more technical posts on Building Multi-Agent Systems . People building on AI consistently underestimate the cross-domain expertise of LLMs while underestimating the coordination and confusion cost of co-execution. This post in some ways is my realization that this is true for AI-augmented human organizational design as well. “But technically a transpose is …” Ok fine then I’ll use the wider, less linear algebra specific, definition: https://www.merriam-webster.com/dictionary/transpose This isn’t a new aspiration. “Pizza teams”, “squads”, “microenterprises” all promised autonomous units within a larger org. Most delivered partial results with significant coordination tradeoffs. What’s structurally different now is that agents collapse the execution gap that previously required those units to either stay small and limited or grow and re-specialize. The loop owner has access to specialist execution without needing specialist headcount. See Decision Fatigue . On top of generalization, I think adaptable is another key characteristic. Ideally what I call ‘ derivative thinkers ’. Cartoon via Nano Banana. This post proposes Company T (“company-transpose” 2 ), an operation that fundamentally flips how software companies organize talent and operate. Why I think this will happen, what will go wrong, and how an AI-native software company probably doesn’t even really look like a ‘software company’. Company T The transposed organization organizes around what I’ll call “loops” rather than traditional specialities. A loop is the full chain of decisions between a problem and a deployed solution, owned by one person. It’s recurring, not one-shot, which is what makes it an org design unit rather than a project. Customer bugs → fixes in prod is a loop. Revenue pipeline → closed deals is a loop. The strongest loops close back to an external source (the customer who reported the bug gets the fix), but internal loops exist too. Here’s today’s org matrix. Rows are problems to solve, columns are specialist roles. Each cell is the person who handles that step. A single problem touches many people: A toy example of the org matrix, where rows are problems and columns are specialists. Traditional organizations organize around specialists. Read down any column: the same set of specialists handle that function across every problem. Bob writes the code, Carol does the design, Dave handles ops. A customer reports a bug and it touches five people across four handoffs. Frank, the person closest to the customer, has zero ability to fix the problem. Bob, the person who can fix it, has no direct relationship with the customer. Now transpose it. Same grid, but read across any row: the same generalist closes every step of that problem. A toy example of the “transposed” org matrix, where rows are loops and columns are specialties. Transposed organizations organize around loops. Bold means native skill; regular means AI-augmented. Alice was a PM; now she owns the full product loop, with agents handling the code, design, and ops she couldn’t do before. Eve was in sales; now she closes the full revenue loop. Frank was in support; now he diagnoses and fixes issues rather than just logging them. The loop owner’s role is closer to an architect than a traditional individual contributor. They’re not writing code the way an engineer writes code. They’re directing agent execution, making architectural and judgment calls across the full chain, deciding what to build and what “good” looks like before delivery. The primary output is decisions, not artifacts. Bob, Carol, and Dave didn’t disappear. Their roles shifted from direct execution to platform: encoding their specialist taste into the systems, guardrails, and context that every loop owner’s agents rely on. While this has always been the promise of platform teams, AI makes it more tractable because the encoding target is now a context layer that agents can interpret. It’s still hard, but it’s now the dedicated role rather than a side-effect. Bob doesn’t write the fix for Alice’s customer bug anymore, but he built the testing framework and code quality standards that Alice’s agents run against. Carol encoded the design system that ensures Eve’s agent-generated demos meet the brand bar. Dave built the deployment pipeline and observability stack that Frank’s agents use to ship and monitor fixes. Their expertise became infrastructure. They run their own loops now (platform reliability, design systems, etc.), and their output compounds across every other loop without requiring direct coordination with any of them. The relay chain between them collapsed into a shared context layer, made possible by agents that can absorb and apply specialist judgment at the point of execution rather than requiring the specialist to be in the chain. Today this works for some loops better than others. The gap between agent-assisted and specialist-quality output is real, and it narrows unevenly. A PM shipping production code via agents is not yet the same as a staff engineer shipping it. But the trajectory is directional, and the organizational question is whether you begin to restructure around where the capability is heading, wait until it arrives, or continue to assume things will stay where they are today. Thanks for reading Shrivu’s Substack! Subscribe for free to receive new posts and support my work. With loops… The relay chain collapses Every handoff in a relay chain has (at least) three costs: delay (waiting for the next person to pick it up), context loss (the next person only gets a summary of what the last person understood), and coordination overhead (syncs, updates, ticket management). These costs can be so deeply embedded in how companies operate that often we’ve stopped seeing them as costs. In a loop, there’s nobody to hand off to. The same person who heard the customer describe the problem is the one deploying the fix. There’s no “let me loop in engineering.” The context is native because the same human carried it the entire way. For external-facing loops, the output ships directly back to the source of the problem: customer reports a bug, customer gets a fix in prod. Prospect asks a question, prospect gets a live demo built for their use case. For internal loops, the closure is the same structure: system hits a scaling wall, the infra loop owner ships the fix; new hire joins, the people-development loop owner ramps them into a functioning loop owner. The topology is the same whether the source is a customer or the organization itself. This also changes accountability in ways that are hard to overstate. When six people touch a customer issue, responsibility diffuses. “X is working on it” is a phrase that defacto implies nobody specific is responsible for the outcome. When one person owns the full loop, they own the outcome. The best is the default Innovation in a transposed org is encouraged and diffuses rapidly. It’s each loop owner getting better at their end-to-end vertical and being creative about how they close it. The product loop owner who finds a novel way to diagnose customer problems, the revenue loop owner who invents a new demo format that converts better: these emerge from depth within a loop, not from a separate R&D department. That doesn’t mean loop owners innovate in isolation. Shared skills and harnesses propagate ideas between loops through agents. Demo hours, brownbags, and on-the-loop peer review across loop owners propagate them between humans. The innovation surface is the combination both depth within a loop plus breadth across the shared context layer. The codified institutional knowledge (standards, conventions, judgment artifacts) compounds across every loop without requiring direct coordination between them. This doesn’t happen automatically. Skill and context governance, deciding which artifacts propagate, resolving conflicts between competing approaches, maintaining quality as the context layer grows, becomes a critical organizational capacity in an AI-native company. The company feels smaller When you reorganize around loops, the company compresses. Not necessarily fewer people, but fewer coordination surfaces. Each loop operates with a degree of independence that looks a lot like a startup within a company 3 , shared infrastructure, shared values, shared context, but autonomous execution. The weekly meeting where twelve teams give status updates starts to feel unnecessary when each loop is self-contained enough to not need the other eleven to unblock it. This also changes what “management” means. In a function-based org, managers exist partly to coordinate across the relay chain, to make sure the handoff from Product to Engineering actually happens, that priorities align, that nothing falls through the cracks. In a loop-based org, that coordination job mostly evaporates. What remains is setting direction (which loops should exist, what problems they should target) and developing people (training the taste, agency, and judgment that make loops work). The hard realities… Wider loops, not infinite loops The claim is “one person closes a much wider loop,” not “one person does everything.” An Account Owner who closes the full revenue loop is not also the person who takes a customer bug to a fix in production. The loops are distinct; they share infrastructure and context, not responsibilities. The common binding constraint is cross-domain taste. A person can only hold good judgment across so many domains simultaneously. Every artifact on the loop that doesn’t meet the bar is slop, and there’s no ( in the loop ) downstream review to catch it. This is what determines how wide a loop can get: not execution capacity (agents handle that), but the breadth of judgment one person can maintain with quality. The pool of people who can close a full product loop, from customer problem to deployed fix, with taste at every step, is smaller than the pool of people who can do any one step well. Training taste is hard; you can teach someone a new framework in a week, but teaching them to recognize when an architecture decision will cause problems six months from now takes years of pattern-matching. When a loop is too wide for one person’s taste to cover, you pair loops. A revenue loop that requires both deep technical credibility and relationship selling splits into a technical loop co-owner and a commercial one working the same customer from complementary angles. Coordination compresses, it doesn’t vanish This mirrors what happens in agent systems. A single agent with tools has zero internal coordination cost. But the moment you have two agents sharing a database or a file system, you need orchestration. You’ve reduced coordination from O(n) handoffs per problem to O(k) dependencies between loops, where k is much smaller than n. Two loops sharing a dependency (a demo environment, a roadmap, a brand promise, a production database) still create coordination costs between them. Between loops, you still need shared context, shared infrastructure, and someone who holds the picture of how the loops interact. That “someone” is probably the exec team or an uber-architect. Humans like working with humans Some steps in a loop are irreducibly human. Enterprise sales requires being in a room with a business leader. People management requires the kind of trust that doesn’t transfer through an API. These touchpoints are an additional hard floor on how much a loop can compress. A revenue loop can collapse AE, SE, and Deal Desk into one person, but that person still needs to show up to the dinner, still needs to have the relationship, still needs to read the room. The agent handles the demo, the quote, the technical deep-dive. It doesn’t handle the handshake. This also means that the loops with the most human touchpoints can be the ones that compress the least. Support for a self-serve product can compress dramatically. Enterprise sales to Fortune 500 accounts, less so. The transpose is not uniform across the matrix. Decisions per day are finite A human can make a fixed number of high-quality decisions per day 4 . Agents handle execution, but every loop still requires judgment calls: what to prioritize, when to ship, whether the output meets the bar. The number of loops a person can own is also bounded by their decision capacity. And not all decisions are equal. Thirty deployment calls are different from five calls that each blend technical judgment, customer empathy, and business risk. The weight and variety of decisions matters as much as the count. A loop that requires constant high-stakes calls across multiple domains drains capacity faster than one with routine decisions in a familiar domain. This means you can’t just keep adding loops to a person until they’re doing everything. At some point, decision fatigue degrades the quality of every loop they touch. The right load is the number of loops where taste stays high, and that number is probably lower than most executives think. The incentives aren’t set up for this Even if you buy the model, most organizations are actively incentivizing against it. Performance frameworks often reward functional depth: you get promoted for being a better engineer, not for closing a wider loop. Compensation structures may assume specialization. Career ladders push for a specific function to climb. The latent motivation of what people think their job is, “I’m a designer,” “I’m in sales”, cements as identity-level, not just structural. Transposing the org requires transposing the reward functions. Performance need to measure loop ownership and outcomes, not functional output. Compensation needs to reward breadth of judgment, not depth of specialization. And the bar needs to move continuously: what counted as “full loop ownership” three months ago is the baseline today, because the agents keep getting better and the definition of what one person can drive execution on keeps expanding. Expectations ratchet every N months. Organizations that don’t explicitly reset this bar will find their people settling into a comfortable local maximum that’s already behind. One person, one loop, one bus factor If one person owns an end-to-end loop and they leave, you lose the entire capability. Specialist orgs have redundancy baked in: three engineers all know the billing system, so losing one is survivable. In a transposed org, the product loop owner carries the full context of their vertical. The mitigation is structural, and it introduces two roles that survive the transpose: loop managers and trainees. A loop manager owns the people-development loop: new hire → ramped loop owner is their recurring end-to-end responsibility. They set the targets for their loops (what problems to attack, what the bar looks like), develop the people running them, and step in when someone is out or ramping. They don’t coordinate handoffs, because there are no handoffs. They develop loop owners. Training someone into a loop is the harder problem. In a specialist org, onboarding is narrow: learn one tool, one codebase, one function. In a loop, the new person needs to develop judgment across the full chain. The ramp could look something like: shadow the current loop owner, run the loop with training wheels (the loop manager reviews output and flags where taste is off), then take full ownership as the manager steps back. The shared organizational context, the system guardrails, skill files, and encoded judgment that every loop inherits, means the trainee doesn’t start from zero. They start from the accumulated taste of the institution. The more the organization invests in making its context explicit and portable, the lower the risk on any individual loop and the faster new people ramp. How big is Company T , and who works there? If the transpose compresses loops and agents handle execution, why have a company at all? Why not a single “make money” loop run by one person with a swarm of agents? You can. And for many problems, a solo operator with agents will outperform a team. But companies still exist for the things that don’t fit inside a single loop: pooled trust, legal liability, shared infrastructure, the cross-org context layer that makes every loop better. A cluster of individuals with taste, sharing context and compounding each other’s judgment, outperforms the same individuals operating independently. The size of a transposed company is something like: Size = sum of all loops, where each loop is bounded by: The generalist bound. Can you hire someone capable of closing the full loop with taste? The wider the loop, the rarer the person. The human touchpoint floor. How many steps in the loop require a human talking to another human? These are the execution steps agents can’t absorb. The decision capacity ceiling. How many high-judgment calls does the loop require per day, and how heavy are they? Weight and variety matter as much as count. The volume threshold. A support loop for 10 customers and one for 10,000 are different loops entirely. At volume, loops split into sub-verticals, each still owned end-to-end. The authority surface. Can the person actually close the loop? Deploy access, customer access, spending authority. Without these, wider ownership is just more steps in the same relay chain. The shared surface. How many dependencies does this loop share with other loops? A production database, a brand promise, a deployment pipeline: each creates a coordination edge. More loops sharing more surfaces means more governance overhead, even with zero handoffs within any single loop. What infrastructure does Company T need? The loop only works if agents can actually provide the specialist capabilities at every step. That means the right agent tooling, the right context systems, and the right permissions model that gives loop owners the authority to actually close their loops. Just how generalist can a person be? The system creates incentives for maximal generalists, people who can close the widest possible loop with taste. But there’s presumably some bound on how many domains a person can hold enough judgment in simultaneously. Or maybe post-ai-neo-SaaS does just look like a group of micro-CEOs. Who owns organizational taste? A company of micro-CEOs each closing their own loops will develop their own judgment about what good looks like. Some of that divergence is healthy: the support loop owner knows what good support feels like better than the exec does. But the product a company sells still needs to be cohesive (at least I think so). The customer shouldn’t experience three different philosophies depending on which loop they touch. What gets decided at the loop level versus the org level, and how you maintain a shared thesis across independent loop owners without re-introducing the coordination overhead you just eliminated, is a core architectural challenge of the transposed org.

0 views
ava's blog 2 weeks ago

using AI to inflate your ego

Personally, I’m open to retrying AI use cases every now and then. I’ve written about it before, and I freely share the fails and wins in chats I am part of. In my case, it’s no use to endorse it based on hype online, nor to try it once and keep my opinion fixed on that experience. I’m expected to engage with it somewhat at work, and independently, I want to know what I can expect from these tools so I can make better decisions and write better when it comes to the data protection impact these tools have. No use to stick my head in the sand when my desired career path is touched so heavily by it. What bewilders me is how many people seem to use the tool (and topic in general) to inflate their ego. I don’t just mean the literal sycophancy displayed in the model outputs, but also in the conversation around its use. There’s a group of people that are saying they do very important, difficult and smart work every day thanks to AI, in a pace and way humans just can’t. The gist of it is: “ I am better than you because I use AI, and have more productive output than you, and do more difficult work. The fact I need AI to do it means my work is very demanding, very admirable and at the bleeding edge, and humans could never do it like this, or have an output this fast. The fact that you don't want or need to use AI for your work must mean it's low-value. ” Often, they remain very vague about what that work even is, so it's hard from the outside to verify that. On the other hand, there are also people who do the inverse: They don’t plainly say that AI performed badly in their use cases, so it’s not useful for them, but instead, it becomes a way to prove that the work is so difficult and demanding that AI could just never do that. Something like: “ Behold, I am god’s gift to research and problem solving, and the machine cannot beat my perfect brain. The fact that you are able to use AI for your work must mean you are stupid and your work is easy, since AI can, at best, only do stupid and easy work. ” Both of these groups then make sweeping generalizations of what other people should do. The former group tends warn that “ you’ll get left behind! ”. It’s such a pathetic cope. It looks like people who were never the top at any skill in their environment, but now think they can finally have an edge thanks to adopting early and shitting out as much as possible in the quest to "learn" or hit some kind of jackpot, attract the right eyes. They have to cling to the fantasy that the nay-sayers will have a disadvantage somehow just so they can feel justified and special. But tell me, were you left behind when you started using Excel late? Was it bad you only learned office stuff when you needed it? Were you not able to catch up? Chances are, it makes no difference and with some effort and a workshop or YouTube videos, you can use the tools equally well. In the case of LLMs, using it has never been easier. You can just use plain, natural language! No submenus, settings, buttons, search operators and the like to remember. It‘s designed to be easy . Prompt engineering is and was always a scam. There are no secret incantations you only learn in a 500 Euro class. Anyone can use the tool, learn, and refine. It’s embarrassing to pretend otherwise. Coworkers that have trouble with basic Outlook and Word do surprisingly well with ChatGPT. And why wouldn't they? They have spoken a natural language all their life and have probably trained multiple other new employees in their career; they know how to explain standards and expectations, and how to explain tasks, to a human and a tool. The other group I mentioned is so weirdly dismissive based on their attempts at a very niche, still unstable use cases. I understand criticism that’s about directly advertised claims by the companies that aren’t fulfilled, or commonly seem use cases online that just don’t actually seem to reliably work; I wrote about the same thing in the past, and how the free models available are not capable enough to do many of the advertised things we are inundated with. What I don’t understand is thinking “ The LLM couldn’t generate a PDF with all my branding included and a table with this and that and accurate graphs and footnotes with sources. That means it’s not even useful to create an email draft, or for grandma’s grocery shopping list, and you shouldn’t use it for a motivational letter. ” Why can’t there be nuance? It obviously sucks bad for some complex stuff, but it really hits the corporate bullshit text creation just right. Don’t tell me I don’t get it - I recently tried out what it would recommend for a business card and it said I should use a transparent plastic card to signal transparency in my work. Of course I see how stupid it can be, even for some simple stuff. I get how it could royally screw up grandma's shopping list. But for me, both of the groups previously identified also ignore that most people simply aren’t in these high-stakes positions, interested in these hobbies or working these jobs. Many have no need or interest to vibecode some custom solution for their smart home or a family app that rewards homework time of the kids with gaming time automatically just to sell it to VCs or make a SaaS out of it, and they aren’t researchers or problem solvers coding complicated stuff or writing the next bleeding-edge paper in the field. They aren't hustlers scared of being outpaced by competition. Many people on this planet are taxi and bus drivers, nurses, kindergarteners, cleaners, cashiers, baristas, warehouse workers, construction workers, and the like. Or doing a boring secretary job that is about writing e-mails and sending out meeting details via buttons, using templates or pre-generated e-mails. They’re some boomers or part-time parents who aren’t that good with tech or don’t need much of it and pass office time clicking a couple buttons. What are you optimizing for, when you realistically only work like four hours of your eight hours a day and it’s the easiest work ever, just following protocol? They sure as hell aren’t interested in automating themselves out of a job, and they don’t wanna work anything else or do something more demanding. They wanna earn money with the least amount of effort and with the least amount of changing their workflow, and they don’t particularly care for computers or hustle. But if they can get out of some annoying text-based stuff like some e-mail aspect, maybe they’ll use it. And that's fine! They shouldn't be told by some AI fans that them not letting AI take over everything is making them a redundant NPC that has nothing to offer, or told by AI haters that doing easy work that AI can actually somewhat do means they're doing worthless work. The funny thing is: Their jobs often are just easy enough that it is faster and more foolproof to do it themselves than attempt a vibecoded or generated solution, while also having many use cases that work most reliably at this point and can actually be recommended. For example: Writing a short email thanking your boss for something is faster done by yourself than typing the prompt; but asking an LLM to make your angry email disagreeing with your superior sound nicer and more diplomatic works. My coworker can’t vibecode a solution to let AI enter text fields in the database automatically, but she can ask ChatGPT how to hide cells in Excel (nevermind that a search engine could also do this). I definitely am in that boat of “ no use, better done quickly myself ” with the core part of my job. So I just don’t understand why so many people need to brag that they’re moving the needle so much with their daily work either by using or not using AI, and subtly also shitting on people whose jobs are either replaceable with AI or aren’t fit for AI use, which I allege many fall into! It can’t be everyone that has such an unusual, high impact knowledge worker job where AI is either the magic enabler or not capable enough. I mean for fucks sake, seems like most of them posting the stuff are students, trainees, junior devs, or vague office job. It’s like people use this controversial topic to present themselves as less expendable and more important than they actually are. There’s also a group of people who won’t intellectually engage with the topic at all because they just do what everyone else does. Their personal podcast idols have about AI? Better give it all the data and put together a self-improvement plan and let it talk you through some journaling prompts. They don’t wanna discuss the bad sides because the people they admire love it. In my experience, they’re also very easily impressed with shoddy work just because it’s written in a charismatic way. “ This was groundbreaking ” and it’s something a Tumblr girl would have posted at age 14. All your friends hate AI? Better not touch it, out of fear of social repercussions. They can’t talk with you about the bullshit it did last time they tried it, or ethical, privacy, or environmental concerns, because they just never actually cared to develop an opinion aside from not wanting to be hated by their circle. That’s boring and people pleaser behavior. I think you look silly if you have no deeper reason to not use something, no interesting arguments. A tangent about arguments: I no longer care about whether images created by AI are good or bad and I don’t care about water or electricity usage. That is because the capabilities as well as resource usage can and will likely improve, and to me, are more representative of missing regulation and a shitty government than the tech itself; it’s better happening in a context more removed from the actual core of the tool and in how the industry needs to be regulated. I want to be more precise in what is actually the fault of the tool vs. the fault of the region many of these services are located in, and its political problems. If you claim to hate the tool, but only for the fact that it makes soulless images and it starts making better ones, what then? I'm sure you won't suddenly have no concerns! You usually hate the tech for other reasons than that, so we should focus on these better arguments instead. I think it’s much more interesting to debate whether it is art or not, about responsibility in war or accidents, or focus on the privacy aspect, the intellectual theft, the e-waste, job market effects and so on. Additionally, if we truly focus on electricity and water use (irrespective of regulation, placement, and other factors that cause issues of droughts and rising prices thanks to data centers), I think we would quickly have to argue against the terabytes of useless bullshit we all hurl onto the net to be stored for ages, take up space, and are another reason for more data centers and people’s increased use of their devices. Even your well-meaning blog post about enjoying a good sandwich counts, or your favorite cat video. I don’t want to discuss an intellectual bar or importance metric that online content has to clear before it can be uploaded because of our precious resources, because it would hit most of us, and it would hit art and marginalized voices. If we haven’t ever seriously discussed looking critically at each search engine use, each video we watch etc. as something potentially excessive that uses too much resources compared to how useful it was, I don’t know if this is the right way to start. I think for many, it’s only okay to start that conversation because it’s about something they don’t (yet?) use. It’s hypocritical, as many are not ready to give up their other online consumption behaviors for resource reasons either, because they don’t even cease them when mental health and privacy are harmed. 🤡 Lastly, there is some weird ego stuff going on about talking or not talking about AI. “You hate AI, yet you talk about it. Curious!” “Why do you wanna focus on something negative?” “The more you talk about it, the more you speak it into existence.” I don’t need to speak it into existence; billion-dollar industries funnel money into the bubble and force it into every device and software and ad. Don’t be disingenuous. Each industry, or art form (if you believe AI art is art) needs its critics. And as AI fans love to bring up, every new invention has had its moral panic, so if that also applies here, why are you mad? On the other side: Boohooo, you avoid the word "AI" to “ not give it more power ”; fine, have fun self-censoring for virtue signalling reasons, Mr. I-did-it-with-all-ten-fingers-and-a-few-braincells. I will keep writing about it, because everywhere I look, I just see people exploiting both ends of the spectrum for views and money, making extreme claims to get engagement. Who screams the loudest and makes the most absolute judgments is seen as more correct, after all. I will write about the AI Act, about labeling requirements, and more of the spectacular failures and okay-ish results I’ve had, though, and I will have to name the beast. And I don't care to read weird ego boosting shit swirling around elsewhere. Reply via email Published 28 Mar, 2026

0 views