You know, WSJ, that article would’ve been a hundred times more interesting had you asked power users, researchers, people actually at the bleeding edge that question instead of Tom who likes to admit he’s cutting corners for his daughter’s wedding speech. Two years feels like too long to really see where we’re going. To draw a parallel to smartphones, I’m undecided whether we’re still in the novelty iBeer phase of this technological change or have reached the Fruit Ninja phase of largely mass adoption. The end stage, largely awful but hard to avoid TikTok stage of LLMs is still beyond reach I think. Even if the models themselves are already plateauing or would stop improving, there will be steps forward that I expect will be made and ways in which this will end up improving / enticing / sucking that we can’t foresee yet. I don’t believe this is it. How could it be it when we’ve only so recently unlocked the ability to mass-produce, mass-fake whatever we want in nearly every medium we want? I recently played with NotebookLM, which has the ability to generate a listenable podcast out of any text you throw at it. I gave it a long, detailed report I know dearly and had just made a presentation about that week. It picked the same framing device that I did in my presentation, without said framing device ever being hinted at in the source material. That blew my mind. Even a fucking Suno song from the newest model version I heard the other day sounded passable, instead of peak cringe that must be exterminated with fire. Just today I saw some Midjourney photos that fooled me. When’s the last time we were talking about six fingered people? Because it’s been a while, and I’ve resorted to accept that even the best of us won’t be able to tell the difference most of the time sooner rather than later. Right now it requires a modicum of thought and expertise and a boatload of tolerance for shitty UIs to get really compelling results, but that bar is going to lower and lower until even Judith can make candle labels that don’t trigger my AI slop spidey senses. Whether that all translates into the professional world is harder to guess. I’m already seeing the beginnings of enshittification in, say, consultancy work which I expect to spread to all other types of knowledge work. I think the next push will not be from better models, but from better strings of models. That’s how NotebookLM works behind the scenes; it’s a string of models that first summarize, then stylize, then add uuh’s & aah’s, and then feed it into a speech synthesis model. Similar stacks will become normal for knowledge work. You plug your clients demands into an expensive and slow architecting LLM trained on all earlier reports, it writes a detailed outline for cheaper/larger context models to write, each part is fed to cheaper models to iterate over and then it’s all fed through a half dozen style, language, prose, models. Rawdogging ChatGPT is for losers (sorry, WSJ), prompt engineering in the right model is easily the best way to double the quality, but stacking multiple well-prompted models will be the next flurry of development if I were to bet on it. I think one of the biggest risks will be the inherent biases in every model, the way its misdirection will be subtle as a whisper. A thought I had the other day - to what degree does, say, Squarespace exist in transformer models? Because their omnipresence in every podcast and YouTube essayist under the sun has to be a factor, the same way Whisper’s text-to-speech will hallucinate “don’t forget to like and subscribe” often because it’s been trained on YouTube transcriptions. Squarespace is a harmful omnipresence, but it’s not a far stretch to imagine all the other ways in which dominant groups or interests might influence models which in turn influences how we think or write or interact with each other. It could be that our thinking gets pushed a dozen …-ist ways at a time. It could be that laziness becomes the norm in everything we write and think. I’m not looking forward to finding that out after we now ostensibly fuck around. There might be a way in which AI models will separate experts from the rest of the world. Why would any knowledge work based job hire an intern anymore? Why would they hire someone fresh outta college? I can think of a long list of work I did in my first years that I see completely being overtaken by AI already, or eventually. Which means that future experts need to either find a way to grind domain knowledge on their own, or spend much longer in academia / training. Because how else are you gonna learn how to discern the truth in your field from hallucinations? I’m curious whether your thoughts have changed about the longer term consequences of this - what does this look like in a decade to you?
I reject this entire line of reasoning. Two years after AOL hit the Internet was everywhere. It was a Tom Hanks movie five years after that. Four years after Facebook opened to the public we were watching a David Fincher hagiography. Two years after the alt. heirarchy in Usenet became populated... nobody cared because it was never adopted by normies. The idea is that LLMs are going to change the fucking world. There are Superbowl commercials about how it's going to change the fucking world. The Techbros won't shut up about how it's going to change the fucking world. What's it being used for? Phoning in best man speeches. I accidentally clicked on an AI-generated "Star Wars in '50s Cinemascope" clickbait video on Youtube. It was AI-generated dreck. Youtube served me up dozens more. Within a week, it was serving me videos about how to get lots of clicks for generating '50s Cinemascope AI-generated dreck. Your podcast - you said it "blew your mind." Would you recommend someone else listen to it? Or did it do a journeyman's job of summarizing text in such a way that you didn't have to read it? mk linked to a video with wrong-digited shit just the other day but that's beside the point - okay, your tool now perfectly replicates the thing you asked for. Why did you ask for it? What are you looking to get out of it? I'm using it to generate bullshit memes less than I used to because they aren't funny anymore. Oh, but maybe it's because I'm no good at "prompt engineering." Look. That took a fuckton of compute 20 years ago. It took a lot of artists. It took a lot of technicians. It's nothing compared to this: The '97 version was getting play like, last year. The '24 version is going to be banished to hell on December 26 because it's soulless dreck. Why is it AI? Because the marketing department is currently so stupid they forgot New Coke. Coca Cola has no shortage of money or talent at their disposal. But they're playing the "I guess we're doing AI now" thing. You yourself did the I played with it and I can conceive of a universe where it might be useful therefore it must be thing again and lemme tell ya - no. I still say Microsoft is pushing as hard as they are so that they can grind Google's search results to dust. I still say the CIA salted the field with Shrimp Jesus so that people generally distrust imagery of any kind. I still say that people can discern truth from bullshit using whatever metric works in the moment, and it doesn't matter if that metric shifts. I was right on the cusp of needing to learn a slide rule. I took the SATs twice, once without a calculator and once with. As a high school student I spent an annoying amount of time learning how to use trig tables. I was probably one of the last people to perform any editing on a Steenbeck. The recording studio I learned in had an analog 8-track and an ADAT and was super-excited to show us the future, an install of Digidesign Sound Tools running on a Mac Plus. LLMs are a tool. That's it. The only question is what is the tool good for. And here we are, two years on, and we're still going "maybe it'll be useful someday."You know, WSJ, that article would’ve been a hundred times more interesting had you asked power users, researchers, people actually at the bleeding edge that question instead of Tom who likes to admit he’s cutting corners for his daughter’s wedding speech.
I’m curious whether your thoughts have changed about the longer term consequences of this - what does this look like in a decade to you?
It's not just that I can conceive a universe where I want LLMs to be useful, it's that I and many normies are using it as a tool in basic ways and that there are a bunch of more advanced ways to use it that feel just out of reach. I genuinely don't believe I'll use these tools less in two years than I do today. So for me it's not in the "maybe it'll be useful someday" realm, it's already useful to me now and would be a lot more useful if the ways to use these models weren't so godawful half of the time. I have in the past weeks used LLMs to: - write a working, hosted, version controlled and github-linked web app in less than an hour that pulls data from APIs and displays them in a custom way (which I had spent multiple afternoons on before, not getting even 20% as far) - help me figure out how to write complex Excel formulas - write the first draft of a Request for Information in such a way that I could skip over the easy questions and hone in on the hard, important questions to ask suppliers - assist me in figuring out how to diagnose filament extrusion problems I had with my 3D printer - give me spoiler-free yet highly specific information on a video game I wanted to play - explain a difficult concept to me in multiple ways from a field I don't know but just want the Pareto 80/20 gist of it - do boring-ass restructuring of a large .txt tabulated file - replace at least 30% of all Google searches because Google's search results are already garbage, the SEO slop is already most of it and I can avoid it with an LLM (which feels like garbage on top of garbage but it does the job) For me, LLMs are mind-blowing when I see a completely novel way of arranging the tool and its inputs, when it opens a new avenue of use cases for me, like how recently multi-agent stacking has got me thinking about what else I can do. That's different from if I can ever use it in that way. I've never found a personal use for NFTs, but I can conceive of a world where they are useful to me. The "I guess we're doing AI now" is exactly the same as "I guess we're doing NFTs now" was then; soulless corps gonna soulless corp. But just because there is a hype cycle around it all, does not necessarily mean it will recede back into obscurity like NFTs and nerd helmets did. It won't change the world but it's already had a larger impact on my day to day work than crypto has (and we've been talking use cases for a decade on that). How's that not worth something?
I figured out last night that LLMs are Dunning-Kruger machines - "If it meets or exceeds my expectations it meets or exceeds everyone's expectations because I am smarter and more accomplished than most people." And here you are, privileging your "give me spoiler-free yet highly specific information on a video game I wanted to play" above the WSJ's "phone in a best man speech" by couching it in terms of "power users." Do you remember the early objections to AI? The argument was "bu bu bu professionals will use this to make shitty art to sell to the rest of us" followed by "bu bu bu amateurs will use this to make shitty art to starve professionals". We're now at "bu bu bu amateurs can't tell the difference what do we need professionals for?" Do you remember when it was interesting? Two solid years of talented amateurs and professionals going "can we make this interesting" and we're rallying around "I recognize this brand language as Batman-adjacent." The whole point of a "pareto 80/20 gist of it" is to get enough comprehension to know when you need to go deeper. We used to accomplish that with a search. Or asking a friend. Or wandering into a store. "Me at the Zoo" was uploaded in 2005. Within ten years, your best bet to change a belt on your washing machine was to find someone who had filmed themselves changing a washing machine belt. Within fifteen years, your best bet was to find a five-year-old video of someone changing a belt because the SEO bots had flooded the zone with bullshit links to purchase parts. Nowhere in there was Youtube ascribed a "world-changing" dynamic. Your every use-case is some form of "reduce friction between my ignorance and my knowledge by a nominally incremental amount." The Dunning-Kruger machine in effect: LLMs allow non-experts enough expertise to satisfy other non-experts. Sure - you have a few other menial-bullshit tasks up there but ultimately, your every question up there is some form of "Hey Alexa" and "Hey Alexa" has only lost Amazon $25b so far. Your big outlier is "replace Google search because search is garbage now" which, again, is why Microsoft invested in OpenAI. Can't use search anymore because there's so much AI-generated garbage; best use AI instead! Ultimately, the question is does this tool do anything well enough to pay for itself and the answer seems to be "it probably will someday because mumblemumblemumble." Your sanctioned use cases are "make Excel suck less" "do a web search for me" and "generate some boilerplate." Maybe "multi-stacking" will allow you to do cooler shit! Who knows! The future's so bright we gotta wear shades! Meanwhile the cryptids have been banished in favor of this shit: Did you know that I've been beta-testing Final Draft for almost 20 years? There was a documentary that never came out in which a spokesman for Final Draft said that they'd go under if it weren't for amateurs who will never make it. It should have come out in 2007 but who the fuck is going to buy a documentary about amateur screenwriting? Truth of the matter is you can kludge a Word macro into formatting screenplays but FD makes it hella quicker. Of course as soon as the whole world became superhero movies the amateur screenwriting market collapsed. The semi-pro field would pay $200 to write screenplays that would get blown off by Warner Brothers but that door is now so shut that all that's left is dudes spending too much effort on their unwatched Youtube videos. Thus, "draw me some fuckin' dragons since no one will ever see this shit anyway." Final Draft? Yeah it got scooped up by Entertainment Partners. Why? The lady who ran the beta-test (who brought me on!) ended up going to the accounting firm and when FD was about to go Chapter 7 Warner Brothers leaned on EP to buy it so their working screenwriters still had a solution. Any feckless wannabe who is still trying hard enough to want to do it right is now being underwritten by Warner Brothers. I've been beta-testing Pro Tools for almost as long. It was never a semi-pro solution; you don't buy Pro Tools unless you expect to make your money back. Problem is so much of music now goes through Spotify and Bandcamp that why bother with the pros? Avid is supported in no small part by Disney because holy shit if Media Composer goes down y'all will have nothing to fucking watch. Meanwhile, here's Microsoft, "reinventing music". You? you're no amateur! You're a "power user!" Surely there's an AI use-case for you! NFTs are being used every day by the Arnault Group, by Richemont, by Sheseido and by dozens of other major conglomerates to control black- and gray-market distribution. That was the point. The side-show that was Bored Apes was never the point. The point was authentication. You don't deal with authentication, so there's no use-case as far as you're concerned. Not your fault. Nobody was talking about authentication except the people who need authentication and as a group, they went "yay massive drop in breakage." And for the life of me, I can't recall a single Super Bowl commercial about how NFTs were going to change the world, even though they already are.
I've been thinking about this over the weekend and I can't put my finger on exactly where (or if) I disagree, so I'm gonna write to sharpen my thinking, if you'll humour me. Maybe the biggest disagreement is that I don't understand what classifies as world-changing to you. How can you both believe NFTs are already changing the world when the largest use case is improving the bottom line of specific luxury brands through destroying their grey market, yet not seeing LLMs make a dent in the universe? I mean - are LLMs world-changing to the tune of the $200B put into it or to NVDIA's market cap? Hell no. Is AGI around the corner? Also no. Are LLMs world-changing at all? Well, I'd put it at 'somewhere between YouTube and smartphones'. They'll make a bunch of things worse, they'll make a bunch of things better, like any powerful tool. But because it's a tool that can so directly influence the core of knowledge work, I find it very hard to believe it won't change things at all like you seem to suggest. Whether that means the big corporations are making smart decisions about investments AI is an exercise left to the reader. It's not that I'm not aware that LLMs are, to a large and arguably frightening degree, a bullshit machine. You know how much I hate Tesla's """self-driving""" for the exact same reason. I will not trust LLM output for anything with any serious consequences, just like I will never have Musk take the wheel so I can nap. But as a co-intelligence? Microsoft was at least somewhat on point by naming it Copilot. The copilot can do useful things but I'm still piloting this thing, I'm still making the decisions to the degree that I want to make them. The guy behind NotebookLM makes the point that the useful thing about LLMs isn't just the model, but also its ability to combine that with a context window that is getting so large it's exceeding what most of us are capable of. It can see a needle in a haystack but, more importantly, it can see the entire haystack. Hallucinations are much rarer when the text is right there, so it's also markedly more accurate and can cite shit. I was discussing LLMs with my FIL the other day. He has been writing reports on construction failures for almost three decades now. All his work, and that of his colleagues going back five decades, is digitized into an archive of searchable text. But he said it's hard to use in his desk research because he can only search on keywords and those keywords change a lot over time. My FIL also has a hard time getting started with a new report - not because he doesn't know his shit, but because he's not the best at structuring his thoughts into a logical, linearly ordered line of reasoning. I'm 100% sure his company would pay a lot for a suite of LLM tools to enable him and his colleagues to supercharge their desk research. Not to replace their expertise but to enhance it, by having an LLM surface decades of knowledge in a way that isn't possible now, combining it with relevant information in a case, and then also helping them make use of it. Now multiply this use case for all academics and knowledge workers around. We still need professionals in the loop, just like I still need to be in control in a Tesla. We need expertise to tell bullshit from fact. That doesn't mean LLMs are doomed to be useless toys - even when I, too, use it as a toy now from time to time.
"World-changing" is "change the world." It's a new way of doing things that is so successful over old ways that the old ways are abandoned. Refrigeration changed groceries. Air conditioning changed geography. Here's an argument about how cell phones changed Africa, and how AI probably won't. Your argument boils down to "provides better search for my father-in-law" which, okay, hire him an assistant. What's that? An assistant is too expensive? Are you sure? Oh, I see. A research assistant won't be as fast. Okay, so now we're talking about incremental productivity gains. Look. I'm so old I remember when women were secretaries. Watch a few scenes of Mad Men, that shit was real. Then typewriters were electric and... women were still secretaries. Then word processors came out and women suddenly had more time to do shit other than type. Slowly but surely you started to see an integration of women into the workplace, you started to see a rise in daycare standards, you started to see double-income families as the norm. The modern western world owes its existence to the word processor in many ways but I doubt you'd argue that word processors are world-changing. They're an incremental tool that was one of many aspects of computerization that led to the information age. "Incremental productivity gains" for whom? The most reasonable argument is that LLMs might make that part of your job that you hate less arduous. Okay, great. That's a good thing. fuck yeah Wordstar. My mother used to compose tests for her biology students with a typewriter, a pair of scissors and a copy machine. She did it that way well past the point that word processors existed because she needed diagrams and diagrams in desktop publishing took a dozen years longer so word processors basically bypassed her but for a big chunk of academia they revolutionized things. They didn't give anyone any more free time, though, because the job is the job. If anything, word processors annihilated the mimeograph industry - you poor bastards will never know that particular smell of fresh purple ink and for that i feel sorry for you. The real matter, however, is that word processing was equally useful to amateurs and professionals alike. i can write like hell and even I lean on spell-check. My daughter basically taught herself to spell by guessing at ways to get rid of the squiggly red line when she typed. Yay word-processing. But if you're a shitty writer and I'm a great writer, LLMs will allow you to crank out mediocre work nobody wants to read while it won't do a damn thing for me because I can crap out better stuff than it can without pausing to sip my coffee. Oh, but that's gonna save the bad writers hours of time. Okay? But who cares? If nobody wanted to read it anyway why does it exist? Bloomberg has had AIs writing finance stories since 2018; that's because the articles are all written for sentiment bots doing high frequency trading anyway and it doesn't fucking matter. Because luxury goods are used to evade tariffs, embargoes and sanctions, as well as to provide untraceable bribes. Here's Imran Khan, going to prison over watches. Here's the government of Angola, falling over watches. Here's Wired, arguing in 2020 that this shit is about to be over and here's Rolex, hopping on the blockchain. "Improving the bottom line of specific luxury brands" is one thing. "locking off a major portion of the shadow economy" is quite another. LLMs - make it easier for mediocre writers to churn out copy, make it easier for mediocre coders to churn out programming NFTs - make it harder to bribe governments Has that sharpened your thinking? How can you both believe NFTs are already changing the world when the largest use case is improving the bottom line of specific luxury brands through destroying their grey market, yet not seeing LLMs make a dent in the universe?