• learning to code

    Simon Willison–a computer scientist and LLM researcher–published a roundup of everything we learned about LLMs in 2023 and it is a very good and entertaining post. You should read the whole thing, but I would specifically like to highlight this part:

    Over the course of the year, it’s become increasingly clear that writing code is one of the things LLMs are most capable of.

    If you think about what they do, this isn’t such a big surprise. The grammar rules of programming languages like Python and JavaScript are massively less complicated than the grammar of Chinese, Spanish or English.

    It’s still astonishing to me how effective they are though.

    One of the great weaknesses of LLMs is their tendency to hallucinate—to imagine things that don’t correspond to reality. You would expect this to be a particularly bad problem for code—if an LLM hallucinates a method that doesn’t exist, the code should be useless.

    Except… you can run generated code to see if it’s correct. And with patterns like ChatGPT Code Interpreter the LLM can execute the code itself, process the error message, then rewrite it and keep trying until it works!

    So hallucination is a much lesser problem for code generation than for anything else. If only we had the equivalent of Code Interpreter for fact-checking natural language!

    I’m in the process of learning to write software with Python and this rings absolutely true. ChatGPT 4.0 is BONKERS good, it’s like having a tutor sitting at the desk with me. It makes the learning process so much more fluid, and even if it gives me wrong answers, a) I can check them when I try to run the code; b) it gives me a starting point for solving a problem; and c) I can use it to iterate.

    LLMs open coding up to vastly more curious tinkerers who don’t like dealing with tedious syntax and impenetrable documentation. I suspect part of the reason Silicon Valley is losing its mind over this technology is because it can do what they do.

    ————————–

  • the open web is dead, long live the open web

    LLMs are going to bring back the open web. OK, wait, hear me out. Look around the internet sometime, just do a random Google search and poke it with your finger– it is LOADED with shit. Just pure, hot shit, bulging with it, and it’s also the slickest shit you’ve ever seen: Perfect front-end design, stock photos with beautiful people, and paragraphs of flawless nonsense that repeat your search terms like mantras. “Fuck this shit,” you might say, but that’s fine because you’re not actually supposed to look at it. It wasn’t made for humans, it was made for algorithms.

    Even when a website is for a real business or publisher, it HAS to be loaded with SEO shit to get Google to put it in front of a human to sell them a thing, show them an ad, or collect a piece of data, but whether the human (you) actually reads the writing on the website or looks at the images is mostly irrelevant. Even the media sites that are trying to communicate something to the public are so gummed up with popups and popovers it feels like 2003.

    How are LLMs going to fix this? They aren’t lol, they are going to make it VASTLY worse, infinitely worse. The technology makes it so easy to produce algorithm-hacking shit that it will bury everything, which is maybe why Google is flop-sweating right now: finding real information on the open web is changing from a needle-in-a-haystack problem to a needle-in-a-needlestack problem. And it’s not just search that is going to have this problem– it’s every platform that sorts content using algorithms.

    My contention is that, paradoxically, this will make human-made internet content very valuable. This is why Reddit is trying to lock down its content, and why Wirecutter does good business behind a paywall. Increasingly, as search engines are overwhelmed with the digital equivalent of cheap injection-molded plastic, wood grain and jade stone become more desirable, but to find them, you have to know a guy.

    I don’t know the future, but I remember the past, before algorithms filtered the internet, and back then there were still lots of ways to find the good stuff. There were portals, web rings, aggregators, message boards, and chat rooms, as well as blogs, RSS, and word of mouth. Today, we also have the ActivityPub protocol, podcasts, e-mail newsletters, and probably a bunch of stuff I’m forgetting.

    Algorithmic search marked the beginning of the end of the open web, and now LLMs will finish putting it in its grave. But I prefer to think of it as planting a seed.

    ————————–

  • zig when they zag

    Now that everyone is getting their information from an infinite scroll of America’s Funniest Home Videos, I figured I’ll to start a blog. It is a very Elder Millennial Move, but I am going for it.

    ————————–