š¼ Making AI āusefulā is our biggest mistake
Did you know that we could have just been having fun with it instead?
Hello screen addicts, and welcome to my braincast.
Some predictions for a few of my readers who I know personally: Lisa, you are reading this on a coffee break. Alice you are opening this up on Monday morning. Jo you are not reading this at all because you have a tiny baby and I respect that.
Loads of people unsubscribed after I pointed out that Swifties represent the strongest social network we have at the moment in Elon Musk is just Taylor Swift for men. To me, this is a good thing ā I am filtering out the people who arenāt ready to hear the TRUTH. Anyway, subscribe for more stuff like this. Ta.
Iāve made an effort to use and play with generative AI tools a lot over the last couple of months; I think itās generally better when Iām critiquing something that I actually have experience with, perhaps you agree lol.
Something I canāt help noticing is how, when confronted with a system that is purported to be almost limitless in its capabilities, its limitations become glaringly obvious. Iāve realised that the main draw of a lot of generative AI tools ā like ChatGPT and Stable Diffusion and whatever else ā is not what they can produce, but that they can produce it with natural language.
But the problem here is, the way we interface with gen AI systems at the moment is mostly by awkwardly typing out a prompt. The way prompts are constructed are far from natural; in fact, writing prompts is something you have to learn. There are subreddits and listicles like this one which tell you how to write the best prompts to achieve a certain goal. The listicle I linked is about how to reliably produce pixel art; the prompts all go something like this: ā16-bit pixel art, cyberpunk cityscape, neon lights, and futuristic vehicles āar 16:9 ās 2500 āupbeta āq 3ā and, in some cases, donāt even produce something that looks like pixel art. So the language you use to interface with these machines is generally not natural, and you cannot guarantee good outputs.
And, as usual, another problem is that generative AI tools tend to be muddled and show-boaty ā which only adds to how underwhelming it all is. Even if Googleās Gemini demo wasnāt half-faked to look more impressive, and the machine could, in an instant, recognise a bad drawing of a duck on a post-it ā who cares? Is that supposed to beā¦ useful? Watching this ādemoā was like watching a parent praising their toddler for all the new things theyād learned at school that day. And then youāve got Grok, a text generator remade in Elon Muskās image (rude, obnoxious, unfunny) but it turned out to be too woke, reminding its redpilled incel userbase that trans women are in fact real women, and that no, it would not say a racial slur, even if it meant saving humanity from a nuke (which is pretty dumb). Fact: just as with other biases, the only way to scrub your training data of the woke mind virus is by scrubbing it out of humans first.
I think a lot of the disappointment also stems from the fact that generative AI systems represent the first piece of āuser-friendlyā technology which is infinitely difficult to control. Itās funny because the people who create these systems are often violently addicted to controlling things with machine-readable categorisations and fully auditable breadcrumbs. This reminds me of a very strange conversation I had about the game of Go with a tech guy the other day. He kept asking me if you could create an algorithm that could score a Go game accurately every time, and I said yes ā because of course you can. For some reason, he just couldnāt accept this; he continued to insist that āGo doesnāt actually have rulesā. Actually, it has exceedingly simple rules: you can place a piece anywhere on the board; it cannot be moved unless it is captured by your opponent, and then it is removed from the board entirely. Thatās it. The rest is just vibes. This man, clearly obsessed with reducing everything (even a fun game!) to a process, was unable to parse the fact that he just didnāt understand Go enough to create his pointless algorithm. Which is fine! The rules of Go are very very easy to grasp ā the gameplay and strategy is not.
Iām telling you about this weird interaction because this manās attitude perfectly exemplifies the tech bro mindset: āthere is just no way that this thing isnāt completely 100% knowable with the addition of machinesā. This attitude is incompatible with generative AI. The whole point is that you donāt actually know exactly how the neural network operates, and therefore you have no idea what youāre going to end up with when you generate an output. The unpredictability is whatās good about it; itās not a shortcoming that we need to iron out. If youāre trying to produce a new piece of writing with gen AI, and you know exactly what you want it to say, you may as well write it yourself. Getting AI to generate something exactly as you want it is impossible.
The creators of AI systems will conflate ālimitless capabilitiesā with randomness, and frustrated users will have to deal with that randomness when they thought they were getting something that could reliably automate boring tasks. The even dumber thing about this is that itās not even that random anymore. Since the launch of ChatGPT a year ago, OpenAI et al have constrained their models with significant guardrails so that itās much harder to produce harmful content, and easier to create predictable outputs. It also protects companies from reputational damage and lawsuits, but whatever.
Sorry, Iām going to have to talk about Go again.
, one of my favourite people on Substack, recently wrote a piece about optimising for the best outcomes in games. He talked about how AlphaGo (the AI which has beat professional human players at Go) will certainly win very easily, but it will also make moves that look really rubbish and boring to human commentators. Thatās because AlphaGo is making moves that will maximise its chance of winning, rather than its chance of completely obliterating its opponent. But humans are much more likely to optimise for spectacle, rather than just winning: āItās better to be winning by 50 points than to be winning by 1 point because this bigger margin protects us against the variance in our noisy, imperfect predictions about how the game will unfold.ā Lantz goes on to say that humans will opt to ācrush our enemiesā and in doing so, āare we overlooking the best way of maximizing the chance of getting the thing we want, because we have mistaken this barbaric proxy for the actual thing we want, like idiots?āThe way AlphaGo approaches a game of Go is similar to the ways in which generative AI systems insist we should be making content: blandly. The content restrictions and guardrails embedded within models are only theoretically good for harm reduction, and are definitely good for ensuring the production of flat, boring, samey content. It has been proven over and over again that if someone wants to create something offensive or controversial with DALL-E, they can if they spend long enough on it. They also could do it in photoshop if they wanted! This is not a new problem.
The problem is that weāve decided that AI tools are meant to be useful. This is completely wrong. If weāre just going to program a ālimitlessā machine to generate predictable outputs for work purposes (ew), then what are we doing? I thought the point was to have something that appears to express itself like a human, but does so in a hilarious and whimsical way; I really donāt want an infinite content machine to only ever give me the closest, safest, most underwhelming approximation of what I ask for ā I want it to give me something I never would have thought of. Otherwise itās just boring ffs.
Actually the most useful take on AI. From now on I will use it to maximize my enjoyment and minimize my productivity.