Our new series about generative AI and its use in games suggests some words to use
If you’ve ever tried to play a hardcoreRPGthat’s way above your brain’s pay grade, or got lost in tutorials that use complicated words and bizarre jargon, then you’ve probably felt right at home reading headlines about AI recently. Why are people angry that this character has seven fingers?Why does Nvidia want me to talk to a robot about ramen? Why is everyone saying AI is smart when it still can’t manage its Classical Era luxury resource economy in Civilization properly? In this new series, we’re going to explore what ‘generative AI’ is, why it’s arrived now in the games industry, and what it might mean for people who make, write about and play games in the future.
Today, ‘generative AI’ most often means a machine learning system that has been trained to produce some kind of creative content, especially artistic outputs like art, music or writing. These AI systems can usually be ‘prompted’ by writing a request in plain language, which is where we get famous systems such asMidjourney, which you can get images from simply by describing what you want with words. Using AI to create art or music isn’t new. In fact, it stretches back decades, long before my PhD. But it’s shot into the popular consciousness now that machine learning models have provided simple interfaces and higher-fidelity results. New AI results went viral from time to time on social media, thenthey popped up aslittle demos you could try yourself, and before long it began to feel like a product. Then, suddenly, itwasa product.
It might seem like generative AI is already taking over the games industry. Earlier this year, GDC released their annualState of the Industry survey, in which they claimed that 31% of surveyed developers used generative AI in their workplaces, and 49% of studios were using it. But ‘generative AI’ can describe anything from people using AI image generators to create every single art asset in their game, down to people who just useChatGPTto write their emails. Generative AI is not a single concept, and using it for one area – such as programming – might feel and work very differently to using it in another – translating dialogue, say. Regardless of any studies, headlines or big press releases, generative AI is a big, messy, controversial idea, and it won’t affect all parts of the games industry in the same way.
To help you think about some of these differences, I’ve got some suggestions for new words we can use to talk about generative AI systems. The first is ‘online’ versus ‘offline’ systems (which I’m borrowing fromresearch on procedural generation). Online systems generate content while you’re playing the game –AI Dungeonis an example of an online generative AI system, because it writes in real-time while you’re playing. Offline systems are more for use during development, like the use of generated AI portraits in the indie detective gameThe Roottrees Are Dead. Portraits and other artwork were generated using Midjourney and added to the game, but the game itself doesn’t generate anything. As you’d expect, online systems are a lot riskier because developers can’t test every possibility in advance, but they can also lead to more exciting and innovative game designs. Offline systems are easier to test, secure and validate, which might make them more popular with big studios who can’t afford to take risks with unpredictable technology in live games.
Another way we can categorise generative AI systems is between “visible” and “invisible” systems. Visible systems produce content that you directly feel the effect of – things like art or music - while invisible systems generate content that the average player might not be as aware of. For example, some programmers useGitHub Copilot, a generative AI system that can write small sections of program code. If someone used Copilot to write the multiplayer networking code for your favouriteMMORPG, you almost certainly would never hear about it (unless something went wrong). The same goes for many aspects of game development that we don’t necessarily see directly – if the finance department uses ChatGPT to compile their monthly reports, for example, or if concept artists produce some pre-alpha artwork usingDALL-E. The visibility of a generative AI system may be increasingly important as backlash against the use of AI tools rises, because developers may feel safer employing generative AI in less visible ways that players are less likely to feel the presence of.
The third category, and maybe the most important one, is whether the AI is “heavy” or “light” – thanks to my colleague and studentYounès Rabiifor suggesting the names for this one. Lots of the most famous generative AI tools, like ChatGPT or Midjourney, have been trained on billions of images or documents that were scraped from all across the Internet; they’re what I callheavy. Not only is thislegally murky– something we’ll come back to in the next part of this series – but it also makes the models much harder to predict. Recently it’s come to light that some of these models have a lot ofillegal and disturbing materialin their training data, which isn’t something that publishers necessarily want generating artwork in their next big blockbuster game. But lighter AI can also be built and trained on smaller collections of data that have been gathered and processed by hand. This can still produce great results, especially for really specialised tasks inside a single game.
But that kind of hindsight is still a long way off for generative AI, and in the meantime there’s a storm of court cases, hopeful start-ups and confused game design pitches for us to weather. Tomorrow, we’ll sail headfirst into the choppy waters of ethics and the law. Until then, keep counting those fingers, and don’t talk to any strange robots.
In Part 2, The Art Of The Steal, we discuss whyprotecting art from generative AIhas never been more vital or important.