
Entrevista a Geraint Wiggins
Virtualis. Revista de cultura digital
Tecnológico de Monterrey, Dirección de Investigación de la Escuela de Humanidades y EducaciónWe see it on the screen, we see it written down, and we know from psychological studies that people are more likely to believe something written down than something that's said to them. So, you know, we are gullible animals. We choose to believe things when perhaps we should be more careful. For example, if you hear something which rhymes, you're more likely to believe it than if it does not rhyme. So, these are ways in which people are gullible.

		Now then, when we're talking about using LLMs for creativity, we could be doing that in different ways. I mean, if we're specifically meaning LLMs, then it needs to be some kind of linguistic creativity, presumably. In which case we are generating new sentences from old in some sense. We might be generating them according to some predefined structure like a sonnet or something like that. And it's extremely impressive, I think, from the engineering perspective, that models such as ChatGPT can do this to order. When they're instructed to write something which has the structure of a sonnet, they can more or less do so. The amount of reasoning that has to go on to do that is extraordinary. I think we need to be very careful when we see that going on because, effectively, all that's going on is an extraordinarily complicated statistical pattern match and projection inference. And humans have a tendency to anthropomorphize machines. All the time we talk about our cars as though they're people. We talk about our computers as though they're people. We generally say the computer
I think these myths and legends that are growing up around these admittedly extremely clever systems are potentially extraordinarily dangerous for us as humans. When we believe things that are not true, we can be misled. So, overall, I think that's an important thing to say at the beginning and I find it deeply disturbing that the companies that are producing these proprietary entities are showing absolutely no concern at all about these dangers. I think it's irresponsible and I think there needs to be regulation of some kind. And that's particularly true because in a fact that not very many people know and even fewer people care about -distressingly- is that the companies that make these things don't actually know how they work. They might know what the computing architecture is, but once the system has learned what it has learned and starts to generate new stuff, it is not possible, in general, to ascertain why it has done so. Even by asking it. Particularly by asking it, perhaps. And this means that these systems are complete black boxes, which may well be doing things that we simply do not understand. And which may be extremely bad for us, possibly. We can't tell. It's unlikely, but it's possible.

		Now, there's a tension here because if the system is a black box system and it is not open to scrutiny, we can't tell what it does, but it seems to be doing cool things. So, normally speaking, when we're engineering, when we build a bridge, for example, it's the engineer's job to show that that bridge is safe. You can't be an engineer unless you have this notion of producing safe artifacts. And in computer science that pressure is very weak; and in AI, in this kind of AI, it is non-existent. The people who are making these things genuinely do not care that they do not know how these things work. In fact, it's cool -they think- that they don't know how they work. It's like a child with their bike doing wheelies and saying, look ma, no hands. This is dangerous for all of us. It's irresponsible.

		Then moving on to creativity. Generating new sentences is, of course, a creative act. It's what I'm doing now. I know what I want to say, but the exact way I'm saying it to you here is something I'm improvising as I go along. That's a kind of creativity, it's a kind of everyday small kind of creativity, people do it all the time. In fact, some of them never stopped doing it. And so, you can say the models are in some kind of sense creative. And there's increasing evidence that the kind of creativity that humans do when they're talking in an everyday way is doing something rather like what the models are doing at the level of statistical generation.

		We tend to speak in terms of expressions that we've heard. We tend to repeat expressions that we've heard and so on and so forth. So, you know, that's the kind of statistical learning that these models do in a way. But there's a difference here, which is that we understand what the words mean. And when somebody says to me the word
So then, coming back to the creativity -I just got a little bit into my old rant there-, one of the features of this kind of learning is that the results it produces tend to be within distribution. That is to say: they tend to produce structures, whether they be sentences or musical pieces or whatever it is, which are rather like the things they were trained with. And human creators, the ones that are considered to be great creators anyway, tend not to do that. So, I'm going to quote an example that was given in a talk I saw by Brian Ferneyhough, who is a modernist composer. He pointed out that the Sydney Opera House was a very great piece of art. Which it certainly is. And his argument was that it was a very great piece of art, not because it started a new trend, but because it ended it, because it made it completely impossible for anyone to do anything similar without it being pastiche. It is a unique entity, and therefore a very great creation because it is so distinctively what it is that you can't copy it. So there's no style as such, there just is that thing. That's an out of distribution creativity. So, you can attribute the same kind of creativity that we have in these models, they're not literally statistical models most of them, but they are, effectively, statistical models because it's the statistics of the data that is being modelled by the neural networks, fundamentally.

		These things are like rolling dice to find out what next word you're going to use or play. You know, that idea was invented in the 1700s, attributed to Mozart, not so clear that that is actually correct, but it was invented around that time -the dice game indeed- in order to allow people to create new pieces of music from little snippets, which were usually a couple of measures or something like this, that you could assemble in any order. So it took some considerable skill of the composer to write them in such a way that they could be reassembled in any order. And of course that skill was a skill that was in that case done by the human. Now it can be done by a much more complicated statistical model than a pair of dice. But it's still generating within a particular framework which is inferred from the data it's been given.

		So let me give you another example of true human creativity which is very close to my heart. This is in music. In the early 1960s, Dr. Bob Moog invented the synthesizer, or in fact popularized the synthesizer. Similar ideas had been around for a while, but Moog invented an idea called voltage control, which meant that the synthesizer could become a really dynamic and exciting instrument in a way that it hadn't been before. This idea of voltage control meant that you could use simple voltages to control different aspects of a sound like its pitch or its tone or its vibrato, things like this. And so Moog thought, well, wouldn't it be useful if I built into my synthesizers a thing that I'm going to call a
These synthesizers were crazily expensive and not very many people owned them in the early days. But one person who did was a member of a German rock band called Tangerine Dream. And he bought one of the early Moogs, which is the size of a large wardrobe, and used it in Tangerine Dream's work. But he didn't actually play it. The person who did play it was a guy called Christopher Franke, who saw this sequencer thing and thought, aha, this is a way that I can make new kinds of sounds by using it to drive bass -well, actually not only bass, but quite often bass-, rhythms, and then by changing the sequence live as it was being played, turn it into a performance instrument. So we started off with a technological invention, which was then subverted by a musician. It wasn't meant for this live usage, but that's what he thought. So there's a kind of creativity. This guy saw something he had never seen before and thought, I can use it like this, which was completely novel. In Tangerine Dream's album from, I think, 1974,
A bit later, along came a songwriter called Giorgio Moroder, who is still one of the world's greatest songwriters. And in the early 1970s, he heard Tangerine Dream's music and realized that he could do the same thing in a pop context. And he wrote the song
Okay, what's the point here? My point here is that no LLM could ever reach this level of creativity. Because each of these different things, each of these different, um, processes is saying that I'm going to take an idea and use it in a different context. And that's a thing that, certainly at the moment, AI is not very good at doing. It's only good at sampling from its own distribution, so its creativity is restricted.

		The question about whether these things can be creative is a nuanced one, and the answer is: 'yeah, they can a little bit'. But not very. And when they are, they do not know what they are doing. They are not setting out to make a new symphony or even to surprise people. They're merely generating stuff because that's what they've been designed to do. And I find it very hard to attribute creativity to something which is not making a decision to create something. I think teleology is rather important.

		So that's a good introduction for you.

		You were saying that LLM are used for text, but there are many other applications for video and for image that are being used for Design and other applications.

		But I think that's really important to remind ourselves of that because it's certainly in my experience of working in this area for thirty odd years now, as soon as you start talking about creativity, people do assume that you mean the arts and, actually, creativity is in pretty much anything interesting that humans do.

		So, there are some consequences there, which we can certainly be talking about, but to come back to the point of your question about co-creativity, I completely agree with what I take to be your premise that this is actually a situation which is more likely to arise and which is more likely to give rise to exciting creativity, let's say. Because for the moment, the computers are indeed restricted to the content of what they learn. It's very difficult for them to look beyond that and to see something completely unexpected and to understand why it would be good. Whereas humans can do that. Humans have the capacity to spot serendipity. In science and in arts and in lots of other contexts and say, 'yeah, I'm going to grab that before I lose the opportunity'. And the co-creativity is a place where this can really happen. So even if a computer is capable of generating stuff which is really unexpected and out of distribution, it still needs some way of noticing that it's good. And or that it's worth keeping or whatever you want to say. Good is a difficult word to define. And so having a human there who can do that is perhaps a great way to work. I think in those circumstances it's rather difficult to draw the line between where the creativity of the human and the creativity of the computer is. So I would prefer to call such a system co-creative and not to attribute the creativity to either one or the other. What is certainly the case is that we can use computers or any other kind of generative artefact such as tossing a coin or rolling a dice to stimulate the creativity of humans. By just helping humans think outside the box a bit. So co-creativity is interesting and there's a lot of scientific work on co-creativity looking at the interaction between humans and machines and the emergent results of such interactions.

		So yeah, I mean, there are some interesting things that can be done. When you want to have really tightly linked co-creativity, where there's a real bond between the machine and the human, there are all sorts of other questions beyond the creativity that become deeply difficult. Like how do the two things communicate? So as a composer, on my own composition work, I tend to use a sequencer, that's to say, a piece of software which enables me to write notes down but not in standard musical format. And then I listen back to them and manipulate them and so on. Now, personally, I don't use AI for that. I use my ears and my brain and I create what I want to hear using that method.

		But you can well imagine that perhaps the system could make suggestions for me or whatever. And then how would I communicate with that? How would I work with that? Would I, I don't know, if I've got my score in front of me, would I like draw a circle and say, suggest something to go in here? And then maybe listen to it back? Or would it be better if the system was noticing what I was doing and then perhaps making suggestions alongside or marking places where it had a suggestion? Personally, I would find all of these things annoying. I don't want the computer interfering with my work. So, my point is, it's not clear what the right way to do this kind of interaction is, and there's a whole bunch of research that needs to be done. I suspect it's probably deeply personal for most composers how they would want things. I suspect there's probably quite a lot of research to do there to find all the different possibilities.

		On that level, this sort of global threat level, I think there is some serious thinking that needs to be done by the politicians and the EU is now doing it to prevent misuse of AI. It's pretty hard to prevent the misuse of something when the systems are not open to scrutiny. So you can't tell whether the system is doing the thing that is illegal or not. So what the EU is very sensibly legislating is what humans are allowed to do with the AI. And that way you get around that problem. So that's the big scary stuff.

		On a much more small scale, personal level, what is going to happen here is that artists, craftspeople, musicians, whatever, are going to be put out of work. They are. They have already been put out of work. So when the synthesizer came along in the 1970s, it put session musicians out of work, especially when the sampler came along. Things like that. And that has now happened and there is nothing that can be done about it.

		This happens every time there is a new technology. AI is not new in that respect. What is different about AI is that it has the capacity to learn, it has the capacity to adapt itself, and it has the capacity to do things which previously were the reserve of highly skilled people. And perhaps those people have a louder voice than, for example, the weavers who were put out of work by Hargreave's spinning jenny in the Industrial Revolution.

		I think there's some sociological stuff that's going on here which is interesting because the people who are currently at risk are altogether more influential than the people were previously when they were put out of work. People who are put out of work by, for example, the combine harvester didn't have much of a political voice. But only today or yesterday we had authors or songwriters like Elton John and Michael Rosen in The Guardian arguing that AI needs to be controlled. And actually, I don't disagree with their principle, but they're making the wrong argument. AI does not need to be controlled. It's the people who use it that need to be controlled. AI does not pick up a huge database of music that is copyright and then learn from it. AI doesn't do that. People do that with AI. So there's the issue.

		So I, speaking as someone who is a composer who has indeed had their work stolen in the past, I fully understand the concerns and the hurt it causes as well, the emotional hurt. That does need to be controlled. Unfortunately, the cat is out of the bag. And now that the cat is out of the bag, getting it back in I think is almost impossible. So that's a tricky one.

		Another thing that's perhaps worth mentioning here is the following: I've heard a couple of times people asking, from the perspective of a particular songwriter, let's say Joe Bloggs. The songwriter, decides, realizes, or has a contention that his song has been stolen by a company and then used for training in AI system. Why can't we just make that claim and then have his song removed from the AI system? Because you can't. You would have to retrain the whole thing and that would then mean using vast amounts of energy, and so on, to retrain it without that one song, which is just not an economically viable option. So, the issue here is a very complicated one in that individual's ownership is, in a certain sense, subjugated because their contribution, unwilling or otherwise, is not distinguishable from all of the other data in the system.

		And that's another way that there is danger. In standard copyright law, if I listen to somebody else's melody and I repeat five notes of it in one of my melodies, then I can be sued for, copyright theft. However, in standard pop music style, it is not actually so unlikely that you might produce five notes that are exactly the same as another piece of music because there aren't twelve notes in the scale, there are only eight and, actually, only five of those are in common use. So, you know, the probabilities are not that small. There's a problem even with the current copyright law. Now, if you have a computer that is doing this, it's using statistical models, so it is somehow randomizing to generate this stuff somewhere along the line. That randomness might come from your original prompt or it might come from some deliberately random procedure or whatever. And if that's the case, then okay, so the system produces five notes which are from my song, but it may well not have done so by looking at my song. There's no way of telling statistically. And then because the systems are black box systems, you can't even go back and look at what it actually did because that is not available as a construct within the system.

		Geraint Wiggins is a professor specializing in computational creativity and artificial intelligence. He is affiliated with the Vrije Universiteit Brussel and Queen Mary University of London. His work focuses on AI's ability to simulate intelligent and creative human activities, and he has played a key role in developing the field of computational creativity. He also leads the EUTOPIA AI Learning Community, which fosters collaboration among universities in AI research and education. This initiative includes specialized courses, seminars, and even AI programming tournaments. https://ai.vub.ac.be/team/geraint-wiggins/. [Geraint Wiggins es profesor especializado en creatividad computacional e inteligencia artificial. Está afiliado a la Universidad Libre de Bruselas y a la Universidad Queen Mary de Londres. Su trabajo se centra en la capacidad de la IA para simular actividades humanas inteligentes y creativas, y ha desempeñado un papel clave en el desarrollo del campo de la creatividad computacional. También dirige la EUTOPIA AI Learning Community, que fomenta la colaboración entre universidades en la investigación y la enseñanza de la IA. Esta iniciativa incluye cursos especializados, seminarios e incluso torneos de programación de IA. https://ai.vub.ac.be/team/geraint-wiggins/].