I think where people misunderstand these LLMs is they think they have creativity. But they don't. They cannot think of something new. They can simply rehash. So it is useless for any actual creative work. Once you have the creative idea, it can generate a few paragraphs which sound OK but you need to carefully double check.
I think that better philosophers than us debated this question for thousands or years but I'll give my take. Suppose that creativity means to find a particular place in the n-dimensional space of all possible ideas. Artists find some interesting places in that space in part because they started from other interesting places nearby. They know those nearby places and we don't, because we trained to do a different job. However we have our share of creativity.
Algorithms like GPT or Stable Diffusion explore that n-dimensional space starting from possibly more places than a single human has experienced. They will miss many of the starting points that any human experiences, because of limits of those algorithms' interfaces with the world. However they can still find interesting places. Maybe it's not creativity as we mean it, but they can show us interesting places that no artist has found before.
I spent about 10 hours one day trying to convince my professional artist friend that AI could produce legitimate works of art, against his insistence that they could at best produce "kitsch" collections of art-like patterns, the kind of stuff that belongs on the wall at malls and not being discussed by academics in galleries.
I have to say that, although I still disagree with him, I was unable to get GPT to produce midjourney prompts which I would consider creative. Even when I prompted GPT to take the perspective of a known artist, it produced an endless series of banal landscapes, cliche compositions, overplayed metaphors, and insipid imagery.
Midjourney only really accels when a human is involved with the prompting.
I had to concede to him at the end of the discussion that /today's/ AI are not creative.
I firmly believe though that the inclusion of all mankind's perspectives possesses the raw material to generate creative ideas. My understanding is that current AI lack the ability to "turn off" the vanilla oversaturation of averages that pervade when you are exposed to all stimuli. One day, perhaps soon, there will be AI able to combine ideas they are exposed to without trying to combine ALL of the ideas at once.
how is that different from humans ? how common is truly original thought in humans ? to my knowledge humans work exactly the same, we "simply" rehash and mix existing ideas and repeat ending up with "novel" concepts. To common people these ideas may seem original, while to an expert it's just rehashed information.
original work often does not mean doing something already done in a slightly different way (paint a different set of sunflowers than Van Gogh)
What these models do is more similar to pastiche, if we really want to compare it to some technique.
But the value of the pastiche depends on the intent and its perception by the viewer(s), because there's nothing inherently original in it.
So basically what models produce has no value, unless we are able to attach some to it. [1]
If we asked chat-gpt to analyze something it has produced, it will probably say that it is "similar to" or "in resemblance of" but it's unlikely that it would say "this is the work of genius, how original! lovely!"
[1] edit: you don't read "pizza maker creates fabulous art on AI" but <person who's already in the business> won a contest of <some art form> submitting something created by <AI of your choosing>. Why? Because they know how to market it and can rely on other people believing that's their creation. Nobody says it upfront "I will submit an AI generated work", because they know it won't be judged the same way. The pizza maker was probably trying it for fun or to make a new logo for the pizza place or simply has no instrument to assign it a value and convince other people that it is true (including using their professional card).
It is very wrong to compare human creative process with a very elaborate text generator (AI).
If you write some simple code to generate art procedurally with randomness and it creates really beautiful pieces sometimes, was your software being creative?
I'd think if it had the ability to generate text as it does now, then iteratively apply the sorts of value judgments humans do on said output, and refine/discard/preserve accordingly, it wouldn't necessarily be so far from "genuine" creativity. Indeed maybe it does already do something like that.
I've played around a bit with getting it to write stories etc., and they do often seem quite "creative", the problem is eventually they often end up making little sense or contain fairly obvious contradictions or non-sequiturs in a way I wouldn't expect to see in the output of a typical human author (and certainly not a skilled writer). Indeed I'm not sure ChatGPT has the ability to formulate any sort of longer "story arc" within which to frame the text it generates. But I also suspect it will gradually be able to develop that capacity with future refinements.
This isn't true. You can ask it basic logic problems that it's never seen before, and it will apply the rules of logic to them. It can also identify correctly which rules of logic would make sense to apply in more complex situations, even when it doesn't get the answer right straight away. At doing this stuff GPT4 is better than GPT 3.5 which is better than previous GPTs. I fully expect that future models will be able to tackle more complex applications of logic to new domains successfully.
If you use only examples that weren't in its training set, you'll get to its limits quickly, but basic level first order logic is definitely within its ability.
It is a side effect of human language and a huge corpus of text it was trained on. I know that people sometimes use this kind of reasoning, words have attached meaning and then you can do some simple reasoning. I would say though, that it is not a real reasoning. It can do some of the things real reasoning can do, but the same as with people it can get attached to the meaning of words too much. When a person can't go beyond that they are called lazy or sloppy thinkers. It is a heuristic, not a real reasoning. I think it is a cheaper processing and it is of course useful, but it has its limits. For me it will mean that ChatGPT will not get to see some patterns that in creative works would be obvious for a creative human.
As you said, when things are not in its training set it can struggle. If there is a plausible looking text for the question I asked it will give it to me, that's how it is designed. For example I asked it about Windows command line debugger - CDB. It gave me an example command line for it: cdb -c "your-app" -o "logfile". It is very wrong. -c requires an argument which are debugging commands to run on start, -o is to attach to all created attached processes. Real command line looks something like this: cdb -logo "logfile" "your-app" (and it still does not exactly behave as you would imagine having experience with Unix CLI). The problem ChatGPT has with CDB is probably, because it has much much bigger corpus on Unix-like command line tools and because the documentation for CDB is abysmal. From this ChatGPT session I would have more examples.
I'm not saying it is useless. It just is not designed to do that. It might improve or there might be an another algorithm needed on top or instead of what it does use.
For me it is like a kind of a step up from a search engine. It helps me to find something to start with. When I get to some details it is often wrong. I get a starting point from it and then find a proper source for the rest.
It can reason but it has limited ability to transfer its reasoning skills into other domains.
For example, it can take code and add types to it. This involves a lot of reasoning ability. It can do this because it’s been trained on a vast amount of code. But it can’t yet fully transfer those reasoning abilities outside the narrow domain of code.
What is a domain? You talk of it like you know what it means. These things have no concept of a domain. If they use reasoning they use it all over the place. It is up to us to judge domainness.
Domain as in a context, the type of data that they were trained on. Code vs text. English vs Russian. Python vs APL.
For instance, OpenAI found that GPT4 is much better at reasoning in some human languages than others. It is best at reasoning in English, but struggles reasoning in less resourced languages.
There is clearly some context-independent reasoning going on (i.e. generalization) otherwise the model would not be able to reason at all in languages that it hasn’t seen a particular problem in. But there also appears to be a large context-dependent factor.
It’s a step towards something that can act like it reasons. I agree that many people perceive its capabilities inaccurately, but ChatGPT is an impressive machine. Just think how long it took us to evolve into being able to reason(billions of years). Here, we(humanity) went from first transistors to ChatGPT in, what, 100 years? The next 100 years might bring true AI.
In fiction it is often the case that an AI can reason better than humans do, but it doesn't understand emotions. But we now in general that emotions and reading of emotions is simpler than general problem solving. A child picks up on parent's emotions without extensive training. Animals can sense them. The fear response is a basic instinct. I would imagine it should be easier to make a machine being able to almost perfectly recognize emotions than general reasoning or this big heuristic machine which is GPT. I guess it all goes to the training set available.
just like humans. I'm not saying chatgpt can do everything humans can and I'm not saying it's AGI, but it certainly is overlapping with what the human brain cam do
Obviously we can say things like "What about a car that ran on pencils" but real, useful, entropy comes from inferences and abduction rather than the new.
The technology provides very interesting and impressive results, but to me it seems people are interpreting very wrongly its underlying process, trying hard to see emerging consciousness and whatever any time the result doesn’t manifestly show that it’s not having such a characteristic.
I tested GPT on its ability to handle word creation with suffixes. Some results are really interesting, but sometime it pretends a word is a sample of some suffix use when the word doesn’t encompass it. I’m rather confident that this kind of artifact can be overcome, I won’t slander the technology because it still has some obvious weakness. But I’m afraid that once overcome, the lake of snake will convince most people all the more that this impressive technological achievement is more than what it is actually.
I have positively had GPT-4 create genuinely creative answers. "More creative" is even listed as one of its traits in the ChatGPT model drop-down.
Ironically I think anyone who believes LLMs cannot be creative are themselves uncreative in their prompt crafting and use of the model, I've even seen this from senior developers and experienced writers. It's painful, like watching your mom try to find the downloads folder or enter terrible search engine keywords.
What do humans who exhibit "creativity" typically do but rehash themes and ideas from others? Trends rather slowly evolve as they reverberate back and forth between thousands of humans -- whether in music, painting, clothing, books, memes, ...
If an artist just thinks of something completely new not inspired by others it only appeals to a small minority. It is a minor portion of human creativity, and avoided by perhaps most humans.