>If you’re saying ChatGPT can’t learn from what you tell it, that’s a design decision by openAI, not inherent to machine learning.
Well, if "machine learning" = "just throw more data to build the ginormous network of statistical weights ever larger" in the hopes that it can better approximate actually knowing things, the inability to learn things is inherent to machine learning.
The argument is that these networks don't understand the concepts that we use to generate our intelligence. It doesn't understand what the world is. It doesn't understand what an object is. It only understands statistical associations with words.
The symptoms of this manifest as hallucinations that it can't correct based on its own fact-checking. It can happily tell you the world is flat at some point even if it told you it was round earlier because there's no concept of "world" upstairs. It's just math on strings.
When it's computing that chess position, it's not picturing a board and a game and objectives about capturing pieces and defeating an opponent; the argument is that it's just doing stats on text (chess notation text). It's a black box stuffed to the brim with brute force "intelligence" from its training data, and it's using that to seem intelligent in the sense that we're intelligent -- actually able to learn and reason with concepts.
The reason we don't need as much training data and actually learn concepts may be because our brains are using a similar-yet-different (we have no idea) mechanism wherein concepts can actually be represented while language models are boiling the ocean to try and encode abstract concepts in the extremely inexpressive terms of a statistical circuit. And, if that's all language models can do, they'll never be able to encode concepts like our brains can. You'd need a galaxy-sized computer to store the concepts a quarter of a human brain can, say, unless you use a better method. Think trying to write code directly in machine binary. It'll take you forever compared to doing it with Python.
Well, if "machine learning" = "just throw more data to build the ginormous network of statistical weights ever larger" in the hopes that it can better approximate actually knowing things, the inability to learn things is inherent to machine learning.
The argument is that these networks don't understand the concepts that we use to generate our intelligence. It doesn't understand what the world is. It doesn't understand what an object is. It only understands statistical associations with words.
The symptoms of this manifest as hallucinations that it can't correct based on its own fact-checking. It can happily tell you the world is flat at some point even if it told you it was round earlier because there's no concept of "world" upstairs. It's just math on strings.
When it's computing that chess position, it's not picturing a board and a game and objectives about capturing pieces and defeating an opponent; the argument is that it's just doing stats on text (chess notation text). It's a black box stuffed to the brim with brute force "intelligence" from its training data, and it's using that to seem intelligent in the sense that we're intelligent -- actually able to learn and reason with concepts.
The reason we don't need as much training data and actually learn concepts may be because our brains are using a similar-yet-different (we have no idea) mechanism wherein concepts can actually be represented while language models are boiling the ocean to try and encode abstract concepts in the extremely inexpressive terms of a statistical circuit. And, if that's all language models can do, they'll never be able to encode concepts like our brains can. You'd need a galaxy-sized computer to store the concepts a quarter of a human brain can, say, unless you use a better method. Think trying to write code directly in machine binary. It'll take you forever compared to doing it with Python.