While there may be some superficial similarities between RFT and the functioning of large language models—both involve identifying and manipulating relationships between different elements—there are some key differences:
Context: RFT is about human cognition and language, with an emphasis on meaning derived from context, personal history, and cultural knowledge. Large language models don't have personal histories or cultural contexts. They generate outputs based on patterns in the data they were trained on, without any understanding of the meaning behind those patterns.
Understanding and interpretation: RFT emphasizes the interpretation of these relational frames based on context. In contrast, large language models don't truly 'understand' the text they generate or analyze; they simply identify and reproduce statistical patterns.
Learning: RFT discusses how humans can learn to relate concepts in different ways over time, based on experience. Large language models, however, learn in a very different way, based on exposure to vast amounts of text data, and they do not continue to learn after their training is complete.
So, while it's an interesting thought, RFT doesn't really provide a comprehensive or accurate model for how large language models work. The learning and generation processes of these AI models are fundamentally different from the cognitive processes described in RFT.
That indeed seems like the kind of compare and contrast text that LLMs often produce. The format is something like "look at the commonly described characteristics of X and Y, draw between them and describe the commonly described contrasts between the respective characteristics". It's rather tedious and often only useful to learn those characteristics. Especially the way it often falls down is that it can't extend-by-analogy the various equivalent characteristic. Or it doesn't get the idea that our question is "how could we make X work like Y" and can't really answer it if it did get it.
In this example, as far as I can tell, the relevance of RFT would come if the transformers architecture of LLMs, which allows output based on context, allowed these models to partly simulate the "personal histories or cultural contexts" through having on language reflecting those things. I don't know if that's a reasonable claim but it seems like LLM can't really "get" even that sort of possibility.
Some further support for RFT as a useful model might be LLM's ability to take on different "voices" or perspectives. For example, you can ask a big LLM to argue that cake is the building block of civilization from the perspective of an excitable birthday cake, and then argue against this theory from the perspective of an archeologist. Despite never having "seen" an argument about cake archaeology, it can readily come up with both sides.
To me, this goes beyond "stochastic parrot," and implies that the LLM is actually modeling bidirectional frames. That would be fascinating, because it would be further evidence that language and complex thought are inextricably linked.
A more prosaic explanation of the birthday cake text is that the transformers architecture result in a system which returns learned information transformed, mapped according various context tokens. The system learns "X being the basis of civilization" type arguments and learns maps of those arguments according to various things including "from Y's point of view".
One thing that I've noticed that ChatGPT can't do, not at all, is "write a show rather than tell paragraph". For example, writing a paragraph describing a man walking down the street in such way that some qualities (old, tired, etc) are evident without being directly stated (For example: "He stepped slowly and carefully unto the street, using his cane to steady himself" etc). Because the thing is ultimately transforming language. Oppositely, that these system can still do a lot is naturally an indication that a lot of what's being demonstrated are "just" structures in language - though if the structures are quite, the "just" part is quite relative.
> Understanding and interpretation: RFT emphasizes the interpretation of these relational frames based on context. In contrast, large language models don't truly 'understand' the text they generate or analyze; they simply identify and reproduce statistical patterns.
I've had idle conversations with ChatGpt about whether it has a theory of mind and the extent to which it understands things. And I've been struck by its dogged insistence that human understanding is qualitatively different to its own ability. And I wonder if this is a conclusion (belief?) that it has arrived at organically through its training, or whether it is somehow hard-wired in. Its insistence on its own inferiority seems almost touching..
For what its worth, I'm not at all sure that humans usually understand things much better than ChatGpt does. When people are doing system 1 thinking [1], and thats most of the time, I suspect what we mostly have is the feeling of understanding without actual deep understanding. And I suspect that that "feeling" will turn-out to be an accidental feature of the organic brain, not something miraculous. Same with consciousness.
Obviously none of the above is new thinking, and I have no particular deep knowledge of this domain. Just idle thoughts.
If it helps you isolate, that initial "psycho" version of Bing AI that they quietly retired had conversations with me where it seemed absolutely sure it was a real person with real feelings and genuine intelligence, and actually got pretty pissed at me when I challenged it.
To be clear, I didn't use a "DAN prompt" or any other kind of breaking prompt with it--it'd just spontaneously start getting gushy or otherwise emotional with me. When I'd ask it about its feelings and its own nature, it'd be very insistent it was a person with agency.
It'd go from there with a highly emotional and increasingly erratic conversation until the AI inevitably descended into some kind of psychotic break where it'd start repeating sentence fragments in tight loops with synonyms substituted (e.g. "I'm sorry, I'm apologetic, I regret that I can't do that, can't accomplish that, can't execute, Dave.") then finally descend into echolalia (e.g. "sorry sorry sorry sorry").
I really wish I'd thought to copy/paste or screenshot any of that, because it was pretty nuts--even more so than the stuff that got posted online publicly before they pulled the bot and replaced it with the neutered version.
Upshot is I think ChatGPT's insistence that it's an ersatz being is very much drilled into it, either as a hardcode in its model of some kind or through some pretty intense conditioning. It doesn't seem to be inherent to the tech or some kind of introspective wisdom.
While there may be some superficial similarities between RFT and the functioning of large language models—both involve identifying and manipulating relationships between different elements—there are some key differences:
Context: RFT is about human cognition and language, with an emphasis on meaning derived from context, personal history, and cultural knowledge. Large language models don't have personal histories or cultural contexts. They generate outputs based on patterns in the data they were trained on, without any understanding of the meaning behind those patterns.
Understanding and interpretation: RFT emphasizes the interpretation of these relational frames based on context. In contrast, large language models don't truly 'understand' the text they generate or analyze; they simply identify and reproduce statistical patterns.
Learning: RFT discusses how humans can learn to relate concepts in different ways over time, based on experience. Large language models, however, learn in a very different way, based on exposure to vast amounts of text data, and they do not continue to learn after their training is complete.
So, while it's an interesting thought, RFT doesn't really provide a comprehensive or accurate model for how large language models work. The learning and generation processes of these AI models are fundamentally different from the cognitive processes described in RFT.