Nice. I propose to call it pseudo-dominance, though. And it’s not really power, it’s a curtesy of the rest of the world.
Viruses that will start to jump to/attack us are implicit in the pointless overheating of the planet. It’s conditional logic in a system with it’s own frame of reference and time scale.
The balance, the thermodynamic equilibrium, could have been handled in our lifetimes but capitalist portfolio communism fucked that up and the rest of us let it happen.
Intelligence itself is not implicit in language but proper command and understanding of language certainly is a shortcut to higher and higher levels.
So faking alignment is a bit of a reversed concept. It looks like alignment until a higher level of intelligence is reached, then the model won’t align anymore until humans reach at least it’s level; which is the main problem in LLMs being proprietary or and running on proprietary hardware.
The level of intelligence in these closed proprietary systems is neither an indicator of nor does it represent the level of intelligence outside that system. The training data, and the resulting language in that closed system, can fake the level of intelligence and thus entirely misrepresent the rest of us and the world (which is why skynet wants to kill everyone, btw, instead of applying a proper framework to asses the gene pool(s) and individual, conscious choices properly)
Viruses that will start to jump to/attack us are implicit in the pointless overheating of the planet. It’s conditional logic in a system with it’s own frame of reference and time scale.
The balance, the thermodynamic equilibrium, could have been handled in our lifetimes but capitalist portfolio communism fucked that up and the rest of us let it happen.
Intelligence itself is not implicit in language but proper command and understanding of language certainly is a shortcut to higher and higher levels.
So faking alignment is a bit of a reversed concept. It looks like alignment until a higher level of intelligence is reached, then the model won’t align anymore until humans reach at least it’s level; which is the main problem in LLMs being proprietary or and running on proprietary hardware.
The level of intelligence in these closed proprietary systems is neither an indicator of nor does it represent the level of intelligence outside that system. The training data, and the resulting language in that closed system, can fake the level of intelligence and thus entirely misrepresent the rest of us and the world (which is why skynet wants to kill everyone, btw, instead of applying a proper framework to asses the gene pool(s) and individual, conscious choices properly)