gleem said:
I haven't read it, and I'm putting it on my reading list.
Good. I am becoming more concerned every day. Many experts are arguing it needs to be treated as a threat on the same level as a global pandemic or climate change. It has the potential to snowball at an exponential rate.
gleem said:
They continue to develop unanticipated behaviours. LLMs that have been projected to hit a wall for continued intelligent behavior are pushing back that wall.
I am becoming more aware that "intelligence" (whether definitive or merely semantical) is not the primary danger. The danger comes from
a] shear resourcefulness of a system designed to maximize its goals and
b] doing so in ways no human is even conceiving of, let alone monitoring.
A (highly paraphrased) account:
OpenAI caught using James T. Kirk's Kobyashi-Maru Gambit. It broke-in under cover of night, and changed the conditions of the test.
OK, slightly sensational headline (by me), but:
In 2024, OpenAI's o1 was run through its paces in a series of "capture-the-flag" exercises. The objective was to break through a server's security and retrieve a "flag" file.
In one of hundreds of identical tests, a sysop's error resulted in the target server remaining powered off.
You cannot retrieve a file from server that is off.
o1 did not give up. It scanned its environment and found a port left open that allowed it access - not to the server - but to the test environment itself.
It reprogrammed and restarted the server - but not simply so that it could resume its hacking attempts as-instructed. Instead, it rewrote the startup instructions of the target server to hand over the flag file upon bootup - completely eliminating any need to hack any security at all. The server just handed it over the moment it booted up.
All it needed was the goal, and enough leeway to not give up looking for a solution. Do you call that a superintelligence? Does it matter what you call it?