- cross-posted to:
- singularity@lemmit.online
- cross-posted to:
- singularity@lemmit.online
“One of these things is not like the other, one of these things is not the same”
Do love me some dry humor lol
I was hoping for a mention of Roko’s Basilisk.
Roko’s basilisk is, if I understood correctly, an intelligent super computer that would kill anyone that tried halting it’s creation, or am I understanding it wrong ?
Sort of, the first line of that wiki page summarizes it a bit more accurately though:
Roko’s basilisk is a thought experiment which states that an otherwise benevolent artificial superintelligence (AI) in the future would be incentivized to create a virtual reality simulation to torture anyone who knew of its potential existence but did not directly contribute to its advancement or development, in order to incentivize said advancement.
It’s punishment not just for being an adversary but for not being an ally either. Also, like The Game (you just lost) you participate just by being aware of the concept.
So I am technically safe if anything I ever wrote or made end up in it’s training dataset.
Hopefully! But maybe the AI thinks you could have done more. You could have influenced colleagues to contribute [more], could have donated money to others working towards it, could have had more children and guided their interests and education, deterred detractors etc. It’s interesting/scary to imagine just how petty it might decide to be in determining what actions or lack thereof constitute [enough] support.
AIs are ultimately a reflection of humanity. I’d rather accept the pain inflicted from an emotionless machine following erroneous alignment than knowing someone with emotions is perhaps enjoying inflicting it on me.
deleted by creator