If only AI safety research had a mechanism this clear. "We have proof that building the machine will kill everybody, so get to work making a provably safe version."
"AI safety" is essentially incoherent. It's like trying to build an all-purpose chemistry lab that can't produce explosives.
"AI safety" is essentially incoherent. It's like trying to build an all-purpose chemistry lab that can't produce explosives.