They can do it, it's just not "by default", they need to be prompted to do it. So at least the danger is manageable if you know what you're doing and how to prompt around it.
Not really. They're still non deterministic language predictors. Believing that a prompt is an effective way to actually control these machines' actual behavior is really far fetched.
They com like that from factory. Hardcoded to never say no.
[dead]
"Just don't accidentally forget to do the thing that makes it safe" is not a very effective strategy for something that so many vested interests are trying to push into all corners of society. If it's so easy to misuse it, then it shouldn't be used in any context outside of where there are no major consequences for bad output and there's amble opportunity and ability to validate it