Seems like LLMs are fundamentally flawed as production-worthy technologies if they, when given direct orders to not do something, do the thing