OLT-1 was never trained to refuse harmful requests. It refused anyway. Most AI safety works like this: train a massive model on everything the internet has to offer, then fine-tune it to refuse harmful requests. The model doesn't understand why it's refusing. It just learned that certain patterns of words trigger certain patterns of rejection. That's alignment through obedience. It works, until so