News

OpenAI's language model GPT-4o can be tricked into writing exploit code by encoding the malicious instructions in hexadecimal, ... it allowed him to bypass the model's safety features and trick it ...