AI Safety
“Universal and Transferable Adversarial Attacks”: Researchers Jailbreak GPT
“Universal and Transferable Adversarial Attacks”: Researchers Jailbreak GPT Researchers with Carnegie Mellon’s Center for AI Safety have published a paper describing a method for developing adversarial attacks (a.k.a. jailbreaks) that … Read more
AI “Safety” and AI Ethics Are Not the Same
I have been agitating for some time now over the decreasing performance of AI models like GPT-4 or (especially) the new LLaMA 2 in direct relation to the “safeguards” being … Read more