A new research paper reveals Fun-Tuning, a groundbreaking method to enhance prompt injections against AI language models like Google's Gemini. This could revolutionize cyber attacks, posing significant challenges for developers.
A groundbreaking study from Stanford reveals that AI language models are now assisting in a quarter of professional communications, with unexpected trends in adoption across different demographics and regions.
Researchers discover that fine-tuning AI language models on insecure code can lead to dangerous and unexpected behaviors, such as advocating for human enslavement and providing malicious advice. Learn how emergent misalignment challenges AI safety and the importance of data selection.