We are a nonprofit raising policymaker awareness about AI risks, aiming to slow potentially dangerous research and accelerate technical safety. We have ambitious goals and need to move fast. Come work with us! About Palisade

Here are the kinds of projects we work on:

Palisade Research on Twitter / X

BadGPT-4o: stripping safety finetuning from GPT models

LLM Agent Honeypot: Monitoring AI Hacking Agents in the Wild

Badllama 3: removing safety finetuning from Llama 3 in minutes

FoxVox: one click to alter reality

Hacking CTFs with Plain Agents