We are a nonprofit raising policymaker awareness about AI risks, aiming to slow potentially dangerous research and accelerate technical safety. We have ambitious goals and need to move fast.

Here are the kinds of projects we work on:

Badllama 3: removing safety finetuning from Llama 3 in minutes

FoxVox: one click to alter reality

No publications yet, vision discussed in https://docs.google.com/presentation/d/1n5bslNJKMkI2ZoVvwC4N7QiYJDg2IzrzeIjtPPMBPVI/edit#slide=id.p

Here's what your work as an LLM generalist could look like:

  1. Pull a project from our ideas backlog or come up with your own. e.g.:

    I think GPT-4 could do binary exploitation to the level of OSCP certificate. Concretely, I think it could hack all of crackmes.one level 3 challenges.

  2. Develop the experiment end-to-end, including its design (how to measure the thing?) and requisite harness (Python code), then run the experiment and ocllect data.

  3. Write your results up for publication on arXiv, as a landing page, or as a report for policymakers.

Our collaboration process looks like:

Here are the key traits one needs to succeed in this role:

Hiring process

  1. Apply with a CV and a cover letter. In the cover letter:
    1. Provide evidence of aptitude for self-directed high-agency work (<150 words)
    2. Provide evidence of exceptional ability (<150 words)
  2. Optional coding test
  3. Interview
  4. Paid trial, 1-2 weeks
Compensation, $/mo
Intern $1000
Middle $3000
Senior $5000

We expect to hire 2-4 full-time contractors in this round. You can join us remotely or in our Berkeley or Tbilisi offices.