Github: Kwaai Alignment GitHub
Our current goal this year is to produce a survey paper in alignment research and run small experiments.
Right now, we’re working on the survey and small experiments. If you’d like to help out with either, reach out and we can get you started.
That's okay! We welcome all different backgrounds. If you are interested in this area and willing to help, then that is great.
Contributions: I want to run a small experiment I read in a paper, can I add my code and findings somewhere?
Yes, create a new branch, add your changes using the below format to /Implementations
, and open a PR.
Please create a folder named {experiment_name}
inside the /Implementations
directory. The structure should look like this:
/Implementations
│
└───{/experiment_name}
│ │ README.md
│ │ relevant_paper_1.pdf
│ │ relevant_paper_1.pdf
│ │
│ └───code
│ │ script_1.py
│ │ script_2.py
│ │ ...
Feel free to add more specific instructions or scripts based on the particular experiment and its requirements.
- Research Priorities for Robust and Beneficial AI: Read here
- Concrete Problems in AI Safety (2016): Read here
- Unsolved Problems in ML Safety (2022): Read here