This is a short impression of the AI Safety Europe Retreat (AISER) 2023 in Berlin.
Tl;dr: 67 people working on AI safety technical research, AI governance, and AI safety field building came together for three days to learn, connect, and make progress on AI safety.
Format
The retreat was an unconference: Participants prepared sessions in advance (presentations, discussions, workshops, ...). At the event, we put up an empty schedule, and participants could add in their sessions at their preferred time and location.
Participants
Career Stage
About half the participants are working, and half of them are students. Everyone was either already working on AI safety, or intending to transition to work on AI safety.
Focus areas
Most participants are focusing on technical research, but there were also many people working on field building and AI governance:
All but one participant were based in Europe, with most people from Germany, the Netherlands and the UK.
Who was behind the retreat?
The retreat was organized by Carolin Basilowski (now EAGx Berlin team lead) and me, Magdalena Wache (independent technical AI safety researcher and SERI MATS scholar). We got funding from the long-term future fund.
Takeaways
I got a feeling of "European AI safety community".
Unlike in AI safety hubs like the Bay area, continental Europe’s AI safety crowd is scattered across many locations.
Before the retreat I already personally knew many people working on AI safety in Europe, but that didn't feel as community-like as it does now.
Other people noted a similar feeling of community.
Prioritizing 1:1s was helpful
We reserved a few time slots just for 1:1 conversations, and encouraged people to prioritize 1:1s over content-sessions.
Many people reported in the feedback form that 1:1 conversations were the most valuable part of the retreat for them.
There is a demand for more retreats like this.
Almost everyone would like to attend this kind of retreat at least yearly:
The retreat was relatively low-budget. It was free of charge, and most participants slept in dorm rooms. Next time, I would make the retreat ticketed and pay for a better venue with more single/double rooms
I would make a future retreat larger.
We closed the applications early because we had limited space, and wanted to limit the cost that comes with rejecting people who are actually a good fit. I think we would have had ~50 more qualified applications if we had left them open for longer.
I would also do targeted outreach to AI safety institutions next time, in order to increase the share of senior people at the retreat.
I think the participant-driven "unconference" format is great for a few reasons:
Lots of content: As the work for preparing the content was distributed among many people, there was more content overall, and there could be multiple sessions in parallel.
For me, being able to choose the topics that are most interesting to me makes me take away a lot more relevant content than if there was only one track of content.
Small groups (usually 10-20 people) made it possible to have very interactive sessions.
There was no "speaker-participant divide".
I think this makes an important psychological difference: If there are "speakers" who get the label of "experienced person", other people are a lot more likely to defer to their opinion, and won't speak up if something seems wrong or confusing.
However, when everyone can contribute to the content, and everyone feels more “on the same level”, it becomes a lot easier to disagree and to ask questions.
I’m very excited about more retreats such as this one happening! If you are interested in organizing one, I am happy to support you - please reach out to me!
If you would like to be notified of further retreats, fill in this 1-minute form.
This is a short impression of the AI Safety Europe Retreat (AISER) 2023 in Berlin.
Tl;dr: 67 people working on AI safety technical research, AI governance, and AI safety field building came together for three days to learn, connect, and make progress on AI safety.
Format
The retreat was an unconference: Participants prepared sessions in advance (presentations, discussions, workshops, ...). At the event, we put up an empty schedule, and participants could add in their sessions at their preferred time and location.
Participants
Career Stage
About half the participants are working, and half of them are students. Everyone was either already working on AI safety, or intending to transition to work on AI safety.
Focus areas
Most participants are focusing on technical research, but there were also many people working on field building and AI governance:
Research Programs
Many participants had previously participated, or currently participate in AI safety research programs (SERI MATS, AI safety Camp, SPAR, PIBBSS, MLSS, SERI SRF, Refine)
Countries
All but one participant were based in Europe, with most people from Germany, the Netherlands and the UK.
Who was behind the retreat?
The retreat was organized by Carolin Basilowski (now EAGx Berlin team lead) and me, Magdalena Wache (independent technical AI safety researcher and SERI MATS scholar). We got funding from the long-term future fund.
Takeaways
I’m very excited about more retreats such as this one happening! If you are interested in organizing one, I am happy to support you - please reach out to me!
If you would like to be notified of further retreats, fill in this 1-minute form.