Let's list the constraints:
- Basically, people know how to code but don't really know ML
- But at the same time, 3 days is enough time to learn a notion or understand and improve a template
- If we don't want it to be a hard to evaluate, the evaluation metric must be clear
- The topic must be related to AI safety, bonus if the topic is useful for ia safety
- The topic must be difficult, and must be able to keep strong people busy even for 3 days
- It must be possible to work on colab
- bonus if the number of ways to approach the topic is large so that participants do not all implement the same thing
- bonus if it's not too painful to organize
The goal of the hackathon is to find smart students who might be motivated to pursue in technical AI safety
I PayPal 50 euros for any idea I end up using in the hackathon
Do a literature survey for the latest techniques on detecting if a image/prose text/piece of code is computer-generated or human-generated. Apply it to a new medium (i.e. if it's an article about text, borrow techniques to apply it to images, or vice-versa).
Alternatively, take the opposite approach and show AI safety risks. Can you train a system that looks very accurate, but gives incorrect output on specific examples that you choose during training? Just as one idea, some companies use face recognition as a key part of their security system. Imagine a face recognition system that labels 50 "employees" that are images of faces you pull from the internet, including images of Jeff Bezos. Train that system to correctly classify all the images, but also label anyone wearing a Guy Fawkes mask as Jeff Bezos. Think about how you would audit something like this if a malicious employee handed you a new set of weights and you were put in charge of determining if they should be deployed or not.
Ah, I misinterpreted your question. I thought you were looking for ideas for your team that was participating in the hackation, not as the organizer of the hackation.
In my experience, most hackathons are judged qualitatively, so I wouldn't worry about ideas (mine or others') without a strong metric