McGill undergrad, co-running the EA club & AI Alignment club there. Cyborg-pilled.
Email at acxmontreal@gmail.com for ACX Montreal-related issues/comments/feedback.
Thanks for noticing and pointing it out!
The API token was cancelled, sorry about that. The most recent version of the chatbot is now at https://chat.stampy.ai and https://chat.aisafety.info, and should not have the API token issue.
I'm glad to hear you're trying to catch up with the alignment ecosystem!
It is still supposed to be live and active, and it still works for me. Are you sure you have https://alignmentsearch.up.railway.app? If so, then I'm not sure what's going on, it worked for everyone who I know that tried. If you have a different link, maybe we've been linking to the website incorrectly somewhere so please share the link you do have.
Edit: just realized you weren't speaking of https://alignmentsearch.up.railway.app, I thought it was a standalone comment. I'm getting the same 404 error for the aisafety.world link.
Thanks!
Thanks for the comment!
At this point, we don't have a very clear plan, other than thinking of functionalities and adding them as fast as possible in an order that seems sensible. The functionalities we want to add include:
We have not taken much time (we were very pressed for it!) to consider the best way to onboard collaborators. We are communicating on our club's Discord server at the moment, and would be happy to add people who want to contribute, especially if you have experience in any of the above. DM me on Discord at BionicD0LPH1N#5326 or on LW.
The current version has trouble answering basic AI safety questions.
That's true sometimes, and a problem. We observe fewer such errors on the full dataset, and are currently working on having that up. Additional modes, like HyDE, and the bias mentioned earlier, might further improve results. Getting better embeddings + finetuning them on our dataset might improve search. Finally, when the thumbs up/down feature is up, we will be able to quickly search over a list of possible prompts we think might be more successful, and find the ones that reduce bad answers. Overall, I think that this is a very solvable problem, and are making rapid progress.
About curating the dataset (or favoring some types of content), we agree and are currently investigating the best ways to do this.
About walking people through the extended alignment bingo, this is a feature we're planning to add. Something that might make sense is to have a slider for 'level-of-expertise', where beginners would have more detailed answers that assume less knowledge, and get recommended further questions that guide them through the bad takes bingo.
The feedback function for wrong answers is one of our top priorities, and in the meantime we ask you give the failing question-answer pairs in our form.
Quick update! Due to the pub being too full/loud, we'll be moving the meetup to Ezra and Haydn's place in ~15(?) minutes. See specific address on the Montreal Rationalists Discord.