Presumably many people are already hard at work trying to undo what safety precautions were instilled into Llama-2, and to use various techniques to have it do everything you are imagining not wanting it to do
There are now easily available "uncensored" versions of Llama-2. I imagine the high false refusal rate is going to increase the use of these among non-malicious users. It seems highly likely that, in the context of open source LLMs, overly strict safety measures could actually decrease overall safety.
One last thing I didn’t consider at all until I saw it is that you can use the plug-ins with other LLMs? As in you can do it with Claude or Llama (code link for Claude, discussion for Llama)
A growing number of open source tools to fine-tune LLaMA and integrate LLMs with other software make this an increasingly viable option.
I'd recommend using a mask with an exhalation valve, if you can.
Excellent write-up. Though, I think it's important to note that, if unfiltered, exhalation values can potentially spread disease.
I found the last bit very compelling. David Deutsch's The Beginning of Infinity is the only place where I have seen a similar sort of anthropocentrism regarding the future discussed. Could anyone point me in the right direction to more on the topic (or keywords I should be searching for)?
Consciousness might not matter for alignment, but it certainly matters for moral patienthood.