Could LLMs Learn to Detect Bias Autonomously, Like Tesla’s Self-Driving Cars?
LLMs are shaped by human curator making decisions about what data is “ethical” or “safe.” This curation introduces biases, focusing on politically salient issues like fairness for marginalized groups while often overlooking subtler distortions, such as selective outrage or perfectionist comparisons. What if LLMs could judge input themselves, learning from...