The Core Idea
What if we could test whether language models truly understand meaning, rather than just matching patterns? Here's a simple thought experiment:
- Create two artificial languages (A and B) that bijectively map to the same set of basic concepts R'
- Ensure these languages are designed independently (no parallel texts)
- Test if an LLM can translate between them without ever seeing translations
If successful, this would suggest the model has learned to understand the underlying meanings, not just statistical patterns between languages. Theoretically, if Language A and Language B each form true mappings (MA and MB) to the same concept space R', then the model should be able to perform translation through the composition MA·MB^(-1), effectively... (read 332 more words →)