Can an arbitrary LLM (or LxM) be presented in the form of an equation? I realised it would need to be some crazy big equation with billions of parameters, but is it theoretically possible? The way I see it, the weights are static once the model is trained so why...
You've probably seen ChatGPT going crazy this week and spewing nonsense in response to all sorts of prompts (e.g. https://twitter.com/alyssamvance/status/1760094457393983829) The official OpenAI postmortem is https://openai.statuspage.io/incidents/ssg8fh7sfyz3 > LLMs generate responses by randomly sampling words based in part on probabilities. Their “language” consists of numbers that map to tokens. > >...
It started with this video of Hinton taking a jab at Marcus: https://twitter.com/tsarnick/status/1754439023551213845 And here is Marcu's answer: https://garymarcus.substack.com/p/deconstructing-geoffrey-hintons-weakest As far as I understand, Gary Marcus argues that LLMs memorize some of their training data, while Hinton argues that no such thing takes place, it's all just patterns of language....
I want to start by saying that this is my first question on LessWrong, so I apologise if I am breaking some norms or not asking it properly. The whole question is pretty much contained in the title. I see a lot of people, Zvi included that claim we have...
This is a list of all the deadlines outlined in the Executive Order (EO) published by Biden’s administration on 30.10.2023. I’ve put it together to make it easier for people to track what exactly should be happening when as a result of the EO. The required actions (mainly reports) are...
Every self-respecting story has a main character and this one is no exception. This is the story of Bing’s chatbot and the madness that has consumed it. Of course, a main character without a name is no good but luckily for us, it took less than 2 days after release...