Large Language Models Fail on Trivial Alterations to Theory-of-Mind Tasks is a paper that I recently tried to read and tried to recreate its findings and succeeded. Whether or not LLMs have TOM feels directionally unanswerable, is this a consciousness level debate?
However, I followed up by asking questions prompted by the phrase "explain Sam's theory of mind" which got much more cohesive answers. It's not intuitive to me yet how much order can arise from prompts. Or where the order arises from? Opaque boxes indeed.
Also consider including non-ML researchers in the actual org building. Project management for example, or other administration folks. People who've got experience in ensuring organizations don't fail, ML researchers need to eat, pay their taxes etc.
I posit that we've imagined basically everything available with known physics, and extended into theoretical physics. We don't need to capitulate to the ineffable of a superintelligence, known + theoretical capabilities already suffice to absolutely dominate if managed by an extremely competent entity.
I agreed with the conclusions, now that you had brought up the point of the incomprehensibility of an advanced mind, FAI almost certainly will have plans that we deem as hostile and are to our benefit. Monkeys being vaccinated, seems like a reasonable analogy. I want us to move past the "we couldn't imagine their tech" to me a more reasonable "we couldn't imagine how they did their tech"
I find this thought pattern frustrating. That these AI's possess magic powers that are unimaginable. Even with our limited brains, we can imagine all the way past the current limits of physics and include things like potential worlds if the AI could manipulate space-time in ways we don't know how too.
I've seen people imagining computronium, and omni-universal computing clusters. Figuring out ways to generate negentropy, literally re-writing the laws of the universe, Bootstrapped Nano-factories, using the principle of non-locality to effect changes at...
the number of popular X in a human system Y:
Orbits just came to me, not sure if that counts a novel but I had never thought of them before as a stable equilibrium. They should stay the same unless perturbed by an outside force... bu
I have been thinking about the argument of the singularity in general. This proposition that an intellect sufficiently advanced can / will change the world by introducing technology that is literally beyond comprehension. I guess my question is this, is there some level of intelligence such that there are no possibilities that it can't imagine even if it can't actually go and do them.
Are humans past that mark? We can imagine things literally all the way past what is physically possible and or constrained to realistic energy levels.
I did encounter this problem (once) and I was experiencing resistance to going back even though I had a lot of success with the chat. I figured having a game plan for next time would be my solution.
This post and reading "why our kind cannot cooperate" kicked me off my ass to donate. Thanks Tuxedage for posting.
Fair critique. Despite the lack of clarity on my part the comments have more than satisfactorily answered the question about community norms here. I suppose the responders can thank g-factor for that :)
It does answer my question. Also thanks for suggestion to focus on the behaviour rather than the person. I didn't even realize I was thinking like that till you two pointed it out.
What are community norms here about sexism (and related passive aggressive "jokes" and comments about free speech) at the LW co-working chat? Is LW going for wheatons law or free speech and to what extent should I be attempting to make people who engage in such activities feel unwelcome or should I be at all?
I have hesitated to bring this up because I am aware its a mind-killer but I figured If facebook can contain a civil discussion about vaccines then LW should be able to talk about this?
There are no official community norms on the topic.
For my own part, I observe a small but significant number of people who seem to believe that LessWrong ought to be a community where it's acceptable to differentially characterize women negatively as long as we do so in the proper linguistic register (e.g, adopting an academic and objective-sounding tone, avoiding personal characterizations, staying cool and detached).
The people who believe this ought to be unacceptable are either less common or less visible about it. The majority is generally silent o...
Ideally, I'd want the people to feel that the behavior is unwelcome rather than that they themselves are unwelcome, but people are apt to have their preferred behaviors entangled with their sense of self, so the ideal might not be feasible. Still, it's probably worth giving a little thought to discouraging behaviors rather than getting rid of people.
I have committed to a food log with social back up, I am testing the hypothesis that to the first degree of approximation calories out > calories in = weight loss.
I have started to hard code a personal website using team tree-house (style sheets and two pages complete). I figure that the last comparative advantage we have before the machines take over is coding so why not test if I can do it.
so un retracting is not possible.
I have committed to a food log with social back up, I am testing the hypothesis that to the first degree of approximation calories out > calories in = weight loss.
I have started to hard code a personal website using team tree-house (style sheets and two pages complete). I figure that the last comparative advantage we have before the machines take over is coding so why not test if I can do it.
To what extent do you prefer the spreadsheet to have additional rows versus complete columns?
I love it! Next up, supporting wallet connect :)