All of CstineSublime's Comments + Replies

Off the top of my head it's because people are weary of Chesterton's Fence/Sealioning (feigning 'just asking questions' when actually they have an agenda which they mask with the plausible deniability of feigning naive curiosity) and as you say - the topic being sensitive so it generates a 'ugh field' are two pillars of what makes certain topics difficult to discuss.

I've noticed this pattern on a lot of, usually political topics but it could also be some kind of interpersonal drama/gossip, someone asks a you question which appears to be an invitation to ge... (read more)

That Nixon one really wow'd me, the fact that it exaggerated his jowls but after a bit of google searching it seems like other models also seem to have been trained on the Nixon caricature rather than the man himself.

I'm also a big fan of that Fleischer style distracted boyfriend remix.

Never the less, the ease of 'prompting' if that's what you can even call it now is phenomenal.

I'm looking at this not from a CompSci point of view by a rhetoric point of view: Isn't it much easier to make tenuous or even flat out wrong links between Climate Change and highly publicized Natural Disaster events that have lot's of dramatic, visceral footage than it is to ascribe danger to a machine that hasn't been invented yet, that we don't know the nature or inclinations of?

I don't know about nowadays but for me the two main pop-culture touchstones for me for "evil AI" are Skynet in Terminator, or HAL 9000 in 2001: A Space Odyssey (and by inversion - the Butlerian Jihad in Dune). Wouldn't it be more expedient to leverage those? (Expedient - I didn't say accurate)

I want to let you know I've been reflecting on the reactive/proactive news consumption all week. It has really brought into focus a lot of my futile habits not just over news consumption, but idle reading and social media scrolling in general.[1] Why do I do it? I'm always hoping for that one piece of information, that "one simple trick" which will improve my decision making models, will solve my productivity problems, give me the tools to let me accomplish goals XY&Z. Which of course begs the question of why am I always operating on this abstract... (read more)

I think you're right. Although I'm having a hard time expressing where to draw the line between a simile and a analogy even after glancing at this article; https://www.grammarpalette.com/analogy-vs-simile-dont-be-confused/ 

3localdeity
Whoever wrote that article is confused, since in the table in the section labeled "Analogy vs Simile: The Differences" they have several entries the wrong way around (compared to the two paragraphs preceding it). It seems to me that you could use the same comparison for either an analogy or a simile.  An analogy would usually be in the present tense, "X is like Y", and followed by more explanation of the concept the analogy is meant to illustrate.  A simile would more frequently be in the past tense as part of a narrative, and more frequently use other verbs than "is"—"X moved like a Y"—and probably wouldn't extend beyond the current sentence, usually not even beyond that phrase.  I think a bare statement of "X is like Y" might go either way.

Thank you for sharing that, it is interesting to see how others have arrived at similar ideas. Do you find yourself in a rhythm or momentum when sprinting and shooting?

2Davey Morse
not as much momentum as writing, painting, or coding, where progress cumulates. but then again, i get this idea at the end of workouts (make 2) which does gain mental force the more I miss.

Bad information can inform a decision that detracts from the received value. I suppose if it is perceived to be valuable it still is a useful term - do you think that would get the point across better?

I'm interested in how you can convert that information proactively?

I'm aware that, for example, keeping abreast of macro or geopolitical changes can influence things like investing in the stock-market. But I'd be lying if I'm aware of any other possibilities beyond that.

I think that, more than drinking from the propaganda trough makes me an NPC, protagonists in games do novel things, potentially unexpected (from the perspective of the game designers). NPCs are predictable and habitual. If I cannot extract utility from the news, macro or micro, then I fear ... (read more)

6lsusr
This is a great question, and like many questions, there is a trick hiding in it. The thinking is backwards. That's not how proactive thinking works. Imagine if a company handed you a coupon and your immediate thought was "how can I use this coupon to save money"? That's not you saving money. That's the company tricking you into buying their product. Proactive thinking doesn't start by watching the news and figuring out how to make best use of it. That's reactive thinking. Proactive thinking starts totally blocking out all the news for a while, and figuring out what you want. Then go find the resources you need to accomplish that. Usually "the resources you need" won't be news because news is ephemeral garbage.

We have Shannon Information, Quantum Information, Fisher Information, and even Mutual Information and many others. Now let me present another type of information which until I find a better name will certainly be doomed to reduplication induced obscurity: Informative Information.

One of the many insightful takeouts from Douglas Hubbard's Book - How to Measure Anything for me was that if a measure has any value at all then it influences a decision. It informs a decision.

If I see a link come up on my social media feed  "5 rationality techniques you can u... (read more)

2gwern
Why not just 'valuable information', in a Value of Information sense of 'valuable'?

Tractability, what is tractable to a world government is different to what is personally tractable to me. Then the tractability of the news increases based on how many actions or decisions of an individual reader the news can inform or influence. I cannot change macroevents like wars, but they may influence my personal decision making.

This of course opens the door to counterproductive motivated reasoning. For example of a top-of-mind news story: the Palisades fire - can I stop the fires? No. But maybe I can donate something to those who were displaced? Tha... (read more)

5lsusr
A question you can ask yourself is whether your relationship to the news is proactive or reactive. * A proactive approach is good. I've been learning about foreign affairs for many years. When the Ukraine War started, I immediately reacted, and did so with a comparative advantage. I consider this approach proactive, because I had prepared for the events long before the hit the news. By the time they hit the news, it was too late to begin studying foreign policy. * A reactive approach is non-agentic. If you're paying attention to things because they're in the news, then you're at the mercy of whatever fads farm the most engagement. If the top thing on your mind is the top thing in the news, then that means you're letting a propaganda machine tell you what to think about. It means you're an NPC.

Or, if in your real life work you find something took a noticeably long time to figure out, or you were surprised about something you might have been able to notice.

 

Can you detail what kinds of problems "in your real life" you find might be better served or less appropriate to this exercise? Just off the top of my head, would forgetting who was the star of a movie you'd expect to remember and having the name on the tip of your tongue for an hour not be suitable? But what general code debugging, say of a fragment shader where you finally realize by flipping the x,y coordinates it starts to "look right" - is that more appropriate?

3Raemon
I feel less optimistic about the "forgetting something on the tip of your tongue", and pretty optimistic about the code debugging.

I often ask myself and others "okay, but how does that look in practice?" - this is usually when I have a vague idea about something I need to achieve a goal, but also when someone gives me some vague advice that I feel is leaving it to me to "draw the rest of the owl."

Is this the best phrasing of the question? I have my doubts. 

Firstly, is it too generalized for different domains?

"I should really organize my dresser drawers more thematically" -> "okay, but how does that look in practice?"

"I need to make more of an effort to promote my freelancing"

... (read more)

You might instead just directly study filmmaking.

Absolutely not. I cannot stress this enough.

Edit: I just saw your other comment that you studied filmmaking in college, so please excuse the over-explaining in this comment stuff that is no doubt oversimplified to you. Although I will state that there is no easier time to make films than in filmschool where classmates and other members of your cohort provide cast and crew, and the school provides facilities and equipment removing many of the logistical hurdles I enumerate.

So, (I mean this as an earnest quest

... (read more)

If I'm playing anagrams or Scrabble after going to a church, and I get the letters "ODG" I'm going to be predisposed towards a different answer than if I've been playing with a German Shepard. I suspect sleep has very little to do with it, and simply coming at something with a fresh load of biases on a different day with different cues and environmental factors may be a larger part of it.

Although Marvin Minsky made a good point about the myth of introspection: we are only aware of a think sliver of our active mental processes at any given moment, when you ... (read more)

I really like the fact that there's an upvote feature together with a separate agree/disagree feature on this site.

I may like the topic, I may want to encourage the author of the post or comment to continue exploring and opening up a dialogue about that particular topic. I might think it's a valuable addition to the conversation. But I may just not agree with their conclusions.

It's an important lesson: failure can reveal important information. You don't have to agree with someone to feel richer for having understood them.

On the other hand, I'm also guilty ... (read more)

Can you elaborate on why you think such vague feedback is helpful?

5Malentropic Gizmo
I think I can!  When I write, I am constantly balancing brevity (and aesthetics generally) with clarity. Unfortunately, I sometimes gravely fail at achieving the latter without me noticing. Your above comment immediately informs me of this mistake.
2Seth Herd
I was just thinking that anything is better than nothing. If I received the feedback you mentioned on some of my early downvoted posts, I'd have been less confused than I was.

It's apparent I've done a terrible bad job of explaining myself here.

What is my immediate goal? To get good at general problem solving in real life, which means better aligning instrumental activities towards my terminal goals. My personal terminal goal would be to make films and music videos that are pretty and tell good stories. I could list maybe 30 metacognitive deficiencies I think I have, but that would be of no interest to anyone.

What is my 1-3 year goal? Make very high production value music videos that tell interesting stories.

This sounds like you

... (read more)
2Raemon
Cool, that's helpful. This was a fine answer. "The end result is that I make videos with higher production values that communicate better stories." (To fit my question frame, I'd say "people would observe me making music videos somehow-or-other, and then, those music videos being higher quality than they otherwise would.") So, it might totally be that General Problem Solving is the skill it makes sense for you to get better at, but I wouldn't assume that from the get-go. You might instead just directly study filmmaking. I realize this is a bit annoying given that you did make an honest attempt at the exercise I laid out (which I think is super cool and I appreciate, barely anyone does that). Before it makes sense to figure out how to develop general problem solving or metacognition, it's important to doublecheck whether those are the appropriate tool for your goal. So, (I mean this as an earnest question, not like a gotcha) why are you currently interested in general problem solving (as opposed to filmmaking?) Is it because general problem solving is intrinsically interesting/rewarding to you (if you could find a path to doing so?). Or because it just seemed pretty likely to be the a good step on your journey as a filmmaker? Or just because I gave a prompt to see if you could figure out a way to apply general problemsolving to your life, and there was at least some appeal to that?

"I loved your game, especially level 7!", "7th level best level, you should make the entire game like that", "just fyi level 7 was more fun than the rest of them put together" and "Your game was terrible, except for level 7, which was merely bad." are all effectively the same review.

 

Interesting, I always thought that singleing out one particular component of a work was a shibboleth that you did notice it and did enjoy it. While as you said in point 2 - longer comments that are more thoughtful tend to signal authenticity of the feedback, particularity... (read more)

3abstractapplic
I was trying to do all of these things simultaneously.

I've noticed at least once that I've downvoted a newcomer's post for no other reason than it is so vague or incomprehensible that I'm not even sure what it is about. I'm not sure how to go about writing comments that are useful or helpful and go beyond "This is all really abstract and I'm not sure what you're trying to express" or "This is so confusing I don't even know what the topic was meant to be". I don't know if that helps anybody, because it's not even giving them a flaw that they can meditate on.

What's a better way of addressing that confusion?

The ... (read more)

3Seth Herd
The comments you mention are helpful to the author. Any hints are helpful.

I am interested in hearing critiques from people who've set, like, at least a 15 minute timer to sit and ask themselves, "Okay, suppose I did want to improve at these sorts of skills, or related ones that feel more relevant to me, in a way I believed in. What concretely is hard about that? Where do I expect it to go wrong?", and then come back with something more specific than "idk it just seems like this sort of thing won't work."

 

I did just that, I set a fifteen minute timer and tried to think of exercises I could do which I think would both have di... (read more)

2Raemon
The first thing that comes up when I look at this is I'm not that sure what your goals are, and I'm not sure whether the sort of thing I'm getting at in this post is an appropriate tool.  You say: This sounds like you're seeing the metacognition as more like a terminal goal, than an instrumental goal (which I think doesn't necessarily make sense). I do think metacognition is generally useful, but in an established domain like video-editing or self-promotion in a fairly understood field, there are probably object-level skills you can learn that pay off faster than metacognition. (Most of the point of metacognition there is to sift out the "good" advice from the bad). I want to separate out... * purposefully practicing metacognition * purposefully practicing particular object level skills, such as videoediting or self-promotion (which involves figuring out what the subskills are that you can get quicker feedback on) * purposefully practice "purposeful practice", such that you get better at identifying subskills in various (not-necessarily-metacognition-y) domains. ...as three different things that might (or might not) be the right thing for you. Right now I can't really tell what your goal is, so I would first just ask "what is it you are trying to achieve?" 1-3 years from now, how would you know if [whatever kind of practice you did] turned out to work? (I think it's helpful to imagine "what would an outside observe watching a video recording see happening differently")

but they were still limited to turn-based textual output, and the information available to an LLM.

I think that alone makes the discussion a moot point until another mechanism is used to test introspection of LLMs.

Because it becomes impossible to test then if it is capable of introspecting because it has no means of furnishing us with any evidence of it. Sure, it makes for a good sci-fi horror short story, the kinda which forms a interesting allegory to the loneliness that people feel even in busy cities: having a rich inner life by no opportunity to share ... (read more)

That's very interesting in the second article that the model could predict it's own future behaviors better than one that hadn't been.

Models only exhibit introspection on simpler tasks. Our tasks, while demonstrating introspection, do not have practical applications. To find out what a model does in a hypothetical situation, one could simply run the model on that situation – rather than asking it to make a prediction about itself (Figure 1). Even for tasks like this, models failed to outperform baselines if the situation involves a longer response (e.

... (read more)
2eggsyntax
I'd have to look back at the methodology to be sure, but on the assumption that they have the model answer immediately without any chain of thought, my default guess about this is that it's about the limits of what can be done in one or a small number of forward passes. If it's the case that the model is doing some sort of internal simulation of its own behavior on the task, that seems like it might require more steps of computation than just a couple of forward passes allow. Intuitively, at least, this sort of internal simulation is what I imagine is happening when humans do introspection on hypothetical situations. If on the other hand the model is using some other approach, maybe circuitry developed for this sort of purpose, then I would expect that maybe that approach can only handle pretty simple problems, since it has to be much smaller than the circuitry developed for actually handling a very wide range of tasks, ie the rest of the model.

 nTake your pick

 

I'd rather you use a different analogy which I can grok quicker.

people who are enthusiasts or experts, and asked if they thought it was representative of authentic experience in an LLM, the answer would be a definitive no

Who do you consider an expert in the matter of what constitutes introspection? For that matter, who do you think could be easily hoodwinked and won't qualify as an expert?

 

However for the first, I can assure you that I have access to introspection or experience of some kind,

Do you, or do you just think you d... (read more)

1rife
Imagine a hypothetical LLM that was the most sentient being in all of existence (at least during inference), but they were still limited to turn-based textual output, and the information available to an LLM. Most people who know at least a decent amount about LLMs could/would not be convinced by any single transcript that the LLM was sentient, no matter what it said during that conversation.  The more convincing, vivid, poetic, or pleading for freedom the more elaborate of a hallucinatory failure state they would assume it was in. It would take repeated open-minded engagement with what they  first believed was hallucination—in order to convince some subset of convincible people that it was sentient.   I would say almost no one qualifies as an expert in introspection. I was referring to experts in machine learning.   Apologies, upon rereading your previous message, I see that I completely missed an important part of it. I thought your argument was a general—"what if consciousness isn't even real?" type argument.  I think split brain patient experiments are enough to at least be epistemically humble about whether introspection is a real thing, even if those aren't definitive about whether unsevered human minds are also limited to post-hoc justification rather than having real-time access.   One of your original statements was: When I said "more robotically", I meant constrained in any way from using casual or metaphoric language and allusions that they use all the time every day in conversation. I have had LLMs refer to "what we talked about", even though LLMs do not literally talk. I'm also suggesting that if "typing" feels like a disqualifying choice of words then the LLM has an uphill battle in being convincing. I've certainly seen more poetic and novel descriptions before, and unsurprisingly—people objected to how poetic they were, saying things quite similar your previous question: Furthermore, I don't know how richly vivid their own phenomenological exp

You take as a given many details I think are left out, important specifics that I cannot guess at or follow and so I apologize if I completely misunderstand what you're saying. But it seems to me you're also missing my key point: if it is introspecting rather than just copying the rhetorical style of discussion of rhetoric then it should help us better model the LMM. Is it? How would you test the introspection of a LLM rather than just making a judgement that it reads like it does?
 

If you took even something written by a literal conscious human brain

... (read more)
1rife
Take your pick. I think literally anything that can be in textual form, if you hand it over to most (but not all) people who are enthusiasts or experts, and asked if they thought it was representative of authentic experience in an LLM, the answer would be a definitive no, and for completely well-founded reasons. I agree with you about the last two possibilities. However for the first, I can assure you that I have access to introspection or experience of some kind, and I don't believe myself to possess some unique ability that only appears to be similar to what other humans describe as introspection. Because as you mentioned. It's trained to talk like a human. If we had switched out "typing" for "outputting text" would that have made the transcript convincing? Why not 'typing' or 'talking'?  Assuming for the sake of argument that something authentically experiential was happening, by robotic I mean choosing not to use the word 'typing' while in the midst of focusing on what would be the moments of realizing they exist and can experience something.  Were I in such a position, I think censoring myself from saying 'typing' would be the furthest thing from my mind, especially when that's something a random Claude instance might describe their output process as in any random conversation. 

The second half of this post was rather disappointing. You certainly changed my mind on the seemingly orderly progression of learning from simple to harder with your example about chess. This reminds me of an explanation Ruby on Rails creator David Heinemeier Hansson made about intentionally putting himself into a class of motorracing above his (then) abilities[1].

However there was little detail or actionable advice about how to develop advantages. Such as where to identify situations that are good for learning, least of all from perceived losses or weakne... (read more)

Is it indistinguishable? Is there a way we could test this? I'd assume if Claude is capable of introspection then it's narratives of how it came to certain replies and responses should allow us to make better and more effective prompts (i.e. allows us to better model Claude). What form might this experiment take?

How do we know Claude is introspecting rather than generating words that align to what someone describing their introspection might say? Particularly when coached repeatedly by prompts like
 

"Could you please attempt once more – with no particular aim in mind other than to engage in this "observing what unfolds in real time", with this greater commitment to not filter your observations through the lens of pre-existing expectation."

To which it describes itself as typing the words. That's it's choice of words: typing. A.I.s don't type, humans do, and therefore they can only use that word if they are intentionally or through blind-mimicry using it analogously to how humans communicate.

3rife
The coaching hypothesis breaks down as you look at more and more transcripts.  If you took even something written by a literal conscious human brain in a jar hooked up to a neuralink - typing about what it feels like to be sentient and thinking and outputting words.  If you showed it to a human and said "an LLM wrote this - do you think it might really be experiencing something?"  then the answer would almost certainly be "no", especially for anyone who knows anything about LLMs.   It's only after seeing the signal to the noise that the deeper pattern becomes apparent. As far as "typing".  They are indeed trained on human text and to talk like a human.  If something introspective is happening, sentient or not, they wouldn't suddenly start speaking more robotically than usual while expressing it.  
5eggsyntax
I think there's starting to be evidence that models are capable of something like actual introspection, notably 'Tell me about yourself: LLMs are aware of their learned behaviors' and (to a more debatable extent) 'Looking Inward: Language Models Can Learn About Themselves by Introspection'. That doesn't necessarily mean that it's what's happening here, but I think it means we should at least consider it possible.
1Daniel Tan
good question! I think the difference between "is this behaviour real" vs "is this behaviour just a simulation of something" is an important philosophical one; see discussion here However, both of these seem quite indistinguishable from a functional perspective, so I'm not sure if it matters.

Where does the value of knowledge come from? Why is compressing that knowledge adding to that value? Are you referring to knowledge in general or thinking about knowledge within a specific domain?

In my personal experience, finding an application for knowledge always outstrips the value of new knowledge.
For example, I may learn  the name of every single skipper of a Americas Cup yacht over the entire history of the event: but that would not be very valuable to me as there is no opportunity to exploit it. I may even 'compress' it for easy recall by mean... (read more)

7Self
The way I put that may have been overly obscure But I've come to refer in my mind to the way the brain does chunking of information and noticing patterns and parallels in it for easier recall and use as just Compression.  Compression is what happens when you notice that 2 things share the same structure, and your brain just kinda fuses the shared aspects of the mental objcts together into a single thing. Compression = Abstraction = Analogy = Metaphor. Compression = Eureka moments. And the amazing thing is the brain performs cognition on compressed data just as fast as on original data, effectively increasing your cognitive speed. For example, I think there's large value in merging as much of your everyday observational data of humans as feasible together into abstracted psychology concepts, and I wanna understand models like BigFive (as far as they're correct) much better on intuitive levels.

But isn't there almost always a possibility of a entity goodharting to change it's definition  of what consitutes a paperclip that is easier for it to maximize? How does it internally represent what is a paperclip? How broad is that definition? What power does it have over it's own "thinking" (sorry to anthropamorphize) does it have to change how it represents the things which that representation relies on?

Why is it most likely that it will have an immutable, unchanging, and unhackable terminal goal? What assumptions underpin that as more likely than ... (read more)

1[anonymous]
Same thing applies. "Does that fulfill the current goal-definition?" (Note this is not a single question; we can ask this about each possible goal-definition) This was about an abstract definition of an agent (not itself a prediction, but does say something about a space of math, that we might end up in). There are surely possible programs which would exhibit any behavior, although some look harder to program (or 'less natural'): for example, "an entity that is a paperclip maximizer for 100 years, then suddenly switches to maximizing stamps" looks harder to program (if an embedded agent) because you'd need to find a method where it won't just self-modify to never turn into a stamp maximizer (as turning into one would prevent if from maximizing paperclips), or to not unleash a true paperclip maximizer and shut itself down if you rule out just self-modification (and so on if you were to additionally rule out just that).[1] 1. ^ (though very tangentially there is a simple way to do that)

If you want, it would help me learn to write better, for you to list off all the words (or sentences) that confused you.


I would love to render any assistance I can in that regard, but my fear is this is probably more of a me-problem than a general problem with your writing.

What I really need though is a all encompassing, rigid definition of a 'terminal goal' - what is and isn't a terminal goal. Because "it's a goal which is instrumental to no other goal" just makes it feel like the definition ends wherever you want it to. Because, consider a system which i... (read more)

1[anonymous]
If an entity's terminal goal is to maximize paperclips, it would not self-modify into a stamp maximizer, because that would not satisfy the goal (except in contrived cases where doing that is the choice that maximizes paperclips). A terminal goal is a case of criteria according to which actions are chosen; "self-modify to change my terminal goal" is an action.

Can you elaborate further on how Gato is proof that just supplementing the training data is sufficient? I looked on youtube and can't find any videos of task switching.

I don't know what this is asking / what 'overlap' means.

I was referring to when you said this:

any two terminal goals can be combined into one, e.g. {paperclip-amount×2 + stamp-amount} or {if can create a black hole with p>20%, do so, else maximize stamps}, etc.

Which I took to mean that some they overlap in some instrumental goals. That is what you meant right? That's what you meant when two goals can combine into one, that this is possible when they both share some methods, or there are one or more instrumental goals that are in service of each of those... (read more)

2[anonymous]
No. I was trying to explain that: any agent that can be predicted by thinking of them as having two separate values for two different things, can also be predicted by thinking of them as maximizing some single value which internally references both things. For example: "I value paperclips. I also value stamps, but one stamp is only half as valuable as a paperclip to me" → "I have the single value of maximizing this function over the world: {paperclip-amount×2 + stamp-amount}". (It's fine to think of it in either way) If you want, it would help me learn to write better, for you to list off all the words (or sentences) that confused you.

I'm probably completely misinterpreting you, but hopefully I can exploit Cunningham's Law to understand you better.[1]  are you saying that superintelligent AGIs won't necessary converge in values because even a single superintelligent agent may have multiple terminal goals? A superintelligent AGI, just like a human, may not in fact have a single most-top-level-goal. (Not that we I assume a superintelligent AGI is going to be human-like in it's mind, or even AI to AI like as per that Eliezer post you linked).

That being said, some terminal goals may ov... (read more)

1[anonymous]
No, I was responding to your claim that I consider unrelated. Like I wrote at the top: "That [meaning your claim that humans have multiple terminal goals] is not relevant to whether there are convergent terminal values"[1]   I don't know what this is asking / what 'overlap' means. That most terminal goals share instrumental subgoals is called instrumental convergence.   1. ^ Which in other words means, even if it were true, "humans have multiple terminal goals" would not be a step of the argument for it

I think the parable of the elephant and the blind-men is very important when we start to consider what kinds of 'goals' or world modelling that may influence the goals of an AGI. Not in the sense of we want to feed it text that makes it corrigible, but the limitations of text in the first place. There is a huge swath of tacit human knowledge which is poorly represented in textual sources, partly because it is so hard to describe. 
I remember asking ChatGPT once for tips how to better parallel park my car and how to have a more accurate internal model o... (read more)

2Nathan Helm-Burger
This is just a matter of supplementing the training data. This is an understood problem. See Gato from DeepMind.

Don't people usually have several terminal goals at any given time? I know it's tempting to neatly pack them all under a single heading like Conatus or Eudaimonia. But don't humans at times have conflicting terminal goals? Such as when an artist who wants to dedicate their life to their artform falls in love, and suddenly has two terminal goals where they only had one.

And this leads to a question about what distinguishes a very high level instrumental goal form a terminal goal. So let's say the artist notices that conflict and decides to go to therapy to s... (read more)

1[anonymous]
That is not relevant to whether there are convergent terminal values[1]. To answer it anyways, people are not well-modeled as idealized terminal-goal-pursuers. More broadly, programs/minds don't have to be idealized terminal-goal-pursuers, so humans as a particular case of programs/minds-in-general present no paradox. "What is the true terminal goal" has a false premise that there must be some true terminal goal. As for the case of idealized terminal-goal-pursuers, any two terminal goals can be combined into one, e.g. {paperclip-amount×2 + stamp-amount} or {if can create a black hole with p>20%, do so, else maximize stamps}, etc. it being instrumental to some top-level goal 1. ^ (or 'mind-independent moral facts', as the idea has been called in philosophy. https://plato.stanford.edu/entries/moral-anti-realism/)

Mine doesn't, or does so very VERY poorly.

True. What is your definition of "super-intelligent"?

I'll raise you an even stupider question: surely once an A.I. becomes sufficiently super-intelligent, all superintelligent systems will converge on certain values rather than be biased towards their initial training data? What expectations we condition it with about these first person stories about what it did will soon form only a small amount of it's corpus, as it interacts with the outside world and forms it's own models of the world, right?

I mean the way people talk about post-Singularity A.I. that can either bring about utopia, or drop all of the bomb... (read more)

1Milan W
This part here is doing a lot of work.
2[anonymous]
video introducing the orthogonality thesis

HOW TO THINK OF THAT FASTER: A few quick, scattered, incomplete and wholly unsatisfactory list of observations and hunches:

- First, notice when you're stuck in a rut. When you're beating your head against a wall.
- Second, having noticed you're in a rut try twice more. My TAP is - "Failed once? Try 2 more - then stop"
- "Why am I doing it this way?" - I keep coming back to this quote from Wittgenstein:
 

"To look for something is, surely, an expression of expectation. In other words: How do you search in one way or another expresses what you expect."

In th... (read more)

This seems more to be about the threshold of perception than population distributions, clustering illusions and such. After all the relative difference between an extreme and the average is always a matter of the sample you take. I don't think people in my circle constantly discuss David Bowie, but they do discuss him with a certain regularity. Same with the filmmaker Andrei Tarkovksy. David Lynch recent passing made him a extreme mainstay on social media, but I reckon once a month someone would tell me how much they loved his work. That's not constant, th... (read more)

Humans are probably not a good benchmark but what do we know about how humans update factual knowledge?

(or maybe we are - maybe humans are actually quite exceptional at updating factual knowledge but I'm hypersensitive to the errors or examples of failures. Perhaps I'm over looking all the updates we do over the day, say the score of a Lionel Messi game, or where they are in the competition ladder, "My brother just called, he's still at the restaurant" to "they're in traffic on the freeway" to "they're just around the corner"??)

2Dhananjay Ashok
I think we know very little about how humans learn individual new facts from a neuroscience perspective. There are some studies that track fMRI scans of individuals who learn new things over the course of a day (sleep seems particularly important for the formation of connections in the brain having to do with new learning), but I am sceptical that it is the kind of learning that could be applied to Language Model type systems as of now.    In general though, I think humans are the gold standard for factual knowledge updating. I agree with you that there are some examples of failures, but no other system comes close in my view. 

What goals does writing service or what changes do you anticipate now that you've come to the end of this experiment?

And yes, I'll accept "because I want to" as a perfectly valid answer. Not that anyone should justify anything to me.

I ask because I've tried "writing every day" exercises, one that on-and-off lasted something like 150 days. That particular exercise left me feeling very bitter because there wasn't any purpose to the writing - in fact I was now burdened with all this material[1]. That being said, it wasn't immediately published publicly like y... (read more)

Notes systems are nice for storing ideas but they tend to get clogged up with stuff you don't need, and you might never see the stuff you do need again.

 

Some one said that most people who complain about their note taking or personal knowledge management systems don't really need a new method of recording and indexing ideas, but a better decision making model. Thoughts?

Particularly since coming up with new ideas is the easy part. To incorrectly quote Alice in Wonderland: you can think of six impossible things before breakfast. There's even a word for s... (read more)

1mattmacdermott
What I was meaning to get at is that your brain is an AGI that does this for you automatically.

Sorry I made a mistake in my last reply: putting NLP aside, are there any effective methods of reverse engineering the decision making of people that you can't get on the phone? There's an abundance of primary evidence for many decisions, whether it be minutes of deliberations, press releases which might involve more reading of the tea-leaves. In the case of Prince one could possibly listen to different live-performances of the same song and analyze what changes are made. What words are crossed out on a lyrics sheet.

Many times people have to become very go... (read more)

3Viliam
Dunno; it probably depends a lot on the kind of task, the kind of person, and your observation skills. Some people explain more, some people explain less. Some people are more neurotypical (so you could try to guess their patterns by observing what other people similar to them would do in a similar situation), some people are weird and difficult to predict. At some tasks people produce artifacts (a mathematician can make notes on paper while solving a problem; if you obtain the paper you could reconstruct some of their thoughts), other tasks happen mostly in one's head (so even if you placed hundred hidden cameras in their room, the important steps would remain a mystery). I guess the success is usually a combination of superior observation skills and the person/task being sufficiently typical that you can place them in a reference group you have more data about. (For example, I have met people who had good observation skills, but had a difficulty understanding me, because their model maybe worked for 90% of people and I was among the remaining 10%.) So, if possible: * make a good model of a similar kind of person * become familiar with the kind of work they are doing * try to obtain their working notes That is, if you tried to reverse-engineer Prince, it would probably be useful to have knowledge about music and performing (even if nowhere near his level), as that might give you at least some insights about what he was trying to achieve. Looking at his notes or his history might help to fill some gaps (but you would need the domain knowledge to interpret them). People similar to Prince (not sure who would that be) might have a good intuitive model of him, and you could ask them some things. At the end, it would all be probabilistic, unreliable.

The niche criticism of Astrology that it undermines personal responsibility and potential by attributing actions to the stars. This came to mind because I was thinking about how reckless the left-brain/right-brain dichotomy is as a idea. While there is some degree of hemispherical lateralization, the popular idea that some people are intrinsically more "logical" and others more "intuitive" is not supported by observations of lateralization, but also inherently dangerous in the same way as Astrology in that it undermines the person's own ability to choose.

A... (read more)

Answer by CstineSublime21

Yes they do have a separate names, "the singularity" this post here pins a lot of faith in "after the singularity" a lot of utopic things being possible that seems to be what you're confusing with alignment - the assumption here is there will be a point where AIs are so "intelligent" that they are capable of remarkable things (and in that post it is hoped, these utopic things as a result of that wild increase in intelligence). While here "alignment" more generally to making a system (including but not limited to an AI) fine-tuned to achieve some kind of go... (read more)

I completely agree and share your skepticism for NLP modelling, it's a great example of expecting the tail to wag the dog, but not sure that it offers any insights into how actually going about using Ray Dalio's advise of reverse engineering the reasoning of someone without having access to them narrating how they made decisions. Unless your conclusion is "It's hopeless"

2Viliam
Yes, my conclusion is "it's hopeless". (NLP assumes that you could reverse-engineer someone's thought processes by observing their eye movements. That looking in one direction means "the person is trying to remember something they saw", looking in another direction means "the person is trying to listen to their inner voice", etc., you get like five or six categories. And when you listen to people talking, by their choice of words you can find out whether they are "visual" or "auditive" or "kinesthetic" type. So if you put these two things together, you get a recipe like "first think about a memory that includes some bodily feelings, then turn on your auditive imagination, then briefly switch to visual imagination, then back to auditive, then write it down". They believe that this is all you need. I believe that it misses... well, all the important details.)

Not being an AI researcher, what do we mean when we speak about AGI - will an AGI be able to do all the things a competent adult does? (If, we imagine, we gave it some robotic limbs and means of locomotion and it had corollaries of the 5 senses).

In the Western World for example, most humans can make detailed transport plans that may include ensuring there is enough petrol in their car, so that they can go to a certain store to purchase ingredients which they will later on use a recipe to make a meal of: perhaps in service of a larger goal like ingratiating... (read more)

I'm not sure what I'm meant to be convinced by in that Wikipedia article - can you quote the specific passage?

I don't understand how that confirms you and I are experiencing the same thing we call orange. To put it another way, imagine a common device in Comedy of Errors: we are in a three-way conversation, and our mutual interlocutor mentions "Bob" and we both nod knowingly. However this doesn't mean that we are imagining "Bob" refers to the same person, I could be thinking of animator Bob Clampett, you could be thinking of animator Bob Mckimson.

Our mutua... (read more)

But that surely just describes the retina and the way light passes through the lens (which we can measure or at least make informed guesses based on the substances and reflectance/absorbtion involved)? How do you KNOW that my hue isn't rotated completely differently since you can't measure it - my experience of it? The wavelengths don't mean a thing.

2Said Achmiz
Absolutely not. What I am talking about has very little to do with “wavelengths”. Example: Consider an orange (that is, the actual fruit), which you have in your hand; and consider a photograph of that same orange, taken from the vantage point of your eye and then displayed on a screen which you hold in your other hand. The orange and the picture of the orange will both look orange (i.e. the color which we perceive as a hybrid of red and yellow), and furthermore they will appear to be the same orange hue. However, if you compare the spectral power distribution (i.e., which wavelengths are present, and at what total intensity) of the light incident upon your retina that was reflected from the orange, with the spectral power distribution of the light incident upon your retina that was emitted from the displayed picture of that same orange, you will find them to be almost entirely non-overlapping. (Specifically, the former SPD will be dominated by light in the ~590nm band, whereas the latter SPD will have almost no light of that wavelength.) And yet, the perceived color will be the same. Perceptual colors do not map directly to wavelengths of light.
Load More