Every so often I come back to a daydream about an investigation game (think Her Story, Immortality, Gone Home, Digital: A Love Story, Return of the Obra Dinn, etc.) but with a premise like you're in a small town and have access to a bunch of messy data from all over the place, mostly in Excel files but also receipts, security cam footage, maps, stuff like that.
Maybe you're a a hacker or an AI that was able to gain illicit access to everything, but now you have to clean up and correlate everything without outside help (i.e. no ability to ask things like "Wh...
What would have presumably given much different results would be Claude Sonnet 4.5, which is actually a lot less sycophantic by all reports (I’m a little worried it agrees with me so often, but hey, maybe I’m just always right, that’s gotta be it.)
Now you've got me wondering if I'm being reverse-sycophantic, and have been trained to say things Claude would agree with?
I agree with the idea of looking at customer response management (CRM) systems for ideas. This talk feels like a pretty good overview of that idea: https://www.youtube.com/watch?v=8jwiABwGC6c
Because you hopefully may enjoy the ideas, I've been kind of tackling this from a hobbyist perspective:
Lately I'm drawing inspiration from this articles, and imagining that I'm "building myself a skrode": https://medium.com/@greyboi/building-a-skrode-initial-thoughts-a195c4a0663d - and also the original story, A Fire Upon the Deep where skroderiders are introduced. Chiefly, the idea that your own skrode is something that is DEEPLY personalized and customizable, including p...
I think I feel the same sort of 'What if we just said EVERYTHING deserves welfare?' thought. I care for my birds, but I also care for my plants, and care for my books, each in their own way.
Like, if someone built this small skin-device-creature, and then someone else came along and smashed it then burned the pieces, I think I would be a little sad for the universe to have 'lost' that object. So there's SOEMTHING there that is unrelated to "can it experience pain?", for me.
Explanation
(The post describes a fallacy where you rule out a few specific members of a set using properties specific to those members, and proceed to conclude that you've ruled out that entire set, having failed to consider that it may have other members which don't share those properties. My comment takes specific examples of people falling into this fallacy that happened to be mentioned in the post, rules out that those specific examples apply to me, and proceeds to conclude that I'm invulnerable to this whole fallacy, thus committing this fallacy.
(Unle
There was a little Animal Crossing mod that made the rounds a little more 'gently' than I expected.
I think the trick here might be a game that runs a local, small, known-ethically-sourced model, but even if we had more than the one (Comma) that's still a lot of ire to overcome before you can even get to the elevator pitch for the game.
This three-factor framework reminds me of an idea from this: https://www.shamusyoung.com/twentysidedtale/?p=12768
If I had only attended one of these places, I probably would have concluded, “This place is what being a Christian is all about”. But these three points form a plane, and by moving around on that plane I can view Christianity from a lot of different angles and extrapolate a lot of other kinds of churches.
This idea of "third options which break overfit 1D mental models" has stuck with me for a big portion of my life now.
I'd like to relate this old blog post I found recently after my love affair with A Fire Upon the Deep: https://medium.com/@greyboi/building-a-skrode-initial-thoughts-a195c4a0663d
I’ve been wondering about addressing this [working memory] slippage with tech. For instance, as a youngster I used to code on tiny screens, but nowadays I use as many large monitors as I can get; they help me retain context, taking the load off my working memory. So tech can definitely help if you use it well.
One of the things I’m thinking about primarily is adopting a prosthetic memory.
I think you would enjoy this video from Alan Moore, where he starts off with a similarly fascinating word-by-word analysis: https://www.youtube.com/watch?v=ft8eO67auCs
Though I appreciate the reference at the end, I think an important part of this is that it's also so that when you meet a hero you can do more with this skill. You can "see" and engage with the real person. A person as real and mundane as every other person, with all the good and bad that implies.
I actually think "see" is too limited an analogy, because this really involves all your senses and reasoning, but it's also true that I feel it has a close connection to what artists call "learning to see", like maybe it's using the same mental circuits.
You can le...
Discovering that an alien species has bred a group of humans into what a pug is to a wolf would be absolutely horrific.
Makes me think of All Tomorrows: https://www.youtube.com/watch?v=imNtSPM3-r4
There's no Darwinian selective pressure to favor agents who engage in acausal trades.
I think I would make this more specific- there's no external pressure from that other universe, sort of by definition. So for acausal trade to still work you're left only with internal pressure.
The question becomes, "Do one's own thoughts provide this pressure in a usefully predictable way?"
Presumably it would be have to happen necessarily, or be optimized away. Perhaps as a natural side effect of having intelligence as all, for example. Which I think would be similar in argument as, "Do natural categories exist?"
I sat down one morning and for fun started trying to translate The Art of War from scratch, by simply going character by character and looking up the etymology and historical usage of each. Took me about two hours to get through the first page that way, and that was enough to be entertaining so I stopped there.
But, I noticed something, which reminds me of this "bodhi" vs. "enlightenment" contrast.
The text starts out by explaining five core concepts that make up "the art of war". The first one, I've normally seen written in English like this, from the trans...
This may seem like it's coming out of left field, but reading A Fire Upon the Deep a few weeks ago helped me find a calm perspective on this idea. In-universe the characters straight up have some of these discussions over the course of the book, and there's so much of all this stuff happening "just off screen".
The story is in part about the folly and impossibility of something as easy and comforting as "trust" between agents and systems in radically different scales and realities. Yet they are forced to coexist and interact regardless.
I haven't read his ot...
https://www.youtube.com/watch?v=_OpxrtUwjNw - This is a little fan project I did of that short story, as a sort of a radio play. I've never had it be relevant to a conversation before!
I was simply trying to decorate a compliment, so I suppose I will stop doing that 🤔 (EDIT: from a later vantage point, I think I now see it's better to say "sorry for adding a distraction" rather than passively projecting blame.)
I think you make an important point in this context- understanding that all the emotions you're "feeling" are still coming from you, not from them.
"A monk rowed out to the middle of a calm lake to meditate. A while later, they were bumped into and interrupted by another boat! The monk opened their eyes in anger, ready to chide the other monk for being so careless and making them so angry... to find the other boat empty. The anger was inside them, not from another monk."
Probably need a different name for goodharting.
This is honestly the biggest concept I struggle with trying to share and teach and raise familiarity with at work, in many contexts beyond just AI safety. There are some adjacent concepts like the cobra effect that are close, but are also just close enough to be distracting.
I appreciated this perspective from a prominent SCP author (Sam Hughes, who wrote and established the 'Antimemetics' stories and "sub-genre" on SCP): https://qntm.org/chatscp
..."But doesn't this whole scenario sound like an SCP?"
A couple of people suggested this. An LLM which bamboozles certain types of user with paranoid fantasy until they spiral into psychosis? That sounds like science fiction! It sounds like something out of the SCP project!
Okay, so, no? Because an SCP has to be anomalous in some way and this is clearly actually happening. Four years ago
This story captures a lot of bundled feelings I have which I want to try to put into words, though words are imperfect. This is the part of the story which I feel mirrored in my own:
My mom was fighting something, and she was fighting (and teaching me to fight) harder and earlier than I knew at the time, giving me tools and perspectives that most people don't get until they are adults (if they get them at all). Now I also see that a very few people even saw that she was fighting anything, much less what it was, and few see it now, and we just can't seem to ...
I notice I apply this lesson to the design of data entry forms/surveys in general as well: you need an 'other' option much more often than you would think, or one ends up with messy survey data: extra comments and thoughts crammed into the wrong questions wherever users can find an opening. EDIT: Upon further reflection, I also remember that I've had conversations during the rollout of surveys which included the solution "let's ask them to self-classify in another question" at multiple times in unrelated projects over the years.
The clouds are sort of permanent- as they are filled in, whatever degree of detail it's at is where it stays in my head, wherever they're "stored", and it just sort of sticks in there. It feels a lot like just putting energy into some sort of "progress bar", my best thinking I really do in just a meditation sort of behavior- closed eyes, relaxed pose, slow breathing [I'll shift back and forth into this as I'm reading challenging books for instance, or when working on a project there's a lot of "lean back and sit quietly for a moment, staring into the middl...
I'm with you on "music feels language-like", I think even just looking at spectrograms of music and speech, and comparing those to the spectrograms of random soundscapes, makes it visible that music at least plays with the same types of rhythmic, pitch, and formant patterns in our "sound view", but they have a difference similar to the difference between how a textbook "plays with the patterns of letters and punctuation" to convey an idea, and a Celtic knot "plays with patterns" to create a pattern that is sort of just... intrinsically nice to have around?...
This give me a feeling I would like to express via reference:
...At school, Doug finds that everyone there is dressed as him and becomes weirded out by the fact. The others also tell Doug that he is rocking the "Dylan Farnum" look. But Doug tells them that he always dresses like that. Doug has his mind stuck to the new fashion trend all day and he finally becomes fed up with the others saying that he is copying Dylan Farnum. So, he invites them all into his room and shows them his closet of clothes to prove that he is not copying Dylan Farnum. This, however,
I'd like to apply because I experiment a lot specifically on variation within an area of concept space, though I'll have to assemble a portfolio. Either way, I'd also like to throw this idea out there for consideration:
It would be nice for this to be paired with a commitment to use only ethically sourced image models, and to use that to help explain what that even means.
I don't know why everyone is making this so complicated when there's a clear disqualifying factor for me: Miyazaki himself has said that they did not consent to be trained on, would not have consented to being trained on, and do not want anyone making Ghibli art, and all of this was known before Sam Altman started pushing Ghibliffication. There are other factors too, but this one by itself is already sufficient for me.
EDIT: I see a lot of upvotes and disagreement on this comment, which I think I agree with. I should have clarified, this is personally disqu...
For sure! Much like the AI safety scorecard, no one is out of the red, but it seems like some of the older publishing house type companies are trying to respect existing content licensing institutions. However, I've seen many creators and artists complain that it doesn't matter; it's already too overshadowed by the actions of OpenAI et al.
I love this idea, it feels like it would also work for a lot of non-fiction, and I could see this being a part of a traditional book club too.