The subset of people who are Anki users and members of the competitive conspiracy might be interested in the Anki high score list addon I wrote: Ankichallenge
According to their site Jaan Tallinn is not the CEO but chairman of the board. Zvi Mowshowitz is the CEO.
4) Less than 1%; While internal evaluation is important, I primarily measure based on empirical behavior - I go without shaving my legs, I don't mind wearing stained clothing, I'll happily sit on wet grass... 2 or 3 seems just flat-out more plausible to me. This also fails to explain why I'm both low-status AND seem to be entirely content with that.
5) Less than 1%; I consciously believe that truth and understanding are LOW status, and I'd be rather baffled to be mistaken about that self evaluation.
Thank you though, I had not considered either :)
I go without shaving my legs, I don't mind wearing stained clothing, I'll happily sit on wet grass
There are Communities where this is high status behavior. But i presume you would have considered this if you were to belong to such a community.
A lot of these links are about how to get started coding. I am already a good programmer, maybe a damn good programmer. But I know there's a level above mine, and probably several. How do I get to the next level? What does the next level look like? Are there any resources that talk about how to become a great programmer?
Paul Graham's writings on programming and learning Haskell leveled me up. Although i suspect you are already at that level.
Our universe has not enough atoms or energy to destroy 3^^^^^3 paperclips.
Simulated paperclips. Because the Paperclipper is in a box.
I am extrapolating.
As I thought. I disagree that such effects are possible in reasonable time-frames (no-one is going to read constantly for a month) and may be totally impossible. I see no reason to think any work of fiction can lead to such a distortion of reality.
Groups of people are not that much harder to manipulate than single persons.
They are if you're trying to weaponize novels. If they work shifts then you cannot exploit the effects you claim for reading a novel for 24 hrs straight. They can watch each other and, if one of them is visibly compromised, prevent them from freeing the AI. A single individual is probably easier to manipulate, assuming a total lack of supervision and safeguards.
Simulated paperclips.
Now we get to the question how detailed the paperclips have to be for the paperclipper to care. I expect the paperclipper to only care when the paperclips are simulated individually and we can't simulate 3^^^^^^3 paperclips individually.
I see no reason to think any work of fiction can lead to such a distortion of reality.
I see no reason to think works of fiction that lead to such a distortion of reality are impossible.
from the answers it can derive a fairly accurate map of how it looks outside the matrix.
Which is a good thing, because we really do have such powers and we really don't value paperclips.
Another possible way would be to write a novel that is so interesting that the guard doesn't put it down and that leaves him in so a confused state that he types in the code, thinking he saves princess Foo from the evil lord Bar.
... were you seriously that confused or are you extrapolating to a "supercharged" novel?
I am sure any other well written book read for 24 hours would have a similar effect. I think it is likely that a potential guard is at most 2 orders of magnitude less vulnerable to such things than i was at that time. That's not enough against an AI that has 6 orders of magnitude more optimization power.
I somehow doubt there would be a single, full-time guard.
Which is a good thing, because we really do have such powers and we really don't value paperclips.
Our universe has not enough atoms or energy to destroy 3^^^^^3 paperclips.
... were you seriously that confused or are you extrapolating to a "supercharged" novel?
I am extrapolating.
I somehow doubt there would be a single, full-time guard.
Groups of people are not that much harder to manipulate than single persons.
Why would it believe us that we are able to destroy 3^^^^^3 paperclips?
Because we have magical powers from outside the matrix and don't value paperclips.
"arguing" is to narrow a word for describing the possibilities the AI has. For example it could manipulate us emotionally. It could write us a novel that leaves us in a very irrational state and then give us a bogus, but effective on us, argument for why we should let it out.
It would have to argue that destroying humanity and replacing it with paperclips was a good thing. Not impossible, sure, but easier to guard against.
I once read the fifth Harry Potter book nonstop for 24 hours and for a couple of hours afterwards i had difficulties distinguishing between me and Harry Potter.
That sounds like more a side effect of reading the same thing "nonstop for 24 hours" than a property of the book, unless you know of anyone else that happened to?
Because we have magical powers from outside the matrix [...].
The AI is vastly more smarter than we and can communicate with us. So it asks us questions which sound innocent to us, but from the answers it can derive a fairly accurate map of how it looks outside the matrix.
It would have to argue that destroying humanity and replacing it with paperclips was a good thing.
The goal of the AI is to have the guard execute the code that would let the AI access the outside world. Arguing with us could be one way to archive this goal. Although i agree it sounds like a unlikely way to succeed. Another possible way would be to write a novel that is so interesting that the guard doesn't put it down and that leaves him in so a confused state that he types in the code, thinking he saves princess Foo from the evil lord Bar.
A super smart AI who wants to reach this goal very badly will likely come up with a whole bunch of other possible ways. Some of which i would never consider even if i spent the next 4 decades thinking about that.
That sounds like more a side effect of reading the same thing "nonstop for 24 hours" than a property of the book [...]
Yes. I am sure any other well written book read for 24 hours would have a similar effect. I think it is likely that a potential guard is at most 2 orders of magnitude less vulnerable to such things than i was at that time. That's not enough against an AI that has 6 orders of magnitude more optimization power.
No, I mean it can hardly argue it would be in our best interest to let it out of the box, can it? And we can always threaten to destroy 3^^^^^3 paperclips if it wont cooperate, which is handy.
Why would it believe us that we are able to destroy 3^^^^^3 paperclips?
"arguing" is to narrow a word for describing the possibilities the AI has. For example it could manipulate us emotionally. It could write us a novel that leaves us in a very irrational state and then give us a bogus, but effective on us, argument for why we should let it out.
I once read the fifth Harry Potter book nonstop for 24 hours and for a couple of hours afterwards i had difficulties distinguishing between me and Harry Potter. It seems likely that a author who is a millions times smarter than Rowling and who has it as explicit goal, could write a novel that leaves me with far bigger misconceptions.
Boxing an AI of unknown friendliness may be a bad idea, but how about one known to be unfriendly? A paperclipper sounds pretty easy to manipulate and I think it would have a harder time persuading any guards to let it out. Am I missing something?
Could it be that you are confusing the complexity of a utility function of an agent with its optimization power? A super intelligent paperclipper has a simple utility function, but would have no problem reasoning about humans in great enough detail to find out what it has to say to get the guard to let it out of the box.
View more: Next
Subscribe to RSS Feed
= f037147d6e6c911a85753b9abdedda8d)
--Marcus Aurelius, Meditations, 8.50
If you get one bitter cucumber, asking for its cause may be a waste of time. But if you get a lot of bitter cucumbers, spending some time on changing that might give net positive utility.