Virtually no-one differentiates between a 4 and a 5 on these kind of surveys. Either they are the kind of person who "always" puts 5, or they "never" do.
With Rat. Adjacent or other overthibkers, you can give more specific anchors (eg 5 = this was the best pairing). Or you can have specific unsealed questions (ie:
Bonus points in a dating context: by being specific and authentic you drive away people who won't be compatible. In the egg example, even if the second party knows nothing about the topic, they can continue the conversation with "I can barely boil water, so I always take a frozen meal in to work" or "I don't like eggs, but I keep pb&j at my desk" or just swipe left and move on to the next match.
Follow up question: is this a permanent gain or temporary optimization (eg without further intervention, what scores would the subject get in 6 months?)
We know for sure that eating well and getting a good night's sleep dramatically improves performance on a wide array of mental tasks. It's not a stretch to think other interventions could boost short term performance even higher.
For further study: Did the observed increase represent a repeatable gain, or an optimization? Within-subject studies show a full SD variation between test sessions for many subjects, so I would predict that "a set of interventions" could produce a "best possible score" for an individual but hit rapid diminishing returns.
Communication bandwidth: if you find that you’re struggling to understand what the person is saying or get on the same page as them, this is a bad sign about your ability to discuss nuanced topics in the future if you work together.
Just pulling this quote out to highlight the most critical bit. Everything else is about distinguishing between BS and ability to remember, understand, and communicate details of an event (note: this is a skill not often found at the 100 IQ level). That second thing isn't necessarily a job requirement for all positions (eg sales, entry level positions), but being comfortable talking with your direct reports is always critical.
Mr. Pero got fewer votes than either major party candidate. Not a ringing endorsement. And I didn't say the chances were quite low, I said they were zero*. Which is at least 5 orders of magnitude difference from "quite low" so I don't think we agree about his chances.
*technically odds can't be zero, but I consider anything less likely than "we are in a simulation that is subject to intervention from outside" to be zero for all decision making purposes.
There is an actual 0% chance that anyone other than the Democratic or Republican nominee (or thier replacement in the event of death etc.) becomes president. Voting for/supporting any other candidate has, historically, done nothing to support that candidate's platform in the short or long term. If you find both options without merit, you should vote for your preferred enemy:
Note one weakness of this technique. An LLM is going to provide what the average generic written account would be. But messages are intended for a specific audience, sometimes a specific person, and that audience is never"generic median internet writer." Beware WIERDness. And note that visual/audio cues are 50-90% of communication, and 0% of LLM experience.
Agree that closer to reality would be one advisor, who has a secret goal, and player A just has to muddle through against an equal skill bot with deciding how much advice to take. And playing like 10 games in a row, so the EV of 5 wins can be accurately evaluated against.
Plausible goals to decide randomly between:
Arguing against A doesn't support Not A, but arguing against Not Not A is arguing against A (while still not arguing in favor of Not A) - albeit less strongly than arguing against A directly. No back translation is needed, because arguments are made up of actual facts and logic chains. We abstract it to "not A" but even in pure Mathematics, there is some "thing" that is actually being argued (eg, my grass example).
Arguing at a meta level can be thought of as putting the object level debate on hold and starting a new debate about the rules that do/should govern the object level domain.
Alice: grass is green -> grass isn't not green Bob: the grass is teal -> the grass is provably teal Alice: your spectrometer is miscalibrated -> your spectrometer isn't not miscalibrated.
...
I'm having trouble with the statement {...and has some argument against C'}. The point of the double negative translation is that any argument against not not A is necessarily an argument against A (even though some arguments against A would not apply to not not A). And the same applies to the other translation - Alice is steelmanning Bob's argument, so there shouldn't be any drift of topic.
Consider a scale that runs from "authentic real life" to "Lotus eater box" At any point along that scale, you can experience euphoria. At the Lotus Eater end, it is automatic. At the real life end, it is incidental. "Games" fall towards the Lotus Eater end of the spectrum, not as far as slot machines, but further from real life than Exercise or Eating Chocolate. Modern game design is about exploiting what is known about what brains like, to guide the players through the (mental) paths necessary to generate happy chems. They call it "being Fun" but that's j...
90% of games are designed to be fun. Meaning the point is to stimulate your brain to produce feel-good chemicals. No greater meaning, or secret goal. To do this, they have goals, rules, and other features, but the core loop is very simple:
When the brain generates good feelings, it usually has reasons for doing that, which a game designer has to be aware of. If you keep trying to make it generate good feelings without respecting the deeper purposes of the source of the feelings, afaik it generally stops working after a bit.
My aspiration is to make games that are compatible with living in real life. It's a large underserved market.
I don't think the assumption of equal transaction costs holds. If I want to fill in some potholes on my street, I can go door to door and ask for donations - which costs me time but has minimal and well understood costs to the other contributors. If I have to add "explain this new thing" and "keep track of escrow funds" and "cycling back and telling everyone how the project funding is going, and making them re-decide if/how much to contribute" that is a whole bunch of extra costs.
Also, of the public good is not quantum (eg, I could fix anywhere from 1-10 o...
A clarification:
Consider the premises (with scare quotes indicating technical jargon):
The original poster here is questioning statement 1, presenting evidence that "good" people act in bad faith too often for it to be evidence of "evil."
However, I belive the original poster is using a more broad definition of "Acting in Bad Faith" than the people who support premise 1.
That definition, concisely, would be "engaging in behavior that is recognized in context a...
Just don't use the term "conspiracy theory" to describe a theory about a conspiracy. Popular culture has driven "false" into the definition of that term, and wishful appeals to bare text doesn't make that connection go away. It hurts that some terms are limited in usability, but the burden of communication falls on the writer.
Setting aside the object level question here, trying to redefine words in order to avoid challenging connotations is a way to go crazy.
If someone is theorizing about a conspiracy, that's a conspiracy theory by plain meaning of the words. If it's also true, then the connotation about conspiracy theories being false is itself at least partly false.
The point is to recognize that it does belong in the same class, and how accurate/strong those connotations are for this particular example of that reference class, and letting connotations shift to match as ...
The innocent explanation is that the SS got back to him just before some sort of 90 day deadline, and he did the math. In which case the tweet could have been made out of ignorance, like flashing an "OK" sign in the "White Power" orientation. It's not easy to keep up with all the dog whistles out there.
Still political malpractice to not track and avoid those signals, though. If you "accidentally" have a rainbow in the background of a campaign photo, that counts as aligning with the LGBTQ+ crowd - same thing with putting "88" in a campaing post & Natzis.
So, the tweet aligns hos campaign with the Natzis, but might have done it accidentally.
Neurotypicals have weaker preferences regarding textures and other sensory inputs. By and large, they would not write, read, or expect others to be interested in a blow-by-blow of asthetics. Also, at a meta level, the very act of writing down specifics about a thing is not neurotypical. Contrast this post with the equivalent presentation in a mainstream magazine. The same content would be covered via pictures, feeling words, and generalities, with specific products listed in a footnote or caption, if at all. Or consider what your neurotypical friend's Face...
The real answer is that you should minimize the risk that you walk away and leave the door open for hours, and open it zero times whenever possible. The relative heat loss from 1 vs many separate openings is not significantly different from each-other, but it is much more than 0, and the tail risk of "all the food gets warm and spoils" should dominate the decisions
I don't thunk your model is correct. Opening the fridge causes the accumulated cold air to fall out over a period of a few (maybe 4-7?) seconds, after which it doesn't really matter how long you leave it open, as the air is all room temp. The stuff will slowly take heat from the room temp air, at a rate of about 1 degree/minute. Once the door is closed, it takes a few minutes (again, IDK how long) to get the air back to 40F, and then however long to extract the heat from the stuff. If you are chosing between "stand there with it open" and "take something o...
It's not. The original Nash construction is that player N picks a strategy that maximizes thier utility, assuming all other players get to know what N picked, and then pick a strategy that maximizes thier own utility given that. Minimax as a goal is only valid for atomic game actions, not complex strategies - Specifically because of this "trap"
There is a more fundamental objection: why would a set of 1s and 0s represent (given periodic repetition in 1/3 of the message, so dividing it into groups of 3 makes sense) specifically 3 frequencies of light and not
I think the key facility of am agent vs a calculator is the capability to create new short term goals and actions. A calculator (or water, or bacteria) can only execute the "programming" that was present when it was created. An agent can generate possible actions based on its environment, including options that might not even have existed when it was created.
I think even these first rough concepts have a distinction between beliefs and values. Even if the values are "hard coded" from the training period and the manual goal entry.
Being able to generate short term goals and execute them, and see if you are getting closer to your long tern goals is basically all any human does. It's a matter of scale, not kind, between me and a dolphin and AgentGPT.
In summary: Creating an agent was apparently already a solved problem, just missing a robust method of generating ideas/plans that are even vaguely possible.
Star Trek (and other Sci fi) continues to be surprisingly prescient, and "Computer, create an adversary capable of outwitting Data" creating an agen AI is actually completely realistic for 24th century technology.
Our only hopes are:
Years ago, I advocated banning crypto as a means of limiting the damage AI could do, thinking an advanced AI might be able to mine or hack exchanges (eg by guessing passwords of "lost" bitcoins) and accumulate wealth (ie power).
Apparent it could also just make a meme coin, and generate billions from nothing, given a sufficiently edgy coin (for example, AI itself)
https://www.axios.com/2025/01/19/donald-trump-crypto-billionaire
I am once again humbly suggesting that all un-regulated currency, especially distributed ledgers, be banned worldwide as a precautionary measure.
Another example that reality (especially anything involving technology) is not constrained by the need to be realistic. What SF author would dare write a story with meme coins, much less one in which the meme coins involved AIs like Claude?