Brain Speed Test
I've been experimenting with various nootropics lately, and wanted some way to have a frame of reference as to whether they work or not. So I put together this little tool:
It's pretty rudimentary for the time being, but I'm definitely open to feedback on 1. Ways to improve it, 2. Different tests you'd like to see.
A collection of Stubs.
In light of SDR's comment yesterday, instead of writing a new post today I compiled my list of ideas I wanted to write about, partly to lay them out there and see if any stood out as better than the rest, and partly so that maybe they would be a little more out in the wild than if I hold them until I get around to them. I realise there is not a thesis in this post, but I figured it would be better to write one of these than to write each in it's own post with the potential to be good or bad.
Original post: http://bearlamp.com.au/many-draft-concepts/
I create ideas at about the rate of 3 a day, without trying to. I write at about a rate of 1.5 a day. Which leaves me always behind. Even if I write about the best ideas I can think of, some good ones might never be covered. This is an effort to draft out a good stack of them so that maybe it can help me not have to write them all out, by better defining which ones are the good ones and which ones are a bit more useless.
With that in mind, in no particular order - a list of unwritten posts:
From my old table of contents
Goals of your lesswrong group – As a guided/workthrough exercise in deciding why the group exists and what it should do. Help people work out what they want out of it (do people know)? setting goals, doing something particularly interesting or routine, having fun, changing your mind, being activists in the world around you. Whatever the reasons you care about, work them out and move towards them. Nothing particularly groundbreaking in the process here. Sit down with the group with pens and paper, maybe run a resolve cycle, maybe talk about ideas and settle on a few, then decide how to carry them out. Relevant links: Sydney meetup, group resources (estimate 2hrs to write)
Goals interrogation + Goal levels – Goal interrogation is about asking <is this thing I want to do actually a goal of mine> and <is my current plan the best way to achieve that>, goal levels are something out of Sydney Lesswrong that help you have mutual long term goals and supporting short term goal. There are 3 main levels, Dream, Year, Daily (or approximate) you want dream goals like going to the moon, you want yearly goals like getting another year further in your degree and you want daily goals like studying today that contribute to the upper level goals. Any time you are feeling lost you can look at the guide you set out for yourself and use it to direct you. (3hrs)
How to human – A zero to human guide. A guide for basic functionality of a humanoid system. Something of a conglomeration of maslow, mental health, so you feel like shit and system thinking. Am I conscious?Am I breathing? Am I bleeding or injured (major or minor)? Am I falling or otherwise in danger and about to cause the earlier questions to return false? Do I know where I am? Am I safe? Do I need to relieve myself (or other bodily functions, i.e. itchy)? Have I had enough water? sleep? food? Is my mind altered (alcohol or other drugs)? Am I stuck with sensory input I can't control (noise, smells, things touching me)? Am I too hot or too cold? Is my environment too hot or too cold? Or unstable? Am I with people or alone? Is this okay? Am I clean (showered, teeth, other personal cleaning rituals)? Have I had some sunlight and fresh air in the past few days? Have I had too much sunlight or wind in the past few days? Do I feel stressed? Okay? Happy? Worried? Suspicious? Scared? Was I doing something? What am I doing? do I want to be doing something else? Am I being watched (is that okay?)? Have I interacted with humans in the past 24 hours? Have I had alone time in the past 24 hours? Do I have any existing conditions I can run a check on - i.e. depression? Are my valuables secure? Are the people I care about safe? (4hrs)
List of common strategies for getting shit done – things like scheduling/allocating time, pomodoros, committing to things externally, complice, beeminder, other trackers. (4hrs)
List of superpowers and kryptonites – when asking the question “what are my superpowers?” and “what are my kryptonites?”. Knowledge is power; working with your powers and working out how to avoid your kryptonites is a method to improve yourself. What are you really good at, and what do you absolutely suck at and would be better delegating to other people. The more you know about yourself, the more you can do the right thing by your powers or weaknesses and save yourself troubles.
List of effective behaviours – small life-improving habits that add together to make awesomeness from nothing. And how to pick them up. Short list: toothbrush in the shower, scales in front of the fridge, healthy food in the most accessible position in the fridge, make the unhealthy stuff a little more inacessible, keep some clocks fast - i.e. the clock in your car (so you get there early), prepare for expected barriers ahead of time (i.e. packing the gym bag and leaving it at the door), and more.
Stress prevention checklist – feeling off? You want to have already outsourced the hard work for “things I should check on about myself” to your past self. Make it easier for future you. Especially in the times that you might be vulnerable. Generate a list of things that you want to check are working correctly. i.e. did I drink today? Did I do my regular exercise? Did I take my medication? Have I run late today? Do I have my work under control?
Make it easier for future you. Especially in the times that you might be vulnerable. – as its own post in curtailing bad habits that you can expect to happen when you are compromised. inspired by candy-bar moments and turning them into carrot-moments or other more productive things. This applies beyond diet, and might involve turning TV-hour into book-hour (for other tasks you want to do instead of tasks you automatically do)
A p=np approach to learning – Sometimes you have to learn things the long way; but sometimes there is a short cut. Where you could say, “I wish someone had just taken me on the easy path early on”. It’s not a perfect idea; but start looking for the shortcuts where you might be saying “I wish someone had told me sooner”. Of course the answer is, “but I probably wouldn’t have listened anyway” which is something that can be worked on as well. (2hrs)
Rationalists guide to dating – Attraction. Relationships. Doing things with a known preference. Don’t like unintelligent people? Don’t try to date them. Think first; then act - and iteratively experiment; an exercise in thinking hard about things before trying trial-and-error on the world. Think about places where you might meet the kinds of people you want to meet, then use strategies that go there instead of strategies that flop in the general direction of progress. (half written)
Training inherent powers (weights, temperatures, smells, estimation powers) – practice makes perfect right? Imagine if you knew the temperature always, the weight of things by lifting them, the composition of foods by tasting them, the distance between things without measuring. How can we train these, how can we improve. Probably not inherently useful to life, but fun to train your system 1! (2hrs)
Strike to the heart of the question. The strongest one; not the one you want to defeat – Steelman not Strawman. Don’t ask “how do I win at the question”; ask, “am I giving the best answer to the best question I can give”. More poetic than anything else - this post would enumerate the feelings of victory and what not to feel victorious about, as well as trying to feel what it's like to be on the other side of the discussion to yourself, frustratingly trying to get a point across while a point is being flung at yourself. (2hrs)
How to approach a new problem – similar to the “How to solve X” post. But considerations for working backwards from a wicked problem, as well as trying “The least bad solution I know of”, Murphy-jitsu, and known solutions to similar problems. Step 0. I notice I am approaching a problem.
Turning Stimming into a flourish – For autists, to make a presentability out of a flaw.
How to manage time – estimating the length of future tasks (and more), covered in notch system, and do tasks in a different order. But presented on it's own.
Spices – Adventures in sensory experience land. I ran an event of spice-smelling/guessing for a group of 30 people. I wrote several documents in the process about spices and how to run the event. I want to publish these. As an exercise - it's a fun game of guess-the-spice.
Wing it VS Plan – All of the what, why, who, and what you should do of the two. Some people seem to be the kind of person who is always just winging it. In contrast, some people make ridiculously complicated plans that work. Most of us are probably somewhere in the middle. I suggest that the more of a planner you can be the better because you can always fall back on winging it, and you probably will. But if you don't have a plan and are already winging it - you can't fall back on the other option. This concept came to me while playing ingress, which encourages you to plan your actions before you make them.
On-stage bias – The changes we make when we go onto a stage include extra makeup to adjust for the bright lights, and speaking louder to adjust for the audience which is far away. When we consider the rest of our lives, maybe we want to appear specifically X (i.e, confident, friendly) so we should change ourselves to suit the natural skews in how we present based on the "stage" we are appearing on. appear as the person you want to appear as, not the person you naturally appear as.
Creating a workspace – considerations when thinking about a “place” of work, including desk, screen, surrounding distractions, and basically any factors that come into it. Similar to how the very long list of sleep maintenance suggestions covers environmental factors in your sleep environment but for a workspace.
Posts added to the list since then
Doing a cost|benefit analysis - This is something we rely on when enumerating the options and choices ahead of us, but something I have never explicitly looked into. Some costs that can get overlooked include: Time, Money, Energy, Emotions, Space, Clutter, Distraction/Attention, Memory, Side effects, and probably more. I'd like to see a How to X guide for CBA. (wikipedia)
Extinction learning at home - A cross between intermittent reward (the worst kind of addiction), and what we know about extinguishing it. Then applying that to "convincing" yourself to extinguish bad habits by experiential learning. Uses the CFAR internal Double Crux technique, precommit yourself to a challenge, for example - "If I scroll through 20 facebook posts in a row and they are all not worth my time, I will be convinced that I should spend less time on facebook because it's not worth my time" Adjust 20 to whatever position your double crux believes to be true, then run a test and iterate. You have to genuinely agree with the premise before running the test. This can work for a number of committed habits which you want to extinguish. (new idea as at the writing of this post)
How to write a dating ad - A suggestion to include information that is easy to ask questions about (this is hard). For example; don't write, "I like camping", write "I like hiking overnight with my dog", giving away details in a way that makes them worth inquiring about. The same reason applies to why writing "I'm a great guy" is really not going to get people to believe you, as opposed to demonstrating the claim. (show, don't tell)
How to give yourself aversions - an investigation into aversive actions and potentially how to avoid collecting them when you have a better understanding of how they happen. (I have not done the research and will need to do that before publishing the post)
How to give someone else an aversion - similar to above, we know we can work differently to other people, and at the intersection of that is a misunderstanding that can leave people uncomfortable.
Lists - Creating lists is a great thing, currently in draft - some considerations about what lists are, what they do, what they are used for, what they can be used for, where they come in handy, and the suggestion that you should use lists more. (also some digital list-keeping solutions)
Choice to remember the details - this stems from choosing to remember names, a point in the conversation where people sometimes tune out. As a mindfulness concept you can choose to remember the details. (short article, not exactly sure why I wanted to write about this)
What is a problem - On the path of problem solving, understanding what a problem is will help you to understand how to attack it. Nothing more complicated than this picture to explain it. The barrier is a problem. This doesn't seem important on it's own but as a foundation for thinking about problems it's good to have sitting around somewhere.
How to/not attend a meetup - for anyone who has never been to a meetup, and anyone who wants the good tips on etiquette for being the new guy in a room of friends. First meetup: shut up and listen, try not to be too much of an impact on the existing meetup group or you might misunderstand the culture.
Noticing the world, Repercussions and taking advantage of them - There are regularly world events that I notice. Things like the olympics, Pokemon go coming out, the (recent) spaceX rocket failure. I try to notice when big events happen and try to think about how to take advantage of the event or the repercussions caused by that event. Motivated to think not only about all the olympians (and the fuss leading up to the olympics), but all the people at home who signed up to a gym because of the publicity of the competitive sport. If only I could get in on the profit of gym signups...
leastgood but only solution I know of - So you know of a solution, but it's rubbish. Or probably is. Also you have no better solutions. Treat this solution as the best solution you have (because it is) and start implementing it, as you do that - keep looking for other solutions. But at least you have a solution to work with!
Self-management thoughts - When you ask yourself, "am I making progress?", "do I want to be in this conversation?" and other self management thoughts. And an investigation into them - it's a CFAR technique but their writing on the topic is brief. (needs research)
instrumental supply-hoarding behaviour - A discussion about the benefits of hoarding supplies for future use. Covering also - what supplies are not a good idea to store, and what supplies are. Maybe this will be useful for people who store things for later days, and hopefully help to consolidate and add some purposefulness to their process.
list of sub groups that I have tried - Before running my local lesswrong group I partook in a great deal of other groups. This was meant as a list with comments on each group.
If you have nothing to do – make better tools for use when real work comes along - This was probably going to be a poetic style motivation post about exactly what the title suggests. Be Prepared.
what other people are good at (as support) - When reaching out for support, some people will be good at things that other people are not. For example - emotional support, time to spend on each other, ideas for solving your problems. Different people might be better or worse than others. Thinking about this can make your strategies towards solving your problems a bit easier to manage. Knowing what works and what does not work, or what you can reliably expect when you reach out for support from some people - is going to supercharge your fulfilment of those needs.
Focusing - An already written guide to Eugine Gendlin's focusing technique. That needs polishing before publishing. The short form: treat your system 1 as a very powerful machine that understands your problems and their solutions more than you do; use your system 2 to ask it questions and see what it returns.
Rewrite: how to become a 1000 year old vampire - I got as far as breaking down this post and got stuck at draft form before rewriting. Might take another stab at it soon.
Should you tell people your goals? - This thread in a post. In summary: It depends on the environment, the wrong environment is actually demotivational, the right environment is extra motivational.
Meta: this took around 4 hours to write up. Which is ridiculously longer than usual. I noticed a substantial number of breaks being taken - not sure if that relates to the difficulty of creating so many summaries or just me today. Still. This experiment might help my future writing focus/direction so I figured I would try it out. If you see an idea of particularly high value I will be happy to try to cover it in more detail.
Crude measures
A putative new idea for AI control; index here.
Partially inspired by as conversation with Daniel Dewey.
People often come up with a single great idea for AI, like "complexity" or "respect", that will supposedly solve the whole control problem in one swoop. Once you've done it a few times, it's generally trivially easy to start taking these ideas apart (first step: find a bad situation with high complexity/respect and a good situation with lower complexity/respect, make the bad very bad, and challenge on that). The general responses to these kinds of idea are listed here.
However, it seems to me that rather than constructing counterexamples each time, we should have a general category and slot these ideas into them. And not only have a general category with "why this can't work" attached to it, but "these are methods that can make it work better". Seeing the things needed to make their idea better can make people understand the problems, where simple counter-arguments cannot. And, possibly, if we improve the methods, one of these simple ideas may end up being implementable.
Crude measures
The category I'm proposing to define is that of "crude measures". Crude measures are methods that attempt to rely on non-fully-specified features of the world to ensure that an underdefined or underpowered solution does manage to solve the problem.
To illustrate, consider the problem of building an atomic bomb. The scientists that did it had a very detailed model of how nuclear physics worked, the properties of the various elements, and what would happen under certain circumstances. They ended up producing an atomic bomb.
The politicians who started the project knew none of that. They shovelled resources, money and administrators at scientists, and got the result they wanted - the Bomb - without ever understanding what really happened. Note that the politicians were successful, but it was a success that could only have been achieved at one particular point in history. Had they done exactly the same thing twenty years before, they would not have succeeded. Similarly, Nazi Germany tried a roughly similar approach to what the US did (on a smaller scale) and it went nowhere.
So I would define "shovel resources at atomic scientists to get a nuclear weapon" as a crude measure. It works, but it only works because there are other features of the environment that are making it work. In this case, the scientists themselves. However, certain social and human features about those scientists (which politicians are good at estimating) made it likely to work - or at least more likely to work than shovelling resources at peanut-farmers to build moon rockets.
In the case of AI, advocating for complexity is similarly a crude measure. If it works, it will work because of very contingent features about the environment, the AI design, the setup of the world etc..., not because "complexity" is intrinsically a solution to the FAI problem. And though we are confident that human politicians have some good enough idea about human motivations and culture that the Manhattan project had at least some chance of working... we don't have confidence that those suggesting crude measures for AI control have a good enough idea to make their idea works.
It should be evident that "crudeness" is on a sliding scale; I'd like to reserve the term for proposed solutions to the full FAI problem that do not in any way solve the deep questions about FAI.
More or less crude
The next question is, if we have a crude measure, how can we judge its chance of success? Or, if we can't even do that, can we at least improve the chances of it working?
The main problem is, of course, that of optimising. Either optimising in the sense of maximising the measure (maximum complexity!) or of choosing the measure that is most extreme fit to the definition (maximally narrow definition of complexity!). It seems we might be able to do something about this.
Let's start by having AI create sample a large class of utility functions. Require them to be around the same expected complexity as human values. Then we use our crude measure μ - for argument's sake, let's make it something like "approval by simulated (or hypothetical) humans, on a numerical scale". This is certainly a crude measure.
We can then rank all the utility functions u, using μ to measure the value of "create M(u), a u-maximising AI, with this utility function". Then, to avoid the problems with optimisation, we could select a certain threshold value and pick any u such that E(μ|M(u)) is just above the threshold.
How to pick this threshold? Well, we might have some principled arguments ("this is about as good a future as we'd expect, and this is about as good as we expect that these simulated humans would judge it, honestly, without being hacked").
One thing we might want to do is have multiple μ, and select things that score reasonably (but not excessively) on all of them. This is related to my idea that the best Turing test is one that the computer has not been trained or optimised on. Ideally, you'd want there to be some category of utilities "be genuinely friendly" that score higher than you'd expect on many diverse human-related μ (it may be better to randomly sample rather than fitting to precise criteria).
You could see this as saying that "programming an AI to preserve human happiness is insanely dangerous, but if you find an AI programmed to satisfice human preferences, and that other AI also happens to preserve human happiness (without knowing it would be tested on this preservation), then... it might be safer".
There are a few other thoughts we might have for trying to pick a safer u:
- Properties of utilities under trade (are human-friendly functions more or less likely to be tradable with each other and with other utilities)?
- If we change the definition of "human", this should have effects that seem reasonable for the change. Or some sort of "free will" approach: if we change human preferences, we want the outcome of u to change in ways comparable with that change.
- Maybe also check whether there is a wide enough variety of future outcomes, that don't depend on the AI's choices (but on human choices - ideas from "detecting agents" may be relevant here).
- Changing the observers from hypothetical to real (or making the creation of the AI contingent, or not, on the approval), should not change the expected outcome of u much.
- Making sure that the utility u can be used to successfully model humans (therefore properly reflects the information inside humans).
- Make sure that u is stable to general noise (hence not over-optimised). Stability can be measured as changes in E(μ|M(u)), E(u|M(u)), E(v|M(u)) for generic v, and other means.
- Make sure that u is unstable to "nasty" noise (eg reversing human pain and pleasure).
- All utilities in a certain class - the human-friendly class, hopefully - should score highly under each other (E(u|M(u)) not too far off from E(u|M(v))), while the over-optimised solutions - those scoring highly under some μ - must not score high under the class of human-friendly utilities.
This is just a first stab at it. It does seem to me that we should be able to abstractly characterise the properties we want from a friendly utility function, which, combined with crude measures, might actually allow us to select one without fully defining it. Any thoughts?
And with that, the various results of my AI retreat are available to all.
Tools want to become agents
In the spirit of "satisficers want to become maximisers" here is a somewhat weaker argument (growing out of a discussion with Daniel Dewey) that "tool AIs" would want to become agent AIs.
The argument is simple. Assume the tool AI is given the task of finding the best plan for achieving some goal. The plan must be realistic and remain within the resources of the AI's controller - energy, money, social power, etc. The best plans are the ones that use these resources in the most effective and economic way to achieve the goal.
And the AI's controller has one special type of resource, uniquely effective at what it does. Namely, the AI itself. It is smart, potentially powerful, and could self-improve and pull all the usual AI tricks. So the best plan a tool AI could come up with, for almost any goal, is "turn me into an agent AI with that goal." The smarter the AI, the better this plan is. Of course, the plan need not read literally like that - it could simply be a complicated plan that, as a side-effect, turns the tool AI into an agent. Or copy the AI's software into a agent design. Or it might just arrange things so that we always end up following the tool AIs advice and consult it often, which is an indirect way of making it into an agent. Depending on how we've programmed the tool AI's preferences, it might be motivated to mislead us about this aspect of its plan, concealing the secret goal of unleashing itself as an agent.
In any case, it does us good to realise that "make me into an agent" is what a tool AI would consider the best possible plan for many goals. So without a hint of agency, it's motivated to make us make it into a agent.
Best causal/dependency diagram software for fluid capture?
I've found most graphing software too clunky, or having too much mental friction, for my purpose of creating graphically represented plans, to convert written diagrams into digital form, or to do preference inference based on the structure of my goals (amongst other things).
So far the only tool that I've seen that reduces this friction is GraphViz [1], since I think I can literally just list down connection after connection in markup, with no care for structure or reasonableness, and then prune connections after I see how the entire thing looks. Point and click is for suckers.
However, I also like the approach of Freemind that quickly outputs a visual map that is easily traversable; but it doesn't do much for me when the causality is more involved.
Are there any alternatives that anyone is aware of?
[1] If you are not familiar with GraphViz, see this amusing introduction that maps the social network in R. Kelly's hit hip hopera, "Trapped in the Closet".
Tool/Agent distinction in the light of the AI box experiment
This article poses questions on the distinction between Tool AGI and Agent AGI, which was described very concisely by Holden Karnofsky in his recent Thoughts on the Singularity Institute post:
In short, Google Maps is not an agent, taking actions in order to maximize a utility parameter. It is a tool, generating information and then displaying it in a user-friendly manner for me to consider, use and export or discard as I wish.
For me, this instantly raised one question: What if a Tool AGI becomes/is self-aware (which, for the purposes of this post, I define as “able to have goals that are distinct from the goals of the outside world”) and starts manipulating its results in a way that is non-obvious to its user? Or, even worse: What if the Tool AGI makes its user do things (which I do not expect to be much more difficult than succeding in the AI box experiment)?
My first reaction was to flinch away by telling myself: “But of course a Tool would never become self-aware! Self-awareness is too complex to just happen unintentionally!”
But some uncertainty survived and was strenghtened by Eliezer's reply to Holden:
[Tool AGI] starts sounding much scarier once you try to say something more formal and internally-causal like "Model the user and the universe, predict the degree of correspondence between the user's model and the universe, and select from among possible explanation-actions on this basis."
After all, “Self-awareness is too complex to just happen unintentionally!” is just a bunch of English words expressing my personal incredulity. It's not a valid argument.
So, can we make the argument, that self-awareness will not happen unintentionally?
If we can't make that argument, can we stop Tool AGIs from potentially becoming a Weak Agent AGI which acts through its human user?
If we can't do that, how meaningful is the distinction between a Weak Agent AGI (a.k.a. Tool AGI) and an Agent AGI?
For more, see the Tools versus Agents post by Stuart_Armstrong, which points to similar questions.
Tools versus agents
In his critique of the Singularity Institute, Holden Karnofsky presented a distinction between an AI functioning as a tool versus one functioning as an agent. In his words, a tool AI would
(1) Calculate which action A would maximize parameter P, based on existing data set D. (2) Summarize this calculation in a user-friendly manner, including what Action A is, what likely intermediate outcomes it would cause, what other actions would result in high values of P, etc.
In contrast, an agent AI would:
(1) Calculate which action, A, would maximize parameter P, based on existing data set D. (2) Execute Action A.
The idea being that an AI, asked to "prevent human suffering", would come up with two plans:
- Kill all human.
- Cure all diseases, make everyone young and immortal.
Then the agent AI would go out and kill everyone, while the tool AI would give us the list and we would pick the second one. In the following, I'll assume the AI is superintelligent, and has no other objectives than what we give it.
= 783df68a0f980790206b9ea87794c5b6)

Subscribe to RSS Feed
= f037147d6e6c911a85753b9abdedda8d)