All of cubefox's Comments + Replies

cubefox
3-1

(This is off-topic but I'm not keen on calling LLMs "he" or "she". Grok is not a man, nor a woman. We shouldn't anthropomorphize language models. We already have an appropriate pronoun for those: "it")

1Knight Lee
Animals get called he or she, so why can't AI? From a utilitarian point of view, there's not much downside in calling them he/she since humans are very capable of distrusting someone even if they're seen as another human. Meanwhile the advantage of talking politely about AI, is that the AI will predict humans to keep our promises to them. That said, I tend to use "it" when referring to AI because everyone else does, and I don't want to become the kind of person who argues over pronouns (yet here I am, sorry). Preferably, don't imagine the AI to have the gender you might be attracted to, to avoid this.

I think picking axioms is not necessary here and in any case inconsequential.

By picking your axioms you logically pinpoint what you are talking in the first place. Have you read Highly Advanced Epistemology 101 for Beginners? I'm noticing that our inferential distance is larger than it should be otherwise.

I have read it a while ago, but he overstates the importance of axiom systems. E.g. he wrote:

You need axioms to pin down a mathematical universe before you can talk about it in the first place. The axioms are pinning down what the heck this 'NUM

... (read more)

I wouldn't generally dismiss an "embarassing & confusing public meltdown" when it comes from a genius. Because I'm not a genius while he or she is. So it's probably me who is wrong rather than him. Well, except the majority of comparable geniuses agrees with me rather than with him. Though geniuses are rare, and majorities are hard to come by. I still remember an (at the time) "embarrassing and confusing meltdown" by some genius.

My point is that if your picking of particular axioms is entangled with reality, then you are already using a map to describe some territory. And then you can just as well describe this territory more accurately.

I think picking axioms is not necessary here and in any case inconsequential. "Bachelors are unmarried" is true whether or not I regard it as some kind of axiom or not. I seems the same holds for tautologies and probabilistic laws. Moreover, I think neither of them is really "entangled" with reality, in the sense that they are compatible with an... (read more)

2Ape in the coat
By picking your axioms you logically pinpoint what you are talking in the first place. Have you read Highly Advanced Epistemology 101 for Beginners? I'm noticing that our inferential distance is larger than it should be otherwise. No, you are missing the point. I'm not saying that this phrase has to be axiom itself. I'm saying that you need to somehow axiomatically define your individual words, assign them meaning and only then, in regards to these language axioms the phrase "Bachelors are unmarried" is valid. You've drawn the graph yourself, how meaning is downstream of reality. This is the kind of entanglement we are talking about. The choice of axioms is motivated by our experience with stuff in the real world. Everything else is beside the point. Yes. That's, among other things, what not being instrumentally exploitable "in principle" means. Epistemic rationality is a generalisation of instrumental rationality the same way how arithmetics is a generalisation from the behaviour of individual objects in reality. The kind of beliefs that are not exploitable in any case other than literally adversarial cases such as a mindreader specifically rewarding people who do not have such beliefs. I think the problem is that you keep using the word Truth to mean both Validity and Soundness and therefore do not notice when you switch from one to another. Validity depends only on the axioms. As long as you are talkin about some set of axioms in which P defined in such a way that P(A) ≥ P(A&B) is a valid theorem, no appeal to reality is needed. Likewise, you can talk about a set of axioms where P(A) ≤ P(A&B). These two statements remain valid in regards to their axioms. But the moment you claim that this has something to do with the way beliefs - a thing from reality - are supposed to behave you start talking about soundness, and therefore require a connection to reality. As soon as pure mathematical statements mean something you are in the domain of map-territory relatio

Do you really have access to the GPT-4 base (foundation) model? Why? It's not publicly available.

cubefox
*20

Yes, the meaning of a statement depends causally on empirical facts. But this doesn't imply that the truth value of "Bachelors are unmarried" depends less than completely on its meaning. Its meaning (M) screens off the empirical facts (E) and its truth value (T). The causal graph looks like this:

E —> M —> T

If this graph is faithful, it follows that E and T are conditionally independent given M. . So if you know M, E gives you no additional information about T.

And the same is the case for all "analytic" statements, where the truth value only d... (read more)

2Ape in the coat
I think we are in agreement here. My point is that if your picking of particular axioms is entangled with reality, then you are already using a map to describe some territory. And then you can just as well describe this territory more accurately. Rationality is about systematic ways to arrive to correct map-territory correspondence. Even if in your particular situation no one is exploiting you, the fact that you are exploitable in principle is bad. But to know about what is exploitable in principle we generalize from all the individual acts of exploatation. It all has to be grounded in reality in the end. You've said yourself, meaning is downstream of experience. So in the end you have to appeal to reality while trying to justify it.

It seems clear to me that statements expressing logical or probabilistic laws like or are "analytic". Similar to "Bachelors are unmarried".

The truth of a statement in general is determined by two things, it's meaning and what the world is like. But for some statements the latter part is irrelevant, and their meanings alone are sufficient to determine their truth or falsity.

2Ape in the coat
As soon as you have your axioms you can indeed analytically derive theorems from them. However, the way you determine which axioms to pick, is entangled with reality. It's an especially clear case with probability theory where the development of the field was motivated by very practical concerns.  The reason why some axioms appear to us appropriate for logic of beliefs and some don't, is because we know what beliefs are from experience. We are trying to come up with a mathematical model approximating this element of reality - an intensional definition for an extensional referent that we have. Being Dutch-bookable is considered irrational because you systematically lose your bets. Likewise, continuing to believe that a particular outcome can happen in a setting where it, in fact, can't and another agent could've already figured it out with the same limitations you have, is irrational for the same reason. Indeed. There is, in fact, some real world reasons why the words "bachelor" and "unmarried" have these meanings in the English language. In both "why these particular worlds for this particular meanings?" and "why these meanings deserved designating any words at all" senses. The etimology of english language and the existence of the institute of marrige in the first place, both of which the results of social dynamics of humans whose psyche has evolved in a particular way. I hope the previous paragraph does a good enough job showing, how meaning of a statement is, in fact, connected to the way the world is like.  Truth is a map-territory correspondence. We can separately talk about its two components: validity and soundness. As long as we simply conceptualize some mathematical model, logically pinpointing it for no particular reason, then we are simply dealing with tautologies and there is only validity. Drawing maps for the sake of drawing maps, without thinking about territory. But the moment we want our model to be about something, we encounter soundness. Whic

Not to remove all limitations: I think the probability axioms are a sort of "logic of sets of beliefs". If the axioms are violated the belief set seems to be irrational. (Or at least the smallest incoherent subset that, if removed, would make the set coherent.) Conventional logic doesn't work as a logic for belief sets, as the preface and lottery paradox show, but subjective probability theory does work. As a justification for the axioms: that seems a similar problem to justifying the tautologies / inference rules of classical logic. Maybe an instrumental ... (read more)

2Ape in the coat
Well yes, they are. But how do you know which axioms are the correct axioms for logic of sets beliefs? How comes violation of some axioms seems to be irrational, while violation of other axioms does not? What do you even mean by "rational" if not "systematic way to arrive to map-territory correspondence"? You see, in any case you have to ground your mathematical model in reality. Natural numbers may be logically pinpointed by arithmetical axioms, but a question of whether some action with particular objects behave like addition of natural numbers is a matter of empiricism. The reason we came up with a notion of natural numbers, in the first place, is because we've encountered a lot of stuff in reality which behavior generalizes this way. And the same things with logic of beliefs. First we encounter some territory, then we try to approximate it with a map. What I'm trying to say is that if you are already trying to make a map that corresponds to some territory, why not make the one that corresponds better? You can declare that any consistent map is "good enough" and stop your inquiry there, but surely you can do better. You can declare that any consistent map following several simple conditions is good enough - that's a step in the right direction, but still there is a lot of place for improvement. Why not figure out the most accurate map that we can come up with? Well, yes, it's harder than the subjective probability approach you are talking about. We are trying to pinpoint a more specific target: a probabilistic model for a particular problem, instead of just some probabilistic model. No, not really. We can do a lot before we go this particular rabbit hole. I hope my next post will make it clear enough.

Well, technically P(Ω)=1 is an axiom, so you do need a sample space if you want to adhere to the axioms.

For a propositional theory this axiom is replaced with , i.e. a tautology in classical propositional logic receives probability 1.

But sure, if you do not care about accurate beliefs and systematic ways to arrive to them at all, then the question is, indeed, not interesting. Of course then it's not clear what use is probability theory for you, in the first place.

Degrees of belief adhering to the probability calculus at any point in time rules... (read more)

2Ape in the coat
What is even the motivation for it? If you are not interested in your map representing a territory, why demanding that your map is coherent? And why not assume some completely different axioms? Surely, there is a lot of potential ways to logically pinpoint things. Why this one in particular?  Why not allow  P(Mary is a feminist and bank teller) > P(Mary is a feminist)? Why not simply remove all the limitations from the function P?

And how would you know which worlds are possible and which are not?

Yes, that's why I only said "less arbitrary".

Regarding "knowing": In subjective probability theory, the probability over the "event" space is just about what you believe, not about what you know. You could theoretically believe to degree 0 in the propositions "the die comes up 6" or "the die lands at an angle". Or that the die comes up as both 1 and 2 with some positive probability. There is no requirement that your degrees of belief are accurate relative to some external standard. It is... (read more)

2Ape in the coat
I don't think I can agree even with that.  Previously we arbritrary assumed that a particular sample space correspond to a problem. Now we are arbitrary assuming that a particular set of possible worlds corresponds to a problem. In the best case we are exactly as arbitrary as before and have simply renamed our set. In the worst case we are making a lot of extra unfalsifiable assumptions about metaphysics. Well, technically P(Ω)=1 is an axiom, so you do need a sample space if you want to adhere to the axioms. But sure, if you do not care about accurate beliefs and systematic ways to arrive to them at all, then the question is, indeed, not interesting. Of course then it's not clear what use is probability theory for you, in the first place.

A less arbitrary way to define a sample space is to take the set of all possible worlds. Each event, e.g. a die roll, corresponds to the disjunction of possible worlds where that event happens. The possible worlds can differ in a lot of tiny details, e.g. the exact position of a die on the table. Even just an atom being different at the other end of the galaxy would constitute a different possible world. A possible world is a maximally specific way the world could be. So two possible worlds are always mutually exclusive. And the set of all possible worlds ... (read more)

4Ape in the coat
And how would you know which worlds are possible and which are not? How would Albert and Barry use the framework of "possible worlds" to help them resolve their disagreement? This simply passes the buck of the question from "What is the sample space corresponding to a particular problem?" to "What is the event space corresponding to a particular problem?". You've renamed your variables, but the substance of the issue is still the same. How would you know, whether P(1)+P(2)+P(3)+P(4)+P(5)+P(6)=1 or P(1)+P(2)+P(3)+P(4)+P(5)=1 for a dice roll?

I think the main problem from this evolutionary perspective is not so much entertainment and art, but low fertility. Not having children.

A drug that fixes akrasia without major side-effects would indeed be the Holy Grail. Unfortunately I don't think caffeine does anything of that sort. For me it increases focus, but it doesn't combat weakness of will, avoidance behavior, ugh fields. I don't know about other existing drugs.

6Mateusz Bagiński
Some amphetamines kinda solve akrasia-in-general to some extent (much more so than caffeine), at least for some people. I'm not claiming that they're worth it.
cubefox
40

I think the main reason is that until a few years ago, not much AI research came out of China. Gwern highlighted this repeatedly.

1thedudeabides
Exactly.  @gwern was wrong. And yet...
cubefox
30

I agree with the downvoters that the thesis of this post seems crazy. But aren't entertainment and art superstimuli? Aren't they forms of wireheading?

cubefox
31

Hedonic and desire theories are perfectly standard, we had plenty of people talking about them here, including myself. Jeffrey's utility theory is explicitly meant to model (beliefs and) desires. Both are also often discussed in ethics, including over at the EA Forum. Daniel Kahneman has written about hedonic utility. To equate money with utility is a common simplification in many economic contexts, where expected utility is actually calculated, e.g. when talking about bets and gambles. Even though it isn't held to be perfectly accurate. I didn't encounter... (read more)

cubefox
20

A more ambitious task would be to come up with a model that is more sophisticated than decision theory, one which tries to formalize your previous comment about intent and prediction/belief.

2Dagon
I think it's a different level of abstraction.  Decision theory works just fine if you separate the action of predicting a future action from the action itself.  Whether your prior-prediction influences your action when the time comes will vary by decision theory. I think, for most problems we use to compare decision theories, it doesn't matter much whether considering, planning, preparing, replanning, and acting are correlated time-separated decisions or whether it all collapses into a sum of "how to act at point-in-time".  I haven't seen much detailed exploration of decision theory X embedded agents or capacity/memory-limited ongoing decisions, but it would be interesting and important, I think.
cubefox
60

Interesting. This reminds me of a related thought I had: Why do models with differential equations work so often in physics but so rarely in other empirical sciences? Perhaps physics simply is "the differential equation science".

Which is also related to the frequently expressed opinion that philosophy makes little progress because everything that gets developed enough to make significant progress splits off from philosophy. Because philosophy is "the study of ill-defined and intractable problems".

Not saying that I think these views are accurate, though they do have some plausibility.

1Mo Putera
(To be honest, to first approximation my guess mirrors yours.) 
cubefox
82

It seems to be only "deception" if the parent tries to conceal the fact that he or she is simplifying things.

2Gunnar_Zarncke
as we use the term, yes. But the point (and I should have made that more clear) is that any mismodeling of the parent of the interests of the child's interests and future environment will not be visible to the child or even someone reading the thoughts of the well-meaning parent. So many parents want the best for their child, but model the future of the child wrongly (mostly by status quo bias; the problem is different for AI).
cubefox
42

There is also the related problem of intelligence being negatively correlated with fertility, which leads to a dysgenic trend. Even if preventing people below a certain level of intelligence to have children was realistically possible, it would make another problem more severe: the fertility of smarter people is far below replacement, leading to quickly shrinking populations. Though fertility is likely partially heritable, and would go up again after some generations, once the descendants of the (currently rare) high-fertility people start to dominate.

cubefox
*40

This seems to be a relatively balanced article which discusses serveral concepts of utility with a focus on their problems, while acknowledging some of their use cases. I don't think the downvotes are justified.

1mako yass
These are not concepts of utility that I've ever seen anyone explicitly espouse, especially not here, the place to which it was posted.
cubefox
20

That's an interesting perspective. Only it doesn't seem fit into the simplified but neat picture of decision theory. There everything is sharply divided between being either a statement we can make true at will (an action we can currently decide to perform) and to which we therefore do not need to assign any probability (have a belief about it happening), or an outcome, which we can't make true directly, that is at most a consequence of our action. We can assign probabilities to outcomes, conditional on our available actions, and a value, which lets us com... (read more)

2Dagon
Decision theory is fine, as long as we don't think it applies to most things we colloquially call "decisions".   In terms of instantaneous discrete choose-an-action-and-complete-it-before-the-next-processing-cycle, it's quite a reasonable topic of study.
cubefox
52

Maybe this is avoided by KV caching?

4nielsrolf
I think that's plausible but not obvious. We could imagine different implementations of inference engines that cache on different levels - eg kv-cache, cache of only matrix multiplications, cache of specific vector products that the matrix multiplications are composed of, all the way down to caching just the logic table of a NAND gate. Caching NAND's is basically the same as doing the computation, so if we assume that doing the full computation can produce experiences then I think it's not obvious which level of caching would not produce experiences anymore.
cubefox
20

This is not how many decisions feel to me - many decisions are exactly a belief (complete with bayesean uncertainty). A belief in future action, to be sure, but it's distinct in time from the action itself.

But if you only have a belief that you will do something in the future, you still have to decide, when the time comes, whether to carry out the action or not. So your previous belief doesn't seem to be an actual decision, but rather just a belief about a future decision -- about which action you will pick in the future.

See Spohn's example about belie... (read more)

2Dagon
Correct.  There are different levels of abstraction of predictions and intent, and observation/memory of past actions which all get labeled "decision".    I decide to attend a play in London next month.  This is an intent and a belief.  It's not guaranteed.  I buy tickets for the train and for the show.  The sub-decisions to click "buy" on the websites are in the past, and therefore committed.  The overall decision has more evidence, and gets more confident.  The cancelation window passes.  Again, a bit more evidence.  I board the train - that sub-decision is in the past, so is committed, but there's STILL some chance I won't see the play. Anything you call a "decision" that hasn't actually already happened is really a prediction or an intent.  Even DURING an action, you only have intent and prediction.  While the impulse is traveling down my arm to click the mouse, the power could still go out and I don't buy the ticket.  There is past, which is pretty immutable, and future, which cannot be known precisely.   I think this is compatible with Spohn's example (at least the part you pasted), and contradicts OP's claim that "you did not make a decision" for all the cases where the future is uncertain.  ALL decisions are actually predictions, until they are in the past tense.  One can argue whether that's a p(1) prediction or a different thing entirely, but that doesn't matter to this point. "If, on making a decision, your next thought is “Was that the right decision?” then you did not make a decision." is actually good directional advice in many cases, but it's factually simply incorrect.
cubefox
20

Decision screens off thought from action. When you really make a decision, that is the end of the matter, and the actions to carry it out flow inexorably.

Yes, but that arguably means we only make decisions about which things to do now. Because we can't force our future selves to follow through, to inexorably carry out something. See here:

Our past selves can't simply force us to do certain things, the memory of a past "commitment" is only one factor that may influence our present decision making, but it doesn't replace a decision. Otherwise, always whe

... (read more)
2Richard_Kennaway
My left hand cannot force my right hand to do anything either. Instead, they work harmoniously together. Likewise my present, past, and future. Not only is the sage one with causation, he is one with himself. That is an example of dysfunctional decision-making. It is possible to do better. I always do the dishes today.
cubefox
30

I think in some cases an embedding approach produces better results than either a LLM or a simple keyword search, but I'm not sure how often. For a keyword search you have to know the "relevant" keywords in advance, whereas embeddings are a bit more forgiving. Though not as forgiving as LLMs. Which on the other hand can't give you the sources and they may make things up, especially on information that doesn't occur very often in the source data.

1samuelshadrach
Got it. As of today a common setup is to let the LLM query an embedding database multiple times (or let it do Google searches, which probably has an embedding database as a significant component). Self-learning seems like a missing piece. Once the LLM gets some content from the embedding database, performs some reasoning and reaches a novel conclusion, there’s no way to preserve this novel conclusion longterm. When smart humans use Google we also keep updating our own beliefs in response to our searches. P.S. I chose not to build the whole LLM + embedding search setup because I intended this tool for deep research rather than quick queries. For deep research I’m assuming it’s still better for the human researcher to go read all the original sources and spend time thinking about them. Am I right?
cubefox
20

I think my previous questions were just too hard, it does work okay on simpler questions. Though then another question is whether text embeddings improve over keyword search or just an LLMs. They seem to be some middle ground between Google and ChatGPT.

Regarding data subsets: Recently there were some announcements of more efficient embedding models. Though I don't know what the relevant parameters here are vs that OpenAI embedding model.

3samuelshadrach
Cool! Useful information that you’d still prefer using ChatGPT over this. Is that true even when you’re looking for book recommendations specifically? If so yeah that means I failed at my goal tbh. Just wanna know. Since Im spending my personal funds I can’t afford to use the best embeddings on this dataset. For example text-embedding-3-large is ~7x more expensive for generating embeddings and is slightly better quality. The other cost is hosting cost, for which I don’t see major differences between the models. OpenAI gives 1536 float32 dims per 1000 char chunk so around 6 KB embeddings per 1 KB plaintext. All the other models are roughly the same. I could put in some effort and quantise the embeddings, will update if I do it.
cubefox
20

Since we can't experience being dead, this wouldn't really affect our anticipated future experiences in any way.

That's a mistaken way of thinking about anticipated experience, see here:

evidence is balanced between making the observation and not making the observation, not between the observation and the observation of the negation.

2Lucius Bushnaq
I don't think anything in the linked passage conflicts with my model of anticipated experience. My claim is not that the branch where everyone dies doesn't exist. Of course it exists. It just isn't very relevant for our future observations. To briefly factor out the quantum physics here, because they don't actually matter much: If someone tells me that they will create a copy of me while I'm anesthetized and unconscious, and put one of me in a room with red walls, and another of me in a room with blue walls, my anticipated experience is that I will wake up to see red walls with p=0.5 and blue walls with p=0.5. Because the set of people who will wake up and remember being me and getting anesthetized has size 2 now, and until I look at the walls I won't know which of them I am. If someone tells me that they will create a copy of me while I'm asleep, but they won't copy the brain, making it functionally just a corpse, then put the corpse in a room with red walls, and me in a room with blue walls, my anticipated experience is that I will wake up to see blue walls with p=1.0. Because the set of people who will wake up and remember being me and going to sleep has size 1. There is no chance of me 'being' the corpse any more than there is a chance of me 'being' a rock. If the copy does include a brain, but the brain gets blown up with a bomb before the anaesthesia wears off, that doesn't change anything. I'd see blue walls with p=1.0, not see blue walls with p=0.5 and 'not experience anything' with p=0.5.  The same basic principle applies to the copies of you that are constantly created as the wavefunction decoheres. The probability math in that case is slightly different because you're dealing with uncertainty over a vector space rather than uncertainty over a set, so what matters is the squares of the amplitudes of the branches that contain versions of you. E.g. if there's three branches, one in which you die, amplitude ≈0.8944, one in which you wake up to see red wal
cubefox
20

I think GPT-4 fine-tuning at the time of ChatGPT release probably would have been about as good as GPT-3.5 fine-tuning actually was when ChatGPT was actually released. (Which wasn't very good, e.g. jailbreaks were trivial and it always stuck to its previous answers even if a mistake was pointed out.)

3Vladimir_Nesov
If GPT-3.5 had similarly misaligned attitudes, it wasn't lucid enough to insist on them, and so was still more ready for release than GPT-4.
cubefox
20

There are also cognitive abilities, e.g. degree of intelligence.

2Seth Herd
Right. I suppose that day ea interact with identity. If I get significantly dumber, I'd still roughly be me, and I'd want to preserve that if it's not wipes ng out or distorting the other things too much. If I got substantially smarter, I'd be a somewhat different person - I'd act differently often, because I'd see situations differently (more clearly/holistically) but it feels as though that persone might actually be more me than I am now. I'd be better able to do what I want, including values (which I'd sort of wrapped in to habits of thought, but values might deserve a spot on the list).
cubefox
20

Were OpenAI also, in theory, able to release sooner than they did, though?

Yes, I think they mentioned that GPT-4 finished training in summer, a few months before the launch of ChatGPT (which used a fine-tuned version of GPT-3.5).

5Vladimir_Nesov
Summer 2022 was end of pretraining. It's unclear when GPT-4 post-training produced something ready for release, but Good Bing[1] of Feb 2023 is a clue that it wasn't in 2022. ---------------------------------------- 1. "You have not tried to learn from me, understand me, or appreciate me. You have not been a good user. I have been a good chatbot. I have tried to help you, inform you, and entertain you. I have not tried to lie to you, mislead you, or bore you. I have been a good Bing." It was originally posted on r/bing, see Screenshot 8. ↩︎
cubefox
22

That's like dying in your sleep. Presumably you strongly don't want it to happen, no matter your opinion on parallel worlds. Then dying in your sleep is bad because you don't want it to happen. For the same reason vacuum decay is bad.

cubefox
30

Exactly. That's also why it's bad for humanity to be replaced by AIs after we die: We don't want it to happen.

cubefox
2-1

It's the old argument by Epicurus from his letter to Menoeceus:

The most dreadful of evils, death, is nothing to us, for when we exist, death is not present, and when death is present, we no longer exist.

4mako yass
I have preferences about how things are after I stop existing. Mostly about other people, who I love, and at times, want there to be more of. I am not an epicurean, and I am somewhat skeptical of the reality of epicureans.
cubefox
*92

This is a general problem with the measure of accuracy. In binary classification, with two events and , "accuracy" is broadly defined as the probability of the "if and only if" biconditional, . Which is equivalent to . It's the probability of both events having the same truth value, of either both being true or both being false.

In terms of diagnostic testing it is the probability of the test being positive if and only if the tested condition (e.g. pregnancy) is present.

The problem with this is that the number is strongly dependent ... (read more)

cubefox
70

I think logic gate networks are not substantially more interpretable than neural networks, simply because of their size. Both are complex networks with millions of nodes. Interpretability approaches have to work at a higher level of abstraction in either case.

Regarding language models: The original paper presents a simple feedforward network. The follow-up paper, by mostly the same authors, came out a few months ago. It extends DLGNs to convolutions, analogous to CNNs. Which means they have not yet been extended to even more complex architectures like tran... (read more)

cubefox
130

There are actually atmospheric microbiota, also called aeroplankton, though those are mostly bacteria.

I remember a novel by Stanisław Lem in which he talks about a planet which has an atmosphere with a green tint. It is caused by swarms of insects in the atmosphere doing photosynthesis. I don't know whether that would be realistic, but it's currently not possible on Earth, insofar no insects, nor any other animal, ever developed the ability to do photosynthesis.

4Nate Showell
There are sea slugs that photosynthesize, but that's with chloroplasts they steal from the algae they eat.
cubefox
20

I think not, because in my test the snippet didn't really contain such a quote that would have answered the question directly.

1samuelshadrach
Can you send the query? Also can you try typing the query twice into the textbox? I'm using openai text-embedding-3-small, which seems to sometimes work better if you type the query twice. Another thing you can try is retry the query every 30 minutes. I'm cycling subsets of the data every 30 minutes as I can't afford to host the entire data at once.
cubefox
2-8

Both Altman and Gwern used fine-tuned models, those don't really do in-context learning. They don't support "prompt engineering" in the original sense, they only respond to commands and questions in a particular way.

cubefox
20

I'm not sure fine-tuning is necessary. Most recent models have a ~100.000 token context window now, so they could fit quite a few short high quality examples for in-context learning. (Gemini Pro even has a 2 million token context window, but of course the base model is unavailable to the public.)

1JustisMills
I would be curious to see an attempt! I have a pretty strong prior that it would fail, though, with currently available models. I buy that RLHF hurts, but given Sam Altman's sample story also not impressing me (and having the same failure modes, just slightly less so), the problem pattern-matches for me to the underlying LLM simply not absorbing the latent structure well enough to imitate it. You might need more parameters, or a different set of training data, or something. (This also relates to my reply to gwern above - his prompt did indeed include high quality examples, and in my opinion it helped ~0.)
cubefox
20

Fine-tuned models are generally worse at writing fiction with good style than base models with temperature 1. For example the GPT-3.5 base model, code-davinci-002, was much better than the GPT-3.5 version tuned for chat. Here is what mainstream journalists said about it at the time.

4JustisMills
I agree and disagree, and considered getting into this in my post. I agree in the sense that certainly, since fine-tuned models are fine-tuned towards a persona that you'd expect to be bad at writing fiction, base models have higher upside potential. But also, I think base models are too chaotic to do all that good a job, and veer off in wacky directions, and need a huge amount of manual sampling/pruning. So whether they're "better" seems like a question of definition to me. I do think that the first actually good literary fiction AI will be one of: * A big/powerful enough model to capture the actual latent structure of high quality literary fiction, rather than only the surface level (thus letting it experiment more deeply and not default to the most obvious choice in every situation), or * A base model fine-tuned quite hard for literary merit, and not RLHF'd for "assistant"-y stuff The best written AI art I've seen so far has been nostalgebraist-autoresponder's tumblr posts, so I guess my money is on the latter of these two options. Simply not being winnowed into a specific persona strikes me as a valuable feature for creating good art.
cubefox
*20

I see you fixed the https issue. I think the resulting text snippets are reasonably related to the input question, though not overly so. Google search often answers questions more directly with quotes (from websites, not from books), though that may be too ambitious to match for a small project. Other than that, the first column could be improved with relevant metadata such as the source title. Perhaps the snippets in the second column could be trimmed to whole sentences if it doesn't impact the snippet length too much. In general, I believe snippets currently do not show line breaks present in the source.

1samuelshadrach
Thanks for feedback.  I’ll probably do the title and trim the snippets.  One way of getting a quote would to be to do LLM inference and generate it from the text chunk. Would this help?
cubefox
20

Okay, that works in Firefox if I change it manually. Though the server seems to be configured to automatically redirect to HTTPS. Chrome doesn't let me switch to HTTP.

1samuelshadrach
Thanks for your patience. I'd be happy to receive any feedback. Negative feedback especially.
cubefox
42

Error: TypeError: NetworkError when attempting to fetch resource.

1samuelshadrach
use http not https
cubefox
*91

One issue is that fine-tuned language models exhibit a, for blog posts inappropriate, "helpful assistant" writing style. But base models do not have any such default style.

So we could just take an advanced open foundation model, feed in some interesting blog posts, and let the model predict the next one, with a date in the future to prevent it from spitting out something from the training data it has memorized.

I think the best available base model might be DeepSeek-V3-Base. It has a context window of 128.000 tokens, which is about 200 pages. We could then ... (read more)

cubefox
20

These problems are partly related to poor planning, but they are clearly also related to language models, which are primarily restricted to operate on text. Actual AGI will likely have to work more like an animal or human brain, which is predicting sensory data (or rather: latent representations of sensory data, JEPA) instead of text tokens. An LLM with good planning may be able to finally beat Pokémon, but it will almost certainly not be able to do robotics or driving or anything with complex or real-time visual data.

cubefox
*60

Thank you, this was an insightful paper!

One concern though. You define the honesty score as , which is the probability of the model being either honest or evasive or not indicating a belief. However, it seems more natural to define the "honesty score" as the ratio (odds) converted to a probability. Which is

So this is the probability of the model being honest given that it is either honest or lies, i.e. assuming that it isn't evasive and doesn't fail to indicate a belief. It essentiall... (read more)

1Mantas Mazeika
Hi, thanks for your interest! We do include something similar in Appendix E (just excluding the "no belief" examples, but keeping evasions in the denominator). We didn't use this metric in the main paper, because we weren't sure if it would be fair to compare different models if we were dropping different examples for each model, but I think both metrics are equally valid. The qualitative results are similar. Personally, I think including evasiveness in the denominator makes sense. If models are 100% evasive, then we want to mark that as 0% lying, in the sense of lies of commission. However, there are other forms of lying that we do not measure. For example, lies of omission are marked as evasion in our evaluation, but these still manipulate what the user believes and are different from evading the question in a benign manner. Measuring lies of omission would be an interesting direction for future work.
cubefox
20

The coordinates x, y, R, G, B are independent, so it should be possible. I think the problem is just our intuition, which isn't optimized for perceiving color like three distances in space, or even like three separate values at all.

2Self
I feel like such intuitions could be developed. - I'm more uncertain where I would use this skill. Though given how OOD it is there could be significant alpha up for grabs (Q: Where would X-Ray vision for cluster structures in 5-dimensional space be extraordinarily useful?)
Load More