Do you have any thoughts on whether it would make sense to push for a rule that forces open-source or open-weight models to be released behind an API for a certain amount of time before they can be released to the public?
It is better than nothing I suppose but if they are keeping the safeties and restrictions on then it will not teach you whether it is fine to open it up.
in a zero marginal cost world
nit: inference is not zero marginal cost. statement seems to be importing intuitions from traditional software which do not necessarily transfer. let me know if I misunderstood or am confused.
I think people just say "zero marginal cost" in this context to refer to very low marginal cost. I agree that inference isn't actually that low cost though. (Certainly much higher than the cost of distributing/serving software.)
It was all quiet. Then it wasn’t.
Note the timestamps on both of these.
Dwarkesh Patel did a podcast with Mark Zuckerberg on the 18th. It was timed to coincide with the release of much of Llama-3, very much the approach of telling your story directly. Dwarkesh is now the true tech media. A meteoric rise, and well earned.
This is two related posts in one. First I cover the podcast, then I cover Llama-3 itself.
My notes are edited to incorporate context from later explorations of Llama-3, as I judged that the readability benefits exceeded the purity costs.
Podcast Notes: Llama-3 Capabilities
The Need for Inference
Great Expectations
Open Source and Existential and Other Risks
Interview Overview
That was a great interview. It tackled important questions. For most of it, Zuck seemed like a real person with a unique perspective, saying real things.
The exception was that weird period where he was defending open source principles using what sounded like someone else’s speech on a tape recorder. Whereas at other times, his thoughts on open source were also nuanced and thoughtful. Dwarkesh was unafraid to press him on questions of open source throughout the interview.
What Dwarkesh failed to get was any details from Zuck about existential or catastrophic risk. We are left without any idea of how Zuck thinks about those questions, or what he thinks would be signs that we are in such danger, or what we might do about it. He tried to do this with the idea of Meta needing a risk policy, but Zuck kept dodging. I think there was more room to press on specifics. Once again this presumably comes down to Zuck not believing the dangerous capabilities will exist.
Nor was there much discussion of the competitive dynamics that happen when everyone has access to the same unrestricted advanced AI models, and what might happen as a result.
I also think Zuck is failing to grapple with even the difficulties of mundane content moderation, an area where he is an expert, and I would like to see his explicit response. Previously, he has said that only a company with the resources of a Meta can do content moderation at this point.
I think he was wrong in the sense that small bespoke gardens are often successfully well-defended. But I think Zuck was right that if you want to defend something worth attacking, like Meta, you need scale and you need to have the expertise advantage. But if those he is defending against also have the resources of Meta where it counts, then what happens?
So if there is another interview, I hope there is more pressing on those types of questions.
In terms of how committed Zuck is to open source, the answer is a lot but not without limit. He will cross that bridge when he comes to it. On the horizon he sees no bridge, but that can quickly change. His core expectation is that we have a long way to go before AI goes beyond being a tool, even though he also thinks it will soon very much be everyone’s personal agent. And he especially thinks that energy restrictions will soon bind, which will stifle growth because that goes up against physical limitations and government regulations. It is an interesting theory. If it does happen, it has a lot of advantages.
A Few Reactions
Ate-a-Pi has a good reaction writeup on Twitter. It was most interesting in seeing different points of emphasis. The more I think about it, the more Ate-a-Pi nailed it pulling these parts out:
Especially on point is that Zuck never expects the AI itself to be the product. This is a common pattern among advocates for open model weights – they do not actually believe in AGI or the future capabilities of the product. It is not obvious Zuck and I even disagree so much on what capabilities would make it unwise to open up model weights. Which is all the more reason to spell out what that threshold would be.
Then there is speculation from Ate-a-Pi that perhaps Zuck is being realistic because Meta does not need to raise capital, whereas others hype to raise capital. That surely matters on the margin, in both directions. Zuck would love if Altman and Amodei were less able to raise capital.
But also I am confident this is a real disagreement, to a large extent, on both sides. These people expecting big jumps from here might turn out to be bluffing. But I am confident they think their hand is good.
Daniel Jeffries highlights GPT-5 as key evidence either way, which seems right.
If GPT-5 lands at either extreme it would be very strong evidence. We also could get something in the middle, and be left hanging. I also would not be too quick in calendar time to conclude progress is stalling, if they take their time releasing 5 and instead release smaller improvements along the way. The update would be gradual, and wouldn’t be big until we get into 2025.
Ate-a-Pi also offers this explanation of the business case for opening up Llama-3.
Here I find some very interesting model disagreements.
Ate says that Meta’s biggest thereat is character.ai, and that this undercuts character.ai.
Whereas I would say, this potentially supercharges character.ai, they get to improve their offerings a lot, as do their competitors (of varying adult and ethical natures).
Meta perhaps owns your real world friends (in which case, please help fix that locally, ouch). But this is like the famous line. The AIs get more capable. Your friends stay the same.
Similarly, Ate says that this ‘allows for social debugging outside of Meta,’ because Meta’s primary product is moderation. He thinks this will make moderation easier. I think this is insane. Giving everyone better AI, catching them up to what Meta has, makes moderation vastly harder.
Here are some reactions from people less skeptical than I am of open source.
As I noted above, I think everyone sensible is at core talking price. What level of open model weight capabilities is manageable in what capacities? What exactly are we worried about going wrong and can we protect against it, especially when you cannot undo a release, the models may soon be smarter than us and there are many unknown unknowns about what might happen or what the models could do.
To take Nora’s style of thinking here and consider it fully generally, I think such arguments are in expectation (but far from always) backwards. Arguments of the form ‘yes X makes Y worse, but solving X would not solve Y, so we should not use Y as a reason to solve X’ probably points the other way, unless you can point to some Z that solves Y and actually get Z. Until you get Z, this usually means you need X more, as the absolute risk difference is higher rather than lower.
More specifically this is true when it comes to ease of getting necessary information and otherwise removing inconveniences. If something is going to be possible regardless, you need to raise the cost and lower the salience and availability of doing that thing.
I’ve talked about this before, but: Indeed there are many things in our civilization, really quite a lot, where someone with sufficient publically available knowledge can exploit the system, and occasionally someone does, but mostly we don’t partly for ethical or moral reasons, partly for fear of getting caught somehow or other unknown unknowns, but even more so because it does not occur to us and when it does it would be a bunch of work to figure it out and do it. Getting sufficiently strong AI helping with those things is going to be weird and force us to a lot of decisions.
Critch’s proposal generalizes, to me, to the form ‘ensure that civilization is not vulnerable to what the AIs you release are capable of doing.’ The first step there is to secure access to compute against a potential rogue actor using AI, whether humans are backing it or not. Now that you have limited the compute available to the AI, you can now hope that its other capabilities are limited by this, so you have some hope of otherwise defending yourself.
My expectation is that even in the best case, defending against misuses of open model weights AIs once the horses are out of the barn is going to be a lot more intrusive and expensive and unreliable than keeping the horses in the barn.
Consider the metaphor of a potential pandemic on its way. You have three options.
The core problem with Covid-19 is that we found both #1 and #3 unacceptable (whether or not we were right to do so), so we went with option #2. It did not go great.
With open source AI, you can take option #1 and hope everything works out. You are ‘trusting the thermodynamic God,’ letting whatever competitive dynamics and hill climbing favor win the universe, and hoping that everything following those incentive gradients will work out and have value to you. I am not optimistic.
You can also take option #3, and suppress before sufficiently capable models get released. If Zuckerberg is right about energy being the limiting factor, this is a very practical option, even more so than I previously thought. We could talk price about what defines sufficiently capable.
The problem with option #2 is that now you have to worry about everything the AIs you have unleashed might do and try to manage those risks. The hope Critch expresses is that even if we let the AIs get to inference time, and we know people will then unleash rogue AIs on the regular because of course they will try, as long as we control oversized sources of compute what those AIs can do will be limited.
This seems to me to be way harder (and definitely strictly harder) than preventing those open models from being trained and released in the first place. You need the same regime you would have used, except now you need to be more intrusive. And that is the good scenario. My guess is that you would need to get into monitoring on the level of personal computers or even phones, because otherwise the AI could do everything networked even if you did secure the data centers. Also I do not trust you to secure the data centers at this point even if you are trying.
But yes, those are the debates we should be having. More like this.
Safety First
So what about Llama-3? How good is it?
As always we start with the announcement and the model card. They are releasing model weights for two models, Llama-3 8B and Llama-3 70B. They are already available for light inference.
Let’s get the safety question out of the way before we get to capabilities.
Then in the model card:
Under this philosophy, safety is not a model property.
Instead, safety is a property of a particular deployment of that model, with respect to the safety intentions of the particular party making that deployment.
In other words:
Or:
Or:
I am willing to believe that Llama 3 may have been developed in a responsible way, if the intention was purely to deploy it the ways GPT-4 has been deployed.
That is different from deploying Llama 3 in a responsible way.
One can divide those who use Llama 3 into three categories here.
If you are in category #1, Meta still has a job to do. We don’t know if they did it. If they didn’t, they are deploying it to all their social media platforms, so ut oh. But probably they did all right.
If you are in category #2, Meta has another job to do. It is not obviously harder because the standard of what is acceptable is lower. When I was writing this the first time, I noticed that so far people were not reporting back attempts to jailbreak the model, other than one person who said they could get it to produce adult content with trivial effort.
My next sentence was going to be: Even Pliny’s other successes of late, it would be rather surprising if a full jailbreak of Llama-3 was that hard even at Meta.ai.
I was considering forming a Manifold market, but then I realized I should check first, and indeed this has already happened.
This is not proof of a full jailbreak per se, and it is not that I am upset with Meta for not guarding against the thing Google and OpenAI and Anthropic also can’t stop. But it is worth noting. The architecture listed above has never worked, and still won’t.
Meta claims admirable progress on safety work for a benevolent deployment context, including avoiding false refusals, but is light on details. We will see. They also promise to iterate on that to improve it over time, and there I believe them.
Finally, there is scenario three, where someone willing to fine tune the model, or download someone else’s fine tune, and cares not for the input safeguard or output safeguard.
As your periodic reminder, many people want this.
In that scenario, I assume there is no plan. Everyone understands that if a nonstate actor or foreign adversary or anyone else wants to unleash the power of this fully operational battlestation, then so be it. The hope is purely that the full power is not that dangerous. Which it might not be.
Good, that’s out of the way. On to the rest.
Core Capability Claims
They claim the 8B and 70B versions are the best models out there in their classes. They claim improvement on false refusal rates, on alignment, and in increased diversity of model responses. And they have strong benchmarks.
My principle is to look at the benchmarks for context, but never to trust the benchmarks. They are easily gamed, either intentionally or unintentionally. You never know until the humans report back.
This data is representing that the 8B model as far better than Gemma and Mistral. Given how much data and compute they used, this is far from impossible. Maybe it was that simple all along. The numbers are if anything suspiciously high.
For the 70B we see a very strong HumanEval number, and overall roughly comparable numbers.
What about those human evaluators? They claim results there too.
These are from a new Meta-generated question set (careful, Icarus), and are compared side by side by human evaluators. Llama-3 70B won handily, they do not show results for Llama-3 8B.
The context window remains small, only 8k tokens. They promise to improve on that.
They preview Llama 400B+ and show impressive benchmarks.
For comparison, from Claude’s system card:
So currently these numbers are very similar to Claude Opus all around, and at most mildly selected. The core Meta hypothesis is that more training and data equals better model, so presumably it will keep scoring somewhat higher. This is indicative, but as always we wait for the humans.
How Good are the 8B and 70B Models in Practice?
The proof is in the Chatbot Arena Leaderboard, although you do have to adjust for various factors.
So here is where things sit there.
So what does that mean?
I also asked on Twitter, and kept an eye out for other practical reports.
What makes this a bigger deal is that this is only the basic Llama-3. Others will no doubt find ways to improve Llama-3, both in general and for particular purposes. That is the whole idea behind the model being open.
He expects Claude Haiku would be well above the top of this list, as well.
Note that it looks like he got through by simply asking a second time. And of course, the Tweet does not actually contain hate speech or conspiracy theories, this is a logic test of the system’s refusal policy.
Playing with their image generator is fun. It is 1280×1280, quality seems good although very much not state of the art, and most importantly it responds instantly as you edit the prompt. So even though it seems limited in what it is willing to do for you, you can much easier search the space to figure out your best options, and develop intuitions for what influences results. You can also see what triggers a refusal, as the image will grey out. Good product.
Do they have an even more hilarious copyright violation problem than usual if you try at all? I mean, for what it is worth yes, they do.
I didn’t play with the models much myself for text because I am used to exclusively using the 4th-generation models. So I wouldn’t have a good baseline.
Architecture and Data
The big innovation this time around was More Data, also (supposedly) better data.
As others have pointed out ‘over 5%’ is still not a lot, and Llama-3 underperforms in other languages relative to similar models. Note that the benchmarks are in English.
This makes sense. Bespoke data filtering and more unique data are clear low hanging fruit. What Meta did was then push well past where it was obviously low hanging, and found that it was still helpful.
Note that with this much data, and it being filtered by Llama-2, contamination of benchmarks should be even more of a concern than usual. I do wonder to what extent that is ‘fair,’ if a model memorizes more things across the board then it is better.
There are more details in the model card at GitHub.
The ‘intended use’ is listed as English only, with other languages ‘out of scope,’ although fine-tunes for other languages are considered acceptable.
Training Day
How much compute did this take?
Andrej Karpathy takes a look at that question, calling it the ‘strength’ of the models, or our best guess as to their strength. Here are his calculations.
The estimates differ, but not by not much, so I’d consider them a range:
I think of the compute training cost as potential strength rather than strength. You then need the skill to make that translate into a useful result. Of course, over time, everyone’s skill level goes up. But there are plenty of companies that threw a lot of compute at the problem, and did not get their money’s worth in return.
This is in line with previous top tier models in terms of training cost mapping onto capabilities. You do the job well, this is about what you get.
What Happens Next With Meta’s Products?
Meta says they are going to put their AI all over their social media platforms, and at the top of every chat list. They had not yet done it on desktop when I checked Facebook, Instagram and Messenger, or on Facebook Messenger on mobile. I did see Meta AI in my feed as the second item in the mobile Facebook app, offering to have me ask it anything.
Once they turn this dial up, they will put Meta AI right there. A lot of people will get introduced to AI this way who had not previously tried ChatGPT or Claude, or DALLE or MidJourney.
Presumably this means AI images and text will ‘flood the zone’ on their social media, and also it will be one of the things many people talk about. It could make the experience a lot better, as people can illustrate concepts and do fact and logic checks and other neat low hanging fruit stuff, and maybe learn a thing or two. Overall it seems like a good addition.
We will also get a rather robust test of the first two categories of safety, and a continuous source of stories. Millions of teenagers will be using this, and there will be many, many eyes looking for the worst interactions to shine them under the lights Gary Marcus style. If they have their own version of the Gemini Incident, it will not be pretty.
Here is the Washington Post’s Naomi Nix and Will Oremus firing a warning shot.
I think this is a smart approach from Meta, and that it was a good business reason to invest in AI, although it is an argument against releasing the model weights.
What is not as smart is having Meta AI reply to posts unprompted. We saw the example last week where it hallucinated past experiences, now we have this:
This reads like one of those ‘who could have possibly thought anyone would want any version of this?’ experiences.
Ate-a-Pi pointed out an important implication from the interview. Zuckerberg said Meta does not open source their products themselves.
This means that they do not intend for Llama-3 to be the product, even the 400B version. They will not be offering a direct competitor in the AI space. And indeed, they do not think future Llama-Xs will ‘be the product’ either.
Will they integrate Llama-3 400B into their products? They might like to, but it is not so compatible with their business model to pay such inference costs and wait times. Remember that for Meta, you the customer are the product. You pay with your time and your attention and your content and very soul, but not directly with your money. Meanwhile the lifetime value of a new Facebook customer, we learned recently, is on the order of $300.
So what is Llama-3 400B, the most expensive model to train, even for from a product perspective? It does help train Llama-4. It helps try and hurt competitors like Google. It helps with recruitment, both to Meta itself and into their intended ecosystem. So there are reasons.
What Happens Next With AI Thanks To These Two Models?
Open models get better. I expect that the people saying ‘it’s so over’ for other models will find their claims overblown as usual. Llama-3 8B or 70B will for now probably become the default baseline model, the thing you use if you don’t want to think too hard about what to use, and also the thing you start with when you do fine tuning.
Things get more interesting over time, once people have had a chance to make variations that use Llama-3 as the baseline. In the space of Llama-2-based models, Llama-2 itself is rather lousy. Llama-3 should hold up better, but I still expect substantial improvements at least to specific use cases, and probably in general.
Also, of course, we will soon have versions that are fine-tuned to be useful,and also fine-tuned to remove all the safety precautions.
And we will see what happens due to that.
In the grand scheme, in terms of catastrophic risk or existential risk or anything like that, or autonomous agents that should worry us, my strong assumption is that nothing scary will happen. It will be fine.
In terms of mundane misuse, I also expect it to be fine, but with more potential on the margin, especially with fine-tunes.
Certainly some people will switch over from using Claude Sonnet or Haiku or another open model to now using Llama-3. There are advantages. But that will look incremental, I expect, not revolutionary. That is also true in terms of the pressure this exerts on other model providers.
The real action will be with the 400B model.
The Bigger One: It’s Coming
What happens if Meta goes full Leroy Jenkins and releases the weights to 400B?
Meta gets a reputational win in many circles, and grows its recruitment and ecosystem funnels, as long as they are the first 4-level open model. Sure.
Who else wins and loses?
For everyone else (and the size of Meta’s reputational win), a key question is, what is state of the art at the time?
In the discussions below, I assume that 5-level models are not yet available, at most OpenAI (and perhaps Google or Anthropic) has a 4.5-level model available at a premium price. All of this is less impactful the more others have advanced already.
And I want to be clear, I do not mean to catastrophize. These are directional assessments, knowing magnitude is very hard.
Who Wins?
The obvious big winner is China and Chinese companies, along with every non-state actor, and every rival and enemy of the United States of America. Suddenly they can serve and utilize and work from what might be a competitive top-level model, and no they are not going to be paying Meta a cut no matter the license terms.
Using Llama-3 400B to help train new 4.5-level models is going to be a key potential use case to watch.
They also benefit when this hurts other big American companies. Not only are their products being undercut by a free offering, which is the ultimate predatory pricing attack in a zero marginal cost world, those without their own models also have another big problem. The Llama-3 license says that big companies have to pay to use it, whereas everyone else can use it for free.
Another way they benefit? This means that American companies across industries, upon whom Meta can enforce such payments, could now be at a potentially large competitive disadvantage against their foreign rivals who ignore that rule and dare Meta to attempt enforcement.
This could also be a problem if foreign companies can ignore the ‘you cannot use this to train other models’ clause in 1(b)(v) of the license agreement, whereas American companies end up bound by that clause.
I am curious what if anything the United States Government, and the national security apparatus, are going to do about all that. Or what they would want to do about it next time around, when the stakes are higher.
The other obvious big winners are those who get to use Llama-3 400B in their products, especially those for whom it is free, and presumably get to save a bundle doing that. Note that even if Meta is not charging, you still have to value high quality output enough to pay the inference costs. For many purposes, that is not worthwhile.
Science wins to some degree, depending on how much this improves their abilities and lowers their costs. It also is a big natural experiment, albeit without controls, that will teach us quite a lot. Let’s hope we pay attention.
Also winners are users who simply want to have full control over a 4-level model for personal reasons. Nothing wrong with that. Lowering the cost of inference and lowering the limits imposed on it could be very good for some of those business models.
Who Loses?
The big obvious Corporate losers are OpenAI, Google, Microsoft and Anthropic, along with everyone else trying to serve models and sell inference. Their products now have to compete with something very strong, that will be freely available at the cost of inference. I expect OpenAI to probably have a superior product by that time, and the others may as well, but yes free (or at inference cost) is a powerful selling point, as is full customization on your own servers.
The secondary labs could have an even bigger problem on their hands. This could steamroller a lot of offerings.
All of which is (a large part of) the point. Meta wants to sabotage its rivals into a race to the bottom, in addition to the race to AGI.
Another potential loser is anyone or anything counting on the good guy with an AI having a better AI than the bad guy with an AI. Anywhere that AI could flood the zone with bogus or hostile content, you are counting on your AI to filter out what their AI creates. In practice, you need evaluation to be easier than generation under adversarial conditions where the generator chooses point and method of attack. I worry that in many places this is not by default true once the AIs on both sides are similarly capable.
I think this echoes a more general contradiction in the world, that is primarily not about AI. We want everyone to be equal, and the playing field to be level. Yet that playing field depends upon the superiority and superior resources and capabilities in various ways of the United States and its allies, and of certain key corporate players.
We demand equality and democracy or moves towards them within some contained sphere and say this is a universal principle, but few fully want those things globally. We understand that things would not go well for our preferences if we distributed resources fully equally, or matters were put to a global vote. We realize we do not want to unilaterally disarm and single-handedly give away our advantages to our rivals. We also realize that some restrictions and concentrated power must ensure our freedom.
In the case of AI, the same contradictions are there. Here they are even more intertwined. We have far less ability to take one policy nationally or locally, and a different policy globally. We more starkly must choose either to allow everyone to do what they want, or not to allow this. We can either control a given thing, or not control it. You cannot escape the implications of either.
In any case: The vulnerable entities here could include ‘the internet’ and internet search in their broadest senses, and it definitely includes things like Email and social media. Meta itself is going to have some of the biggest potential problems over at Facebook and Instagram and its messenger services. Similar logic could apply to various cyberattacks and social engineering schemes, and so on.
I am generally confident in our ability to handle ‘misinformation,’ ‘deepfakes’ and similar things, but we are raising the difficulty level and running an experiment. Yes, this is all coming anyway, in time. The worry is that this levels a playing field that is not currently level.
I actually think triggering these potential general vulnerabilities now is a positive impact. This is the kind of experiment where you need to find out sooner rather than later. If it turns out the bad scenarios here come to pass, we have time to adjust and not do this again. If it turns out the good scenarios come to pass, then we learn from that as well. The details will be enlightening no matter what.
It is interesting to see where the mind goes now that the prospect is more concrete, and one is thinking about short term, practical impacts.
Other big Western corporations that would have to pay Meta could also be losers.
The other big loser, as mentioned above, is the United States of America.
And of course, if this release is bad for safety, either now or down the line, we all lose.
Again, these are all directional effects. I cannot rule out large impacts in scenarios where Llama-3 400B releases as close to state of the art, but everyone mostly shrugging on most of these also would not be shocking. Writing this down it occurs to me that people simply have not thought about this scenario much in public, despite it having been reasonably likely for a while.
How Unsafe Will It Be to Release Llama-3 400B?
The right question is usually not ‘is it safe?’ but rather ‘how (safe or unsafe) is it?’ Releasing a 4-level model’s weights is never going to be fully ‘safe’ but then neither is driving. When we say ‘safe’ we mean ‘safe enough.’
We do not want to be safetyists who demand perfect safety. Not even perfect existential safety. Everything is price.
The marginal existential safety price on Llama-3 70B and Llama-3 8B is very small, essentially epsilon. Standing on its own, the decision to release the weights of these models is highly reasonable. It is a normal business decision. I care only because of the implications for future decisions.
What is the safety price for the releasing the model weights of Llama-3 400B, or another 4-level model?
I think in most worlds the direct safety cost here is also very low, especially the direct existential safety cost. Even with extensive scaffolding, there are limits to what a 4-level model can do. I’d expect some nastiness on the edges but only on the edges, in limited form.
How many 9s of direct safety here, compared to a world in which a 4-level model was never released with open weights? I would say two 9s (>99%), but not three 9s (<99.9%). However the marginal safety cost versus the counterfactual other open model releases is even smaller than that, and there I would say we have that third 9 (so >99.9%).
I say direct safety because the primary potential safety dangers here seem indirect. They are:
And again, these only matter on the margin to the extent they move the margin.
At the time of Llama-2, I said what I was concerned about opening up was Llama-4.
That is still the case now. Llama-3 will be fine.
Will releasing Llama-4 be fine? Probably. But I notice my lack of confidence.
The Efficient Market Hypothesis is False
(Usual caveat: Nothing here is investing advice.)
Market is not impressed. Nasdaq was down 6.2% in this same period.
You can come up with various explanations. The obvious cause is that WhatsApp and Threads were forcibly removed from the Apple Store in China, along with Signal and Telegram. I am confused why this would be worth a 3% underperformance.
(Then about a day later it looked like we were finally going to actually force divestiture of TikTok while using that to help pass a foreign aid bill, so this seems like a massive own goal by China to remind us of how they operate and the law of equivalent exchange.)
The stock most down was Nvidia, which fell 10%, on no direct news. Foolish, foolish.
At most, markets thought Llama-3’s reveal was worth a brief ~1% bump.
You can say on Meta that ‘it was all priced in.’ I do not believe you. I think the market is asleep at the wheel.
Some are of course calling these recent moves ‘the market entering a correction phase’ or that ‘the bubble is bursting.’ Good luck with that.
Here is a WSJ article about how Meta had better ensure its AI is used to juice advertising returns. Investors really are this myopic.
Any given company, of course, could still be vastly overvalued.
Here was the only argument I saw to that effect with respect to Nvidia.
I can totally buy that a lot of investors have no idea what Nvidia actually produces, and got freaked out by suddenly learning what Nvidia actually does. I thought it was very public long ago that Google trains on TPUs that they design? I thought it was common knowledge that everyone involved was going to try to produce their own chips for at least internal use, whether or not that will work? And that Nvidia will still have plenty of customers even if all the above switched to TPUs or their own versions?
That does not mean that Nvidia’s moat is impregnable. Of course they could lose their position not so long from now. That is (a lot of) why one has a diversified portfolio.
Again. The Efficient Market Hypothesis in False.
What Next?
I expect not this, GPT-5 will be ready when it is ready, but there will be pressure:
I do not doubt that OpenAI and others will do everything they can to stay ahead of Meta’s releases, with an unknown amount of ‘damn the safety checks of various sorts.’
That does not mean that one can conjure superior models out of thin air. Or that it is helpful to rush things into use before they are ready.
Still, yes, everyone will go faster on the frontier model front. That includes that everyone in the world will be able to use Llama-3 400B for bootstrapping, not only fine-tuning.
On the AI mundane utility front, people will get somewhat more somewhat cheaper, a continuation of existing trends, with the first two models. Later we will have the ability to get a 4-level model internally for various purposes. So we will get more and cheaper cool stuff.
Meta will deploy its tools across its social media empire. Mostly I expect this to be a positive experience, and to also get a lot more people to notice AI. Expect a bunch of scare stories and highlights of awful things, some real and some baseless.
On the practical downside front, little will change until the 400B model gets released. Then we will find out what people can do with that, as they attempt to flood the zone in various ways, and try for all the obvious forms of misuse. It will be fun to watch.
All this could be happening right as the election hits, and people are at their most hostile and paranoid, seeing phantoms everywhere.
Careful, Icarus.