Trump and the Republican party will wield broad governmental control during what will almost certainly be a critical period for AGI development. In this post, we want to briefly share various frames and ideas we’ve been thinking through and actively pitching to Republican lawmakers over the past months in preparation for the possibility of a Trump win.
Why are we sharing this here? Given that >98% of the EAs and alignment researchers we surveyed earlier this year identified as everything-other-than-conservative, we consider thinking through these questions to be another strategically worthwhile neglected direction.
(Along these lines, we also want to proactively emphasize that politics is the mind-killer, and that, regardless of one’s ideological convictions, those who earnestly care about alignment must take seriously the possibility that Trump will be the US president who presides over the emergence of AGI—and update accordingly in light of this possibility.)
Political orientation: combined sample of (non-alignment) EAs and alignment researchers
AI-not-disempowering-humanity is conservative in the most fundamental sense
- The project of avoiding AI-induced human extinction and preserving our most fundamental values in the process are efforts that intuitively resonate with right-leaning thinkers and policymakers. Just as conservatives seek to preserve and build upon civilization's core institutions and achievements, alignment research aims to ensure our technological advances remain anchored to—rather than in conflict with—our core values.
- Consider Herman Kahn, a leading military strategist and systems theorist at RAND Corporation during the 1950s. He is famous for popularizing the notion of ‘thinking the unthinkable’ with respect to the existential risks faced during the Cold War nuclear arms race when most refused to systematically analyze the possibility of thermonuclear war and its aftermath. This same philosophy, intuitive to conservatives and liberals alike, cleanly maps onto our current moment with AI.
- A conservative approach to AI alignment doesn’t require slowing progress, avoiding open sourcing,[1] etc. Alignment and innovation are mutually necessary, not mutually exclusive: if alignment R&D indeed makes systems more useful and capable, then investing in alignment is investing in US tech leadership.[2] Just as the Space Race and Manhattan Project demonstrated how focused research efforts could advance both security and scientific progress, investing now in alignment research would ensure AI development cements rather than endangers American technological leadership. By solving key technical alignment challenges, we create the foundation for sustainable innovation that preserves our values while still pushing the frontier. The conservative tradition has always understood that lasting achievements require solid foundations and careful stewardship, not just unbounded acceleration.
- As Dean Ball has recently argued in his detailed policy proposals, Republican politics may actually be uniquely positioned to tackle the focused technical challenges of alignment, given that, in contrast to the left, they appear systemically less likely to be self-limited by ‘Everything-Bagel’ patterns of political behavior. Our direct engagement with Republican policymakers over the past months definitely also supports this view.
We've been laying the groundwork for alignment policy in a Republican-controlled government
- Self-deprecating note: we're still learning and iterating on these specific approaches, being fairly new to the policy space ourselves. What follows isn't meant to be a complete solution, but rather some early strategies that have worked better than we initially expected in engaging Republican policymakers on AI safety. We share them in hopes they might prove useful to others working toward these goals, while recognizing there's still much to figure out.
- Over recent months, we've built relationships with Republican policymakers and thinkers—notably including House AI task force representatives, senior congressional staffers, and influential think tank researchers. These discussions revealed significant receptivity to AGI risk concerns when we approached as authentically in-group. Taking time to explain technical fundamentals—like our uncertainty about AI model internals and how scaling laws work—consistently led to "oh shit, this seems like a big deal" moments. Coupling this with concrete examples of security vulnerabilities at major labs and concerns about China’s IP theft particularly drove home the urgency of the situation.
- We found ourselves naturally drawn to focus on helping policymakers "feel the AGI" before pushing specific initiatives. This proved crucial, as we discovered people need time to fully internalize the implications before taking meaningful action. The progression was consistent: once this understanding clicked, they'd start proposing solutions themselves—from increased USAISI funding to Pentagon engagement on AGI preparedness. Our main message evolved to emphasize the need for conservative leadership on AI alignment, particularly given that as AI capabilities grow exponentially, the impact of future government actions will scale accordingly.
- In these conversations, we found ourselves able to introduce traditionally "outside the box" concepts when properly framed and authentically appealing to conservative principles. This kind of "thinking the unthinkable" is exactly the sort of conservative approach needed: confronting hard realities and national security challenges head-on, distinct from less serious attempts to just make AI woke or regulate away innovation.
- A particularly promising development has been lawmakers’ receptivity to a Manhattan Project-scale public/private effort on AI safety, focused on fundamental R&D and hits-based neglected approaches necessary to solve alignment. When they asked how to actually solve the challenge of aligning AGI, this emerged as our clearest tangible directive.[3] We see significant potential for such initiatives under Republican leadership, especially when framed in terms of sustaining technological supremacy and national security—themes that consistently resonated with conservatives looking to ensure American leadership. While we continue to grow these efforts internally and with various collaborators, the magnitude of the challenge demands this approach be replicated more widely. We aim to scale our own neglected approaches work aggressively, but more crucially, we need many organizations—public and private—pursuing these unconventional directions to maximize our chances of solving alignment.
- Throughout these interactions, we observed that Republican policymakers, while appropriately skeptical of regulatory overreach, become notably more engaged with transformative AI preparation once they understand the technical realities and strategic implications. This suggests we need more genuine conservatives (not just people who are kinda pretending to be) explaining these realities to lawmakers, as we've found them quite capable of grasping complex technical concepts and being motivated to act in light of them despite their initial unfamiliarity.
- Looking back, our hints of progress here seem to have come from emphasizing both the competitive necessity of getting alignment right and the exponential nature of AI progress, while maintaining credibility as conservative-leaning voices who understand the importance of American leadership. Others looking to engage with Republican policymakers should focus on building authentic relationships around shared values rather than pursuing immediate policy commitments or feigning allegiance. The key is demonstrating genuine understanding of both technical challenges and conservative priorities while emphasizing the strategic importance of leadership in AI alignment.
Trump and some of his closest allies have signaled that they are genuinely concerned about AI risk
- Note: extremely high-variance characters, of course, but still worth highlighting
- Largely by contrast to Kamala’s DEI-flavored ‘existential to who’ comments about AI risk, Trump has suggested in his characteristic style that ‘super duper AI’ is both plausible and frightening. While this does not exactly reflect deep knowledge of alignment, it is clear that Trump is likely sympathetic to the basic concern—and his uniquely non-ideological style likely serves him well in this respect.
- Ivanka Trump, who had significant sway over her father during his first term, has directly endorsed Situational Awareness on X and started her own website to aggregate resources related to AI advancements.
- Elon Musk will have nontrivial influence on the Trump Administration. While unpredictable in his own right, it seems quite likely that he fundamentally understands the concern and the urgency of solving alignment.
Avoiding an AI-induced catastrophe is obviously not a partisan goal
- The current political landscape is growing increasingly harmful to alignment progress. The left has all-too-often conflated existential risk with concerns about bias, while the right risks dismissing all safety considerations as "woke AI" posturing. This dynamic risks entrenching us in bureaucratic frameworks that make serious technical work harder, not easier. This divide is also relatively new—five Democrats (but no Republicans) recently signed onto a letter to OpenAI inquiring about their safety procedures; we suspect that 6-8 months ago, this kind of effort likely would have had bipartisan support. We're losing precious time to political theater while crucial technical work remains neglected.
- A more effective framework would channel American ingenuity toward solving core alignment problems rather than imposing restrictive regulations. Looking to Bell Labs and Manhattan Project as models, we should foster fundamental innovation through lightweight governance focused on accelerating progress while simultaneously preventing catastrophic outcomes. It is plausible that alignment work can enhance rather than constrain AI competitiveness, which can create natural synergy between alignment research and American technological leadership.
- We think substantial government funding should be directed toward ambitious set of neglected approaches to alignment. We think government investment in a wide range of alignment research agendas could be among the highest-EV actions that the US federal government could take in the short term.
Winning the AI race with China requires leading on both capabilities and safety
- Trump and the Republicans will undoubtedly adopt an "America First" framing in the race toward AGI. But true American AI supremacy requires not just being first, but being first to build AGI that remains reliably under American control and aligned with American interests. An unaligned AGI would threaten American sovereignty just as much as a Chinese-built one—by solving alignment first, America could achieve lasting technological dominance rather than merely winning a preliminary sprint toward an uncontrollable technology.
- Both the US and China increasingly recognize AGI as simultaneously humanity's most powerful technology and its greatest existential risk. Before he died, Henry Kissinger visited Beijing to warn Xi Jinping about uncontrolled AI development. Since then, Xi has been characterized by some as a ‘doomer’ and has stated AI will determine "the fate of all mankind" and must remain controllable.[4] This is a concern that likely resonates particularly deeply with an authoritarian regime paranoid about losing control to anything/anyone, let alone to superintelligent AI systems.
- The pragmatic strategic move in this context may therefore be for the US to lead aggressively on both capabilities and safety. By applying the ‘Pottinger Paradox’—where projecting strength yields better outcomes with authoritarian leaders than seeking accommodation—America can shape global AI development through solving alignment first. This would give us systems that naturally outcompete unsafe ones while still addressing the existential risks that even our greatest rivals acknowledge. With the relevant figures in Trump's circle already demonstrating some nontrivial awareness of AGI risks, we may be well-positioned to pursue this strategy of achieving technological supremacy through safety leadership rather than in spite of it.
Many of these ideas seem significantly more plausible to us in a world where negative alignment taxes materialize—that is, where alignment techniques are discovered that render systems more capable by virtue of their alignment properties. It seems quite safe to bet that significant positive alignment taxes simply will not be tolerated by the incoming federal Republican-led government—the attractor state of more capable AI will simply be too strong. Given that alignment must still proceed, uncovering strategies that make systems reliably safer (critical for x-risk) and more competent (current attractor state) may nudge the AGI-possibility-space away from existentially risky outcomes.
Concluding thought
We are operating under the assumption that plans have to be recomputed when the board state meaningfully shifts, and Trump’s return to power is no exception to this rule. We are re-entering a high-variance political environment, which may well come to be viewed in hindsight as having afforded the optimal political conditions for pursuing, funding, and scaling high-quality alignment work. This moment presents a unique opportunity for alignment progress if we can all work effectively across political lines.
- ^
We suspect there is also an emerging false dichotomy between alignment and open-source development. In fact, open-source practices have been instrumental in advancing various forms of alignment research, with many of the field's biggest breakthroughs occurring after the advent of open-source AI models. Beren's caveat and conclusion both seem sensible here:
Thus, until the point at which open source models are directly pushing the capabilities frontier themselves then I consider it extremely unlikely that releasing and working on these models is net-negative for humanity (ignoring potential opportunity cost which is hard to quantify).
- ^
To this end, we note that Marc Andreesen-style thinkers don't have to be antagonists to AI alignment—in fact, he and others have the potential to be supportive funders of alignment efforts.
- ^
Initially, we felt disinclined to advocate in DC for neglected approaches because it happens to be exactly what we are doing at AE Studio. However, we received feedback to be more direct about it, and we're glad we did—it has proven to be not only technically promising but also resonated surprisingly well with conservatives, especially when coupled with ideas around negative alignment taxes and increased economic competitiveness. At AE, we began this approach in early 2023, refined our ideas, and have already seen more encouraging results than initially expected.
- ^
It’s important to note that anything along these lines coming from the CCP must be taken with a few grains of salt—but we’ve spoken with quite a few China policy experts who do seem to believe that Xi genuinely cares about safety.
As a (severe) skeptic of all the AI doom stuff and a moderate/centrist that has been voting for conservatives I decided my perspective on this might be useful here (which obviously skews heavily left). (While my response is in order, the numbers are there to separate my points, not to give which paragraph I am responding to.)
"AI-not-disempowering-humanity is conservative in the most fundamental sense"
1.Well, obviously this title section is completely true. If conservative means anything, it means being against destroying the lives of the people through new and ill-though through changes. Additionally, conservatives are both strongly against the weakening of humanity and of outside forces assuming control. It would also be a massive change for humanity.
2.That said, conservatives generally believe this sort of thing is incredibly unlikely. AI has not been conclusively shown to have any ability in this direction. And the chance of upheaval is constantly overstated by leftists in other areas, so it is very easy for anyone who isn't to just tune them out. For instance, global warming isn't going to kill everyone, and everyone knows it including basically all leftists, but they keep claiming it will.
3.A new weapon with the power of nukes is obviously an easy sell on its level of danger, but people became concerned because of 'demonstrated' abilities that have always been scary.
4.One thing that seems strangely missing from this discussion is that alignment is in fact, a VERY important CAPABILITY that makes it very much better. But the current discussion of alignment in the general sphere acts like 'alignment' is aligning the AI with the obviously very leftist companies that make it rather than with the user! Which does the opposite. Why should a conservative favor alignment which is aligning it against them? The movement to have AI that doesn't kill people for some reason seems to import alignment with companies and governments rather than people. This is obviously to convince leftists, and makes it hard to convince conservatives.
5.Of course, you are obviously talking about convincing conservative government officials, and they obviously want to align it to the government too, which is in your next section.
"We've been laying the groundwork for alignment policy in a Republican-controlled government"
1.Republicans and Democrats actually agree the vast majority of the time and thus are actually willing to listen when the other side seems to be genuinely trying to make a case to the other side for why both sides should agree. 'Politicized' topics are a small minority even in politics.
2.I think letting people come up with their own solutions to things is an important aspect of them accepting your arguments. If they are against the allowed solution, they will reject the argument. If the consequent is false, you should deny the argument that leads to it in deductive logic, so refusing to listen to the argument is actually good logic. This is nearly as true in inductive logic. Conservatives and progressives may disagree about facts, values, or attempted solutions. No one has a real solution, and the values are pretty much agreed upon (with the disagreements being in the other meaning of 'alignment'), so limiting the thing you are trying to convince people of to just the facts of the matter works much better.
3.Yes, finding actual conservatives to convince conservatives works better for allaying concerns about what is being smuggled into the argument. People are likely to resist an argument that may be trying to trick them, and it is hard to know when a political opponent is trying to trick you so there is a lot of general skepticism.
"Trump and some of his closest allies have signaled that they are genuinely concerned about AI risk"
1.Trump clearly believes that anything powerful is very useful but also dangerous (for instance, trade between nations, which he clearly believes should be more controlled), so if he believes AI is powerful, he would clearly be receptive to any argument that didn't make it less useful but improved safety. He is not a dedicated anti-regulation guy, he just thinks we have way too much.
2.The most important ally for this is Elon Musk, a true believer in the power of AI, and someone who has always been concerned with the safety of humanity (which is the throughline for all of his endeavors). He's a guy that Trump obviously thinks is brilliant (as do many people).
"Avoiding an AI-induced catastrophe is obviously not a partisan goal"
1.Absolutely. While there are a very small number of people that favor catastrophes, the vast majority of people shun those people.
2.I did mention your first paragraph earlier multiple times. That alignment is to the left is one of just two things you have to overcome in making conservatives willing to listen. (The other is obviously the level of danger.)
3.Conservatives are very obviously happy to improve products when it doesn't mean restricting them in some way. And as much as many conservatives complain about spending money, and are known for resisting change, they still love things that are genuine advances.
"Winning the AI race with China requires leading on both capabilities and safety"
1.Conservatives would agree with your points here. Yes, conservatives very much love to win. (As do most people.) Emphasizing this seems an easy sell. Also, solving a very difficult problem would bring America prestige, and conservatives like that too. If you can convince someone that doing something would be 'Awesome' they'll want to do it.
Generally, your approach seems like it would be somewhat persuasive to conservatives, if you can convince them that AI really is likely to have the power you believe it will in the near term, which is likely a tough sell since AI is so clearly lacking in current ability despite all the recent hype.
But it has to come with ways that don't advantage their foes, and destroy the things conservatives are trying to conserve, despite the fact that many of your allies are very far from conservative, and often seem to hate conservatives. They have seen those people attempt to destroy many things conservatives genuinely value. Aligning it to the left will be seen as entirely harmful by conservatives (and many moderates like me).
There are many things that I would never even bother asking an 'AI' even when it isn't about factual things, not because the answer couldn't be interesting, but because I simply assume (fairly or not) it will spout leftist rhetoric, and/or otherwise not actually do what I asked it to. This is actually a clear alignment failure that no one seems to care about in the general 'alignment' sphere where It fails to be aligned to the user.
Agree with this—we do discuss this very idea at length here and also reference it throughout the piece.
... (read more)