Quoting from Rohin Shah's Alignment Newsletter (with light editing):
In the US, the National Security Commission on AI released their report to Congress. The full pdf is over 750 pages long, so I have not read it myself, and instead I’m adding in some commentary from others. In their newsletter, CSET says that highlights include:
- A warning that the U.S. military could be at a competitive disadvantage within the next decade if it does not accelerate its AI adoption. The report recommends laying the foundation for widespread AI integration by 2025, comprising a DOD-wide digital ecosystem, a technically literate workforce, and more efficient business practices aided by AI.
- A recommendation that the White House establish a new “Technology Competitiveness Council,” led by the vice president, to develop a comprehensive technology strategy and oversee its implementation.
- A recommendation that the U.S. military explore using autonomous weapons systems, provided their use is authorized by human operators.
- A proposal to establish a new Digital Service Academy and a civilian National Reserve to cultivate domestic AI talent.
- A call to provide $35 billion in federal investment and incentives for domestic semiconductor manufacturing.
- A recommendation to double non-defense AI R&D funding annually until it reaches $32 billion per year, and to triple the number of National AI Research Institutes.
- A call for reformed export controls, coordinated with allies, on key technologies such as high-end semiconductor manufacturing equipment.
- A recommendation that Congress pass a second National Defense Education Act and reform the U.S. immigration system to attract and retain AI students and workers from abroad.
While none of the report’s recommendations are legally binding, it has reportedly been well-received by key members of both parties.
Matthew van der Merwe also summarizes the recommendations in Import AI; this has a lot of overlap with the CSET summary so I won't copy it here.
Jeff Ding adds in ChinAI #134:
"[I]f you make it past the bluster in the beginning — or take it for what it is: obligatory marketing to cater to a DC audience hooked on a narrow vision of national security — there’s some smart moderate policy ideas in the report (e.g. chapter 7 on establishing justified confidence in AI systems)."
In email correspondence, Jon Rodriguez adds some commentary on the safety implications:
"1. The report acknowledges the potential danger of AGI, and specifically calls for value alignment research to take place (pg. 36). To my knowledge, this is one of the first times a leading world government has called for value alignment.
2. The report makes a clear statement that the US prohibits AI from authorizing the launch of nuclear weapons (pg. 98).
3. The report calls for dialogues with China and Russia to ensure that military decisions made by military AI at "machine speed" does not lead to out-of-control conflict escalation which humans would not want (pg. 97)."
And here's a very boiled down version of the executive summary (with the phrasings of the recommendations being the exact phrasings the report uses as headings):
- The NSCAI recommends the government take the following 7 actions to "Defend America in the AI Era":
- Defend against emerging AI-enabled threats to America’s free and open society
- Prepare for future warfare
- Manage risks associated with AI-enabled and autonomous weapons
- Transform national intelligence
- Scale up digital talent in government
- Establish justified confidence in AI systems
- Present a democratic model of AI use for national security
- The NSCAI also recommends the government take the following 8 actions to "Win the Technology Competition":
- Organize with a White House–led strategy for technology competition
- Win the global talent competition
- Accelerate AI innovation at home
- Implement comprehensive intellectual property (IP) policies and regimes
- Build a resilient domestic base for designing and fabricating microelectronics
- Protect America’s technology advantages
- Build a favorable international technology order
- Win the associated technologies competitions
(Like Rohin, I've not read the full report - I've just read the executive summary, read all mentions of "nuclear", and listened to some episodes of the accompanying podcast series.)
Do "we" pay too little attention to things like this report?
Epistemic status: Just thinking aloud. As usual, these are my personal views only.
I found the executive summary of the report, the accompanying podcast episodes I listened to, and the commentary Rohin collected interesting regarding:
- Topics like AI development, AI risk, AI governance, and how this intersects with national security and international relations
- How the US government is thinking about about framing those things, how the government might react to or try to influence those things, etc.
- Obviously the US government is massive and does not act as a single, coherent agent. But this report should still provide one interesting window into its thinking.
I've found what I've read more interesting for (2) than for (1), though I expect the full report would teach me a lot more about (1) as well. And I expect many other Forum readers would find the report - or at least its executive summary - useful for learning about those things as well. In particular, the report and accompanying podcast episodes seemed to have substantially different framings, focuses, and tones to what the longtermist/EA/rationalist communities talk about in relation to AI.
Yet there seems to be no previous link-post to this report on the EA Forum or LessWrong. (The section of the Alignment Newsletter linked to above serves a similar function, but doesn't provide a dedicated title and link just for this report.) Nor does there seem to be much discussion of the report on other posts on those sites. I also hadn't really considered looking at the report myself until last week, despite knowing it existed.
I wonder if this is a sign that (parts of) the longtermist, EAs, and rationalist communities are paying insufficient attention to the available info on how (a) other communities and (b) relevant powerful actors other than AI labs are thinking about these issues? (I think there are also other signs of this, and that similar criticisms/concerns have been raised before.)
Counterpoints: I've seen the report mentioned by some professional longtermist researchers (that was how it came to my attention). And the report is very long and (unsurprisingly) doesn't seem to be primarily focused on things like existential risk from AI. And it makes sense for people to specialise. So maybe it's unsurprising that things like this report would be discussed mostly by a handful of professional specialists, rather than in venues like the EA Forum or LessWrong.
Counter-counterpoint: The Forum and LessWrong aren't just like regular blogs for semi-laypeople; they're also where a nontrivial fraction of longtermist thinking occurs and is disseminated.
Thus concludes this round of thinking aloud.