AI Safety Newsletter #43: White House Issues First National Security Memo on AI Plus, AI and Job Displacement, and AI Takes Over the Nobels

By Center for AI Safety, Corin Katzke, AlexaPanYue, Dan H @ 2024-10-28T16:02 (+6)

This is a linkpost to https://newsletter.safe.ai/p/ai-safety-newsletter-43-white-house

Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required.

Listen to the AI Safety Newsletter for free on Spotify or Apple Podcasts.

Subscribe here to receive future versions.


White House Issues First National Security Memo on AI

On October 24, 2024, the White House issued the first National Security Memorandum (NSM) on Artificial Intelligence, accompanied by a Framework to Advance AI Governance and Risk Management in National Security.

The NSM identifies AI leadership as a national security priority. The memorandum states that competitors have employed economic and technological espionage to steal U.S. AI technology. To maintain a U.S. advantage in AI, the memorandum directs the National Economic Council to assess the U.S.’s competitive position in:

The Intelligence Community must make gathering intelligence on competitors' operations against the U.S. AI sector a top-tier intelligence priority. The Department of Energy will launch a pilot project to evaluate AI training and data sources, while the National Science Foundation will distribute compute through the National AI Research Resource pilot.

The NSM and Framework outline AI safety and security practices. The AI Safety Institute at the National Institute of Standards and Technology becomes the primary U.S. Government point of contact for private sector AI developers. The AI Safety Institute must:

The Department of Energy's National Nuclear Security Administration will lead classified testing for nuclear and radiological risks, while the National Security Agency will evaluate cyber threats through its AI Security Center.

The framework prohibits government AI use for:

National security agencies must also:

All requirements take effect immediately, with initial implementation deadlines beginning within 60 days. The memorandum’s fact sheet is available here

AI and Job Displacement

In this story, we look at recent work on the future of AI job displacement. 

Brookings projects widespread job disruption from GPT-4 capabilities. A study by the Brookings Institution investigates the potential impacts of current generative AI on the American workforce, using estimates shared by OpenAI. Key findings include:

These results assume GPT-4 capabilities and moderate innovation in system autonomy and applications. Importantly, it “does not attempt to project future capability enhancements from next-generation AI models likely to be released”—meaning that the actual effects of AI on job displacement are likely to be even greater than those above.

GPT-4o can outperform human CEOs in some respects. A recent experiment conducted by researchers at Cambridge found that GPT-4o can significantly outperform human CEOs in strategic decision-making tasks. The study, which involved 344 participants navigating a gamified simulation of the U.S. automotive industry, found that:

Traditionally, executive positions have been thought to be insulated from AI automation due to the generalist and strategic nature of executive work. However, this experiment suggests that current frontier AI systems have potential to automate (parts of) even these positions.

Physical labor is not safe from AI. The jobs currently most exposed to frontier AI capabilities involve cognitive labor such as writing and coding. However, if investment is any indication, jobs involving physical labor might be next. 

Fei Fei Li’s startup, World Labs, raised $230 million last month. The startup describes itself as a “spatial intelligence company building Large World Models to perceive, generate, and interact with the 3D world.”

Robotics is seeing similar levels of investment. Earlier this month, Tesla previewed “Optimus,” a line of humanoid robots in development. The robots at the event were piloted by humans—but the robotics were impressive and moved stably, and could eventually be piloted by a sufficiently capable AI.

Economic policy implications of AGI. Anton Korinek wrote a working paper for the National Bureau of Economic Research on the economic policy challenges posed by AI. He writes that job displacement resulting from the development of AGI “will necessitate a fundamental reevaluation of our economic structures, social systems, and the meaning of work in human society.”

The paper identifies several challenges the development of AGI would pose for economic policy, including:

AI Takes Over the Nobels

Two of the 2024 Nobel Prizes were awarded to AI researchers. In this story, we discuss their implications for AI in science and AI safety. 

Geoffrey Hinton, widely known as “the Godfather of AI”, received the Nobel Prize in Physics. Hopfield, a physicist who shares the award, created Hopfield networks which improved associative memory in neural nets. Hinton is credited with Boltzmann machines, a technique building on Hopfield’s, which helped pretrain networks by backpropagation. This then spawned recent explosive progress in machine learning.

Illustrations of Hopfield and Hinton. Source.

Demis Hassabis and John Jumper, both Google Deepmind scientists, were awarded the Nobel Prize in Chemistry. They are recognized for developing AlphaFold2, an AI model trained to predict protein structure from amino acid sequences. Presented in 2020, the model has now succeeded in predicting the structure of virtually all known proteins, previously thought to be extremely difficult for human researchers.

The Nobel decisions suggest that AI is increasingly driving scientific progress. Both the physics and chemistry prize statements highlight the impact of machine learning on their fields. The physics statement reads: “in physics we use artificial neural networks in a vast range of areas, such as developing new materials with specific properties”. Likewise, the Chemistry committee noted that “AlphaFold2 has been used by more than two million people from 190 countries” and for a myriad of scientific applications. 

As AI capabilities for scientific research improve, future scientific breakthroughs might become increasingly dominated by AI. 

Hinton’s Nobel win could bolster credibility for AI safety. In post-award interviews and press conferences, Hinton has used much of his time to discuss and promote AI safety. In particular, he has emphasized challenges in controlling advanced AI systems and avoiding catastrophic outcomes, and called for more safety research from labs, governments, and young researchers. 

The Nobel Prize for Hinton's achievements in AI will hopefully lend more credibility to his advocacy for AI safety. 

Government

Industry

Research

See also: CAIS website, X account for CAIS, our $250K Safety benchmark competition, our new AI safety course, and our feedback form

You can contribute challenging questions to our new benchmark, Humanity’s Last Exam. The deadline is November 1st—the end of this week!

The Center for AI Safety is also hiring for several positions, including Chief Operating Officer, Director of Communications, Federal Policy Lead, and Special Projects Lead.

Double your impact! Every dollar you donate to the Center for AI Safety will be matched 1:1 up to $2 million. Donate here.