November 07, 2024

The United States Must Win The Global Open Source AI Race

On Nov. 1, Reuters reported that Chinese researchers, including ones affiliated with the People’s Liberation Army (PLA), used one of Meta’s Llama models for military purposes last year. The news led to a quick and robust reaction from many, including U.S. policymakers, arguing for further restrictions on open source AI. Michael McCaul, Chairman of the House Foreign Affairs Committee, said the recently proposed ENFORCE ACT–a bill that could effectively prohibit American AI developers from releasing open-weight models–was necessary to “keep American AI out of China’s hands.”

Unlike models such as OpenAI’s ChatGPT or Anthropic’s Claude, the Llama family of language models are “open-weight,” meaning that their weights–the numbers that define its functionality–are available for anyone to download for free online. Other well-known open source AI providers include Mistral, based in France, and Falcon, developed in the United Arab Emirates. For years, debates have been raging n the strategic benefits and risks of two AI ecosystems: one that is based primarily on proprietary and closed-source AI systems and one that is supportive of open source.

As AI becomes increasingly integrated into the world’s digital infrastructure, the importance of open source AI will grow, too, as it is likely to be a key building block in driving AI’s global diffusion and adoption.

While public access to open-weight models does represent a real tradeoff between control, security, and innovation – as the Llama example underscores – the story is more complicated. Critics of open source models fail to recognize the key role these models will play in securing U.S. security interests in the long term. Rather than focusing on the risks of open source AI, policymakers should ask whether the world should rely on U.S.-developed AI – or the increasingly capable open source models from China.

The Risks of Open Source AI

Those who are more skeptical of open source AI argue that the best way to mitigate AI’s negative impacts and security risks is to develop new regulations and restrict its global distribution. Threat actors can modify open models and remove critical safety features, creating new security risks. Moreover, as open models can be run on anyone’s hardware, the original developer cannot monitor its usage for dangerous or harmful applications in ways that closed model providers can (at least, in theory). It is for this reason that closed-source AI companies are investing vast resources to prevent the theft or export of their model weights.

Read the full article on Just Security.

  • Reports
    • March 20, 2025
    Countering the Digital Silk Road: Indonesia

    This year marks the 10th anniversary of the Digital Silk Road (DSR), China’s ambitious initiative to shape critical digital infrastructure around the world to advance its geop...

    By Vivek Chilukuri & Ruby Scanlon

  • Video
    • March 20, 2025
    How Secure Is America’s AI Advantage?

    https://www.youtube.com/watch?v=7njJkH7XRa8...

    By Paul Scharre

  • Commentary
    • March 17, 2025
    The Development of an Artificial Intelligence (AI) Action Plan

    Strengthening and securing America’s AI dominance is crucial for U.S. national security and economic competitiveness...

    By Vivek Chilukuri, Michael Depp, Bill Drexel, Janet Egan, Paul Scharre, Josh Wallin, Becca Wasser & Caleb Withers

  • Podcast
    • March 17, 2025
    ChinaTalk: Building Compute in America

    Despite leading the world in AI innovation, there’s no guarantee that America will rise to meet the challenge of AI infrastructure. Specifically, the key technological barrier...

    By Tim Fist & Jordan Schneider

View All Reports View All Articles & Multimedia