ΑΙhub.org
 

Meta now allows military agencies to access its AI software. It poses a moral dilemma for everybody who uses it


by
25 November 2024



share this:

camouflage pattern
By Zena Assaad, Australian National University

Meta will make its generative artificial intelligence (AI) models available to the United States’ government, the tech giant has announced, in a controversial move that raises a moral dilemma for everyone who uses the software.

Meta last week revealed it would make the models, known as Llama, available to government agencies, “including those that are working on defence and national security applications, and private sector partners supporting their work”.

The decision appears to contravene Meta’s own policy which lists a range of prohibited uses for Llama, including “[m]ilitary, warfare, nuclear industries or applications” as well as espionage, terrorism, human trafficking and exploitation or harm to children.

Meta’s exception also reportedly applies to similar national security agencies in the United Kingdom, Canada, Australia and New Zealand. It came just three days after Reuters revealed China has reworked Llama for its own military purposes.

The situation highlights the increasing fragility of open source AI software. It also means users of Facebook, Instagram, WhatsApp and Messenger – some versions of which use Llama – may inadvertently be contributing to military programs around the world.

What is Llama?

Llama is a collation of large language models – similar to ChatGPT – and large multimodal models that deal with data other than text, such as audio and images.

Meta, the parent company of Facebook, released Llama in response to OpenAI’s ChatGPT. The key difference between the two is that all Llama models are marketed as open source and free to use. This means anyone can download the source code of a Llama model, and run and modify it themselves (if they have the right hardware). On the other hand, ChatGPT can only be accessed via OpenAI.

The Open Source Initiative, an authority that defines open source software, recently released a standard setting out what open source AI should entail. The standard outlines “four freedoms” an AI model must grant in order to be classified as open source:

  • use the system for any purpose and without having to ask for permission
  • study how the system works and inspect its components
  • modify the system for any purpose, including to change its output
  • share the system for others to use with or without modifications, for any purpose.

Meta’s Llama fails to meet these requirements. This is because of limitations on commercial use, the prohibited activities that may be deemed harmful or illegal and a lack of transparency about Llama’s training data.

Despite this, Meta still describes Llama as open source.

The intersection of the tech industry and the military

Meta is not the only commercial technology company branching out to military applications of AI. In the past week, Anthropic also announced it is teaming up with Palantir – a data analytics firm – and Amazon Web Services to provide US intelligence and defence agencies access to its AI models.

Meta has defended its decision to allow US national security agencies and defence contractors to use Llama. The company claims these uses are “responsible and ethical” and “support the prosperity and security of the United States”.

Meta has not been transparent about the data it uses to train Llama. But companies that develop generative AI models often utilise user input data to further train their models, and people share plenty of personal information when using these tools.

ChatGPT and Dall-E provide options for opting out of your data being collected. However, it is unclear if Llama offers the same.

The option to opt out is not made explicitly clear when signing up to use these services. This places the onus on users to inform themselves – and most users may not be aware of where or how Llama is being used.

For example, the latest version of Llama powers AI tools in Facebook, Instagram, WhatsApp and Messenger. When using the AI functions on these platforms – such as creating reels or suggesting captions – users are using Llama.

The fragility of open source

The benefits of open source include open participation and collaboration on software. However, this can also lead to fragile systems that are easily manipulated. For example, following Russia’s invasion of Ukraine in 2022, members of the public made changes to open source software to express their support for Ukraine.

These changes included anti-war messages and deletion of systems files on Russian and Belarusian computers. This movement came to be known as “protestware”.

The intersection of open source AI and military applications will likely exacerbate this fragility because the robustness of open source software is dependent on the public community. In the case of large language models such as Llama, they require public use and engagement because the models are designed to improve over time through a feedback loop between users and the AI system.

The mutual use of open source AI tools marries two parties – the public and the military – who have historically held separate needs and goals. This shift will expose unique challenges for both parties.

For the military, open access means the finer details of how an AI tool operates can easily be sourced, potentially leading to security and vulnerability issues. For the general public, the lack of transparency in how user data is being utilised by the military can lead to a serious moral and ethical dilemma.The Conversation

Zena Assaad, Senior Lecturer, School of Engineering, Australian National University

This article is republished from The Conversation under a Creative Commons license. Read the original article.




The Conversation is an independent source of news and views, sourced from the academic and research community and delivered direct to the public.
The Conversation is an independent source of news and views, sourced from the academic and research community and delivered direct to the public.

            AIhub is supported by:



Subscribe to AIhub newsletter on substack



Related posts :

monthly digest

AIhub monthly digest: February 2026 – collective decision making, multi-modal learning, and governing the rise of interactive AI

  27 Feb 2026
Welcome to our monthly digest, where you can catch up with AI research, events and news from the month past.

The Good Robot podcast: the role of designers in AI ethics with Tomasz Hollanek

  26 Feb 2026
In this episode, Tomasz argues that design is central to AI ethics and explores the role designers should play in shaping ethical AI systems.

Reinforcement learning applied to autonomous vehicles: an interview with Oliver Chang

  25 Feb 2026
In the third of our interviews with the 2026 AAAI Doctoral Consortium cohort, we hear from Oliver Chang.

The Machine Ethics podcast: moral agents with Jen Semler

In this episode, Ben and Jen Semler talk about what makes a moral agent, the point of moral agents, philosopher and engineer collaborations, and more.

Extending the reward structure in reinforcement learning: an interview with Tanmay Ambadkar

  23 Feb 2026
Find out more about Tanmay's research on RL frameworks, the latest in our series meeting the AAAI Doctoral Consortium participants.

The Good Robot podcast: what makes a drone “good”? with Beryl Pong

  20 Feb 2026
In this episode, Eleanor and Kerry talk to Beryl Pong about what it means to think about drones as “good” or “ethical” technologies.

Relational neurosymbolic Markov models

and   19 Feb 2026
Relational neurosymbolic Markov models make deep sequential models logically consistent, intervenable and generalisable

AI enables a Who’s Who of brown bears in Alaska

  18 Feb 2026
A team of scientists from EPFL and Alaska Pacific University has developed an AI program that can recognize individual bears in the wild, despite the substantial changes that occur in their appearance over the summer season.



AIhub is supported by:







Subscribe to AIhub newsletter on substack




 















©2026.02 - Association for the Understanding of Artificial Intelligence