ΑΙhub.org
 

Meta now allows military agencies to access its AI software. It poses a moral dilemma for everybody who uses it


by
25 November 2024



share this:

camouflage pattern
By Zena Assaad, Australian National University

Meta will make its generative artificial intelligence (AI) models available to the United States’ government, the tech giant has announced, in a controversial move that raises a moral dilemma for everyone who uses the software.

Meta last week revealed it would make the models, known as Llama, available to government agencies, “including those that are working on defence and national security applications, and private sector partners supporting their work”.

The decision appears to contravene Meta’s own policy which lists a range of prohibited uses for Llama, including “[m]ilitary, warfare, nuclear industries or applications” as well as espionage, terrorism, human trafficking and exploitation or harm to children.

Meta’s exception also reportedly applies to similar national security agencies in the United Kingdom, Canada, Australia and New Zealand. It came just three days after Reuters revealed China has reworked Llama for its own military purposes.

The situation highlights the increasing fragility of open source AI software. It also means users of Facebook, Instagram, WhatsApp and Messenger – some versions of which use Llama – may inadvertently be contributing to military programs around the world.

What is Llama?

Llama is a collation of large language models – similar to ChatGPT – and large multimodal models that deal with data other than text, such as audio and images.

Meta, the parent company of Facebook, released Llama in response to OpenAI’s ChatGPT. The key difference between the two is that all Llama models are marketed as open source and free to use. This means anyone can download the source code of a Llama model, and run and modify it themselves (if they have the right hardware). On the other hand, ChatGPT can only be accessed via OpenAI.

The Open Source Initiative, an authority that defines open source software, recently released a standard setting out what open source AI should entail. The standard outlines “four freedoms” an AI model must grant in order to be classified as open source:

  • use the system for any purpose and without having to ask for permission
  • study how the system works and inspect its components
  • modify the system for any purpose, including to change its output
  • share the system for others to use with or without modifications, for any purpose.

Meta’s Llama fails to meet these requirements. This is because of limitations on commercial use, the prohibited activities that may be deemed harmful or illegal and a lack of transparency about Llama’s training data.

Despite this, Meta still describes Llama as open source.

The intersection of the tech industry and the military

Meta is not the only commercial technology company branching out to military applications of AI. In the past week, Anthropic also announced it is teaming up with Palantir – a data analytics firm – and Amazon Web Services to provide US intelligence and defence agencies access to its AI models.

Meta has defended its decision to allow US national security agencies and defence contractors to use Llama. The company claims these uses are “responsible and ethical” and “support the prosperity and security of the United States”.

Meta has not been transparent about the data it uses to train Llama. But companies that develop generative AI models often utilise user input data to further train their models, and people share plenty of personal information when using these tools.

ChatGPT and Dall-E provide options for opting out of your data being collected. However, it is unclear if Llama offers the same.

The option to opt out is not made explicitly clear when signing up to use these services. This places the onus on users to inform themselves – and most users may not be aware of where or how Llama is being used.

For example, the latest version of Llama powers AI tools in Facebook, Instagram, WhatsApp and Messenger. When using the AI functions on these platforms – such as creating reels or suggesting captions – users are using Llama.

The fragility of open source

The benefits of open source include open participation and collaboration on software. However, this can also lead to fragile systems that are easily manipulated. For example, following Russia’s invasion of Ukraine in 2022, members of the public made changes to open source software to express their support for Ukraine.

These changes included anti-war messages and deletion of systems files on Russian and Belarusian computers. This movement came to be known as “protestware”.

The intersection of open source AI and military applications will likely exacerbate this fragility because the robustness of open source software is dependent on the public community. In the case of large language models such as Llama, they require public use and engagement because the models are designed to improve over time through a feedback loop between users and the AI system.

The mutual use of open source AI tools marries two parties – the public and the military – who have historically held separate needs and goals. This shift will expose unique challenges for both parties.

For the military, open access means the finer details of how an AI tool operates can easily be sourced, potentially leading to security and vulnerability issues. For the general public, the lack of transparency in how user data is being utilised by the military can lead to a serious moral and ethical dilemma.The Conversation

Zena Assaad, Senior Lecturer, School of Engineering, Australian National University

This article is republished from The Conversation under a Creative Commons license. Read the original article.




The Conversation is an independent source of news and views, sourced from the academic and research community and delivered direct to the public.
The Conversation is an independent source of news and views, sourced from the academic and research community and delivered direct to the public.




            AIhub is supported by:


Related posts :



#AAAI2025 workshops round-up 1: Artificial intelligence for music, and towards a knowledge-grounded scientific research lifecycle

  18 Mar 2025
We hear from the organisers of two workshops at AAAI2025 and find out the key takeaways from their events.

The Good Robot podcast: Re-imagining voice assistants with Stina Hasse Jørgensen and Frederik Juutilainen

  17 Mar 2025
Eleanor and Kerry chat to Stina Hasse Jørgensen and Frederik Juutilainen about an experimental research project that created an alternative voice assistant.

Visualizing research in the age of AI

  14 Mar 2025
Felice Frankel discusses the implications of generative AI when communicating science visually.

#IJCAI panel on communicating about AI with the public

  13 Mar 2025
A recording of this session at IJCAI2024 is now available to watch.

Interview with Tunazzina Islam: Understand microtargeting and activity patterns on social media

  11 Mar 2025
Hear from Doctoral Consortium participant Tunazzina about her research on computational social science, natural language processing, and social media mining and analysis

Microsoft cuts data centre plans and hikes prices in push to make users carry AI costs

  10 Mar 2025
Microsoft is trying to recoup the costs by raising prices, putting ads in products, and cancelling data centre leases

Report on the future of AI research

  07 Mar 2025
Find out more about a report released by the AAAI 2025 Presidential Panel.

Andrew Barto and Richard Sutton win 2024 Turing Award

  06 Mar 2025
Pair are recognised for their pioneering reinforcement learning research.




AIhub is supported by:






©2024 - Association for the Understanding of Artificial Intelligence


 












©2021 - ROBOTS Association