The rapid development of military artificial intelligence requires clear democratic governance rather than ad-hoc decisions by governments or private companies. As AI systems begin to influence intelligence analysis, cyber operations, logistics, and possibly battlefield decisions, the rules that control their use are becoming increasingly important. The article emphasizes that these decisions should be debated publicly and defined through law and policy rather than informal negotiations between technology companies and government officials.
The discussion is illustrated by a dispute between the United States Department of Defense and the AI company Anthropic. The company reportedly refused to allow its AI models to be used for certain purposes, such as domestic surveillance of U.S. citizens or fully autonomous military targeting. Government officials argued that such restrictions should be determined by public law and military policy rather than being built directly into private AI systems. This conflict highlights a larger question: who should decide the limits of military AI—governments, companies, or democratic institutions like legislatures.
The article explains that both sides have valid concerns. Governments must ensure that AI tools can support national defense and lawful military operations. At the same time, technology companies worry about ethical risks and the potential misuse of powerful AI systems. When decisions about these technologies happen through procurement disputes or private contracts, the public has little visibility into how the rules are being set. As a result, policies that affect national security and civil liberties may be shaped without democratic accountability.
The author concludes that military AI governance should be established through transparent institutions. Legislatures should create clear laws, defense agencies should develop detailed operational doctrine, and companies and civil society should participate in structured policy discussions. By placing guardrails in formal law rather than private negotiations, societies can ensure that AI used in warfare remains accountable, lawful, and aligned with democratic values.