During a war with national mobilization, that would make sense. Or in a country like China. This kind of coercion is not an expected part of democratic rule.
It has always been a part of democratic rule, in peacetime and war. All telco's share virtually all of their technology with the government. Governments in europe and elsewhere routinely requisition services from many of their large corporations. I think it's absurd to think llm's can meaningfully participate in realworld cmd+ctrl systems and the government already has access to ml-enhanced targeting capabilities. I really have no idea what dod normies think of ai, other than that it's infinitely smarter than them, but that's not saying much.
Not the same thing. The parent comment was talking about government “requisiting” services as in forceful compliance, takeovers, not collaboration or regulatory compliance.
If you're referring to telcos sharing their tech with government there are a few examples of Ericsson working with the Swedish military:
> Brigadier-General Mattias Hanson, CIO, Swedish Armed Forces, says: “Strengthening Sweden’s militarily and acting as part of a collective defense requires us to increase our defensive capabilities. We need to utilize the latest technology and all the innovative power of the Swedish private sector. Sweden has unique skills and capabilities in both telecoms and defense technology..." [0]
The question of whether or not the government should be able to use AI for targeting without the involvement of humans is a wartime question, since that is the only time the military should be killing people.
Under such a scenario, requisition applies, and so all of this talk is moot.
The fact that the military is killing people without a declaration of war is the problem, and that's where energy and effort should be directed.
Edit:
There's a yet larger question on whether any legal constraints on the military's use of technology even makes sense at all, since any safeguards will be quickly yielded if a real enemy presents itself. As a course of natural law, no society will willingly handicap its means of defense against an external threat.
It follows then that the only time these ethical concerns apply is when we are the aggressor, which we almost always are. It's the aggression that we should be limiting, not the technology.
You could view various non-proliferation agreements as a legislative constraint on military technology.
Same for chemical and biologicals. Those do prove your point that the law will be ignored if expedient. But it doesn't invalidate the notion of a society putting constraints on itself.
During a war with national mobilization, that would make sense. Or in a country like China. This kind of coercion is not an expected part of democratic rule.