Tag Archives: Militarism

Killer Robots: Applying arms-control frameworks to autonomous weapons

October 5, 2021  |  Zachary Kallenborn

Brookings

“Mankind’s earliest weapons date back 400,000 years—simple wooden spears discovered in Schöningen, Germany. By 48,000 years ago, humans were making bows and arrows, then graduating to swords of bronze and iron. The age of gunpowder brought flintlock muskets, cannons, and Gatling guns. In modern times, humans built Panzer tanks, the F-16 Fighting Falcon, and nuclear weapons capable of vaporizing cities.

Today, humanity is entering a new era of weaponry, one of autonomous weapons and robotics.

The development of such technology is rapidly advancing and poses hard questions about how their use and proliferation should be governed. In early 2020, a drone may have been used to attack humans autonomously for the first time, a milestone underscoring that robots capable of killing may be widely fielded sooner rather than later. Existing arms-control regimes may offer a model for how to govern autonomous weapons, and it is essential that the international community promptly addresses a critical question: Should we be more afraid of killer robots run amok or the insecurity of giving them up?

The current state of autonomy

The first use of an autonomous weapon to kill is thought to have occurred in March of 2020 in Libya, but what actually happened in remains murky. According to a UN report, a Turkish-made Kargu-2 drone is reported to have autonomously “hunted down” members of the Libyan National Army. If the manufacturer’s claims are correct, the Kargu-2 can use machine learning to classify objects, apparently allowing it to “autonomously fire-and-forget.” Turkey denies using the Kargu-2 in this way, though seems to acknowledge the Kargu-2 can be used autonomously. Regardless of whether Kargu-2 was used autonomously in the episode in Libya, the claim that Kargu-2 can be autonomous is plausible on its face.

The United Nations report about the Kargu-2 caused an uproar. Sensationalist headlines compared the Kargu-2 to a “Terminator-style AI drone” that “hunted down human targets without being given orders.” These stories conjured images of out of control, sentient robots killing as they saw fit. To be blunt, that is nonsense. Although artificial intelligence—technically a super intelligent narrow AI—can beat the world’s best human chess and Go players, that is far from a generalized, human-level intelligence like the Terminator. In fact, a sticky note is enough to convince a cutting edge machine vision system that an apple is an iPod.

But simple autonomy is not that hard, and autonomous weapons have been a feature of warfare for centuries. Autonomy is about machines operating without human control. The weapon just needs a sensor, a way to process sensor information, and activate the harmful payload. During the American Civil War, Confederate forces deployed the “Rains Patent,” a simple landmine made of sheet iron with a brass cap sealed in beeswax to protect the fuse. When Union soldiers put sufficient pressure on the Rains Patent, it exploded.

Modern autonomous weapons play a real, but relatively limited role in military operations. The Ottawa Convention banned anti-personnel mines, but anti-vehicle and sea mines are still used. Loitering munitions are somewhere between a drone and a missile, hovering above a battlefield and striking targets that meet various designations. The U.S. Phalanx close-in weapon system, the Israeli Iron Dome, and various active defense systems defend against incoming missiles and other close range risks with varying degrees of autonomy. Along the demilitarized zone between South and North Korea, South Korea has deployed the SGR A-1 gun turret, which reportedly has an optional fully autonomous mode. This is just the beginning.

Though hype in the view of some, artificial intelligence has won over the world’s great military powers as the next great military technology. The U.S. National Security Commission on AI recently concluded that “properly designed, tested, and utilized AI-enabled and autonomous weapon systems will bring substantial military and even humanitarian benefit.” The Chinese People’s Liberation Army believes AI could fundamentally change the character of warfare. Russian President Vladimir Putin’s claim—that the world’s AI leader “will become the ruler of the world”— has become cliché.

Such excitement has translated to new research, prototypes, and increasingly operational autonomous weapons with increasing degrees of sophistication. The Defense Advanced Research Projects Agency—the U.S. military’s high risk, high-reward research and development center that helped birth the internet and GPS—ran a virtual dogfight between an F-16 Fighting Falcon and an artificial intelligence last year. The AI beat the human in each of five rounds. China is doing the same, with similar results. The United StatesChina, and Russia are all developing loyal wingman drones: unmanned semi-autonomous or autonomous aircraft that support manned aircraft. The pilot provides strategic decisions, while the artificial intelligence manages the details.

Growing autonomy is closely tied to the rise of unmanned platforms. Numerous states are testing, building, and deploying a wide range of unmanned aircraft, shipssubmarines, and tanks. Unmanned platforms require remote orders to achieve their mission. That’s tough when militaries cannot provide the full staff needed and pilots burn out from overwork. Plus, enemies seek to jam, manipulate, or otherwise interfere with the signals from the pilots to the drone. The more that unmanned platforms can operate without human control, the less need for those signals and the people sending them.

States are integrating unmanned platforms into drone swarms, and in May, Israel became the world’s first to deploy a swarm in combat. In a true drone swarm, the drones communicate and collaborate, forming a single weapons platform. While drone swarms are not necessarily autonomous weapons, no human could control 10,000 drones without artificial intelligence helping. Israel’s groundbreaking use of a drone swarm appears to have consisted of an unknown number of small drones equipped with a mixture of sensors and weapons. Israel’s swarm use is just the beginning. India tested a 75-drone swarm last year, and earlier this year, South Africa’s Paramount Group revealed a swarming system of 41-kilogram long-range drones that cruise at more than 100 miles per hour. Russia, meanwhile, is designing swarms for anti-submarine warfare. Numerous other states are developing other swarm applications.

Assuming these trends continue, autonomous weapons will increasingly enter the battlefield. For some, that’s terrifying.

In his Christmas address of 1937, the Archbishop of Canterbury posed a prescient question: “Who can think without horror of what another widespread war would mean, waged as it would be with all the new weapons of mass destruction?” Arms control debates are rooted in fear, and treaties to control the spread of weapons have sought to address them. Advocates of arms-control treaties fear the consequences of horrific weapons of war spreading widely. Opponents fear what might happen if their adversaries build such weapons, but they cannot. These dueling fears animate everything from gun debates around the dinner table to nuclear arms debates at the United Nations. The proliferation of autonomous weapons creates new fears: that autonomous weapons might misidentify a civilian as a soldier and killing him, that autonomous weapons will provide an enemy state a decisive edge in war.

In an autonomous weapon, the system decides when to engage by processing environmental stimuli. Landmines, for example, use simple pressure sensors—the sensor sensitivity determines whether the heft of a tank or the hands of a child are enough to trigger the explosion. Conversely, an anti-radar loitering munition homes in on radar signals. The risk of error—and by extension the arms control concern—depend on the type of environmental stimuli, how the stimuli is processed, and the type of decisions made.

Emerging autonomous weapons using machine learning process stimuli in more complex ways. Machine learning systems rely on large amounts of data to draw conclusions about what the system observes. But the data dependence also makes them brittle. Color differences, tree branches, or foggy days may confound the ability of the system to correctly identify a target. Although some states may adopt robust verification and testing programs to increase reliability, others may not. As autonomous weapons are deployed in larger numbers, arms control advocates fear a higher likelihood of something going horrifyingly wrong.

As autonomous weapons scale into massive drone swarms, the uncontrolability and potential for mass harm create a new weapon of mass destruction. Imagine 1,000 Slaughterbots flitting about a city, deciding who to kill. And that’s not terribly outlandish: India wants to build a swarm of 1,000 drones operating without human control. And the Naval Postgraduate School is modeling swarms of up to a million drones, operating underwater, on the ocean’s surface, and in the air. Particularly nefarious governments might equip the drones with facial recognition to assassinate regime opponents or carry out ethnic cleansing. States have adopted a wide range of policies to reduce similar risks from traditional weapons of mass destruction, including export controls, arms control treaties, and deterrent and coercive threats. If drone swarms are weapons of mass destruction, they deserve similar risk reduction policies.

At the same time, militaries see great value in the development of autonomous weapons. Autonomous weapons offer speed. A typical human takes 250 milliseconds to react to something they see. An autonomous weapon can respond far faster—Rheinmetall Defense’s Active Defense System can react to incoming rocket-propelled grenade in less than one millisecond. According to General George Murray, head of the U.S. Army’s Future Command, that speed may be necessary to defend against massive drone swarms. These weapons may be the difference between survival and defeat. Giving them up in an arms control treaty would be foolish.

Militaries also dispute the risk of error. Humans get tired, frustrated, and over-confident. That creates mistakes. Autonomous weapons have no such emotion, and advocates of military AI applications argue a reduced error-rate makes pursuing the technology a moral imperative. Plus, artificial intelligence can improve aiming. That reduces collateral harm. For example, Israel reportedly used an artificial intelligence-assisted machine gun to assassinate an Iranian nuclear scientist without hitting the scientist’s wife inches away. So, in their view, what are arms control advocates really afraid of?

Work on this issue is ongoing. Diplomats have debated autonomous weapons issues under the United Nations Convention on Conventional Weapons Group of Governmental Experts on Lethal Autonomous Weapons since 2014. These meetings have accomplished little to create an international treaty on autonomous weapons, but they have helped clarify state positions, brought greater attention to the topic, and better articulated concerns regarding autonomous weapons. Arms control advocates have called for bans and new treaties, but these vary in scope. Groups like the Campaign to Stop Killer Robots argue all autonomous weapons must be banned. Others, like the International Committee of the Red Cross, have a more nuanced view, focusing on “unpredictable” weapons. Human Rights Watch estimates 30 states have endorsed a complete ban.Great military powers have resisted a new arms-control regime, arguing existing international law is sufficient to cover autonomous weapons. Researchers have also floated alternatives to arms control treaties, such as norms and bilateral and multi-lateral confidence-building measures.

The global community must now resolve the tension of fear between arms-control and military advocates. That means serious debate on which types of autonomous weapon offer the most military value and which present the most risk to civilians and noncombatants. Weapons with high risk to civilians and low military value should form the basis of conversations around risk reduction.

Existing arms control treaties offer models to address these complexities. The Ottawa Convention on Anti-Personnel Landmines narrowly focuses on anti-personnel landmines, excluding anti-vehicle landmines that require high pressure to detonate. An autonomous weapon treaty might focus on anti-personnel weapons using machine learning, given the challenges of distinguishing farmers from soldiers. A more precise treaty may allow military powers to separate weapons they fear giving up from the weapons arms control advocates fear proliferating. This may make getting an okay from those powers easier.

Autonomous weapons might also be tiered based on characteristics that make them more or less risky, akin to the Chemical Weapons Convention’s schedules. The convention divides chemical agents into three schedules, based on their historical use as chemical weapons and use for civilian purposes. Chemicals in each schedule have different restrictions placed upon them. Autonomous weapons could also be tiered based on the risk the weapons pose, particularly the risk to civilian populations if the weapon errs and the likelihood of an error. Defensive turrets used at sea to defend against incoming missiles would likely be of lowest risk, while offensive weapons targeting people using machine learning would be a higher tier. Autonomous chemical, biological, radiological, and nuclear weapons are the highest risk, and should never be used.

Debate is needed on the best policy approaches to stem the proliferation of the highest risk weapons, and reduce broader global risks. Existing global discussion has focused on whether international treaties should ban the weapons, but that’s just a start. Even if autonomous weapons are banned in whole, in part, or not at all, governments must consider how to ensure they are not inadvertently exported to states not party to the ban. Restricting access to terrorist groups is an extra, different problem as autonomous weapons are simple enough to be made as a classroom project. And if a new international treaty is established, an obvious question is: How can it be given teeth? If a state uses a banned autonomous weapon, should they suffer retaliatory diplomatic or economic sanctions? When, if ever, should the United Nations Security Council endorse military action?

The era of killer robots is here. What comes next is up to the world.

Zachary Kallenborn is a research affiliate with the Unconventional Weapons and Technology Division of the National Consortium for the Study of Terrorism and Responses to Terrorism, a policy fellow at the Schar School of Policy and Government, and a U.S. Army Training and Doctrine Command “Mad Scientist.” 

 


Every US President has committed War Crimes (since WWII), Here’s why….

KnowDrones was founded in 2012 to inform the American public about the illegality, immorality and dreadful human consequences of U.S. drone attacks in order to bring about: (1) a complete halt to drone attacks; and (2) an international ban on weaponized drones and military and police drone surveillance.

WBAY Channel 2: [Afghani] Refugee speaks about conditions at Fort McCoy

Published: Sep. 13, 2021 at 11:15 PM CDT

original article…

Refugee speaks about conditions at Fort McCoy 

 

“MADISON, Wis. (WKOW/WBAY) – One of the thousands of refugees who are currently housed at Wisconsin’s Fort McCoy.

Due to continued fear of Taliban retaliation, the refugee’s identity has been concealed.

He says he’s grateful to be safe, but adds that life at Fort McCoy has been a struggle.

Some refugees say they don’t have a change of clothes, and they are dealing with food shortages that force them to ration.

“It is not like that all the time. But if you’re a little bit late, you know, go to the chow hall, you know, most of their, like, four or five times happened, you know, they’re out of food. Or either they have just one other thing, you know, maybe boiled carrots, or little bit rice,” he told our Madison ABC affiliate, WKOW.

His wife, who doesn’t live at the base, also talked about the experience.

“He is not a complainer, he’s not one of those kind of people that is ungrateful at any shape or form. If anything, you know, he’s so grateful to be here,” she said.

The refugee says he asked officials about the food shortages, and was told they didn’t expect this many people to come to the base.

He also says some people need warmer clothes, adding it gets chilly at night.

For a complete list of donation items being accepted and more information, click here. To coordinate donations big or small, or interested in being a drop-off location, email [email protected].

Our Hearts are Weary ~ September 11 Madison Action

MOBILIZATION AGAINST MILITARISM

Sept. 11, 2021

Downtown Madison Farmer’s Market

On Sept. 11, the 20th anniversary of 9/11, our hearts are weary after at least 20 years of the “forever wars” and skyrocketing Pentagon budgets.  So many innocent children, women, and men have been killed, in Iraq, in Afghanistan, in Yemen, and so many other places in the Middle East and Africa.  And we are no closer to bringing peace.  War WAS NOT the answer!


End the ever-increasing military profiteering

End the violations of human rights, natural rights and legal rights


8:00 am – noon:  FLY KITES, NOT DRONES Join us at King St. corner of the square. The Wisconsin Network for Peace and Justice table will have kites, leaflets, buttons, and other information.

10:00 am:   ALL ARE WELCOME AS WE gather for a rally outside Senator Baldwin’s office at 30 West Mifflin St. (near the State St. corner of the square).

Speakers include: 

Bonnie Block activist and rebel with Wisconsin Coalition to Ground the Drones and End the Wars

Vicki Berenson, human being for peace and justice, WNPJ, Friends, Safe Skies Clean Water Wisconsin Coalition 

Allen Ruff, historian, concerned citizen and activist 

 

Senator Baldwin, stop voting for the War Machine, fund human needs, and lead us to a peace economy!

The U.S. government spends more on the military than the next top eleven countries combined.  And we are no closer to bringing peace!

The F-35 Joint Strike Fighter jet program, one of the biggest boondoggles in military history, are looking to come to Madison.  And we are no closer to bringing peace!

As we move closer and closer to ecocide, the military remains the biggest source of pollution on earth.  And we are no closer to bringing peace!

Suicide, PTSD, and moral injury is a huge problem for our veterans who have been emotionally, psychologically, and spiritually damaged from our wars.  And we are no closer to bringing peace!

Senator Tammy Baldwin, along with so many other Democrats and Republicans, does not represent our desire to bring an end to the “forever wars”.  We have no party for peace.  We only have a War Party!  Sen. Baldwin actively worked to bring the F-35s to Madison, she does not provide strong support for policies to make the world green, she backs economic policies that support militarism, and she continues to vote for the obscene military budget.

 

We have had enough of these crimes and injustice.   Senator Baldwin needs to be a strong leader for the people.  She needs to stand up to the War Party!

There will be a nonviolent civil resistance action in front of Baldwin’s office.  If you are interested or want more information, contact [email protected]

 

Sponsored by:  Wisconsin Coalition to Ground the Drones and End the Wars

If you would like your group to co-sponsor, contact [email protected]

‘We can’t bomb our way out of a pandemic’

John Nichols | ‘We can’t bomb our way out of a pandemic,’ says Pentagon critic Mark Pocan

24 August 2021
Cap Times

Original Link 


“U.S. Rep. Mark Pocan is a voice of sanity amidst all the partisan wrangling over the messy withdrawal of U.S. troops from Afghanistan.

Of course, he’s been frustrated by the disorderly evacuation process, and he has been outspoken in arguing for the rapid resettlement of refugees.

“The United States has a duty to honor its commitment to the Afghanis who helped our efforts, such as guides and translators,” he said. “We must ensure these refugees are allowed to reach our shores as safely and as quickly as possible.”

Yet he remains a thoughtful supporter of the decision to end the U.S. mission in Afghanistan.

A longtime critic of U.S. military adventurism, Pocan supports the withdrawal as an acknowledgement of the reality that nation building of the sort that was initiated almost two decades ago by the administration of Vice President Dick Cheney and President George W. Bush was never going to work.

“You can’t go into a country like Afghanistan and expect to leave it with Dunkin Donuts and Disneylands,” he said. “And too often I think that that is what we try to do.”

What should we do?


Yes. Cut the spending on profiteering waste and fraud. It is wrong for the US government to violate the rights of other people.

The US government is a gangster, not the world police.  Murdering people to steal their resources or other reasons is a violation of their natural rights, human rights and likely their legal rights.  Americans are not the only people who enjoy liberty.

#WarIsARacket

#Accountability for all.


~ Ratified Treaties have the POWER of US law  ~

 

This Constitution, and the Laws of the United States which shall be made in Pursuance thereof; and all Treaties made, or which shall be made, under the Authority of the United States, shall be the supreme Law of the Land; and the Judges in every State shall be bound thereby, any Thing in the Constitution or Laws of any State to the Contrary notwithstanding. 


~ Charter of the United Nations ~

Chapter VII — Action with respect to Threats to the Peace, Breaches of the Peace, and Acts of Aggression


~ Convention II Article 2 Geneva Conventions ~ 


Shatter the Myths of Empire – Matt McKenna