Should we ban killer robots
- Super Nova
- Posts: 11786
- Joined: Sat Dec 15, 2007 12:49 am
- Location: Overseas
Should we ban killer robots
I don't think we can.
Military advantage no matter what the risk will be pursued. While I admire the risks being highlighted... they will be developed. If they get out of control... well we will go to war with them. terminator here we come.
Also we need this as the first step to going to the stars. It is a step in the evolution of thinking machines that are fully autonomous. Killing each other will help refine this technology.
Ban Killer Robots before They Become Weapons of Mass Destruction
We need an international agreement to prevent the development of autonomous weapons before they threaten global security
Last week the Future of Life Institute released a letter signed by some 1,500 artificial intelligence (AI), robotics and technology researchers. Among them were celebrities of science and the technology industry—Stephen Hawking, Elon Musk and Steve Wozniak—along with public intellectuals such as Noam Chomsky and Daniel Dennett. The letter called for an international ban on offensive autonomous weapons, which could target and fire weapons without meaningful human control.
This week is the 70th anniversary of the atomic bombing of the Japanese cities of Hiroshima and Nagasaki, together killing over 200,000 people, mostly civilians. It took 10 years before the physicist Albert Einstein and philosopher Bertrand Russell, along with nine other prominent scientists and intellectuals, issued a letter calling for global action to address the threat to humanity posed by nuclear weapons. They were motivated by the atomic devastation in Japan but also by the escalating arms race of the Cold War that was rapidly and vastly increasing the number, destructive capability, and efficient delivery of nuclear arms, draining vast resources and putting humanity at risk of total destruction. They also note in their letter that those who knew the most about the effects of such weapons were the most concerned and pessimistic about their continued development and use.
The Future of Life Institute letter is significant for the same reason: It is signed by a large group of those who know the most about AI and robotics, with some 1,500 signatures at its release on July 28 and more than 17,000 today. Signatories include many current and former presidents, fellows and members of the American Association of Artificial Intelligence, the Association of Computing Machinery and the IEEE Robotics & Automation Society; editors of leading AI and robotics journals; and key players in leading artificial-intelligence companies such as Google DeepMind, Facebook, and IBM’s Watson team. As Max Tegmark, Massachusetts Institute of Technology physics professor and a founder of the Future of Life Institute, told Motherboard, “This is the AI experts who are building the technology who are speaking up and saying they don’t want anything to do with this.”
Autonomous weapons pose serious threats that, taken together, make a ban necessary. There are concerns whether AI algorithms could effectively distinguish civilians from combatants, especially in complex conflict environments. Even advanced AI algorithms would lack the situational understanding or the ability to determine whether the use of violent force was appropriate in a given circumstance or whether the use of that force was proportionate. Discrimination and proportionality are requirements of international law for humans who target and fire weapons but autonomous weapons would open up an accountability gap. Because humans would no longer know what targets an autonomous weapon might select, and because the effects of a weapon may be unpredictable, there would be no one to hold responsible for the killing and destruction that results from activating such a weapon.
Then, as the Future of Life Institute letter points out, there are threats to regional and global stability as well as humanity. The development of autonomous weapons could very quickly and easily lead to arms races between rivals. Autonomous weapons would reduce the risks to combatants, and could thus reduce the political risks of going to war, resulting in more armed conflicts. Autonomous weapons could be hacked, spoofed and hijacked, and directed against their owners, civilians or a third party. Autonomous weapons could also initiate or escalate armed conflicts automatically, without human decision-making. In a future where autonomous weapons fight autonomous weapons the results would be intrinsically unpredictable, and much more likely lead to the mass destruction of civilians and the environment than to the bloodless wars that some envision. Creating highly efficient automated violence is likely to lead to more violence, not less.
There is also a profound moral question at stake. What is the value of human life if we delegate the responsibility for deciding who lives and who dies to machines? What kind of world do we want to live in and leave for our children? A world in which AI programs and robots have the means and authority to use violent force and kill people? If we have the opportunity to create a world in which autonomous weapons are banned, and those who might use them are stigmatized and held accountable, do we not have a moral obligation to work toward such a world?
We can prevent the development of autonomous weapons before they lead to arms races and threaten global security and before they become weapons of mass destruction. But our window of opportunity for doing so is rapidly closing.
For the past two years, the Campaign to Stop Killer Robots has been urging the United Nations to ban autonomous weapons. The U.N.’s Convention on Certain Conventional Weapons (CCW) has already held two expert meetings on the issue, and our coalition of 54 nongovernmental organizations* from 25 countries is encouraging the CCW to advance these discussions toward a treaty negotiation. We very much welcome the support from this letter but we must continue to encourage the states represented at the CCW to move forward on this issue.
The essential nature of an arms race involves states acting to improve their own short-term interests at the expense of their own and global long-term benefits. As the letter from Einstein and Russell makes clear:
“We have to learn to think in a new way. We have to learn to ask ourselves not what steps can be taken to give military victory to whatever group we prefer, for there no longer are such steps; the question we have to ask ourselves is: What steps can be taken to prevent a military contest of which the issue must be disastrous to all parties?”
We must continue to demand that of our leaders and policy makers work together with other nations to preempt the threats posed by autonomous weapons by banning their development and use, before we witness the mass destruction they threaten to bring.
http://www.scientificamerican.com/artic ... struction/
Military advantage no matter what the risk will be pursued. While I admire the risks being highlighted... they will be developed. If they get out of control... well we will go to war with them. terminator here we come.
Also we need this as the first step to going to the stars. It is a step in the evolution of thinking machines that are fully autonomous. Killing each other will help refine this technology.
Ban Killer Robots before They Become Weapons of Mass Destruction
We need an international agreement to prevent the development of autonomous weapons before they threaten global security
Last week the Future of Life Institute released a letter signed by some 1,500 artificial intelligence (AI), robotics and technology researchers. Among them were celebrities of science and the technology industry—Stephen Hawking, Elon Musk and Steve Wozniak—along with public intellectuals such as Noam Chomsky and Daniel Dennett. The letter called for an international ban on offensive autonomous weapons, which could target and fire weapons without meaningful human control.
This week is the 70th anniversary of the atomic bombing of the Japanese cities of Hiroshima and Nagasaki, together killing over 200,000 people, mostly civilians. It took 10 years before the physicist Albert Einstein and philosopher Bertrand Russell, along with nine other prominent scientists and intellectuals, issued a letter calling for global action to address the threat to humanity posed by nuclear weapons. They were motivated by the atomic devastation in Japan but also by the escalating arms race of the Cold War that was rapidly and vastly increasing the number, destructive capability, and efficient delivery of nuclear arms, draining vast resources and putting humanity at risk of total destruction. They also note in their letter that those who knew the most about the effects of such weapons were the most concerned and pessimistic about their continued development and use.
The Future of Life Institute letter is significant for the same reason: It is signed by a large group of those who know the most about AI and robotics, with some 1,500 signatures at its release on July 28 and more than 17,000 today. Signatories include many current and former presidents, fellows and members of the American Association of Artificial Intelligence, the Association of Computing Machinery and the IEEE Robotics & Automation Society; editors of leading AI and robotics journals; and key players in leading artificial-intelligence companies such as Google DeepMind, Facebook, and IBM’s Watson team. As Max Tegmark, Massachusetts Institute of Technology physics professor and a founder of the Future of Life Institute, told Motherboard, “This is the AI experts who are building the technology who are speaking up and saying they don’t want anything to do with this.”
Autonomous weapons pose serious threats that, taken together, make a ban necessary. There are concerns whether AI algorithms could effectively distinguish civilians from combatants, especially in complex conflict environments. Even advanced AI algorithms would lack the situational understanding or the ability to determine whether the use of violent force was appropriate in a given circumstance or whether the use of that force was proportionate. Discrimination and proportionality are requirements of international law for humans who target and fire weapons but autonomous weapons would open up an accountability gap. Because humans would no longer know what targets an autonomous weapon might select, and because the effects of a weapon may be unpredictable, there would be no one to hold responsible for the killing and destruction that results from activating such a weapon.
Then, as the Future of Life Institute letter points out, there are threats to regional and global stability as well as humanity. The development of autonomous weapons could very quickly and easily lead to arms races between rivals. Autonomous weapons would reduce the risks to combatants, and could thus reduce the political risks of going to war, resulting in more armed conflicts. Autonomous weapons could be hacked, spoofed and hijacked, and directed against their owners, civilians or a third party. Autonomous weapons could also initiate or escalate armed conflicts automatically, without human decision-making. In a future where autonomous weapons fight autonomous weapons the results would be intrinsically unpredictable, and much more likely lead to the mass destruction of civilians and the environment than to the bloodless wars that some envision. Creating highly efficient automated violence is likely to lead to more violence, not less.
There is also a profound moral question at stake. What is the value of human life if we delegate the responsibility for deciding who lives and who dies to machines? What kind of world do we want to live in and leave for our children? A world in which AI programs and robots have the means and authority to use violent force and kill people? If we have the opportunity to create a world in which autonomous weapons are banned, and those who might use them are stigmatized and held accountable, do we not have a moral obligation to work toward such a world?
We can prevent the development of autonomous weapons before they lead to arms races and threaten global security and before they become weapons of mass destruction. But our window of opportunity for doing so is rapidly closing.
For the past two years, the Campaign to Stop Killer Robots has been urging the United Nations to ban autonomous weapons. The U.N.’s Convention on Certain Conventional Weapons (CCW) has already held two expert meetings on the issue, and our coalition of 54 nongovernmental organizations* from 25 countries is encouraging the CCW to advance these discussions toward a treaty negotiation. We very much welcome the support from this letter but we must continue to encourage the states represented at the CCW to move forward on this issue.
The essential nature of an arms race involves states acting to improve their own short-term interests at the expense of their own and global long-term benefits. As the letter from Einstein and Russell makes clear:
“We have to learn to think in a new way. We have to learn to ask ourselves not what steps can be taken to give military victory to whatever group we prefer, for there no longer are such steps; the question we have to ask ourselves is: What steps can be taken to prevent a military contest of which the issue must be disastrous to all parties?”
We must continue to demand that of our leaders and policy makers work together with other nations to preempt the threats posed by autonomous weapons by banning their development and use, before we witness the mass destruction they threaten to bring.
http://www.scientificamerican.com/artic ... struction/
Always remember what you post, send or do on the internet is not private and you are responsible.
- freediver
- Posts: 3487
- Joined: Fri Dec 14, 2007 10:42 pm
- Contact:
Re: Should we ban killer robots
You can already get autonomous guns, mounted on a slab of concrete. They shoot anything that moves. Not sure why they need the "intelligence" angle to make it scary.
- IQS.RLOW
- Posts: 19345
- Joined: Mon Mar 08, 2010 10:15 pm
- Location: Quote Aussie: nigger
Re: Should we ban killer robots
You shouldn't be scared by intelligence.freediver wrote:You can already get autonomous guns, mounted on a slab of concrete. They shoot anything that moves. Not sure why they need the "intelligence" angle to make it scary.
Just because you are lacking it is no reason to feel frightened by it.
Quote by Aussie: I was a long term dead beat, wife abusing, drunk, black Muslim, on the dole for decades prison escapee having been convicted of paedophilia
- AiA in Atlanta
- Posts: 7258
- Joined: Mon Sep 12, 2011 11:44 pm
Re: Should we ban killer robots
when has man ever shown much if any intelligence about technology? can't imagine we will start now ...
- Rorschach
- Posts: 14801
- Joined: Wed Jun 06, 2012 5:25 pm
Re: Should we ban killer robots
Man has shown much intelligence about technology... he creates it.
DOLT - A person who is stupid and entirely tedious at the same time, like bwian. Oblivious to their own mental incapacity. On IGNORE - Warrior, mellie, Nom De Plume, FLEKTARD
- Neferti
- Posts: 18113
- Joined: Wed Jan 12, 2011 3:26 pm
Re: Should we ban killer robots
SciFi nonsense. Some people read too many SciFi novels.
Wasn't there something the other day about seeing a "woman" on the Moon?
Virgin Mary on a slice of toast ... crying real blood tears and so forth. Miracles.
There is NO HEAVEN. God is a myth.
Wasn't there something the other day about seeing a "woman" on the Moon?
Virgin Mary on a slice of toast ... crying real blood tears and so forth. Miracles.
There is NO HEAVEN. God is a myth.
- Super Nova
- Posts: 11786
- Joined: Sat Dec 15, 2007 12:49 am
- Location: Overseas
Re: Should we ban killer robots
You're kidding .. right.Neferti~ wrote:SciFi nonsense. Some people read too many SciFi novels.
.
This will happen and will happen and while they may not be too smart it will happen in the next 10 years.
We already have drones doing their own thing except push the kill button. They will be empowered with the authority to determine their own target and kill it. What will happen when the blue screen of death occurs...........
Always remember what you post, send or do on the internet is not private and you are responsible.
- Neferti
- Posts: 18113
- Joined: Wed Jan 12, 2011 3:26 pm
Re: Should we ban killer robots
WW3 will END the World, remember? I honestly think we are due for another World War. Not everybody would die but remember "On the Beach"?Super Nova wrote:You're kidding .. right.Neferti~ wrote:SciFi nonsense. Some people read too many SciFi novels.
.
This will happen and will happen and while they may not be too smart it will happen in the next 10 years.
We already have drones doing their own thing except push the kill button. They will be empowered with the authority to determine their own target and kill it. What will happen when the blue screen of death occurs...........
http://www.imdb.com/title/tt0053137/
- AiA in Atlanta
- Posts: 7258
- Joined: Mon Sep 12, 2011 11:44 pm
Re: Should we ban killer robots
You really have a knack for pointing out the obvious so let me rephrase my sentence for you Roach: Man hasn't shown much, if any, intelligence around the use of technology.Rorschach wrote:Man has shown much intelligence about technology... he creates it.
- Rorschach
- Posts: 14801
- Joined: Wed Jun 06, 2012 5:25 pm
Re: Should we ban killer robots
Wrong again AiA and for again very obvious reasons... our world is run and exists due to technology.
DOLT - A person who is stupid and entirely tedious at the same time, like bwian. Oblivious to their own mental incapacity. On IGNORE - Warrior, mellie, Nom De Plume, FLEKTARD
Who is online
Users browsing this forum: No registered users and 72 guests