Should we fear artificial intelligence?
an event to discuss whether we can be optimistic about AI: can we trust that it than they actually are and systematically oppose technological progress.
Should We Fear the Robot Revolution? (The Correct Answer is Yes
In their model technological progress proceeds on two fronts: automation and the creation of new more complex tasks that only human labor can perform. A repre-.
GAFAM a progress and a danger for civilization
17 avr. 2019 All these companies are great users of personal data but the abuses that we start to measure better and better
Fear and Loathing of Technological Progress? Leveraging Science
“Technological change and productivity are primary sources of our unprecedented wealth but many persons have not shared in that abundance. We recommend
Facing your Fears - How to Do Exposure
Make sure to track your progress. See the Facing Fears form which will help you identify how anxious you were before and after facing the feared situation
HOW CAN HUMANS KEEP THE UPPER HAND?
The hopes and fears sparked by artificial intelligence today have raised it to mythological yet we are all too aware of the risks the creation of such.
Should we fear transition risks? A review of the applied literature
22 mars 2021 Should we thus fear transition risks? ... been on “sunset industries” like fossils given the slow progress of low-carbon investments. This.
Accelerating progress toward the reduction of adolescent pregnancy
There is plenty of evidence in the Region that we are capable of successfully confronting public health challenges. The Americas are leading the world in the
To what extent should we endow machines with intelligence?
Do you think that progress is always positive? When you hear the phrase “Artificial Intelligence” what do you have in ... R.A : Should we fear AI ?
Should we fear artificial intelligence?
Many experts believe that we are confronted with an inflection point in research will very likely be reduced in favour of more rapid progress and ...
Idea of progress: should we fear progress? - Fiche - Laura Debourg
So we are going to look at this question: "should we fear progress?" I On the one hand I think progress is essential to evolve and it is indispensable for
Idea of progress: Should We fear progress ? - Fiche - martymai
I'm going to talk about the « idea of progress However some people can consider it as a negative thing because progress may go beyond ethical morals
[PDF] Should We Fear the Robot Revolution? (The Correct Answer is Yes)
In their model technological progress proceeds on two fronts: automation and the creation of new more complex tasks that only human labor can perform A repre-
[PDF] Should we fear artificial intelligence? - European Parliament
The path to a more automated society in which the only complex intelligence is not human is not exempt from damages and fear Our ancestrally pessimistic bias
The idea of progress: Should we fear the progress - Essaysclub
Thenin a movie Splice starring by Vincenzo Natali show us that the engineers manage to mix DNA of different animals to create a nex species for medical use So
[PDF] Should We Fear Derivatives?
The derivative must be worth the same as the replicating portfolio if financial markets are frictionless since otherwise there is an opportunity to make a risk
[PDF] Fear and Loathing of Technological Progress? Leveraging Science
31 jan 2018 · The answer proposed here is that there are two ways in which technological change can be leveraged to support the implementation of the
[PDF] Afraid? Of What? - cloudfrontnet
12 jui 2018 · When we fear one another our neighbors become the “other” and their problems are allowed to fester In a time where ideologies and invisible
Fear and The Inhibition of Progress PDA Journal of Pharmaceutical
This is a PDF-only article The first page of the PDF of this article appears below PDA members receive access to all articles published in the current
![Should we fear artificial intelligence? Should we fear artificial intelligence?](https://pdfprof.com/Listes/18/7669-18EPRS_IDA_2018_614547_EN.pdf.pdf.jpg)
Should we fear
artificial intelligence?Should we fear artificial intelligence?
In-depth Analysis
March 2018
PE 614.547
STOA - Science and Technology Options Assessment
2AUTHORS
Peter J. Bentley, University College London
Miles Brundage, University of Oxford
Thomas Metzinger, Johannes Gutenberg University of Mainz With a foreword by María Teresa Giménez Barbat, MEP and an introduction by Philip Boucher, Scientific Foresight Unit (STOA)STOA ADMINISTRATOR RESPONSIBLE
Philip Boucher
Scientific Foresight Unit (STOA)
Directorate for Impact Assessment and European Added Value Directorate-General for Parliamentary Research Services European Parliament, Rue Wiertz 60, B-1047 BrusselsE-mail: STOA@ep.europa.eu
LINGUISTIC VERSION
Original: EN
ABOUT THE PUBLISHER
To contact STOA or to subscribe to its newsletter please write to: STOA@ep.europa.eu This document is available on the Internet at: http://www.ep.europa.eu/stoa/Manuscript completed in March 2018
Brussels, © European Union, 2018
DISCLAIMER
This document is prepared for, and addressed to, the Members and staff of the European Parliament asbackground material to assist them in their parliamentary work. The content of the document is the sole
responsibility of its author(s) and any opinions expressed herein should not be taken to represent an
official position of the Parliament. Reproduction and translation for non-commercial purposes are authorised, provided the source is acknowledged and the European Parliament is given prior notice and sent a copy.Picture credit: © José María Beroy
PE 614.547
ISBN 978-92-846-2676-2
doi: 10.2861/412165QA-01-18-199-EN-N
Should we fear the future of artificial intelligence? 3Table of contents
1. Foreword ........................................................................................................................................... 4
2. Introduction ....................................................................................................................................... 5
3. The Three Laws of Artificial Intelligence: Dispelling Common Myths .................................... 6
4. Scaling Up Humanity: The Case for Conditional Optimism about Artificial Intelligence .. 13
5. Remarks on Artificial Intelligence and Rational Optimism ..................................................... 19
6. Towards a Global Artificial Intelligence Charter ....................................................................... 27
STOA - Science and Technology Options Assessment
41. Foreword
María Teresa Giménez Barbat, MEP
For some years now, artificial intelligence (AI), has been gaining momentum. A wave of programmesthat get the maximum performance out of latest generation processors are obtaining spectacular results.
One of the most outstanding AI applications is voice recognition: while the first models were awkward
and marked by constant defects, they are now capable of responding correctly to all sorts of userrequests in the most diverse situations. In the field of image recognition, remarkable advances are also
being made, with programs able to recognise figures ² and even cats ² in online videos now being
adapted for the software to control the autonomous cars set to invade our streets in the coming years.
Today, we cannot imagine a future in Europe without advanced AI that will impact more and morefacets of our lives, from work to medicine, and from education to interpersonal relations. In February
2017, the European Parliament approved a report with recommendations for the European Commission
on civil law rules for robotics. Many Members of Parliament (MEPs) heard a series of curiousH[SUHVVLRQV SRVVLNO\ IRU POH ILUVP PLPH ŃRQŃHSPV VXŃO MV ´LQPHOOLJHQP MXPRQRPRXV URNRPµ and even
´HOHŃPURQLŃ SHUVRQMOLP\µB
Any future legislation in this field that aims to be truly useful, favouring progress and benefitting the
biggest possible number of citizens, needs to be based on a dialogue with experts. This concern lies at
the heart of my request to the Science and Technology Options Assessment (STOA) Panel to organisean event to discuss whether we can be optimistic about AI: can we trust that it will benefit society? We
succeeded in bringing together a panel headed up by the Harvard psychology professor and scientificauthor Steven Pinker. He was accompanied by Peter John Bentley, computational scientist from
University College London; MLOHV %UXQGMJH IURP 2[IRUG 8QLYHUVLP\·V )XPXUH RI +XPMQLP\ HQVPLPXPH book Here be dragons, and the philosopher, Thomas Metzinger, from the University of Mainz and us texts providing the basis for the following collection.What the reader holds is a collection of papers dealing with some of the ideas I consider particularly
useful for politicians and legislators. For instance, it is essential not to give in to the temptation to
legislate on non-existent problems. The path to a more automated society, in which the only complexintelligence is not human, is not exempt from damages and fear. Our ancestrally pessimistic bias makes
us see things in a worse light than they actually are and systematically oppose technological progress,
and also gives us the ability to generate e[RUNLPMQP IHMUV VXŃO MV POH LGHM POMP M ´VXSHULQPHOOLJHQŃHµ RLOO
inevitably turn against +XPMQLP\ MQG PULJJHU M ´SRVP-OXPMQµ IXPXUHB $ŃŃRUGLQJ PR 3HPHU %HQPOH\
author of the text The Three Laws of Artificial Intelligence, this myth that AI may constitute an existential
threat for humanity is one of the most widespread and at the root of numerous misunderstandings. AIconsists of mathematical algorithms limited to searching for patterns: the belief that AI may lead to
robots wishing to dominate the world has no basis in reality, but is mere science fiction.Another noteworthy idea is that AI will drive and develop a society of well-being. ´There are myriad
possible malicious uses of AIµ, explains Miles Brundage, but if a series of conditions described in his
article Scaling Up Humanity: The Case for Conditional Optimism about Artificial Intelligence converge, we
can be very optimistic. AI will enable the solution of complex issues and will be attributed the
responsibility for certain decisions, thus avoiding prejudice or abuse. AI will be of spectacular economic
the additional economic value resulting from AI can be cautiously estimated at 30 billion dollars. Thomas Metzinger identifies some of the most important challenges he sees in the future of AI, and proposes a set of accompanying practical recommendations for how the EU could respond. Certainly,we will have to coexist with different degrees of AI. We hope that between us all, we can to overcome
most of our fears and better understand a technology that is already shaping our future. Should we fear the future of artificial intelligence? 52. Introduction
Philip Boucher
Humans are, on the whole, living longer and healthier lives than ever before. For many, these basic measures are enough to conclude that the world is becoming a better place. However, when we look atthe headlines, it is clear that there remains a great deal of human suffering. Indeed, if we consider the
growing threats of climate change, rising sea levels and mass extinction, as well as nuclear threats and
political instability, some would find few reasons to be cheerful. Depending upon which variables weprioritise (equality, biodiversity, violence, poverty, CO2 levels, conflict, ozone layer depletion), and how
we measure them, we can make rational arguments for optimistic or pessimistic views on the future of humanity.The picture is equally mixed when we consider new technologies, such as artificial intelligence (AI),
which are predicted to have a huge impact on the future of humanity, for better or worse. For example,
AI could bring substantial benefits to several aspects of our lives, from weather predictions to cancer
diagnostics. At the same time, concerns have been raised that it could threaten many jobs and take over
important decision-making processes without transparency. Well-known figures have joined both sides of the debate. For example, Elon Musk shared concerns thatAI posed an existential threat to the human race, while Bill Gates countered that the technology will
make us more productive and creative. Beyond the headlines, however, both Gates and Musk recognisethat AI presents a wide range of opportunities and challenges, and both call for reflection on how we
can manage its development in a way that maximises its benefits without exposing us to danger.Our hopes and fears about AI are not only about far-flung futures. They are often about PRGM\·V $H
which already has a substantial influence on our lives, and seemingly for both better and worse. For example, AI is part of both the problem and solution to fake news. AI algorithms have been used to support more impartial criminal justice, yet are accused of racial bias. While nobody can predict how AI will develop in the future, it seems that we will encounter manychallenges and opportunities, some more serious than others. If there were a single rational position on
the future of AI, it would certainly be more nuanced than unbridled optimism or crippling fear. Until
we know more about the impacts of AI and the capabilities of humanity to respond to them, it is important to create spaces where we can observe, reflect and debate the issues and, where necessary,prepare appropriate responses. This debate must remain open to a wide range of disciplines. The science
and engineering community has an important role to play, particularly in considering the boundaries of what is technically possible. On the other hand, understanding the development and impact of technology in society requires social scientific expertise. No discipline has a monopoly on wisdom. It is in this context that, on 19 October 2017, STOA hosted a workshop at the European Parliament toconsider whether it is rational to be optimistic about AI. Steven Pinker (Harvard University) opened the
event with a lecture on the broad concept of rational optimism. This was followed by four speakers from
different disciplines ² Peter J. Bentley, a computer scientist from University College London,
statistician from Chalmers University, and Thomas Metzinger, a philosopher from Johannes Gutenberg University of Mainz ² who presented their own positions on whether we should fear AI. The livelydebate remains available online, and we are very pleased that the four speakers agreed to refine their
perspectives into individual position papers which are published together in this collection. We gave
the authors carte blanche to set out their arguments on their own terms and in their own style, with the
aim of making a useful contribution to ongoing debates about AI in the parliamentary community and beyond. Given the increasing attention to the subject amongst MEPs and citizens alike, there will be many more debates and publications in the years to come.STOA - Science and Technology Options Assessment
63. The Three Laws of Artificial Intelligence: Dispelling Common
MythsPeter J. Bentley
Introduction
Artificial intelligence (AI) is fashionable today. After some notable successes in new AI technologies,
and new applications, it is seeing a resurgence of interest, which has resulted in a surge of opinions
from many disciplines. These include from laypeople, politicians, philosophers, entrepreneurs and professional lobbyists. However, these opinions rarely include those from the people who understandAI the most: the computer scientists and engineers who spend their days building the smart solutions,
applying them to new products, and testing them. This article provides the views of a computer scientist
experienced in the creation of AI technologies in an attempt to provide balance and informed opinion on the subject.Debunking Myths
One of the most extraordinary claims that is oft-repeated, is that AI is somehow a danger to humankind,
HYHQ MQ ´H[LVPHQPLMO POUHMPµB 6RPH ŃOMLP POMP MQ AI might somehow develop spontaneously and
ferociously like some exponentially brilliant cancer. We might start with something simple, but theintelligence improves itself out of our control. Before we know it, the whole human race is fighting for
its survival (Barrat, 2015).It all sounds absolutely terrifying (which is why many science fiction movies use this as a theme). But
despite earnest commentators, philosophers, and people who should know better than spreading thesestories, the ideas are pure fantasy. The truth is the opposite: AI ² like all intelligence ² can only develop
VORRO\ XQGHU MUGXRXV MQG SMLQIXO ŃLUŃXPVPMQŃHVB HP·V QRP HMV\ NHŃRPLQJ ŃOHYHr.There have always been two types of AI: reality and fiction. Real AI is what we have all around us ² the
voice-recognising Siri or Echo, the hidden fraud detection systems of our banks, even the number-plate
reading systems used by the police (Aron, 2011; Siegel, 2013; Anagnostopoulos, 2014). The reality of AI
is that we build hundreds of different and highly-specialised types of smart software to solve a million
different problems in different products. This has been happening since the birth of the field of AI,
which is contemporary with the birth of computers (Bentley, 2012). AI technologies are already
embedded within software and hardware all around us. But these technologies are simply clever tech. They are the computational equivalents to cogs and springs in mechanical devices. And like a brokencog or loose spring, if they fail then that particular product might fail. Just as a cog or spring cannot
magically turn itself into a murderous killing robot, our smart software embedded within their products
cannot turn itself into a malevolent AI. Real AI saves lives by helping to engage safety mechanisms (automatic braking in cars, or even self-driving vehicles). Real AI helps us to optimise processes or predict failures, improving efficiency and
reducing environmental waste. The only reason why hundreds of AI companies exist, and thousandsof researchers and engineers study in this area, is because they aim to produce solutions that help people
and improve our lives (Richardson, 2017).The other kind of AI ² comprising those super-intelligent general AIs that will kill us all ² is fiction.
Research scientists tend to work on the former kind of AI. But because this article needs to providebalance in favour of rational common sense, the following sections will dispel several myths in this area.
HQ POLV MUPLŃOH H RLOO LQPURGXŃH ´7OUHH IMRV RI $Hµ MV M RM\ PR H[SOMLQ RO\ POH P\POV MUH IMQPMVPLŃMO LI
in AI, simplified for the layperson. Should we fear the future of artificial intelligence? 7 Myth 1: A self-modifying AI will make itself super-intelligent.6RPH ŃRPPHQPMPRUV NHOLHYH POMP POHUH LV VRPH GMQJHU RI MQ $H ´JHPPLQJ ORRVHµ MQG ´PMNLQJ LPVHOI VXSHU-
The first law of AI tells us why this is not going to happen.First law of AI: Challenge begets intelligence.
From our research in the field of artificial life (ALife) we observe that intelligence only exists in order to
overcome urgent challenges. Without the right kinds of problems to solve, intelligence cannot emergeor increase (Taylor et al., 2014). Intelligence is only needed where those challenges may be varied and
unpredictable. Intelligence will only develop to solve those challenges if its future relies on its success.
To make a simple AI, we create an algorithm to solve one specific challenge. To grow its intelligence
into a general AI, we must present ever-more complex and varied challenges to our developing AI, anddevelop new algorithms to solve them, keeping those that are successful. Without constant new
challenges to solve, and without some reward on success, our AIs will not gain another IQ point.AI researchers know this all too well. A robot that can perform one task well, will never grow in its
abilities without us forcing it to grow (Vargas et al., 2014). For example, the automatic number plate
recognition system used by police is a specialised form of AI designed to solve one specific challenge ²
reading car number plates. Even if some process were added to this simple AI to enable it to modifyitself, it would never increase its intelligence without being set a new and complex challenge. Without
an urgent need, intelligence is simply a waste of time and effort. Looking at the natural world this is
illustrated in abundance ² most challenges in nature do not require brains to solve them. Only very few
organisms have needed to go to the extraordinary efforts needed to develop brains. Even fewer develop
highly complex brains.The first law of AI tells us that artificial intelligence is a tremendously difficult goal, requiring exactly the right
conditions and considerable effort. There will be no runaway AIs, there will be no self-developing AIs out of our
control. There will be no singularities. AI will only be as intelligent as we encourage (or force) it to be, under
duress. As an aside, even if we could create a super-intelligence, there is no evidence that such a super- intelligent AI would ever wish to harm us. Such claims are deeply flawed, perhaps stemming fromobservations of human behaviour, which is indeed very violent. But AIs will not have human
intelligence. Our real future will almost certainly be a continuation of the situation today: AIs will co-
evolve with us, and will be designed to fit our needs, in the same way that we have manipulated crops,
cattle and pets to fit our needs (Thrall et al., 2010). Our cats and dogs are not planning to kill all humans.
Likewise, a more advanced AI will fit us so closely that it will become integrated within us and our societies. It would no more wish to kill us than it would kill itself. Myth 2: With enough resources (neurons/computers/memory) an AI will be more intelligent than humans.FRPPHQPMPRUV ŃOMLP POMP ´PRUH LV NHPPHUµB HI M OXPMQ NUMLQ OMV M OXQGUHG NLOOLRQ QHXURns, then an AI
with a thousand billion simulated neurons will be more intelligent than a human. If a human brain is equivalent to all the computers of the Internet, then an AI loose in the Internet will have humanintelligence. In reality, it is not the number that matters, it is how those resources are organised, as the
second law of AI explains. Second law of AI: Intelligence requires appropriate structure.7OHUH LV QR ´RQH VL]H ILPV MOOµ IRU NUMLQ VPUXŃPXUHVB (MŃO NLQG RI ŃOMOOHQJH UHTXLUHV M QHR GHVLJQ to solve
it. To understand what we see, we need a specific kind of neural structure. To move our muscles, we need another kind. To store memories, we need another. Biology shows us that you do not need manySTOA - Science and Technology Options Assessment
8 neurons to be amazingly clever. The trick is to organise them in the right way, building the optimal algorithm for each problem (Garner and Mayford, 2012).JO\ ŃMQ·P RH XVH PMPOV PR PMNH $HV"
We do use a lot of clever maths and because of this some Machine Learning methods produce
predictable results, enabling us to understand exactly what these AIs can and cannot do. However, most
practical solutions are unpredictable, because they are so complex and they may use randomness within
their algorithms meaning that our mathematics cannot cope, and because they often receive
unpredictable inputs. While we do not have mathematics to predict the capabilities of a new AI, we do
have mathematics that tells us about the limits of computation. Alan Turing helped invent theoretical
computer science by telling us about one kind of limit ² we can never predict if any arbitrary algorithm
LQŃOXGLQJ MQ $H RLOO HYHU OMOP LQ LPV ŃMOŃXOMPLRQV RU QRP 7XULQJ 1E37B JH MOVR OMYH POH ´1R )UHH IXQŃO
7KHRUHPquotesdbs_dbs33.pdfusesText_39
[PDF] compte rendu tp spectroscopie uv visible
[PDF] quiproquo entre deux amis
[PDF] l'avare quiproquo entre harpagon et valere
[PDF] exemple de quiproquo dans l'avare
[PDF] quiproquo comique
[PDF] faut il préférer le bonheur ? la vérité dissertation
[PDF] faut il préférer la vérité ? l illusion dissertation philosophie
[PDF] émigration italienne aux etats unis
[PDF] attractivité académie
[PDF] académie déficitaire enseignant 2016
[PDF] académie déficitaire enseignant 2017
[PDF] académies les moins demandées capes
[PDF] détachement sur contrat
[PDF] détachement fonction publique territoriale vers fonction publique d'état