- From: Paola Di Maio <paola.dimaio@gmail.com>
- Date: Wed, 17 Jun 2020 09:34:49 +0800
- To: ProjectParadigm-ICT-Program <metadataportals@yahoo.com>
- Cc: W3C AIKR CG <public-aikr@w3.org>
- Message-ID: <CAMXe=SquLkNUHb747PhEq=PKBTG+F5ru4XA8xxrz1f=W70WV+g@mail.gmail.com>
Thanks for sharing. The bottom line perhaps is can humans, or anything be trustworthy? Especially when most of the tricks we do play upon ourselves - in the sense that I have never seen a model of infallibility in the human realm On Wed, Jun 17, 2020 at 12:30 AM ProjectParadigm-ICT-Program < metadataportals@yahoo.com> wrote: > An excerpt from the EU EDRI watchdog organizations. > > EDRi-gram > > fortnightly newsletter about digital civil rights in Europe > > EDRi-gram 18.11, 10 June 2020 > Read online: https://edri.org/edri-gram/18-11/ > > ======================================================================= > Contents > ======================================================================= > > 1. Can the EU make AI “trustworthy”? No – but they can make it just > 2. COVID-Tech: the sinister consequences of immunity passports > 3. UK: Stop social media monitoring by local authorities > 4. Cryptocurrency scammers flood Facebook users with manipulative ads > 5. SHARE’s campaign bears fruit: Google appoints Serbian representatives > 6. Recommended Action > 7. Recommended Reading > 8. Agenda > 9. About > > ======================================================================= > 1. Can the EU make AI “trustworthy”? No – but they can make it just > ======================================================================= > > Today, 4 June 2020, European Digital Rights (EDRi) submitted its answer > to the European Commission’s consultation on the AI White Paper. On top > of our response, in our additional paper we outline recommendations to > the European Commission for a fundamental rights- based AI regulation. > You can find our consultation response, recommendations paper, and > answering guide for the public here. > > How to ensure a “trustworthy AI” has been highly debated since the > European Commission launched its White Paper on AI in February this > year. Policymakers and industry have hosted numerous conversations about > “innovation”, “Europe becoming a leader in AI”, and promoting a “Fair > AI”. > > Yet, a “fair” or “trustworthy” artificial intelligence seems a far way > off. As governments, institutions and industry swiftly move to > incorporate AI into their systems and decision-making processes - grave > concerns remain as to how these changes will impact people, democracy > and society as a whole. > > EDRi’s response outlines the main risks AI poses for people, communities > and society, and outlines recommendations for an improved, truly > ‘human-centric’ legislative proposal on AI. We argue that the EU must > reinforce the protections already embedded in the General Data > Protection Regulation (GDPR), outline clear legal limits for AI by > focusing on impermissible use, and foreground principles of collective > impact, democratic oversight, accountability, and fundamental rights. > Here’s a summary of our main points. > > Put people before industrial policy > > A ‘human centric’ approach to AI requires that considerations of safety, > equality, privacy, and fundamental rights are the primary factors > underpinning decisions as to whether to promote or invest in AI. > > However, the European Commission’s White Paper proposal takes as a point > of a departure the inherent economic benefits of promoting AI, > particularly in the public sector. Promoting AI in the public sector as > a whole, without requiring scientific evidence to justify the need or > the purpose of such applications in some potentially harmful situations, > is likely to have the most direct consequences on everyday peoples’ > lives, particularly on marginalised groups. > > Despite wide ranging applications that could advance our societies (such > as some uses in the field of health), we have also seen the vast > negative impacts of automated systems at play at the border, in > predictive policing systems which exacerbate overpolicing of racialised > communities, in ‘fraud detection’ systems which target poor, working > class and migrant areas, and countless more examples. All such examples > highlight the potentially devastating consequences AI systems can have > in the public sector, contesting the case for ‘promoting the uptake of > AI.’ These examples highlight the need for AI regulation to be rooted in > a human-centric approach. > > "The development of artificial intelligence technology offers huge > potential opportunities for improving our economies and societies, but > also extreme risks. Poorly-designed and governed AI will exacerbate > power imbalances and inequality, increase discrimination, invade privacy > and undermine a whole host of other rights. EU legislation must ensure > that cannot happen. Nobody's rights should be sacrificed on the altar of > innovation." said Chris Jones, Statewatch. > > Address collective harms of AI > > The vast potential scale and impact AI systems challenges existing > conceptions of harm. Whilst in many ways we can view the challenges > posed by AI as fundamental rights issues, often the harms perpetrated > are much broader, disadvantaging communities, economy, democracy and > entire societies. From the impending threat of mass surveillance as a a > result of biometric processing in publicly-accessible spaces, to the use > of automated systems or ‘upload filters’ to moderate content on social > media, to severe disruptions to the democratic process, we see the > impact goes far beyond the level of the individual. One specificity of > regulating AI is the need to address societal-level harms. > > Prevent harms by focusing on impermissible use > > Just as the problems with AI are collective and structural, so must be > the solutions. The European Commission’s White Paper outlines some > safeguards to address ‘high-risk’ AI, such as training data to correct > for bias and ensuring human oversight. Whilst these safeguards are > crucial, they will not address the irreparable harms which will result > from a number of uses of AI. > > “The EU must move beyond technical fixes for the complex problems posed > by AI. Instead, the upcoming AI regulation must determine the legal > limits, impermissible uses or ‘red-lines’ for AI applications. This is a > necessary step for a people-centered, fundamental rights-based AI” says > Sarah Chander, Senior Policy Adviser, EDRi. > > The EDRi network lists some of the impermissible uses of AI: > > - indiscriminate biometric surveillance and biometric capture and > processing in public spaces [1] > - use of AI to solely determine access to or delivery of essential > public services (such as social security, policing, migration control) > - uses of AI which purport to identify, analyse and assess emotion, > mood, behaviour, and sensitive identity traits (such as race, > disability) in the delivery of essential services > - predictive policing > - autonomous lethal weapons and other uses which identify targets for > lethal force (such as law and immigration enforcement) > > "The EU must ensure that states and companies meet their obligations and > responsibilities to respect and promote human rights in the context of > automated decision-making systems. EU institutions and national > policymakers must explicitly recognise that there are legal limits to > the use and impact of automation. No safeguard or remedy would make > indiscriminate biometric surveillance or predictive policing acceptable, > justified or compatible with human rights" said Fanny Hidvegi, Europe > Policy Manager at Access Now. > > Require democratic oversight for AI in the public sphere > > The rapidly increasing deployment of AI systems presents a major > governance issue. Due to the (designed) opacity of the systems, the > complete lack of transparency from governments when such systems are > deployed for use in public, essential functions, and the systematic lack > of democratic oversight and engagement – AI is furthering the ‘power > asymmetry between those who develop and employ AI technologies, and > those who interact with and are subject to them [2]. > > As a result, decisions impacting public services will be more opaque, > increasingly privately owned, and even less subject to democratic > oversight. It is vital that the EU’s regulatory proposal on AI addresses > this – implementing mandatory measures of democratic oversight for the > procurement and deployment of AI in the public sector and essential > services. More, the EU must explore methods of direct public engagement > on AI systems. In this regard, authorities should be required to > specifically consult marginalised groups likely to be disproportionately > impacted by automated systems. > > Implement the strongest possible fundamental rights protections > > Regulation on AI must reinforce, rather than replace, the protections > already embedded in the General Data Protection Regulation (GDPR). The > European Commission has the opportunity to complement these protections > with safeguards for AI. To put people first and provide the strongest > possible protections, all systems should complete mandatory human rights > impact assessments. This assessment should evaluate the collective, > societal, institutional and governance implications the system poses, > and outline adequate steps to mitigate this. > > "The deployment of such systems for predictive purposes comes with high > risks on human rights violations. Introducing ethical guidelines & > standards for the design and deployment of these tools is welcome, but > not enough. Instead, we need the European Union and Member States to > ensure compliance with the applicable regulatory frameworks, and draw > clear legal limits to ensure AI is always compatible with fundamental > rights." says Eleftherios Chelioudakis from Homo Digitalis. > > EDRi’s position calls for fundamental rights to be prioritised in the > regulatory proposal for all AI systems, not only those categorised as > ‘high-risk’. We argue AI regulation should avoid creating loop-holes or > exemptions based on sector, size of enterprise, or whether or not the > system is deployed in the public sector. > > "It is crucial for the EU to recognize that the adoption of AI > applications is not inevitable. The design, development and deployment > of systems must be tested against human rights standards in order to > establish their appropriate and acceptable use. Red lines are thus an > important piece of the AI governance puzzle. Recognizing impermissible > use at the outset is particularly important because of the > disproportionate, unequal and sometimes irreversible ways in which > automated decision making systems impact societies." said Vidushi Marda, > Senior Programme Officer, at ARTICLE 19 > > The rapid uptake of AI will fundamentally change our society. From a > human rights’ perspective, AI systems have the ability to exacerbate > surveillance and intrusion into our personal lives, fundamentally alter > the delivery of public and essential services, vastly undermine vital > data protection legislation, and disrupt the democratic process. > > For some, AI will mean reinforced, deeper harms as such systems feed and > embed existing processes of marginalisation. For all, the route to > remedies, accountability, and justice will be ever-more unclear, as this > power asymmetry further shifts to private actors, and public goods and > services will be not only automated, but privately owned. > > There is no “trustworthy AI” without clear red-lines for impermissable > use, democratic oversight, and a truly fundamental rights-based approach > to AI regulation. The European Union’s upcoming legislative proposal on > artificial intelligence (AI) is a major opportunity change this; to > protect people and democracy from the escalating economic, political and > social issues posed by AI. > > Footnotes: > > [1] EDRi (2020). ‘Ban Biometric Mass Surveillance: A set of fundamental > rights demands for the European Commission and Member States’ > > https://edri.org/wp-content/uploads/2020/05/Paper-Ban-Biometric-Mass-Surveillance.pdf > > > [2] Council of Europe (2019). ‘Responsibility and AI DGI(2019)05 > Rapporteur: Karen Yeung > https://rm.coe.int/responsability-and-ai-en/168097d9c5 > > Read more: > > EDRi Consultation Response: European Commission Consultation on the > White Paper on Artificial Intelligence > https://edri.org/wp-content/uploads/2020/06/AI_EDRiConsultationResponse.pdf > > EDRI Recommendations for a Fundamental-rights based Artificial > Intelligence Regulation: Addressing collective harms, democratic > oversight and impermissable use > https://edri.org/wp-content/uploads/2020/06/AI_EDRiRecommendations.pdf > > Access Now Consultation Response: European Commission Consultation on > the White Paper on Artificial Intelligence > https://www.accessnow.org/EU-white-paper-consultation > > Bits of Freedom (2020). ‘Facial recognition: A convenient and efficient > solution, looking for a problem?’ > > https://www.bitsoffreedom.nl/2020/01/29/facial-recognition-a-convenient-and-efficient-solution-looking-for-a-problem/ > > > EDRi (2020). ‘Ban Biometric Mass Surveillance: A set of fundamental > rights demands for the European Commission and Member States’ > > https://edri.org/wp-content/uploads/2020/05/Paper-Ban-Biometric-Mass-Surveillance.pdf > > > Privacy International and Article 19 (2018). ‘Privacy and Freedom of > Expression in the Age of Artificial Intelligence’ > > https://www.article19.org/wp-content/uploads/2018/04/Privacy-and-Freedom-of-Expression-In-the-Age-of-Artificial-Intelligence-1.pdf > > > ======================================================================= > 2. COVID-Tech: the sinister consequences of immunity passports > > > Milton Ponson > GSM: +297 747 8280 > PO Box 1154, Oranjestad > Aruba, Dutch Caribbean > Project Paradigm: Bringing the ICT tools for sustainable development to > all stakeholders worldwide through collaborative research on applied > mathematics, advanced modeling, software and standards development >
Received on Wednesday, 17 June 2020 01:35:41 UTC