There is a growing recognition in the UK, Europe and globally that AI specific legal and governmental standards are needed. This page contains links to the most important documents when considering AI and Machine Learning by geographical area and organisation.
Here we outline the various committees, parliamentary reports and case law on AI in the United Kingdom.
The Council of Europe is responsible for the European Convention on Human Rights (ECHR) and the European Court of Human Rights (ECtHR) and it has developed specific human rights standards for many years. We analyse the AI, specific case law and initiatives.
A detailed outline available of European Union AI work.
A brief overview of the work being undertaken by the United Nations.
An outline of AI initiatives, cases and laws in Europe .
An outline of AI initiatives outside the UK and Europe including in the US.
Whilst at present, the UK has no AI specific equality and human rights framework designed to tackle discriminatory technology, there is a data protection framework which also covers algorithms and machine learning (as outlined here) and there are various developments in the pipeline (as outlined below).
There are also numerous governmental and quasi governmental bodies which are interested in or concerned with the development of AI/ML. One commentator has note that there are the following relating to AI –
The Centre for Data Ethics and Innovation; the AI Council; the Office for AI; the House of Lords Select Committee on AI; the House of Commons Inquiry on Algorithms in Decision-Making; the Alan Turing Institute; the National Data Guardian; the Information Commissioner’s Office; a proposed new digital regulator; departmental directorates]; the Office for Tackling Injustices; the Regulatory Horizons Council; … Ofcom;1 NHSX, a new health National AI Lab; AI monitoring by the Health and Safety Executive’s … Foresight Centre; AI analysis from the Government Office for Science; the Office for National Statistics’ Data Science Campus; and the Department of Health and Social Care’s code of conduct for data-driven health and care technology.Veale, Michael. 2019. “A Critical Take on the Policy Recommendations of the EU High-level Expert Group on Artificial Intelligence.” LawArXiv. October 28. Footnotes omitted.
This list is useful but it omits other bodies key to AI/ML issues and some are of less relevance than others. The main bodies this hub follows are as set out below –
Centre for Data Ethics and Innovation
The Government announced the formation of a Centre for Data Ethics and Innovation (CDEI) in 2018. Its terms of reference indicate that it will be formulating advice on best practice but they do not specifically refer to equality and human rights. Its work programme can be seen here.
Most importantly the Government announced that the Centre will conduct an investigation into potential for discriminatory bias in algorithmic decision-making in society. The announcement can be seen here. Its recent call for evidence is available here.
The Centre published two landscape summaries on 19 July 2019 – Landscape summary: bias in algorithmic decision-making and Landscape summary: online targeting. These are important summaries of the situation in the UK. However the understanding of the law relating to equality and non-discrimination is relatively basic and in our view does not adequately address the difference between direct and indirect discrimination. The summaries approach the respective issues from the point of view of fairness which is highly relevant but is not a substitute for carrying out a proper proportionality review.
In September 2019, it published a report looking at the way in which AI is being used in personal insurance. A copy of the report is available here.
Office for Artificial Intelligence
The UK also has an Office for Artificial Intelligence. This is a joint unit of the Departments of Business Energy and Industrial Strategy (BEIS) and of Digital, Culture, Media and Sport (DCMS). It has been doing some interesting work with the Open Data Institute on Data Trust issues.
The OAI has published draft Guidelines for AI Procurement . These emphasise the importance of beig aware of relevant legislation and codes of practice, including data protection and equality law. While the Guidance is useful on data issues it says almost nothing about equality law issues.
Committee on Standards in Public Life
In addition, an “Artificial Intelligence (AI) and Public Standards Review” has recently been announced by the Committee on Standards in Public Life. Details are available here.
The Information Commissioner’s Office (ICO)
The UK Information Commissioner’s Office (ICO) is developing its approach to auditing and supervising AI applications. This includes considering how how AI can play a part in maintaining or amplifying human biases and discrimination as outlined within its blog series which is available here.
The ICO has just launched its own “Tech and innovation hub” in which it promises to collate all its work in this area.
The Surveillance Camera Commissioner (SCO) and the Surveillance Camera Code of Practice
- encourage compliance with the surveillance camera code of practice
- review how the code is working
- provide advice to ministers on whether or not the code needs amending.
The SCO has no enforcement or inspection powers and works with relevant authorities to make them aware of their duty to have regard to the code. The code is not applicable to domestic use in private households. The commissioner also must consider how best to encourage voluntary adoption of the code by other operators of surveillance camera systems
What the SCO is responsible for:
- providing advice on the effective, appropriate, proportionate and transparent use of surveillance camera systems
- reviewing how the code is working and if necessary add others to the list of authorities who must have due regard to the code
- providing advice on operational and technical standards
- encouraging voluntary compliance with the code
However the SCO is not responsible for:
- enforcing the code
- inspecting CCTV operators to check they are complying with the code
- providing advice with regard to covert surveillance
- providing advice with regard to domestic CCTV systems
The Commissioner for the Retention and Use of Biometric Material (‘the Biometrics Commissioner’) was established by the Protection of Freedoms Act 2012. The statute introduced a new regime to govern the retention and use by the police of DNA samples, profiles and fingerprints. The commissioner is independent of government, and is required to –
- keep under review the retention and use by the police of DNA samples, DNA profiles and fingerprints
- decide applications by the police to retain DNA profiles and fingerprints (under section 63G of the Police and Criminal Evidence Act 1984)
- review national security determinations which are made or renewed by the police in connection with the retention of DNA profiles and fingerprints
- provide reports to the Home Secretary about the carrying out of his functions
National Data Guardian for Health and Social Care
The National Data Guardian for Health and Social Care was set up by the Health and Social Care (National Data Guardian) Act 2018, to promote the provision of advice and guidance about the processing of health and adult social care data in England. The Act imposes a duty on public bodies within the health and adult social care sector (and private organisations who contract with them to deliver health or adult social care services) to have regard to the National Data Guardian’s guidance. The Guardian has conducted a consultation on proposed work whcih is now closed but the details and response are available here.
Parliamentary select committees have however taken a much more pro-active lead in establishing a framework for discussing equality and human rights issues relating to AI and machine learning. They show a growing campaign for regulation and control within an ethical framework. The relevant reports are –
- House of Commons Science and Technology Committee The work of the Biometrics Commissioner and the Forensic Science Regulator Nineteenth Report of Session 2017–19 17 July 2019. This report concludes that the Government should issue a moratorium on the current use of facial recognition technology and that there should be no further trials until a legislative framework has been introduced and guidance on trial protocols, and an oversight and evaluation system, has been established.
- House of Commons Science and Technology Committee: Biometrics strategy and forensic services Fifth Report of Session 2017–19, 23 May 2018.
- House of Commons Science and Technology Committee: Algorithms in decision making Fourth Report of Session 2017–19, 15 May 2018.
- House of Commons Culture Media and Sport Committee:
Disinformation and ‘fake news’: 18 February 2018.
The All Party Parliamentary Group on Artificial Intelligence has produced numerous reports since it was set up in 2017. While these are not official Parliamentary documents they are an important resource indicating how Parliamentarians are addressing the issues raised by AI and ML.
Everyone working on AI/ML should be aware of this passage from the judgment of the future Senior Law Lord, Lord Browne-Wilkinson in Marcel and Others v Commissioner of Police of the Metropolis and Others written nearly 30 years ago…
…if the information obtained by the police, the Inland Revenue, the social security offices, the health service and other agencies were to be gathered together in one file, the freedom of the individual would be gravely at risk. The dossier of private information is the badge of the totalitarian state. Apart from authority, I would regard the public interest in ensuring that confidential information obtained by pubic authorities from the citizen under compulsion remains inviolate and incommunicable to anyone as being of such importance that it admitted of no exceptions and overrode all other public interests. The courts should be at least as astute to protect the public interest in freedom from abuse of power as in protecting the public interest in the exercise of such powers. 2 W.L.R. 1118 , at 1130, see also  Ch. 225 at 264.
Until the summer of 2019, few UK cases referred to artificial intelligence, or had any detailed consideration from an equality or human right perspective. That changed with the judgment of the Administrative Court on the 4th September 2019 in R. (o.t.a Bridges) v The Chief Constable of South Wales Police
This case concerned a challenge brought by a member of Liberty to the use of automatic facial recognition (AFR) technology by the South Wales Police (SWP). The police used a system which scanned the public to see if there were faces which matched watch lists. The watch lists concerned different categories of seriousness.
Challenges were brought on three major fronts: a breach of Article 8 of the European Convention on Human Rights, a breach of Data Protection laws; and a breach of the Public Sector Equality Duty (PSED) contained in section 149 of the Equality Act 2010.
The facts were weak. Nothing adverse happened to Mr Bridges and it was not even clear that his face had ever been photographed by the facial recognition technology. It was accepted that if it had been his biometric data would have been destroyed immediately it was found not to match data on the watch lists. Since he was not on the watch lists this would have happened almost immediately.
The Court summarised for the press why the case was dismissed in this way
The Court concluded that SWP’s use of AFR Locate met the requirements of the Human Rights Act. The use of AFR Locate did engage the Article 8 rights of the members of the public whose images were taken and processed  – . But those actions were subject to sufficient legal controls, contained in primary legislation (including the Data Protection legislation), statutory codes of practice, and the SWP’s own published policies  – , and were legally justified  – . In reaching its conclusion on justification, the Court noted that on each occasion AFR Locate was used, it was deployed for a limited time, and for specific and limited purposes. The Court also noted that, unless the image of a member of the public matched a person on the watchlist, all data and personal data relating to it was deleted immediately after it had been processed. On the Data Protection claims, the Court concluded that, even though it could not identify members of the public by name (unless they appeared on a watchlist), when SWP collected and processed their images, it was collecting and processing their personal data  – . The Court further concluded that this processing of personal data was lawful and met the conditions set out in the legislation, in particular the conditions set out in the Data Protection Act 2018 which apply to law enforcement authorities such as SWP  – . The Court was also satisfied that before commencing the trial of AFR Locate, SWP had complied with the requirements of the public sector equality duty  – . The Court concluded that the current legal regime is adequate to ensure the appropriate and non-arbitrary use of AFR Locate, and that SWP’s use to date of AFR Locate has been consistent with the requirements of the Human Rights Act, and the data protection legislation .
This case provides a helpful guide to the way cases such as this are to be analysed. The outcome really reflects the fact that the court was impressed with the care and preparation that had gone into the deployment of AFR. In particular the public had been warned about its use.
One issue thought that the court had which is not reflected in the press summary above is the recommendation given by the court that the product of the AFR should not be used without checking by a person.
Thus, SWP may now… wish to consider whether further investigation should be done into whether the NeoFace Watch software may produce discriminatory impacts. When deciding whether or not this is necessary it will be appropriate for SWP to take account that whenever AFR Locate is used there is an important failsafe: no step is taken against any member of the public unless an officer (the systems operator) has reviewed the potential match generated by the software and reached his own opinion that there is a match between the member of the public and the watchlist face.See paragraph 156
“… This new and intrusive technology has the potential, if used without the right privacy safeguards, to undermine rather than enhance confidence in the police…”ICO Statement – 4 September 2019
“… Up until now, insofar as there has been a public debate, it has been about the police trialling of facial image matching in public places and whether this is lawful or whether in future it ought to be lawful. As Biometrics Commissioner I have reported on these police trials and the legal and policy question they have raised to the Home Secretary and to Parliament. However, the debate has now expanded as it has emerged that private sector organisations are also using the technology for a variety of different purposes. Public debate is still muted but that does not mean that the strategic choices can therefore be avoided, because if we do so our future world will be shaped in unknown ways by a variety of public and private interests: the very antithesis of strategic decision making in the collective interest that is the proper business of government and Parliament.Biometrics Commissioner response to court judgment on South Wales Police’s use of automated facial recognition technology. Published 10 September 2019
The use of biometrics and artificial intelligence analysis is not the only strategic question the country presently faces. However, that is no reason not to have an informed public debate to help guide our lawmakers. I hope that ministers will take an active role in leading such a debate in order to examine how the technologies can serve the public interest whilst protecting the rights of individuals citizens to a private life without the unnecessary interference of either the state or private corporations. As in 2012 this again is about the ‘protection of freedoms’…”
One key issue, relating to the use of AI/ML by public bodies, concerns the question whether a machine can take a decision. In Khan Properties Ltd v The Commissioners for Her Majesty’s Revenue & Customs it was held that tax penalties had to be determined by
“…a flesh and blood human being who is an officer of the HMRC”Judgment at  – , and at 
The judge noted that Parliament said expressly when a machine alone could make a decision as in section 2 of the Social Security Act 1998. This approach has been reviewed in a later Tax Tribunal case Barry Gilbert v The Commissioners for Her Majesty’s Revenue & Customs  UKFTT 0437 (TC), 2018 WL 04006232.
In cases governmental bodies have made extensive use of AI/ML, albeit with a human interface between the output and the person affected, it may be necessary to ask whether a human has actually made a decision. If the involvement has been minimal, for instance where the machine has done all the work and is completely relied on by the official, this may not be lawful.
Other UK cases concerned with AI include –
- Lowmoore Nursing Home Limited v Miss C Smith, UKEAT/0239/15/JOJ, 21 June 2016.
- Gibbs v Westcroft Health Centre Employment Tribunal, 3 December 2014,  12 WLUK 110.
These first two cases address issues related to the application of the so called Bradford Formula, an early approach to using AI techniques to manage employee absence.
- Privacy International v Secretary of State for Foreign and Commonwealth Affairs (Rev 1), 23 July  UKIPTrib IPT_15_110_CH.
There are also cases which address issues such as e-disclosure –
- Pyrrho Investments Ltd v MWB Property Ltd & Ors  EWHC 256 (Ch)
Council of Europe
The Council of Europe (CoE) is responsible for the European Convention on Human Rights (ECHR) and the European Court of Human Rights (ECtHR) and it has developed specific human rights standards for many years. In particular –
- The CoE has a website dedicated to addressing human rights issues raised by AI which can be accessed here.
- The CoE has a dedicated “MSI-AUT Committee of experts on Human Rights Dimensions of automated data processing and different forms of artificial intelligence”.
- In September 2019 the MSI-AUT Committee published “A study of the implications of advanced digital technologies (including AI systems) for the concept of responsibility within a human rights framework”, see here.
- The CoE’s Committee of Ministers have proposed to start work on new legislation to address AI at its October 2019 Meeting. We will report on this proposal as it develops.
- On the 13 February 2019 the CoE published its Declaration by the Committee of Ministers on the Manipulative Capabilities of Algorithmic Processes.
- The CoE’s European Ethical Charter on the Use of Artificial Intelligence in Judicial Systems of 3 December 2018 contains five key principles to avoid discrimination and to ensure AI is human – centric.
- On the 7 March 2018 the CoE published its Recommendation CM/Rec(2018)2 of the Committee of Ministers to member States on the roles and responsibilities of internet intermediaries.
- The CoE’s Convention 108: The Convention for the Protection of Individuals with regard to Automatic Processing of Personal Data is significant as a basis for the EU’s GDPR. This has been updated and is now known as Convention 108+.
European Court of Human Rights
The UK gives effect to the jurisprudence of the European Court of Human Rights through the Human Rights Act 1998. To date however, no judgment of the European Court of Human Rights has specifically addressed AI and equality and non-discrimination issues. Nonetheless it is important to recall that this court will normally take into account all relevant work of the Council of Europe, so it is to be expected that the provisions of the European Ethical Charter will be very important for it.
Some European Court of Human Rights judgments have considered intelligence gathering and its consequences through AI and machine learning –
- Big Brother Watch v United Kingdom (58170/13), 13 September 2018  9 WLUK 157.
- Szabo v Hungary (37138/14) 12 January 2016  1 WLUK 80; (2016) 63 E.H.R.R. 3.
- Zakharov v Russia (47143/06) 4 December 2015,  12 WLUK 174; (2016) 63 E.H.R.R. 17; 39 B.H.R.C. 435.
- Weber v Germany (54934/00) 2 June 2006  6 WLUK 28; (2008) 46 E.H.R.R. SE5.
- Catt v. United Kingdom (43514/15) 24 January 2019  1 WLUK 241; (2019) 69 E.H.R.R. 7 – This case concerns the obligation to delete personal data. The context was police of records of an elderly man’s participation in peaceful demonstrations organised by an extremist protest group. The indefinite retention of this data infringed his right to respect for his private life under ECHR art.8. While there were reasons to collect his personal data in the first place (he having aligned himself with a violent group) there were no effective safeguards and no reason to retain his data for an indefinite period. Moreover some of the data should have attracted a heightened level of protection as it concerned the complainant’s political opinions. It is interesting that the ECtHR held that some automated searching of the police data base could have been used to find the entries relating to the complainant and therefore make the process of removing the data that had been unlawfully kept and easier task.
The European Union’s Fundamental Rights Agency (FRA) published #BigData: Discrimination in data-supported decision making in September 2018. Later that year in December it published Preventing unlawful profiling today and in the future: a guide.
in June 2019, the FRA published a Focus paper: Data quality and artificial intelligence – mitigating bias and error to protect fundamental rights which usefully addresses the problem of systems based on incomplete or biased data and shows how they can lead to inaccurate outcomes that infringe on people’s fundamental rights, including discrimination.
On the 8th April the European Commission its communication to the European Parliament, the Council, the European Economic and Social Committee and the Committee of the Regions, entitled Building Trust in Human-Centric Artificial Intelligence.
The European Commission has published AI Ethical Guidelines developed by the High-Level Expert Group on Artificial Intelligence (AI HLEG) . These guidelines are based on the following key requirements when AI is in use –
- Human agency and oversight: AI systems should enable equitable societies by supporting human agency and fundamental rights, and not decrease, limit or misguide human autonomy.
- Robustness and safety: Trustworthy AI requires algorithms to be secure, reliable and robust enough to deal with errors or inconsistencies during all life cycle phases of AI systems.
- Privacy and data governance: Citizens should have full control over their own data, while data concerning them will not be used to harm or discriminate against them.
- Transparency: The traceability of AI systems should be ensured.
- Diversity, non-discrimination and fairness: AI systems should consider the whole range of human abilities, skills and requirements, and ensure accessibility.
- Societal and environmental well-being: AI systems should be used to enhance positive social change and enhance sustainability and ecological responsibility.
- Accountability: Mechanisms should be put in place to ensure responsibility and accountability for AI systems and their outcomes.
In June 2019, the AI HLEG published its second paper entitled “Policy and investment recommendations for trustworthy Artificial Intelligence” which is available here. This paper repeatedly emphasises the importance of building a FRAND (fair reasonable and non-discriminatory) approach, and proposes regulatory changes, arguing that the EU –
Adopt a risk-based governance approach to AI and an ensure an appropriate regulatory framework Ensuring Trustworthy AI requires an appropriate governance and regulatory framework. We advocate a risk-based approach that is focused on proportionate yet effective action to safeguard AI that is lawful, ethical and robust, and fully aligned with fundamental rights. A comprehensive mapping of relevant EU laws should be undertaken so as to assess the extent to which these laws are still fit for purpose in an AI-driven world. In addition, new legal measures and governance mechanisms may need to be put in place to ensure adequate protection from adverse impacts as well as enabling proper enforcement and oversight, without stifling beneficial innovation.Key Takeaways, paragraph 9
In the summer 2019, the European Commission has said that it will launch a pilot phase involving a wide range of stakeholders. Companies, public administrations and organisations are encouraged now to sign up to the European AI Alliance and receive a notification when the pilot starts.
Following the pilot phase, in early 2020, the AI expert group will review the assessment lists for the key requirements, building on the feedback received. Building on this review, the European Commission proposes to evaluate the outcome and propose any next steps.
These steps are not just important for the European Commission – the European Council emphasised how important these would be for the future development of the Digital Europe programme in its communication of the 11th February 2019.
Court of Justice of the European Union
The first mention of artificial intelligence in the Court of Justice of the European Union (CJEU) was back in 1986 when Advocate General Slynn gave an Opinion that a computer capable of undertaking rudimentary AI was a “scientific machine”. This does not seem to be controversial in retrospect! Since then only three other Opinions have mentioned AI and none has yet made any very important comment on its impact on Equality and Human Rights.
- C-434/15 Asociación Profesional Elite Taxi 11 May 2017 Opinion of AG Szpunar.
- Case C‑99/16 Lahorgue 9 February 2017 Opinion of AG Wathelet.
- C-28/08 P Commission v Bavarian Lager 15 October 2009 Opinion of AG Sharpston.
Other important cases concerned with digital integrity include –
- Glawischnig-Piesczek, C-18/18, judgment of the 3 October 2019 in which the CJEU held that notwithstanding the fact that a host provider such as Facebook is not liable for stored information if it has no knowledge of its illegal nature or if it acts expeditiously to remove or to disable access to that information as soon as it becomes aware of it, that exemption does not, prevent the host provider from being ordered by a court of a Member State, to terminate or prevent an infringement, including by removing the illegal information or by disabling access to it. However, the directive prohibits any requirement for the host provider to monitor generally information which it stores or to seek actively facts or circumstances indicating illegal activity.
- Planet49, C-673/17 judgment of the 1 October 2019. In this case the CJEU ruled that the consent which a website user must give to the storage of and access to cookies on his or her equipment is not validly constituted by way of a pre-checked checkbox which that user must deselect to refuse his or her consent. That decision is unaffected by whether or not the information stored or accessed on the user’s equipment is personal data. It held that EU law aims to protect the user from any interference with his or her private life, in particular, from the risk that hidden identifiers and other similar devices enter those users’ terminal equipment without their knowledge. The Court noted that consent must be specific so that the fact that a user selects the button to participate in a promotional lottery is not sufficient for it to be concluded that the user validly gave his or her consent to the storage of cookies. Furthermore, the information that the service provider must give to a user must include the duration of the operation of cookies and whether or not third parties may have access to those cookies.
- LSG-Gesellschaft zur Wahrnehmung von Leistungsschutzrechten, C-557/07, Order of the Court of 19 February 2009
Initiatives from major international organisations
The UN has many pages dealing with AI issues and the Centre for Policy Research and the UN University havealso developed proposals for the proper use of AI.
The UN University discusses AI & Global Governance.
The OECD’s 36 member countries, along with Argentina, Brazil, Colombia, Costa Rica, Peru and Romania, agreed the OECD Principles on Artificial Intelligence on the 22nd May 2019.
These were based on the work of an expert group formed by more than 50 members from governments, academia, business, civil society, international bodies, the tech community and trade unions. The Principles comprise five values-based principles for the responsible deployment of trustworthy AI and five recommendations for public policy and international co-operation. They aim to guide governments, organisations and individuals in designing and running AI systems in a way that puts people’s best interests first and ensuring that designers and operators are held accountable for their proper functioning.
The AI Principles have the backing of the European Commission, whose high-level expert group has produced Ethics Guidelines for Trustworthy AI, It is proposed that the OECD’s digital policy experts will build on the Principles in the months ahead to produce practical guidance for implementing them.
While not legally binding, existing OECD Principles in other policy areas have proved highly influential in setting international standards and helping governments to design national legislation. The key provisions are
1.2.Human-centred values and fairness
a) AI actors should respect the rule of law, human rights and democratic values, throughout the AI system lifecycle. These include freedom, dignity and autonomy, privacy and data protection, non-discrimination and equality, diversity, fairness, social justice, and internationally recognised labour rights.
b) To this end, AI actors should implement mechanisms and safeguards, such as capacity for human determination, that are appropriate to the context and consistent with the state of art.
1.3.Transparency and explainability
AI Actors should commit to transparency and responsible disclosure regarding AI systems. To this end, they should provide meaningful information, appropriate to the context, and consistent with the state of art:
i.to foster a general understanding of AI systems,
ii.to make stakeholders aware of their interactions with AI systems, including in the workplace,
iii.to enable those affected by an AI system to understand the outcome, and,
iv.to enable those adversely affected by an AI system to challenge its outcome based on plain and easy-to-understand information on the factors, and the logic that served as the basis for the prediction, recommendation or decision.
1.4.Robustness, security and safety
a) AI systems should be robust, secure and safe throughout their entire lifecycle so that, in conditions of normal use, foreseeable use or misuse, or other adverse conditions, they function appropriately and do not pose unreasonable safety risk.
b) To this end, AI actors should ensure traceability, including in relation to datasets, processes and decisions made during the AI system lifecycle, to enable analysis of the AI system’s outcomes and responses to inquiry, appropriate to the context and consistent with the state of art.
c)AI actors should, based on their roles, the context, and their ability to act, apply a systematic risk management approach to each phase of the AI system lifecycle on a continuous basis to address risks related to AI systems, including privacy, digital security, safety and bias.
AI actors should be accountable for the proper functioning of AI systems and for the respect of the above principles, based on their roles, the context, and consistent with the state of art.
The G20 adopted a ministerial statement on AI at its meeting in Japan on the 8th and 9th June 2019. The statement covered a range of issues but the key provision dealing with AI (which adopted the OECD Principles on Artificial Intelligence) is as follows –
17. Recognizing the efforts undertaken so far by all stakeholders in their respective roles including governments, international organizations, academia, civil society and the private sector, and mindful of how technology impacts society, the G20 endeavors to provide an enabling environment for human-centered AI that promotes innovation and investment, with a particular focus on digital entrepreneurship, research and development, scaling up of startups in this area, and adoption of AI by MSMEs which face disproportionally higher costs to adopt AI.
18. We recognize that AI technologies can help promote inclusive economic growth, bring great benefits to society, and empower individuals. The responsible development and use of AI can be a driving force to help advance the SDGs and to realize a sustainable and inclusive society, mitigating risks to wider societal values. The benefits brought by the responsible use of AI can improve the work environment and quality of life, and create potential for realizing a human-centered future society with opportunities for everyone, including women and girls as well as vulnerable groups.
19. At the same time, we also recognize that AI, like other emerging technologies, may present societal challenges, including the transitions in the labor market, privacy, security, ethical issues, new digital divides and the need for AI capacity building. To foster public trust and confidence in AI technologies and fully realize their potential, we are committed to a human- centered approach to AI, guided by the G20 AI Principles drawn from the OECD Recommendation on AI, which are attached in Annex and are non-binding.
This Annex includes the following principles of “inclusive growth, sustainable development and well-being”, “human-centered values and fairness”, “transparency and explainability”, “robustness, security and safety” and “accountability”. The Annex also offers guidance for consideration by policy makers with the purpose of maximizing and sharing the benefits from AI, while minimizing the risks and concerns, with special attention to international cooperation and inclusion of developing countries and underrepresented populations.
20. In pursuing human-centered AI, G20 members recognize the need to continue to promote the protection of privacy and personal data consistent with applicable frameworks. The G20 also recognizes the need to promote AI capacity building and skills development. We will each continue to strive for international cooperation and endeavor to work together with appropriate fora in areas such as research and development, policy development and information sharing through the G20 Repository of Digital Policies and other open and collaborative efforts.
‘G7 Innovation Ministers’ Statement on Artificial Intelligence’, 28 March 2018
Initiatives in European Countries
A National Strategy for AI was published in March 2019.
On the 1st May 2019 the Danish Competition and Consumer Agency (DCCA) established a”Center for Digital Platforms” to strengthen the enforcement of competition rules against digital platforms. It is intened that the Center will be a hub for the DCCA to analyze the impact of big data, ML, AI and algorithms. Further information here.
There are numerous initiatives in France including the Villani Report (2018), France Intelligence Artificielle (2017) and a CNIL report, The ethical matters raised by algorithms and artificial intelligence (2017)
Germany’s Federal Government set up the Data Ethics Commission (Datenethikkommission) on 18 July 2018. It asked the Commission key questions concerning algorithm-based decision-making (ADM), AI and data. In October 2019 the Commission published its Opinion (available in English). The Opinion opened by stating that –
Humans are morally responsible for their actions, and there is no escaping this moral dimension. Humans are responsible for the goals they pursue, the means by which they pursue them, and their reasons for doing so. Both this dimension and the societal conditionality of human action must always be taken into account when designing our technologically shaped future. At the same time, the notion that technology should serve humans rather than humans being subservient to technology can be taken as incontrovertible fact. Germany’s constitutional system is founded on this understanding of human nature, and it adheres to the tradition of Europe’s cultural and intellectual history. Digital technologies have not altered our ethical framework – in terms of the basic values, rights and freedoms enshrined in the German Constitution and in the Charter of Fundamental Rights of the European Union. Yet the new challenges we are facing mean that we need to reassert these values, rights and freedoms and perform new balancing exercises. With this in mind, the Data Ethics Commission believes that the following ethical and legal principles and precepts should be viewed as indispensable and socially accepted benchmarks for action
The Opinion made numerous recommendations as to the way forward for the German Federal Republic to deal with identified harms. In relation to discrimination the Opinion stated –
Consideration should be given to expanding the scope of anti-discrimination legislation to cover specific situations in which an individual is discriminated against on the basis of automated data analysis or an automated decision-making procedure. In addition, the legislator should take effective steps to prevent discrimination on the basis of group characteristics which do not in themselves qualify as protected characteristics under law, and where the discrimination often does not currently qualify as indirect discrimination on the basis of a protected characteristic.Recommendation 53
The Data Protection Commission is the national independent authority in Ireland responsible for upholding the fundamental right of individuals in the European Union (EU) to have their personal data protected.
There is also an Irish Development Agency with a program on AI as explained here at Ireland AI Island.
The Government is undertaking a public consultation on a national strategy on AI. Given how important Ireland is as the base for major tech companies in Europe this will be closely watched. It is available here The consultation closes on the 7th November 2019.
In Dwyer v Commissioner of An Garda Siochana & ors  IEHC 685 the Irish High Court reviewed the circumstances in which a democracy may tolerate the State mandated electronic surveillance of every citizen who uses a telephone device and considered whether the Irish Police (“An Garda Síochána”) could consider that information.
An AI strategy was published in March 2018 – L’intelligenza artificiale al servizio del cittadino.
An AI strategy was published in 2019 – Artificial Intelligence Strategy: A Vision of the Future.
In the Netherlands there is considerable concern about the use of Governmental data bases to target the poor. The relevant legislation is available here. Litigation has been started to challenge the legality of this in the so – called SyRI case. The UN Special Rapporteur on extreme poverty has filed an important amicus brief on the implications for poor people of the digitisation of welfare benefits in the Netherlands which is available here. The Dutch Government response can be seen here.
An AI strategy was published in May 2018 – National orientation for artificial intelligence.
Initiatives in other parts of the world
In mid 2019, the Australian government started a consultation process in relation to an AI ethics framework. More information is here. On the 7th November 2019 the Department for Innovation Industry and Science published the Federal Government’s Finalised Ethics Guidelines. These draw on the work of other bodies such as IEEE. See the link to the Institute’s work here.
Eight principles have been enunciated, each of which is developed in some detail in the full Guidelines, In summary these are –
 Human, social and environmental wellbeing: Throughout their lifecycle, AI systems should benefit individuals, society and the environment.
 Human-centred values: Throughout their lifecycle, AI systems should respect human rights, diversity, and the autonomy of individuals.
 Fairness: Throughout their lifecycle, AI systems should be inclusive and accessible, and should not involve or result in unfair discrimination against individuals, communities or groups.
 Privacy protection and security: Throughout their lifecycle, AI systems should respect and uphold privacy rights and data protection, and ensure the security of data.
 Reliability and safety: Throughout their lifecycle, AI systems should reliably operate in accordance with their intended purpose.
 Transparency and explainability: There should be transparency and responsible disclosure to ensure people know when they are being significantly impacted by an AI system, and can find out when an AI system is engaging with them.
 Contestability: When an AI system significantly impacts a person, community, group or environment, there should be a timely process to allow people to challenge the use or output of the AI system.
 Accountability: Those responsible for the different phases of the AI system lifecycle should be identifiable and accountable for the outcomes of the AI systems, and human oversight of AI systems should be enabled.
The full guidelines can be seen here.
A Directive on Automated Decision making was introduced in Canada in April 2019 with a requirement that organisations comply no later than April 2020. It is available here. In addition, an “Algorithmic Impact Assessment” (AIA) has been developed which is a questionnaire designed to help organisations assess and mitigate the risks associated with deploying an automated decision system. The AIA also helps identify the impact level of any automated decision system. It is available here.
Canada adopted a Digital Charter on the 21st May 2019, that aims to address challenges posed by digital and data transformation; see here.
China published its ‘AI Strategy‘ in July 2017. There is a useful paper published by the Oxford Internet Institute written on the 1st September 2019 analysing the strategy and the steps which have been taken which can be found here.
‘National Strategy for Artificial Intelligence‘, June 2018
‘Artificial Intelligence Technology Strategy‘, March 2017
The Personal Data Commission, created by the Singapore government, published in January 2019 its first edition of “A Proposed Model AI Governance Framework (Model Framework)“. an accountability-based framework to help chart the language and frame the discussions around harnessing AI in a responsible way.
In April 2019, the US proposed the Algorithmic Accountability Act 2019. The bill that would require major tech companies to detect and remove any discriminatory biases embedded in their computer models.
New York: The New York City Council introduced, in November 2018,
a local law in relation to automated decision systems used by agencies which require the creation of a task force that provides recommendations on how information on agency automated decision systems may be shared with the public and how agencies may address instances where people are harmed by agency automated decision systems.
Washington State: A proposal to amend the Revised Code of Washington State (RCW) (that is the compilation of all permanent laws now in force in that State – see here) was discussed on the 23 January 2019 as Senate Bill 5528. It proposes a moratorium on the use of facial recognition AI: see here). It was referred to the State Committees on Environment, Energy & Technology and on Innovation, Technology & Economic Development.
There are also various US wide policies:
‘Department of Defense: artificial intelligence, big data and cloud taxonomy’, Govini, December 2017
Committee on Technology, National Science and Technology Council, Executive Office of the President ‘Preparing for the Future of Artificial Intelligence‘, October 2016.