This is a House of Commons Committee report, with recommendations to government. The Government has two months to respond.
Science, Innovation and Technology Committee
Governance of artificial intelligence (AI)
Date Published: 28 May 2024
This is the full report, read the report summary.
1. Since the publication of our interim Report examining the governance of artificial intelligence (AI) in August 2023, debates over how to regulate the development and deployment of AI have continued. These debates have often centred around the twelve challenges of AI governance we identified in our interim Report.1
2. Jurisdictions including the UK,2 European Union3 and United States4 have begun to establish regulatory regimes to govern—to varying degrees and using different approaches—the development and deployment of AI, ahead of the anticipated launch of new, more advanced models during the months ahead.5
3. Whilst an international consensus around AI governance has not yet been reached, the topic has risen up the agenda of various international fora, including the G7,6 United Nations7 and at the AI Safety Summit organised by the UK Government at Bletchley Park in November 2023.8
4. With AI’s growing ubiquity, the capability of existing models, and the ongoing conversation about how best to regulate the development and deployment of the technology, there is much to reflect on in our second Report.
5. We launched our Inquiry on 20 October 2022, to examine the impact of AI on different areas of society and the economy; whether and how AI and its different uses should be regulated; and the Government’s AI governance proposals.9 We have received and published over 100 written submissions and taken oral evidence from 33 witnesses, including Government Ministers and officials, AI researchers, businesses, civil society, and professionals affected by the technology.
6. In August 2023 we published an interim Report,10 to which the Government responded in November of that year.11 We have also visited the United States, where we met with representatives from public bodies, private companies and research institutions in Boston and Washington, D. C.; and the European Union institutions in Brussels. We are grateful to everyone who has contributed to our Inquiry.
7. This Report examines domestic and international developments in the governance and regulation of AI since the publication of our interim Report. It also revisits the Twelve Challenges of AI Governance we identified in our interim Report and suggests how they might be addressed by policymakers.
8. With a General Election approaching we have sought to make this Report futureproof and believe that our conclusions and recommendations will remain applicable to future Administrations. It is important that the timing of the General Election does not stall necessary efforts by the Government, developers and deployers of AI to increase the level of public trust in a technology that has become a central part of our everyday lives.
9. As AI has become an increasingly ubiquitous, general-purpose technology, debates over the societal and economic implications have continued. In this Chapter, we will examine the increasingly widespread deployment of AI, the positive and negative consequences, and the ‘case in favour’ of the technology.
10. In our interim Report, published in August 2023, we examined the potential benefits of the deployment of AI models and tools in healthcare provision, medical research and education.12 Since then, we have seen AI used across a growing number of societal and economic activities—as one analysis has observed, 2024 has been the deployment phase for AI.13
11. The UK Government has consistently emphasised the benefits associated with the deployment of AI in the public and private sectors, a topic we will return to in Chapter 5. The Secretary of State for Science, Innovation and Technology, Rt. Hon. Michelle Donelan MP, described AI to us as “… a foundational technology that interlinks with all the other technologies”.14
12. The Secretary of State’s depiction of AI as a foundational technology has been borne out in the expanding range of sectors where it has improved existing processes and either offered or already delivered tangible productivity gains, by augmenting and assisting skilled workers.15
13. One sector highlighted to us as well-placed to capitalise on the benefits associated with the deployment of AI is financial services, a key driver of growth for the UK economy.16 Nikhil Rathi, Chief Executive of the Financial Conduct Authority (FCA), has said that whilst not new, “AI in markets today brings models incorporating deep learning and neural networks capable of analysing large datasets and highlighting intricate patterns”, which facilitate “… synchronised, automated order placements”.17
14. Cyber security is another sector where the analytical capabilities of the technology have been deployed. NCC Group, a cyber security firm, told us that AI was “… used by cyber defenders to analyse large data sets at scale, support threat intelligence and mimic the behaviours of cyber attackers, so that organisations can understand and prepare for potential attacks”.18 Leading AI developers have built generative AI models that can assist and augment the work undertaken by human cyber security analysts.19
15. However, these benefits cannot be realised without incurring risks. The FCA told us that “AI’s potential for autonomous decision-making brings with it potentially serious challenges to governance processes, because it puts in to question the ownership and responsibility for decision making”.20 It also said that the increasing complexity of AI models “… will require a greater focus on testing, validation and explainability… built on strong accountability principles”.21
16. In the cyber security field, NCC Group told us that in addition to the benefits of AI, it was “… lowering the barrier of entry into cybercrime, making it easier for cyber attackers to successfully target victims and widening the availability of voice cloning, deep fakes and social engineering bots”.22 The National Cyber Security Centre, part of GCHQ, has said that “AI is already being used in malicious cyber activity and will almost certainly increase the volume and impact of cyber attacks—including ransomware—in the near term”.23
17. Since the publication of our interim Report, increased attention has been paid to the environmental impact of the development and use of AI models and tools, particularly their electricity and water consumption.24 Researchers at the University of California, Riverside have estimated that running 10–50 ‘inferences’—or queries—using OpenAI’s GPT-3 model can equate to consuming “… 500 millilitres of water, depending on when and where the model is hosted. GPT-4, the model currently used by ChatGPT, reportedly has a much larger size and hence likely consumes more water than GPT-3”.25
18. In its 2024 environmental sustainability report, Microsoft confirmed that its overall carbon emissions had risen by 29.1% since 2020. It attributed this to “… the construction of more datacenters and the associated embodied carbon in building materials, as well as hardware components such as semiconductors, servers, and racks”.26 Whilst Microsoft and other leading developers such as Google have set targets to reduce their emissions and energy consumption by the end of the decade,27 it is nevertheless notable that Sam Altman, CEO of OpenAI, said in January 2024 that “we still don’t appreciate the energy needs of this technology”.28
19. Government communications have referred to both the positive and negative consequences associated with the increasing deployment of AI. The Prime Minister, Rt. Hon. Rishi Sunak MP, has said that “… the more we learn about frontier technologies like AI, the more they widen our horizons… the possibilities are extraordinary”.29 The Secretary of State for Science, Innovation and Technology similarly described improved AI capabilities as a “… once-in-a-generation opportunity for the British people to revolutionise our public services for the better and to deliver real, tangible, long-term results for our country”.30
20. At the same time, the Government response to its AI White Paper consultation detailed three categories of AI-related risk—societal harms, misuse risk, and autonomy risk31—and an AI Safety Institute has been established in order to “… minimise surprise to the UK and humanity from rapid and unexpected advances… by developing the sociotechnical infrastructure needed to understand the risks of advanced AI and support its governance”.32 We will examine the Government’s approach to AI governance, the deployment of AI in the public and private sectors and the role of the AI Safety Institute in Chapters 3, 4, 5 and 6 of this Report.
21. Our interim Report identified Twelve Challenges of AI governance that we said could potentially complicate the process of ensuring that policy could deliver the beneficial consequences of AI whilst also safeguarding the public interest and preventing known potential harms, both societal and individual.33 We will offer some potential solutions to these Challenges in Chapter 8 of this Report.
22. If governed appropriately, we believe that AI can deliver on its significant promise, to complement and augment human activity. The Government has articulated the case for AI: better public services, high quality jobs and a new era of economic growth driven by advances in AI capabilities.
23. The Government is right to emphasise the potential societal and economic benefits to be won from the strategic deployment of AI. However, as our interim Report highlighted, the challenges are as clear as the potential benefits, and these benefits cannot be realised without public trust in the technology.
24. The Government should certainly make the case for AI but should equally ensure that its regulatory framework addresses the Twelve Challenges of AI Governance that we have identified in our interim Report; and offer potential solutions to in this Report.
25. In February 2024 the Government set out further details of its regulatory approach to AI in the form of its response to a consultation on its AI White Paper,34 which was published in March 2023.35 In this Chapter we will examine the consultation response and the likelihood of AI-specific legislation in the UK following the General Election.
26. The consultation response confirmed that the Government would pursue the principles-based approach it proposed in its March 2023 AI White Paper and examined in our interim Report.36 Five high-level principles—”safety, security and robustness; appropriate transparency and explainability; fairness; accountability and governance; and contestability and redress”37—underpinned the Government’s approach, which it has described as a combination of “… cross-sectoral principles and a context-specific framework, international leadership and collaboration, and voluntary measures on developers”.38
27. The Government has said that its intended framework has been developed in order to avoid “… unnecessary blanket rules that apply to all AI technologies, regardless of how they are used. This is the best way to ensure an agile approach that stands the test of time”.39 Existing regulators have been asked to implement the five high-level principles in their respective sectors.40
28. A key question for our Inquiry has been whether the Government should bring forward AI-specific legislation. Our interim Report pointed out that the period leading up to the General Election would be the final opportunity to enact legislation before “… late 2025—more than two years from now and nearly three years from the publication of the [AI] White Paper”.41
29. In the March 2023 AI White Paper the Government said that it anticipated legislating, at a minimum, to establish ‘due regard’ duties for existing regulators in relation to its five high-level principles.42 Our interim Report said that “… [this] commitment alone—in addition to any further requirements that may emerge—suggests that there should be a tightly-focussed AI Bill in the November 2023 King’s Speech”.43
30. In its response to our interim Report, received in November 2023, the Government said that “… rather than rushing to legislate, we want to simultaneously learn about model capabilities and risks, while also carefully considering the frameworks for action”.44 In December 2023 the Secretary of State for Science, Innovation and Technology told us that the Government would not bring forward an AI-specific Bill before the General Election, given the time that it would likely take to become law:
The key here is timing. We are not saying that we would never legislate in this space. Of course we would… every Government will have to legislate eventually. The point is that we do not want to rush to legislate and get this wrong. We do not want to stifle innovation.45
31. The Government response to the AI White Paper consultation confirmed this and argued that the proposed “… non-statutory approach currently offers critical adaptability—especially while we are still establishing our approach”.46 The Government has also emphasised the importance of the safety-related voluntary commitments secured from leading AI developers ahead of the AI Safety Summit at Bletchley Park,47 and the ongoing safety testing work being undertaken by the AI Safety Institute.48 We will return to the role of the AI Safety Institute in Chapter 6.
32. Despite the expressed preference for a principles-based approach, the AI White Paper consultation response confirmed that the Government would bring forward legislation targeted at the most capable, general-purpose AI models and tools if such legislation became necessary, specifically if:
… we determined that existing mitigations were no longer adequate and we had identified interventions that would mitigate risks in a targeted way… if we were not sufficiently confident that voluntary measures would be implemented effectively by all relevant parties and if we assessed that risks could not be effectively mitigated using existing legal powers.49
In April 2024 it was reported by the Financial Times that Government officials were “… beginning to craft new legislation to regulate artificial intelligence”, and that “… such legislation would likely put limits on the production of large language models, the general-purpose technology that underlies AI products such as OpenAI’s ChatGPT”.50
33. The next Government should stand ready to introduce new AI-specific legislation, should an approach based on regulatory activity, existing legislation and voluntary commitments by leading developers prove insufficient to address current and potential future harms associated with the technology.
34. The Government should in its response to this Report provide further consideration of the criteria on which a decision to legislate will be triggered, including which model performance indicators, training requirements such as compute power or other factors will be considered.
35. The next Government should commit to laying before Parliament quarterly reviews of the efficacy of its current approach to AI regulation, including a summary of technological developments related to its stated criteria for triggering a decision to legislate, and an assessment whether these criteria have been met.
36. In addition to existing legislation and voluntary measures, the Government’s high-level principles have begun to be translated into sector-specific action by regulators.51 Our interim Report detailed different views about the capacity of individual regulators to respond to the growing use of AI,52 and the Government has also acknowledged that different regulators are at different stages of readiness.53 In this Chapter we assess the role of regulators in the UK’s AI governance framework.
37. Since the publication of our interim Report we have examined the preparedness of regulators to respond to AI’s increasing prevalence, including the Competition and Markets Authority (CMA), Financial Conduct Authority (FCA), Information Commissioner’s Office (ICO) and Ofcom; as well as the Digital Regulation Cooperation Forum (DRCF), described as a co-ordination mechanism or “connective tissue” by its Chief Executive, Kate Jones.54
38. Whilst all four regulators told us that they were well-placed to respond to the growing use of AI in their respective sectors,55 Nikhil Rathi, Chief Executive of the FCA, has described the challenge of “… managing the balance between transparency, fairness to firms and competitiveness”, particularly in the announcement of investigations.56 Dame Melanie Dawes, Chief Executive of Ofcom, the communications and online safety regulator, said that she was in favour of “… the Government having a policy function to scan where there may be gaps in the regulatory landscape”.57
39. A regulatory gap analysis was recommended in our interim Report,58 and in its response to the AI White Paper consultation the Government said that it recognised:
… the need to assess the existing powers and remits of the UK’s regulators to ensure they are equipped to address AI risks and opportunities in their domains and implement the principles in a consistent and comprehensive way. We will, therefore, work with government departments and regulators to analyse and review potential gaps in existing regulatory powers and remits.59
40. We welcome confirmation that the Government will undertake a regulatory gap analysis to determine whether regulators require new powers to respond properly to the growing use of AI, as recommended in our interim Report. However, as the end of this Parliament approaches, there is no longer time to bring forward any updates to current regulatory remits and powers, should they be discovered to be necessary. This could constrain the ability of regulators to properly implement the Government’s AI principles and undermine the UK’s overall approach.
41. The next Government should conduct and publish the results its regulatory gap analysis as soon as is practicable. If the analysis identifies any legislation required to close regulatory gaps, this should be brought forward in time for it to be enacted as soon as possible after the General Election.
42. The Government has said that regulators will need to coordinate with each other when implementing its AI principles,60 whilst our interim Report concluded that it would likely need to establish “a more well-developed central coordinating function” than that proposed in the AI White Paper.61
43. Will Hayter, Executive Director for Digital Markets at the CMA, told us that the four DRCF member regulators “… all understand throughout our digital regulation activities that we need to talk to one another, sharing expertise and attempting to be more than the sum of our parts”.62 Similarly, four national bodies in the field of health—the Care Quality Commission, Health Research Authority, Medicines and Healthcare Products Regulatory Agency and National Institute for Health and Care Excellence—have launched a joint Artificial Intelligence and Digital Regulations Service, which “… aims to clearly set out the information and guidance that developers and adopters need to follow to develop safe, innovative technologies in health and social care”.63
44. In its response to the AI White Paper consultation, the Government confirmed that it would set up “a steering committee with government representatives and key regulators to support knowledge exchange and coordination on AI governance”.64 A first iteration of guidance for regulators on how to implement the five principles has also been published.65
45. The general-purpose nature of AI will, in some instances, lead to regulatory overlap, and a potential blurring of responsibilities. This could create confusion on the part of consumers, developers and deployers of the technology, as well as regulators themselves. The steering committee that the Government has said it will establish should be empowered to provide guidance and, where necessary, direction to help regulators navigate any overlapping remits, whilst respecting the independence of the UK’s regulators.
46. The regulatory gap analysis being undertaken by the Government should identify, in consultation with the relevant regulators and co-ordinating entities such as the Digital Regulation Cooperation Forum and the AI and Digital Regulations Service, areas where new AI models and tools will necessitate closer regulatory co-operation, given the extent to which some uses for AI, and some of the challenges these can present—such as accelerating existing biases—are covered by more than one regulator. The gap analysis should also put forward suggestions for delivering this co-ordination, including joint investigations, a streamlined process for regulatory referrals, and enhanced levels of information sharing.
47. Some regulators, such as the FCA, are “… funded entirely by the fees [they] charge regulated firms”.66 Similarly, Ofcom receives fees from “… the companies we regulate… these could be broadcasters, telecoms providers, or firms in the postal sector”. The fees are based on company revenues and the amount of work undertaken by Ofcom in their sectors, although an overall spending cap is applied through the Government Spending Review process.67 Other regulators, such as the Information Commissioner’s Office, supplement fees with a Government grant-in-aid.68
48. The picture is similarly mixed in other jurisdictions. In the European Union (EU), some decentralised agencies such as the European Medicines Agency are funded through a combination of regulatory fees and contributions from the EU budget;69 whilst other such as the EU Intellectual Property Office are “… financed through registration fees without imposing any burden on the EU or its taxpayers”.70 In the United States, bodies such as the Federal Trade Commission and Food and Drug Administration rely upon a combination of user fees and budget requests submitted to Congress.71
49. In a Report examining regulator’s performance, published in February 2024, the House of Lords Industry and Regulators Committee expressed concern that some regulators in the UK “… appear not to have sufficient resources to carry out their existing functions effectively, while others have had their responsibilities extended without an increase in resources to match”.72
50. Our Inquiry heard that some regulators would require additional support to help them meet the challenges posed by the growing prevalence of AI in their sectors.73 Dame Melanie Dawes, Chief Executive of Ofcom, told us that Ofcom had experienced:
… a flat cash budget cap from the Treasury for many years now. I think at some point that will start to create real constraints for us. We have become very good at driving efficiency, but if the Government were to ask us to do more in the field of AI, we would need new resources to be able to do that.74
51. Similar concerns were noted by the Committee on Standards in Public Life (CSPL), an advisory non-departmental public body that has examined the implications of AI’s increasing prevalence, including for regulators.75 The CSPL has said that whilst many regulators had identified AI as a strategic priority, “… some regulators told us that because they operate under restricted financial resources, the speed and scale at which they can address the implications of AI is limited”.76
52. The Secretary of State for Science, Innovation and Technology and Sarah Munby, Permanent Secretary at the Department for Science, Innovation and Technology (DSIT) told us that the Government would continue to support regulators as required.77 In its AI White Paper consultation response, the Government said that it had asked a number of regulators to set out publicly how they intend to respond to the use of AI in their respective sectors.78
53. In addition to a “… central function to support effective risk monitoring, regulator coordination, and knowledge exchange” the AI White Paper consultation response announced £10 million in funding to “jumpstart regulators’ AI capabilities” and “… help our regulators develop cutting-edge research and practical tools to build the foundations of their AI expertise and everyday ability to address AI risks… “.79 It has also highlighted the role of the DRCF in the dissemination of best practice.80
54. If this £10 million were divided equally amongst the 14 regulators who were asked to publish their intended approaches to AI by the end of April 2024,81 each would receive an amount equivalent to approximately 0.0085% of the reported annual UK turnover of Microsoft in the year to June 2023.82
55. The increasing prevalence and general-purpose nature of AI will create challenges for the UK’s sectoral regulators, however expert they may be. The AI challenge can be summed up in a single word: capacity. Ofcom, for example, is combining implementation of a broad new suite of powers conferred on it by the Online Safety Act 2023, with formulating a comprehensive response to AI’s deployment across its wider remit. Others will be required to undertake resource-intensive investigations and it is vital that they are able, and empowered, to do so. All will be required to pay greater attention to the outputs of AI tools in their sectors, whilst paying due regard to existing innovation and growth-related objectives.
56. The announced £10 million to support regulators in responding to the growing prevalence of AI is clearly insufficient to meet the challenge, particularly when compared to the UK revenues of leading AI developers.
57. The next Government must announce further financial support, agreed in consultation with regulators, that is commensurate to the scale of the task. It should also consider the benefits of a one-off or recurring industry levy, that would allow regulators to supplement or replace support from the Exchequer for their AI-related activities.
58. The Government has paid increasingly close attention to the deployment of AI in the public sector and announced a number of initiatives to deliver its ambitions. In this Chapter, we assess the Government’s efforts to support increased uptake of AI in the public and private sectors.
59. A National Audit Office (NAO) report found in 2023 that the Central Digital and Data Office (part of the Cabinet Office), DSIT and HM Treasury began to develop a strategy for AI adoption in the public sector. The NAO found that “high-level activities and timescales” had been agreed, but the draft strategy did not “… set out which department has overall ownership of the strategy and accountability for its delivery or how it will be funded and resourced. Performance measures are also still to be determined”.83
60. In November 2023 the Deputy Prime Minister and Chancellor of the Duchy of Lancaster, and Secretary of State in the Cabinet Office (henceforth Deputy Prime Minister), Rt. Hon. Oliver Dowden CBE MP, announced the establishment of an Incubator for Artificial Intelligence (i.AI). This is comprised of an initial team of 30 technical experts “… to design and implement AI solutions across government departments to drive improvements in public service delivery”,84 and was subsequently confirmed that the i.AI team would increase in size to 70, having instigated 10 pilot programmes since its launch.85
61. The NAO has found that i.AI will require an estimated £101 million in funding over five years between 2024–25 and 2028–29.86 Examples of its work to date have included i.AI assisting the Public Sector Fraud Authority in the development of AI-enabled fraud detection tools.87
62. Announced by the Chancellor of the Exchequer, Rt. Hon. Jeremy Hunt MP, in June 2023 and detailed at subsequent fiscal events,88 the public sector productivity programme has been described as seeking to secure “the potential productivity benefits from applying AI to routine tasks across the public sector”, in services such as education and policing.89
63. At the March 2024 Spring Budget the Chancellor announced that £4.2 billion of funding would be allocated to “… the strongest productivity releasing projects that departments have identified through the programme to date”.90 These include the deployment of AI to improve or automate existing processes, such as a £3.4 billion by 2030 for a “technological and digital transformation” of the NHS.91
64. HM Treasury also confirmed in a paper published alongside the Budget that it would, working in partnership with the Cabinet Office and i.AI, assist in the delivery of “… AI adoption plans for every department in time for the next Spending Review and… expand the application of automation and AI across the range of priority areas”.92
65. The Government has outlined mechanisms to coordinate departmental AI activity, including an Inter-Ministerial Group and the designation of lead AI Ministers across all departments “… to bring together work on risks and opportunities driven by AI in their sectors and to oversee implementation of frameworks and guidelines for public sector usage of AI”.93
66. Both the Cabinet Office and DSIT have announced initiatives to underpin adoption of AI across the public sector. The former has confirmed that it will “… [improve] digital infrastructure and access to data sets, and [develop] centralised standards”,94 whilst the Central Digital and Data Office, part of the Cabinet Office, has published guidance on the use of generative AI in Government.95
67. Government Digital Service, also part of the Cabinet Office, has experimented with a generative AI chatbot, GOV.UK Chat, but found that its “… answers did not reach the highest level of accuracy demanded for a site like GOV.UK, where factual accuracy is crucial”.96 The Cabinet Office and Infrastructure Project Authority have also encouraged “… responsible experimentation with AI to find solutions to the biggest challenges in public projects”.97
68. The Government response to the AI White Paper consultation confirmed that use of the Algorithmic Transparency Recording Standard (ATRS), a tool designed to increase transparency from public sector organisations about how they use algorithmic tools to support decision-making,98 would become a requirement for all Government departments during 2024 and “… across the broader public sector over time”.99
69. The Government has also launched a series of initiatives intended to support private sector adoption of AI. The Secretary of State for Science, Innovation and Technology and the Prime Minister’s Special Adviser on Business and Investment, Lord Petitgas, established an AI Opportunity Forum with business leaders. It discussed interrelated issues such as “… AI culture and skills of organisations in the UK, how they manage governance, awareness, and risks of the technology, and the availability of data”.100
70. The Forum has also discussed the challenges that businesses have faced when seeking to adopt AI, including access to data and access to compute101—two Challenges of AI Governance highlighted in our interim Report and discussed in Chapter 6 of this Report. The Government has said that the Forum will produce “a product that will inspire businesses, whether they are a Silicon Roundabout start-up or a family-run firm, to start using AI”, but had yet to confirm a publication date when our Report was finalised.102
71. The Chancellor also announced the launch of an SME Digital Adoption Taskforce at the Spring Budget 2024, along with an upskilling fund pilot targeted at SMEs. The pilot, applications for which closed on 31 May, will make £6.4 million of grant funding available in 2024–25.103 Successful applicants may receive up to 50% of the cost of “… training which supports employees to develop their technical skills and/or understanding of AI to be able to develop, deploy, or use AI in their role”.104
72. DSIT has produced guidance on AI assurance, which detailed how organisations could “… measure and evaluate their systems and communicate that their systems are trustworthy and aligned with relevant regulatory principles”.105 It has also supported the establishment of an AI and Digital Hub aimed at “… innovators with queries concerning cross-regulatory AI and digital issues”, led by the Digital Regulation Cooperation Forum, a co-ordinating body.106
73. AI can be used to increase productivity and augment the contributions of human workers in both the public and private sectors. We welcome the establishment of i.AI and the focus on AI deployment set out in the public sector productivity programme; as well as initiatives to increase business adoption such as the AI and Digital Hub.
74. The next Government should drive safe adoption of AI in the public sector via i.AI, the National Science and Technology Council and designated lead departmental Ministers for AI.
75. In its response to this Report, the Government should confirm the full list of public sector pilots currently being led or supported by i.AI, the criteria that determined i.AI pilot project selections, how it intends to evaluate their success and decide whether to roll them out more widely, and what other pilots are planned for the remainder of 2024.
76. i.AI should undertake an assessment of the existing civil service workforce’s AI capability, identify areas of the public sector that would benefit the most from the use of AI and where value for money can be delivered, set out how potential risks associated with its use should be mitigated, and publish a detailed AI public sector action plan. Progress against these should be reported to Parliament on an annual basis and through regular written or oral statements by Ministers.
77. The requirement for Government departments to use the Algorithmic Transparency Recording Standard should be extended to all public bodies sponsored by Government departments, from 1 January 2025.
78. In our interim Report we welcomed the establishment of the then-Foundation Model Taskforce.107 The Taskforce has subsequently become a permanent entity as the AI Safety Institute,108 and assumed a key role in the UK’s AI governance framework. In this Chapter we examine the role of the Institute, its priorities, and the AI Safety Summits it has informed.
79. Shortly after the publication of our interim Report, the Foundation Model Taskforce was renamed the Frontier Model Taskforce and announced that several leading AI researchers and technical organisations had joined or partnered with it.109 In its first progress report, the Taskforce said that it had offered these individuals:
… the opportunity to fundamentally alter society’s approach to tackling risks at the frontier of AI. These researchers and engineers will bring their skills towards giving the government the capability to work directly on frontier AI models and evaluate their risks—through model evaluations, red-teaming,110 and other aspects of safety infrastructure.111
To date, the Institute has recruited both academic researchers, from institutions including Cambridge University, the University of Oxford and Harvard University; and technical experts with experience at leading developers, including Google DeepMind, Microsoft and OpenAI.112
80. It is a credit to the commitment of those involved that the AI Safety Institute has been swiftly established, with an impressive and growing team of researchers and technical experts recruited from leading developers and academic institutions. The next Government should continue to empower the Institute to recruit the talent it needs.
81. The need to evaluate the most advanced AI models from a safety perspective was emphasised by the Prime Minister, Rt. Hon. Rishi Sunak MP, during a speech in June 2023. He said that for the UK “… leading on AI also means leading on AI safety”.113 In its second progress report the then-Taskforce said that it was “… critical that frontier AI systems are developed safely and that the potential risks of new models are rigorously and independently assessed for harmful capabilities before and after they are deployed”.114
82. The Prime Minister said that his vision for the UK as a leader in AI would be realised via three strands of work: that undertaken by the Taskforce, the pursuit of global cooperation, and the deployment of AI “… to improve people’s lives”.115 The first two have been initiated through the work of the AI Safety Institute and the organisation of an AI Safety Summit at Bletchley Park in November 2023.116 The Government set five objectives for the Summit:
83. Shortly before the Summit, the Prime Minister confirmed that the Frontier AI Taskforce would become a permanent body, the AI Safety Institute.118 Seven leading AI developers—Amazon, Anthropic, Google DeepMind, Inflection, Meta, Microsoft and OpenAI—published their safety policies ahead of the Summit.119 These, together with a collection of discussion papers prepared by the Government and the Taskforce,120 informed the roundtable discussions held over two days at Bletchley Park.121
84. Speaking to us after the Summit, the Prime Minister’s Summit representative and Chair of the Advanced Research and Invention Agency, Matt Clifford CBE, said that in addition to placing AI safety on the international agenda, the Summit had achieved four substantive outcomes:
85. In May 2024, the AI Seoul Summit saw 16 leading developers agree to implement eight commitments focused on mitigating the “severe risks” posed by the most advanced AI; with operational updates promised by an early 2025 Summit, to be hosted by France.127 These commitments are intended to deliver three outcomes:
86. The AI Safety Institute has said that it “… is not a regulator and will not determine government regulation”; but would focus on the evaluation of advanced AI, drive foundational AI safety research, and facilitate information exchange between national and international actors.129 Emran Mian, Director General for Digital Technologies and Telecoms at DSIT, explained to us why the Institute would undertake safety evaluations:
… the companies are doing quite a lot of the testing. That is great and positive. It is right that they should do it and that they should contribute to the development of the science on safety; but relying solely on the companies gives us some pause for thought, both because of the commercial imperatives that the companies may have in the rush to market, but also in the classic issue that arises in these kinds of safety conversations, about companies marking their own homework.130
The Institute has emphasised that it would not “… designate any particular AI system as ‘safe’ … [nor] hold responsibility for any release decisions”.131
87. In its third progress report the Institute confirmed that it had begun testing models,132 and that these had been selected based on “… estimates of the risk of a system possessing harmful capabilities, using inputs such as compute used for training, as well as expected accessibility”.133 In May 2024, the Institute released the results of its evaluations of five publicly-available models,134 and has also announced a programme of joint work with its counterpart in the United States that will include “… at least one joint testing exercise on a publicly accessible model”.135
88. Exactly which model, or models, the Institute will undertake pre-release testing on was not confirmed by the Secretary of State for Science, Innovation and Technology when questioned in the House of Commons,136 and had not been confirmed at the time our Report was finalised. It was reported by the news outlet Politico in April that only Google DeepMind had allowed access to its Gemini model for pre-release testing, and that Anthropic, Meta and OpenAI had yet to grant access to as-yet unreleased models.137
89. Although the Institute is not a regulator, it has undeniably played a decisive role in shaping the UK’s regulatory approach to AI. We commend the work of the Institute and its researchers in facilitating and informing the ongoing international conversation about AI governance.
90. However, we are concerned by suggestions that the Institute has been unable to access as-yet unreleased AI models to perform the pre-deployment safety testing it was set up to undertake. If true, this would undermine the delivery of the Institute’s mission and its ability to increase public trust in the technology.
91. In its response to this Report, the Government should confirm which models the AI Safety Institute has undertaken pre-deployment safety testing on, the nature of the testing, a summary of the findings, whether any changes were made by the model’s developers as a result, and whether any developers were asked to make changes but declined to do so.
92. The Government should also confirm which models the Institute has been unable to secure access to, and the reason for this. If any developers have refused access—which would represent a contravention of the reported agreement at the November 2023 Summit at Bletchley Park—the Government should name them and detail their justification for doing so.
93. The UK is not the only jurisdiction to have dedicated significant thought and resource to the development of its AI governance framework. The United States and the European Union have over the course of our Inquiry debated and developed their own regulatory regimes. In this Chapter we will examine the steps taken by policymakers in Washington, D. C. and Brussels.
94. The United States (US) is home to leading AI developers and start-ups such as OpenAI and Anthropic.138 It has led each edition of the Tortoise AI Index, a ranking of countries by AI implementation, innovation and investment.139 According to data from the Organisation for Economic Co-operation and Development (OECD), the US since 2012 has accounted for an average of 56% of global annual venture capital investment in AI and led the ranking for this metric in all but two years during the same period.140
95. Leading ‘Big Tech’ firms based in the US have also emerged as significant investors in AI,141 with Microsoft having reportedly committed up to $13 billion to OpenAI, the developer behind ChatGPT, including a $10 billion investment announced in January 2023.142
96. Just as the UK secured voluntary commitments from leading AI developers,143 the White House announced in July 2023 that Amazon, Anthropic, Google, Inflection, Meta, Microsoft, and OpenAI had agreed to make “… voluntary commitments… to help move toward safe, secure, and transparent development of AI technology”.144 A further eight companies signed up to the commitments in September.145
97. The White House has described the voluntary commitments as intended “… to advance a generative AI legal and policy regime” and said that it envisioned them “… remain[ing] in effect until regulations covering substantially the same issues come into force”.146 The commitments included:
98. Although wide-ranging, the scope of the commitments was limited. The White House said that “… where commitments mention particular models, they apply only to generative models that are overall more powerful than the current most advanced model produced by the company making the commitment”.148
99. Direct industry engagement in the form of the voluntary commitments was followed in October 2023 by the announcement of a Presidential Executive Order on the safe, secure and trustworthy development and use of AI,149 which although primarily applicable to Federal departments and agencies, was notable for its utilisation of the Defense Production Act, a law passed during the Cold War that has afforded Presidents “… significant emergency authority to control domestic industries”.150
100. The Executive Order detailed 26 requirements across eight areas, including standards for AI safety and security, advancing equity and civil rights, and ensuring responsible and effective Government use of AI.151 It placed safety testing, risk mitigation and reporting requirements on developers of the most powerful models, with any model trained using compute power above a set threshold required to comply.152
101. The White House has stated its desire to pursue the successful passage of bipartisan AI-specific legislation through the United States Congress,153 but as the November 2024 Presidential election nears the Federal legislative process has slowed and the prospect of new legislation being passed reduced.154
102. In the absence of Federal legislation, policymakers in states such as California,155 Colorado,156 Florida157 and New York158 have introduced bills and measures targeted at the development and deployment of AI. According to LexisNexis, a data analytics company, “as of January 11 [2024], 89 bills referring to AI had been pre-filed or introduced in 20 states… [in addition to] more than 100 AI bills that are being carried over from last year”.159
103. California’s Privacy Protection Agency has also published draft proposals for a regime to establish individual “… opt-out rights, pre-use notice requirements and access rights which would enable state residents to obtain meaningful information on how their data is being used for automation and AI tech”.160
104. In December 2023 political agreement was reached between representatives of the European Parliament, the Council of the European Union and the European Commission on the European Union (EU) AI Act.161 This was followed by the approval of ambassadors from the 27 EU member states in February 2024,162 and a European Parliament ratification vote in March 2024.163 This represented the culmination of a legislative process that began in April 2021 with the publication of draft proposals by the Commission.164
105. The AI Act takes a ‘horizontal’,165 risk-based approach, with AI uses categorised into four levels of risk. The EU institutions have also agreed specific requirements for general-purpose AI models. The risk categories are set out below.
106. Examples of uses deemed to present minimal risk included AI-assisted recommender systems, which suggest or recommend additional products to consumers,166 or spam filters. The Commission has said that “the vast majority” of uses for AI will fall into this category and will not be subject to any additional requirements beyond existing legislation.167
107. An Annex to the AI Act listed uses to be designated as high-risk.168 Examples included tools used in recruitment, the judicial system or democratic processes.169 The Commission has said that the list will be updated as appropriate.170 High-risk uses will be permitted subject to compliance with requirements such as “… risk-mitigation systems, high quality of data sets, logging of activity, detailed documentation, clear user information, human oversight, and a high level of robustness, accuracy and cybersecurity”.171
108. Once fully in force the AI Act will ban a small number of uses that have been deemed to pose an unacceptable level of risk, including:
109. The AI Act introduces transparency requirements to ensure that citizens are made aware whenever they interact with a machine, such as in the use of chatbots. It also introduced a requirement for “… AI generated content… to be labelled as such, and users need to be informed when biometric categorisation or emotion recognition” tools are deployed.175
110. Under the provisions of the AI Act general-purpose models, including those that underpin generative AI tools such as GPT-4176 or Google Gemini177 will be subject to specific requirements designed to ensure transparency and provide reassurances to those who deploy them.178
111. For models that have been trained using compute power above a certain threshold and that are deemed to potentially “… pose systemic risks, there will be additional binding obligations related to managing risks and monitoring serious incidents, performing model evaluation and adversarial testing”.179 These will be put into practice “… through codes of practices developed by industry, the scientific community, civil society and other stakeholders together with the Commission”.180
112. The AI Act “… will be fully applicable 24 months after entering into force”,181 likely meaning during June 2026.182 Some provisions, including the bans on prohibited uses outlined above, will enter into force by the end of 2024 and others, such as the provisions relating to general-purpose AI, are expected to apply from June 2025.183 It will apply to “public and private actors inside and outside the EU as long as the AI system is placed on the Union market, or its use affects people located in the EU”.184
113. Enforcement of the AI Act has been made a joint responsibility of designated authorities in each EU member state,185 and a new European AI Office within the Commission, which has been tasked with enforcing the provisions that apply to general-purpose AI.186
114. The International Association of Privacy Professionals (IAPP), a trade body, has pointed out that successful implementation of the AI Act will require skilled staff in each member state authority, with “… sufficient expertise in fundamental rights law, personal data protection and others”.187 The same can be said of the Commission’s AI Office, which has begun to recruit staff.188
115. Some member states, such as Spain, have established new entities to act as their designated authority. Others, such as Luxembourg, Ireland and the Netherlands, have assigned the task to existing public bodies or Government departments.189 The IAPP has highlighted the uneven states of readiness among European regulators, with enforcement of the AI Act described by some as “a tall order”.190 Non-compliance can be punishable by fines of up to €35 million or 7% of the offender’s total worldwide annual turnover in the previous financial year, whichever is higher.191
116. The scope of the AI Act is not unlimited. The following uses are exempt:
We will examine the use of AI in a military, defence, and national security context in Chapter 8 of this Report.
117. Speaking prior to the finalisation of the AI Act, the Secretary of State for Science, Innovation and Technology said that she had heard “… deep concerns among some of its member states that it will stifle innovation”.193 She also argued that the AI Act’s risk-based approach was “… quite a blunt tool… we are taking an approach that is a much more context based”.194
118. The Computer and Communications Industry Association, a trade body, has said that the impact of the AI Act “… needs to be closely monitored to avoid overburdening innovative AI developers with disproportionate compliance costs and unnecessary red tape”.195 French President Emmanuel Macron has also argued that the EU “… can decide to regulate much faster and much stronger than our major competitors. But we will regulate things that we will no longer produce or invent. This is never a good idea”.196 Hugh Milward of Microsoft UK told us that the EU’s proposed approach was “… a model of how not to do it”, citing similar concerns.197
119. Beyond the EU AI Act, the Commission has proposed an Artificial Intelligence Liability Directive that would create “… uniform rules for certain aspects of non-contractual civil liability for damage caused with the involvement of AI systems”.198 In late 2023 the EU institutions also agreed a new Directive on Liability for Defective Products to replace the existing Product Liability Directive, which will account for the growing prevalence of AI and allow individuals to bring claims against product and in some cases product component manufacturers.199
120. China ranked second in the latest Tortoise Global AI Index, behind the United States.200 In a submission to our inquiry Dr Steve Rolf, a research fellow at the University of Sussex, told us that China wanted to increase the size of its digital economy:
… to 10 per cent of GDP by 2025 and to this end commercial AI applications have been successfully embedded in areas like smart cities programs, education and healthcare, and autonomous vehicles.201
121. Statistics released by the Chinese authorities showed a notable increase in imports of data processing equipment and computer chips in the first quarter of 2024,202 and it has been reported that the Chinese state has offered significant subsidies to encourage the development of AI start-ups whilst maintaining close oversight of the approval of new models for release.203
122. Matt Sheehan, a fellow at the Carnegie Endowment for International Peace, has described China as being “… in the midst of rolling out some of the world’s earliest and most detailed regulations governing AI”.204 He has termed China’s approach to AI as the first example of its authorities “… having to do a trade-off between two Communist party goals of sustaining AI leadership and controlling information”.205
123. From a geopolitical perspective, Dr Rolf’s submission to our inquiry described China’s overall regulatory approach as “… substantially driven by rivalry with the United States for technological supremacy in digital technology and AI”.206 Shortly after the publication of our interim Report, the Government confirmed that China had been invited to the AI Safety Summit at Bletchley Park, with Foreign Secretary Rt. Hon. James Cleverly MP describing a strategy “… to engage [with China] where it is in the U.K.’s national interest”.207
124. China was among the signatories to the Bletchley Declaration announced at the 2023 AI Safety Summit,208 and has subsequently engaged with the United States on AI-related topics at researcher, business and diplomatic levels, according to media reports.209 We will return to the International Coordination Challenge in Chapter 8 of this Report.
125. Both the UK and EU have emphasised the importance of technical industry standards. The European Commission has described the AI Act as a framework that “… leaves the concrete technical solutions and operationalisation primarily to industry-driven standards”.210
126. Similarly, in its response to the AI White Paper consultation the Government said that it would continue to support UK participation in standards fora “… to both leverage the benefits of global technical standards here in the UK and deliver global digital technical standards shaped by democratic values”.211
127. In a submission to our inquiry, the British Standards Institution (BSI), which represents the UK view on European and international technical standards organisations, said that “… standards and accreditation offer a business and consumer friendly alternative to regulation and enable interoperability in complex supply-chains”, by “… underpinning high-level regulatory objectives with technical or framework specificities that can be adopted by businesses”.212
128. The BSI said that the UK should remain an active participant in international standards fora, “… to learn from the approach of others, ensure alignment where possible, and to influence others in adopting a use-case -specific approach to risk based AI regulation”.213 An AI Standards Hub has been established by the BSI, the Alan Turing Institute and National Physical Laboratory; with support from DSIT “… to support stakeholders to understand and engage with AI standardisation and strengthen AI governance practices domestically and internationally”.214
129. In our interim Report we highlighted moves by both the United States and European Union to develop their own approaches to AI governance. The subsequent White House Executive Order and the EU AI Act are clear attempts to secure competitive regulatory advantage.
130. It is true that the size of both the United States and European Union markets may mean that ‘the Washington effect’ and ‘Brussels effect’—referring to the de facto standardising of global regulatory approaches, potentially to the detriment of the UK’s distinct approach—will apply to AI governance. Nevertheless, the distinctiveness of the UK’s approach and the success of the AI Safety Summit have underlined the significance of its current and future role.
131. Both the US and EU approaches to AI governance have their downsides. The scope of the former only imposes a requirement on Federal bodies and relies on voluntary commitments from leading developers. The latter has been criticised for its top-down, prescriptive approach and the potential for uneven implementation across different member states.
132. The UK is entitled to pursue an approach that considers developments in other jurisdictions but does not unthinkingly replicate them. However, where there are lessons to be learned from other jurisdictions, the next Government should be willing to apply them.
133. The UK has a long history of encouraging technological innovation by offering a stable, expert regulatory environment coupled with clear industry standards. The current Government is therefore right to have encouraged the growth of a strong AI sector in the UK, engaged with leading developers through the AI Safety Institute and future Summits, and participated in international standards fora. This international agenda should be continued by the next Government, and coupled with the swift establishment of a domestic framework that sufficiently addresses the Twelve Challenges of AI Governance highlighted in our interim Report.
134. In our interim Report, published in August 2023, we identified a sense in many jurisdictions, including the UK, “… that the pace of development of AI requires an urgent response from policymakers if the public interest is not to be outstripped by the pace of deployment”.215 We also found that the process of determining a coherent policy response to the increasing prevalence of AI was being complicated by “… the reality that the optimal responses to all of the challenges AI gives rise to are not always—at this stage—obvious”.216
135. Our interim Report provided an initial contribution to these important discussions. It identified Twelve Challenges that AI governance frameworks must meet and called on the Government and regulators to address them through the UK’s approach.217
136. Although progress has been made in the UK and other jurisdictions, we believe that the Twelve Challenges identified in our interim Report still apply. In this Chapter we suggest potential solutions to each of them. These should not be viewed as fully-formed policy responses, but as representing this Committee’s ‘starter for ten’ in a complex, rapidly developing area.
Developers and deployers of AI models and tools must not merely acknowledge the presence of inherent bias in datasets, they must take steps to mitigate its effects
137. Our interim Report highlighted how developers and researchers relied on data to test, train, operate and refine AI models and tools,218 and that these datasets contained inherent bias.219
138. The Government has acknowledged that AI can “… entrench bias and discrimination” and said that it “… is working closely with the Equality and Human Rights Commission and ICO [Information Commissioner’s Office] to develop new solutions to address bias and discrimination in AI systems”.220
139. Other public bodies have also taken steps to address the Bias Challenge. The Office of the Police Chief Scientific Adviser has produced a Covenant for Using AI in Policing,221 whilst the Metropolitan Police has commissioned independent testing and analysis of the performance its facial recognition tool, procured from NEC, by the National Physical Laboratory.222 In the field of health, the Medicines and Healthcare products Regulatory Agency has said that it will require approval applications for new medical devices, many of which are AI-assisted, to detail how they will address bias.223
140. AI can entrench and accelerate existing biases. The current Government, future administrations and sectoral regulators should require deployers of AI models and tools to submit them to robust, independent testing and performance analysis prior to deployment.
141. Model developers and deployers should be required to summarise what steps they have taken to account for bias in datasets used to train models, and to statistically report on the levels of bias present in outputs produced using AI tools. This data should be routinely disclosed in a similar way to company pay gap reporting.
Privacy and data protection frameworks must account for the increasing capability and prevalence of AI models and tools, and ensure the right balance is struck
142. Our interim Report highlighted the need to balance the protection of privacy and the potential benefits to be gained from the deployment of AI, particularly in sectors such as law enforcement.224
143. Information Commissioner John Edwards told us that the Information Commissioner’s Office (ICO), one of the UK’s principal privacy regulators, had paid close attention to the implications of AI’s increasing prevalence for some time and had sought “… to ensure that all parts of the supply chain in AI—whether they are developing models, training models or deploying retail applications of them” were aware of their obligations under existing data protection regulations.225
144. In a paper that set out its strategic approach to AI, the ICO said that many of the risks associated with the deployment of AI “… derive from how data—and specifically personal data—is used in the development and deployment of AI systems”.226 It said that by enforcing data protection law, it would ensure that “organisations who are accountable for the processing of personal data are expected to identify the risks, mitigate them and be able to demonstrate how they achieve this”.227 In sensitive areas such as facial recognition technology, the ICO has said that “… deployments must be proportionate and strike the correct balance between privacy intrusion and the purpose they are seeking to achieve”.228
145. Regulators and deployers should ensure that the right balance is maintained between the protection of privacy and pursuing the potential benefits of AI. Determining this balance will depend on the context in which the technology is being deployed, with reference to the relevant laws and regulations.
146. Sectoral regulators should publish detailed guidance to help deployers of AI strike the balance between the protection of privacy and securing the technology’s intended benefits. In instances where regulators determine that this balance has not been met, or where the relevant laws or regulatory requirements have not been met, it should impose sanctions or prohibit the use of AI models or tools.
Those who use AI to misrepresent others, or allow such misrepresentation to take place unchallenged, must be held accountable
147. Our interim Report noted how new AI-assisted tools had significantly expanded “opportunities for malign actors to ‘pass off’ content as being associated with particular individuals or organisations when it is in fact confected”.229 Since then, the extent to which the Misrepresentation Challenge is present across society and the economy has been further underlined.
148. The increasing prevalence of AI-assisted tools capable of producing pornographic ‘deepfake’ images and videos targeted at women and girls has been highlighted by Glamour magazine.230 Committee members participated in an event in Parliament that discussed the findings of the Glamour Consent Survey undertaken together with Refuge, a charity that provides specialist support for women and children experiencing domestic violence,231 and how the powers given to Ofcom, the online safety regulator, under the Online Safety Act 2023 should be implemented in a way that addresses the increasing threat posed by online deepfakes.232
149. The Government subsequently brought forward an amendment to the Criminal Justice Bill before Parliament, which would criminalise the creation of sexually explicit deepfake images without consent, or the installation of equipment to enable someone to do so.233 The amendment was added to the Bill on 15 May,234 but as our Report was finalised it was unclear whether its remaining stages would be completed prior to the dissolution of Parliament.
150. We welcome the Government amendment to the Criminal Justice Bill as a necessary step towards ensuring the UK’s legal framework reflects the current state of technological development and protects citizens, primarily women and girls, from the consequences of AI-assisted misrepresentation, including deepfake pornography. Should the Bill’s remaining stages fail to be completed prior to the dissolution of Parliament, the next Government must introduce similar provisions as soon as is practicable after the General Election.
151. The Misrepresentation Challenge has also become increasingly visible in our politics as the General Election expected this year approaches.235 Deepfake audio and video clips that misrepresented the Leader of the Opposition, Rt. Hon. Sir Keir Starmer MP236 and the Mayor of London, Sadiq Khan237 were widely viewed and circulated on online platforms. The National Cyber Security Centre, a part of GCHQ, said in its 2023 annual review that:
… rather than presenting entirely new risks, it is AI’s ability to enable existing techniques which poses the biggest threat. For example: large language models will almost certainly be used to generate fabricated content, AI-created hyperrealistic bots will make the spread of disinformation easier and the manipulation of media for use in deepfake campaigns will likely become more advanced.238
152. The Government has established a Defending Democracy Taskforce intended “… to reduce the threat of foreign interference in our democracy by bringing together a wide range of expertise across government, the intelligence community and industry”.239 It has also cited provisions in existing legislation, such as the Elections Act 2022, National Security Act 2023 and Online Safety Act 2023 as capable of helping to address the Misrepresentation challenge.240
153. Some online platforms and technology companies have also accepted their role in addressing AI-generated content that aims to interfere in the democratic process. At the February 2024 Munich Security Conference, a group of companies including Google, Meta, Microsoft, OpenAI, TikTok and X, formerly Twitter, announced “… a set of commitments to deploy technology countering harmful AI-generated content meant to deceive voters”.241
154. The Government and regulatory authorities, informed by the work of the Defending Democracy Taskforce, should safeguard the integrity of the upcoming General Election campaign in their approach to the online platforms that host deepfake content which seeks to exert a malign influence on the democratic process. If these platforms are found to have been slow to remove such content, or to have facilitated its spread, regulators must take stringent enforcement action—including holding senior leadership personally liable and imposing financial sanctions.
155. A cross-Government public awareness campaign should be launched to inform the public about the growing prevalence of AI-assisted misrepresentation, the potential consequences, what the Government is doing to address the Challenge, and what steps individuals can take to protect themselves online.
Access to data, and the responsible management of it, are prerequisites for a healthy, competitive and innovative AI industry and research ecosystem
156. Our interim Report highlighted the extent to which AI developers and researchers alike rely on access to high-quality data, and the potential competition concerns that increased market consolidation could pose.242 Since then, the Competition and Markets Authority (CMA) has opened an investigation into the relationship between Microsoft and OpenAI, specifically whether it “… has resulted in a relevant merger situation and, if so, the impact that the merger could have on competition in the UK”.243
157. Sarah Cardell, Chief Executive of the CMA, has observed that leading developers hold “… strong positions in one or more critical inputs for upstream model development, while also controlling key access points or routes to market for downstream deployment”, a situation that has raised concerns “… that they could leverage power up and down the value chain” to the detriment of free and open competition.244
158. Data is chief among these critical inputs, due to the volume required to train current AI models. Whilst the banking duty of confidentiality is well-established in the UK,245 Nikhi Rathi, Chief Executive of the Financial Conduct Authority, has highlighted how “safe data sharing can benefit firms, markets and consumers”.246 Air Street Capital, a UK-based venture capital firm that invests in AI-first technology and life science companies, has said that the UK “could consider creating a national data bank… [using] data from the BBC, government departments, our universities, and other sources”, which could then be made available to developers.247
159. At the so-called ‘frontier’ of AI a small group of leading developers are responsible for and accruing significant benefits from the development of advanced models and tools—thanks in part to their ability to access the necessary training data. This potential dominance is arguably to the detriment of free and open competition.
160. As the regulator responsible for promoting competitive markets and tackling anti-competitive behaviour, the CMA should identify abuses of market power and use its powers to stop them. This could take the form of levying fines or requiring the restructuring of proposed mergers.
161. AI models and tools rely on access to high-quality input data. The phrase ‘garbage in, garbage out’ is not new, but it is particularly applicable to AI. The potential for human error and bias notwithstanding, deployers should not solely rely on outputs produced with AI tools to determine their decision-making, particularly in areas that could affect the rights and standing of the individuals or entities concerned, such as insurance decisions or recruitment. These algorithmic decisions should always be reviewed and verified by trained humans, and those affected should have the right to challenge these decisions—a process that should also be human-centred.
162. The Government and future administrations should support the emergence of more AI startups in the UK by ensuring they can access the high-quality datasets they need to innovate. This could involve facilitating access to anonymised public data from data.gov.uk, the NHS and BBC via a National Data Bank, subject to appropriate safeguards.
Democratising and widening access to compute is a prerequisite for a healthy, competitive and innovative AI industry and research ecosystem
163. As one analysis described, “… data is the raw material that is processed by compute; put differently, compute is the ‘engine’ fuelled by large amounts of data”.248 Access to both has become essential to developing and deploying at scale many of the AI tools available today.249
164. The Government has identified access to compute as key to the further development of AI-related research and industry in the UK, and has announced the establishment of an AI Research Resource and a new cluster of supercomputers.250 The 2024 Spring Budget said that the Government would set out “… how access to the UK’s cutting edge public compute facilities will be managed, so that both researchers and innovative companies are able to secure the computing power they need… “.251 The Advanced Research and Invention Agency has also launched a programme aimed at reducing the cost of AI hardware, Scaling Compute.252
165. We welcome the Government’s moves to establish a dedicated AI Research Resource and a cluster of supercomputers but are concerned that it has yet to set out further details of how researchers and startups will be able to access the compute they need to maximise the potential benefits of AI across society and the economy.
166. The Government, or its successor administration, should publish an action plan and proposed deliverables for both the AI Research Resource and its cluster of supercomputers, and further details of the terms under which researchers and innovative startups will be able to access them. It should also undertake a feasibility study into the establishment of a National Compute Cluster that could be made available to researchers and startups.
We should accept that the workings of some AI models are and will remain unexplainable and focus instead on interrogating and verifying their outputs
167. Our interim Report described how new AI models and tools “… have increasingly become ‘black boxes’, that is, their decision-making processes are not explainable”.253 “Appropriate transparency and explainability” was one of the five high-level principles set out in the March 2023 AI White Paper.254
168. The emergence of large language models (LLMs) in particular has encapsulated the Black Box Challenge. A paper prepared by an expert group to inform discussions at of the AI Seoul Summit in May 2024 described how even “… researchers currently cannot generate human-understandable accounts of how general-purpose AI models and systems arrive at outputs and decisions”.255
169. In a Report examining LLMs, the House of Lords Communications and Digital Committee described them as “… very complex and poorly understood; [they] operate blackbox decisionmaking; datasets are so large that meaningful transparency is difficult … ”.256
170. The Black Box Challenge is one of the most paradigm-shifting consequences of AI, as it upends our well-established reliance on explainability and understanding. Given the complexity of currently available and in all likelihood future models, the starting point should be an acknowledgement how little we can understand about how many AI models produce their outputs, an acceptance that new ways of thinking will be required, and a regulatory approach that accounts for the impossibility of total explainability.
171. The regulators charged with implementing the Government’s high-level AI governance principles should, in their approach to these models and tools, prioritise testing and verifying their outputs, as well seeking to establish—whilst accepting the difficulty of doing so with absolute certainty—how they arrived at them.
The question should not be ‘open’ or ‘closed’, but rather whether there is a sufficiently diverse and competitive market to support the growing demand for AI models and tools
172. Our interim Report found differing views as to whether the code, training data and weights of AI models should be freely available.257 We referred to this as the Open-Source Challenge.
173. Since the publication of our interim Report, these debates have intensified. Leading developers such as Google and OpenAI have made their most advanced AI models proprietary,258 whilst Meta has differentiated itself by making its Llama 2 and Llama 3 models more openly available—although, as TechCrunch has noted, in a move designed to safeguard against competitors Meta does not permit other developers to use Llama to train generative AI models, whilst app developers with over 700 million monthly users are required to seek a commercial licence, which Meta could refuse.259
174. In its Report examining LLMs, the House of Lords Communications and Digital Committee said that “the UK has particular strengths in mid-tier businesses and will benefit most from a combination of open and closed source technologies”.260 The importance of having a range of ‘open’ and ‘closed’ models available has also been underlined by the Competition and Markets Authority.261
175. The open-source approach has underpinned many technological breakthroughs, including the Internet and AI. Whilst some providers of products and services, such as AI models and their applications, will want to keep elements of their offerings proprietary, a healthy AI marketplace should be sufficiently diverse to support both ‘open’ and ‘closed’ options. The volume of investment flowing into AI developers of all types of models, rather than one or the other, is evidence of this market diversity.
176. When procuring AI models for deployment in the public sector the Government and public bodies should utilise those best suited to the task.
177. The Secretary of State for Science, Innovation and Technology told us that the debate associated with the Open-Source Challenge should be more nuanced:
… our concerns should be around the capability. Sometimes, depending on the capability, you will be less concerned if it is open source. If the capability presents more risk, you will be more concerned.262
178. These risks include those detailed in a submission to our Inquiry by the Internet Watch Foundation (IWF), which highlighted the use of AI tools—mostly but not exclusively open-source—to generate child sexual abuse imagery. The IWF recommended “… greater regulatory oversight of open-source models and the data sets they are built on before they are released”, and the insertion of safeguards by developers to prevent such content being generated with their proprietary models.263
179. The Government has said that “pre-deployment testing could inform the deployment options available for a model and change the risk prevention steps required of organisations prior to the model’s release”. It has also committed to engaging closely with the open-source community and experts on the question of open release.264
180. The Government should in its response to this Report tell us how it will ensure law enforcement and regulators are adequately resourced to respond to the growing use of AI models and tools to generate and disseminate harmful and illegal content.
The Government should broker a fair, sustainable solution based around a licensing framework governing the use of copyrighted material to train AI models
181. Our interim Report detailed concerns about “the ‘scraping’ of copyrighted content from online sources without permission” in order to train AI models,265 and representatives from the creative industries told us that they hoped to reach a mutually beneficial solution with the AI sector, potentially in the form of a licensing framework for the use of copyrighted content to train models and tools.266
182. In the summer of 2023 the Intellectual Property Office (IPO), an executive agency of the Government, convened a working group comprised of representatives from the technology, creative and research sectors, with a view to agreeing a voluntary code of practice on copyright and AI.267
183. In subsequent months the number of relevant legal proceedings in various jurisdictions has continued to increase, with the ongoing case filed by the New York Times against OpenAI and Microsoft over their alleged use of copyrighted work among the most high-profile examples.268
184. In its response to the AI White Paper consultation, the Government confirmed that the working group convened by the IPO had been unable to agree a code of practice, and that Ministers from DSIT and the Department for Culture, Media and Sport would bring forward proposals “… to ensure the workability and effectiveness of an approach that allows the AI and creative sectors to grow together in partnership”.269 It has subsequently said that it “… will focus on greater transparency from AI developers and ensure that AI outputs are properly attributed”, in co-operation with international counterparts.270
185. The growing volume of litigation relating to alleged use of works protected by copyright to train AI models and tools, and the value of high-quality data needed to train future models, has underlined the need for a sustainable framework that acknowledges the inevitable trade-offs and establishes clear, enforceable rules of the road. The status quo allows developers to potentially benefit from the unlimited, free use of copyrighted material, whilst negotiations are stalled.
186. The current Government, or its successor administration, should ensure that discussions regarding the use of copyrighted works to train and run AI models are concluded and an implementable approach agreed. It seems inevitable that this will involve the agreement of a financial settlement for past infringements by AI developers, the negotiation of a licensing framework to govern future uses, and in all likelihood the establishment of a new authority to operationalise the agreement. If this cannot be achieved through a voluntary approach, it should be enforced by the Government, or its successor administration, in co-operation with its international partners.
Determining liability for AI-related harms is not just a matter for the courts—Government and regulators can play a role too
187. Our interim Report discussed the “increasingly complex and international supply chains for AI models and tools”, and the Challenge this created in terms of the distribution of liability for harms.271 In its response to the AI White Paper consultation, the Government identified “… how to allocate liability across the supply chain” as one of the key questions that policymakers would have to answer through their regulatory approaches to AI.272
188. The Government has acknowledged that many respondents to its AI White Paper consultation “… endorsed further government intervention to ensure the fair and effective allocation of liability”.273 It has said that it will seek expert input on whether updates to civil or criminal liability frameworks are required to account for the continued emergence of advanced AI models and tools.274 This is expected to inform the decision on whether to introduce AI-specific legislation discussed in Chapter 3.
189. Nobody who uses AI to inflict harm should be exempted from the consequences, whether they are a developer, deployer, or intermediary. The next Government together with sectoral regulators should publish guidance on where liability for harmful uses of AI falls under existing law. This should be a cross-Government undertaking. Sectoral regulators should ensure that guidance on liability for AI-related harms is made available to developers and deployers as and when it is required. Future administrations and regulators should also, where appropriate, establish liability via statute rather than simply relying on jurisprudence.
Education is the primary tool for policymakers to respond to the growing prevalence of AI, and to ensure workers can ask the right questions of the technology
190. Our interim Report detailed different perspectives on the impact of AI on employment and found that whatever the outcome, planning ahead should be the key task for policymakers.275
191. Economists Daron Acemoglu and Simon Johnson have argued that policymakers “… must recognise that there is no singular, inevitable path of development for new technology”. The key question, in their view, should be “… what policies would put AI development on the right path, with greater focus on enhancing what all workers can do?”.276
192. The Government has acknowledged the importance of preparing UK workers for an AI-enabled economy, from both a skills and employment rights perspective. It has argued that whilst the impact of AI will be felt differently in different sectors, “… we can be confident that we will need new AI-related skills through national qualifications and training provision”.277
193. AI is already changing the nature of work, and as the technology evolves this process is likely to accelerate, placing some jobs at risk. At the same time, there are productivity benefits to be won, provided people are equipped with the skills to fruitfully utilise AI. This is a process that should begin in the classroom, and through greater prioritisation of initiatives such as the Lifetime Skills Guarantee and digital Skills Bootcamps.
194. The current Government, or its successor, should commission a review into the possible future skills and employment consequences of AI, along the lines of the 2017 Taylor Review of modern working practices which examined the landscape, suggested ideas for debate and has resulted in legislative change. It should also in its response to this Report tell us how it will ensure workers whose jobs are at risk of automation will be able to retrain and acquire the skills necessary to change careers.
A global governance regime for AI may not be realistic nor desirable, even if there are economic and security benefits to be won from international co-operation
195. The global nature of AI governance debates were highlighted in our interim Report,278 and have since been underlined by the proliferation of international-level initiatives to establish a consensus around governance frameworks for the development and deployment of the technology.
196. Notable among these was the inaugural AI Safety Summit organised at Bletchley Park and discussed in Chapter 6 of this Report. The UK has also participated in ongoing initiatives by the Council of Europe, G7, G20, Organisation for Economic Co-operation and Development and United Nations; and signed a number of bilateral partnerships involving joint work on AI and AI governance.279
197. At the AI Seoul Summit in 2024, a group of jurisdictions including the UK, United States and European Union, announced that they would collaborate more closely on matters relating to AI safety, innovation and inclusivity. The Seoul Declaration recognised “… the importance of interoperability between AI governance frameworks in line with a risk-based approach to maximize the benefits and address the broad range of risks from AI”.280 A Ministerial communique published at the conclusion of the Summit confirmed that 27 nations had agreed to “develop shared risk thresholds for frontier AI development and deployment… as part of a wider push to develop global standards to address specific AI risks”.281
198. The current Government has said that it sees value in pursuing “… coherence between our AI governance frameworks to ensure that businesses can operate effectively in both the UK and wider global markets”.282
199. However, since our interim Report the race between different jurisdictions to secure competitive advantage, often underpinned by geopolitical calculations, has become increasingly visible. It is notable for example that Peng Xiao, the Chief Executive of G42, a leading developer based in the United Arab Emirates, told the Financial Times in December 2023 that it would no longer use hardware suppliers from China, in order to maintain and deepen its links with United States-based partners including Microsoft and OpenAI.283
200. A subsequent $1.5 billion investment by Microsoft in G42 was announced in April 2024, and described by the Financial Times as “… part of Washington’s efforts to achieve supremacy over Beijing in the development of artificial intelligence and other sensitive technologies”.284 G42 was among the developers to sign on to the voluntary commitments announced at the May 2024 AI Seoul Summit.285
201. In a national security context, the North Atlantic Treaty Organisation (NATO) adopted an AI Strategy in October 2021.286 It sets out six principles for the responsible use of AI and has been operationalised by the piloting of AI in NATO areas of operation “… as diverse as cyber defence, climate change and imagery analysis”.287
202. We welcome the organisation of the AI Safety Summit at Bletchley Park and commend the Government on bringing many key actors together. We look forward to subsequent Summits and hope that the consensus and momentum delivered at Bletchley Park can be maintained.
203. However, looking beyond the AI safety discussion, we do not believe that harmonisation for harmonisation’s sake should be the end goal of international AI governance discussions. A degree of distinction between different regulatory regimes is, in our view, inevitable. Such distinction may be motivated by geopolitics, but it may also simply be a case of healthy economic competition.
204. Future AI Safety Summits must focus on the establishment of international dialogue mechanisms to address current, medium- and longer-term safety risks presented by the growing use of AI; and the sharing of best practice to ensure its potential benefits are realised in all jurisdictions. This should not set us on the road to a global AI governance regime—we are unconvinced that such a prospect is either realistic or desirable.
Existential AI risk may not be an immediate concern but it should not be ignored, even if policy and regulatory activity should primarily focus on the here and now
205. Our interim Report highlighted debates over the security implications of AI’s increasing prevalence, and over the existential risks that it may or may not pose.288 These were described in the March 2023 AI White Paper by the Government as “… high impact but low probability”.289
206. Since the publication of the AI White Paper and our interim Report, the Government has highlighted the most serious potential risks associated with the advancing capability of AI, through the establishment of the AI Safety Institute and the organisation of the AI Safety Summit, discussed in Chapter 6 of this Report. The AI Safety Institute has since said that its testing work will include examination of the potential for autonomous systems, which it has defined as those “… that are deployed to act semi-autonomously in the real world. This includes the ability for these systems to autonomously replicate, deceive humans and create more powerful AI models”.290
207. A discussion paper written to inform discussions at the inaugural AI Safety Summit and published by the Government described ‘loss of control’ risks posed by advanced AI, and categorised these as:
208. A subsequent paper prepared ahead of the AI Seoul Summit in May 2024 concluded that:
The worst outcomes could see the emergence of risks like large-scale unemployment, general-purpose AI-enabled terrorism, or even humanity losing control over general-purpose AI systems. There is no consensus among experts about how likely these risks are and when they might occur.292
Professor Stuart Russell, a professor of computer science at the University of California, Berkeley who contributed to the paper, has spoken of the need for Governments to agree “… a treaty that compels developers to write an off-switch into future and existing software”,293 in light of findings that in some cases “… AI agents could have a tendency to ‘seek power’ by accumulating resources, interfering with oversight processes, and avoiding being deactivated, because these actions help them achieve their given goals”.294
209. The debate over the existential risk—or lack of it—posed by the increasing prevalence of AI has attracted significant attention. However, the Government’s initial assessment, that such existential risks are high impact but low probability, appears to be accurate. Nevertheless, given the potential consequences should risks highlighted by the AI Safety Institute and other researchers be realised, it is right for Governments to continue to engage with experts on the issue.
210. When implementing the principles set out in the AI White Paper regulatory activity should be focused on here-and-now impacts. Assessing and responding to existential risk should primarily be the responsibility of the UK’s national security apparatus, supported by the AI Safety Institute.
211. Should the acuteness of existential AI risk be judged to have increased, discussions regarding the implications and possible response should take place in international fora, such as AI Safety Summits.
1. With a General Election approaching we have sought to make this Report futureproof and believe that our conclusions and recommendations will remain applicable to future Administrations. It is important that the timing of the General Election does not stall necessary efforts by the Government, developers and deployers of AI to increase the level of public trust in a technology that has become a central part of our everyday lives. (Paragraph 8)
2. If governed appropriately, we believe that AI can deliver on its significant promise, to complement and augment human activity. The Government has articulated the case for AI: better public services, high quality jobs and a new era of economic growth driven by advances in AI capabilities. (Paragraph 22)
3. The Government is right to emphasise the potential societal and economic benefits to be won from the strategic deployment of AI. However, as our interim Report highlighted, the challenges are as clear as the potential benefits, and these benefits cannot be realised without public trust in the technology. (Paragraph 23)
4. The Government should certainly make the case for AI but should equally ensure that its regulatory framework addresses the Twelve Challenges of AI Governance that we have identified in our interim Report; and offer potential solutions to in this Report. (Paragraph 24)
5. The next Government should stand ready to introduce new AI-specific legislation, should an approach based on regulatory activity, existing legislation and voluntary commitments by leading developers prove insufficient to address current and potential future harms associated with the technology. (Paragraph 33)
6. The Government should in its response to this Report provide further consideration of the criteria on which a decision to legislate will be triggered, including which model performance indicators, training requirements such as compute power or other factors will be considered. (Paragraph 34)
7. The next Government should commit to laying before Parliament quarterly reviews of the efficacy of its current approach to AI regulation, including a summary of technological developments related to its stated criteria for triggering a decision to legislate, and an assessment whether these criteria have been met. (Paragraph 35)
8. We welcome confirmation that the Government will undertake a regulatory gap analysis to determine whether regulators require new powers to respond properly to the growing use of AI, as recommended in our interim Report. However, as the end of this Parliament approaches, there is no longer time to bring forward any updates to current regulatory remits and powers, should they be discovered to be necessary. This could constrain the ability of regulators to properly implement the Government’s AI principles and undermine the UK’s overall approach. (Paragraph 40)
9. The next Government should conduct and publish the results its regulatory gap analysis as soon as is practicable. If the analysis identifies any legislation required to close regulatory gaps, this should be brought forward in time for it to be enacted as soon as possible after the General Election. (Paragraph 41)
10. The general-purpose nature of AI will, in some instances, lead to regulatory overlap, and a potential blurring of responsibilities. This could create confusion on the part of consumers, developers and deployers of the technology, as well as regulators themselves. (Paragraph 45)
11. The steering committee that the Government has said it will establish should be empowered to provide guidance and, where necessary, direction to help regulators navigate any overlapping remits, whilst respecting the independence of the UK’s regulators. (Paragraph 45)
12. The regulatory gap analysis being undertaken by the Government should identify, in consultation with the relevant regulators and co-ordinating entities such as the Digital Regulation Cooperation Forum and the AI and Digital Regulations Service, areas where new AI models and tools will necessitate closer regulatory co-operation, given the extent to which some uses for AI, and some of the challenges these can present—such as accelerating existing biases—are covered by more than one regulator. The gap analysis should also put forward suggestions for delivering this co-ordination, including joint investigations, a streamlined process for regulatory referrals, and enhanced levels of information sharing. (Paragraph 46)
13. The increasing prevalence and general-purpose nature of AI will create challenges for the UK’s sectoral regulators, however expert they may be. The AI challenge can be summed up in a single word: capacity. Ofcom, for example, is combining implementation of a broad new suite of powers conferred on it by the Online Safety Act 2023, with formulating a comprehensive response to AI’s deployment across its wider remit. Others will be required to undertake resource-intensive investigations and it is vital that they are able, and empowered, to do so. All will be required to pay greater attention to the outputs of AI tools in their sectors, whilst paying due regard to existing innovation and growth-related objectives. (Paragraph 55)
14. The announced £10 million to support regulators in responding to the growing prevalence of AI is clearly insufficient to meet the challenge, particularly when compared to the UK revenues of leading AI developers. (Paragraph 56)
15. The next Government must announce further financial support, agreed in consultation with regulators, that is commensurate to the scale of the task. It should also consider the benefits of a one-off or recurring industry levy, that would allow regulators to supplement or replace support from the Exchequer for their AI-related activities. (Paragraph 57)
16. AI can be used to increase productivity and augment the contributions of human workers in both the public and private sectors. We welcome the establishment of i.AI and the focus on AI deployment set out in the public sector productivity programme; as well as initiatives to increase business adoption such as the AI and Digital Hub. (Paragraph 73)
17. The next Government should drive safe adoption of AI in the public sector via i.AI, the National Science and Technology Council and designated lead departmental Ministers for AI. (Paragraph 74)
18. In its response to this Report, the Government should confirm the full list of public sector pilots currently being led or supported by i.AI, the criteria that determined i.AI pilot project selections, how it intends to evaluate their success and decide whether to roll them out more widely, and what other pilots are planned for the remainder of 2024. (Paragraph 75)
19. i.AI should undertake an assessment of the existing civil service workforce’s AI capability, identify areas of the public sector that would benefit the most from the use of AI and where value for money can be delivered, set out how potential risks associated with its use should be mitigated, and publish a detailed AI public sector action plan. Progress against these should be reported to Parliament on an annual basis and through regular written or oral statements by Ministers. (Paragraph 76)
20. The requirement for Government departments to use the Algorithmic Transparency Recording Standard should be extended to all public bodies sponsored by Government departments, from 1 January 2025. (Paragraph 77)
21. It is a credit to the commitment of those involved that the AI Safety Institute has been swiftly established, with an impressive and growing team of researchers and technical experts recruited from leading developers and academic institutions. (Paragraph 80)
22. The next Government should continue to empower the Institute to recruit the talent it needs. (Paragraph 80)
23. Although the Institute is not a regulator, it has undeniably played a decisive role in shaping the UK’s regulatory approach to AI. We commend the work of the Institute and its researchers in facilitating and informing the ongoing international conversation about AI governance. (Paragraph 89)
24. However, we are concerned by suggestions that the Institute has been unable to access as-yet unreleased AI models to perform the pre-deployment safety testing it was set up to undertake. If true, this would undermine the delivery of the Institute’s mission and its ability to increase public trust in the technology. (Paragraph 90)
25. In its response to this Report, the Government should confirm which models the AI Safety Institute has undertaken pre-deployment safety testing on, the nature of the testing, a summary of the findings, whether any changes were made by the model’s developers as a result, and whether any developers were asked to make changes but declined to do so. (Paragraph 91)
26. The Government should also confirm which models the Institute has been unable to secure access to, and the reason for this. If any developers have refused access—which would represent a contravention of the reported agreement at the November 2023 Summit at Bletchley Park—the Government should name them and detail their justification for doing so. (Paragraph 92)
27. In our interim Report we highlighted moves by both the United States and European Union to develop their own approaches to AI governance. The subsequent White House Executive Order and the EU AI Act are clear attempts to secure competitive regulatory advantage. (Paragraph 129)
28. It is true that the size of both the United States and European Union markets may mean that ‘the Washington effect’ and ‘Brussels effect’—referring to the de facto standardising of global regulatory approaches, potentially to the detriment of the UK’s distinct approach—will apply to AI governance. Nevertheless, the distinctiveness of the UK’s approach and the success of the AI Safety Summit have underlined the significance of its current and future role. (Paragraph 130)
29. Both the US and EU approaches to AI governance have their downsides. The scope of the former only imposes a requirement on Federal bodies and relies on voluntary commitments from leading developers. The latter has been criticised for its top-down, prescriptive approach and the potential for uneven implementation across different member states. (Paragraph 131)
30. The UK is entitled to pursue an approach that considers developments in other jurisdictions but does not unthinkingly replicate them. However, where there are lessons to be learned from other jurisdictions, the next Government should be willing to apply them. (Paragraph 132)
31. The UK has a long history of encouraging technological innovation by offering a stable, expert regulatory environment coupled with clear industry standards. The current Government is therefore right to have encouraged the growth of a strong AI sector in the UK, engaged with leading developers through the AI Safety Institute and future Summits, and participated in international standards fora. This international agenda should be continued by the next Government, and coupled with the swift establishment of a domestic framework that sufficiently addresses the Twelve Challenges of AI Governance highlighted in our interim Report. (Paragraph 133)
32. AI can entrench and accelerate existing biases. The current Government, future administrations and sectoral regulators should require deployers of AI models and tools to submit them to robust, independent testing and performance analysis prior to deployment. (Paragraph 140)
33. Model developers and deployers should be required to summarise what steps they have taken to account for bias in datasets used to train models, and to statistically report on the levels of bias present in outputs produced using AI tools. This data should be routinely disclosed in a similar way to company pay gap reporting. (Paragraph 141)
34. Regulators and deployers should ensure that the right balance is maintained between the protection of privacy and pursuing the potential benefits of AI. Determining this balance will depend on the context in which the technology is being deployed, with reference to the relevant laws and regulations. (Paragraph 145)
35. Sectoral regulators should publish detailed guidance to help deployers of AI strike the balance between the protection of privacy and securing the technology’s intended benefits. In instances where regulators determine that this balance has not been met, or where the relevant laws or regulatory requirements have not been met, it should impose sanctions or prohibit the use of AI models or tools. (Paragraph 146)
36. We welcome the Government amendment to the Criminal Justice Bill as a necessary step towards ensuring the UK’s legal framework reflects the current state of technological development and protects citizens, primarily women and girls, from the consequences of AI-assisted misrepresentation, including deepfake pornography. (Paragraph 150)
37. Should the Bill’s remaining stages fail to be completed prior to the dissolution of Parliament, the next Government must introduce similar provisions as soon as is practicable after the General Election. (Paragraph 150)
38. The Government and regulatory authorities, informed by the work of the Defending Democracy Taskforce, should safeguard the integrity of the upcoming General Election campaign in its approach to the online platforms that host deepfake content which seeks to exert a malign influence on the democratic process. If these platforms are found to have been slow to remove such content, or to have facilitated its spread, regulators must take stringent enforcement action—including holding senior leadership personally liable and imposing financial sanctions. (Paragraph 154)
39. A cross-Government public awareness campaign should be launched to inform the public about the growing prevalence of AI-assisted misrepresentation, the potential consequences, what the Government is doing to address the Challenge, and what steps individuals can take to protect themselves online. (Paragraph 155)
40. At the so-called ‘frontier’ of AI a small group of leading developers are responsible for and accruing significant benefits from the development of advanced models and tools—thanks in part to their ability to access the necessary training data. This potential dominance is arguably to the detriment of free and open competition. (Paragraph 159)
41. As the regulator responsible for promoting competitive markets and tackling anti-competitive behaviour, the CMA should identify abuses of market power and use its powers to stop them. This could take the form of levying fines or requiring the restructuring of proposed mergers. (Paragraph 160)
42. AI models and tools rely on access to high-quality input data. The phrase ‘garbage in, garbage out’ is not new, but it is particularly applicable to AI. (Paragraph 161)
43. The potential for human error and bias notwithstanding, deployers should not solely rely on outputs produced with AI tools to determine their decision-making, particularly in areas that could affect the rights and standing of the individuals or entities concerned, such as insurance decisions or recruitment. These algorithmic decisions should always be reviewed and verified by trained humans, and those affected should have the right to challenge these decisions—a process that should also be human-centred. (Paragraph 161)
44. The Government and future administrations should support the emergence of more AI startups in the UK by ensuring they can access the high-quality datasets they need to innovate. This could involve facilitating access to anonymised public data from data.gov.uk, the NHS and BBC via a National Data Bank, subject to appropriate safeguards. (Paragraph 162)
45. We welcome the Government’s moves to establish a dedicated AI Research Resource and a cluster of supercomputers but are concerned that it has yet to set out further details of how researchers and startups will be able to access the compute they need to maximise the potential benefits of AI across society and the economy. (Paragraph 165)
46. The Government, or its successor administration, should publish an action plan and proposed deliverables for both the AI Research Resource and its cluster of supercomputers, and further details of the terms under which researchers and innovative startups will be able to access them. It should also undertake a feasibility study into the establishment of a National Compute Cluster that could be made available to researchers and startups. (Paragraph 166)
47. The Black Box Challenge is one of the most paradigm-shifting consequences of AI, as it upends our well-established reliance on explainability and understanding. Given the complexity of currently available and in all likelihood future models, the starting point should be an acknowledgement how little we can understand about how many AI models produce their outputs, an acceptance that new ways of thinking will be required, and a regulatory approach that accounts for the impossibility of total explainability. (Paragraph 170)
48. The regulators charged with implementing the Government’s high-level AI governance principles should, in their approach to these models and tools, prioritise testing and verifying their outputs, as well seeking to establish—whilst accepting the difficulty of doing so with absolute certainty—how they arrived at them. (Paragraph 171)
49. The open-source approach has underpinned many technological breakthroughs, including the Internet and AI. Whilst some providers of products and services, such as AI models and their applications, will want to keep elements of their offerings proprietary, a healthy AI marketplace should be sufficiently diverse to support both ‘open’ and ‘closed’ options. The volume of investment flowing into AI developers of all types of models, rather than one or the other, is evidence of this market diversity. (Paragraph 175)
50. When procuring AI models for deployment in the public sector the Government and public bodies should utilise those best suited to the task. (Paragraph 176)
51. The Government should in its response to this Report tell us how it will ensure law enforcement and regulators are adequately resourced to respond to the growing use of AI models and tools to generate and disseminate harmful and illegal content. (Paragraph 180)
52. The growing volume of litigation relating to alleged use of works protected by copyright to train AI models and tools, and the value of high-quality data needed to train future models, has underlined the need for a sustainable framework that acknowledges the inevitable trade-offs and establishes clear, enforceable rules of the road. The status quo allows developers to potentially benefit from the unlimited, free use of copyrighted material, whilst negotiations are stalled. (Paragraph 185)
53. The current Government, or its successor administration, should ensure that discussions regarding the use of copyrighted works to train and run AI models are concluded and an implementable approach agreed. It seems inevitable that this will involve the agreement of a financial settlement for past infringements by AI developers, the negotiation of a licensing framework to govern future uses, and in all likelihood the establishment of a new authority to operationalise the agreement. If this cannot be achieved through a voluntary approach, it should be enforced by the Government, or its successor administration, in co-operation with its international partners. (Paragraph 186)
54. Nobody who uses AI to inflict harm should be exempted from the consequences, whether they are a developer, deployer, or intermediary. The next Government together with sectoral regulators publish guidance on where liability for harmful uses of AI falls under existing law. This should be a cross-Government undertaking. Sectoral regulators should ensure that guidance on liability for AI-related harms is made available to developers and deployers as and when it is required. Future administrations and regulators should also, where appropriate, establish liability via statute rather than simply relying on jurisprudence. (Paragraph 189)
55. AI is already changing the nature of work, and as the technology evolves this process is likely to accelerate, placing some jobs at risk. At the same time, there are productivity benefits to be won, provided people are equipped with the skills to fruitfully utilise AI. This is a process that should begin in the classroom, and through greater prioritisation of initiatives such as the Lifetime Skills Guarantee and digital Skills Bootcamps. (Paragraph 193)
56. The current Government, or its successor, should commission a review into the possible future skills and employment consequences of AI, along the lines of the 2017 Taylor Review of modern working practices which examined the landscape, suggested ideas for debate and has resulted in legislative change. It should also in its response to this Report tell us how it will ensure workers whose jobs are at risk of automation will be able to retrain and acquire the skills necessary to change careers. (Paragraph 194)
57. We welcome the organisation of the AI Safety Summit at Bletchley Park and commend the Government on bringing many key actors together. We look forward to subsequent Summits and hope that the consensus and momentum delivered at Bletchley Park can be maintained. (Paragraph 202)
58. However, looking beyond the AI safety discussion, we do not believe that harmonisation for harmonisation’s sake should be the end goal of international AI governance discussions. A degree of distinction between different regulatory regimes is, in our view, inevitable. Such distinction may be motivated by geopolitics, but it may also simply be a case of healthy economic competition. (Paragraph 203)
59. Future AI Safety Summits must focus on the establishment of international dialogue mechanisms to address current, medium- and longer-term safety risks presented by the growing use of AI; and the sharing of best practice to ensure its potential benefits are realised in all jurisdictions. This should not set us on the road to a global AI governance regime—we are unconvinced that such a prospect is either realistic or desirable. (Paragraph 204)
60. The debate over the existential risk—or lack of it—posed by the increasing prevalence of AI has attracted significant attention. However, the Government’s initial assessment, that such existential risks are high impact but low probability, appears to be accurate. Nevertheless, given the potential consequences should risks highlighted by the AI Safety Institute and other researchers be realised, it is right for Governments to continue to engage with experts on the issue. (Paragraph 209)
61. When implementing the principles set out in the AI White Paper regulatory activity should be focused on here-and-now impacts. Assessing and responding to existential risk should primarily be the responsibility of the UK’s national security apparatus, supported by the AI Safety Institute. (Paragraph 210)
62. Should the acuteness of existential AI risk be judged to have increased, discussions regarding the implications and possible response should take place in international fora, such as AI Safety Summits. (Paragraph 211)
Stephen Metcalfe, in the Chair
Chris Clarkson
Dame Tracey Crouch
James Davies
Katherine Fletcher
Draft Report (Governance of artificial intelligence), proposed by the Chair, brought up and read.
Ordered, That the draft Report be read a second time, paragraph by paragraph.
Paragraphs 1 to 211 read and agreed to.
Summary agreed to.
Resolved, That the Report be the Third Report of the Committee to the House.
Ordered, That the Chair make the Report to the House.
Ordered, That embargoed copies of the Report be made available, in accordance with the provisions of Standing Order No. 134.
The Committee adjourned.
The following witnesses gave evidence. Transcripts can be viewed on the inquiry publications page of the Committee’s website.
Professor Michael Osborne, Professor of Machine Learning and co-founder, University of Oxford and Mind Foundry; Michael Cohen, DPhil candidate in Engineering Science, University of OxfordQ1–54
Mrs Katherine Holden, Head of Data Analytics, AI and Digital Identity, techUK; Dr Manish Patel, CEO, Jiva.aiQ55–96
Adrian Joseph, Chief Data and AI Officer, BT Group; Jen Gennai, Director, Responsible Innovation, Google; Hugh Milward, General Manager, Corporate, External and Legal Affairs, Microsoft UKQ97–143
Professor Dame Wendy Hall, Regius Professor of Computer Science, University of Southampton; Professor Sir Nigel Shadbolt, Professorial Research Fellow in Computer Science and Principal, Jesus College, University of OxfordQ144–173
Professor Andrew Hopkins, Chief Executive, ExscientiaQ174–222
Professor Delmiro Fernandez-Reyes, Professor of Biomedical Computing, University College London, Adjunct Professor of Paediatrics, University of Ibadjan; Professor Mihaela van der Schaar, John Humphrey Plummer Professor of Machine Learning, Artificial Intelligence and Medicine, and Director, Cambridge Centre for AI in Medicine, Cambridge UniversityQ223–262
Professor Rose Luckin, Professor of Learner Centred Design, University College London, Director, Educate; Daisy Christodoulou, Director of Education, No More MarkingQ263–294
Dr Matthew Glanville, Head of Assessment Principles and Practice, The International Baccalaureate; Joel Kenyon, Science Teacher and Community Cohesion Lead, Dormers Wells High School, Southall, LondonQ295–326
Jamie Njoku-Goodwin, CEO, UK Music; Paul Fleming, General Secretary, EquityQ327–373
Coran Darling, Associate, Intellectual Property and Technology, DLA Piper; Dr Hayleigh Bosher, Senior Lecturer in Intellectual Property Law, Brunel UniversityQ374–411
Lindsey Chiswick, Director of Intelligence, Metropolitan Police; Dr Tony Mansfield, Principal Research Scientist, National Physical LaboratoryQ412–506
Michael Birtwistle, Associate Director, AI and data law & policy, Ada Lovelace Institute; Dr Marion Oswald, Senior Research Associate for Safe and Ethical AI and Associate Professor in Law, The Alan Turing Institute and Northumbria UniversityQ507–538
Dame Melanie Dawes, Chief Executive, Ofcom; Will Hayter, Senior Director, Digital Markets Unit, Competition and Markets Authority (CMA)Q539–601
John Edwards, Information Commissioner, Information Commissioner’s Office; Kate Jones, Chief Executive, Digital Regulation Cooperation Forum; Jessica Rusu, Chief Data, Information and Intelligence Officer, Financial Conduct AuthorityQ602–652
Matt Clifford, Prime Minister’s representative, AI Safety Summit; Emran Mian, Director General, Digital Technologies and Telecoms, Department for Science, Innovation and TechnologyQ653–755
Rt Hon Michelle Donelan MP, Secretary of State, Department for Science, Innovation and Technology; Sarah Munby, Permanent Secretary, Department for Science, Innovation and TechnologyQ756–822
The following written evidence was received and can be viewed on the inquiry publications page of the Committee’s website.
GAI numbers are generated by the evidence processing system and so may not be complete.
1 ACT | The App Association (GAI0018)
2 ADS (GAI0027)
3 AI & Digital Healthcare Group, Centre for Regulatory Science and Innovation, Birmingham (University Hospitals Birmingham NHS Foundation Trust/University of Birmingham) (GAI0055)
4 AI Centre (GAI0037)
5 AI Governance Limited (GAI0050)
6 Abrusci, Dr Elena (Lecturer , Brunel University London); and Scott, Dr Richard Mackenzie-Gray (Postdoctoral Fellow, University of Oxford) (GAI0038)
7 Academy of Medical Sciences (GAI0072)
8 Ada Lovelace Institute (GAI0086)
9 Alfieri, Joseph (GAI0062)
10 Alliance for Intellectual Property (GAI0118)
11 Assuring Autonomy International Programme (AAIP), University of York.; McDermid, Professor John ; Calinescu, Professor Radu ; MacIntosh, Dr Ana ; Habli, Professor Ibrahim ; and Hawkins, Dr Richard (GAI0044)
12 BCS - Chartered Institute for Information Technology (GAI0022)
13 BILETA (GAI0082)
14 BT Group (GAI0091)
15 Belfield, Mr Haydn (Academic Project Manager, University of Cambridge, Leverhulme Centre for the Future of Intelligence & Centre for the Study of Existential Risk); igeartaigh, Dr Seán Ó hÉ (Acting Director and Principal Researcher, University of Cambridge, Centre for the Study of Existential Risk & Leverhulme Centre for the Future of Intelligence); Avin, Dr Shahar (Senior Research Associate, University of Cambridge, Centre for the Study of Existential Risk); ndez-Orallo, Prof José Herná (Professor, Universitat Politècnica de València); and Corsi, Giulio (Research Associate, University of Cambridge, Leverhulme Centre for the Future of Intelligence)) (GAI0094)
16 Big Brother Watch (GAI0088)
17 Bosher, Dr Hayley (Reader in Intellectual Property Law and Associate Dean, Brunel University) (GAI0128)
18 British Standards Institution (BSI) (GAI0028)
19 Burges Salmon LLP (GAI0064)
20 CBI (GAI0115)
21 CENTRIC (GAI0043)
22 Carnegie UK (GAI0041)
23 Center for AI and Digital Policy (GAI0098)
24 Chiswick, Lindsey (Director of Intelligence, Metropolitan Police) (GAI0121)
25 Clement-Jones, Lord (Digital Spokesperson for the Liberal Democrats, House of Lords); and Darling, Coran (GAI0101)
26 Cohen, Michael (DPhil Candidate, University of Oxford); and Osborne, Professor Michael (Professor of Machine Learning, University of Oxford) (GAI0046, GAI0116)
27 Collins, Dr Philippa (Senior Lecturer in Law, University of Bristol); and Atkinson, Dr Joe (Lecturer in Law, University of Sheffield) (GAI0074)
28 Committee on Standards in Public Life (GAI0110)
29 Competition and Markets Authority (GAI0124)
30 Compliant & Accountable Systems Research Group, Department of Computer Science & Technology, University of Cambridge; and Compliant & Accountable Systems Research Group, Department of Computer Science & Technology, University of Cambridge (GAI0106)
31 Connected by Data (GAI0052)
32 Copyright Alliance (GAI0097)
33 Creative Commons (GAI0015)
34 Crockett, Professor of Computational Intelligence Keeley (Professor of Computational Intelligence, Manchester Metropolitan University) (GAI0020)
35 DeepMind (GAI0100)
36 Dennis, Bobbie (Policy and Public Affairs Officer, Internet Watch Foundation (IWF)) (GAI0130)
37 Department for Digital, Culture, Media and Sport; and Department for Business, Energy and Industrial Strategy (GAI0107)
38 Edwards, Professor Rosalind (Professor of Sociology , University of Southampton); Gillies, Professor Val (Professor of Social Policy and Criminology , University of Westminster); Gorin, Dr. Sarah (Assistant Professor , University of Warwick); and Ducasse, Dr. Hélène Vannier (Senior Research Fellow , University of Southampton) (GAI0035)
39 Electoral Commission (GAI0129)
40 Employers Lawyers Association (GAI0031)
41 Equity (GAI0065)
42 Financial Conduct Authority (GAI0125)
43 Fotheringham, Kit (Postgraduate Researcher, University of Bristol) (GAI0042)
44 GSK (GAI0067)
45 Google (GAI0099)
46 Hopgood, Professor Adrian (Professor of Intelligent Systems, University of Portsmouth) (GAI0030)
47 Imperial College London Artificial Intelligence Network (GAI0014)
48 Information Commissioner’s Office (ICO) (GAI0112)
49 Institute for the Future of Work (GAI0063)
50 Institute of Physics and Engineering in Medicine (IPEM) (GAI0051)
51 Joshi, Ruchika (Harvard Kennedy School) (GAI133)
52 Leslie, Professor David (Director of Ethics and Responsible Innovation Research, The Alan Turing Institute; and Professor of Ethics, Technology and Society, Queen Mary University of London) (GAI0113)
53 Liberty (GAI0081)
54 Library and Archives Copyright Alliance (GAI0120)
55 Loughborough University (GAI0070)
56 Mason, Mr Shane (Freelance consultant, n/a) (GAI0006)
57 Microsoft (GAI0083)
58 Milestone Systems (GAI0132)
59 Minderoo Centre for Technology and Democracy, University of Cambridge (GAI0032)
60 NCC Group (GAI0040)
61 NICE; HRA; MHRA; and CQC (GAI0076)
62 National Physical Laboratory (GAI0053)
63 Ofcom (GAI0126)
64 Oswald, Dr Marion (GAI0012)
65 Oxford Internet Institute (GAI0058)
66 Oxford Internet Institute, University of Oxford; University of Exeter; Oxford Internet Institute, University of Oxford; Oxford Internet Institute, University of Oxford; and Oxford Internet Institute, University of Oxford (GAI0024)
67 Patelli, Dr Alina (Senior Lecturer in Computer Science, Aston University) (GAI0095)
68 Protect Pure Maths (GAI0117)
69 Public Law Project (GAI0069)
70 Publishers Association (GAI0102)
71 Pupils 2 Parliament (GAI0096)
72 Queen Mary University London (GAI0073)
73 RELX (GAI0033)
74 Reed, Professor Chris (Professor of Electronic Commerce Law, Centre for Commercial Law Studies, Queen Mary University of London) (GAI0059)
75 Richie, Dr Cristina (lecturer, TU Delft) (GAI0001, GAI0002)
76 Rolf, Dr Steve (Research Fellow, The Digital Futures at Work (Digit) Centre, University of Sussex Business School) (GAI0104)
77 Rolls-Royce plc (GAI0109)
78 Sage Group (GAI0108)
79 Salesforce (GAI0105)
80 Sanchez-Graells, Professor Albert (Professor of Economic Law, University of Bristol Law School) (GAI0004)
81 School of Informatics, University of Edinburgh (GAI0079)
82 Scott, Mr. Michael (Chair of Trustees, Home-Start Nottingham) (GAI0005)
83 Sense about Science (GAI0078)
84 TUC (GAI0060)
85 Tang, Dr Guan H (Senior Lecturer, Centre for Commercial Law Studies, Queen Mary University of London) (GAI0077)
86 TechWorks (GAI0068)
87 Tessler, Leonardo (PhD in law Candidate, University of Montreal) (GAI0092)
88 The Alliance for Intellectual Property (GAI0103)
89 The Institution of Engineering and Technology (IET) (GAI0021)
90 The LSE Law School, London School of Economics.; and The LSE Law School, London School of Economics. (GAI0036)
91 The Nutrition Society (GAI0007)
92 The Royal Academy of Engineering (GAI0039)
93 The Royal College of Radiologists (RCR) (GAI0087, GAI0131)
94 Thorney Isle Research (GAI0016)
95 Tripathi, Mr Karan (Research Associate , University of Sheffield); and Tzanou , Dr Maria (Senior Lecturer in Law, University of Sheffield) (GAI0047)
96 Trustpilot (GAI0054)
97 Trustworthy Autonomous Systems Hub; The UKRI TAS Node in Governance & Regulation; and The UKRI TAS Node in Functionality (GAI0084)
98 UK BioIndustry Association (GAI0026)
99 UK Dementia Research Institute (GAI0111)
100 UKRI (GAI0114)
101 United Nations Association UK; Article 36; Women’s International League for Peace and Freedom UK; and Drone Wars UK (GAI0090)
102 University of Glasgow (GAI0057)
103 University of Sheffield (GAI0017)
104 University of Surrey (GAI0075)
105 Wayve (GAI0061)
106 Which? (GAI0049)
107 Whittlestone, Dr Jess (Head of AI Policy, Centre for Long-Term Resilience); and Moulange, Richard (PhD student , MRC Biostatistics Unit, University of Cambridge) (GAI0071)
108 Wudel, Alexandra (Political Advisor, German Parliament); Gengler, Eva (PhD Student, FAU Nürnberg); and Center for Feminist Artificial Intelligence (GAI0013)
109 Wysa Limited (GAI0093)
110 medConfidential (GAI0011)
111 techUK (GAI0045)
112 Zurich UK (GAI0127)
All publications from the Committee are available on the publications page of the Committee’s website.
Number |
Title |
Reference |
1st |
The antimicrobial potential of bacteriophages |
HC 328 |
2nd |
Insect decline and UK food security |
HC 326 |
1st Special |
The governance of artificial intelligence: interim report: Government response to the Committee’s Ninth report of Session 2022–23 |
HC 248 |
Number |
Title |
Reference |
1st |
Pre-appointment hearing for the Executive Chair of Research England |
HC 636 |
2nd |
UK space strategy and UK satellite infrastructure |
HC 100 |
3rd |
My Science Inquiry |
HC 618 |
4th |
The role of Hydrogen in achieving Net Zero |
HC 99 |
5th |
Diversity and Inclusion in STEM |
HC 95 |
6th |
Reproducibility and Research Integrity |
HC 101 |
7th |
UK space strategy and UK satellite infrastructure: reviewing the licencing regime for launch |
HC 1717 |
8th |
Delivering nuclear power |
HC 626 |
9th |
The governance of artificial intelligence: interim report |
HC 1769 |
Number |
Title |
Reference |
1st |
Direct-to-consumer genomic testing |
HC 94 |
2nd |
Pre-appointment hearing for the Chair of UK Research and Innovation |
HC 358 |
3rd |
Coronavirus: lessons learned to date |
HC 92 |
Number |
Title |
Reference |
1st |
The UK response to covid-19: use of scientific advice |
HC 136 |
2nd |
5G market diversification and wider lessons for critical and emerging technologies |
HC 450 |
3rd |
A new UK research funding agency |
HC 778 |
1 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, summary
2 A pro-innovation approach to AI regulation, CP 815, Department for Science, Innovation and Technology, 29 March 2023
3 Artificial intelligence act: Council and Parliament strike a deal on the first rules for AI in the world, Council of the European Union, 9 December 2023
4 Executive Order on the Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence, The White House, 30 October 2023
5 Frontier AI Taskforce: second progress report, GOV.UK, 30 October 2023
6 G7 nations to harness AI and innovation to drive growth and productivity, GOV.UK, 15 March 2024
7 Interim Report: Governing AI for Humanity, United Nations AI Advisory Body, 22 December 2023
8 Countries agree to safe and responsible development of frontier AI in landmark Bletchley Declaration, GOV.UK, 1 November 2023
9 MPs to examine regulating AI in new inquiry, UK Parliament, 20 October 2022
10 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769
11 Science, Innovation and Technology Committee, First Special Report of Session 2023–24, The governance of artificial intelligence: interim report: Government response to the Committee’s Ninth report, HC 248
12 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, Chapter 3
13 AI poised to begin shifting from ‘excitement’ to ‘deployment’ in 2024, Goldman Sachs, 17 November 2023
15 State of AI Report, Air Street Capital, 12 October 2023
16 State of the sector: annual review of UK financial services 2023, City of London Corporation and HM Treasury, 4 July 2023
17 Collaborate to compete: why we must all embrace a growth mindset, Financial Conduct Authority, 18 October 2023
19 Is artificial intelligence the solution to cyber security threats?, Financial Times, 16 January 2024
20 Financial Conduct Authority (GAI0125)
21 Financial Conduct Authority (GAI0125)
23 The near-term impact of AI on the cyber threat, National Cyber Security Centre, 24 January 2024
24 How much electricity does AI consume? The Verge, 16 February 2024;
25 How much water does AI consume? The public deserves to know, OECD.AI, 30 November 2023
26 Microsoft 2024 Environmental Sustainability Report, Microsoft, 15 May 2024, p. 5
27 Microsoft’s emissions jump almost 30% as it races to meet AI demand, Financial Times, 15 May 2024; Net-zero carbon, Google, accessed 23 May 2024
28 Sam Altman says the future of AI depends on breakthroughs in clean energy, The Verge, 19 January 2024
29 PM speech at London Tech Week, GOV.UK, 12 June 2023
30 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 3
31 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 17
32 Frontier AI Taskforce: second progress report, Department for Science, Innovation and Technology, 30 October 2023
33 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, summary
34 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024
35 A pro-innovation approach to AI regulation, CP 815, Department for Science, Innovation and Technology, 29 March 2023
36 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, Chapter 5
37 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 13
38 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 7
39 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 13
40 A pro-innovation approach to AI regulation, CP 815, Department for Science, Innovation and Technology, 29 March 2023, p. 6
41 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, para 105
42 A pro-innovation approach to AI regulation, CP 815, Department for Science, Innovation and Technology, 29 March 2023, p. 6
43 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, para 106
44 Science, Innovation and Technology Committee, First Special Report of Session 2023–24, The governance of artificial intelligence: interim report: Government response to the Committee’s Ninth report, HC 248, p. 8
46 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 14
47 Leading frontier AI companies publish safety policies, GOV.UK, 27 October 2023
48 World leaders, top AI companies set out plan for safety testing of frontier as first global AI Safety Summit concludes, GOV.UK, 2 November 2023
49 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 33
50 UK rethinks AI legislation as alarm grows over potential risks, Financial Times, 15 April 2024
51 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 13
52 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, para 97
53 A pro-innovation approach to AI regulation, CP 815, Department for Science, Innovation and Technology, 29 March 2023, p. 15
55 Competition and Markets Authority (GAI0124), Financial Conduct Authority (GAI0125), Information Commissioner’s Office (GAI0112), Ofcom (GAI0126)
56 Navigating the UK’s Digital Regulation Landscape: Where are we headed? Financial Conduct Authority, 22 April 2024
58 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, para 104
59 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 15
60 A pro-innovation approach to AI regulation, CP 815, Department for Science, Innovation and Technology, 29 March 2023, p. 42
61 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, para 103
63 Artificial Intelligence and Digital Regulations Service launches, NHS Health Research Authority, 7 March 2023
64 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 15
65 Implementing the UK’s AI Regulatory Principles: Initial Guidance for Regulators, Department for Science, Innovation and Technology, 6 February 2024
66 About the FCA, Financial Conduct Authority, accessed 23 May 2024
67 How is Ofcom funded? Ofcom, 25 April 2024
68 How we are funded, Information Commissioner’s Office, accessed 23 May 2024
69 Funding, European Medicines Agency, accessed 23 May 2024
70 About us, European Union Intellectual Property Office, accessed 23 May 2024
71 Congressional Budget Justification Fiscal Year 2025, Federal Trade Commission, accessed 23 May 2024; FY 2025 FDA Budget Summary, Food and Drug Administration, accessed 23 May 2024
72 House of Lords Industry and Regulators Committee, First Report of Session 2023–24, Who watches the watchdogs? Improving the performance, independence and accountability of UK regulators, HL Paper 56, para 134
73 A pro-innovation approach to AI regulation, CP 815, Department for Science, Innovation and Technology, 29 March 2023, p. 62
75 Committee on Standards in Public Life (GAI0110)
76 Artificial intelligence and public standards: an update on progress made against our 2020 recommendations, Committee on Standards in Public Life, 6 March 2024
78 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 14
79 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 7
80 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 55
81 Regulators’ strategic approaches to AI, Department for Science, Innovation and Technology, 1 May 2024
82 Jobs boost at Microsoft as UK revenues hit £8bn, The Times, 27 April 2024
83 Use of artificial intelligence in government, National Audit Office, HC 612, p. 18
84 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 21
85 Deputy Prime Minister speech on AI for Public Good, GOV.UK, 29 February 2024
86 Use of artificial intelligence in government, National Audit Office, HC 612, p. 17
87 “Criminals should be aware” says Minister as Government upgrades AI fraud detection tool, GOV.UK, 14 March 2024
88 Hunt announces ‘most ambitious public sector productivity review ever’, Civil Service World, 13 June 2023
89 Chancellor to cut admin workloads to free up frontline staff, GOV.UK, 18 November 2023
90 Spring Budget 2024, HC 560, HM Treasury, 6 March 2024, p. 31
91 Spring Budget 2024, HC 560, HM Treasury, 6 March 2024, pp. 31–35
92 Seizing the opportunity: delivering efficiency for the public, HM Treasury, 6 March 2024, p. 32
93 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 17
94 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 22
95 Generative AI Framework for HMG, Cabinet Office and Central Digital & Data Office, 18 January 2024
96 The findings of our first generative AI experiment: GOV.UK Chat, Inside GOV.UK, 18 January 2024
97 Government to harness the power of AI to improve public project delivery under new framework, GOV.UK, 20 March 2024
98 Algorithmic Transparency Recording Standard Hub, Central Digital and Data Office and Department for Science, Innovation and Technology, accessed 23 May 2024
99 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, pp. 22, 40
100 Business and tech heavyweights to boost productivity through AI, GOV.UK, 25 January 2024
101 AI Opportunity Forum holds first meeting, GOV.UK, 15 February 2024
102 AI Opportunity Forum holds penultimate meeting, GOV.UK, 9 May 2024
103 Spring Budget 2024, HC 560, HM Treasury, 6 March 2024, p. 61
104 Flexible AI Upskilling Fund pilot: open for applications, Department for Science, Innovation and Technology, accessed 23 May 2024
105 Introduction to AI Assurance, Department for Science, Innovation and Technology, 12 February 2024, p. 15
106 AI and Digital Hub, Digital Regulation Cooperation Forum, accessed 23 May 2024
107 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, para 108
108 Introducing the AI Safety Institute, CP 960, Department for Science, Innovation and Technology, 2 November 2023. The Institute will receive a continuation of the Taskforce’s 2024–25 funding as an annual amount for the remainder of this decade.
109 Frontier AI Taskforce: first progress report, Department for Science, Innovation and Technology, 7 September 2023
110 Red-teaming is a military term, described by the Ministry of Defence as intended to challenge existing thinking by seeking “… an external viewpoint separate to that of ‘home team’ decision-makers and problem solvers”. In AI model development, red-teaming processes subject models to technical attacks to identify weaknesses and create a more robust product. Google’s Red Team “… consists of a team of hackers that simulate a variety of adversaries, ranging from nation states and well-known Advanced Persistent Threat groups to hacktivists, individual criminals or even malicious insiders”.
111 Frontier AI Taskforce: first progress report, Department for Science, Innovation and Technology, 7 September 2023
112 Frontier AI Taskforce: first progress report, Department for Science, Innovation and Technology, 7 September 2023; Frontier AI Taskforce: second progress report, 30 October 2023; AI Safety Institute: third progress report, Department for Science, Innovation and Technology, 5 February 2024
113 PM London Tech Week speech, GOV.UK, 12 June 2023
114 Frontier AI Taskforce: second progress report, 30 October 2023
115 PM London Tech Week speech, GOV.UK, 12 June 2023
116 AI Safety Summit: introduction, Department for Science, Innovation and Technology, 31 October 2023
117 AI Safety Summit: introduction, Department for Science, Innovation and Technology, 31 October 2023
118 Prime Minister’s speech on AI: 26 October 2023, GOV.UK, 26 October 2023
119 Leading frontier AI companies publish safety policies, GOV.UK, 27 October 2023
120 Capabilities and risks from frontier AI: A discussion paper on the need for further research into AI risk, Department for Science, Innovation and Technology, 25 October 2023; Future Risks of Frontier AI: Which capabilities and risks could emerge at the cutting edge of AI in the future?, Government Office for Science, 27 October 2023; Emerging processes for frontier AI safety, Department for Science, Innovation and Technology, 27 October 2023; Safety and Security Risks of Generative Artificial Intelligence to 2025, HM Government, 27 October 2023
121 AI Safety Summit 2023: Roundtable Chairs’ Summaries, 1 November, Department for Science, Innovation and Technology, 1 November; AI Safety Summit 2023: Roundtable Chairs’ Summaries, 2 November, Department for Science, Innovation and Technology, 3 November 2023
122 The Bletchley Declaration by Countries Attending the AI Safety Summit, 1–2 November 2023, Department for Science, Innovation and Technology, 1 November 2023
123 ‘State of the Science’ Report to Understand Capabilities and Risks of Frontier AI: Statement by the Chair, Department for Science, Innovation and Technology, 2 November 2023
124 The developers were Amazon Web Services, Anthropic, Google, Google DeepMind, Inflection AI, Meta, Microsoft, Mistral AI, Open AI and xAI
125 Safety Testing: Chair’s Statement of Session Outcomes, Department for Science, Innovation and Technology, 2 November 2023
126 World leaders, top AI companies set out plan for safety testing of frontier as first global AI Safety Summit concludes, GOV.UK, 2 November 2023; UK and Republic of Korea to build on legacy of Bletchley Park, GOV.UK, 12 April 2024; Qq. 658–659
127 Frontier AI Safety Commitments: AI Seoul Summit 2024, Department for Science, Innovation and Technology, 21 May 2024
128 Frontier AI Safety Commitments: AI Seoul Summit 2024, Department for Science, Innovation and Technology, 21 May 2024
129 Introducing the AI Safety Institute, CP 960, Department for Science, Innovation and Technology, p. 8
131 Introducing the AI Safety Institute, CP 960, Department for Science, Innovation and Technology, p. 9
132 AI Safety Institute: third progress report, Department for Science, Innovation and Technology, 5 February 2024
133 AI Safety Institute approach to evaluations, Department for Science, Innovation and Technology, 9 February 2024
134 Advanced AI evaluations at AISI: May update, AI Safety Institute, 20 May 2024
135 UK & United States announce partnership on science of AI safety, GOV.UK, 2 April 2024
136 HC Deb, 17 April 2024, col 288 (Commons Chamber)
137 Rishi Sunak struggles to implement his ‘landmark’ AI testing deal, Politico, 19 April 2024
138 Meet the generative AI startups pulling in the most cash, PitchBook, 18 October 2023
139 The Global AI Index, Tortoise, accessed 23 May 2024
140 Live data: top countries in VC investments in AI by industry, OECD.AI visualisations powered by JSI using data from Preqin, accessed 23 May 2024
141 Big Tech outspends venture capital firms in AI investment frenzy, Financial Times, 29 December 2023
142 How Microsoft’s multibillion-dollar alliance with OpenAI really works, Financial Times, 15 December 2023
143 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 7
144 FACT SHEET: Biden-Harris Administration Secures Voluntary Commitments from Leading Artificial Intelligence Companies to Manage the Risks Posed by AI, The White House, 21 July 2023
145 FACT SHEET: Biden-Harris Administration Secures Voluntary Commitments from Eight Additional Artificial Intelligence Companies to Manage the Risks Posed by AI, The White House, 12 September 2023
146 Voluntary AI Commitments, The White House, 12 September 2023, p. 1
147 Voluntary AI Commitments, The White House, 12 September 2023, pp. 1–3
148 Voluntary AI Commitments, The White House, 12 September 2023, p. 1
149 Executive Order on the Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence, The White House, 30 October 2023
150 What Is the Defense Production Act? Council on Foreign Relations, 22 December 2021
151 FACT SHEET: President Biden Issues Executive Order on Safe, Secure, and Trustworthy Artificial Intelligence, The White House, 30 October 2023
152 Executive Order on the Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence, The White House, 30 October 2023
153 FACT SHEET: Biden-Harris Administration Secures Voluntary Commitments from Eight Additional Artificial Intelligence Companies to Manage the Risks Posed by AI, The White House, 12 September 2023
154 Capitol Hill stunner: 2023 led to fewest laws in decades, Axios, 18 December 2023
155 California’s privacy watchdog eyes AI rules with opt-out and access rights, TechCrunch, 27 November 2023
156 The Colorado AI Act: What you need to know, International Association of Privacy Professionals, 21 May 2024
157 What States are Making Moves in US AI Regulation in 2024? Holistic AI, 11 January 2024
158 What States are Making Moves in US AI Regulation in 2024? Holistic AI, 11 January 2024
159 State AI Legislation Off to Quick Start in 2024, LexisNexis, 16 January 2024
160 California’s privacy watchdog eyes AI rules with opt-out and access rights, TechCrunch, 27 November 2023
161 Commission welcomes political agreement on Artificial Intelligence Act, European Commission, 9 December 2023
162 EU countries give crucial nod to first-of-a-kind Artificial Intelligence law, Euractiv, 2 February 2024
163 Artificial Intelligence Act: MEPs adopt landmark law, European Parliament, 13 March 2024
164 Regulation of the European Parliament and the Council laying down harmonised rules on artificial intelligence (Artificial Intelligence Act) and amending certain Union legislative acts, European Commission, 21 April 2021
165 AI in the EU and UK: two approaches to regulation and international leadership, UK in a changing Europe, 26 January 2023
166 Recommendation System, NVIDIA, accessed 23 May 2024
167 Artificial intelligence: questions and answers, European Commission, 12 December 2023
168 Proposal for a Regulation of the European Parliament and of the Council laying down harmonised rules on artificial intelligence (Artificial Intelligence Act) and amending certain Union legislative acts, Annex III, Council of the European Union, p. 248
169 Artificial intelligence: questions and answers, European Commission, 12 December 2023
170 Artificial intelligence: questions and answers, European Commission, 12 December 2023
171 Commission welcomes political agreement on Artificial Intelligence Act, European Commission, 9 December 2023
172 Artificial intelligence: questions and answers, European Commission, 12 December 2023
173 AI in policing and security, Parliamentary Office of Science and Technology, 29 April 2021
174 Artificial intelligence: questions and answers, European Commission, 12 December 2023
175 Artificial intelligence: questions and answers, European Commission, 12 December 2023
176 GPT-4, OpenAI, accessed 23 May 2024
177 Gemini, Google, accessed 23 May 2024
178 Commission welcomes political agreement on Artificial Intelligence Act, European Commission, 9 December 2023
179 Artificial intelligence: questions and answers, European Commission, 12 December 2023
180 Commission welcomes political agreement on Artificial Intelligence Act, European Commission, 9 December 2023
181 Artificial intelligence: questions and answers, European Commission, 12 December 2023
182 Commission presses governments to appoint AI regulators, Euronews, 3 April 2024
183 Commission presses governments to appoint AI regulators, Euronews, 3 April 2024
184 Artificial intelligence: questions and answers, European Commission, 12 December 2023
185 As the EU AI Act enters into force, focus shifts to countries’ oversight appointments, Euronews, 12 March 2024
186 Commission Decision Establishing the European AI Office, European Commission, 24 January 2024
187 Will the EU AI Act work? Lessons learned from past legislative initiatives, future challenges, International Association of Privacy Professionals, 17 April 2024
188 Commission to look for head of AI Office only when law is fully approved, Euronews, 20 March 2024
189 As the EU AI Act enters into force, focus shifts to countries’ oversight appointments, Euronews, 12 March 2024
190 European regulators discuss AI Act enforcement, fines, International Association of Privacy Professionals, 10 April 2024
191 Artificial intelligence: questions and answers, European Commission, 12 December 2023
192 Artificial intelligence: questions and answers, European Commission, 12 December 2023
195 Generative AI Thriving in Competitive EU Market, New Study Finds, Computer and Communications Industry Association, 21 March 2024
196 EU’s new AI Act risks hampering innovation, warns Emmanuel Macron, Financial Times, 11 December 2023
198 Liability rules for Artificial Intelligence, European Commission, 28 September 2022
199 What can you expect from the new product liability directive? Covington, 14 March 2024
200 The Global AI Index, Tortoise, accessed 23 May 2024
202 China’s trade returns to growth on back of AI equipment imports, Financial Times, 9 May 2024
203 China offers AI computing ‘vouchers’ to its underpowered start-ups, Financial Times, 4 March 2024
204 China’s AI Regulations and How They Get Made, Carnegie Endowment for International Peace, 10 July 2023, p. 3
205 China to lay down AI rules with emphasis on content control, Financial Times, 11 July 2023
207 UK foreign secretary confirms China invite to AI summit, Politico, 19 September 2023
208 The Bletchley Declaration by Countries Attending the AI Safety Summit, 1–2 November 2023, GOV.UK, 1 November 2023
209 US companies and Chinese experts engaged in secret diplomacy on AI safety, Financial Times, 11 January 2024; White House science chief signals US-China co-operation on AI safety, Financial Times, 25 January 2024; US and China to hold first talks to reduce risk of AI ‘miscalculation’, Financial Times, 13 May 2024
210 Artificial intelligence: questions and answers, European Commission, 12 December 2023
211 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 52
212 British Standards Institution (GAI0028)
213 British Standards Institution (GAI0028)
214 Use of artificial intelligence in Government, National Audit Office, HC 612, p. 31; About the AI Standards Hub, AI Standards Hub, accessed 23 May 2024
215 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, para 42
216 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, para 42
217 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, Chapter 4
218 Creative Commons (GAI0015), Institution of Engineering and Technology (GAI0021)
220 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 20
221 Covenant for Using Artificial Intelligence in Policing, National Police Chiefs Council, 28 September 2023
222 Facial recognition technology in law enforcement equitability study: final report, National Physical Laboratory, 5 April 2023
223 New action to tackle ethnic and other biases in medical devices, GOV.UK, 12 March 2024
224 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, para 49
226 Regulating AI: the ICO’s Strategic Approach, Information Commissioner’s Office, 30 April 2024, p. 4
227 Regulating AI: the ICO’s Strategic Approach, Information Commissioner’s Office, 30 April 2024, p. 5
228 Regulating AI: the ICO’s Strategic Approach, Information Commissioner’s Office, 30 April 2024, p. 7
229 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, para 54
230 It’s not just Taylor Swift; all women are at risk from the rise of deepfakes, Glamour, 31 January 2024
231 We asked thousands of GLAMOUR readers about sexual consent, from sexual assault to deepfaking. Here’s what they said…, Glamour, 28 February 2024
232 How the Online Safety Act will help to protect women and girls, Ofcom, 29 November 2023
233 Government cracks down on ‘deepfake’ creation, GOV.UK, 19 April 2024
234 HC Deb, 15 May 2024, col 374 (Commons Chamber)
235 Rishi Sunak suggests general election in second half of year, BBC News, 4 January 2024
236 Deepfake video shows Keir Starmer promoting an investment scheme, Full Fact, 16 November 2023
237 No evidence clip of Sadiq Khan supposedly calling for ‘Remembrance weekend’ to be postponed is genuine, Full Fact, 10 November 2023
238 Annual Review 2023: Making the UK the safest place to live and work online, National Cyber Security Centre, 14 November 2023, p. 40
239 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, pp. 22–23; JCNSS launches inquiry on Defending Democracy with UK election expected this year, 1 February 2024
240 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, pp. 22–23
241 Tech Accord to Combat Deceptive Use of AI in 2024 Elections, Munich Security Conference, 16 February 2024
242 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, para 57
243 CMA seeks views on Microsoft’s partnership with OpenAI, Competition and Markets Authority, 8 December 2023
244 Opening remarks at the American Bar Association (ABA) Chair’s Showcase on AI Foundation Models, GOV.UK, 11 April 2024
245 ICO statement on banks sharing and gathering personal information, Information Commissioner’s Office, 26 July 2023
246 Navigating the UK’s Digital Regulation Landscape: Where are we headed? Financial Conduct Authority, 22 April 2024
247 The UK LLM opportunity: Air Street at the House of Lords, Air Street Press, 28 September 2023
248 Computing Power and the Governance of Artificial Intelligence, Girish Sastry, Lennart Heim, Haydn Belfield, Markus Anderljung, Miles Brundage, Julian Hazell, Cullen O’Keefe, Gillian K. Hadfield, Richard Ngo, Konstantin Pilz, George Gor, Emma Bluemke, Sarah Shoker, Janet Egan, Robert F. Trager, Shahar Avin, Adrian Weller, Yoshua Bengio, Diane Coyle, 14 February 2024, p. 7
249 Computing Power and the Governance of Artificial Intelligence, Girish Sastry, Lennart Heim, Haydn Belfield, Markus Anderljung, Miles Brundage, Julian Hazell, Cullen O’Keefe, Gillian K. Hadfield, Richard Ngo, Konstantin Pilz, George Gor, Emma Bluemke, Sarah Shoker, Janet Egan, Robert F. Trager, Shahar Avin, Adrian Weller, Yoshua Bengio, Diane Coyle, 14 February 2024, p. 10
250 Technology Secretary announces investment boost making British AI supercomputing 30 times more powerful, GOV.UK, 1 November 2023
251 Spring Budget 2024, HC 560, HM Treasury, 6 March 2024, p. 61
252 Scaling Compute, Advanced Research and Invention Agency, accessed 23 May 2024
253 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, para 61
254 A pro-innovation approach to AI regulation, CP 815, Department for Science, Innovation and Technology, 29 March 2023, p. 6
255 International Scientific Report on the Safety of Advanced AI: Interim Report, Department of Science, Innovation and Technology, 17 May 2024, p. 83
256 House of Lords Communications and Digital Committee, First Report of Session 2023–24, Large Language Models and Generative AI, HL Paper 54, para 159
257 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, para 65–67
258 OpenAI co-founder on company’s past approach to openly sharing research: ‘We were wrong’, The Verge, 15 March 2023; Google Gemma: because Google doesn’t want to give away Gemini yet, The Verge, 21 February 2024
259 Meta releases Llama 3, claims it’s among the best open models available, TechCrunch, 18 April 2024
260 House of Lords Communications and Digital Committee, First Report of Session 2023–24, Large Language Models and Generative AI, HL Paper 54, para 40
261 AI Foundation Models update paper, Competition and Markets Authority, 11 April 2024, p. 23
263 Internet Watch Foundation (GAI0130)
264 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, pp. 33–34
265 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, para 68
267 The government’s code of practice on copyright and AI, GOV.UK, 29 June 2023
268 The Times Sues OpenAI and Microsoft Over A.I. Use of Copyrighted Work, The New York Times, 27 December 2023
269 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 19
270 Department for Culture, Media and Sport, Council for Science and Technology Report: Harnessing Research and Development (R&D) in the UK Creative Industries, April 2024, p. 4
271 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, para 72–74
272 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, pp. 25–26
273 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 57
274 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 34
275 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, para 75–77
276 Daron Acemoglu, Simon Johnson: Rebalancing AI, International Monetary Fund, 30 November 2023
277 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, p. 18
278 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, para 78–81
279 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, pp. 37–39; G7 nations to harness AI and innovation to drive growth and productivity, GOV.UK, 15 March 2024; UK & United States announce partnership on science of AI safety, GOV.UK, 2 April 2024
280 Seoul Declaration for safe, innovative and inclusive AI by participants attending the Leaders’ Session: AI Seoul Summit, 21 May 2024, Department for Science, Innovation and Technology, 21 May 2024. The signatory jurisdictions were Australia, Canada, the European Union, France, Germany, Italy, Japan, the Republic of Korea, the Republic of Singapore, the United Kingdom, and the United States
281 New commitment to deepen work on severe AI risks concludes AI Seoul Summit, GOV.UK, 22 May 2024
282 A pro-innovation approach to AI regulation: Government response to consultation, CP 1019, Department for Science, Innovation and Technology, 6 February 2024, pp. 39
283 UAE’s top AI group vows to phase out Chinese hardware to appease US, Financial Times, 7 December 2023
284 US seeks alliance with Abu Dhabi on artificial intelligence, Financial Times, 20 April 2024
285 Frontier AI Safety Commitments: AI Seoul Summit 2024, Department for Science, Innovation and Technology, 21 May 2024
286 An Artificial Intelligence Strategy for NATO, NATO Review, 25 October 2021
287 NATO starts work on Artificial Intelligence certification standard, NATO, 7 February 2023
288 Science, Innovation and Technology Committee, Ninth Report of Session 2022–23, The Governance of Artificial Intelligence: interim report, HC 1769, para 82
289 A pro-innovation approach to AI regulation, CP 815, Department for Science, Innovation and Technology, 29 March 2023, p. 50
290 AI Safety Institute: third progress report, Department for Science, Innovation and Technology, 5 February 2024
291 Capabilities and risks from frontier AI: A discussion paper on the need for further research into AI risk, Department for Science, Innovation and Technology, 25 October 2023, p. 26
292 International Scientific Report on the Safety of Advanced AI: Interim Report, Department for Science, Innovation and Technology, 17 May 2024, p. 83
293 Apocalypse a real risk unless smart tech has ‘kill switch’, warns AI godfather, The Times, 24 September 2023
294 International Scientific Report on the Safety of Advanced AI: Interim Report, Department for Science, Innovation and Technology, 17 May 2024, p. 52