Artificial intelligence (AI), the delicate balance between regulation and innovation, has become a focal point for policymakers, technologists, and ethicists alike. How do nations approach AI governance, from Europe’s stringent AI Act to diverse US proposals?
I believe it is crucial to establish an AI policy and guidelines to safeguard humanity, privacy, well-being, and accurate understanding. Concurrently, it should strive to harmonise innovation, longevity, openness, and effectiveness. The article is my perspective on the policy and the world’s movement toward a thoughtful direction.
“One of the dangers in the 21st century is that machine learning and artificial intelligence will make centralised systems much more efficient than distributed systems, and dictatorships might become more efficient than democracies.” – Yuval Noah Harari, bestselling author, historian, and philosopher.
Over the last few years, AI systems have proliferated across different industries. According to McKinsey, 72% of organisations now use some form of artificial intelligence (AI), representing a 17% increase since 2023.
A recent study by the IBM Institute for Business Value (IBM IBV) revealed that 96% of leaders acknowledge that implementing generative AI could increase the risk of security breaches. Surprisingly, despite this awareness, only 24% of ongoing generative AI projects are currently adequately secured.
AI governance legislation and policies that match the rapid growth and diversity of AI technologies. These efforts include comprehensive legislation, targeted laws for specific use cases, national AI strategies, and voluntary guidelines and standards.
While there is no one-size-fits-all approach to regulating AI, common patterns in these efforts can be observed. AI ‘s transformative nature poses a challenge for governments to balance innovation with the risks regulation. Therefore, AI governance often begins with a national strategy or ethics policy rather than immediate legislation, ensuring a thoughtful approach to integrating AI into society while managing potential risks.
AI regulation and policy framework: A global overview
Developing AI policies to reduce the potential negative impacts and risks of AI technologies involves addressing issues like privacy, security, fairness, and democracy.
The EU AI Act, which was approved in December 2023, establishes a framework for regulating artificial intelligence (AI) technologies in European Union member countries. This legislation aims to ensure the responsible development and use of AI systems by providing clear guidelines for ethical standards, transparency, and safety. Under this Act, AI applications are categorised into different risk levels, from low-risk to high-risk, and strict requirements are imposed to protect people’s rights and safety.
1. It can manage challenges related to privacy, security, fairness, and even democracy.
As AI technology advances, safeguarding privacy, security, and civil liberties becomes increasingly important. Governments worldwide, such as the EU with GDPR are implementing stringent policies to protect privacy, security, and civil liberties in the era of advancing AI technology.
2. It is to promote responsible development of AI systems that are safe, secure, trustworthy, and aligned with human values and ethical principles.
The General Data Protection Regulation (GDPR) in Europe establishes strict data protection and privacy rules for companies that handle the personal data of EU citizens, including those using AI systems. It enforces data minimisation, purpose limitation, and accountability to protect privacy and civil liberties.
The California Consumer Privacy Act (CCPA) gives California residents the right to know what personal data is being collected about them, who it’s being sold to, and the ability to access and delete their information.
3. It can help safeguard privacy, civil liberties, and other fundamental rights that AI systems may impact
Ensuring AI is developed responsibly and ethically is crucial as its influence expands. Policies and guidelines are being crafted to align AI with human values and ethical standards.
The OECD Principles on AI advocate for strong, safe, and fair AI systems, ensuring they are developed in accordance with human values and ethical standards. They emphasise transparency, accountability, and the protection of human rights.
4. It can create an environment that supports AI innovation while addressing potential risks, striking a balance between regulation and technological progress.
The EU AI guidelines propose a framework for AI’s ethical development and use, focusing on human agency, technical robustness, privacy, transparency, diversity, non-discrimination, societal well-being, and accountability.
Balancing regulation and innovation is key in AI governance. Policies aim to encourage technological progress while addressing potential risks.
5. It can establish frameworks for accountability in AI development and deployment, including standards for transparency and explainability.
The proposed AI Act aims to regulate high-risk AI applications while promoting innovation. It establishes a risk-based approach to regulation, encouraging technological progress while addressing potential risks.
Ensuring AI systems are fair, transparent, and accountable is crucial for building public trust and preventing misuse. This proposed legislation requires companies to conduct impact assessments for automated decision systems to ensure they are fair, transparent, and accountable. It focuses on mitigating biases and ensuring explainability in AI systems.
6. It can focus on catalysing AI research nationally and internationally and supporting fair, open, and competitive AI ecosystems.
The National AI Research Resource Task Force – United States initiative aims to create shared computing and data infrastructure to support AI research across the country, promoting an open and competitive AI ecosystem.
7. It can significantly manage the impact and equitable outcomes of various sectors of society, including labour markets, healthcare, and education.
AI for Good Global Summit – United Nations summit focuses on leveraging AI to address global challenges and promote equitable outcomes in sectors like healthcare, education, and labour markets. It encourages collaboration among governments, industry, and academia.
8. It can foster international collaboration, establish common standards or principles for responsible and governed AI, and promote global collective efforts.
In the global landscape of AI governance, international collaboration is key to establishing unified standards and ensuring the responsible deployment of AI technologies.
The G20 countries have united behind principles that advocate for the AI’s ethical and trustworthy developmentI. By fostering collaboration, these principles aim to harmonise global AI standards and encourage collective efforts in addressing shared challenges.
9. It can build public confidence in the technology by demonstrating that its development and use are responsibly managed.
Building public confidence in AI requires robust frameworks prioritising ethical practices and transparent governance.
Singapore’s Model AI Governance framework offers clear guidelines for organisations to implement ethical AI practices. It aims to bolster public trust in AI technologies and their applications by ensuring transparency and accountability.
10. It can strategically develop a way to share public datasets, measure and evaluate AI technologies, understand workforce needs, and expand public-private partnerships.
Open access to government data fuels AI innovation and supports evidence-based decision-making. Open Government Data Act – United States: Mandating the accessibility of federal data, this act facilitates the evaluation of AI technologies. It promotes public-private collaborations that harness this data to drive societal benefits and advancements in AI.
11. It can emphasise developing accountable AI systems and considering the interplay between AI and broader society.
Developing accountable AI systems and considering societal impacts is crucial for ethical AI deployment. IEEE Global Initiative on Ethics of Autonomous and Intelligent Systems: This initiative sets standards for responsible AI development. It addresses the broader societal implications of AI technologies, ensuring they align with human values and ethical principles.
12. It can strategically encourage collaboration between government, universities, and industry to accelerate advances in AI.
Collaboration between sectors is essential for accelerating AI research and realising its potential for societal transformation. AI and Society 5.0 – Japan: This initiative promotes synergy between government, academia, and industry to advance AI research. Promoting innovation ecosystems aims to create a society where AI contributes significantly to human well-being and economic growth.
13. It can guide AI’s responsible and effective use in government operations and public services.
Efficient deployment of AI in public services requires frameworks that ensure responsible and effective implementation. AI in Government Act (proposed) – United States: This act seeks to enhance the federal government’s use of AI. It focuses on deploying AI technologies responsibly to improve public services and operational efficiency.
14. It can provide a framework for addressing new challenges and opportunities as they arise to adapt to rapid technological change.
Adapting to the dynamic landscape of AI technologies involves strategic investments in research and development. Canada: Canada’s AI strategy prioritises agility in responding to technological advancements. By investing in AI R&D and nurturing talent, it aims to address emerging challenges and opportunities in artificial intelligence.
15. It can shape market dynamics, influence industry practices, and support fair competition in AI development and deployment while considering economic factors.
Promoting fair competition is essential for fostering innovation and ensuring a level playing field in AI markets. UK: The Competition and Markets Authority (CMA) plays a pivotal role in regulating AI technologies to prevent monopolistic practices. Ensuring fair competition supports a diverse and competitive AI market that benefits consumers and drives technological progress.
16. It can aim to address the increasing demand for AI expertise and improve efforts to advance the AI R&D workforce.
Investing in AI education and training is important for developing a skilled workforce capable of harnessing AI’s potential. France: This initiative aims to meet the growing demand for AI expertise by funding educational programmes and promoting AI literacy. It seeks to empower individuals and organisations to thrive in the AI-driven economy by supporting workforce development.
17. It can involve integrating AI innovations and early education into academics to cultivate a workforce needed for the future.
Efforts to integrate AI into education and workforce development aim to prepare individuals for future challenges and opportunities. Estonia has integrated AI into its education system to cultivate a future-ready workforce, focusing on early education in AI and related technologies.
18. It can address specific sector needs. For example, healthcare, finance, and transportation may require tailored AI policies to address their unique challenges and opportunities.
The FDA provides regulatory frameworks for AI in healthcare, ensuring AI-based medical devices are safe, effective, and beneficial for patient care. The UK Financial Conduct Authority (FCA) regulates AI applications in finance, ensuring they are used responsibly to prevent fraud, manage risks, and protect consumers.
AI Regulations and Policies: Regional case studies
Globally, several federal governments and unions have launched significant AI-related initiatives that seek to balance legislation and AI’s natural evolution. These programmes cover a wide range of strategies, regulations, and rules that promote creativity while guaranteeing the development of AI in an ethical, safe, and transparent manner, showing a thoughtful approach to AI’s future.
The timeline vividly illustrates a noticeable increase in global efforts and collaboration to establish policies and regulations for artificial intelligence (AI) from 2020 onward. What’s particularly reassuring is the proactive nature of several major economies in implementing specific measures to oversee the development and utilisation of AI.
Here is the list of significant government policies and programs to promote AI while maintaining policies around it:
European Union: Leading the Charge in AI Policy and Innovation
AI policies and support mechanisms are evolving rapidly, with distinct approaches emerging across countries and regions. The European Union (EU) is at the forefront of this evolution, actively developing comprehensive AI policies, regulations, and support schemes that distinguish it from other regions.
Timeline for a Coherent AI Framework:
October 2020: The European Council invited the Commission to propose ways to increase investments in AI research and provide a clear definition of high-risk AI systems.
April 2021: The Commission proposed a regulation aimed at harmonising rules to improve trust and foster development (AI Act), along with a coordinated plan that includes joint actions for the Commission and member states.
December 2022: The Council adopted its position on the AI Act to ensure that AI systems in the EU market are safe and respect fundamental rights and Union values.
December 2023: The Council and the European Parliament agreed on the AI Act, emphasising fundamental rights and EU values.
May 2024: The Council approved the AI Act to harmonise rules on artificial intelligence, taking a risk-based approach with stricter rules for higher-risk AI applications.
Programs and Support Packages for AI Innovation
The EU has implemented various programs and packages to encourage AI innovation, commercialisation, and research, including:
AI Factories: These are equipped with powerful computers for machine learning, training large AI models, and supporting AI startups through robust data policies.
GenAI4EU Initiative: Aim to boost the adoption of generative AI in key industrial sectors and enhance the EU’s talent pool.
Financial Support: Substantial funding is provided through Horizon Europe and the Digital Europe Programme, demonstrating the EU’s commitment to AI development.
AI Ecosystems and Startups: The focus is on creating thriving AI ecosystems and supporting startups to become global frontrunners in trustworthy AI development by creating common European data spaces to facilitate AI innovation.
United Kingdom: Evolving AI Governance and Innovation
Alan Turing (1912-1954), often hailed as a founder of AI, introduced the Turing machine and the Turing Test, pivotal contributions that shaped the field of artificial intelligence. His groundbreaking work laid the foundation for modern computing and inspired generations of researchers and innovators in the UK and beyond.
Timeline of UK’s Influential Role in Global AI Governance
2018: The House of Lords Select Committee on Artificial Intelligence recommends measures to promote workers joining the AI sector. The UK government and industry agree on a £1 billion deal to boost AI adoption.
2020: The Lords Liaison Committee reviews AI adoption progress, finding no need for cross-cutting legislation.
2021: The UK government unveils its ambition to establish Britain as a global AI superpower with a new national AI strategy.
2022: The government signals a light regulatory approach to AI.
2023: A light-touch AI regulation White Paper is released, and the UK hosts an international AI Safety Summit culminating in the Bletchley Declaration, signed by 28 countries and the EU.
2024: Based on feedback, the UK maintains a principles-based framework for existing regulators. The ICO initiates a consultation on generative AI, and the UK participates in G7 AI discussions.
Comprehensive Approach to Fostering AI Innovation
The UK’s strategy focuses on fostering AI innovation, supporting entrepreneurs, and ensuring responsible AI development across sectors:
National AI Strategy (2021): Aims to position the UK as a global AI leader, investing in long-term AI ecosystem needs and ensuring equitable benefits across sectors and regions.
AI Safety Institute at NIST: Develops AI standards for national security, public safety, and individual rights.
Responsible AI UK (RAI UK): A £31 million program led by the University of Southampton, addressing AI challenges in health, social care, law enforcement, and financial services.
BridgeAI Program: Facilitates adoption of trusted AI solutions in agriculture, construction, creative industries, and transport/logistics.
Alan Turing Institute: National champion for AI research and innovation, enhancing its role through regional investments to become a truly national institute.
United States: Shaping AI Policy and Innovation
The US leads global efforts in AI research and development with a comprehensive approach that emphasises innovation, ethics, security, and global leadership. These initiatives highlight the country’s commitment to advancing AI while ensuring responsible deployment and fostering public trust.
Timeline of Key AI Policies and Initiatives
December 2020: Directive establishes the AI Center of Excellence and guides AI acquisition across federal agencies.
January 2021: National AI Strategic Plan is developed to coordinate federal AI activities and guide research investment.
October 2022: Framework and blueprint for the AI Bill of Rights outline principles for AI system design, use, and deployment.
January 2023: NIST releases the AI Risk Management framework to manage associated risks.
September 2023: Leading AI companies commit to product safety, security, and public trust.
October 2023: Executive order establishes AI safety and security standards, mandates safety test sharing, and promotes privacy, civil rights, equity, innovation, and workforce support.
Programs Promoting AI Innovation and Development
The US government has introduced several programs to foster AI innovation across sectors:
National AI Research Resource (NAIRR) Pilot: Provides researchers and students access to essential AI resources including computing, data, software, models, and training.
AI Workforce Development Pilot Program: Aims to train over 500 new AI researchers by 2025 through partnerships with national laboratories and higher education institutions.
Entrepreneurial Fellowships Program: Supports entrepreneurs with mentorship, stipends, and research tools over a two-year period.
National Defense Science and Engineering Graduate (NDSEG) Fellowship: Offers three-year fellowships to graduate students in AI and strategic research disciplines.
EducateAI NSF Initiative: Invites proposals to advance inclusive computing education and integrate AI-focused curricula in educational institutions.
AI Safety Institute at NIST: Develops standards ensuring national security, public safety, and individual rights.
AI Testbed Programs: Collaboration between National Laboratories, NIST, NAIRR Pilot, Department of Energy, and the private sector to develop security risk assessment tools and testing environments for AI systems.
India: Advancing AI through Strategic Initiatives
India’s AI market is rapidly expanding, projected to reach $17 billion by 2027, with substantial growth anticipated by 2025. Leveraging its position as the world’s third-largest talent pool for AI, India is focusing on fostering innovation and adoption through strategic policies and initiatives.
Key Milestones in India’s AI Development
2018: NITI Aayog unveils the National Strategy for Artificial Intelligence (AIforAll) to harness AI for economic growth and social development.
2019: Ministry of Electronics and Information Technology (MeitY) establishes a committee to advise on a national AI program.
2020: NITI Aayog proposes a $1.3 billion plan for the AIRAWAT cloud computing platform and AI Centers of Excellence at leading educational institutes.
2021: Launch of the AI portal with NASSCOM to share AI developments; introduction of AI curriculum in schools by Ministry of Education and CBSE.
2022: Expansion of INDIAai into a comprehensive platform for AI-related developments; government plans to establish three AI Centers of Excellence.
2023: Inclusion of AI Centres of Excellence in the national budget; RAISE events support AI startups in fundraising and innovation showcasing.
2024: The government is working on regulating AI to contain risks and implement guard rails and expected to be released in 2024.
Initiatives Driving AI Growth in India
National AI Strategy (AIforAll): Aims to utilise AI for economic and social development.
Promotion of AI Research: Funding initiatives and institutions support AI research across India.
Skills Development: Various programs focus on enhancing AI talent through education and training.
Sector-Specific Applications: AI is targeted at sectors like agriculture, healthcare, education, and smart cities.
Startup Ecosystem: Vibrant support for AI-focused startups through numerous incubators and accelerators.
Digital India Initiative: Comprehensive program to digitally empower society, indirectly supporting AI development.
Japan: Advancing AI Governance and Innovation
Japan is intensifying its focus on AI governance while actively participating in shaping global frameworks. The country’s efforts span domestic policy development and international collaboration, positioning Japan as a key player in the global AI landscape.
Key Milestones in Japan’s AI Development
April 2016: Proposal of basic AI research and development rules at the G-7 ministers’ meeting.
May 2023: Launch of the “Hiroshima AI Process” at the G-7 Hiroshima Summit.
October 2023: Establishment of international guiding principles and a code of conduct for advanced AI system developers by the G-7.
December 2023: Presentation of draft guidelines for AI-related businesses aligned with G-7 principles.
February 2024: Establishment of the Japan AI Safety Institute to advance AI safety research.
April 2024: Collaboration with the United States on AI development under the Hiroshima AI Process; launch of the “Hiroshima AI Process Friends Group” at the OECD ministerial meeting.
Initiatives Driving AI Growth in Japan
Society 5.0 Strategy: Integrates AI and robotics to address societal challenges through industry-academia collaboration.
Top Universities in AI Research: Institutions like the University of Tokyo, Kyoto University, and Tokyo Institute of Technology lead in AI research and entrepreneurship support.
Japan AI Alliance: Involves established businesses, startups, and government to drive AI innovation across sectors.
Education and Skills Development: Investments in educational programs to equip workers with AI skills; partnerships with universities and research labs.
Agile Regulation: Emphasises a risk-based approach to AI regulation, focusing on maximising benefits while managing risks.
AI Guidelines for Business: Voluntary AI risk management tool based on the Hiroshima AI Process principles.
International Engagement: Actively participates in G-7 and OECD frameworks for international AI governance.
Subnational Initiatives: Local entities gather knowledge, develop AI use-cases, establish sandboxes, and form partnerships to address sector-specific AI implications.
China: Advancing AI Governance and Regulation
China has demonstrated a growing commitment to AI governance and regulation, evolving through comprehensive policies and laws aimed at addressing ethical concerns, managing algorithms, and overseeing content generation.
Key Milestones in China’s AI Development and Regulation
Late 1970s: Initiation of China’s AI development following Deng Xiaoping’s economic reforms prioritising science and technology.
July 2017: Release of the “New Generation AI Development Plan” by the State Council, outlining a roadmap for AI governance regulations up to 2030.
June 2019: Issuance of the “Governance Principles for New Generation AI” by the National New Generation AI Governance Expert Committee, proposing eight principles for AI governance.
December 2020: Publication of the “Outline for Establishing a Rule-of-Law-Based Society (2020–2025)” by the CCP Central Committee, addressing issues like recommendation algorithms and deep fakes.
September 2021: Release of “Guiding Opinions on Strengthening Overall Governance of Internet Information Service Algorithms” by the Cyberspace Administration of China (CAC) and other bodies, along with “Ethical Norms for New Generation AI” by the National New Generation AI Governance Expert Committee.
December 2021: Issuance of the “Provisions on the Management of Algorithmic Recommendations in Internet Information Services” by the CAC and other agencies, introducing the concept of an “algorithm registry.”
March 2022: Implementation and enforcement of national, regional, and local level AI regulations.
January 2023: Proposal of legislation known as the “Deep Synthesis Provisions” aimed at regulating deepfake and generative technologies nationally.
May 2023: Implementation of interim measures governing generative AI, effective from August 15, 2023.
Initiatives Driving AI Governance and Regulation in China
Comprehensive Governance Framework: China’s approach involves a systematic framework covering ethical norms, algorithm management, and content generation oversight.
Strategic Planning: Long-term plans like the “New Generation AI Development Plan” provide a structured approach to AI governance up to 2030.
Ethical Standards: The establishment of ethical norms and principles ensures responsible AI development and deployment.
Legislative Efforts: Continuous development and refinement of laws and regulations tailored to emerging AI technologies like deepfakes and generative AI.
Implementation and Enforcement: Rigorous implementation and enforcement mechanisms at national, regional, and local levels ensure compliance with AI regulations.
International Influence: China’s policies and guidelines contribute to global AI governance discussions and frameworks.
The future of innovation with AI regulation and governance
The development and implementation of robust regulatory and governance frameworks will significantly impact future artificial intelligence (AI) developments. The rapid development of AI technologies is impacting several industries, including healthcare, finance, education, and transportation. However, these developments also raise important moral, societal, and financial issues. Effective regulation is essential to mitigate risks like bias, privacy violations, and unanticipated effects and to ensure that AI systems are created and implemented responsibly.
It is crucial that governance frameworks are inclusive and adaptable, bringing together leaders from business, government, academia, and civil society. By collaborating, we can establish norms and guidelines that strike a balance between innovation and the public interest. To ensure that AI applications are just and fair, it is essential to have frameworks that not only encourage but also enforce accountability and openness in AI decision-making processes. This is vital for building and maintaining trust among the public in AI technologies.
Additionally, the establishment of standardised regulations that protect against regulatory arbitrage and promote international norms requires close international cooperation. By creating an atmosphere where ethical issues and technical progress exist, we can fully utilise AI while preserving moral principles and the welfare of society. Therefore, the ability to successfully negotiate the complexity of governance and legislation will determine the direction of AI innovation in the future.
Dilip Pungliya is a business leader, Artificial Intelligence consultant, blockchain advisor, metaverse solution expert, data leader, technologist, and business, process, & technology architect. As a board member and significant shareholder of Ztudium, Dilip brings a wealth of experience in business leadership and data technology. In his role as the Managing Partner of the ztudium Group, he benchmarks his strategic acumen in steering effective strategy and framework development for the company. Dilip also plays a pivotal role in his family’s limited company in India, VPRPL, where he oversees operations and strategic planning. His professional journey includes impactful collaborations with esteemed organisations such as Shell, the Department for Environment Food and Rural Affairs, Deutsche Bank, ICBC Standard Bank Plc, BNP Paribas, and HSBC Investments. Beyond his professional endeavours, Dilip is deeply committed to philanthropy and charitable work, particularly during the global challenges presented by the COVID-19 pandemic.