Asia: different approaches to the risk-reward balance
28 November 2024
AI regulation in Asia is diverse and dynamic, reflecting the varying interests, policy goals, and cultural values of different jurisdictions. While some jurisdictions, such as Singapore and Hong Kong, have adopted a soft touch, voluntary approach to AI regulation, relying on general and sector-specific guidance frameworks, others, such as China, have adopted a more prescriptive and technology-specific approach, introducing mandatory regulations and measures to address the risks associated with different aspects of AI. Other jurisdictions, such as Vietnam, Taiwan, South Korea, and Japan, have also begun to propose and enact laws and guidelines to regulate AI, demonstrating a growing recognition of the potential and importance of AI for economic and social development, and the need to regulate the risks of AI. At the regional level, ASEAN has also issued a guide on AI governance and ethics, which aims to encourage the alignment and interoperability of AI regulatory frameworks across Southeast Asia, and to support the development and adoption of trustworthy and human-centric AI in the region.
This article provides a comparative overview of the current and emerging AI regulatory frameworks in Asia, focusing on the following jurisdictions: Singapore, Hong Kong, China, Vietnam, Taiwan, South Korea, Japan, and ASEAN.
Singapore has been a pioneer in promoting and adopting AI in the region, launching its National AI Strategy in 2019, which was further updated in 2023. The Strategy aims to position Singapore as a global hub for AI innovation and deployment, and to leverage AI to enhance social and economic outcomes. To support this vision, Singapore has adopted a voluntary approach to AI regulation, providing guidance to interested parties on the ethical and responsible use of AI, without any legally binding effects.
The key frameworks and guidelines include:
Sector-specific regulators have also developed frameworks that apply to particular industries. For example, the Monetary Authority of Singapore has launched a project with industry partners to develop a risk framework for the use of generative AI in financial sectors. This will supplement the guidance on Principles of Fairness, Ethics, Accountability and Transparency (FEAT) in the use of Artificial Intelligence and Data Analytics in Singapore's Financial Sector released in 2018.
The intention behind Singapore's approach thus far is to encourage the use of AI in accordance with its national strategies, while at the same time allowing flexibility and innovation for organisations to adopt AI solutions that suit their needs and contexts. However, compliance with current guidance is entirely voluntary which means that developers and users of AI may at times struggle with a lack of certainty around the enforceability of the various frameworks and guidelines. At the same time, however, they will need to be aware of existing laws which potentially impact specific aspects of AI use.
Hong Kong: Ethical Guidelines and Sectoral Regulations
Hong Kong, like Singapore, does not have any overarching AI-specific legislation. It too has taken a similar voluntary approach to regulation in this area. A number of guidance notes have been published by government bodies to steward and facilitate the ethical use of AI technology, especially where it impacts personal data protection and public sector projects.
The key guidance notes include:
The aim underpinning each of these guidance notes is the provision of a common reference point and a balanced approach for organisations to adopt AI in a manner that respects privacy, ethics, and human rights, while simultaneously encouraging innovation and competitiveness. However, as the guidance notes are not legally binding, their effectiveness and impact is heavily dependent on the extent to which stakeholders understand, adopt and follow them, and will also hinge on the willingness of the various regulators and stakeholders to coordinate their approaches. Moreover, given the breadth of the potential applications of AI, the somewhat narrower focus of the guidance thus far and rapid pace of change in the industry generally, it may prove impossible for the guidance notes to address all the emerging and complex issues and risks arising from AI. Matters such as safety, security, accountability, and liability, in particular, may require more specific and comprehensive legal and regulatory responses.
As in other jurisdictions, organisations may also need to reconcile AI regulation with pre-existing laws, such as the Personal Data (Privacy) Ordinance, the Copyright Ordinance, and the Trade Descriptions Ordinance.
China is a frontrunner in AI in both the regulation of AI as well as its development and deployment. Although there is also currently no general or overarching AI law in China, regulators (the Cybersecurity Administration of China in particular) have in recent years introduced mandatory technology-specific regulations and measures to address the risks associated with different aspects of AI, in particular algorithmic recommendations, deep synthesis, and generative AI (see our previous article here). These measures have legal effect and apply to both domestic and foreign providers and users of AI services in China.
The key regulations and measures include:
The current AI governance regime in China appears to target specific issues arising from AI technology whilst still promoting the broader development of the technology as a whole. The burden of regulatory compliance is placed largely on AI service providers as the upholders of the security and quality of their AI services. Unlike the EU's overarching legislation and risk-based approach, the measures currently in force in China target specific AI technologies with the ultimate penalty of prosecution for non-compliance in accordance with public security and criminal laws.
Looking ahead, more comprehensive legislation to regulate AI is expected to be introduced in mainland China, as indicated by the State Council's 2024 Legislative Work Plan, which includes a draft AI law to be submitted for deliberation before the end of the year. As a result, AI service providers active in mainland China need to take steps to comply with current regulations where applicable and stay informed of further AI regulatory developments.
Vietnam, with its ambitions to become a regional hub for AI, is one of the few jurisdictions in Asia that has recently proposed a draft law for the specific regulation of AI. On 2 July 2024, the Ministry of Information and Communication released a draft digital technology industry law to regulate digital technology products and services, including AI for public consultation. The draft law is expected to be submitted to the National Assembly for approval in 2025.
The draft legislation proposes to regulate AI in the following manner:
The public consultation, which concluded in September 2024, took place against the backdrop of the Ministry's acknowledgement at a press conference in May 2024 of the significant export revenue generated by Vietnamese digital technology enterprises, and the need to swiftly put legislation in place to encourage domestic digital technology firms to do business abroad.
As a result, the draft law reflects Vietnam's recognition of the need to balance the potential importance of AI for economic and social development against need to address the ethical and legal challenges it poses. However, the draft legislation is still at an early stage and may be subject to further revisions and clarifications before it becomes law. Moreover, the potential implications of AI in areas such as data protection, intellectual property, consumer protection, and liability rules may require the AI-specific legislation to be harmonised with the existing regulatory regime in Vietnam.
Taiwan does not have a single comprehensive law regulating AI, although – as in other jurisdictions - some aspects of AI may fall within the reach of existing legislation such as the Personal Data Protection Act, the Copyright Act, and the Consumer Protection Act. However, a draft AI basic law has been proposed by the National Science and Technology Council of Taiwan (NSTC).
The draft basic law sets out fundamental principles concerning research into and use of AI. These principles include sustainable development and well-being, human autonomy, protecting the privacy of personal data, establishing security measures to ensure the safety of AI systems, maintaining transparency and explainability, avoiding the risk of bias and discrimination and enhancing accountability. It is proposed that these would only apply to the release of AI applications in an effort to support innovation in the research and development stages. The basic law also proposes that the Ministry of Digital Affairs promotes a risk classification framework in line with international standards. This will potentially be influenced by the EU AI Act and the draft US Algorithmic Accountability Act of 2022.
The public consultation period for the draft basic law ended in September 2024. The draft law is expected to be sent to the Cabinet for review shortly before being sent to the Legislative Yuen for deliberation. In the meantime, some sector regulators have issued guidelines or regulations for the use of AI in their respective domains. For example, the Financial Supervisory Commission has issued Guidelines for AI Applications in the Finance Industry, which provide principles and standards for financial institutions to follow when adopting AI technology, such as conducting risk assessments, ensuring data quality and security, providing transparency, and establishing governance and audit mechanisms.
Although the draft may change as it goes through the legislative process, it reflects Taiwan's aspirations to establish a legal foundation for the development and governance of AI which aligns with international standards and best practice.
South Korea is another jurisdiction in Asia that has recently proposed a draft law to regulate AI, demonstrating a commitment similar to that in other jurisdictions to foster the AI industry whilst protecting the users of AI services. In February 2023, the Science, ICT, Broadcasting and Communications Committee of the South Korean National Assembly approved the proposed Act on Promotion of the AI Industry and Framework for Establishing Trustworthy AI (AI Bill). The AI Bill is currently under review by the National Assembly of Korea, and as such may be subject to further revisions and the National Assembly's final approval before it becomes law.
The AI Bill consolidates seven AI-related bills introduced since 2022, with two principal goals: promoting the AI industry, and user protection. User protection will be achieved by increasing security through the imposition of stringent notice and certification requirements for high-risk AI applications that could have a significant impact on safety, health or fundamental rights. Examples of high-risk AI systems include those used in healthcare, transportation (including autonomous vehicles) and automated decision-making that have a significant impact on individual rights or obligations. A more permissive approach will be taken for lower-risk systems.
The AI Bill also lays down a statutory basis for issuing ethical guidance for AI, reflecting the principles of human dignity and value, human autonomy and control, transparency and explainability, fairness and non-discrimination, reliability and safety, and social responsibility and public interest. The ethical guidance will be formulated by a newly established AI Ethics Committee, which will also be responsible for reviewing and evaluating the ethical compliance of AI systems and services.
The AI Bill further addresses some specific issues and challenges posed by AI, such as the protection of personal data and intellectual property rights, the establishment of a AI Mediation Committee as a means of dispute resolution, and the promotion of AI education and research.
Overall, the approach of the draft legislation reflects South Korea's recognition of the potential benefits and importance of AI for its economic and social development. The AI Bill adopts a risk-based approach to regulating AI, similar to the EU AI Act, but aims to take a more business-friendly approach by not requiring registration or conformity assessments, or pre-approval from any government authority.
Japan does not have a comprehensive law regulating AI. Existing laws which potentially affect certain aspects of AI include the Act on the Protection of Personal Information, the Copyright Act, and the Unfair Competition Prevention Act. However, Japan has been active in developing and promoting ethical and governance principles for AI, both at the domestic and international levels.
At the domestic level, Japan issued the AI Guidelines for Business in April 2024, which are based on the 10 principles of human-centric AI adopted by the Cabinet Office in 2019. The Guidelines are non-binding and contain recommendations based on 10 guiding principles, such as ensuring safety and security, respecting human dignity and rights, providing transparency and accountability, and promoting innovation and development. The Guidelines also provide specific guidance for advanced AI systems, such as generative AI, systems that are capable of operating without human intervention, adaptive AI and interactive AI.
At the international level, Japan has been leading the Hiroshima AI Process, launched under its presidency of the G7 in May 2023, to introduce a harmonised global governance framework for AI. The Hiroshima AI Process consists of three main components: the Hiroshima International Guiding Principles for all AI Actors, which set out 10 common principles for the ethical and responsible use of AI; the Hiroshima Process International Code of Conduct for Organisations Developing Advanced AI Systems, which provide more detailed and operational guidance for specific types of AI systems, such as generative AI and autonomous AI; and the Hiroshima Process International Cooperation Mechanism, which aims to facilitate the implementation and monitoring of the principles and the code of conduct, as well as to promote dialogue and cooperation among stakeholders.
Japan has also launched a new AI Safety Institute, which will, among other things, implement standards for the development of generative AI. It has already issued sector-specific guidelines and regulations for the use of AI in certain domains, such as finance, healthcare, and education.
Japan's AI Guidelines for Business and the Hiroshima AI Process reflect its aspiration to establish a common and human-centric framework for the development and governance of AI, as well as to align with international standards and best practices. However, the AI Guidelines for Business and the Hiroshima AI Process are not legally binding and as such their effectiveness is limited by the willingness of the various interested regulators and stakeholders to comply with them. It is also possible that, as AI technology develops and its reach widens, the current voluntary framework may not be adequate to cover all the potential implications of AI, and further legal and regulatory action will be required.
ASEAN, which comprises the 10 member states of Brunei, Cambodia, Indonesia, Laos, Malaysia, Myanmar, Philippines, Singapore, Thailand and Vietnam, issued a regional Guide on AI Governance and Ethics in February 2024. The guide is intended to encourage the alignment and interoperability of AI frameworks across Southeast Asia, and to support the development and adoption of trustworthy and human-centric AI in the region.
The ASEAN Guide is framed around seven guiding principles that closely track the OECD AI Principles first adopted in 2019. The guiding principles are that AI should be:
The ASEAN Guide also provides more detailed recommendations and best practices for organisations to adopt in four key areas:
The ASEAN guide includes a risk assessment template to help organisations to assess the probability, nature and severity of harm arising from the use of AI systems and how many people could be affected. The assessment should also scope the feasibility and practicability of human involvement in the decision-making. The ASEAN Guide advises that the each company's governance structure needs to strike the right balance between flexibility and rigidity to reflect its own individual culture, and recommends designing in escalation mechanisms for AI systems and use cases that are of higher risk.
The ASEAN Guide is a voluntary and light touch approach to regulating AI. It encourages organisations to adopt AI in a manner that respects ethics, human rights, and social values, without stifling innovation and competitiveness. However, as the ASEAN Guide is not legally binding, its effectiveness and impact cannot be guaranteed and will be influenced by the willingness of industry and regulators to adhere to its guidance. It is also worth highlighting that it does not cover generative AI and as such it may not address all the emerging and complex issues and risks arising in this sector.
AI is transforming various sectors and industries across the world, raising new opportunities and challenges for businesses, consumers, and regulators. Currently in Asia, different jurisdictions have adopted diverse approaches to regulating AI, reflecting their varying levels of development, policy goals, and cultural values. With AI legislation still in a relatively junior stage, it remains to be seen whether a region-wide approach can be achieved, or whether countries will continue to adopt different models of AI regulation.
The information provided is not intended to be a comprehensive review of all developments in the law and practice, or to cover all aspects of those referred to.
Readers should take legal advice before applying it to specific issues or transactions.