thumbnail image

 

  • Home
  • About APB Forum
  • APB Forum 2025
  • History 
    • 2024
    • 2023
    • 2022
    • 2021
    • 2020
    • Fall 2019
    • Spring 2019
    • Fall 2018
    • Spring 2018
    • Fall 2017
    • Spring 2017
    • Fall 2016
    • Spring 2016
  • Call for Paper
  • Invitation to the 2024 Asia Privacy Bridge Forum

    Recent advancements in AI technology have accentuated the growing importance of data governance and privacy, while also highlighting the need for international cooperation.

    The 13th Asia Privacy Bridge Forum, inconjunction with Privacy Global Edge, will convene under the theme "International Collaborations in Trustworthy AI Governance and Privacy." This forum offers a unique opportunity for learning and growth in your respective fields. It aims to engage in profound discussions on global collaborative strategies to build a happier society in the AI era. The myriad of ethical issues surrounding data protection and privacy, particularly when intertwined with artificial intelligence technologies, necessitate proactive cooperation among nations to strike an equilibrium between technological progress and regulation, thus fostering corporate innovation.

    Consequently, the 13th Asia Privacy Bridge Forum will go beyond the mere exchange of knowledge pertaining to personal information protection. It will serve as a platform for a thorough analysis of the changes and impacts that artificial intelligence technology will have on various aspects of our lives, including work, education, entertainment, and politics. Furthermore, it will provide an opportunity to collectively generate innovative ideas and collaborative measures across these domains.

    We are confident that your active participation will make a substantial contribution to establishing a forum for discussions that will shape a better future through the 13th Asia Privacy Bridge Forum.

    Beomsoo KIM

    Director, Barun ICT Research Center

  • Keynote Speeches

    Section image

    Navigating the Future: AI Governance and Data Privacy

    in the Philippines - A Regulatory Perspective

    Ivin Ronald D.M. Alzona

    Executive Director, National Privacy Commission,

    Republic of the Philippines

  • Summary of Discussion

    DAY 1

    Session 1: Navigation Gen AI and

    Trustworthy AI Governance for the Future

    Moderator : Beakcheol Jang (Yonsei Univ., Korea)

    Singapore's Evolving Approach to AI Governance

    Jason Grant ALLEN

    Associate Professor, Singapore Management University, Yong Pung How School of law, Singapore

    This presentation explores Singapore's evolving AI governance framework, highlighting the

    country's strategic approach to balancing innovation with public trust and safety. As one of

    the most AI-ready jurisdictions globally, Singapore has positioned AI as a key driver of its

    economic development while adopting a collaborative and risk-based governance model.

    The discussion covers key initiatives such as the Model Framework for AI Governance, AI

    Verify Toolkit, and the National AI Strategy (NAIS 1.0 and 2.0), focusing on the alignment

    between government, industry, and research in building a robust AI ecosystem.

    The presentation also delves into Singapore’s "soft-touch" regulatory approach, which

    emphasizes voluntary standards and quasi-regulation, while comparing it with more rules-

    based models such as the EU’s and China's. Special attention is given to sector-specific

    AI governance in finance, through the FEAT Principles and Veritas Toolkit, and technology-

    specific governance for generative AI, addressing issues like content provenance, safety,

    and AI for the public good.

    Additionally, Singapore's role in shaping regional AI governance through ASEAN and its

    global influence in international AI forums are discussed. The future outlook considers the

    potential shift toward more formal regulation as emerging technologies evolve and the need

    for sustained public trust and collaboration in AI governance.

  • Data Protection, Competition, and AI Governance: The Importance of Data Portability and ADM Governance in Data Protection Laws

    Qing HE

    Assistant Professor, Beijing University of Posts and Telecommunications, China

    This presentation addresses the complexities of data use policies and their effects on

    competition, particularly focusing on how these policies may hinder or promote competitive

    dynamics. Although certain data transfer policies, such as those enabling data portability

    rights, have the potential to enhance competition, the practical implementation of these

    policies often falls short in fostering competitive markets.

    The presentation also delves into the governance of Automated Decision-Making (ADM)

    and its relationship with broader AI governance frameworks. Under data protection laws in

    both China and the EU, individuals are granted the right to challenge algorithmic decisions

    that have a significant impact on them, highlighting the role of ADM governance in AI

    regulation. Key aspects explored include legal definitions, protection policies, and liability

    rules. A comparative analysis of ADM governance across the EU, the United States, and

    China is provided, including the scope and definition of automated decision-making, its

    effect on individual rights, and how ADM governance intersects with policies on Generative

    AI. Relevant legislation such as the EU’s AI Act, the U.S. Blueprint for an AI Bill of Rights,

    Biden’s Executive Order, and China’s Personal Information Protection Law and related

    algorithmic provisions are examined in this context.

    Finally, the presentation emphasizes the importance of risk classification in AI systems,

    with a particular focus on legal practices in China. Three case studies are used to illustrate

    the significance of this issue: credit scoring systems within financial services, price

    discrimination in online services, and electronic surveillance and management systems in

    workplace environments. These examples demonstrate the critical need for a structured

    approach to identifying and mitigating risks within AI systems across various sectors.

  • Designing Accountable Community in the Emerging AI period

    Kohei Kurihara

    CEO, Privacy by Design Lab, Japan

    This presentation focuses on delivering key insights to the design community and

    emphasizing accountability in the process of developing AI services and products. In line

    with the emerging AI trend in society, AI developers and providers are increasingly expected

    to take on responsibility, especially as regulatory and societal demands on the supply side

    rise in the coming decades.

    To address this challenging theme, the discussion highlights the crucial role the design

    community plays in enhancing safety and accountability in relationships between diverse

    stakeholders. Additionally, by sharing effective knowledge and experiences, the community

    can prevent unexpected consequences by integrating different perspectives and insights

    early in the process.

    The community comprises various experts and practitioners, deepening mutual literacy and

    occasionally leveraging their work through "connecting the dots" via project collaborations.

    These projects strengthen the trusted networks among parties that share a similar vision,

    contributing to the community’s goals.

    These are the main topics in this presentation. The necessary action in the emerging AI

    period to prevent the unexpected consequences Multi-stakeholder based accountability

    model by sharing diverse experiences and methods Learning and Sharing community

    function to leverage community member synergies in the projects Designing the vision and

    roadmaps with diverse backgrounds beyond the cultures and histories Finding the remarks

    of community benefits against the AI harms As a conclusion, the presenter will show future

    affection with community based authentic relationship building from his past methodology

    and containing the actionable planning to design community network. And he will speak

    about the future community design to boost the designing opportunities in multilateral

    Asian approaches.

  • Session 2: Reconciling Data Protection and Competition Laws in the Age of AI

    Moderator : Hayoung Kim (Yonsei Univ., Korea)

    Taking Stock: Data Protection, Privacy, and Competition Law

    Orla Lynskey

    Professor, University College London, Faculty of Laws, UK

    Data protection and competition law have historically been treated as distinct fields of law

    with clearly demarcated boundaries, and there has been significant resistance to breaking

    down these boundaries. Nevertheless, legal and technical developments (such as Apple's

    use of a privacy defense to defend against allegations of abuse of market power) mean

    that their intersection is now inevitable. This presentation maps out and critically analyzes

    four ways in which these areas of law influence one another. First, data protection law is

    not neutral—its application (or lack of application) affects market dynamics in a way that

    is relevant to competition law. Second, data protection is integrated into competition law

    analysis as part of the consumer welfare benchmark. Third, competition considerations

    influence the interpretation of some data protection concepts, such as consent, and the

    extent of data protection interferences. Finally, the legislature recognizes this intersection

    by imposing limitations on the data processing activities of digital gatekeepers, subject to

    data protection law.

  • Reproduction of Personas with AI and the Right of Publicity

    Kunifumi Saito

    Associate Professor, Faculty of Policy Management, Keio University, Japan

    This presentation examines the relationship between the personality rights and the right of

    publicity in the context of the reproduction of personas using artificial intelligence.

    In the United States, most lawyers consider the right of publicity to be a type of intellectual

    property right like copyright. Recently, however, an argument has emerged that emphasizes

    the similarities with the right to privacy. It classifies the functions of the right into four

    categories: the Right of Performance, the Right of Commercial Value, the Right of Control,

    and the Right of Dignity. It is significant that the similarity between the Right of Commercial

    Value, which is the core of the function, and the trademark right has been pointed out.

    Meanwhile, in 2012, the Japanese Supreme Court positioned the right of publicity as a

    kind of personality right. However, the official commentary to the decision emphasizes

    the similarities between the right of publicity and copyright. And in practice, disputes over

    the right of publicity are assigned to the specialized divisions for intellectual property of

    the courts. In addition, the case law of the lower courts distinguishes between the right

    of publicity and the rights of personality that relate to moral damages, such as the right of

    privacy and the right of likeness.

    Under Japanese law, personal rights cannot be inherited. For this reason, it is believed that

    a celebrity's right of publicity also ceases upon his or her death. In this presentation, we will

    examine the legal rights involved in the reproduction of the persona of the deceased using

    artificial intelligence. In our discussion, we will draw on a theory from the United States that

    focuses on the similarities between the right of publicity and trademark law.

  • Personal Data & Generative AI

    DaeHee Lee

    Professor, Korea University, Law School, Republic of Korea

    The presentation addresses Korea's personal data regime and its related issues concerning

    AI development. Specifically, it focuses on the recently released "Guidelines on Processing

    of Personal Information Publicly Available for the Development and Deployment of AI

    Models" by Korea's Personal Data Protection Commission. The presentation argues that

    personal data concerns should not serve as obstacles to AI development.

  • Session 3: Digital Shield: Safeguarding Privacy

    and Data for Vulnerable Users

    Moderator : Hyojin Jo (Yonsei Univ., Korea)

    Challenges for Non-digital Natives to Protect the Rights of Digital Natives

    Byungsoo Jung

    Director, Children's Rights Division, The Korean Committee for UNICEF, Republic of Korea

    In 1989, Tim Berners-Lee proposed the concept of hypertext called the World Wide Web

    (WWW). That same year, the UNGA unanimously adopted the UN Convention on the Rights

    of the Child (CRC). The WWW and CRC may not seem to have any direct connection, but

    the publication of these two documents has had a profound impact on life, especially for

    children.

    Children have traditionally been marginalized and viewed as a labor force, parental property,

    etc. However, the CRC affirmed that children are subjects of rights. Digital technology has

    also brought about significant changes in the expansion of children's rights. Educational

    materials available online support children's 'self-directed learning,' and 'distance learning'

    ensures equal educational opportunities for vulnerable children. It also facilitates social

    participation. Therefore, children are referred to as digital natives.

    In response to the growing influence of digital technology, the UN Committee on the

    Rights of the Child (the Committee) issued "General Comment No. 25 on Children's Rights

    in Relation to the Digital Environment" in 2021. Children from around the world expressed

    concerns that while digital technology is an indispensable tool in their lives, it exposes them

    to the risk of violence, abuse, misinformation/disinformation, and the collection of personal

    information, which can lead to further risks. The Committee urges all States Parties to

    protect children from harmful content, all forms of violence in the digital environment,

    respect and protect children's privacy, and regulate advertising and marketing in digital

    services that are inappropriate for children.

    UNICEF, the only agency explicitly mandated by the CRC, is also committed to protecting

    children's rights in the digital environment. It has established a strategic framework

    for online child protection and seeks collaboration from various stakeholders, including

    governments, businesses, caregivers, educators, and children. It is also moving quickly to

    provide direction for emerging technologies such as AI guidance.

    UNICEF calls on all stakeholders to make choices and take actions that put children at the

    center. This is similar to how traffic lights and laws were created to bring order to roads that

    had become chaotic and dangerous with the increase of cars. The difference is that 'child

    centered' approaches are built in from the start to reduce trial and error.

  • Children and AI: Key Issues to Consider to Empower and Protect Them

    Steven Edwin Vosloo

    Policy Specialist, Digital Engagement and Protection, UNICEF Innocenti, Italy

    This presentation provides a detailed framework for pseudonymizing unstructured data, critical for privacy and AI applications. Starting with an introduction to the importance of
    pseudonymization in today’s data-driven landscape, it outlines key methodologies for
    handling sensitive information in formats like images, videos, and free text.
    Practical applications across fields such as healthcare, security, and AI development are
    presented, illustrating real-world benefits and challenges. The presentation concludes with
    a step-by-step approach to pseudonymization—spanning preparation, risk assessment,
    processing, and management—designed to foster responsible and compliant data usage in
    an evolving regulatory environment.

  • Safeguarding and Empowering Vulnerable Children in the Digital Age: Save the Children's Global Initiatives

    Jeffrey DeMarco

    Senior Advisor, Protecting Children from Digital Harm, Save the Children's Global Safe Digital Childhood Initiative, UK

    This presentation explores Save the Children’s comprehensive efforts to protect and

    empower vulnerable children online through three key initiatives.

    First, the Safe Digital Childhood Coalition addresses online protection challenges in the

    Global South, where inadequate regulations expose children to online risks. Notable examples

    include the development of Sri Lanka's National Action Plan, aligned with WeProtect Global

    Alliance recommendations, and the SaferKidsPH program in the Philippines, which combats

    online sexual exploitation and abuse.

    Second, the organization promotes digital literacy and inclusive online safety education

    through initiatives such as the IT for Learning/DIGITAL project in India and Indonesia,

    and a cyber safety campaign led by Save the Children Australia across Pacific nations, in

    collaboration with Facebook.

    Finally, Save the Children is leveraging technology to tackle online harms with innovative

    approaches, including an AI-powered project in India aimed at preventing online violence, a

    collaboration with NetClean to detect abuse materials on corporate devices, and the Cloud

    Chaos mobile game developed in Cambodia. Together, these programs highlight a global

    strategy to safeguard children and empower them as responsible digital citizens.

  • DAY 2

    Session 4: Platform Governance and AI Accountability

    Moderator : Jongsoo Yoon (Attorney, Lee & Ko, Republic of Korea)

    Meta's Approach to Responsible AI

    Raina Yeung

    Director of Privacy and Data Policy, Engagement, APAC at Meta, Singapore

    With the rapid evolution of AI technology, including Generative AI, it is essential for

    different stakeholders to ensure that its development and deployment are responsible and

    transparent. This presentation shares Meta's experience in AI developments, including

    the latest introduction of Llama 3.1 and how Meta built AI responsibly. By using these

    products as examples, we aim to emphasize the importance of an open-source approach

    to benefits for safety, security, competition, and innovation in AI developments and explain

    how our approach to responsible AI has continued to guide us in addressing hard questions

    around issues such as privacy and security, fairness and inclusion, robustness and safety,

    transparency and control, and accountability and governance.

  • Responsible AI in Malaysia: The Role of Data Protection Policy

    Jillian Chia

    Attorney, SKRINE, Malaysia

    This presentation focuses on the AI landscape in Malaysia, particularly the regulatory

    environment and proposed plans to regulate AI, as well as the challenges Malaysia faces in

    this area. Additionally, the discussion covers laws that impact the implementation of AI in

    Malaysia, such as the country’s personal data protection and cybersecurity regimes.

  • Regulatory Landscape for Generative AI in Japan: Insights and Outlook

    Hitomi Iwase

    Attorney, Nishimura & Asahi, Japan

    Japan's approach to regulating Generative AI is characterized by a soft law framework,

    while existing laws (such as the Act on the Protection of Personal Information (APPI),

    the Copyright Act, etc.) apply to the development or use of Generative AI depending

    on the industry or the nature of the AI. In April 2024, the Ministry of Internal Affairs and

    Communications and the Ministry of Economy, Trade and Industry issued the "AI Guidelines

    for Businesses." These guidelines provide 10 guiding principles, including fairness,

    transparency, and accountability, as well as practical guidance for AI developers, providers,

    and users.

    This presentation also covers the legal issues surrounding Generative AI, such as potential

    violations of the APPI and copyright infringement, and examines what platforms need to do

    to manage the risks associated with developing and providing Generative AI.

    Session 5: What is Data Sovereignty?

    Global Cross-border Privacy Rules (GCBPRs)

    and Cooperation in Investigation and Enforcement

    Moderator : Kwangbae PARK (Attorney, Lee & Ko, Republic of Korea)

    South Korea's Regulatory Framework for Cross-Border Data Transfer Policies

    Jeongsoo LEE

    Deputy Director, Personal Information Protection Commission, Republic of Korea

    In this presentation, you can expect a comprehensive introduction to Korean legislation concerning cross-border data transfers. It begins with a brief historical overview of the
    legislative framework, followed by an explanation clarifying the scope and application.
    Additionally, the presentation details the amended legislation enacted in September 2023,
    which enhances the mechanisms for safe cross-border transfers. This includes provisions
    for certification and equivalency recognition, which form part of Korea's adequacy system.
    Furthermore, the presentation explores potential future developments in cross-border
    transfer regulations, considering the increasing global demand for such transfers.

    Data Sovereignty in Vietnam: Legal Requirements, Enforcement Trends, and Global CBPRs Interactions

    Huyen-Minh Nguyen

    Senior Associate, BMVN International LLC, Vietnam

    Imposing data localization requirements is one way Vietnam asserts its sovereignty

    over data. The first data localization requirement was introduced in Vietnam under the

    Cybersecurity Law of 2016, which broadly applies to all offshore and onshore enterprises

    providing services on the Internet and processing certain data generated by and pertaining

    to service users in Vietnam. However, due to a lack of guidance from local authorities and

    the absence of a legal mechanism to enforce it, the requirement remained unenforceable

    for years. In 2022, the Government issued Decree 53 to clarify the data localization

    requirements under the Cybersecurity Law of 2016. Decree 53 significantly limits the cases

    in which companies are required to localize their data in Vietnam, with different sets of

    triggering conditions applying to offshore and onshore enterprises.

    This presentation discusses the requirements of the Cybersecurity Law of 2016, Decree

    53, and enforcement trends over the last few years. It also explores several new regulations

    that attempt to introduce additional cross-border data restrictions, such as the Data Law

    and the draft decree guiding the Law on Telecommunications, and how these may interact

    with or hinder the application of Global Cross-border Privacy Rules in Vietnam.

    Global Cross-Border Transfers: A Comparative Analysis of China, Hong Kong, and Beyond

    Dominic Edmondson

    Special Counsel, Baker McKenzie, Hong Kong

    This presentation focuses on the challenges of enabling cross-border data flows while

    complying with data sovereignty laws. First, the discussion covers how conflicting laws

    across countries can complicate data transfers and analyzes data localization requirements

    in various countries, such as China. Next, it analyzes the effectiveness of Global Cross

    Border Privacy Rules (GCBPRs) in facilitating cross-border data transfers and explores the

    challenges of achieving widespread adoption, using the APEC Cross-Border Privacy Rules

    (CBPR) system as an example. The presentation examines its role in enabling secure data

    flows while protecting privacy and provides examples of how this system has been used by

    participating countries.

  • Session 6: Fair Use of Data

    Moderator : Byungnam Lee (Senior Advisor, Kim & Chang, Republic of Korea)

    Exploring Utility and Privacy in Synthetic Data

    Joseph Hyuntae Kim

    Professor, Yonsei University, Department of Applied Statistics, Republic of Korea

    Synthetic data is becoming increasingly popular as a valuable resource for data-driven decision-making and machine learning, particularly in contexts where privacy and data
    security are paramount. However, creating synthetic data requires a careful balance
    between utility—ensuring the data remains useful—and privacy, aimed at safeguarding
    sensitive information from exposure. This presentation delves into these two key aspects of
    synthetic data and illustrates them through an auto insurance example. Additionally, insights
    from industry experience as the CEO of a synthetic data startup are shared to provide
    practical perspectives.

  • Guidelines for Using Pseudonymization for Unstructured Data in South Korea

    Hyunjoon Kwon

    Former Director, Personal Data Secure Usage Division, Korea Internet & Security Agency, Republic of Korea

    NaN

Contact Us

Barun ICT Research Center

50 Yonsei-ro, Seodaemun-gu, Seoul 03722, Korea

 

Tel : +82-2-2123-6694

 

Email : barunict@barunict.kr

    Cookie Use
    We use cookies to ensure a smooth browsing experience. By continuing we assume you accept the use of cookies.
    Learn More