Prof. Yucheng Zhang

1 ETHICAL CHALLENGES OF Al IN BUSINESS

1.1 AI and Ethics

1.1.1 Introduction to ethics of AI

AI has quietly penetrated every aspect of life (The invisible penetration and social impact of artificial intelligence) → This widespread penetration has raised ethical issues concerning human-machine collaboration, privacy, fairness, dignity, governance, and sustainability (Key ethical issues and question framework) → The root cause of these issues lies in AI’s unprecedented data and computing power, enabling it to both benefit society and potentially pose risks (Sources and dual effects of AI power) → Therefore, we need to rethink “what constitutes good technology” from an ethical perspective and clarify the importance of ethics in technological development (Necessity of ethical perspective and definition of ethics) → Finally, by introducing systematic ethical analysis methods, we can translate ethical principles into specific design and governance practices (Method and output)

1.1.1.1 The invisible penetration and social impact of artificial intelligence

  • Scope of penetration: business operations, social interactions, worldview formation, consumer decisions, entertainment preferences, and intimate relationships.

  • Overall assessment: AI systems are ‘quietly and seamlessly’ integrating into every corner of life and changing how we live.

1.1.1.2 Key ethical issues and question framework

Category Covered Topics
Human-Machine Collaboration and Decision Ethics Healthcare, justice, education, designer responsibility
Privacy, Autonomy, and Human Agency Public nature of privacy, challenges to autonomy
Fairness, Justice, and Trust Fairness, justice, trustworthy AI
Human Dignity and Bioethics Autonomous weapons, autonomous driving
Governance and Social Order Algorithmic governance, social control
Sustainability and Future Ethics AI and environment

1.1.1.3 Sources and dual effects of AI power

  • Sources of power: unprecedented data collection capabilities + immense computational power

  • Characteristics: ubiquitous, widely accepted, opaque in operation, and often ‘invisible’ in application

  • Dual effects: accelerating research, driving innovation, and improving life vs. bringing new concerns and challenges

  • Judgment: AI can solve old problems but will also generate new challenges

1.1.1.4 Necessity of ethical perspective and definition of ethics

  • What is ethics: a systematic reflection on moral issues (a branch of philosophy)

  • More than a technical issue: to make AI beneficial to humanity, ethical analysis must be central

  • Meaning of morality: the totality of individuals/groups expressing ‘what is good/just’ through opinions, decisions, and actions

1.1.1.5 Method and output

  • Method: Combine organizational examples and case studies

  • Framework: Introduce mainstream ethical frameworks to structure issues and propose possible solutions

1.1.2 Responsibility in the ethics of technology

AI ethics takes “responsibility” as its core starting point (The close connection between AI ethics and the concept of “responsibility”) → First, clarify the basic connotations and categories of responsibility (role/moral/professional, etc.) (Basic connotations and categories of responsibility) → Then distinguish the two types of responsibility paths in technology ethics (Two types of responsibility in technology ethics) → Passive responsibility: emphasize interpretability and attributability to define fault and undertake liability (Passive responsibility (looking backward / after the fact)) → Active responsibility: embed values and protections (e.g., VSD) in the front-end of design to prevent risks and promote positive outcomes (Active responsibility (looking forward / before the fact)) → Ultimately form forward-looking guidance for technological development, while facing the dual challenges of multiple stakeholders and high uncertainty (Guiding technological development: necessity and challenges)

1.1.2.1 The close connection between AI ethics and the concept of “responsibility”

  • Core issue: Once powerful technologies emerge, who is responsible?

  • Context shift: As AI gains autonomy and decision-making power, the concept of responsibility needs to be re-examined

1.1.2.2 Basic connotations and categories of responsibility

  • General definition: Being accountable for one’s own actions and their consequences

  • Role responsibility: Responsibility tied to specific roles (e.g., parent, contractual role)

  • Example: Moral responsibility vs. professional responsibility

    • Moral responsibility: Obligations, norms, and duties based on moral reasoning

    • Professional responsibility: Grounded in the professional role and must not exceed moral boundaries

1.1.2.3 Two types of responsibility in technology ethics

  • Passive responsibility

  • Active responsibility

1.1.2.4 Passive responsibility (looking backward / after the fact)

  • Applicable scenario: Accountability after an undesirable event

  • Example: In an autonomous driving collision — who is responsible: the driver, the designer, or the manufacturer?

  • Two key elements:

    • Explainability: The ability to explain decisions and actions

    • Blameworthiness: Being a legitimate target of blame, based on common criteria such as:

      • Misconduct: e.g., engineers failed to test the recognition system adequately

      • Causal contribution: acts or omissions that causally contributed to the event (e.g., premature system rollout)

      • Foreseeability: the harm was foreseeable (e.g., failure to test sufficiently caused preventable collisions)

      • Freedom of action: acting without coercion; even under corporate pressure, autonomy implies moral agency

1.1.2.5 Active responsibility (looking forward / before the fact)

  • Core idea: Prevent negative consequences and promote positive outcomes through proactive action

  • Significance: Especially crucial in the design stage of impactful technologies like AI

  • Methodological transformation: Integrating values and ethics into early design and embedding them into technology

  • Value-Sensitive Design (VSD): A systematic approach that treats moral values as design requirements

  • Example: Designing autonomous vehicles with privacy protection as a guiding principle

1.1.2.6 Guiding technological development: necessity and challenges

  • Necessity: Providing directional guidance for technological development

  • Two main challenges:

    • Diverse actors and stakeholders: Conflicting interests with moral relevance
    • Unpredictable processes: Particularly pronounced in emerging, powerful AI systems
  • Current situation: Multiple response strategies exist, but continuous improvement is required

1.2 Case Studies and Moral Frameworks

1.2.1 Ethics of autonomous vehicles

Taking “autonomous driving ethics” as the entry point, we first clarify the scope of discussion and core concept - Operational Autonomy, which refers to the system’s ability to perform tasks without human intervention (Topic positioning and definition of “Operational Autonomy”) → Then evaluate its moral pros and cons from a macro perspective: the improvement of safety and efficiency vs. the risks to privacy, freedom and responsibility (Moral advantages and disadvantages overview) → Focus on the key scenario - the “inevitable crash” issue, and reveal that AI faces real moral choices (Key scenario: inevitable crash) → Guide students to engage in ethical thinking through two typical dilemmas (e.g., “crash into the helmeted person or the unhelmeted person”) (Group discussion — two typical moral dilemmas) → Further analyze using different ethical frameworks: Consequentialism pursues minimal harm, while Deontology emphasizes moral principles and the boundary of rights (Analysis under different ethical frameworks) → Then discuss the complexity of responsibility attribution: the breakage in the responsibility chain among manufacturers, programmers and car owners (Challenges in assigning responsibility) → Furthermore, demonstrate how ethics influences policy-making, taking the principles of the German Ethics Commission as an example to transform theory into governance practice (From ethics to policy: The German case (2016-2017)) → Finally, summarize: Ethical analysis does not provide a single answer, but helps society strike a balance between innovation and responsibility (Summary).

1.2.1.1 Topic positioning and definition of “Operational Autonomy”

  • Discussion value: Use autonomous driving ethics as a lens to explore AI-related issues through multiple frameworks

  • Operational autonomy: The ability of a system to perform tasks without continuous human supervision, adapt to dynamic environments, and respond to unexpected situations; corresponds to levels of autonomous driving (L1–L5):

    • L1: Automation of basic functions such as acceleration and cruise control

    • L2–L4: Gradual enhancement of perception, decision-making, and control across scenarios

    • L5: Full autonomy under all conditions

1.2.1.2 Moral advantages and disadvantages overview

  • Pros: Improve traffic safety (more law-abiding, not fatigued, not intoxicated, not distracted); increase mobility and autonomy for vulnerable groups (e.g., the elderly); improve efficiency and reduce pollution

  • Cons / Risks: Potential violations of privacy and freedom; unclear allocation of responsibility

1.2.1.3 Key scenario: inevitable crash

  • Definition: Situations where a vehicle cannot completely avoid colliding with an obstacle

  • Ethical delegation: Beyond operational control, moral choices (e.g., collision angle, passenger protection priority) are delegated to the system

1.2.1.4 Group discussion — two typical moral dilemmas

  • Scenario 1: Two cyclists crossing the road (A wearing a helmet / B without); the vehicle can only hit one but may choose the target

  • Scenario 2: Two pedestrians crossing ahead; options — collide with pedestrians, or swerve to avoid them but risk severe passenger injury

1.2.1.5 Analysis under different ethical frameworks

  • Consequentialism (Utilitarianism)

    • Principle: Moral value is judged by consequences; choose the option that benefits the majority or minimizes harm

    • Application: In inevitable crashes, “fairly minimize risk and damage”

    • Scenario 1 inference: The system might be programmed to hit the helmeted cyclist (expecting less harm) — a controversial choice that seems to “punish” compliance

  • Deontological Ethics

    • Principle: The morality of an act depends on moral rules, rights, or duties, independent of outcomes

    • Application: For instance, a mother driving her child daily — prioritizing the child’s protection can be viewed as a separate moral duty

    • Tension: It may conflict with the rights of others and universal social rules (e.g., does “Passengers first” harm pedestrians’ rights?) Challenges in Assigning Responsibility

1.2.1.6 Challenges in assigning responsibility

  • Potential responsible parties: Manufacturers, design engineers, vehicle owners

  • Source of difficulty: No one has “complete and direct control” of the system — responsibility chains are fragmented or overlapping

  • Academic approaches: Multiple frameworks coexist, depending on ethical and institutional choices; the “policy vacuum” must be addressed

1.2.1.7 From ethics to policy: The German case (2016-2017)

  • Process: An ethics commission was established and issued 20 guiding principles

  • Content: Deontological ethics (prohibiting discrimination based on personal traits such as gender) + Consequentialism (aiming to minimize risk through design)

1.2.1.8 Summary

  • Ethics does not offer a checklist of absolute answers but serves as a tool for deepening understanding and structuring problems

  • In the context of autonomous driving, ethics is both necessary and useful — guiding responsible design and policymaking

1.2.2 Autonomous weapons and digitalization of conflicts

Exploring “Autonomous Weapon Systems (AWS)” from an ethical perspective: First, clarify its definition and categories — weapon platforms with autonomous navigation, identification, and attack capabilities (Definition and categories) → Then reveal three driving factors: operational efficiency, economic costs, and strategic advantages, which collectively promote its rapid development (Three drivers) → Further distinguish the key differences between remotely piloted drones and fully autonomous weapons, pointing out that the chain of responsibility tends to be vague in the latter (Key distinction: Remotely piloted drones vs. Autonomous weapons) → Subsequently, introduce international debates on use and regulation: proponents emphasize reducing human errors, while opponents warn of out-of-control risks and ethical hazards (Debate on use and regulation) → The use of autonomous weapons is not completely prohibited; the focus of the debate lies in how to restrict and manage them (Regulatory perspectives and policy positions) → At the theoretical level, deontology emphasizes human dignity and moral responsibility, advocating that “life-or-death decisions” should be reserved for humans (Deontological argument (Ethics of duty)) → However, real cases show the dilemma of the responsibility chain: engineers, commanders, and algorithm designers may all be partially responsible, but no one assumes full responsibility (The chain of responsibility dilemma (example)) → In contrast, consequentialists believe that if autonomous weapons can reduce casualties and mistaken killings, they may have positive value (Consequentialist argument) → Finally, the concept of “Meaningful Human Control” is proposed, advocating the retention of sufficient information and intervention opportunities to ensure the continuity of ethical and legal responsibilities (Meaningful human control)

1.2.2.1 Definition and categories

  • Autonomous weapon systems: Capable of navigation, investigation, and attack without continuous human supervision

  • Categories: Aerial systems, robotic sentries, loitering munitions, and other similar platforms

  • Technological evolution: Group intelligent weapons — large numbers of small, inexpensive platforms collaborating to execute complex missions

1.2.2.2 Three drivers

  • Operational: Unmanned or automated systems can act faster and operate in high-risk environments

  • Economic: Reduced need for human operators—–lowers operational costs

  • Strategic: Enhanced military capability and altered geopolitical dynamics

1.2.2.3 Key distinction: Remotely piloted drones vs. Autonomous weapons

  • Remotely piloted drones: Human-operated, maintaining a command-and-control chain that allows clear attribution of responsibility

  • Autonomous weapons: Systems capable of independently selecting and engaging targets — human control is diminished, and responsibility chains become blurred

1.2.2.4 Debate on use and regulation

  • Support (Ronald Arkin): Believes that properly designed autonomous weapons could reduce human battlefield misconduct and improve adherence to international law

  • Opposition (Noel Sharkey): Argues that such systems lack verifiable safeguards; unpredictable machine behavior could exceed human oversight

1.2.2.5 Regulatory perspectives and policy positions

  • 2021/03: The US AI National Security Commission: Does Not Support Complete Ban on Autonomous Weapon Systems Capable of AI

  • Debate focus: Whether and how to restrict or regulate autonomous weapons

1.2.2.6 Deontological argument (Ethics of duty)

  • Basic obligation: Following the principle of human dignity (THL); emphasize the moral responsibility chain and uphold respect for humanity

  • Dignity requires: The decision to take life should remain a uniquely human moral act

  • Proportionality principle: Only attack targets that bring about a military advantage; it is necessary to distinguish between those who are still in active combat and those who have surrendered.

  • Challenge: How can AI identify “intentions of surrender”? How can it avoid making incorrect perceptions?

1.2.2.7 The chain of responsibility dilemma (example)

  • Case of Mistakenly Attacking a School Bus: There are numerous potential responsible parties (engineers, managers, procurement decision-makers, consultants, manual writers, commanders, operating soldiers…)

  • The essence of the problem: No one has a “significant causal contribution”, leading to a gap in attribution

1.2.2.8 Consequentialist argument

  • Supporters argue: Replacing human soldiers could reduce casualties and psychological harm; strict launch conditions can be set

  • Critics/warn:

    • Risk of escalation beyond human reaction time
    • Security vulnerabilities and malicious misuse
    • Uncontrollable escalation and unpredictable behavior of autonomous systems

1.2.2.9 Meaningful human control

  • Background of proposal: Rethinking responsibility in uncertain military environments

  • Connotation: “Meaningful” means having sufficient information and time for human intervention

  • Current status: How to operationalize it remains controversial, but the focus has shifted from “defining autonomy” to “establishing control and supervision”

1.2.3 Digital medicine

Regarding the ethics of digital healthcare, we first point out the new opportunities and risks brought by AI in the medical field: Algorithms promote precision diagnosis and public health, but also bring challenges related to privacy, bias, and trust (Emerging opportunities and challenges) → Then sort out the four major application scenarios of digital health: health monitoring, auxiliary diagnosis, treatment decision prediction, and public health governance, revealing that AI runs through the entire medical process (Four types of applications in digital health/digital medicine) → Focus on AI in diagnosis, emphasizing that its potential lies in improving accuracy, reducing the burden on doctors, and expanding preventive coverage (AI in diagnosis: potential and objectives) → However, there are three core ethical focuses at the same time: privacy and data protection, algorithm representativeness and fairness, transparency and interpretability (Three major ethical focuses) → Therefore, it is necessary to rely on regulations and international guidelines as guarantees. Currently, initiatives from GDPR to OECD/WHO all emphasize human control, privacy minimization, and responsible application (The current situation of regulations and guidelines) → Finally, summarize: Digital healthcare is reshaping medical practice and the doctor-patient relationship. In the future, we should adhere to the principle that “medicine is a social practice” and maintain an ethical balance between technological innovation and patients’ rights (Summary and outlook)

1.2.3.1 Emerging opportunities and challenges

  • IT/AI-enabled diagnosis and treatment (data-driven medical decision-making)

  • Moral Challenges: Global Public Health Policies, Transparency and Credibility, Privacy Protection

1.2.3.2 Four types of applications in digital health/digital medicine

  • Health monitoring (wearable devices / mobile apps / home-based equipment)

  • Auxiliary diagnosis (image and clinical data analysis)

  • Predictive algorithm for treatment decisions

  • Public health applications (epidemiological monitoring, policy tools)

1.2.3.3 AI in diagnosis: potential and objectives

  • Machine learning is highly effective in clinical/imaging analysis, and sometimes outperforms humans.

  • Objective: Improve accuracy, reduce burden on specialists, and expand prevention coverage

1.2.3.4 Three major ethical focuses

  • Privacy and Data Protection

    • Based on large-scale collection/storage/analysis/distribution

    • Informed consent is challenged in the era of big data (as algorithms can infer undisclosed information)

  • The applicability/representativeness of algorithms in medicine

    • Training requires representative data

    • Biomedical data is biased towards white men (in North America/Europe) and harms disadvantaged/underrepresented groups, violating justice

  • Transparency and Explainability

    • ML has a lot of “internal correlation” learning, and the parameters are not explainable (it is a “black box”)

    • Challenges Arise in Determining Doctoral Duties, Patient Informed Consent, and Legal and Ethical Responsibilities

1.2.3.5 The current situation of regulations and guidelines

  • GDPR (2018): Granting the right to refuse fully automated decision-making

  • The current guidelines are mostly abstract; the OECD/WHO has proposed more specific suggestions:

    • Protecting privacy and minimizing negative impacts

    • Digital solutions support rather than replace clinical practice

    • Human control must always be ensured

1.2.3.6 Summary and outlook

  • Digital medicine is reshaping medical practice and medical perspectives

  • The tension between prediction and explanation (correlation vs. causation) remains unresolved

  • The future should be guided by the principle that “medicine is a shared social practice”, and patient rights (both clinical and ethical) should be protected in two dimensions

1.2.4 Sustainability and environmental impact

Turning to the ethics of AI and the environment, this section starts with the “duality” of AI and the environment, pointing out that AI can not only promote sustainable development but also may increase resource consumption and environmental burdens (Overview: The duality of AI and the environment) → Then reveal the contradictions within the sustainability framework: the digital revolution supports the UN SDGs, but harms vulnerable groups due to insufficient data representativeness (The framework of sustainability and its contradictions) → Subsequently, explore the measurement dilemma and the metaphorical trap of “cloud computing” — the cloud is not intangible, but a physical infrastructure that consumes enormous energy and land resources (The measurement dilemma and the metaphorical trap of “The cloud”) → Therefore, ethical evaluation should go beyond quantitative calculation and shift to value measurement, integrating multi-dimensional consequences such as environment, education, and health (Ethical assessment: Beyond quantification to value) → Further discuss distributive justice: the unequal flow of data centers and electronic waste makes poor areas bear environmental and health costs (Justice and distribution) → Facing enterprises’ “green AI” commitments and carbon neutrality goals, we also need to be alert to the risk of “greenwashing” and advocate for transparent and verifiable sustainable actions (Initiatives, Goals, and ‘Greenwashing’ risks) → Finally, summarize: AI sustainability is a systematic issue involving technology, ethics, and politics, requiring multi-level responsibilities and global cooperation to achieve truly green intelligence (Summary)

1.2.4.1 Overview: The duality of AI and the environment

  • Opportunities: Smart grid, satellite data monitoring of deforestation, contributing to SDGs, etc.

  • Risk: The processing of massive data and the disposal of equipment cause environmental and social impacts.

1.2.4.2 The framework of sustainability and its contradictions

  • Core: The Global Connection between the Well-being of Current/Future Generations and Ecosystems

  • 2030 Sustainable Development Goals (SDGs):The digital revolution can play a crucial role, but the lack of representative data has disadvantaged vulnerable groups. This reflects an inherent contradiction.

1.2.4.3 The measurement dilemma and the metaphorical trap of “The cloud”

  • High energy consumption and significant emissions; “The Cloud” requires physical infrastructure (land occupation, energy consumption, emissions)

  • The energy consumption of Alil is significantly reduced (for instance, the training of large-scale NLP models can exceed several times the emissions of an entire vehicle throughout its lifecycle).

1.2.4.4 Ethical assessment: Beyond quantification to value

  • Consequentialist approach: Assess the costs and benefits (including comprehensive consequences such as environmental, educational, health, and energy consumption aspects)

  • Limitations:

    • Value cannot be globally monetized
    • It is difficult to quantify the impact
    • The definition of “return/risk” is not easily unified
  • Direct and Indirect Influences & Relative Pollution

  • It is necessary to assess absolute and relative pollution (compared with social/environmental benefits)

  • Note the indirect effect: For instance, remote working reduces the need for company vehicles but increases the demand for personal vehicles at home.

1.2.4.5 Justice and distribution

  • Resource Utilization and Community Impact: The Effects of Data Centers on Land/Energy and Local Political and Social Structures

  • Disposal of justice; Electronic waste mostly flows to poor countries, with insufficient regulation enforcement - health/environmental hazards

  • Governance requirements: Decisions related to urban space must be transparent, democratic, and subject to public supervision.

1.2.4.6 Initiatives, Goals, and “Greenwashing” risks

  • One of the trustworthy AI standards: Environmental Sustainability (High-Level Expert Group on AI of the European Union)

  • Practical challenges: Difficulty in measurement and implementation; most goals are voluntary - with limited impact

  • Corporate carbon neutrality/net zero commitments often lack detailed implementation plans - there is a risk of “greenwashing”

1.2.4.7 Summary

  • Improving digital pollution and enhancing the sustainability of AI are comprehensive issues involving technology, ethics, and politics

  • It is necessary to establish common moral values and implement multi-level responsibilities (individuals, countries, and the international community).

  • It is necessary to adopt a global perspective and go beyond the simplistic narrative that “energy consumption can be reduced simply by optimizing technology”

1.2.5 Cybersecurity and AI

Moving on to the ethics of cybersecurity, this section first defines basic concepts and “the three elements” — Confidentiality, Integrity, and Availability, which constitute the core pillars of information security (Basic definitions and “The three elements of CIA”) → Then point out that security issues are not only a technical bottom line but also ethical issues, because protecting data means protecting human rights and social order (The connection between safety and ethics) → Subsequently, discuss responsibility and liability: if risks arise due to failure to take protective measures, corresponding moral and legal responsibilities should be borne (Responsibility and liability) → Further reveal value trade-offs: strengthening integrity and monitoring may sacrifice confidentiality, while enhancing privacy may reduce availability (Value trade-offs (the tension between the CIA)) → At the social level, this trade-off extends to the conflict between the government’s security power and citizens’ privacy and freedom, such as ethical reflections after anti-terrorism monitoring (Social-level trade-off: Government power and privacy) → At the same time, it raises the moral controversy of “ethical hackers” and vulnerability disclosure: how to achieve legitimacy between revealing risks and violating privacy (“Ethical hackers” and disclosure ethics) → In the AI context, it is necessary to combine cybersecurity with the four principles of trustworthy AI to ensure the parallel implementation of harm prevention, respect for autonomy, fairness, and interpretability (Cybersecurity in the AI context and the principles of trustworthy AI) → Finally, summarize: Cybersecurity is both a technical issue and an ethical issue. A balance should be achieved between the three CIA elements and human-centric values to build a responsible and trustworthy digital society (1.2.5.8 Summary).

1.2.5.1 Basic definitions and “The three elements of CIA”

  • Confidentiality:Access is granted only to the authorized person

  • Integrity:Prevent unauthorized tampering

  • Availability:The authorizer can access it at any time and it will not be denied

1.2.5.2 The connection between safety and ethics

  • Bottom line: Unauthorized access in itself is inappropriate.

  • On a deeper level: Safety measures will shape technology and social order

1.2.5.3 Responsibility and liability

  • Who is responsible for the security vulnerability? If no protective measures were taken, should they bear some responsibility?

  • Ethical Focus: The Responsibility of Actors for the Risks of Others

1.2.5.4 Value trade-offs (the tension between the CIA)

  • The solution for enhancing integrity (global monitoring) may sacrifice confidentiality

  • At the same time, a solution that ensures high confidentiality and integrity may reduce availability (complex verification)

1.2.5.5 Social-level trade-off: Government power and privacy

  • After 9/11, the “USA Patriot Act” granted broader powers for electronic surveillance

  • Lesson: Safety should not take precedence over other values; each case requires a detailed analysis and weighing

1.2.5.6 “Ethical hackers” and disclosure ethics

  • Authorization penetration can be used in security testing, but the method of disclosure raises ethical concerns

  • Four types of arguments and counterarguments:

    • Information should be free: In conflict with privacy/encouragement
    • Identifying potential risks: Is the approach appropriate (analogy: “police officer breaking into a house”)
    • No alteration = No harm: Unauthorized access is inherently an infringement
    • Curbing Big Brother: Just because the purpose is legitimate doesn’t mean the means must also be legitimate

1.2.5.7 Cybersecurity in the AI context and the principles of trustworthy AI

  • Preventing harm: If the well-being and rights cannot be protected, the plan is ethically inappropriate; it needs to be balanced with usability

  • Respecting human autonomy: Excessive reliance on autonomous AI leads to degradation of human skills and impairment of recognition and decision-making abilities

  • Fairness: Automated intrusion detection may undermine fairness due to bias or errors

  • Explainability: The results of black-box ML are difficult to explain, thus meaningful human control may fail

1.2.5.8 Summary

  • Cybersecurity is both a technical issue and an ethical issue

  • In the context of Al, the three elements of CIA must be integrated with the principles of trustworthy AI, and a detailed consideration and design should be carried out based on specific circumstances

1.3 Problems and Challenges

1.3.1 Human decisions and AI

Shifting focus to AI’s role in human decision-making, AI has begun to replace humans in various human decision-making scenarios (Application scenarios and method evolution) → Its operation follows an automated reasoning process of “learning-modeling-prediction” (The learning–modeling–prediction process of machine learning) → This process brings efficiency in recommendation systems and high-risk decision-making, but also triggers ethical controversies (Recommendations and controversial applications) → The fundamental issue lies in the fact that algorithms are not neutral; instead, they embed biases and limitations (‘Algorithmic positive bias’ and model limitations) → Therefore, we need to establish normative principles to ensure that AI decisions conform to social, legal, and ethical standards (Normative conclusion and transition)

1.3.1.1 Application scenarios and method evolution

  • Application fields: Automated recruitment, business crime prediction tools, medical diagnosis and treatment recommendations, etc

  • Automated Management Decision System: Processes data related to employees, integrating human and machine decision-making

  • Historical comparison

    • Traditional (top-down): Based on employee knowledge and rules
    • Contemporary (bottom-up): Machine learning derives rules from data management experience and infers new knowledge

1.3.1.2 The “learning - modeling - prediction” process of machine learning

  • The core of learning: Adapting to new situations from experience, without explicit instructions

  • Model Construction: By training with large-scale data, a mathematical model can be derived, where the input data will be mapped until the most likely answer emerges

  • Classification as a learning paradigm

    • Mechanism: Predict the category of an unknown object based on a known category set
    • Analogy: The learning process of children in a zoo, which involves “observing examples - correcting mistakes - improving”
  • Decisions are classified: Many real-life decisions can be expressed as classification problems

    • Key prerequisite: Adequate sample size + Good representativeness

1.3.1.3 Recommendations and controversial applications

  • Simple decision-making: Recommendation system (for recruitment interviews, etc.), based on the “similarity principle” to learn user preferences

  • Controversial decision-making case

    • Performance evaluation: “Mathematical Destruction Weapon” assesses teachers based on standardized tests, resulting in the dismissal of excellent teachers
    • Recruitment: Automation of Resume Screening and Interview Evaluation

1.3.1.4 “Algorithmic positive bias” and model limitations

  • Myth: Mathematics/Statistics are not equivalent to neutrality and objectivity

  • Reality: A model is an approximation of reality, involving assumptions/simplifications/errors

  • Representativeness and Bias: If the generation mechanism of the training data is flawed, the model will mistake spurious correlations for genuine associations

  • Overfitting and failure of generalization

    • Phenomenon: The training set performs well, but the test/new scenarios perform poorly

    • Metaphor: Focusing only on depicting a small part of the city map, resulting in the rest of the area being useless

1.3.1.5 Normative conclusion and transition

  • The algorithm should be accurate, reliable, secure, and in line with legal/ethical/social norms

  • To lead into the next part: In the next lecture, we will summarize the common problems and the corresponding solutions

1.3.2 Value of privacy in organization – Part one

Now focusing on privacy ethics in the AI era, this section starts with privacy issues in the age of AI. First, it explains why privacy has become a core ethical focus against the backdrop of growing intelligence and the popularization of surveillance (Why discuss privacy in the era of AI) → Then distinguish the two connotations of privacy: “freedom from interference” at the decision-making level and “data access control” at the information level (The two connotations of privacy) → Further draw on Deborah Johnson’s ideas to put forward key reflections: What is the value of privacy? What will we lose if we lose privacy? And what kind of beings will individuals be shaped into in a society under constant surveillance (Key questions (Inspired by Deborah Johnson)) → Subsequently, analyze the characteristics of information flow in contemporary organizations: large-scale, diverse in type, easy to spread, and difficult to delete. Misinformation will be amplified rapidly, leading to systematic privacy risks (Characteristics of information flow in contemporary organizations) → Then through common arguments against privacy (such as “those with nothing to hide don’t need privacy” or “data sharing benefits everyone”) and their rebuttals, explain that privacy is not only about secrets, but also about fairness, dignity, and autonomy (Common arguments against privacy and their rebuttals) → Finally, summarize: In an information society dominated by AI, privacy is the cornerstone of safeguarding individual freedom and social justice (Summary)

1.3.2.1 Why discuss privacy in the era of AI

  • AI tools make multi-form monitoring possible - directly affecting privacy
  • In the information society, organizations are often described as “monitoring organizations”

1.3.2.2 The two connotations of privacy

  • Decision/Constitutional-level Privacy: The freedom to make decisions without interference in private matters

  • Privacy at the information/infringement level: Control over “access to one’s own information” (such as social disclosure)

1.3.2.3 Key questions (Inspired by Deborah Johnson)

  • What is the value of privacy? What will be lost if privacy is lost?

  • In a surveillance society, how will we be shaped as individuals?

1.3.2.4 Characteristics of information flow in contemporary organizations

  • Scale: The electronic records of employees make it easier to collect/store/share, which leads to an explosive increase in the volume of information

  • Information type: Includes new types of personal data such as transaction generation information

  • Distribution: Internal organizational data can be bought/sold/misappropriated/stealed

  • Persistence: Insufficient deletion incentives in digital storage

  • Error amplification: Errors within the organization spread rapidly and widely, making it difficult to fully trace them

1.3.2.5 Common arguments against privacy and their rebuttals

  • “Only departments with secrets need privacy.”

    • Counterargument: Incorrect data can also be fatal; using inappropriate information for decisions such as recruitment is unfair
  • The employees have given up their privacy, resulting in it becoming worthless

    • Counterargument: Just because the disclosure was made without informing employees or was deliberately chosen (in exchange for data for the sake of benefits), it does not mean that privacy was not valued
  • Collecting data is beneficial for both organizations and individuals

    • Counterargument: The organization may shape users rather than serve them, and inappropriate information is used in decision-making

1.3.2.6 Summary

  • Through examples and counterarguments, it is clearly demonstrated that privacy is of utmost importance in the information society

1.3.3 Value of privacy in organizations – Part two

This section continues to deepen the discussion on privacy ethics, pointing out that privacy is inseparable from fundamental values such as autonomy, equality, and democracy (The interweaving of privacy and fundamental values) → Secondly, from the individual perspective, privacy is the “individual well-being” of employees. It ensures the diversity of relationships between people and enables individuals to freely express themselves in different scenarios (Privacy as an “individual well-being” of employees) → From the organizational perspective, privacy is also a kind of “social well-being”. Priscilla Regan points out that in organizations with a “panopticon-style” surveillance system, continuous observation will weaken autonomy and democratic participation, leading to institutional obedience and the loss of critical ability (Privacy as “social well-being” within an organization (Priscilla Regan)) → Therefore, multi-level response strategies must be adopted: establish a multiple responsibility system involving individuals, organizations, and regulators through industry ethical guidelines, transparent policies, and technical approaches such as “Privacy by Design” (Response strategies and multiple responsibilities) → Finally, summarize: Privacy is not only an individual right, but also a common value for maintaining organizational health and social democracy, which should be jointly protected through the coordination of systems and technologies (Summary)

1.3.3.1 The interweaving of privacy and fundamental values

  • Privacy is interwoven with values such as autonomy, equality, and democracy

1.3.3.2 Privacy as an “individual well-being” of employees

  • Rachels (1975):Privacy maintains the diversity of relationships (asymmetric information shapes different types of relationships)

  • Problem: Often finds itself in a disadvantaged position when balancing against social welfare (such as in the case of the “Patriot Act”)

1.3.3.3 Privacy as “social well-being” within an organization (Priscilla Regan)

  • Panoramic Prison Metaphor (Panopticon)

    • As long as there is a possibility of being monitored, an individual will self-discipline to conform to the “invisible norms”
  • Two major concerns

    • Limited autonomy: Lack of privacy restricts free choice
    • Inconsistent standards: The criteria for evaluation and constraints are invisible and non-negotiable
  • Democratic Shock: Invisible and Non-Consultative Norms Weaken Citizen Criticism and Freedom, Leading to the Shaking of the Democratic Foundation

1.3.3.4 Response strategies and multiple responsibilities

  • Professional associations: Ethical guidelines and industry standards, promoting fair handling

  • Regulation and Policy: Shift from default consent / only opt-out to active opt-in; Transparent Policy

  • Technical Design: Privacy by Design, determining the minimization of data collection/transmission from the very beginning

1.3.3.5 Summary

  • Not only technicians, but everyone should realize the fundamental significance of privacy to life and democracy

1.3.4 Moralization of technologies

This section explores “the moralization of technology”. First, it points out that technology itself is not neutral; instead, it carries the social power structure and value orientation, and design is politics (Technology embodies morality and politics) → Then put forward the concept of “technological mediation”: technology is not a mere tool, but a medium that affects human behavior, perception, and judgment (Technological mediation) → This leads to the connotation of “technological moralization”, that is, actively embedding moral goals in the design stage so that technology can guide humans to behave in a more ethical manner (The connotation of technological moralization) → However, this process has two major underlying problems: first, the result of the interaction between technology and humans is highly uncertain; second, the values in the program are often invisible and difficult to perceive (Two major underlying problems) → Examples show that although moralized technologies such as drunk driving lock systems and speed bumps are well-intentioned, they often encounter three types of backlash — restricted freedom, degraded moral ability, and lack of democratic legitimacy in value embedding (Examples of moralized technologies and three types of backlash) → Therefore, it is necessary to explore how to design ethical AI in a democratic way: emphasize transparency, public participation, and value consensus to prevent negative effects caused by “unexpected mediation” (How to design ethical AI in a democratic manner) → At the same time, clarify the division of labor among multiple parties: designers need to predict technical consequences, users should understand the embedded values, and policy makers need to shift from ex post supervision to ex ante co-creation (Roles and division of labor) → Finally, summarize: Moralized design is a process jointly shaped by humans and technology. Only through open, interpretable, and collaborative ethical design can responsible AI innovation be achieved (Summary)

1.3.4.1 Technology embodies morality and politics

  • Winner:The “racist overpass” designed by Robert Moses - the low bridge with height restrictions to exclude buses, determined who could access the beach

  • Latour:Artificial objects are the bearers of morality (speed bumps “make the moral decision for you” to slow down)

1.3.4.2 Technological mediation

  • The technology is not a neutral method, but rather a mediator that influences actions and perceptions

  • Case: Obstetric Ultrasound

    • Separate the fetus from the maternal representation and consider it as an “independent living entity”

    • Incorporating pregnancy into medical guidelines: The fetus becomes a “potential patient”, and defects become “preventable suffering”

1.3.4.3 The connotation of technological moralization

  • Definition: The deliberate development of technologies for the purpose of shaping moral actions or decisions

  • The relationship with proactive responsibility: Preempting negative outcomes in advance and facilitating positive effects

  • AI relevance: Al has a significant impact on humans and society. The proactive approach is particularly crucial during the design stage

1.3.4.4 Two major underlying problems

  • Uncertainty: The introduction of technology and human-machine interaction is uncertain. AI is often regarded as an “experimental technology”

  • The three types of invisibility problems (Jim Moor)

    • Abuse of Invisibility: Concealed Use of Computers for Evil
    • The invisibility of programming value: Value is implicitly embedded in the code
    • The invisibility of complex computations: Beyond the scale of computations that exceed human understand

1.3.4.5 Examples of moralized technologies and three types of backlash

  • Alcohol lock for cars (locks the car when alcohol is detected) - Even without concerns over accuracy and privacy, the sales resistance remains high

  • The three main reasons why people are opposed to technologically moralized systems

    • Freedom under threat: Worried about being dominated by technology and having one’s dignity undermined
    • Degradation of moral capacity: By entrusting moral decisions to technology, it leads to the atrophy of autonomous moral judgment
    • Democratic Legitimacy: Embedded Values/Choosing Non-Democratic Outcomes Differ from Legal Restrictions

1.3.4.6 How to design ethical AI in a democratic manner

  • How to Democratize Ethical Design: Transparency, public discourse in the process of value embedding, and avoiding “unexpected mediation”

  • Counter example: Rebound effect of energy-efficient light bulbs (installed in more locations, leading to increased total energy consumption)

  • Insight: Design = materialization of morality; there is a need to re-examine the moral responsibility of designers

1.3.4.7 Roles and division of labor

  • Designers: Should not only focus on “desired morality” but also need to predict mediation effects

  • Users/Citizens: Should know who is embedding which values

  • Policymakers: Should not only conduct ex-post regulation but also proactively participate in co-shaping technology

1.3.4.8 Summary

  • Moral decisions are the result of the joint effort of human and object

  • The proactive, democratic and explainable moralized design is the key to responsible innovation in the AI era

1.4 Trustworthy AI

1.4.1 The trustworthy AI at work

Focusing on “Trustworthy AI”, this section first introduces its background and conceptual positioning — derived from Europe’s Ethical Guidelines for Trustworthy AI, emphasizing that AI should not only be technically reliable but also trustworthy in ethical and social dimensions (Background and conceptual orientation) → Then summarize the four core ethical principles: Human Agency and Autonomy, Prevention of Harm, Fairness, and Interpretability, which form the value foundation for evaluating Trustworthy AI (Four ethical principles (European guidelines)) → Seven implementable ethical requirements are derived from these four principles, including Human Oversight, Technical Robustness and Safety, Privacy and Data Governance, Transparency, Diversity and Fairness, Social and Environmental Well-being, and Accountability Systems, providing an implementation framework for AI governance (Seven implementable ethical requirements (Derived from the four principles)) → Finally, point out the key insight: Although technical performance (such as accuracy and robustness) is a necessary condition, it is not sufficient to ensure AI’s trustworthiness; true trust also depends on the implementation of ethical principles and the balance of social responsibilities (Key insight: Trustworthiness is not solely determined by technical indicators)

1.4.1.1 Background and conceptual orientation

  • Issue Prominence:Trustworthy AI is a focal topic in both academic and public discourse. It underpins the Ethics Guidelines for Trustworthy AI in Europe(2019) , which further serves as the ethical foundation for the AI Act (a legally binding instrument that provides unified supervision)

  • Difference between “Trustworthiness” and Human Trustworthiness:

    • Human Trustworthiness: Centered on care, good intentions, keeping commitments, etc
    • System Trustworthiness: (i) Output performance (accuracy, robustness) + (ii) Adherence to ethical principles (responsible design and deployment)

1.4.1.2 Four ethical principles (European guidelines)

  • Human agency & autonomy

  • Prevention of harm

  • Fairness

  • Explainability

1.4.1.3 Seven implementable ethical requirements (Derived from the four principles)

  • Human Agency and Oversight

    • AI should not pose risks to fundamental rights
    • AI can assist in decision-making, while users should retain autonomous choices
    • Provide human-in / on / over-the-loop oversight mechanisms
  • Technical Robustness and Safety

    • Requirements: Accuracy (correctness of prediction/classification/recommendation), repeatability, cross-condition stability
    • Resistance to attacks (security protection), redundancy and contingency mechanisms
  • Privacy and Data Governance

    • Individual data and inferred data used for training and inference
    • Data quality / completeness
    • Access control / minimization
    • Data access agreements (specifying under what conditions access is permitted)
  • Transparency (to be elaborated later): Transparency in data, systems, and business models

  • Diversity, Non-Discrimination, and Fairness (to be elaborated later): Focus on bias identification: mitigation and assessment

  • Social and environmental well-being

    • Continuously assess the impact on society/democracy (especially AI within social platforms)
    • Monitoring environmental sustainability (energy consumption, water consumption, material usage)
    • Intervention when necessary
  • Accountability system

    • Internal and external audits can assess the system; incidents/ negative impacts should be publicly reported

    • Clarify the responsible entities and pay attention to vulnerable groups; Improve the remedial mechanisms

1.4.1.4 Key insight: Trustworthiness is solely determined by technical indicators

  • Technical stability/safety is a necessary but insufficient condition

  • Large Language Model (LLM/ChatGPT) Challenges:

    • May contain harmful biases and disclose sensitive information
    • Opaque (algorithm and data sources, proprietary and closed-source)
    • Significant environmental impact: High energy consumption and carbon footprint during training/inferring; High water consumption in data centers; Hardware relies on rare earth/mineral mining; Ecological damage
  • Conclusion: The trustworthy AI framework enables us to balance ethical and social dimensions during the design and deployment stages, rather than focusing solely on performance

1.4.2 Algorithm fairness at work

Focusing on “Algorithmic Bias and Fairness”, this section first reveals the facts and causes of bias in AI: Algorithms are not neutral; instead, they inherit and amplify social inequalities and structural biases embedded in training data (Facts and causes of bias in AI) → Then classify the types and sources of bias, including data bias (incomplete samples, cultural exclusivity) and label bias (incorrect annotation, crowdsourcing errors) (Types and sources of bias) → Next, elaborate on the multiple definitions and paradigms of fairness: From the perspective of distributive justice, it includes both “equality” (equal opportunities) and “impartiality” (compensation based on differences); at the technical level, it is further divided into two dimensions: “individual fairness” and “group fairness” (Definition and paradigms of fairness) → Further explain the three technical links of fairness intervention: the data preprocessing stage (cleaning and resampling), the model training stage (introducing fairness constraints or regularization), and the result post-processing stage (threshold adjustment or re-ranking), which together form a full-process path for algorithmic fairness (Involvement stage: Three types of technical approaches) → Finally, summarize: AI fairness is not only a technical challenge but also a reflection of social values; only through the parallel implementation of institutional governance and technical means can a trustworthy AI system that balances efficiency and justice be established (Summary)

1.4.2.1 Facts and causes of bias in AI

  • Technology is not neutral: AI often inherits and amplifies biases in data and algorithms

  • Example: Recruitment training data reflects historical reality (most managers are male) → The model favors men, exacerbating gender bias

  • Statistical Perspective

    • Bias: Systematic overestimation or underestimation of the true value
    • Overfitting: Overfitting to training data, leading to poor generalization
    • Underfitting: The model is too simple, leading to insufficient performance
    • In general: Systematic deviation from the “correctness criterion”
  • “Biased Algorithm”: Systematic error in automated processes, producing unfair results for a certain group (e.g., systematically excluding applicants under 25)

1.4.2.2 Types and sources of bias

  • Data Bias

    • Incomplete/inaccurate/collected over too short a period/systematic exclusion
    • Cultural bias infiltration (e.g., generative models often portray “successful people” as white males)
  • Label Bias:

    • Labeling errors/incompleteness (e.g., cats labeled as tigers; Huskies in snowy mountains labeled as wolves)
    • Crowdsourced labeling introduces human bias, leading to errors migrating to models and applications
  • Amplification Effect When Targeting Humans: In arease Face Recognition, spreads across domains like advertising, security, soft biometrics, human-computer interaction, and healthcare; classification + inference decisions impact life outcomes

1.4.2.3 Definition and paradigms of fairness

  • Fairness = non-discrimination (intuitive at the top level), but there is no single definition; multiple frameworks coexist

  • Distribution Principles

    • Equality:Providing the same resources/opportunities
    • Equity:Providing resources/opportunities based on differences to achieve goals
  • Object Levels

    • Individual Fairness: Similar individuals should be treated equally (e.g., candidates with similar qualifications should receive similar evaluations in recruitment)
    • Group Fairness: Groups defined by protected attributes should receive similar opportunities/treatment (e.g., in organ allocation, different ethnic groups should have comparable priority when needs are equal)

1.4.2.4 Involvement stage: Three types of technical approaches

  • Preprocessing Fairness (Data Layer): Cleaning/Resampling/De-sensitization/Causal Re-weighting, etc.; Example: Before training the recruitment model, remove or correct the gender signals that cause discrimination

  • Fairness in the processing (model layer): During training, fairness constraints/regularization are introduced to achieve multi-objective optimization of accuracy and fairness

  • Post-processing fairness (result layer): Adjusting thresholds, reordering, or setting quotas for the output; for example: If the number of female hires is significantly lower than that of males, the prediction will increase the selection rate of females to the target range.

1.4.2.5 Summary

  • Technology reflects culture: AI reflects social structures and may exacerbate inequality

  • Feasible path: Establish a fair model (indicators + processes) compatible with business and law, systematically reduce discrimination, and provide a more just society

1.4.3 Opacity and explainability

This section explores the ethical implications of “explainability and opacity”. First, it clarifies the core concept: Explainability refers to whether the process of a model from input to output is understandable, and its ethical significance lies in ensuring traceable responsibility, identifiable bias, and establishable trust (Concepts and ethical importance) → Further clarify its value positioning:Explainability is an “instrumental value”, which aims to serve higher-level intrinsic values such as safety, fairness, and autonomy, rather than being an end in itself (Value positioning: Instrumental value, not instrument itself) → Then demonstrate the trade-offs between explainability and other values through contextual cases, such as excessive information in autonomous driving may interfere with safety, and “black-box models” in medical triage, although accurate, undermine informed consent—indicating that the degree of interpretation depends on risks and contexts (Examples of contextual trade-offs) → Finally, put forward a practical orientation: The “dynamic balance” between explainability and performance should be achieved based on risk levels and usage scenarios. By combining mechanisms such as auditability, traceability, and human-in-the-loop, AI can be both understandable and effective (Practical orientation: How to ‘strike a balance’)

1.4.3.1 Concepts and ethical importance

  • Explainability: Whether the process and key factors from input to output can be explained.

  • Black Box Problem: The input-output pathway of complex models (e.g., deep neural networks/LLMs) is opaque, even experts struggle to interpret it

  • Why It Matters:

    • Legal/Ethical Risks: High-stakes decisions require justifiability and accountability; black boxes undermine responsibility attribution and error correction

    • Bias Hard to Detect: Opacity combined with historical biases makes errors difficult to locate, endangering the safety of underrepresented groups

    • Trust Eroded: Unexplainable processes are perceived as untrustworthy, triggering suspicion and resistance

1.4.3.2 Value positioning: Instrumental value, not instrument itself

  • Instrumental Value: Helps achieve intrinsic values (e.g., autonomy, safety, fairness, trustworthiness)

  • Inevitability of Trade-offs: Explainability needs to be contextually balanced with other values. One cannot neglect intrinsic values like safety/effectiveness “in the name of explainability”

1.4.3.3 Examples of contextual trade-offs

  • Autonomous Driving

    • Increasing Explainability: Displaying environmental classification details enhances user autonomy and intervention capabilities
    • Risks: Excessive details cause information overload or misintervention, posing safety threats. In this scenario, limited opacity may better align with safety values
  • Hospital Triage (ICU Allocation)

    • Argument A: If a black-box AI is significantly more accurate, it benefits benevolence and efficiency
    • Argument B: Lack of explanation hinders quality verification and the principle of non-maleficence, especially for minority/vulnerable groups
    • Conclusion: Dual decision-making (technical feasibility + ethical evaluation) is needed to select the necessary level of explainability (intra-/inter-model explanation, auditability, interrogability))

1.4.3.4 Practical orientation: How to “strike a balance”

  • Clarify usage scenarios and risk levels (higher risk → higher requirements for explanation and auditing)

  • Integrate auditability, traceability, interrogability, and human-in-the-loop mechanisms

  • Recognizing that explanatory improvements do not equal a panacea, additional measures such as bias management, data governance, and security safeguards are still necessary

1.4.4 Ethics of AI: A paradigmatic change

This section summarizes the paradigm shift in AI ethics. First, it explains the overall context: AI ethics has shifted from focusing on single issues (such as privacy or fairness) to systematic reflection, requiring collaborative governance at the technical, social, and institutional levels (Overall context) → Then point out the three types of key actors and their responsibilities in the paradigm shift: designers and engineering teams, users and citizens, and policy makers (Three types of actors and their responsibilities in paradigm shift) → Finally, put forward the methodological positioning: Ethics should not be regarded as a “list of answers”, but a continuous process of reflection and creation, helping society enhance its judgment and governance capabilities in the complex and uncertain AI ecosystem (Methodological positioning of ethics)

1.4.4.1 Overall context

  • Expansion of the Issue Spectrum: From traditional issues (privacy, responsibility) to new topics (sustainability, digital medicine, etc.)

  • Conclusion Highlights: AI ethical issues cannot be solved by technology alone; a socio-technical perspective is needed (technical solutions must be embedded in broader ethical and social frameworks)

1.4.4.2 Three types of actors and their responsibilities in paradigm shift

  • Designers/Engineering Teams: Should not directly embed “desired moral forms” into technology; instead, they need to anticipate technological mediation and employ moral imagination

  • Users and Citizens: Should recognize “who is embedding which values into AI” and foster public supervision and participation

  • Policymakers/Regulators: Should not only conduct ex-post supervision but also co-shape technology, participate from the outset, and promote public discourse

1.4.4.3 Methodological positioning of ethics

  • Ethics is not an “answer manual/checklist” but a continuous reflective process

  • It relies on creativity and moral imagination to enhance society’s ability to address and govern complex moral issues

2 THE DEVELOPMENT OF AI AND ITS ETHICAL IMPLICATIONS

2.1 The Development of AI

2.1.1 What is AI

This section marks the starting point of the AI Introduction course. Firstly, it points out the contemporary significance of AI and its latest representative achievements—from deep learning to large language models, AI has become a core driving force behind technological, economic, and social transformations (The significance of AI in the era and its recent representative achievements)→ Then it explains the difficulty in defining AI: the boundaries of AI continue to evolve with technological advancement, so this course regards AI as an interdisciplinary “dynamic concept” and focuses on its technological logic and social impacts (The difficulty in defining “What is AI” and the positioning of this course)→ Next, it provides an intuitive definition—“Judging intelligence by behavior”, which means evaluating intelligence by observing whether a machine can demonstrate human-like cognition or decision-making in complex environments (Intuitive definition and the method of ‘Judging intelligence by behavior’)→ Subsequently, it introduces the “AI effect” and Larry Tesler’s view: once a technology is widely mastered, people no longer consider it part of AI, reflecting the relativity and social constructiveness of “intelligence” (The ‘AI effect’ and Larry Tesler’s view)→ Finally, it previews the subsequent structure: the course will take the four historical stages of AI development as the main thread, sort out the evolutionary path from symbolism to machine learning, and then to modern generative AI, laying a historical context for understanding AI ethics and social impacts (Subsequent structure: Historical stages of four periods)

2.1.1.1 The significance of AI in the era and its recent representative achievements

  • Regarded as one of the most fascinating technologies of this century and will dominate at least until 2050

  • Representative Milestones and Systems:

    • AlphaBot:Predicting the physical properties of proteins
    • Google Self-driving Car: Millions of miles without accidents
    • GPT: Generating news text comparable to that of humans
    • Pluribus: Defeating former top five players in unlimited Texas Hold’em

2.1.1.2 The difficulty in defining “What is AI” and the positioning of this course

  • Nowadays, everyone talks about AI, but it is not easy to give a strict answer

  • This course: Traces the evolutionary history of definitions and their impact on society; here only outlines concepts and proposes an analytical framework

2.1.1.3 Intuitive definition and the method of “Judging intelligence by behavior”

  • Intuitive understanding: Endowing machines with the ability of “intelligence” (distinguishing from the natural intelligence of humans/animals)

  • Dilemma: Lack of a unified definition of “intelligence” → Adopt “observable behavior + objective function” for judgment

  • Measure and optimize with formal indicators (relying on mathematics/logic):

    • Travel: Minimize travel time
    • Games/Chess: Maximize winning rate
    • Recognition: Minimize error probability

2.1.1.4 The “AI effect” and Larry Tesler’s view

  • As capabilities become widespread, tasks originally regarded as AI will be “removed” (e.g., OCR)

  • Tesler said: “AI is a technology that has not yet been realized.”

2.1.1.5 Subsequent structure: Historical stages of four periods

  • Before 1956: Predecessor period

  • 1956: The formal birth of AI

  • From Birth to the end of the 20th century: Early successes and failures pave the way for the explosion in the new century

  • Early 21st century to present

2.1.2 Before the birth of AI: Precursors

This section unfolds around the “intellectual origins of AI”. Firstly, it traces the early goals and focus areas—humans have long attempted to simulate reasoning, computation, and intelligent behaviors through mechanical means (Early goals and focus areas)→ Then it introduces the three major milestones in mechanical computing in the 17th century, such as Pascal’s Adding Machine and Leibniz’s Step Reckoner, demonstrating the initial attempts to “replace human cognitive operations with machines” (Three milestones in mechanical computing in the 17th century)→ Entering the 19th century, the ideological foundation before Turing had emerged: Babbage’s Analytical Engine and Ada Lovelace’s concept of programming made “programmability” the core idea of intelligent machines (Emergence of programmable ideas in the 19th century)→ By the mid-20th century, the birth of electronic computers and the theoretical foundation of Turing’s Universal Machine marked the possibility of “general intelligent computing” in the true sense (Mid-20th century: From electronic computing to general computing theory)→ During the same period, McCulloch and Pitts proposed the first generation of artificial neuron models (1943), igniting another intelligent paradigm inspired by biological cognition (The first spark of neural networks (1943))→ Finally, it elevates to the conceptual level: these technological veins collectively lead to an ultimate question—what is the essence of intelligence? Based on what criteria should we judge “whether a machine is thinking”? (Elevated discussion: The essence of intelligence and criteria for judgment)

2.1.2.1 Early goals and focus areas

  • Humans have been attempting to create machines that can “replicate” intelligence for hundreds of years

  • Early exploration of machines and theories focusing on “computing power”

2.1.2.2 Three milestones in mechanical computing in the 17th century

  • 1623 Schickard: Mechanical calculation within six digits (requiring human intervention)

  • 1642 Pascal: Mechanical adder-subtractor (multiplication and division realized through repeated addition and subtraction)

  • 1674 Leibniz: Recursive mechanism for addition/subtraction/multiplication

2.1.2.3 Emergence of programmable ideas in the 19th century

  • Babbage’s “Difference Engine”: Solving polynomials with the finite difference method (materialized by the Scheutz father-son team)

  • Babbage’s “Analytical Engine”: A general computing concept that surpasses the difference engine

2.1.2.4 Mid-20th century: From electronic computing to general computing theory

  • Turing: Proposal of the Turing Machine — simulating any mathematical reasoning with a 0/1 symbol system

  • Church & Turing: Digital computers can simulate any formal reasoning; the Turing Machine is a universal model

  • Convergence with neurobiology, information theory, and cybernetics → the “electronic brain” concept

2.1.2.5 The first spark of neural networks (1943)

  • McCulloch & Pitts: Linear Threshold Unit (LTU), artificial neuron model

  • Outcomes:

    • Neural networks can implement any Boolean function (conjunctive/disjunctive normal forms)
    • Cyclic networks with feedback → dynamic systems with memory

2.1.2.6 Elevated discussion: The essence of intelligence and criteria for judgment

  • Turing further focused on “how to judge an intelligent entity”

  • Laying the ideological groundwork for the Turing Test

2.1.3 Before the birth of AI: Turing test

This section focuses on the ideological logic of the Turing Test. Firstly, it introduces its “problem restatement strategy”: Turing transformed the originally intractable question “Can machines think?” into a more operable one—“Can machines behave like humans in conversations?” (Problem restatement strategy)→ Then it explains the three-person setup of the “Imitation Game”: the interrogator, the human subject, and the machine act as judges and imitators respectively, testing the machine’s intelligent performance through text-based conversations (‘Imitation game’ Three-person game setup)→ Next, it points out the key transformation: shifting from “What is the essence of intelligence?” to “How is intelligence recognized through behavior”, that is, evaluating intelligence through observable behaviors rather than internal mechanisms (Key transformation)→ Finally, it discusses the variants and impacts of the Turing Test, from CAPTCHA to dialogue system evaluation. It has not only shaped AI research methods but also triggered ongoing debates about intelligence, consciousness, and machine ethics (Variants and impacts)

2.1.3.1 Problem restatement strategy

  • Original Question: “Can machines think?” → “Thinking” is difficult to define

  • Turing’s Proposal: Substitute with an operable alternative question

2.1.3.2 “Imitation game” Three-person game setup

  • Roles: A (Male), B (Female), C (Interrogator, gender-neutral)

  • Interaction: C cannot see A/B and judges gender only through text-based questions

  • Goal: A induces C to make a wrong judgment, while B helps C judge correctly

2.1.3.3 Key transformation

  • Replace A with a machine: If C’s misjudgment rate is comparable to that when “a human plays A” → the machine achieves “imitability”

  • Question Rewrite: Does there exist a digital computer that can perform excellently in the Imitation Game?

2.1.3.4 Variants and impacts

  • Multiple test variants have been derived

  • Reverse Turing Test (machine distinguishes human/machine) → The basic principle of CAPTCHA

2.1.4 The birth of AI

This section returns to the formal birth of AI as an independent discipline. Firstly, marked by the 1956 Dartmouth Conference, it illustrates that AI evolved from scattered explorations into a research field with a common agenda (Dartmouth conference (1956, The formal birth of AI))→ Then it explains the proposal of the concept and naming of “Artificial Intelligence”, which signifies the expansion of research goals from “automatic computation” to the grand vision of enabling machines to demonstrate human-like intelligence (Concept and naming)→ Finally, it briefly reviews two major early achievements of AI, such as symbolic reasoning and problem-solving programs, early game-playing and theorem-proving systems, laying the paradigmatic foundation for symbolic AI in the subsequent decades (Two major early achievements)

2.1.4.1 Dartmouth conference (1956, The formal birth of AI)

  • Organizers: Minsky, McCarthy; Initiators include Shannon, Rochester

  • Format: Intensive “brainstorming” lasting 6-8 weeks

  • Participants: Solomonoff, Selfridge, More, Samuel, Newell, Simon, and about ten others

2.1.4.2 Concept and naming

  • Proposal Declaration: “All aspects of learning and intelligence can be precisely described and simulated by machines”

  • McCarthy established the term Artificial Intelligence (deliberately distinguished from “cybernetics”)

2.1.4.3 Two major early achievements

  • 1956 “Logic Theorist” (Newell & Simon)

    • Proved 38 out of the first 52 theorems in Principia Mathematica

    • Provided more concise proofs for some theorems

  • 1959 Samuel’s Checkers Program: Reached the level of an amateur master

    • Impact: Laid the methodological foundation and confidence for subsequent AI technology waves

2.1.5 After the birth of AI: Winters and springs

This section sorts out the periodic ups and downs in AI development. Firstly, it points out that early research made unfulfillable promises due to over-optimism, leading to a serious disconnect between expectations and reality (Over-optimism and failed promises)→ Then it analyzes the reasons for failure, including structural issues such as insufficient computing resources, limited algorithmic capabilities, and misjudgment of the essence of intelligence (Reasons for failure)→ These setbacks ultimately led to funding withdrawal and confidence collapse, contributing to the first “AI winter” in 1974 (Funding withdrawal and the ‘First AI winter’ (1974))→ However, with the rise of expert systems in the 1980s, AI ushered in a short-lived “spring”, driving practical commercial applications through knowledge engineering and rule-based systems (Expert systems bring a ‘Spring’ (1980s))→ Yet as maintenance costs soared and systems became rigid, AI entered the second winter between 1990 and 2000 (The second AI winter (1990–2000))→ Finally, in the 21st century, relying on improved computing power, accumulation of big data, and breakthroughs in machine learning, AI moved towards revival and diversification, entering a new stage of modern industry and research (Revival and diversification after the 21st century)

2.1.5.1 Over-optimism and failed promises

  • 1958 Simon & Newell: Computers will become world chess champions within a decade; discover and prove important theorems within a decade

  • 1965 Simon: Machines will be able to do everything humans can do within two decades

  • 1967 Minsky: “AI problems will be largely solved within a generation”

  • 1970 Minsky: Build a machine with “average human intelligence” within three to eight years

2.1.5.2 Reasons for failure

  • Application difficulty was underestimated: Solvable in small examples, but difficult to handle at real scale

  • Insufficient computing power: Severely limited by the hardware of the time

  • Data scarcity: Fields like computer vision/NLP require massive data, which was difficult to construct/learn at that time

2.1.5.3 Funding withdrawal and the “First AI winter” (1974)

  • Governments (UK), Defense Advanced Research Projects Agency (DARPA), National Research Council (NRC), etc., withdrew funding for “undirected AI research”

  • Funding plummeted, making projects unsustainable

2.1.5.4 Expert systems bring a “Spring” (1980s)

  • Architecture: Knowledge base + inference engine

  • Implementation: Successful enterprise-level applications such as medical diagnosis

  • The boom lasted about a decade

2.1.5.5 The second AI winter (1990-2000)

  • Maintainability/portability issues of expert systems emerged

  • Investment and enthusiasm cooled down again

2.1.5.6 Revival and diversification after the 21st century

  • Causes: Improved computing power + focus on “verifiable” specific problems + rigorous scientific paradigms

  • Result: AI diversified into multiple competitive subfields (each deeply exploring specific problems/methods)

  • Industry: Flourished comprehensively in large-scale engineering and products

2.1.6 After the birth of AI: General vs. narrow AI

This section distinguishes between different types of AI. Firstly, it introduces Artificial General Intelligence (AGI / General AI), emphasizing that its goal is to enable machines to possess cross-domain, human-level understanding and reasoning capabilities (Artificial general intelligence)→ Then it explains that Narrow AI only performs excellently in specific tasks, such as image recognition or recommendation systems, but cannot be transferred to other scenarios (Narrow AI)→ Finally, it summarizes the significance of the shift in research focus from “pursuing general intelligence” to “developing usable narrow-domain intelligence”: it has brought AI from a philosophical vision to practical applications and driven the rapid development of modern AI (Significance of the shift in research focus)

2.1.6.1 Artificial general intelligence (AGI / General AI)

  • Goal: Can understand/learn any intellectual task that humans can accomplish

  • Current Status: No successful instances yet

2.1.6.2 Narrow AI

  • Only handles single or limited tasks (targeting specific scenarios and metrics)

  • Representatives: AlphaGo, IBM Watson (integrates multiple narrow-domain technologies)

2.1.6.3 Significance of the shift in research focus

  • The shift from “pursuing general AI” to “deeply cultivating narrow AI” was the key to emerging from the first AI winter

  • 1990s – early 2000s: The rise of subfields oriented towards verifiable results and commercial value—bringing greater academic and commercial success

2.2 Today’s AI

2.2.1 Definition and goals

This section deepens the AI definition system. Firstly, it introduces the academic classification of AI concepts, refining “intelligence” into several research directions from different dimensions (Deepening and classification system of AI definitions)→ Then it elaborates on four classic definition paths respectively: targeting “Thinking humanly”, attempting to simulate human cognitive structures; taking “Thinking rationally” as the criterion, emphasizing logical reasoning and formalized knowledge; measuring intelligent behavior by “Acting humanly”; and focusing on “Acting rationally”, pursuing the optimal behavior choice in the environment→ Finally, it puts forward the core concept of rational agent, defining AI as a system that can independently perceive, reason and take optimal actions in the environment, laying a unified framework for subsequent AI systems and modeling methods (Core concept of rational agent)

2.2.1.1 Deepening and classification system of AI definitions

  • In recent years, the understanding of the definition of AI has become more in-depth.

  • The two-dimensional framework proposed by Stuart Russell and Peter Norvig:

    • Dimension 1: Thinking vs. Acting

    • Dimension 2: Humanly vs. Rationally

  • Four types of artificial intelligence formed therefrom:

Human-Like Rational
Thinking Thinking Humanly Thinking Rationally
Acting Acting Humanly Acting Rationally

2.2.1.2 Thinking humanly

  • Representative definition: AI aims to automate activities related to human thinking, such as decision-making, problem-solving, and learning (Bellman, 1978)

  • Closely related to cognitive science and neuroscience

  • The goal is to replicate the human thinking process

  • Distinction: Different from using computational models to study mental abilities (which falls into the realm of rational thinking)

2.2.1.3 Thinking rationally

  • Goal: Establish strict reasoning rules to achieve logical thinking (Charniak & McDermott, 1985)

  • Highly relevant to the field of computational logic

  • Limitations:

    • Not all intelligent behaviors can be generated through logical reasoning

    • Therefore, thinking rationally cannot fully cover the field of artificial intelligence

2.2.1.4 Acting humanly

  • Definition: Let computers accomplish tasks that humans are currently better at (Rich & Knight, 1991)

  • Closely related to the Turing Test: Demonstrate behavioral performance as close to humans as possible

  • Characteristic: Focuses on the behavior itself rather than the mechanism of behavior selection

  • Limiting the imitation of human behavior + achieving the ultimate goal of artificial intelligence

2.2.1.5 Acting rationally

  • Definition (Luger & Stubblefield, 1993):

    • Artificial intelligence is a branch of computer science

    • Devoted to automating “intelligent behavior”

    • Intelligence is measured by formal indicators (quantifiable criteria)

  • Differences:

    • Acting Rationally vs. Acting Humanly

    • Humans may make mistakes, while rational programs can avoid errors

  • Examples:

    • Tic-tac-toe, checkers, Rubik’s Cube solving programs

    • All can achieve error-free “rational behavior”

2.2.1.6 Core concept of rational agent

  • The most common definition of modern AI is based on rational agents.

  • Definition of a rational agent:

    • A function that maps perceptions to actions

    • Goal: Maximize the expected utility of the objective function

  • Scope of application: Specific tasks (such as playing chess, image recognition)

  • Perfect rationality: Always take optimal actions

  • Bounded rationality: Limited by computational resources, make the best decisions within feasible ranges

  • Goal of artificial intelligence: Design agents that achieve rational behavior under limited resources

2.2.2 Problems and progresses

This section focuses on the capability levels of agents. Firstly, it introduces five classifications of agent performance levels, ranging from simple reflexive behaviors to advanced agents with learning and reasoning abilities, forming a complete spectrum of AI behavioral capabilities (Five classifications of agent performance levels)→ Then it explains that the structured degree of different problems affects agent design: some tasks have clear rules and definable states, while others involve open and complex environments that require stronger adaptability (Problem characteristics and structured degree)→ Next, it points out which tasks are most challenging at high-human and par-human levels, such as natural language processing, visual understanding, and complex decision-making scenarios (Tasks at high-human and par-human levels)→ Finally, it discusses development limitations and future outlook: data dependence, computing power bottlenecks, and insufficient generalization ability remain current obstacles, but the long-term direction still points to a more autonomous and generalized agent system (Development limitations and future outlook)

2.2.2.1 Five classifications of agent performance levels

  • Based on the core idea of rational agents (i.e., maximizing or minimizing indicators), they are classified into five categories by comparison with human performance:

    • Optimal: Performance reaches the theoretical limit and cannot be improved

    • Super-human: Superior to all humans

    • High-human: Superior to most humans

    • Par-human: Comparable to the average human level

    • Sub-human: Inferior to most humans

2.2.2.2 Problem characteristics and structured degree

  • Problems that can reach optimal or super-human levels are mostly well-structured and mathematically describable

  • Typical examples of optimal agents:

    • Tic-tac-toe, Connect Four, Checkers, Rubik’s Cube

    • Heads-up limited Texas Hold’em (long-term optimal returns)

  • Representative examples of super-human agents:

    • Othello

    • Scrabble

    • Backgammon

    • Jeopardy!

    • Go

    • Heads-up/multiplayer no-limit Texas Hold’em (Chess)

2.2.2.3 Tasks at high-human and par-human levels

  • Examples of high-human agents:

    • Crossword puzzles

    • Dota 2

    • Bridge

    • StarCraft

  • Tasks for par-human or sub-human agents:

    • Par-human: Optical Character Recognition (OCR), image classification, handwritten recognition

    • Sub-human: Object recognition, facial recognition, bipedal walking, speech recognition

2.2.2.4 Development limitations and future outlook

  • Improvements in technology and computing power promote the progress of agents

  • However, it cannot be guaranteed that sub-human tasks will definitely reach human levels

  • Some problems may never surpass human performance

2.2.3 Major milestones

This section presents the key leaps of AI technology in a chronological context. Firstly, it reviews early milestones: from ELIZA to expert systems, AI began to perform pattern matching and rule-based professional reasoning (Early milestones: From ELIZA to expert systems)→ Then it moves to major breakthroughs in the 1990s, such as support vector machines, probabilistic graphical models, and reinforcement learning, laying the mathematical foundation for modern machine learning (Major breakthroughs in the 1990s)→ Subsequently, it demonstrates the rise after 2000: big data, parallel computing power, and deep learning drove revolutionary progress in image recognition, speech recognition, and natural language processing (After 2000: The rise and explosion of modern AI)→ Finally, since 2017, AI has begun to conquer the field of “imperfect information”—systems like AlphaZero and AlphaStar have demonstrated superhuman performance in games, strategies, and high-complexity environments, marking AI’s entry into a new capability landscape (After 2017: AI conquers imperfect information domains)

2.2.3.1 Early milestones: From ELIZA to expert systems

  • ELIZA (1966) :

    • The world’s first chatbot
    • Made artificial intelligence accessible to the public for the first time
    • Triggered a media frenzy and pushed AI to the forefront of research
  • XCON (1980) :

    • An expert system for computer system configuration recommendation
    • Saved the company approximately $25 million annually
    • Fueled the second AI “Spring”

2.2.3.2 Major breakthroughs in the 1990s

  • IBM Deep Blue (1997):

    • Defeated the world chess champion

    • Though relying heavily on brute-force computing, it marked a historic breakthrough for AI

2.2.3.3 After 2000: The rise and explosion of modern AI

  • 2005: Succeeded in autonomous driving at the DARPA Grand Challenge.

  • 2011: IBM Watson Defeated Human Champions in Jeopardy!

    • Could understand natural language problems

    • Did not rely on logical or algebraic reasoning

    • Combined classic NLP techniques with statistical methods

  • 2012: AlexNet Sparked the Deep Learning Revolution

    • Significantly improved image recognition accuracy
  • 2016: Deep learning defeated top human Go players

2.2.3.4 After 2017: AI conquers imperfect information domains

  • 2017 Libratus: Defeated the world’s top 4 poker players and won $1.8 million

  • 2018 Waymo: Launched self-driving taxis in Phoenix

  • 2019 AlphaStar:Defeated 99% of human players in StarCraft II

  • After 2020: GPT-3:

    • Generates texts like stories and poems

    • The language is fluent and natural, on par with human-level

2.2.4 Scientific community

This section examines AI development from an ecological perspective. Firstly, it points out the explosive growth of the AI research community, with the number of scholars, engineers, and interdisciplinary researchers rising rapidly, forming an unprecedented density of innovation (Explosive growth of the research community)→ Then it uses quantitative indicators such as the number of papers, conference influence, and patents to show the continuous rise in research output, indicating that AI has become a core sector in the global scientific research landscape (Quantitative indicators of research output)→ This is followed by the continuous strengthening of cooperation between academia and industry, as large technology companies and universities/research institutions promote the transformation of technology from papers to products through joint laboratories and cooperative projects (Strengthened academic-industrial cooperation)→ Finally, it emphasizes the importance of AI education and talent cultivation: university curriculum systems, professional degrees, and online education jointly build an AI talent training pipeline, providing a human resource foundation for future technological development and governance (AI education and talent cultivation)

2.2.4.1 Explosive growth of the research community

  • In 1956: Only 10 researchers in the AI field

  • Today: The research community has expanded exponentially

2.2.4.2 Quantitative indicators of research output

  • Number of AI scientific papers (2000-2019)

    • Increased by more than 600%
    • Accounted for nearly one-third from 2017 to 2019
  • The number of papers reflects:

    • The number of researchers

    • Invested funds and research popularity

2.2.4.3 Strengthened academic-industrial cooperation

  • Nearly 10% of AI papers in the US involve corporate authors

  • Indicates the close integration of academia and industry

2.2.4.4 AI education and talent cultivation

  • Number of courses: Stanford University had nearly 80 AI courses in 2020

  • Student scale: More than 1,000 students take basic AI courses in top universities every year

  • Reflects the national-level AI application and development potential

2.2.5 Research in industry

This section emphasizes the core position of enterprises in the modern AI ecosystem. Firstly, it points out that AI research has shifted from being academia-led to a pattern where “enterprises are on an equal footing or even leading”, and commercial demands have driven enterprises to increase R&D investment (Rise of AI research in enterprises)→ Then it explains the key role of industry in AI research: mastering large-scale data, computing power resources, and real application scenarios makes it the main engine for promoting technology landing and iteration (Core role of industry in AI research)→ Subsequently, it lists major global AI research enterprises, such as Google, OpenAI, Microsoft, Meta, Amazon, etc., which lead the frontier through foundation models and platform technologies (Major global AI research enterprises)→ At the same time, it also emphasizes the importance of other types of enterprises and interdisciplinary forces, such as companies in medical care, finance, autonomous driving and other fields jointly participating in AI innovation (Other enterprises and interdisciplinary research)→ Finally, it summarizes: the in-depth integration of industry and academia is the key to modern AI development. Only by realizing the cycle of research-application-feedback can AI maintain continuous progress and meet social needs (Importance of integration between industry and research)

2.2.5.1 Rise of AI research in enterprises

  • After 2010: Major companies established AI/ML labs

  • Deep cooperation between enterprises and academic researchers

  • Researchers’ involvement:

    • Academic conferences

    • Paper publications

    • Product innovation and application

2.2.5.2 Core role of industry in AI research

  • AI research has become the core driver of corporate technological innovation

  • Many AI milestones have been achieved by private enterprise labs

2.2.5.3 Major global AI research enterprises

  • Tech giants: Google, Microsoft, Amazon, Apple, IBM, Meta (formerly Facebook)

  • Each lab has an average of over 1,000 researchers

  • Provide awards and funding support to academia

  • Co-founded the “Partnership on AI”

    • Focus on the social responsibility and ethics of AI

    • Baidu joined in 2018, becoming the first Chinese member

    • By 2019, the alliance had more than 100 members (covering academia, enterprises, and non-profit organizations)

2.2.5.4 Other enterprises and interdisciplinary research

  • Other companies with AI labs: Adobe, Criteo, Uber, Intel, McKinsey, Sony, Samsung, BBC, Accenture

  • Promote data-driven innovative products and services

2.2.5.5 Importance of integration between industry and research

  • Data-driven methods + AI tools are key to innovation

  • Without deep cooperation between research and industry, innovation is difficult to achieve

2.3 Research Fields of AI

2.3.1 AI field

This section outlines the overall framework for building AI systems. Firstly, it clarifies its goals and scope: from classification methods to various characteristics, to comprehensively understand the composition and operation of AI systems (Goals and scope)→ Then it proposes a top-level tripartite classification, dividing AI systems into three categories: technical methods, technical tools, and regulatory and organizational issues (Top-level tripartite classification: Technical methods / Technical tools / Regulatory and organizational issues)→ Technical methods are further subdivided into three major subtypes: Perception technology, Action technology, and Cognitive technology (Three major subtypes of technical methods)→ In the dimension of technical tools, it is divided into modeling tools and basic support platforms (Two major types of technical tools)→ Finally, it points out that the regulatory and organizational issues need to focus on ethical compliance, safety risks, human-machine collaboration, and social impacts, laying a structured foundation for subsequent discussions on AI governance (Overview of regulatory and organizational issues)

2.3.1.1 Goals and scope

  • Overview of AI-related subfields

  • Explain classification methods, respective characteristics, and their interrelationships

  • Other courses: Will develop into each subfield separately

2.3.1.2 Top-level tripartite classification: Technical methods / Technical tools / Regulatory and organizational issues

  • Technical methods (theories, models, algorithms)

    • Diverse and require further subdivision

    • Classified by imitated human capabilities: Perception / Cognition / Action

  • Technical tools (software/hardware)

    • Support the development and implementation of AI products

    • Two categories: Methodology and AI platforms

  • Regulatory and organizational issues

    • Ethics and law
    • Impact on labor market
    • Management process optimization and AI project advancement in public and private organizations

2.3.1.3 Three major subtypes of technical methods

  • Perception technology

    • Task: Acquire, process, and understand information from the environment

    • Representatives: Computer Vision (CV), Natural Language Processing (NLP)

  • Action technology

    • Task: Execute actions in the physical world

    • Representative: Robotics

  • Cognitive technology

    • Task: Learning, Knowledge Representation, Reasoning, Planning, Collaboration

    • Representatives: Machine Learning (ML), Knowledge Representation and Reasoning (KRR), Planning and Scheduling, Multi-Agent Systems

2.3.1.4 Two major Types of technical tools

  • Methodology: Support the design and engineering processes of AI products

  • AI platforms: Reusable software libraries/services to promote AI democratization

2.3.1.5 Overview of regulatory and organizational issues

  • Ethics and law

  • Labor market and job structure

  • Processes and governance in public/private organizations

2.3.2 Techniques: Perception

This section introduces the core role of “perception and action” in AI. Firstly, it points out that the two together enable agents to interact with the physical world: perception is responsible for acquiring, processing, and understanding information; action is responsible for performing tasks based on that understanding (Overall role of perception and action)→ Then it moves to the visual perception module, explaining that the goal of Computer Vision (CV) is to gain high-level understanding from images and videos. By converting high-dimensional visual signals into computable representations, it promotes capabilities such as recognition, tracking, and generation based on geometric, physical, and statistical modeling (Perception: Computer vision (CV))→ Finally, it introduces Natural Language Processing (NLP) as another key perception method. Its goal is to process and analyze large-scale natural language data to support tasks such as information extraction, understanding, question answering, translation, and generation, connecting language, knowledge, and intelligent behaviors (Perception: natural language processing (NLP))

2.3.2.1 Overall role of perception

  • Together enable interaction with the physical world

  • Perception: Acquire—Process—Understand information

  • Action: Execute physical tasks

2.3.2.2 Perception: Computer vision (CV)

  • Definition and Goals

    • Gain high-level understanding from images/videos

    • Convert high-dimensional visual data into computable numerical/symbolic representations

    • Model-based understanding grounded in geometry, physics, statistics, and learning theory

  • Subfields

    • Biometrics; Face/Gesture/Posture recognition

    • Image/Video generation and retrieval

    • Video understanding and behavior analysis; motion and tracking

    • Robotics and autonomous driving vision

  • Application examples

    • Automated MRI cancer detection

    • Autonomous driving vision systems

    • GAN image conversion

    • Automated social distancing monitoring

2.3.2.3 Perception: natural language processing (NLP)

  • Definition and goals

    • An interdisciplinary field of linguistics, computer science, and AI

    • Process/analyze large-scale natural language data to extract information and insights

    • Support understanding, classification, organization, and generation

  • Subfields

    • Dialogue systems/Conversational AI; Text generation; Information extraction

    • Machine translation; Question answering systems; Speech processing and synthesis

    • Text classification and sentiment analysis

  • Application examples

    • Smart assistants, Machine translation

    • Document retrieval, Email filtering

2.3.3 Techniques: Action (Robotics)

This section introduces the “action” technology in the AI system, focusing on robotics. Firstly, it emphasizes its interdisciplinary nature: integrating and developing in collaboration with fields such as computer science, mechanical engineering, electrical engineering, control engineering, mechatronics, biology, and software engineering (Action: Robotics)

  • Interdisciplinary integration: Computer science +Mechanical/Electrical/Control/Mechatronics/Biological/Software Engineering, etc.

  • Goals and application scenarios

    • Assist or replace humans in completing tasks

    • Hazardous environments (radioactive, explosive handling), manufacturing processes, extreme environments (space/underwater/high temperature), hazardous/radiation environment cleanup

  • Form and bionics

    • Diverse appearances; some are humanoid to improve acceptance

    • Replicate walking, weightlifting, speech, cognition, etc

    • Development of nature-inspired biorobotics

  • Subfields

    • Human-robot interaction (HRI)

    • Localization, mapping, and navigation (SLAM, etc.)

    • Manipulation control

    • Motion and path planning

  • Application examples

    • Manufacturing, robotic surgery, cleaning operations, etc

2.3.4 Techniques: Cognition

This section systematically sorts out the composition of “cognitive AI technologies”. Firstly, it clarifies its task scope and points out that it is the core module supporting the high-level intelligence of agents (Scope of cognition and AI tasks)→ Then it introduces Machine Learning (ML), whose goal is to enable algorithms to automatically improve performance through data and experience, covering methods such as supervised/unsupervised/reinforcement learning, deep learning, and evolutionary algorithms, and is widely used in prediction, recommendation, maintenance, and anomaly detection (Machine learning (ML))→ Next, it discusses Knowledge Representation and Reasoning (KRR), emphasizing enabling machines to understand and automatically reason about world knowledge in the form of rules, logic, conceptual structures, etc., supporting applications such as expert systems, semantic networks, and automated theorem proving (Knowledge representation and reasoning (KRR))→ Afterwards, it introduces automated planning and scheduling, whose core is to formulate strategies and action sequences for agents, conducting goal optimization and task execution in different environments (modelable vs. dynamic unknown) (Automated planning and scheduling)→ Finally, it presents Multi-Agent Systems (MAS), composed of multiple interacting agents. Through methods such as collaboration, game theory, distributed problem-solving, and mechanism design, it solves complex problems that cannot be handled by a single agent, and is used in economic transaction automation, resource scheduling, and distributed control (Multi-agent systems (MAS))

2.3.4.1 Scope of cognition and AI tasks

  • Covers: Knowledge formation, memory/working memory, judgment/evaluation, reasoning, problem-solving, decision-making

  • Core subfields: Machine learning, KRR, Planning and scheduling, Multi-agent systems

2.3.4.2 Machine learning (ML)

  • Definition: Algorithms that automatically improve through experience and data; build models based on training data for prediction/decision-making

  • Data sources: From perception systems like CV/NLP

  • Relationship with statistics: Closely related to computational statistics but not identical; data mining emphasizes unsupervised exploratory analysis

  • Commercial terms: Predictive analytics/Advanced analytics

  • Subfields/Methods:

    • Unsupervised/Supervised/Reinforcement learning

    • Classification/Regression; Clustering

    • Deep neural networks; Evolutionary learning

    • Online learning and bandit algorithms; Quantum machine learning

  • Applications:

    • Prediction and forecasting (demand/sales volume)

    • Recommendation systems

    • Predictive maintenance

    • Anomaly detection

2.3.4.3 Knowledge representation and reasoning (KRR)

  • Goal: Represent world knowledge in a machine-usable form and enable automatic reasoning

  • Method sources: Draws from human problem-solving research (psychology) and logic

  • Types of reasoning: Rule application, set/subset relationships, etc

  • Subfields:

    • Logic describing; Knowledge engineering; Logic programming

    • Ontology; Automated reasoning and Theorem proving

  • Applications: Semantic Web, expert Systems, etc

2.3.4.4 Automated planning and scheduling

  • Goal: Develop strategies or action sequences for agents/robots/unmanned systems

  • Relationship with decision theory: Optimization under goals and uncertainty

  • Environmental differences:

    • Known and modelable: Offline planning and pre-evaluation are feasible

    • Dynamically unknown: Online adjustment is needed, often combined with iterative trial-and-error and learning

  • Subfields:

    • Deterministic planning; Spatio-temporal system optimization

    • Plan execution and monitoring

    • Integration of planning/scheduling and learning

    • Temporal planning

2.3.4.5 Multi-agent systems (MAS)

  • Definition: A system composed of multiple interacting agents

  • Motivation: Solve problems that are difficult for a single agent to handle

  • Sources of intelligence: Methodology, functionality, procedural methods, search and reinforcement learning, etc

  • Subfields:

    • Adversarial agents

    • Protocol, argumentation and negotiation

    • Coordination and collaboration

    • Distributed problem solving

    • Mechanism design

  • Applications:

    • Automation of economic transactions (microeconomic models)

    • Physical security resource allocation

    • Decentralized control

2.3.5 Technologies

This section outlines the supporting structure of the AI technology system. Firstly, it distinguishes between “technical methods” and “technical tools”: the former focuses on the functions of AI products themselves (such as algorithm and model design), while the latter focuses on the development and deployment processes (engineering and platformization) (Technical methods vs. technical tools)→ Then it emphasizes the engineering methodology of AI projects, pointing out that their high-risk nature requires strict processes, including concept design, blueprint evaluation, PoC verification, and MVP delivery, and draws on the systematic management of software engineering (Methodology (Engineering process and project management))→ Subsequently, it introduces the importance of software and hardware infrastructure: big data and algorithms form the software foundation, while high-computing-power hardware (especially accelerators such as GPUs) significantly determines the efficiency of deep learning training and inference (Infrastructure (Software and hardware))→ It further explains that AI platforms, as key tools integrating software and hardware, provide enterprises with pre-built, drag-and-drop, or low-code tools (such as platforms from Microsoft, Amazon, and Google) to lower development thresholds and continuously expand algorithm capabilities (Artificial intelligence platforms)→ Finally, it discusses AI democratization, whose goal is to enable small and medium-sized enterprises and individuals to carry out AI projects without building their own teams. By opening up modules, data, and cloud infrastructure, it reduces costs, promotes collaboration, and accelerates the popularization of AI, playing a key role in ecological development and industrial implementation (AI democratization)

2.3.5.1 Technical methods vs. technical tools

  • Technical methods: Focus on the functional realization of AI products

  • Technical tools: Focus on development and deployment processes (engineering and platforms)

2.3.5.2 Methodology (Engineering process and project management)

  • AI projects are high-risk and require strict processes

  • Typical stages: Conception — Blueprint and Evaluation — Proof of Concept (PoC) — Minimum Viable Product (MVP)

  • Draw on and customize software engineering practices

2.3.5.3 Infrastructure (Software and hardware)

  • Software infrastructure supported by big data and algorithms

  • Hardware: High computing power demand (especially for deep learning); accelerators like GPUs significantly reduce training/inference costs

2.3.5.4 Artificial intelligence platforms (Key tools integrating software and hardware)

  • Background: Talent shortage leads major companies to launch pre-built/drag-and-drop/no-code AI tools

  • Representatives: Microsoft, Amazon, Google platforms

  • Coverage: Mature components for common applications like CV, NLP

  • Major companies continuously expand the algorithmic capabilities of their platforms

2.3.5.5 AI democratization

  • Concept: Enable small and medium-sized enterprises to carry out AI projects without building their own teams

  • Means: Provide directly usable modules, open data and algorithms, cloud infrastructure

  • Effects:

    • Reduce the entry barrier and overall costs for individuals and organizations

    • Promote community collaboration and talent cultivation in the ecosystem

    • Accelerate the popularization in academia and business

  • Timeline-based coverage of resources: Data, storage/computing, algorithms, deployment, market ecology

  • Importance: Plays a decisive role in the overall success of AI

2.3.6 Regulation and organization issues

This section discusses the governance issues of AI at the social and organizational levels. Firstly, it points out the public value driven by technology: big data, cloud computing, and complex algorithms provide broad application potential for governments and public sectors in healthcare, education, transportation, energy management, and other fields (Technology-driven and public value)→ Then it emphasizes risks and responsibilities, proposes the concept of “Responsible Innovation”, and reminds that large-scale data and automated decision-making may amplify discrimination. Therefore, public sectors should proactively prevent risks and build safe and fair AI systems (Risk and responsibility: Promote ‘Responsible innovation’)→ Subsequently, it defines the scope and subjects of AI ethics, covering the moral behaviors of humans in designing/using AI, as well as the ethical issues of the machines themselves; everyone from engineers to managers needs to assume ethical responsibilities, extending to topics such as superintelligent AI (Scope and subjects of AI ethics)→ Furthermore, it analyzes key legal issues, such as privacy protection, division of algorithmic responsibilities, and establishment of regulatory frameworks, which are core legal challenges for global AI governance (Key issues in legal matters)→ Finally, it discusses the impacts on organizations and the labor market: AI spawns new jobs, promotes the evolution of organizational structures, changes management methods, and brings a trend of “coexistence of job replacement and job transformation” in employment, becoming an important issue for future social transformation (Organizational and labor market impacts)

2.3.6.1 Technology-driven and public value

  • Three major drivers: Big Data, cloud computing platforms, complex ML algorithms

  • Potential government benefits: Healthcare, education, and transportation in smart cities

  • Other public applications: Food supply chain, energy and environmental management

2.3.6.2 Risk and responsibility: Promote “Responsible innovation”

  • Large-scale data and automated decision-making may exacerbate discrimination

  • Public sectors should foster a culture of responsible innovation and proactively prevent potential harms

  • Build ethical, fair, and secure AI systems

2.3.6.3 Scope and subjects of AI ethics

  • AI ethics: A subdomain of technical ethics

  • Two dimensions:

    • Moral behavior of humans in designing/manufacturing/using/disposing of AI systems

    • Ethics of machines themselves (machine ethics)

  • Issue extension: Discussions on superintelligence and the “singularity”

  • Actors: Data scientists, engineers, domain experts, delivery managers, department heads, etc., all need to prioritize ethics and safety

2.3.6.5 Organizational and labor market impacts

  • New occupations and structures: New roles such as AI engineers and ML engineers emerge

  • Experimentation and evolution of organizational structures

  • Management transformation: Help employers adapt to AI paradigms

  • Employment impact:

    • Monitoring and evaluation of machine substitution (not simply “replacement”)

    • Europe’s IT talent shortage (millions), and AI popularization is expected to alleviate it

2.4 AI and National Strategy

2.4.1 The need for national strategies

This section emphasizes the necessity of formulating AI strategies at the national level. Firstly, it points out AI’s comprehensive impact on a country’s future, covering areas such as economic growth, national defense, public security, and privacy protection. AI can generate trillion-dollar-level global value annually, and national competitiveness depends on R&D investment, human capital, and industrial structure (Comprehensive impact of AI on a country’s future)→ Then it analyzes the global national stratification: the United States and China are in the first tier, the United Kingdom, Germany, France, etc., in the second tier, and countries like Italy in the third tier, reflecting differences in national investment, talent, and technological ecosystems (National stratification and Italy’s current status)→ Subsequently, it elaborates on the necessity and goals of formulating national AI strategies: to improve national AI readiness, promote AI adoption in enterprises and public sectors, coordinate development and risks, and protect citizens from technological side effects (such as discrimination and privacy risks caused by lack of regulation) (Necessity and goals of formulating national AI strategies)→ It further introduces the world’s earliest national strategies (e.g., Canada’s 2017 Pan-Canadian AI Strategy) and explains that countries have since formed differentiated priorities, covering layouts in scientific research, talent, ethics, public governance, industrial collaboration, and other aspects (Pioneers and diversified focuses)→ Finally, it summarizes the differences in AI development paths among the three key strategic blocs: the United States, China, and the European Union (Three key strategic bodies (US/China/EU) and their different development perspectives)

2.4.1.1 Comprehensive impact of AI on a country’s future

  • Impact domains: Economic development, national defense construction, public safety and privacy, etc

  • Economic volume (McKinsey 2018): AI unlocks more than $1.5 trillion in global market value annually

  • Determinants of value differences: National preparedness (investment, research, human capital, labor market structure)

2.4.1.2 National stratification and italy’s current status

  • First-tier: United States, China

  • Second-tier: United Kingdom, France, Germany, Australia, Canada, Sweden, Norway, South Korea, etc

  • Third-tier: Italy, Costa Rica, India, Lithuania, Spain, etc

2.4.1.3 Necessity and goals of formulating national AI strategies

  • Necessity: AI development is crucial to nations, so national strategies must be formulated

  • Overall goals:

    • Enhance national preparedness

    • Promote the application and development of AI in enterprises/public administration/education

  • Dual tasks:

    • Coordinate economic development and benefit citizens

    • Protect citizens from side effects (unintended consequences), leading to regulatory demands in ethics, law, etc

2.4.1.4 Pioneers and diversified focuses

  • First national strategy: Canada’s “Pan-Canadian AI Strategy” (2017)

    • Focus: Investment in R&D/technology and talent

    • Main goals: Increase researchers; identify centers of excellence; support leadership in economic/ethical/legal issues; support academic communities

  • Subsequent national strategies: Form differentiated focuses, including research, talent/skills/education, public-private applications, ethics and inclusion, standards and regulations, data and digital infrastructure, etc

2.4.1.5 Three key strategic bodies (US/China/EU) and their different development perspectives

  • United States/China: Postpone regulatory issues (5-10 years) and adopt relatively loose regulation currently

    • United States: Development is mainly driven by large tech companies

    • China: Development is mainly led by the government

  • Europe: Prioritizes regulation before large-scale development

  • Follow-up arrangement: Three separate sessions will analyze the strategies of the United States, China, and the European Union

2.4.2 AI strategy in USA

This section takes the U.S. national AI strategy as a case study. Firstly, it reviews the starting point and goals of the Obama administration: based on a full understanding of AI’s potential, impacts, and risks, formulate a future-oriented national strategic agenda (Starting point and goals (Obama administration period))→ Then it introduces the first report Preparing for the future of AI, which focuses on ethics and regulation, and puts forward suggestions such as opening up data, standardization, and establishing a special committee to monitor AI development in the long term (Report 1: Preparing for the future of AI)→ The second report National AI research and development strategic plan clarifies priority directions such as long-term scientific research investment, human-machine collaboration methods, and system safety and reliability from the perspective of public funding, becoming the cornerstone of the U.S. national AI strategy thereafter (Report 2: National AI research and development strategic plan)→ The third report AI, automation, and the economy evaluates the impact of automation on industries, governments, and the labor market, focusing on how to achieve a cost-benefit balance through structural adjustments (Report 3: AI, automation, and the economy)→ Later, a strategic shift occurred during the Trump administration: it emphasized deregulation, with large technology companies leading the technical route. At the same time, through the White House-tech giant summit, it established goals such as maintaining global leadership, supporting the workforce, encouraging R&D, and removing innovation barriers, reflecting a more market-driven development model (Strategic shift (Trump administration period))

2.4.2.1 Starting point and goals (Obama administration period)

  • Core goals: Understand the potential, impacts, and risks of AI → formulate future-oriented national strategies
  • Three key reports: Establish the agenda and highlight the need for regulation and investment

2.4.2.2 Report 1: Preparing for the future of AI

  • Focus: Ethics and regulation

  • Key recommendations:

    • Data openness and standardization: Make open data and data standards federal priorities (data is the “fuel” for AI)

    • Establish a special committee: Continuously monitor AI development and report to the government regularly

    • Support basic and long-term research: Prioritize investment in AI basic/long-term research

2.4.2.3 Report 2: National AI research and development strategic plan

  • Priority directions for public funding investment:

    • Long-term investment in AI research

    • Develop effective methods for human-machine collaboration

    • Ensure the safety and reliability of AI systems

  • 2019 update: Become the foundation of the current U.S. national strategy

2.4.2.4 Report 3: AI, automation, and the economy

  • Content: Assess the impact of automation on industry, administration, business, education, etc

  • Focus: Labor market impact and restructuring to optimize the benefit-cost balance and promote AI development

2.4.2.5 Strategic shift (Trump administration period)

  • 2017: Relax regulation → large tech companies lead technology development

  • 2018 White House – Tech Giants Summit established four major goals:

    • Maintain global leadership

    • Support AI’s assistance to human labor

    • Promote R&D with public funds

    • Eliminate innovation barriers

  • Supporting measures:

    • Establish an AI Priorities Advisory Committee

    • Establish federal-level industry-academia partnerships

    • Remove multiple regulatory restrictions

  • 2021 Investment: The White House plans to invest over $8 billion in AI-related R&D (an increase of about 50% compared to 2020)

    • Priority areas: National defense and security

    • Applications in other areas: Mainly promoted by enterprises

2.4.3 AI strategy in China

This section introduces China’s national AI development strategy. Starting with the New Generation Artificial Intelligence Development Plan (2017), it clarifies the three-stage overall goals: by 2020, the industry technology reaches the international leading level; by 2025, it leads in key application fields; and by 2030, China becomes the global AI innovation center (Overall goals and timeline)→ Then it presents the economic scale expectations, showing that China’s AI market value will grow from approximately $23 billion to about $150 billion between 2020-2030, reflecting the country’s long-term strategic investment (Economic expectations)→ Next, it elaborates on the regulatory path and talent strategy: China plans to complete the AI-related specification system by 2030, while emphasizing the dual approach of “global recruitment of high-end talents + local talent cultivation,” and improving the industry’s overall level through internal corporate training and ethical supervision (Regulatory path and talent strategy)→ Finally, it explains the technology priority directions and infrastructure layout, focusing on intelligent IoT (autonomous driving, robotics, recognition systems), accelerating breakthroughs in key technologies such as intelligent sensors and neural network chips, and building large science and technology parks with massive public funding to provide systematic support for AI industry development (Technology priority directions and infrastructure)

2.4.3.1 Overall goals and timeline (New generation AI development plan, 2017)

  • Goal: Become a global AI leader

  • Three-phase milestones:

    • 2020: Industrial technology level reaches the level of global competitors

    • 2025: Become a world leader in some application fields

    • 2030: Become a global AI innovation leader

2.4.3.2 Economic expectations (Market value)

  • 2020: ~$23 billion

  • 2025: ~$60 billion

  • 2030: ~$150 billion

2.4.3.3 Regulatory path and talent strategy

  • Regulation: Complete the compilation of norms and rules system by 2030

  • Talent: Emphasize global recruitment of top talents and local labor market development

  • Biggest challenge for enterprises: Lack of qualified talents (reported by >50% of enterprises)

  • Measures: In-company AI training; formulate regulations and ethical guidelines to promote development and application

2.4.3.4 Technology priority directions and infrastructure

  • Priority directions: Intelligent Internet of Things (autonomous driving, robots, recognition systems)

  • Key technologies: Intelligent sensors, neural network chips, intelligent manufacturing

  • Major facility: Establishment of an excellence science and technology park in Beijing (~$2.1 billion)

  • Public funding: ~$8 billion annual public expenditure on AI research and industry

2.4.4 AI strategy in EU

This section takes “human-centered and trustworthy AI” as the core position. Firstly, it emphasizes Europe’s principle for AI development: AI should only be developed when it benefits humans, and risks should be strictly assessed based on a strong regulatory framework (Human-centered and “Trustworthy AI”)→ Then it sorts out the key policy milestones from 2018 to 2021, from the European Artificial Intelligence Strategy to the proposal of the AI Act, gradually translating the vision of “excellence + trustworthiness” into specific systems (Key policy milestones (2018–2021))→ Furthermore, it explains how regulation reversely promotes development: by enhancing user trust, providing legal certainty and unified rules, it expands market space rather than suppressing innovation (Regulation promotes development: Mechanisms and logic)→ On this basis, it introduces a risk-oriented hierarchical regulatory framework, setting prohibitions, mandatory compliance, transparency obligations and regular management requirements for unacceptable risk, high-risk, limited-risk and minimal-risk scenarios respectively (Risk classification and regulatory requirements)→ At the same time, it clarifies the roles of member states in the governance system: establishing national competent authorities and supervisory bodies responsible for implementation, market monitoring, and acting as interfaces with the public and the EU level (Roles of member states and governance framework)→ Finally, it emphasizes building an innovation ecosystem through funding programs, digital infrastructure and public-private partnerships, aiming to spread “excellent and trustworthy” AI applications under strict regulation and a unified market (Investment, infrastructure, and ecosystem)

2.4.4.1 Human-centered and “Trustworthy AI”

  • Principle: AI should only be developed when it benefits humans

  • Approach: Strictly assess benefits and risks before approving specific applications

  • Reliance: Central regulatory framework (regulation first)

2.4.4.2 Key policy milestones (2018-2021)

  • 2018: European Artificial Intelligence Strategy; Member states signed the AI Coordination Plan

  • 2019: Guidelines for Trustworthy AI

  • 2020:

    • AI White Paper: Proposes the “Excellence + Trustworthy” vision to pave the way for regulations
    • Safety and Liability Impact Report: Clarifies open issues related to product safety in AI/IoT/robotics
  • 2021: Proposal for the AI Act: Unify AI rules and establish a strict regulatory framework

  • Position Summary (2021):

    • Vestager:“Trustworthiness is not an option, but a strict requirement.”
    • Breton:AI is a tool, to be used only when beneficial

2.4.4.3 Regulation promotes development: Mechanisms and logic

  • Increased user trust promotes demand from enterprises and public sectors

  • Legal certainty and unified rules allow suppliers to enter broader markets and make products more popular

  • Conclusion: Without avoiding risks, ensuring trustworthiness, and legal safeguards, AI cannot develop effectively

2.4.4.4 Risk classification and regulatory requirements

  • Unacceptable Risk (Prohibited)

    • A few applications that violate European values and infringe on fundamental rights

    • Examples: Government social scoring; exploiting children’s vulnerabilities; subliminal manipulation; real-time remote biometrics for law enforcement in public places (with very few exceptions)

  • High Risk (Mandatory Compliance)

    • Systems that adversely affect safety or fundamental rights

    • Mandatory requirements: Data quality; technical documentation and records; transparency and user information; human oversight; robustness and accuracy; cybersecurity

    • The list can be adjusted as use cases evolve

  • Limited Risk (Transparency Obligation)

    • Example: Scenarios with manipulation risks (e.g., chatbots) – users must be informed of “interacting with a machine”
  • Minimal Risk (Routine Management)

    • Other AI systems are developed and used in accordance with existing laws, with no additional obligations

    • Suppliers can voluntarily follow Trustworthy AI requirements and codes of conduct

2.4.4.5 Roles of member states and governance framework

  • National Competent Authorities: Supervise implementation and enforcement, conduct market monitoring

  • National Supervisory Authorities: Official contact points with the public/stakeholders; represent their countries in the European AI Committee

2.4.4.6 Investment, infrastructure, and ecosystem

  • Funding Channels: Initiatives like “Digital Europe” and “Horizon Europe” (20% of expenditure on digital transformation)

  • Measures: Encourage sharing of information/data/computing infrastructure; public-private cooperation; promote digital innovation and talent mobility (from lab to market)

  • Goal: Promote the diffusion of excellent and trustworthy AI applications under strict regulation and a unified market