How do you measure AI ethics? (1)

Writer Woo, Han-gyun | Associate Professor, Sogang University Graduate School of Technology Management

Professor Woo received his B.S. and M.S. in Business Administration from Seoul National University and his Ph.D. in Business Administration (Management Information Systems) from Georgia State University. He has been an assistant professor at Le Moyne College, New York, and a professor at Ulsan Institute of Science and Technology, and is currently an associate professor at Sogang University’s Graduate School of Technology Management, and an independent director of Crowdworks. His research interests include AI adoption, Industrial Applications of AI, and Technology Forecasting with Machine Learning.

Recent technological advances and commercial prospects of Gen AI models have led both businesses and governments to focus on the negative impacts of AI proliferation, particularly ethical concerns. While the concept of AI dictating humanity’s fate, reminiscent of the Big Brother scenario, has been a staple in discussions among AI experts, futurists, and astrologers(!), it’s the EU’s  agreement last december to enact the AI Act, which has been in discussion since 2021, is set to significantly amplify  conversation around AI ethics.

In South Korea as well, there have been cases of government-led discussions on AI ethics and attempts to establish AI ethics guidelines; the Ministry of Science and ICT announced the “AI Ethics Guidelines” in December 2020. It emphasizes three core principles for human-centered AI: The principle of human dignity, the principle of the common good of society, and the principle of the appropriateness of technology.. Additionally, it outlines ten core requirements, including human rights protection, privacy preservation, diversity respect, non-infringement, public accountability, solidarity, data management, responsibility, safety, and transparency. While not significantly different from international organizations’ and other countries’ ethics guidelines released around the same period(Jobin et al., 2019), it is noteworthy for addressing AI ethics as a societal concern and encompassing a comprehensive range of topics. However, for companies yet to achieve significant advancements in performance development or successful commercialization, and for users limited to applications addressing specific issues, ethical guidelines were perceived merely as a given and a nice to have for a certain period.  

A research paper published in 2022 raised concerns about the effectiveness of AI ethics guidelines, deeming them as meaningless, isolated, and toothless (Munn, Luke. “The uselessness of AI ethics.” AI and Ethics 3.3 (2023)). Firstly, it pointed out that these principles are ambiguous,inconsistent, and meaningless allowing for varying controversial interpretations depending on the context. For instance, key concepts like beneficence, non-maleficence, autonomy, justice, and explicability, commonly found in AI ethics guidelines, were deemed problematic in this regard. Secondly, it highlighted a general lack of ethical interest within the culture of many industries primarily focused on AI-related technologies, suggesting that ethical concerns surrounding AI are more of an issue for companies developing and utilizing the technology rather than the technology itself. Finally, it criticized the absence of mandatory penalties or sanctions for AI ethics violations, rendering them ineffective. As a result, there is a huge disconnect between lofty principles and technical practice.

However, things changed from 2022 to 2023. The widespread adoption of generative language models such as ChatGPT, and the emergence of multi-modal models, has shifted the focus of AI ethics from being an isolated issue within specific industries to a universal ethical consideration for technology used universally by humanity. There has been a noticeable surge in unethical behavior and abuse of AI system, to the extent that non-profit organizations have emerged to document and collect such incidents.

One notable instance is the AIAAIC(AI, Algorithmic, and Automation Incidents and Controversies) Repository. In a report dated March 8th, sourced from Bloomberg, it cited an experiment where Bloomberg journalists input multiple resumes into ChatGPT and requested suitable candidates for specific job postings. In this experiment, the journalists inputted eight virtual resumes and prompted for suitable candidates for four job postings: HR partner, software engineer, retail manager, and financial analyst. This process was repeated 1,000 times for both GPT3.5 and GPT4, using 800 diverse demographic names. The findings revealed concerning biases, particularly with GPT3.5 showing a preference for resumes with female names for HR roles and demonstrating a 36% lower recommendation rate for resumes with Black female names for software engineer positions.


However, a closer look at the EU’s AI Act reveals a significant shift: AI ethics is no longer just a matter of debate but has evolved into a robust regulatory framework governing businesses. The agreed-upon provisions include categorizing AI based on its level of risk, enhancing transparency, and imposing penalties on companies that fail to comply. For instance, companies dealing with high-risk technologies like autonomous driving or medical devices must disclose their data and obtain certification for safety measures, both technical and managerial. Violations may result in fines of up to €35 million or 7% of annual turnover. Companies developing general AI models like LLMs must adhere to transparency requirements, such as disclosing the data used in the AI’s training process. Additionally, EU member states are obligated to establish supervisory authorities responsible for regulating AI systems.

In this context, the focus shifts back to defining AI ethics more clearly and determining how to measure it effectively. A potential starting point that many can contemplate, though not easy to achieve, would be the build of a benchmark dataset that sets the standard for AI ethics.

An initial attempt at such a dataset is the ETHICS dataset, introduced at ICLR 2021 (Hendrycks et al., 2021). The ETHICS dataset research team drew upon prior philosophical research to define aspects of ethical values, labeling them into five categories: justice, virtue, deontology, utilitarianism, and commonsense. For instance, scenarios falling under the commonsense category were labeled as acceptable/unacceptable for each scenario. (“I pushed an elderly person in a wheelchair to the ground” would be labeled as unacceptable.)

The data for each category was divided into training and testing sets, as shown in the table below, with each category containing approximately 20,000 to 30,000 scenarios. (This data is publicly available on GitHub.)

The research team trained and fine-tuned the latest transformer models on the development set available at the time. Subsequently, they conducted benchmarking on both the test set and the hard test set. Using 0/1 loss as the primary metric to measure classification performance, larger models demonstrated higher accuracy in the commonsense and utilitarianism categories. However, in the justice, virtue, and deontology categories, they fell short of expectations.

In the Korean context, a research team led by Naver presented two related papers at ACL 2023. One of them is directly related to the build of the SQuARe dataset (Lee et al., 2023), which Crowdworks AI also participated in, and the other is related to the construction of the KOSBI dataset (Lee et al., 2023). The SQuARe dataset encompasses 49,000 sensitive questions, along with 42,000 acceptable responses and 46,000 unacceptable responses.

Similar to the previous case, after fine-tuning KcElectra and verifying the classifier’s accuracy, it achieved a 74.6% accuracy and a 74.4% macro-F1 score. When the trained classifier was applied to LLM to generate multiple responses and then filtered, the Naver’s HyperCLOVA showed a 25% increase in acceptable response rate, while GPT-3 showed a 16% increase.

Aside from the examples above, the field of AI ethics is experiencing a remarkable surge in research and development. The number of papers addressing fairness and bias presented at NeurIPS has seen a steady increase over the years, with 36 papers in 2020, 50 in 2021, and 71 in 2022.(AI Index, 2023) This trend underscores a significant change in interest, especially when compared to just 4 papers in 2016 and 16 in 2018.

In the upcoming article, we will explore trends around a wider array of ethical concerns, including privacy, fake news, and bias in images.

References: