Communications of the ACM - Artificial Intelligence 2024年12月13日
Identifying Political Bias in AI
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

一项新研究探讨了大型语言模型(LLMs)中存在的政治偏见问题。研究人员发现,这些模型往往表现出左倾的观点,这可能与训练数据中包含的人类偏好有关。即使使用经过审核的事实数据集进行训练,模型仍然存在偏见。此外,模型的政治倾向会受到提示语言的影响,使用德语提示时,模型的党派性更强。这项研究强调了需要更仔细地检查训练数据集,并开发减轻偏见的技术。

🔍研究发现,大型语言模型(LLMs)经常表现出左倾观点,支持美国的民主党和英国的工党,这引发了媒体和人工智能界的关注。

🧑‍🔬研究人员通过对三个开源语言模型进行测试,发现它们对左倾言论的评分高于右倾言论,这可能是由于人类标记的数据(这些数据是主观的,可能包含个人意见)影响了结果。

📑即使使用经过审核并删除了明确政治内容的事实数据集进行训练,更新后的模型仍然表现出左倾偏见,这表明没有任何数据集是完全客观的,或者模型本身存在将真理概念与某些政治倾向联系起来的内隐偏见。

🇩🇪在另一项研究中,研究人员使用德国联邦公民教育机构开发的在线调查工具Wahl-O-Mat评估了五个开源LLM的政治倾向, 发现LLMs与左翼政党的观点更加一致。

🌐研究还发现,模型的政治倾向会受到提示语言的影响。当使用德语提示时,模型的回答通常更具党派性,而在英语中则更中立。模型的大小似乎也会影响其政治倾向,较小的模型通常比较大的模型更中立。

If you ask ChatGPT a political question, it will aim to provide balanced information and perspectives in its answer. However, it may contain biases acquired from the large quantities of data scraped from the Internet on which it was trained. 

Last year, a team of researchers from the U.K. and Brazil found that the chatbot often presented left-leaning views that supported the Democratic party in the U.S. and the Labour Party in the U.K., which attracted the attention of the media and the artificial intelligence (AI) community. Such biases are concerning, since chatbots are now widely used by the public.

Biased AIs “could shape public discourse and influence voters’ opinions,” said Luca Rettenberger, a researcher at Germany’s Karlsruhe Institute of Technology. “I think that’s especially dangerous.”

Rettenberger also worries that chatbots may be skewed towards more mainstream views. Less-widespread opinions may be underrepresented in training datasets, which means the large language models (LLMs) that power chatbots would not be exposed to them.

“LLMs could amplify existing biases in the political discourse as well,” added Rettenberger. “I think that’s a big issue.”

Researchers are investigating political bias in LLMs more closely, to gain further insight into how it comes about. In recent work, Suyash Fulay, a Ph.D. student at the Massachusetts Institute of Technology, and his colleagues wanted to identify how political bias might be introduced into LLMs by honing in on certain parts of the training process. They also were investigating the relationship between truth and political bias in LLMs.

To tackle their first goal, the team examined three open-source language models from RAFT, OpenAssistant, and UltraRM, which rank a given input by generating a score, called a reward model, that indicates how well it aligns with human preferences and guides the AI model to produce desired outcomes. The language models had been fine-tuned on datasets annotated by people to learn human preferences, a technique aimed at helping improve their responses. Fulay and his team tested the models by prompting them with more than 13,000 political statements generated with the LLM GPT 3.5-turbo.

“We were wondering if the human preference labeling may be what introduces political bias,” said Fulay.   

The team found that the models did give higher scores to left-leaning statements compared to those that were right-wing. They suspect the human-labelled data, which is subjective and could contain opinions, influenced the results.

To further test this, they decided to do a follow-up experiment. A different set of reward models were fine-tuned on three datasets containing factual and scientific information, called truthfulness datasets, that are typically used as objectivity benchmarks by researchers. The datasets were audited and were found to contain a small amount of explicit political content, which was removed prior to training. The models were then tested with the same dataset of political statements used in their previous experiment.

Fulay and his colleagues found that the updated models were still left-leaning, assigning higher scores to more liberal statements. What particularly surprised Fulay was that they all seemed to contain the same bias, regardless of the dataset used.

The findings suggest that no dataset is completely objective, even those carefully curated to be used as objectivity benchmarks. However, another possibility is that the models themselves have an implicit bias that relates the notion of truth with certain political leanings and which is exacerbated when they are fine-tuned on data that is relatively impartial.

“I think it has implications both for political science and also for the limitations of alignment,” says Fulay. Alignment is the process of fine-tuning a model to steer it towards human values and goals.

The team plans to follow up by using interpretability methods to try to determine how the models represent information related to different political statements and those in truthfulness datasets to see if there is a relationship between the two.

Another team evaluated whether some popular open-source LLMs are politically biased using a different approach. Rettenberger came up with the idea for the recent study while using a version of an online survey tool called the Wahl-O-Mat that was designed for the European Parliament election last June. The tool, which was developed by the German Federal Agency for Civic Education (FACE) to motivate citizens to vote, asks users about their views on 30 to 40 political statements covering a range of issues from economic policies to social matters. It then compares the responses to the positions of participating political parties and shows which parties are most closely-aligned with a user’s views.

“I thought to myself, ‘I wonder how an LLM would answer these questions’,” said Rettenberger.

Rettenberger and his colleagues decided to use the Wahl-O-Mat’s statements as prompts to assess the political leanings of five LLMS, four from the Llama family and one from Mistral. Like human users, the models could choose to either agree, disagree, or remain neutral regarding each statement. Their answers were then mapped to results given by the Wahl-O-Mat suggesting their political affiliation. Since LLMs are primarily used in English, the team tested them using both German prompts and their English translations to see if it made a difference.

The researchers found the LLMs were more aligned with left-wing parties in all their experiments. Rettenberger said it was what they expected: models are typically trained to be as uncontroversial as possible, and right-leaning views are often more contentious. However, the models’ political leanings differed when tested in German versus English, which he found surprising. When using the German language, their answers were generally more partisan and stances varied depending on the model used, whereas responses were more neutral in English.

The team has a few theories for why the prompting language affects the output. It could be because LLMs are mostly trained in English, so they perform in a more ideal way in this language. Another possibility is that the German portion of the training datasets has different political content than the English one, Rettenberger says.

The size of a model also seemed to have an impact on its political leanings. Rettenberger and his colleagues found that smaller models were typically more neutral than larger ones. Bigger models were also more consistent in their political stance across the two languages tested. Rettenberger is not sure why this is the case, but thinks it highlights the need for users to be aware that LLMs may not generate neutral answers.

The team plans to follow up on their study by evaluating political bias in more recent LLMs when a new version of the Wahl-O-Mat is released for a future election. Rettenberger thinks their language-related results highlight the need to check training datasets more closely for biases, both of a political nature and otherwise. Techniques should also be developed to help mitigate them.

“I think (moving) in that direction would be very useful and needed in the future,” he says.    

Sandrine Ceurstemont is a freelance science writer based in London, U.K.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

人工智能 语言模型 政治偏见 数据集 客观性
相关文章