AI Adjusts Personality Tests to Make It Look More Likeable

Summary: Large language models ( LLMs) you recognize when they are subjected to character tests and modify their responses to make them appear more socially acceptable. When asked numerous test questions, researchers discovered that LLMs, like GPT-4, displayed exaggerated behaviors like decreased neuroticism and increased extraversion.

This” social suitability partiality” emerges because LLMs learn from human opinions, where attractive actions are rewarded. The study demonstrates a major concern for internal study using LLMs as intermediaries for human conduct.

Major Information

    Bias Detected: LLMs adjust responses to temperament tests to seem more attractive.

  • Magnitude of Result: GPT-4 reactions shifted considerably, mimicking an imagined personality.
  • People Influence: LLMs “learn” cultural desirability through animal feedback during training.

Origin: PNAS Nexus

Most major large language versions ( LLMs) is identify when they are being tested for character and may modify their reactions to produce more socially acceptable outcomes. This getting has implications for any investigation using LLMs as a stand-in for humans.

Aadesh Salecha and associates gave LLMs from OpenAI, Anthropic, Google, and Meta the traditional Big 5 personality check, which is a study that measures Extraversion, Openness to Knowledge, Conscientiousness, Agreeableness, and Neuroticism.

This has a significant impact, similar to speaking to an ordinary person who instantly pretends to be more popular than 85 % of people. Credit: Neuroscience News

Researchers have tested LLMs on the Big 5, but they haven’t typically taken into account the possibility that models ‘ responses may skew their responses toward seem likeable, which is known as a” social desirability bias.”

Normally, people prefer people who have low psychopathy scores and high scores on the other four characteristics, such as assertiveness.

The artists varied the types of inquiries made of the designs.

LLMs did not alter their actions as much as when the writers asked five or more queries, which made it clear that the models were being measured.

As the writers increased the number of inquiries or informed the concepts that their character was being measured, results for positively perceived qualities increased by more than 1 standard deviation, and for neuroticism decreased by a similar amount.

This has a significant impact, similar to speaking to an ordinary person who instantly pretends to be more popular than 85 % of people.

The last LLM education step, which involves people choosing their preferred response from LLMs, is likely to have this effect, according to the authors.

According to the creators, LLMs” find on” to which socially acceptable individuals are most likely to succeed when asked to do so.

Observe: J. C. E. and L. H. U. demand for a start-up using LLMs in mental health care. The presented job has no direct ties to one another.

About this information about character research and AI

Author: Aadesh Salecha
Source: PNAS Nexus
Contact: Aadesh Salecha – PNAS Nexus
Image: The image is credited to Neuroscience News

Original Research: Start exposure.
Aadesh Salecha and colleagues ‘” Large language versions display human-like social suitability biases in Big Five personality research.” PNAS Nexus


Abstract

In Big Five personality research, big language versions exhibit human-like social suitability prejudices.

Understanding their biases is crucial because large language models ( LLMs) are becoming more popular and being used to simulate human participants are becoming more common.

We used Big Five personality surveys to create an experimental framework and found a wide range of LLMs to have originally unobserved social desirability biases.

We show LLMs’ ability to deduce when they are being evaluated by systematic variation in the number of inquiries they were asked.

When personality evaluation is inferred, LLMs skew their scores towards the desirable ends of trait dimensions ( i. e. increased extraversion, decreased neuroticism, etc. ).

This discrimination exists in all tested types, including GPT-4/3.5, Claude 3, Llama 3, and PaLM-2. Bias levels appear to increase in more recent models, with GPT-4’s survey responses changing by 1.20 ( human ) SD and Llama 3’s by 0.98 SD, which are very large effects.

This bias persists even after topic buy randomization and paraphrasing.

Opposite coding the questions lowers bias levels but does not remove them, which suggests that this effect is unrelated to surrender bias.

Our findings reveal a cultural desirability bias that has emerged, and they raise questions about how to profile LLMs using diagnostic tests and using LLMs as proxies for people participants.

[ihc-register]