Exploring large language models’ biases in historical knowledge

  • @sarsaparilyptus
    link
    3
    edit-2
    11 months ago

    Here’s a condensed version of the article:

    Language Learning Model Outputs the Same Kind of Data as its Input

    New groundbreaking journalism has discovered that when you analyze a large pluratily of a given language’s written works, you get data that reflects the biases of the people who wrote it all down, most of whom died 100+ years ago. No word yet as to when the author plans to publish illuminating, hard-hitting investigative journalism into the effects of microwaves on popcorn.

    • HeartyBeast
      link
      fedilink
      511 months ago

      That’s the kind of ‘no shit Sherlock’ response that is common when people dismiss studies that have superficiallly ‘obvious’ results. It doesn’t mean the study isn’t worthwhile to see whether hunches stand up to scrutiny.

      For anyone who wants a short summary - here’s ChatGPT’s attempt:

      The article you shared discusses an experiment where the user prompted GPT-4 and Anthropic’s Claude to list the top 10 important historical figures in 10 different languages. The analysis revealed several biases present in the language models’ responses:

      1. Gender bias: Both models disproportionately predicted male historical figures, with GPT-4 generating female figures 5.4% of the time and Claude doing so 1.8% of the time.

      2. Geographic bias: There was a bias towards predicting Western historical figures, with GPT-4 generating figures from Europe 60% of the time and Claude doing so 52% of the time.

      3. Language bias: Certain languages suffered from gender or geographic biases more than others.

      The article also highlights that the models exhibited a skewed view of history, focusing on political and philosophical figures while neglecting the diversity of professions that historical figures can encompass.

      Overall, the analysis calls attention to the lack of representation of women in historical figures and the Western-centric perspective encoded in the language models, emphasizing the importance of being mindful of biases when using such models in various settings.