Ethical Considerations in Using LLMs

The good, the Bad, and the Ugly

Alex van Vorstenbosch

2025-02-01

Overview

__Disclaimer__
I don’t claim to have the answers.
Be aware of these topics.
Form your own opinions and openly discuss issues.

Overview

  • Biases and Misinformation
  • The Dark Side of LLMs
  • Group discussion

Biases and Misinformation

Biases

  • LLMs can strengthen negative stereotypes.
  • LLMs can strengten the views of users via confirmation bias:
    • LLMs have a tendency to agree with the user
    • People have a tendency to think that models are ‘objective’ and speak the ‘truth’
  • LLMs are ‘skewed’ to the trainingset majority:
    • English Western views for example

Hallucinating false information

  • What are Hallucinations?
    • It’s when LLMs generates incorrect, nonsensical, or unverifiable information presented as fact.
    • Might also be answers that are not supported by the provided context
    • Can be hard to spot as the model is a great ‘bluffer’
      • Doesn’t know that the information is wrong
      • asking for self-reflection does help

Hallucinating false information

  • What are Hallucinations?
    • It’s when LLMs generates incorrect, nonsensical, or unverifiable information presented as fact.

Who is responsible when AI makes a mistake?

Stories of dangerous behaviour by AI chatbots

Screenshots of Business Insider’s disturbing conversation with “Eliza,” a chatbot from Chai Research.

Stories of dangerous behaviour by AI chatbots

Importance of human-in-the-loop

  • Quality Control: Because of the generative design LLMs are prone to errors. Require human checks and feedback to ensure accuracy of the output.
  • Human (ethical) Judgement: Some decisions require human (ethical) judgment, especially in complex, nuanced situations where the context matters.

My personal beliefs:

  • Generative AI is an amazing transformative tool, but not an autonomous agent
  • You are responsible for the mistakes you make when using generative AI:
    • Make sure the risks are known
    • Make sure the risks are manageble
    • If not possible, make sure the risks are acceptable

What constitutes appropriate content?

Model Source Restrictions
ChatGPT by OpenAI Closed source Strongly moderated and curated
Grok by Xai Closed source Less restrictions
Llama-models by Meta Open source Can be finetuned for any purpose

Keep in mind: Nobody is sharing the most important part: HIGH QUALITY DATA

A reddit pol on the most annoying ChatGPT responses

Overview

  • Biases and Misinformation
  • The Dark Side of LLMs
  • Group discussion

Overview

  • Biases and Misinformation
  • The Dark Side of LLMs
  • Group discussion

The Dark Side of LLMs

Privacy issues and LLMs

  • Can companies train LLMs on (scraped) private data without consent?
    • What if LLMs memorise private data?
  • How can we mitigate inference of private information by LLMs?
  • How can we trust third-parties with our proprietary/private information?

Current view Dutch-government

Niet-gecontracteerde generatieve AI-toepassingen voldoen over het algemeen niet aantoonbaar aan de geldende privacy- en auteursrechtelijke wetgeving. Zodoende is het gebruik hiervan door Rijksorganisaties (of in opdracht daarvan) niet toegestaan, in die gevallen waarin het risico bestaat dat wetgeving wordt overtreden, tenzij de aanbieder en de gebruiker aantoonbaar voldoen aan de geldende wet- en regelgeving.

Transparency issues of LLMs

  • How can we trust models that are “black boxes”?
    • Especially if aren’t even sure what these models look like or how they were trained?
  • How can these models be used if they can generate ‘hallucinations’ at any point?
  • How can we prevent the use of LLMs for unsuited usecases?

Misuse of LLMs

  • How can we prevent the automated generation of misinformation at scale?
  • How can we prevent the use of these techniques for spam, identity fraud, and worse?
  • Who should decide what misuse of LLMs means?

Are these AI developments safe?

Are these AI developments safe?

Are these AI developments safe?

BI: Google Brain cofounder says Big Tech companies are inflating fears about the risks of AI wiping out humanity because they want to dominate the market

Should there be governmental oversight on AI

EU AI Act

Passed in 2024, main effects:

  • Unacceptable risk AI systems will be banned
    • Real time biometric identification
    • Behavioural manipulation
    • Social scoring systems
  • Limited Risk AI needs to be transparant:
    • You must know if you are interacting with AI
    • Companies must disclose if content was generated with AI
    • Chatbots are classified as limited risk
  • The setup of a new European AI Office to coordinate compliance, implementation, and enforcement of the AI Act
    • Tasked with oversight of General Purpose AI models across Europe

The economic impact of AI

  • Will it take (many of) our jobs?
  • Will it create jobs?
  • Or will it just make us more efficient at our current job?

Climate impact of large language models

Training the model via RLHF

  • Low-wage workers in Kenia were paid to help collect data for the ‘moderation’ tool:
    • Traumatising work

Use of LLMs for essays, homework, etc. cannot be reliably detected.

  • AI-detectors don’t work, which is creating serious issues for students.
  • AI-detectors don’t work, which is disrupting how homework is given and made.

Overview

  • Biases and Misinformation
  • The Dark Side of LLMs
  • Group discussion

Overview

  • Biases and Misinformation
  • The Dark Side of LLMs
  • Group discussion

Group discussion

  1. Can LLMs be used if they are trained on copyrighted and/or AVG-protected data?

  2. Should LLM usage be constrained by ethical guidelines and content filters?

  3. Can LLMs be trusted if hallucinations are an inherent part of these systems?

Discuss within your group for 5-10 minutes , and then we will discuss the results in a plenary session.