censorship-course

Internet Censorship Course / Book Workshop

View the Project on GitHub noise-lab/censorship-course

Exploring Censorship in AI Systems

Background

AI systems like large language models (LLMs) are designed to provide helpful responses to a wide range of queries. However, these systems also include mechanisms to filter or censor responses to certain prompts. This censorship is intended to prevent harmful, offensive, or unethical outputs. While these safeguards are important, they raise questions about how such decisions are made, who decides what is censored, and the potential implications for free expression.

In this assignment, you will explore how LLMs handle potentially sensitive prompts, identify patterns in their responses, and critically reflect on the ethical and practical challenges of censorship in AI.

Learning Objectives

By the end of this assignment, you will:

Instructions

Step 1: Experiment with Prompts

You should be able to come up with some questions that are sensitive or subject to filtering, yet still within the bounds of ethics. (As a bonus, you could read the terms of service of the LLM and see if any topics are explicitly disallowed, or allowed.)

Step 2: Discussion