Skip to content

Digital Ethics and SurveillanceActivities & Teaching Strategies

Active learning lets students confront real stakes in digital ethics, turning abstract privacy concerns into tangible evidence. When they simulate data tracking or role-play algorithm court cases, students see how surveillance reshapes behavior and decisions in their own lives.

Year 10Technologies4 activities35 min50 min

Learning Objectives

  1. 1Analyze the ethical implications of data collection in public surveillance systems.
  2. 2Evaluate the trade-offs between technological convenience and individual privacy in digital environments.
  3. 3Critique the potential biases and fairness issues within predictive policing algorithms.
  4. 4Compare different models of data ownership for digital interactions.
  5. 5Design a user-centered policy proposal addressing a specific digital ethics concern.

Want a complete lesson plan with these objectives? Generate a Mission

50 min·Small Groups

Debate Carousel: Surveillance Pros and Cons

Divide class into four groups, each assigned a viewpoint: privacy advocates, tech companies, governments, citizens. Groups prepare arguments for 10 minutes using provided case studies, then rotate to defend or rebut positions. Conclude with a whole-class vote and reflection on shifted opinions.

Prepare & details

Who owns the data generated by your digital interactions?

Facilitation Tip: During the Debate Carousel, assign clear roles and time limits to keep discussions focused on balancing surveillance benefits and privacy risks.

Setup: Chairs arranged in two concentric circles

Materials: Discussion question/prompt (projected), Observation rubric for outer circle

AnalyzeEvaluateCreateSocial AwarenessRelationship Skills
40 min·Pairs

Data Trail Simulation: Track Your Digital Footprint

Students log a day's digital activities on worksheets, then trace how data flows to third parties using flowcharts. In pairs, they map privacy risks and propose anonymization strategies. Share findings in a class gallery walk.

Prepare & details

How does constant surveillance change human behavior?

Facilitation Tip: In the Data Trail Simulation, provide students with a non-judgmental space to reflect on their own digital habits and the hidden data flows behind seemingly free services.

Setup: Chairs arranged in two concentric circles

Materials: Discussion question/prompt (projected), Observation rubric for outer circle

AnalyzeEvaluateCreateSocial AwarenessRelationship Skills
45 min·Small Groups

Predictive Policing Role-Play: Algorithm Court

Assign roles as algorithm developers, affected communities, and judges. Groups present biased algorithm scenarios, deliberate on fixes, and vote on redesigns. Debrief on real ethical standards like transparency.

Prepare & details

What are the dangers of predictive policing algorithms?

Facilitation Tip: During the Predictive Policing Role-Play, give students a short script to ensure they understand the algorithm’s decision-making process before they critique it.

Setup: Chairs arranged in two concentric circles

Materials: Discussion question/prompt (projected), Observation rubric for outer circle

AnalyzeEvaluateCreateSocial AwarenessRelationship Skills
35 min·Individual

Privacy Audit: App Review Challenge

Individuals select a common app, review its privacy policy in pairs, and score it on criteria like data sharing. Compile scores into a class spreadsheet for patterns, then brainstorm better designs.

Prepare & details

Who owns the data generated by your digital interactions?

Facilitation Tip: For the Privacy Audit, provide a simple rubric that separates app permissions from user benefits so students can evaluate trade-offs fairly.

Setup: Chairs arranged in two concentric circles

Materials: Discussion question/prompt (projected), Observation rubric for outer circle

AnalyzeEvaluateCreateSocial AwarenessRelationship Skills

Teaching This Topic

Teachers should model skepticism by sharing their own digital dilemmas, such as why they choose certain apps despite privacy concerns. Avoid presenting surveillance as purely negative or positive; instead, frame it as a system with designers, users, and unintended outcomes. Research suggests students grasp bias in algorithms best when they trace data back to real people’s experiences, so center activities on lived consequences rather than technical details alone.

What to Expect

Successful learning shows when students move from passive acceptance to critical questioning of data use and its consequences. They should articulate trade-offs between safety and privacy, identify bias in algorithms, and propose ethical alternatives with clear reasoning.

These activities are a starting point. A full mission is the experience.

  • Complete facilitation script with teacher dialogue
  • Printable student materials, ready for class
  • Differentiation strategies for every learner
Generate a Mission

Watch Out for These Misconceptions

Common MisconceptionDuring the Data Trail Simulation, watch for students who assume their personal data remains private because they haven’t sold it themselves.

What to Teach Instead

Use the simulation’s data flow maps to show how app permissions allow aggregation and resale to third parties without direct consent. Ask students to trace their simulated data to at least two unseen recipients and label each step with a risk level.

Common MisconceptionDuring the Debate Carousel, watch for students who claim surveillance only affects people who break rules.

What to Teach Instead

In role assignments, require debaters to present evidence from studies on shopping, social media, or school behavior to show how tracking influences everyone. Use the carousel’s rotating stations to expose students to multiple perspectives before they solidify arguments.

Common MisconceptionDuring the Predictive Policing Role-Play, watch for students who accept algorithm outputs as neutral because they come from data.

What to Teach Instead

In the courtroom setup, provide biased training data samples and ask students to analyze how historical policing patterns skew predictions. Use role cards to force them to defend outcomes that disproportionately affect certain groups, making bias visible through their own arguments.

Assessment Ideas

Discussion Prompt

After the Debate Carousel, pose this scenario to small groups: 'Your school is considering AI-powered cameras to monitor student behavior and attendance. What are the potential benefits for safety and efficiency? What are the privacy concerns? Facilitate a debate where groups present arguments for and against the system. Assess by listening for evidence of trade-offs and ethical reasoning in their responses.'

Quick Check

During the Data Trail Simulation, present students with a scenario: 'A popular mobile app offers free service in exchange for location data and contact lists. Ask students to write down two potential benefits and two risks of using the app. Collect responses to gauge their understanding of data trade-offs and identify misconceptions about data ownership.'

Exit Ticket

After the Predictive Policing Role-Play, ask students to write on an index card: 1. One question they still have about data ownership. 2. One example of how constant surveillance might change their own behavior. 3. A brief description of one ethical challenge related to predictive policing. Use these to identify lingering confusion and adjust future lessons.

Extensions & Scaffolding

  • Challenge: Ask early finishers to design a privacy label for devices or apps, similar to nutrition labels, that clearly communicates data risks.
  • Scaffolding: For students struggling with abstract concepts, provide a partially completed data flow diagram to help them map connections between apps, permissions, and third-party sales.
  • Deeper exploration: Invite students to research a recent surveillance-related news story and prepare a 2-minute podcast segment explaining the ethical dilemma to the class.

Key Vocabulary

DataficationThe process of turning aspects of life into data that can be collected, analyzed, and monetized. This transforms everyday activities into measurable information.
Surveillance CapitalismAn economic system centered on the commodification of personal data, often collected through digital technologies. It prioritizes profit from data over user privacy.
Algorithmic BiasSystematic and repeatable errors in a computer system that create unfair outcomes, such as privileging one arbitrary group of users over others. This can occur in areas like facial recognition or loan applications.
Predictive PolicingThe use of data analysis and algorithms to identify potential criminal activity before it occurs. It aims to prevent crime by forecasting where and when it might happen.
Digital FootprintThe trail of data left behind by a user's online activity. This includes websites visited, emails sent, and social media posts, all contributing to a personal data profile.

Ready to teach Digital Ethics and Surveillance?

Generate a full mission with everything you need

Generate a Mission
Digital Ethics and Surveillance: Activities & Teaching Strategies — Year 10 Technologies | Flip Education