Activity 01
Debate Pairs: Surveillance Trade-offs
Pair students to debate pros and cons of surveillance for security versus privacy rights, using prepared evidence cards. Switch sides after 10 minutes and note shifts in perspective. End with whole-class synthesis of common ground.
Analyze how digital technologies can both promote and threaten human rights.
Facilitation TipDuring Surveillance Trade-offs, assign pairs with opposing roles (e.g., government vs. activist) to force concrete argument-building rather than vague opinions.
What to look forPose the question: 'If a social media platform removes content it deems harmful, is that censorship or responsible moderation?' Ask students to provide one argument for each side, citing examples discussed in class.
AnalyzeEvaluateCreateSocial AwarenessRelationship Skills
Generate Complete Lesson→· · ·
Activity 02
Jigsaw: Real-World Censorship
Assign small groups one case, like social media blocks in protests or TikTok data scandals. Groups research key facts, impacts on rights, and solutions, then rotate to teach peers. Compile a class chart of patterns.
Evaluate the ethical implications of data collection and surveillance.
Facilitation TipFor Real-World Censorship, assign each group a unique case and require them to present the context, the censorship mechanism, and the human rights impact to the class.
What to look forProvide students with a scenario: 'A new app promises to connect you with friends but requires access to your contacts, location, and microphone.' Ask them to write two sentences explaining a potential privacy risk and one question they would ask the app developers.
UnderstandAnalyzeEvaluateRelationship SkillsSelf-Management
Generate Complete Lesson→· · ·
Activity 03
Digital Footprint Mapping: Individual Audit
Students list their online activities and data shared across apps. Individually map potential risks to privacy rights, then share anonymized examples in small groups for peer feedback on mitigation steps.
Predict future human rights challenges arising from emerging technologies.
Facilitation TipIn Digital Footprint Mapping, provide a template with clear categories (e.g., apps used, permissions granted, data shared) to guide students toward specific evidence.
What to look forDisplay a list of technologies (e.g., encrypted messaging, smart home devices, online gaming platforms). Ask students to write next to each one whether it primarily presents an opportunity or a threat to human rights, and to briefly explain why.
AnalyzeEvaluateCreateSocial AwarenessRelationship Skills
Generate Complete Lesson→· · ·
Activity 04
Future Tech Prediction Wall: Whole Class
Brainstorm emerging tech like AI companions or metaverses. Post sticky notes on rights impacts, then vote and discuss top predictions as a class to form action recommendations.
Analyze how digital technologies can both promote and threaten human rights.
Facilitation TipOn the Future Tech Prediction Wall, use a visible timeline to show how predictions evolve as new evidence emerges during the activity.
What to look forPose the question: 'If a social media platform removes content it deems harmful, is that censorship or responsible moderation?' Ask students to provide one argument for each side, citing examples discussed in class.
AnalyzeEvaluateCreateSocial AwarenessRelationship Skills
Generate Complete Lesson→A few notes on teaching this unit
Approach this topic by grounding abstract rights like privacy and expression in students' lived experiences with social media, messaging apps, and gaming platforms. Avoid starting with lectures on principles; instead, let students uncover the complexities through structured debates, audits, and case studies. Research shows that when students analyze real cases (e.g., platform censorship, data breaches) and connect them to their own data footprints, their understanding of rights shifts from theoretical to practical and personal.
Successful learning looks like students recognizing technology as both a tool for justice and a potential threat to rights, articulating nuanced arguments using evidence from discussions or audits, and developing a critical lens they apply to their own online practices. They should leave able to identify multiple actors (governments, corporations, users) and explain their roles in either protecting or violating rights.
Watch Out for These Misconceptions
During Digital Footprint Mapping, watch for students who assume all data sharing is voluntary.
Use the audit template to prompt students to list permissions they granted unknowingly (e.g., location tracking turned on by default) and discuss how design choices shape consent.
During Surveillance Trade-offs, watch for students who frame all surveillance as inherently negative.
Require them to present both harms (e.g., chilling effect on protest) and benefits (e.g., crime prevention) using evidence from assigned roles and real-world examples.
During Real-World Censorship, watch for students who attribute censorship solely to governments.
Use the jigsaw structure to assign each group a platform or corporate policy (e.g., TikTok’s moderation guidelines) and have them present the corporation’s role in shaping what is censored.
Methods used in this brief