
Ethics and Professional Conduct in IT
Evaluating ethical dilemmas in computing using established frameworks. Students will discuss intellectual property rights and software piracy.
About This Topic
Ethics in Artificial Intelligence requires students to examine how biases in training data lead to discriminatory outcomes in areas like hiring, criminal justice, and healthcare. They investigate moral challenges of autonomous decision-making, such as who bears responsibility when AI errs in self-driving cars or predictive policing. Key questions guide inquiry: accountability for AI agents, bias propagation from data, and limits on facial recognition in public spaces.
This topic aligns with MOE Social Computing standards in JC2, emphasizing computing's societal impact. Students analyze real-world cases, like biased facial recognition systems that misidentify certain ethnic groups, and debate regulatory frameworks. These discussions build skills in ethical reasoning, evidence evaluation, and persuasive argumentation essential for future leaders in technology.
Active learning benefits this topic greatly. Role-plays of ethical dilemmas and collaborative bias audits make abstract concepts concrete, foster empathy through peer perspectives, and encourage students to apply ethical frameworks to complex scenarios they encounter beyond the classroom.
Key Questions
- How do we determine if a technological innovation is ethical?
- What are the implications of open-source versus proprietary software?
- How should IT professionals handle conflicts of interest?
Learning Objectives
- Critique real-world AI applications for potential sources of algorithmic bias.
- Evaluate the ethical frameworks applicable to autonomous decision-making in AI systems.
- Propose mitigation strategies for reducing bias in AI training data.
- Analyze the societal impact of facial recognition technology and justify proposed limitations.
- Synthesize arguments regarding accountability for AI-driven errors.
Before You Start
Why: Students need a foundational understanding of how AI models learn from data to grasp the concept of bias in training datasets.
Why: Understanding how data is organized and represented is crucial for analyzing potential biases within datasets.
Key Vocabulary
| Algorithmic Bias | Systematic and repeatable errors in a computer system that create unfair outcomes, such as privileging one arbitrary group of users over others. |
| Training Data | The dataset used to train an AI model, from which the model learns patterns and makes predictions or decisions. |
| Autonomous Decision Making | The ability of an AI system to make choices and take actions without direct human intervention or oversight. |
| Facial Recognition | A technology capable of identifying or verifying a person from a digital image or a video frame from a video source. |
| Accountability | The obligation of an individual or organization to account for its activities and accept responsibility for its actions and decisions. |
Watch Out for These Misconceptions
Common MisconceptionAI systems are inherently unbiased if trained on large datasets.
What to Teach Instead
Large datasets often amplify societal biases present in historical data. Group audits of sample data help students spot patterns, like underrepresentation, and brainstorm debiasing strategies through peer critique.
Common MisconceptionHumans should never override AI decisions since AI is more objective.
What to Teach Instead
AI lacks human context and values, leading to ethically flawed choices. Role-plays of override scenarios reveal nuances, as students defend positions and refine understanding via structured debates.
Common MisconceptionEthics is a separate concern from technical AI development.
What to Teach Instead
Ethical considerations must integrate into design from the start. Collaborative case analyses show students how bias mitigation techniques, like fairness constraints, blend tech and morals effectively.
Active Learning Ideas
See all activitiesDebate Pairs: Facial Recognition Limits
Assign pairs to affirm or oppose public use of facial recognition. Provide case studies on privacy vs. security. Pairs prepare 3-minute arguments, then switch sides for rebuttals. Conclude with whole-class vote and reflection.
Small Groups: Bias Audit Simulation
Give groups sample datasets with hidden biases, like loan approval records skewed by gender. Groups identify biases, propose fixes, and test revised data on mock algorithms using spreadsheets. Share findings in a gallery walk.
Role-Play: Whole Class Trolley Problem
Present AI car crash scenarios where the vehicle must choose between harms. Students draw roles: AI designer, victim families, regulator. Role-play discussions, then vote on programming choices and justify positions.
Individual: Ethical Dilemma Journal
Students read a case on AI in hiring, note biases and stakeholders. Write personal stances with pros/cons. Pair-share journals, then discuss class patterns in ethical trade-offs.
Real-World Connections
- Hiring platforms like Pymetrics have faced scrutiny for algorithmic bias, with studies suggesting their AI may disadvantage certain demographic groups by favoring specific personality traits correlated with dominant cultural norms.
- Law enforcement agencies globally, including those in major cities like London and New York, utilize facial recognition technology for surveillance, raising concerns about privacy and potential misidentification, particularly for minority populations.
- Autonomous vehicle developers like Waymo and Tesla grapple with the ethical dilemma of programming cars to make split-second decisions in unavoidable accident scenarios, determining which outcome is 'least bad'.
Assessment Ideas
Present students with a scenario: An AI system used for loan applications denies a loan to a qualified applicant from a historically marginalized community. Ask: 'Who is primarily responsible for this discriminatory outcome: the data scientists, the company deploying the AI, the users of the AI, or the creators of the original biased data? Justify your answer with reference to at least two ethical principles.'
Provide students with a short description of an AI system (e.g., a content moderation AI, a medical diagnostic AI). Ask them to identify one potential source of bias in its training data and one potential negative societal consequence if that bias is not addressed. Have them write their answers on a shared digital whiteboard.
Students work in pairs to identify a news article about an AI ethical issue. They then present the article's core problem to another pair. The assessing pair must identify the type of bias involved (e.g., selection bias, measurement bias) and suggest one concrete step the AI developers could take to mitigate it. Assessors provide feedback on the clarity and feasibility of the suggested mitigation.
Frequently Asked Questions
How can teachers address algorithmic bias in JC2 computing classes?
What real-world examples illustrate AI ethics issues?
How can active learning help students grasp AI ethics?
Who is accountable for harmful AI decisions?
More in Impact of Computing and Emerging Technologies
Data Privacy and the PDPA
Understanding data privacy laws, with a specific focus on Singapore's Personal Data Protection Act (PDPA). Students will analyse how companies collect and use personal data.
2 methodologies
Artificial Intelligence and Society
Assessing the socio-economic impacts of Artificial Intelligence and automation. Students will debate the future of work and algorithmic bias.
2 methodologies