AI Ethics: Exam-Style Questions
Eight original questions written to match the style and mark allocations of GCSE and A-Level Computer Science ethics papers. Work through them in exam conditions, then reveal the mark scheme.
Ask students: if a company trains a spam filter on emails from 2005, what kinds of modern spam might it miss - and why? Use this to illustrate how historical bias and selection bias can both be present at once.
Present this scenario: an AI loan approval system is 90% accurate for applicants from Group A and 70% accurate for Group B. Is that fair? What would making it equally accurate for both groups require? Use this to surface the tension between group fairness and individual accuracy.
A student scoring 4/4 identifies two distinct concerns, names them precisely, then explains the specific ethical harm: who is affected, how, and why that constitutes an ethical problem rather than just an inconvenience.
The Metropolitan Police began routine use of live facial recognition cameras in London in 2023. Ask: should the public be notified when a facial recognition camera is in use? What would notification change? Explore whether awareness removes the ethical concern or just the surprise.
A patient is deprioritised by the AI and their condition worsens while they wait. Who should be held responsible? Ask students to argue: the doctor who accepted the AI's recommendation, the hospital that deployed the system, or the company that built it. Can responsibility be genuinely shared?
The COMPAS reference is good for developing the bias point but is not required. Credit any accurately described real example of AI in criminal justice contexts. If students cannot recall an example, a hypothetical that correctly describes the mechanism still earns the mark.
Argue both sides: a judge uses COMPAS to help decide between a custodial and a community sentence. The AI recommends prison. Should the judge be able to override it freely? Or does overriding it undermine the point of using AI? What does this tell us about the appropriate role of AI in high-stakes decisions?
Present the Samsung incident: engineers accidentally leaked proprietary code by pasting it into ChatGPT. Ask: who is responsible - the individual, the company that allowed it, or the AI provider that retained the data? Extend: should workplaces be able to ban LLM use? What would need to be true for that ban to be proportionate?
Discuss the ethical implications of this proposal. [8 marks]
Indicative content (potential benefits / counterarguments): Human hiring decisions also contain bias (unconscious bias against names, appearance, accent); AI may actually be more consistent. Algorithmic decisions can in principle be audited and adjusted in ways that human intuition cannot. Efficiency allows large numbers of applicants to be considered fairly.
Indicative conclusion: The proposal as described - with no human review - is ethically indefensible for a decision of this significance. AI screening as a first filter with mandatory human review of borderline and rejected candidates would address many of the concerns while retaining efficiency benefits.
The Amazon hiring AI (discontinued 2018 after it was found to downgrade CVs containing the word "women") is the most directly relevant real-world example. Credit any accurate real example of automated recruitment and its outcomes.
Divide the class: half argue as the company's HR director (defending the AI system), half argue as a rejected candidate's lawyer. After five minutes, swap sides. Debrief: which arguments held up under challenge? Which collapsed? This surfaces the genuinely contested points and the weaker positions students held.
- Support: Provide students with a writing frame: Concern 1 (bias) ... This matters because ... A counterargument is ... However ...
- Stretch: Ask higher-ability students to reference specific legal protections (Equality Act 2010) and explain why the company's proposal creates a compliance risk beyond just an ethical one.
Indicative content (arguments for greater autonomy in some contexts): Speed - in some contexts (fraud detection, medical imaging analysis) human review is too slow; consistency - humans are fatigued and inconsistent in ways machines are not; in lower-stakes contexts, the efficiency gain may outweigh the oversight cost.
Indicative conclusion (Level 4 answers will articulate something like this): The appropriate level of AI autonomy should be proportionate to the stakes involved and the current state of AI reliability. High-stakes, irreversible decisions (criminal justice, healthcare, welfare benefits) require meaningful human oversight regardless of AI capability. Lower-stakes decisions may tolerate more automation. As AI explainability improves, the case for expanding autonomy in specific domains strengthens - but only with auditing mechanisms in place. The EU AI Act's prohibition on autonomous high-risk AI decisions without human oversight reflects this principle.
The EU AI Act (2024) explicitly prohibits certain autonomous AI applications and mandates human oversight for high-risk decisions. A student who references this accurately and explains the principle behind it (proportionality, stakes-based regulation) would be demonstrating Level 4 thinking. Credit any accurate reference to legal or regulatory frameworks governing AI autonomy.
Give students three scenarios: (1) an AI that filters spam email, (2) an AI that decides whether a welfare benefit application is approved, (3) an AI that determines bail conditions for a criminal defendant. Ask: at which point does human oversight become not just preferable but ethically required? What changes between scenario 1 and scenario 3? Use their answers to surface the principles of stakes, reversibility, and accountability.
- Support: Give students the three scenarios above and ask them to place each on a spectrum from "full AI autonomy acceptable" to "human oversight always required." Then ask them to justify their positioning in writing.
- Stretch: Ask students to write a one-paragraph summary of what the EU AI Act says about high-risk AI and explain whether they think the Act goes far enough, too far, or about right.