Mock Trial: A Student Project to Recreate and Critique an AI-Powered Court Decision
project-basedlegalethics

Mock Trial: A Student Project to Recreate and Critique an AI-Powered Court Decision

AAvery Morgan
2026-04-15
17 min read
Advertisement

A definitive student project guide for rebuilding an AI court model, critiquing it, and presenting policy recommendations to a mock oversight board.

Mock Trial: A Student Project to Recreate and Critique an AI-Powered Court Decision

This definitive classroom practice guide shows how to run an interdisciplinary student project where teams rebuild a simplified legal-tech decision model, interrogate its risks, and defend a mitigation plan before a mock oversight board. The goal is not to “gamify” justice; it is to help learners understand how AI systems can influence high-stakes decisions, why human oversight matters, and how to turn technical analysis into practical policy recommendations. If you are also designing project-based learning experiences, you may want to pair this with career exploration strategies for students and agile methods for managing complex student builds.

As a classroom project, this format blends case-based learning, technical modeling, ethics, law, and public speaking. It works in computer science, social studies, data science, pre-law, civic education, and capstone seminars. The most effective versions are structured like a real review process: define the system, test it, audit it, propose fixes, and present findings to a panel that plays the role of governance stakeholders. For background on responsible AI and organizational controls, see AI compliance frameworks and AI-powered workflows in technical environments.

1) Why This Project Works as a Classroom Practice

It turns abstract AI ethics into a concrete decision

Students often learn about bias, fairness, and explainability in the abstract, which can make the topic feel distant and moralizing. A mock trial anchored to a court decision model gives them a specific artifact to inspect: inputs, outputs, thresholds, and consequences. That shift from theory to evidence is powerful because it forces learners to ask what the model is optimizing, what it cannot see, and who bears the downside when it is wrong. In legal tech, those questions are not optional; they are the heart of trust.

It fits interdisciplinary instruction naturally

This project is inherently interdisciplinary because a justice-setting model sits at the intersection of statistics, civics, policy, and rhetoric. Students who are stronger in coding can recreate a simplified scoring algorithm, while students with strengths in writing can draft the critique and recommendation memo. Another group can handle the evidentiary presentation, cross-examination, or visual dashboards. This division of labor mirrors real workplace collaboration and makes the assignment especially effective for mixed-skill classrooms.

It teaches professional judgment, not just technical output

One reason employers value project-based learning is that it reveals how students handle uncertainty. In a model-audit project, a polished accuracy score is not enough; learners must explain tradeoffs, describe failure modes, and defend a mitigation strategy. That combination resembles what hiring managers want in entry-level analysts, policy associates, civic technologists, and product-minded developers. For students building a portfolio, the process also creates a strong artifact they can describe in interviews alongside skills developed through data analysis with Excel and data verification practices.

2) Define the Simulation: What Are Students Recreating?

Use a simplified decision model, not a real criminal justice tool

The project should not replicate an actual proprietary system or present a real court algorithm as if it were a classroom toy. Instead, students should build a simplified model inspired by how risk scoring or recommendation systems work in justice contexts. A good model might take a few variables, such as prior incidents, time since last incident, attendance compliance, or program participation, and generate a low/medium/high recommendation. The point is to demonstrate how a model can appear objective while still encoding assumptions that deserve scrutiny.

Keep the case-based learning scenario realistic

Choose a fictional or anonymized case file with enough detail to make the decision meaningful. For example, students could evaluate whether a hypothetical youth defendant should be recommended for diversion, supervised release, or a restorative program. The dataset should be small enough to understand manually, but large enough to reveal patterns, edge cases, and tradeoffs. If your students need help understanding how a public-facing model can shape user decisions, a useful analogy comes from user interface design in shopping experiences and educational content design, where layout and defaults quietly influence behavior.

Make the model explainable from the start

The best student projects use an intentionally transparent baseline model so learners can inspect every line of logic. A weighted scoring rubric is usually better than a black-box model in a classroom, because the goal is critique and mitigation, not high-performance prediction. Students should be able to show how each input affects the score, where the threshold sits, and what happens if a variable changes by one point. That transparency makes the critique more rigorous and the oversight board presentation more credible.

Project ElementWhat Students BuildWhy It MattersAssessment Focus
Decision modelSimple weighted score or rule-based classifierMakes logic inspectableCorrectness and clarity
Test casesRepresentative and edge-case profilesReveals bias and brittlenessCoverage and reasoning
Critique memoAudit of risks, assumptions, harmsConnects outputs to ethicsEvidence and argument quality
Mitigation planBias checks, human review rules, guardrailsShows practical responsibilityFeasibility and specificity
Board presentationOral defense to mock oversight panelSimulates governance reviewClarity, persuasion, Q&A

3) Build the Project Around Real-World Stakeholders

Assign roles that reflect a real governance workflow

A strong mock trial works best when each team member has a distinct role. One subgroup can act as model developers, another as external auditors, another as counsel for affected residents or defendants, and another as the oversight board. These roles create productive tension, because the same evidence will be interpreted differently depending on each stakeholder’s mandate. The exercise becomes more authentic when students realize that “accuracy” is not the only concern; fairness, transparency, and accountability matter too.

Include policymakers, advocates, and technical reviewers

A realistic oversight board does not consist of one type of expert. Students should encounter a mix of perspectives: a legal scholar focused on due process, a public defender worried about disparate impact, a city administrator focused on efficiency, and a data scientist concerned about calibration. This mirrors how AI governance happens in practice, where the best solution is often negotiated rather than purely technical. For classes looking to simulate organizational review, team composition and role design and digital identity frameworks offer useful parallels for structured accountability.

Use stakeholder briefs to deepen realism

Each role should receive a short brief explaining goals, constraints, and blind spots. For instance, the public-interest advocate may prioritize false positives that over-punish low-risk people, while the court administrator may worry about operational burden. Students often produce stronger critiques when they are forced to argue from a perspective they did not choose. That tension helps them practice empathy, one of the most underrated skills in interdisciplinary legal-tech work.

4) How to Recreate the Simplified AI Decision Model

Start with a transparent rubric or scoring sheet

The simplest model may assign weights to a handful of features, then convert the total into a recommendation. For example, prior incident frequency could count for 40%, recent compliance behavior for 25%, school or work engagement for 20%, and support network stability for 15%. Students can implement the logic in a spreadsheet, Python notebook, or no-code decision tool, depending on the class level. The key is not sophistication; it is inspectability.

Require students to document assumptions

Every model has assumptions, and the project should make them visible. Students should answer questions like: Why is this feature included? Why is it weighted this way? Which data sources are missing? What would happen if the model were used on a different population? These questions connect directly to professional practices in data quality and governance, similar to the logic behind inspection before high-stakes decisions and due diligence checklists.

Test for sensitivity and edge cases

Students should run at least a few “what if” tests. What happens if two people have the same score but very different backgrounds? What if one factor is missing? What if the model overweights one indicator that correlates with socioeconomic status? A good project makes students uncomfortable in a productive way because it shows how easy it is for a seemingly neutral system to become fragile or unjust. If you want to reinforce this idea with other technical domains, see how on-device processing and performance tradeoffs affect system design choices.

5) Run the AI Critique Like a Real Audit

Look for bias, opacity, and proxy variables

The critique phase is where students move from builders to auditors. They should inspect whether the model uses variables that act as proxies for race, class, disability, neighborhood, or language background. They should also ask whether the model can explain itself in human terms, because a system that cannot be explained is difficult to defend in a justice setting. In this phase, students begin to understand why human oversight is central to trustworthy AI rather than an optional add-on.

Separate technical flaws from normative objections

Not every critique is a bug, and not every concern is about code. Some objections are technical, such as poor calibration or missing data. Others are normative, such as whether it is appropriate to use predictive scoring at all in a particular decision context. Students should explicitly label which category each critique belongs to, because that distinction matters when they propose remedies. A system might be technically accurate yet still unacceptable from a civic or ethical standpoint.

Use evidence, not slogans

Students often want to say the system is “biased” without proving it. Push them to show the evidence: compare outcomes across scenarios, identify feature imbalance, or demonstrate how a threshold produces unequal recommendations. The strongest submissions pair quantitative findings with narrative explanation, making the critique understandable to a non-technical audience. That skill is closely related to how professionals present findings in analytics, compliance, and public policy roles.

Pro Tip: Require each team to submit a one-page “model audit trail” that lists inputs, output rules, test cases, observed weaknesses, and one sentence on why each weakness matters to a real person.

6) Design a Mitigation Plan That Would Actually Work

Choose mitigations that change the system, not just the slide deck

A good mitigation plan should do more than add a “human review” line at the end of a presentation. Students need to identify interventions that change the system’s behavior, governance, or operating rules. Examples include reducing the weight of sensitive proxy variables, introducing fairness checks by subgroup, adding uncertainty flags, or requiring manual review for borderline cases. The strongest plans are realistic, specific, and linked to the critique evidence.

Separate short-term fixes from long-term reforms

Students should distinguish between what can be improved immediately and what requires policy change. A short-term fix might be a threshold adjustment or a new audit workflow. A long-term reform could involve changing what data is collected, restricting use cases, or creating a formal appeal process for affected individuals. This framing helps students think like professionals who must balance practicality with governance responsibility.

Build accountability into the mitigation plan

Every mitigation should name an owner, a review cycle, and a success metric. Without accountability, a mitigation plan becomes aspirational language rather than operational change. Students can borrow a project-management mindset from iterative development and combine it with compliance thinking from organizational AI controls. This helps them produce recommendations that an oversight board could plausibly adopt, test, and revisit.

7) Craft Policy Recommendations for the Mock Oversight Board

Translate classroom findings into governance language

Policy recommendations should not simply restate the critique in formal wording. Students need to move from “this model is unfair” to “this model should not be used for high-impact decisions without human override, appeal rights, and periodic disparate-impact review.” That translation from analysis to governance is one of the most important learning outcomes in the project. It teaches students how to speak across disciplines and how to design recommendations that can survive institutional review.

Recommend operational guardrails, not vague principles

Oversight boards respond best to concrete controls. Students might recommend mandatory documentation for training data, external audits every six months, public disclosure of the model’s use case, and a mechanism for appeal when the model affects someone’s access to liberty, services, or opportunities. These are the kinds of controls that make AI governance actionable. For further perspective on responsible organizational practice, compare this with incident response planning and sensitive data pipeline safeguards.

Anticipate objections from the board

The strongest teams show that they have thought ahead about resistance. A board member may say the system is too expensive to audit, too useful to remove, or too established to replace. Students should prepare responses that explain why safeguards are worth the cost, how harms can be measured, and why legitimacy matters in public decision-making. This is where the mock trial becomes a serious lesson in policy design rather than a simple presentation assignment.

8) Assessment Rubric: What Excellent Student Work Looks Like

Evaluate technical quality and ethical reasoning together

Do not grade the project solely on code output or slide aesthetics. A high-quality submission should demonstrate logical model construction, thoughtful critique, realistic mitigation, and persuasive policy reasoning. If students only optimize for one dimension, they miss the point of the assignment. The rubric should reward clarity, evidence, feasibility, and the ability to connect technical design to social impact.

Assess communication as a core competency

Students in legal tech and civic AI need to explain complex ideas to non-experts. The board presentation should therefore be graded on oral clarity, responsiveness to questions, and ability to use plain language without oversimplifying. A team that can explain tradeoffs to policymakers is demonstrating a workplace-ready skill that often distinguishes strong candidates. This is similar to how creators and analysts turn expertise into understandable public-facing work in career development paths.

Use a multi-dimensional scoring model

One practical rubric is to split grading into model design, audit quality, mitigation specificity, policy recommendations, and presentation defense. You can also include a reflection score that measures what students learned about fairness, authority, and uncertainty. Reflection matters because it helps students internalize the lesson that AI systems are socio-technical, not just mathematical. That insight is what turns an assignment into lasting judgment.

9) Implementation Tips for Teachers and Course Designers

Choose the right tools for the class level

In a lower-division class, a spreadsheet-based model may be enough. In an advanced class, students can use Python, Jupyter, or a lightweight dashboard to visualize decisions and test cases. The key is to match the tool to the learning goal so that the software supports critique rather than distracting from it. If you are designing the assignment sequence in a modular way, the logic resembles other structured classroom builds such as project milestones in rapid prototyping.

Sequence the work in stages

Do not assign the whole mock trial at once without checkpoints. Break the project into a model draft, a test-and-audit checkpoint, a mitigation draft, and a board rehearsal. This pacing reduces overload and gives teachers opportunities to correct misunderstandings early. It also helps teams stay aligned across disciplines, especially when technical and non-technical students are collaborating for the first time.

Make room for reflection and revision

Students should revise their work after receiving feedback from classmates or the mock board. Revision teaches them that governance is iterative and that responsible systems improve through review. That lesson is especially valuable in AI education because students often assume the first model is the final model. For additional ideas on iteration and media-rich communication, you can draw inspiration from video-driven engagement strategies and timing and attention dynamics.

10) What Students Learn Beyond the Classroom

They learn to think like auditors and advocates

This project develops a rare combination of skills: technical reasoning, ethical evaluation, and institutional communication. Students learn how to identify risks without becoming cynical and how to propose solutions without pretending technology is neutral. That balance is valuable in internships, policy fellowships, product teams, and graduate programs. It is also one of the clearest ways to show employers that a candidate can handle ambiguity responsibly.

They produce portfolio-ready evidence of judgment

Employers increasingly want proof that candidates can work on meaningful problems, not just complete exercises. A mock trial package can become a portfolio artifact with a model summary, audit memo, mitigation plan, and board presentation deck. If students want to add a presentation layer, they can borrow portfolio framing lessons from digital communication for creatives and audience trust principles from privacy and trust-building strategies. The result is a project that demonstrates both technical fluency and civic maturity.

They understand why oversight is a design feature

The deepest takeaway is that oversight is not an afterthought when AI affects people’s lives. It is part of the product design, part of the workflow, and part of the social contract. Students who finish this project should understand that a model can be useful and still need constraints, transparency, and appeals. That realization is the bridge between classroom practice and responsible participation in the real world.

11) A Sample Timeline for a Two-Week or Four-Week Version

Two-week sprint

In a compressed format, use days 1-2 for case introduction and role assignment, days 3-5 for model building, days 6-7 for testing and audit, days 8-9 for mitigation and policy drafting, and days 10-10 for board presentations and reflection. This version works well in shorter courses or workshops where time is tight and the goal is exposure rather than deep code development. It still produces meaningful discussion if the case, rubric, and board questions are strong.

Four-week deep dive

A longer version allows for stronger research, more testing, and a more formal board hearing. Week 1 can focus on context and baseline model design, week 2 on critique and evidence collection, week 3 on mitigation and policy memo drafting, and week 4 on rehearsal, presentation, and revision. This version is ideal for capstones, honors seminars, or interdisciplinary cohorts because it gives students time to learn the governance context as well as the technical one.

Deliverables checklist

At minimum, students should submit a model document, an audit memo, a mitigation plan, a policy recommendation brief, and a presentation artifact. If possible, require a reflection paragraph on what they would do differently if the project were deployed in a real justice context. You can also have teams attach a short appendix describing assumptions, limitations, and unanswered questions. That appendix often reveals the depth of a team’s reasoning more clearly than the polished slides do.

FAQ

Is this project appropriate for students without coding experience?

Yes. You can run the entire project in a spreadsheet or through a paper-based decision rubric. The most important learning outcomes are critique, evidence, and policy reasoning, not programming complexity.

Should students use a real court dataset?

Usually no, unless you have a strong ethical review process and appropriate data governance. A fictional or anonymized case is safer, easier to teach, and better for focusing on the model critique itself.

How do I stop the project from becoming too political?

Frame it around evidence, governance, and public accountability. Encourage students to make claims that are supported by test cases, clear assumptions, and practical mitigation steps. That keeps the debate grounded in method rather than ideology.

What is the best way to assess the oversight board presentation?

Grade for clarity, evidence, responsiveness, realism, and the quality of the proposed safeguards. Students should be able to explain both what the model does and why its use should be limited or changed.

Can this project work across different subjects?

Absolutely. It fits computer science, social studies, law, philosophy, statistics, and media literacy. In fact, the project is strongest when students bring different disciplinary lenses to the same case.

Conclusion

A mock trial centered on an AI-powered court decision is one of the strongest classroom practices for teaching students how to build, question, and govern consequential systems. It develops technical literacy, ethical reasoning, policy fluency, and presentation skills in one integrated assignment. Most importantly, it shows students that responsible AI work is not just about better models; it is about better decisions, better oversight, and better outcomes for people. If you are designing a broader learning pathway, connect this project with attribution and measurement discipline, communication strategy, and adaptability under pressure so students see how judgment transfers across contexts.

Advertisement

Related Topics

#project-based#legal#ethics
A

Avery Morgan

Senior SEO Content Strategist

Senior editor and content strategist. Writing about technology, design, and the future of digital media. Follow along for deep dives into the industry's moving parts.

Advertisement
2026-04-16T17:39:52.770Z