Machine Learning & Algorithmic Bias — A High-School Lesson Plan

wholerenguru3  (厚仁学术哥)

微信客服
wholerenguru3 (厚仁学术哥)

As a typical high schooler goes about their day, it’s likely that machine learning has played a considerable role: Alexa or Google Home reported the weather as you got up; your news app served you articles that you are most likely to be interested in; at school, Google Docs assisted you by suggesting words as you type; scrolling on Facebook on your way home, it has tagged you in photos your friends uploaded. It is clear that machine learning is having a big impact on our lives.


It is natural to assume that algorithms are neutral and unbiased. Or that a machine learning model, trained on “real world” data will inherently reflect the world at large. However, recent news coverage has demonstrated that machine learning is prone to amplifying existing sexist, racist biases ailing our society. In one often-cited example, Amazon trained an AI recruitment tool only to discover that it is biased against females by discounting any time a resume mentions “women’s” as in “women’s college” or “caption of women’s soccer.” Why? Because the tool is trained on historical data, in other words, resumes from past hiring decisions that skew heavily male.

Amazon says the algorithm was never actually deployed but it rings a bell for how easily dangerous these algorithms can become. Thankfully a lot of research has been devoted to this area. But we at ReadyAI also believe that educating students to proactively think about algorithmic bias is essential, for them to become informed citizens or even future ML engineers.

In this lesson, adapted from WashU, students will explore the ways bias can seep into a machine learning model. Students are asked to imagine if they are data scientists at a bank, whose core business is lending money. The bank’s traditional way of approving loans manually is slow and costly. So, students’ task is to automate this process, to train a model to predict loan approvals. However, the catch is that the dataset is purposely designed to be unbalanced. Women comprise only 15% of the training data.

In the assignment, students will walk through every step in the machine learning workflow in Python from data collection, data cleaning, visualization, model training to testing. In the visualization step, gender mismatch becomes apparent. However, as they continue to train the model (using logistic regression), the validation score comes out to be quite high. As they look a little closer, if validation scores for men and women are calculated separately, a big discrepancy emerges. The assignment demonstrates how model evaluation methods like classification accuracy can be misleadingly high, allowing bias to remain undetected.

Underrepresentation is one of the most common sources of bias in machine learning algorithms. If the data the model is trained on is missing samples from one group, it certainly will not perform equally well for those groups. This is the reason voice assistants have trouble understanding accents. It is also the reason for the above mentioned Amazon recruiting bias against women.

We conclude the lesson by prompting students to consider the impacts of machine learning algorithms if they are not used carefully and fairly. Students also have a discussion on what it means for machine learning models to be fair, and their ideas to help alleviate this issue.

From a hands-on Python assignment, students are learning about algorithmic bias firsthand, in addition to getting familiar with the machine learning workflow and methods such as logistic regression. This lesson is a great introduction for high school students who have some programming experience or are interested in machine learning. The more often and earlier we have a discussion about AI fairness, the longer we hope it will stay on students’ minds as they interact with various AI systems in their daily lives.

Want to try this lesson plan in your classroom? You can access it for free here! If you teach it to your students, let us know what you think by writing to us at info@readyai.org.

美国招生协会
AIRC权威认证

80位
美国双语导师

10年+
名校申请经验

8600+
名校名企录取

wholerenguru3  (厚仁学术哥)

微信客服
wholerenguru3 (厚仁学术哥)

AIRC 权威认证

ICEF 权威认证

NCDA 权威认证

精品服务推荐我要咨询

FITS

资深申请导师 专家专栏
厚仁护学星高端定制留学服务首席专家顾问,在美国、新加坡及中国有多年的教育工作经历。善于帮助学生清晰定位申请局势,深度挖掘学生的优势和潜能,制定个性化最优方案。同时担任美国官方学术委员会AIS (Academic Integrity Seminar) 的常任导师,引导无数学生成功回归学术正轨。经手案例:哥伦比亚大学、埃默里大学、佐治亚理工大学、卡内基梅隆大学、南加州大学、宾夕法尼亚大学、纽约大学、波士顿大学以及UC大学系列名校等。
联系我们  »
                           

美国校园资讯

学业优化
实现梦想

扫码关注 >

厚仁学生中心

实时查看服务进度
阅读文档报告

厚仁APP

扫码关注 >

联系我们

微信24小时在线客服
美国中国8大办公室

扫码关注 >

Scroll to Top