Additional Reading¶
COMPAS Debate¶
This activity replicates the findings of the piece Machine Bias by Julia Angwin et al at Propublica. They also published a more detailed technical description and their data and code.
Northpointe, the people who developed the tool wrote a response to ProPublica and ProPublica wrote a response to Northpointe.
Researchers then used this dataset to do a lot of different studies a few highlights include:
Fairness definitions are incompatible more on this in the Fair Machine Learning Textbook Classification Chapter
COMPAS is only as accurate and fair as untrained people same paper via URI library access
CORELS is a system that produces a list of rules to predict an outcome, when applied to this data it learns a similarly accurate model with only 3 rules (label=1 is predict recidivism)
if ({Prior-Crimes>3}) then ({label=1})
else if ({Age=18-22}) then ({label=1})
else ({label=0})
There are, however, risks to drawing conclusions from this dataset.
Data Analysis, Python,¶
For more on general data analysis with python, see Data Carpentry’s Social Sciences or Ecology lesson.
The code for this activity is hosted on GitHub if you make an account, you can contribute back your ideas or comment on others’ ideas or suggestions.