August 2023 to June 2024
View the Project on GitHub IshanCornick/new_student
What is it?
Which of the following scenarios is an example of computing bias?
A. An email filtering system accurately categorizes emails into spam and non-spam based on a diverse set of features, minimizing false positives and false negatives.
B. A navigation app provides real-time traffic updates and alternate routes to users, considering various factors such as traffic volume, road closures, and weather conditions.
C. An image recognition algorithm identifies objects in photographs with high accuracy, regardless of the gender, ethnicity, or age of the individuals depicted.
D. An automated hiring system consistently favors candidates from specific educational institutions and backgrounds, resulting in the exclusion of qualified applicants from diverse backgrounds.
D is the correct answer:
D. The automated hiring system consistently shows a preference for candidates from particular educational institutions and backgrounds, leading to the exclusion of qualified individuals from diverse backgrounds.
Computing bias can be either intentional or unintentional and as said before, this bias is a result of human biases in development. Here are some examples:
An online example of computing bias is when an HP computer’s facial recognition system couldn’t track the face of someone with darker skin. Why is this and what type of bias is it? Do you think it was intentional or unintentional?
Whether intentional or unintentional, if present, bias toward individuals with darker skin tones could result in increased inequality and unfair treatment among people with different skin colors. This would be a case of unintentional data bias.
Everyday applications like social media and entertainment platforms such as YouTube or TikTok exemplify a common scenario of data collection. These platforms gather insights into user preferences and interests, learning about the content they enjoy. Another instance is when individuals carry a phone while walking, allowing the device to collect data on metrics like step count, location, and other relevant information about the user.
Pre-processing: Writing an algorithm that adjusts to datasets to check for bias before using it as an input. It’s about cleaning up and organizing the raw data and makes it more suitable for building or training models.
Post-processing: This type of strategy is comprised of 3 parts, input correction, classifier correction, output correction
Input Correction:
What it is: Imagine you have a machine learning model that identifies objects in images. Input correction involves making adjustments to the images used to test the model after it has already been trained.
Example: If the model was trained mostly on pictures taken during the day, input correction might involve adjusting the brightness and color balance of images taken at night to make them more comparable.
Classifier Correction:
What it is: This step focuses on fine-tuning the algorithm or model after training to reduce any biases or discrimination it might have inadvertently learned.
Example: Suppose you have a model for hiring decisions, and you notice it tends to favor certain demographics. Classifier correction could involve tweaking the decision-making rules to ensure fair treatment for all groups.
Output Correction:
What it is: After the model makes predictions, output correction involves modifying those predictions to eliminate any biases or unwanted discrimination.
Example: If a language translation model tends to produce more errors when translating sentences from one language to another, output correction might involve adjusting the final translated sentences to be more accurate and fair.
In summary, post-processing is like a three-step check and adjustment process to ensure that a machine learning model behaves fairly and accurately, even after it has been trained. Input correction modifies the testing data, classifier correction fine-tunes the model, and output correction adjusts the final predictions. This helps in addressing issues like biases and discrimination that may arise during the model training process.
Problem: Biased Predictive Policing Algorithm: A city implements a predictive policing algorithm to allocate law enforcement resources more efficiently. However, concerns arise as community members and activists notice that the algorithm appears to disproportionately target certain neighborhoods, leading to over-policing and potential violations of civil rights. Provide a solution to how this situation can be resolved, and how the computing bias can be removed. Explain which method of mitigation you will use and how it works.
Answer: To address and eliminate computing bias in predictive policing algorithms, an effective mitigation strategy would involve classifier correction. This approach entails fine-tuning the algorithm’s parameters to rectify biases and promote equitable treatment across different neighborhoods. Through meticulous adjustments to the classifier, the goal is to minimize any disproportionate targeting, thereby mitigating over-policing and decreasing the likelihood of civil rights violations. Classifier correction serves as a proactive measure that directly tackles the root cause of bias within the algorithm, offering a comprehensive and targeted solution to enhance fairness and equity in the allocation of law enforcement resources.
What is it?
Which of the following scenarios is an example of computing bias?
A. An email filtering system accurately categorizes emails into spam and non-spam based on a diverse set of features, minimizing false positives and false negatives.
B. A navigation app provides real-time traffic updates and alternate routes to users, considering various factors such as traffic volume, road closures, and weather conditions.
C. An image recognition algorithm identifies objects in photographs with high accuracy, regardless of the gender, ethnicity, or age of the individuals depicted.
D. An automated hiring system consistently favors candidates from specific educational institutions and backgrounds, resulting in the exclusion of qualified applicants from diverse backgrounds.
D is the correct answer:
D. The automated hiring system consistently shows a preference for candidates from particular educational institutions and backgrounds, leading to the exclusion of qualified individuals from diverse backgrounds.
Computing bias can be either intentional or unintentional and as said before, this bias is a result of human biases in development. Here are some examples:
An online example of computing bias is when an HP computer’s facial recognition system couldn’t track the face of someone with darker skin. Why is this and what type of bias is it? Do you think it was intentional or unintentional?
Whether intentional or unintentional, if present, bias toward individuals with darker skin tones could result in increased inequality and unfair treatment among people with different skin colors. This would be a case of unintentional data bias.
Everyday applications like social media and entertainment platforms such as YouTube or TikTok exemplify a common scenario of data collection. These platforms gather insights into user preferences and interests, learning about the content they enjoy. Another instance is when individuals carry a phone while walking, allowing the device to collect data on metrics like step count, location, and other relevant information about the user.
Pre-processing: Writing an algorithm that adjusts to datasets to check for bias before using it as an input. It’s about cleaning up and organizing the raw data and makes it more suitable for building or training models.
Post-processing: This type of strategy is comprised of 3 parts, input correction, classifier correction, output correction
Input Correction:
What it is: Imagine you have a machine learning model that identifies objects in images. Input correction involves making adjustments to the images used to test the model after it has already been trained.
Example: If the model was trained mostly on pictures taken during the day, input correction might involve adjusting the brightness and color balance of images taken at night to make them more comparable.
Classifier Correction:
What it is: This step focuses on fine-tuning the algorithm or model after training to reduce any biases or discrimination it might have inadvertently learned.
Example: Suppose you have a model for hiring decisions, and you notice it tends to favor certain demographics. Classifier correction could involve tweaking the decision-making rules to ensure fair treatment for all groups.
Output Correction:
What it is: After the model makes predictions, output correction involves modifying those predictions to eliminate any biases or unwanted discrimination.
Example: If a language translation model tends to produce more errors when translating sentences from one language to another, output correction might involve adjusting the final translated sentences to be more accurate and fair.
In summary, post-processing is like a three-step check and adjustment process to ensure that a machine learning model behaves fairly and accurately, even after it has been trained. Input correction modifies the testing data, classifier correction fine-tunes the model, and output correction adjusts the final predictions. This helps in addressing issues like biases and discrimination that may arise during the model training process.
Problem: Biased Predictive Policing Algorithm: A city implements a predictive policing algorithm to allocate law enforcement resources more efficiently. However, concerns arise as community members and activists notice that the algorithm appears to disproportionately target certain neighborhoods, leading to over-policing and potential violations of civil rights. Provide a solution to how this situation can be resolved, and how the computing bias can be removed. Explain which method of mitigation you will use and how it works.
Answer: To address and eliminate computing bias in predictive policing algorithms, an effective mitigation strategy would involve classifier correction. This approach entails fine-tuning the algorithm’s parameters to rectify biases and promote equitable treatment across different neighborhoods. Through meticulous adjustments to the classifier, the goal is to minimize any disproportionate targeting, thereby mitigating over-policing and decreasing the likelihood of civil rights violations. Classifier correction serves as a proactive measure that directly tackles the root cause of bias within the algorithm, offering a comprehensive and targeted solution to enhance fairness and equity in the allocation of law enforcement resources.