The increasing use of AI, including chatbots like Chat GPT, in various business functions has prompted discussions about the ethical considerations and legal compliance associated with these technologies. In this article, we delve into the recent guidance provided by the U.S. Equal Employment Opportunity Commission (EEOC) regarding AI in HR. The EEOC acknowledges the potential for discrimination and focuses on addressing disparate impact in the workplace.  


We explore the key points from the EEOC guidance and highlight the importance of human oversight in AI implementation to ensure ethical practices and legal compliance. Stay compliant with ever-changing laws and regulations. Connect with one of our friendly HR compliance experts. 


Understanding Disparate Impact and Its Relevance to AI 

Disparate impact refers to a situation where an otherwise neutral policy or practice disproportionately affects individuals from certain protected categories, such as race or sex. The EEOC recognizes that AI tools, including screening software for applicants, can inadvertently lead to negative impacts during recruitment, hiring, promotion, and firing processes. For example, AI tools trained on biased data or influenced by discriminatory factors may exhibit implicit biases, resulting in unfair treatment. 


AI tools trained on biased data or influenced by discriminatory factors have the potential to perpetuate and even amplify existing biases, leading to unfair treatment. The concept of bias in AI systems stems from the fact that these systems learn patterns and make predictions based on the data they are trained on. If the training data is biased or contains discriminatory elements, the AI system may unintentionally incorporate and reinforce those biases in its decision-making processes. 


Bias can manifest in various ways within AI systems 

One common form is algorithmic bias, where the AI system’s predictions or recommendations favor certain groups over others. This can occur when historical data reflects societal biases or systemic inequalities. For example, if an AI tool for resume screening is trained on data that predominantly represents male applicants being selected for certain roles, it may learn to associate certain characteristics more closely with male candidates. Consequently, when evaluating female applicants, the AI system might overlook their qualifications or penalize them for factors such as career gaps due to family responsibilities. 


Another type of bias is representation bias, which occurs when the training data is not diverse or does not adequately represent all relevant groups. If certain demographic groups are underrepresented or excluded from the training data, the AI system may struggle to make accurate predictions or recommendations for those groups, resulting in disparities and unequal treatment. 


Moreover, bias can emerge from the design choices made during the development of AI systems. Human biases and prejudices can inadvertently seep into the design process, consciously or unconsciously. If the individuals involved in developing the AI system hold certain biases, those biases can influence the features selected, the weightings assigned to different variables, or the interpretation of results, thereby introducing bias into the system. 


The implications of biased AI tools are far-reaching. They can perpetuate systemic discrimination and reinforce social inequalities. Unfair treatment resulting from implicit biases in AI systems can affect employment opportunities, promotions, access to resources, and overall well-being. It not only violates principles of fairness and equal opportunity but also has legal ramifications, potentially leading to claims of discrimination under anti-discrimination laws and regulations. 


To mitigate the risk of bias in AI systems, it is essential to address these issues at multiple stages. This includes thoroughly examining the training data for biases, ensuring diverse and representative data sets, regularly auditing and testing the AI models for fairness and bias, and involving multidisciplinary teams with diverse perspectives in the development and deployment processes. Additionally, transparent and explainable AI algorithms can enable better understanding and identification of biases, allowing for appropriate corrective measures. 


Overall, the presence of implicit biases in AI systems highlights the need for ongoing scrutiny, accountability, and human oversight. It is crucial to recognize that AI is a tool that should be guided by ethical considerations, legal compliance, and the principles of fairness and equality. By actively addressing and mitigating bias in AI, we can strive to build more equitable and inclusive systems that promote equal opportunities and avoid perpetuating discriminatory practices. 


Responsibility of Employers 

The EEOC emphasizes that employers bear the responsibility for the actions of their agents, even when using AI tools developed by third-party vendors. If an AI tool exhibits disparate impact or discriminatory effects, the employer will be held accountable. 

Therefore, the EEOC suggests that companies take proactive measures to mitigate potential biases by conducting bias audits or ensuring that vendors have conducted such audits. Employers should inquire about the analysis performed by AI vendors, the factors considered, and the results obtained to assess the presence of any implicit biases in the AI systems. 


Unintended Consequences and Biases in AI 

The vast amount of data on which AI models like Chat GPT rely can inadvertently perpetuate biases present in the training data. For instance, if an AI system trained on predominantly male resumes evaluates female resumes with work gaps due to childbirth, it may unknowingly assign lower grades based on an incomplete understanding of the context. Employers must be cautious of unintentional biases embedded in AI systems and the potential consequences they may have on certain protected classes. 


The Four-Fifths Rule and Indicators of Disparate Impact 

The EEOC highlights the four-fifths rule, which is an indicator of potential disparate impact. According to this rule, if the acceptance rate for a particular group is less than 80% of the rate for the most favored group, it may indicate disparate impact.  


Courts do not universally apply this rule, but it serves as a guideline for evaluating potential biases in AI tools or practices. Employers should be mindful of this rule as an additional measure to identify and address disparate impact issues. 


The Role of Human Oversight 

While AI tools, including Chat GPT, can assist in reducing unconscious biases during the hiring process, they should not replace human judgment and oversight.  


Human involvement is crucial for understanding the context, making informed decisions, and ensuring ethical practices. Employers should exercise due diligence in assessing the potential biases of AI tools and continuously monitor their impact to avoid perpetuating unintended biases. 

Watch our video, Maximizing ChatGPT for HR: Mitigating Risks and Boosting Productivity. 




The EEOC’s guidance on AI in HR emphasizes the importance of addressing disparate impact and the potential for unintended biases. While AI tools can be valuable in reducing human bias, they require human oversight to ensure ethical considerations and legal compliance.  


Employers should conduct or request bias audits from AI vendors, understand the limitations and potential biases of AI systems, and proactively mitigate any adverse impact on protected classes. By striking a balance between AI’s capabilities and human expertise, businesses can navigate the ethical and legal landscape while fostering a fair and inclusive work environment. 


Click here to learn more about how Asure helps you focus on growth while we take care of your HR. 


Unlock your growth potential

Talk with one of experts to explore how Asure can help you reduce administrative burdens and focus on growth.