Browsing by Author "Alexander, Leo"
Now showing 1 - 2 of 2
Results Per Page
Sort Options
Item Making Decisions about Adverse Impact: The Influence of Individual and Situational Differences(2022-12-02) Alexander, Leo; Oswald, Frederick L.Researchers studying adverse impact have focused primarily on the statistical properties of various adverse impact tests, almost completely neglecting the human decision-making processes involved in evaluating the fairness of employee selection decisions. The purpose of this study is to (a) use signal detection theory (SDT) to explore the effect of hiring scenario characteristics (i.e., size of the applicant pool, overall selection ratio, and minority proportion of the applicant pool) on laypeople’s sensitivity in detecting adverse impact, (b) explore how several individual differences (i.e., social desirability bias, risk-taking, neuroticism, and ambivalent sexism) may influence their response bias, and (c) replicate my prior research findings surrounding their sensitivity and response bias in making adverse impact judgments (Alexander, 2021). In the current study, 97 working-age adults recruited from an online panel were shown 57 selection scenarios with varying degrees of difference in selection rates between men and women and asked to decide if each scenario was fair or unfair by the Equal Employment Opportunity Commission’s definition of adverse impact. Participants detected adverse impact beyond chance (d′ = 0.45) and exhibited a slightly conservative response bias (c = 0.18). Mixed-effects probit regression analyses were used to estimate SDT metrics reflecting relationships between various hiring scenario characteristics and individual differences in predicting decisions about adverse impact. Cognitive ability, overall selection ratio, and minority proportion of the applicant pool were all positively related to participant sensitivity. Hostile sexism related positively, and benevolent sexism related negatively, to response bias.Item Embargo Rater Sensitivity and Bias in Adverse Impact Decision-Making: A Signal Detection Theory Approach(2021-12-03) Alexander, Leo; Oswald, Frederick L.The concept of adverse impact (AI) was created to help identify employment discrimination and is defined as a meaningful difference in the selection rates (e.g., hiring, admissions, promotion) between two groups of employees or applicants (e.g., men and women). Researchers studying fairness in organizations have treated AI almost entirely as if it is only a quantitative problem to be addressed by quantitative tests (e.g., statistical testing) and have neglected to study the decision-making processes of individuals charged with the task of evaluating the fairness of employee selection outcomes (detecting AI). The purpose of this study is to understand how well people can detect AI and what factors influences people’s judgements of AI. Signal detection theory (SDT) is a framework that is well-suited to modelling individuals’ perceptions of the fairness of employee selection outcomes (AI) because it was developed to model human decision-making in the presence of uncertainty. In addition to global evaluations of accuracy (whether or not one can accurately determine whether a hiring decision is fair or unfair), SDT allows researchers to calculate people’s sensitivity to AI stimuli (d′; the ability to detect differences in the selection rates of two groups), as well as people’s response bias in doing so (c ;the difference in selection rates between two groups at which a rater tends to decide a selection decision is unfair). In this study, 234 participants (58% women) were asked to decide if 36 selection scenarios where men and women were hired at differing rates were fair (or unfair), based on the Equal Employment Opportunity Commission’s definition of AI. Participants also completed various individual difference measures hypothesized to influence sensitivity: i.e., cognitive ability, numeracy, and graph literacy. Similarly, participants completed various individual difference measures that were hypothesized to influence response bias: i.e., Procedural Justice Beliefs for Others, Distributive Justice Beliefs for Others, and the Ambivalent Sexism Inventory. Other person and environment characteristics were also recorded: i.e., gender and stimulus type (whether hiring decisions were presented as icon array graphs or numeric text). Participants’ sensitivity and response bias in AI decision-making were analyzed using both traditional and mixed-effects-regression-based SDT approaches. Cognitive ability, numeracy, and graph literacy all emerged as significant predictors of participant sensitivity in the analyses. Procedural Justice Beliefs for Others and gender emerged as significant predictors of response bias in the analyses. The implications of these results as well as the suitability of traditional and mixed-effects-regression-based SDT models as a framework for analyzing AI decision-making are discussed.