- Filling Gaps in Race, Ethnicity Data Critical for Health Equity
- Using Social Determinants to Promote Health Equity During a Crisis
Currently, there is significant research emphasis on leveraging “big data” from EHRs to develop machine-learning (ML) algorithms that support clinical decision-making. Many of these algorithms are designed to detect or predict medical conditions and outcomes, such as sepsis-related mortality, pancreatic cancer, and postpartum hemorrhage.
Because these tools affect decisions about who receives care, the type of care they receive, and how the care is provided, it is critical that they do not create or perpetuate biases. Racially or ethnically biased algorithms have significant potential to perpetuate worse outcomes for patients who are already at risk for receiving poorer quality care. In the case of a clinical decision support tool for a life-threatening condition, identifying and correcting bias before deployment is crucial.
However, these tools typically rely on data from patient EHRs, which may contain race and ethnicity data for some patients, but not others, or may have incomplete data as a result of a third party reporting the data or of race and ethnicity response options being too limited. In these cases, assigning race and ethnicity data through one of multiple validated methods can ensure that the algorithm is accurately assessed and adjusted for inequitable performance, the researchers state.
Creating equitable pay-for-performance or value-based payment strategies has similar challenges. Pay-for-performance strategies are designed to incentivize and enhance healthcare quality, but can have the opposite effect, exacerbating disparities for certain groups if these strategies are not assessed for bias and corrected, if necessary.
The authors state that healthcare providers who perform poorly on incentive payout measures are more likely to care for disadvantaged patient populations, meaning that these providers receive smaller reimbursements on average. These smaller reimbursements can prevent providers from making investments to improve care quality and delivery, which may prevent them from keeping pace with other providers.
Therefore, approaches to enforce algorithmic fairness can be used to make sure that payments to providers are not correlated with the racial and ethnic composition of patient populations, the researchers argue. However, these approaches require either knowing or imputing race and ethnic data of a given population.
Once this data is made available or imputed, providers can be categorized into multiple groups based on the estimate racial and ethnic composition of their patient population. Then, performance-based payments could be adjusted to ensure that the total payments directed to providers within each group are equal.
Despite the potential utility of imputed race and ethnicity data, the authors indicate that there are limitations to the methodology. For example, even when imputation methods are highly accurate, some uncertainty about the race and ethnicity of a patient will always be present. Because of this, the authors do not recommend using imputation to make individual-level decisions. In individual cases, self-reported data is the gold standard.
Imputation is also limited as a method to ensure algorithmic fairness and equity in the case of multiracial people, who make up a growing proportion of the US population. The researchers note that most imputation approaches are not designed with multiracial people in mind, but future methodologies could address this by allowing people to have noncategorical race and ethnicity designations.
But overall, the inclusion of race and ethnicity, whether reported or imputed, is still better than “race-blind” approaches for ensuring equitable algorithms, the authors conclude.