Video: We want systems to be fair. AI may be able to help us enforce our values.
Bringing a level of scrutiny to AI processes in healthcare, Marzyeh Ghassemi has some solutions for harmful bias in systems that, as a society, we want to root out.
Demonstrating triage models, Ghassemi talked about labeling and how to audit state-of-the-art AI/ML systems that can perform competitively with human doctors. Beginning with some of the more quotidian data collection processes, she tied those into deeper-level mandates that engineering teams and innovators have, to guard against potentially dangerous outcomes in automations.
“We take a lot of data,” she said, warning that things like false positives can compromise the fairness of clinical procedures.
Ghassemi talked about findings on intersectionality, and how bias so often works in both human-centered and AI-centered systems.
Solving these problems, she said, will require diverse data and diverse teams.
“The question is, how does this do (for) all people?” she said, stressing that just using one sub-section of a populace is not enough to really produce transparency on applicable problems and concerns.
Outlining five stages of a pipeline, Ghassemi mentioned problem selection, data collection, outcome definition, algorithm development and postdeployment considerations.
Looking at this entire life cycle, she said, will help stakeholders to move forward with ethical AI in health, and deal with deeply embedded biases that can otherwise have a negative effect on the fairness that we want in healthcare systems.
In a shocking example of evaluating radiology images, Ghassemi showed how AI can still figure out a person’s self-reported race where a human doctor would not be able to make that prediction.
“It’s not the obvious spurious correlations that you might imagine you could remove from medical imaging data,” she said of AI’s strategic ability to classify the images according to race. “It’s not body mass index, breast density, bone density, it’s not disease distribution. In fact, you can filter this image in a variety of ways until it doesn’t really look like a chest X-ray anymore, and machine learning models can still tell the self-reported race of a patient. This is information that’s incredibly deeply (embedded) in data, and you can’t just remove it simply.”
To illustrate the inner biases that can direct systems unfairly, Ghassemi also showed a chart note automation system that tended to send “belligerent and/or violent” white patients to hospitals, but black patients with the same note to prison.
However, she said, in seeking equitable and just outcomes, engineers can look at prescriptive versus descriptive methods, and work toward safe integration.
“(In) machine learning models that we’re training right now, with the outcome labeling practices, we have created much harsher judgments than if we would have collected labels from humans for the normative setting that we were applying these models to,” she said, noting that changing the labels and the method will change the level of “harshness” in the model’s findings.
Going through some other unfair outcomes including the use of GPT systems, Ghassemi suggested that some of the problems arise when GPT “tells (humans) what to do in a biased way” and described efforts to try to correct a lot of this at the algorithmic and methodical levels. She also presented on how differences in labeling instructions cause human labelers to act in surprisingly diverse ways, and suggested that phenomenon bears a lot more study, in general.
In closing, she reviewed some of the principles that can help us find our way through the challenges confronting clinicians and others who don’t want to be affected by undue bias.
“We can’t just focus on one part of the pipeline,” she said. “We need to consider sources of bias in the data that we collect, including the labels … we need to evaluate our models comprehensively as we’re developing algorithms, and we need to recognize that not all gaps can be corrected, but maybe they don’t have to, if you deploy them intelligently such that when they are wrong, they don’t disproportionately bias care stuff. And by doing this, we think that we can create actionable insights in human health.”
Marzyeh Ghassemi is an Assistant Professor at CSAIL, IMES, & EECS MIT. Ghassemi is an accomplished data scientist and researcher, known for her groundbreaking work at the intersection of machine learning and healthcare. With a deep understanding of data-driven solutions, she has made significant contributions to improving patient outcomes and clinical decision-making.
Read the full article here