Black men were likely underdiagnosed with lung problems because of bias in software, study suggests

Of note (I have done the pulmonary function test as part of my cancer treatments but was completely unaware of the algorithms involved but I could sense the difference between two tests about a year apart):

Racial bias built into a common medical test for lung function is likely leading to fewer Black patients getting care for breathing problems, a study published Thursday suggests. 

As many as 40% more Black male patients in the study might have been diagnosed with breathing problems if current diagnosis-assisting computer software was changed, the study said. 

Doctors have long discussed the potential problems caused by race-based assumptions that are built into diagnostic software. This study, published in JAMA Network Open, offers one of the first real-world examples of how the the issue may affect diagnosis and care for lung patients, said Dr. Darshali Vyas, a pulmonary care doctor at Massachusetts General Hospital.

The results are “exciting” to see published but it’s also “what we’d expect” from setting aside race-based calculations, said Vyas, who was an author of an influential 2020 New England Journal of Medicine article that catalogued examples of how race-based assumptions are used in making doctors’ decisions about patient care.

For centuries, some doctors and others have held beliefs that there are natural racial differences in health, including one that Black people’s lungs were innately worse than those of white people. That assumption ended up in modern guidelines and algorithms for assessing risk and deciding on further care. Test results were adjusted to account for — or “correct” for — a patient’s race or ethnicity. 

One example beyond lung function is a heart failure risk-scoring system that categorizes Black patients as being at lower risk and less likely to need referral for special cardiac care. Another is an equation used in determining kidney function that creates estimates of higher kidney function in Black patients.

The new study focused on a test to determine how much and how quickly a person can inhale and exhale. It’s often done using a spirometer — a device with a mouthpiece connected to a small machine. 

After the test, doctors get a report that has been run through computer software and scores the patient’s ability breathe. It helps indicate whether a patient has restrictions and needs further testing or care for things like asthma, chronic obstructive pulmonary disorder or lung scarring due to air pollutant exposure. 

Algorithms that adjust for race raise the threshold for diagnosing a problem in Black patients and may make them less likely to get started on certain medications or to be referred for medical procedures or even lung transplants, Vyas said.

While physicians also look at symptoms, lab work, X-rays and family histories of breathing problems, the pulmonary function testing can be an important part of diagnoses, “especially when patients are borderline,” said Dr. Albert Rizzo, the chief medical officer at the American Lung Association. 

The new study looked at more than 2,700 Black men and 5,700 white men tested by University of Pennsylvania Health System doctors between 2010 and 2020. The researchers looked at spirometry and lung volume measurements and assessed how many were deemed to have breathing impairments under the race-based algorithm as compared to under a new algorithm.

Researchers concluded there would be nearly 400 additional cases of lung obstruction or impairment in Black men with the new algorithm.

Earlier this year, the American Thoracic Society, which represents lung-care doctors, issued a statement recommending replacement of race-focused adjustments. But the organization also put a call out for more research, including into the best way to modify software and whether making a change might inadvertently lead to overdiagnosis of lung problems in some patients.

Vyas noted some other algorithms have already been changed to drop race-based assumptions, including one for pregnant women that predicts risks of vaginal delivery if the mom previously had a cesarean section.

Changing the lung-testing algorithm may take longer, Vyas said, especially if different hospitals use different versions of race-adjusting procedures and software. 

Source: Black men were likely underdiagnosed with lung problems because of bias in software, study suggests

Demographic skews in training data create algorithmic errors

Of note:

Algorithmic bias is often described as a thorny technical problem. Machine-learning models can respond to almost any pattern—including ones that reflect discrimination. Their designers can explicitly prevent such tools from consuming certain types of information, such as race or sex. Nonetheless, the use of related variables, like someone’s address, can still cause models to perpetuate disadvantage.

Ironing out all traces of bias is a daunting task. Yet despite the growing attention paid to this problem, some of the lowest-hanging fruit remains unpicked.

Every good model relies on training data that reflect what it seeks to predict. This can sometimes be a full population, such as everyone convicted of a given crime. But modellers often have to settle for non-random samples. For uses like facial recognition, models need enough cases from each demographic group to learn how to identify members accurately. And when making forecasts, like trying to predict successful hires from recorded job interviews, the proportions of each group in training data should resemble those in the population.

Many businesses compile private training data. However, the two largest public image archives, Google Open Images and ImageNet—which together have 725,000 pictures labelled by sex, and 27,000 that also record skin colour—are far from representative. In these collections, drawn from search engines and image-hosting sites, just 30-40% of photos are of women. Only 5% of skin colours are listed as “dark”.

Sex and race also sharply affect how people are depicted. Men are unusually likely to appear as skilled workers, whereas images of women disproportionately contain swimwear or undergarments. Machine-learning models regurgitate such patterns. One study trained an image-generation algorithm on ImageNet, and found that it completed pictures of young women’s faces with low-cut tops or bikinis.

https://infographics.economist.com/2021/20210605_GDC200/index.html

Similarly, images with light skin often displayed professionals, such as cardiologists. Those with dark skin had higher shares of rappers, lower-class jobs like “washerwoman” and even generic “strangers”. Thanks to the Obamas, “president” and “first lady” were also overrepresented.

ImageNet is developing a tool to rebalance the demography of its photos. And private firms may use less biased archives. However, commercial products do show signs of skewed data. One study of three programs that identify sex in photos found far more errors for dark-skinned women than for light-skinned men.

Making image or video data more representative would not fix imbalances that reflect real-world gaps, such as the high number of dark-skinned basketball players. But for people trying to clear passport control, avoid police stops based on security cameras or break into industries run by white men, correcting exaggerated demographic disparities would surely help.■

Source: https://www.economist.com/graphic-detail/2021/06/05/demographic-skews-in-training-data-create-algorithmic-errors?utm_campaign=data-newsletter&utm_medium=newsletter&utm_source=salesforce-marketing-cloud&utm_term=2021-06-08&utm_content=data-nl-article-link-1&etear=data_nl_1