Wouldn’t it then make sense to factor skin color out of the equation?
It isn't dark skin, per se, that confuses the algorithms in the software. So, there isn't a line of code that says, in essence, "If subject's complexion is swarthy, then crash!"
Rather, as the excerpt pointed out, more White men than, e.g., dark-skinned women were used to "train" the software.
So, to balance out the software, they could simply use more dark-skinned and female test subjects.
Regards,