It's good that you bring this up, because it's exactly the sort of thing I wanted to discuss. Why do we feel comfortable letting machine learning screen resumes? Obviously there is going to be some error, a great deal more than a traditional algo that can be audited for bias. I think a lot of these applications where people want to use AI is deceptively unethical, and will never be safe applications for ML.
I agree to some extent. I'm not sure whether AI should be used for resume screening, but I'd lean towards no until biases are proven to not be an issue (if that's possible). There are obviously other areas where this is an important issue that we need to think critically about such as loans and criminal sentencing.