AI applications are ubiquitous – and so is their potential to exhibit unintended bias. Algorithmic and automation biases and algorithm aversion all plague the human-AI partnership, eroding trust between people and machines that learn.
But can bias be eradicated from AI?
AI systems learn to make decisions based on training data, which can include biased human decisions and reflect historical or social inequities, resulting in algorithmic bias. The situation is exacerbated when employees uncritically accept the decisions made by their artificial partners. Equally problematic is when workers categorically mistrust these decisions.
Join our panel of industry and academic leaders, who will share their technological, legal, organizational and social expertise to answer the questions raised by emerging artificial intelligence capabilities.
Chelsea Donahue, Rethinc. Labs Assistant Director Chelsea_Donahue@kenan-flagler.unc.edu