Your Genesys Blog Subscription has been confirmed!
Please add email@example.com to your safe sender list to ensure you receive the weekly blog notifications.
Subscribe to our free newsletter and get blog updates in your inbox
Don't Show This Again.
We often talk about the seemingly boundless power of artificial intelligence (AI) when it comes to helping organizations solve difficult problems, realize new efficiencies and achieve very specific business results. For organizations across nearly every industry, it’s quickly gaining traction as the proverbial magic bullet for doing more with less, and even better than before. In customer experience specifically, we also talk about how AI can actually fix what technology “broke” – by facilitating the human-to-human connections that simply got lost in many digitalization efforts.
All of this is great – until it isn’t. Anyone involved in developing and using AI should also be talking about the potential for its extreme side effects – which unfortunately aren’t just fiction. Already there are real examples – whole neighborhoods unfairly targeted by police due to their demographic composition; some candidates having difficulties finding jobs because their profile doesn’t match the company’s idealistic and premeditated ‘right’ fit; and some minority groups facing hardships achieving certain milestones – like being accepted to a university or accessing an affordable loan.
This is all a result of an inherent problem rooted deep within all AI algorithms – biased models. As a society, we have to come to terms with the fact that every model has some degree of bias. How does this happen? Well, it’s easy. If a majority of your customers are males between 50 to 65 years old – the data that feeds your algorithm is already (albeit unintentionally) skewed towards one demographic. Your algorithm likely bases its decisions on this model’s past preferences and behavior, unless you’ve done some work to prevent it. To put it simply, if you don’t use the right data or all of the data, you introduce a bias.
Given that some of the most high-performing deep learning models are complete black boxes today, how does an organization respond? It starts with accountability. From a company’s perspective, there are three pillars to an accountable AI practice:
1: Be responsible for the outcomes: The key here is making sure your AI produces the intended results and taking ownership. You don’t want to be exposed when the result of a score is unethical or unfair. Blaming the algorithm or the data is an indefensible position. AI development is a complex and dynamic process, which is why you need strong guardrails in place to keep up with and stay in control of the models as they are continuously self-learn and evolve.
2: Have the ability to explain the algorithm: When you do not understand how your models make decisions, you allow the algorithm to dictate the reality of many (usually marginalized) people. While it can be exceedingly challenging to determine how a mathematical model reaches a certain conclusion, businesses have to find a way to explain the “why” behind the decisions reached with AI. That is particularly important when those conclusions impact the well-being of an individual and could have legal implications.
3: Show transparency on data sets that are used: First, we should be transparent if a decision is made online by an algorithm or a human being. Both of them will introduce some bias – the machine-generated one will likely be more consistent though. However, transparency is always the way to mitigate bias, especially for models that are difficult to explain. The key is to provide users a view into the data that led to the decision in the first place. To take that concept a step further, consider giving users a voice so they can provide continuous feedback that enables regular adjustments to the model as needed.
In the book, Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy, author Cathy O’Neil was one of the first to shine a light on how automation and AI can destroy people’s lives and perpetuate inequality. She wrote, “Big Data processes codify the past. They do not invent the future. Doing that requires moral imagination, and that’s something only humans can provide. We have to explicitly embed better values into our algorithms, creating Big Data models that follow our ethical lead.”
This is a charge we should all take seriously. It’s not easy, and the paradox is we are beginning to use AI to explain AI because we can’t always do it ourselves. The point is – you can’t just deploy your AI and leave it. You have to be responsible at the time of deployment and accountable throughout its entire lifecycle. There is a constant need to make sure you’re really tuned to what is happening, why it’s happening, the impact its outcomes have on people while also making sure you aren’t inadvertently introducing new biases. It’s a process of continuous improvement – and as they say – we don’t know what we don’t know…yet.
Subscribe to our free newsletter and get blog updates in your inbox.