AI should scare us all, but not for the reasons espoused by Elon Musk.
Companies have moved beyond the hype and raced to build and adopt ‘intelligent’ solutions, and the unintended consequences of AI have become increasingly apparent. From Google’s gender-biased autocomplete to deadly autonomous vehicles to a racially biased algorithm that put caucasians ahead of black patients, there is no shortage of disastrous examples over the last year.
These issues are not new. In 2016, the COMPAS program, widely used in the US to guide sentencing, was found to predict black defendants pose a higher risk of recidivism than they actually do. And, as recently as 2015, an image Google search for the term ‘CEO’ would disproportionately return images of white men. Finally, let’s not forget Tay—Microsoft’s attempt to build a chatbot to converse with and learn from the world that turned into a vile, bullying, hatebot.
So why should you care as a marketer? No doubt it’s good to worry about the greater good. But how does gender bias in natural language generation or discrimination in healthcare risk scoring affect me?
With more than 50% of organizations viewing AI as a priority. This isn’t a hypothetical; this is our new reality, and you need to be prepared to navigate it.
We can all get behind “do no harm”, but when consequences are seemingly unpredictable, how does one bring a code of ethics to the development and deployment of AI solutions? It won’t suffice to have data scientists talking amongst themselves about the ethics of AI. End users and adopters need to be part of these conversations as well.
If all AI models must be trained in some capacity on source data, then we should be wary of our new AI systems learning from the mistakes of our past. As a society we have moved mountains to overcome social biases in the past several decades; we should not let our AI systems learn from the worst of human behavioral patterns and set us back. Therefore, it’s critical for stakeholders to understand how bias manifests itself through data that is, in turn, used to train AI models.
Data complexity vs goal complexity, and the quality conundrum
In a previous post, I introduced the notion of AI as a function of data complexity and analytical goal complexity. As the goal of your AI solution becomes more complex (i.e. real-time recommendations) and as the nature of the data becomes more complex (i.e. larger datasets, natural language, images, and audio), it becomes harder to measure the objective performance of the solution. What is a good recommendation? Are there multiple good recommendations?
It is also harder or less obvious to detect bias within more complex datasets that may be learned by the AI. Yet, AI solutions with complex goals on complex data are rapidly emerging.
The explosion of “Black Box” methods, largely driven by advancement in neural networks, can detect highly complex patterns and achieve state of the art predictive performance. However, the downside to these models is the lack of interpretability.
Let’s say you have a regression model that is predicting total volume of sales based on a number of factors such as date, weather, and historic order volume. Not only can you know the importance the regression model assigns to each factor, it is also relatively straightforward to look at the underlying dataset and point to each factor’s contribution to the ultimate prediction. Unlike a regression model, a neural network has many hundreds of thousands or even billions of parameters that are transformations and permutations of underlying data, mathematically tuned to identify subtle patterns beyond human comprehension.
If the goal of AI is to drive business results, then it is tempting to pursue the models with the highest horsepower and performance. In cases where the risk for unintended consequence is low, then perhaps that is an acceptable choice. Even then, when your boss or organization wants to dig into the ‘why’ behind specific predictions, black-box methods fall short. In the extreme, where bias can introduce significant risk like discriminating against customers or employees, the explainability of the solution becomes even more critical.
Understanding how bias enters into the equation is what makes ethical considerations in AI so challenging. There are many stages and building blocks that go into an AI solution.
Although the challenge with the ethics of AI is that there is no clear answer or guaranteed way to avoid unintended consequences, here are some important questions that any savvy team about to embark on an AI journey should think through:
Finally, I stress the importance of dialogue. It’s everyone’s job to safeguard against bias and negative consequences. Too often I see teams shying away from AI conversations because they’re ‘not techies’, or conversely data scientists operating in isolation. The next time you find yourself having a conversation about AI, don’t just ask about what the model can do, also ask about how and why the model was trained.