Machine learning models can undoubtedly help humans make more informed decisions, with an increasing number of use cases among different industries. By feeding the algorithms with large amounts of data, they are able to identify patterns and make predictions accordingly.
While the cause of predictions of some models are relatively easy to understand by humans, others are hard or even impossible to interpret. This leads to crucial questions when it comes to defining accountability for machine-based decisions, such as who should be held accountable when a self-driving car causes an accident. The more complex the models are, the harder it gets for humans to interpret them and and we’re talking of the models as being “black boxes”, unable to reveal information about their inner processes and possible biased predictions.
Serg Masis, currently a Data Scientist at Syngenta, is the author of the book “Interpretable Machine Learning with Python” and has been working on ways on how we can build more trust in our machine learning models. In this interview, he’ll tell us more about the challenges of building fair models, how we can build trust in our ML products and the future of model interpretability.
"It's about fairness, accountability, transparency, and the many ethical concerns associated with each of these concepts. It's connected to Ethical AI, but not exclusively. Ultimately, it's about making AI systems more trustworthy. After all, trust is a desirable property for every technology because it drives adoption, solidifies reputations, and increases profits."
"We don't use machine learning models to solve simple deterministic problems. For such problems procedural programming would suffice as a solution. Any problem solved by machine learning is only partially known. For instance:
We probably don't have, nor never will have, enough data to support answers to these questions at 100% certainty, so any machine learning solution will be incomplete. Thus we must accept it will be wrong and extremely wrong a certain percentage of the time. It might be even right for the wrong reasons, right today and wrong tomorrow or even tricked into being wrong. To understand all of this and even begin to correct these problems, we ought to do more than assess models with only predictive performance metrics. That's where machine learning interpretability tools can help."
"It's a challenge on several fronts such as:
"Model complexity is often seen as the culprit for all the ills of machine learning but it isn't always. After all, the problems we attempt to solve with machine learning are inherently complex. And for many problems we will find that the only way to improve the solution is to increase model complexity. That being said, it can go too far. For instance, we must wonder if we need to leverage trillions-parameter language models for Natural Language Processing tasks? Perhaps there's a simpler, less brute force way of achieving the same goals. After all, humans only have 86 billion neurons, and we only use a fraction of them for language at any given time. In any case, we can interpret the largest models that exist today, but one example at a time because understanding them holistically is impossible and not really necessary. The challenge remains in defining a comprehensive framework to delineate and prioritize where and how to concentrate our interpretation efforts — this would be especially useful with complex models.
Another big issue is with the training data. Generally, the idea is to train machine learning models that reflect the reality on the ground. But this reality is often biased and so the data is biased too. A data-centric approach would call to scrutinize the data generation process and mitigate these biases accordingly. After all, the models don't have to reflect the truth we HAVE but the truth we WANT. We can fix so much of this with the data alone.
There's so much that can be improved with model interpretability. However, I think it has a promising future, mostly because there will be more opportunities for it to get much more attention in a few years. Right now the nuts and bolts of machine learning are data cleaning, data engineering, training pipelines, and the drudgery of writing all this code to orchestrate training and inference. In coming years, new and better no-code and low-code ML solutions will displace the artisanal ML approach. I believe the best ones will make interpretability prominent because once creating a sophisticated ML pipeline is less than one day's work in a drag and drop interface; we can devote the rest of the time to model interpretability.
For this reason, eventually, every data scientist and machine learning engineer will have to master this topic. My book Interpretable Machine Learning with Python covers introductory and advanced topics."
Are you interested in joining the AI conversation with industry leaders and senior practitioners? Check out the Data Science Salon events 2023, early bird rates are now available.