Artificial intelligence (AI) has already changed the way consumers interact with technology and the way businesses think about big challenges like digital transformation.
In fact, GlobalData research shows that approximately 50 per cent of IT buyers have already prioritised the adoption of AI technologies, and that number is expected to jump to more than 67 per cent over the next two years.
However, there is a growing realisation that good AI is hard to come by and such decisions AI makes, may only appear to be correct, when in reality they harbour unseen biases, based on incorrect or incomplete data.
Many facets of AI such as deep learning (DL) algorithms are in essence a black box, unable to reveal how and why a given decision has been made.
Over the last two weeks, IBM and Google, both took an important next step by introducing tools, capable of building trust and transparency into AI itself. Both offer highly divergent approaches yet neither solves the problem in its entirety.
Google’s new tool, named What-If Tool, allows users to analyse a machine learning (ML) model directly, without any programming. Intended for use long before an AI solution is put into operation, this tool allows users to readily visualise how the outcome of a given ML model will change, according to any number of “what if” scenarios surrounding the model itself or its underlying dataset.
IBM’s new tools evaluate the effectiveness of a given model, based on how the business expects it to behave, explaining its effectiveness and accuracy in natural and business language.
Despite each solution not being enough to solve the overall problem, what these two highly divergent solutions point to, is the necessity of a multi-pronged approach to building trust in AI; first in the underlying data, next in the model and algorithms, and finally, in the final solution running in the wild.
Brad Shimmin is service director at GlobalData, a data and analytics company