The Executive Guide To Artificial Intelligence How Identify And Implement Applications For AI%2

User Manual:

Open the PDF directly: View PDF PDF.
Page Count: 187

DownloadThe Executive Guide To Artificial Intelligence - How Identify And Implement Applications For AI%2
Open PDF In BrowserView PDF
THE

EXECUTIVE GUIDE

TO

ARTIFICIAL INTELLIGENCE

How to identify and implement applications
for AI in your organization

ANDREW BURGESS

The Executive Guide to Artificial Intelligence

Andrew Burgess

The Executive Guide
to Artificial
Intelligence
How to identify and implement
applications for AI in your organization

Andrew Burgess
AJBurgess Ltd
London, United Kingdom

ISBN 978-3-319-63819-5    ISBN 978-3-319-63820-1
https://doi.org/10.1007/978-3-319-63820-1

(eBook)

Library of Congress Control Number: 2017955043
© The Editor(s) (if applicable) and The Author(s) 2018
This work is subject to copyright. All rights are solely and exclusively licensed by the Publisher, whether the
whole or part of the material is concerned, specifically the rights of translation, reprinting, reuse of illustrations,
recitation, broadcasting, reproduction on microfilms or in any other physical way, and transmission or information storage and retrieval, electronic adaptation, computer software, or by similar or dissimilar methodology
now known or hereafter developed.
The use of general descriptive names, registered names, trademarks, service marks, etc. in this publication does
not imply, even in the absence of a specific statement, that such names are exempt from the relevant protective
laws and regulations and therefore free for general use.
The publisher, the authors and the editors are safe to assume that the advice and information in this book are
believed to be true and accurate at the date of publication. Neither the publisher nor the authors or the editors
give a warranty, express or implied, with respect to the material contained herein or for any errors or omissions
that may have been made. The publisher remains neutral with regard to jurisdictional claims in published maps
and institutional affiliations.
Cover illustration: Ukususha/iStock/Getty Images Plus
Printed on acid-free paper
This Palgrave Macmillan imprint is published by Springer Nature
The registered company is Springer International Publishing AG
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland

This book is dedicated to my wonderful wife, Meg, and our two amazing
children, James and Charlie.

Foreword

I remember well the AI work I did whilst at college studying computer
science, how different and fascinating it was and still is. We were set a very
open challenge to write an AI programme on any subject. I decided to
write mine so that it could tell you if the building in a photo was a house,
a flat or a bungalow. Somewhat impractical, but a great learning experience for me, particularly in understanding how AI is different from traditional software.
Although my college days were a number of years ago, since that time the
concept of computers learning has always intrigued me and I have since wondered how long it will take for AI to have a truly widespread impact. In recent
years, we’ve seen massive improvements in processing power, big data collection via sensors and the Internet of Things, cloud services, storage, ubiquitous
connectivity and much more. These technological leaps mean that this is the
right time for AI to become part of the ‘here and now’ and I strongly believe
we will see a dramatic increase in the use of AI over the next few years.
The AI in use today is known as narrow AI because it can excel at thousands
of relatively narrow tasks (e.g. doing internet searches, playing Go or looking
for fraudulent transactions). Things will certainly get even more exciting when
‘general AI’ can outperform humans at nearly every task we do, but we simply
don’t know when this might be, or what the world will then look like. Until
then, what excites me most is how we can apply AI now to solve our day-today problems at home and work.
So why is AI important and how can we use it? Firstly, if you are impatient
(like I am), doing small manual, repetitive tasks on computers simply takes
too much time. I want the computer to do a better job of anticipating my
needs and to just get on with it. If I could, I would prefer to talk to Alexa or
vii

viii

Foreword

Google Assistant and just tell the computer what to do. For example, I would
love to be able to ask Alexa to buy the most convenient train ticket and take
the money out of my account. Compare this to buying a train ticket on any
website, where after something like 50 key strokes you might have bought a
ticket. I don’t think future generations, who are becoming increasingly impatient, will put up with doing these simple and time-consuming tasks. I see my
children and future generations having more ‘thinking time’ and focusing on
things that are outside the normal tasks. AI may in fact free up so much of my
children’s time that they can finally clean up their bedrooms.
In the workplace, how many of the emails, phone calls and letters in a call
centre could be handled by AI? At Virgin Trains, we used AI to reduce the time
spent dealing with customer emails by 85% and this enabled our people to focus
on the personable customer service we’re famous for. Further improvements will
no doubt be possible in the future as we get better at developing conversational
interfaces, deep learning and process automation. One can imagine similar
developments revolutionising every part of the business, from how we hire people to how we measure the effectiveness of marketing campaigns.
So, what about the challenges of AI? One that springs to mind at Virgin is
how to get the ‘tone of voice’ right. Our people are bold, funny and empathetic, and our customers expect this from us in every channel. Conversational
interfaces driven by AI should be no different.
Today it may be a nuisance if your laptop crashes, but it becomes all the
more important that an AI system does what you want it to do if it controls
your car, your airplane or your pacemaker. With software systems that can
learn and adapt, we need to understand where the responsibility lies when
they go wrong. This is both a technical and an ethical challenge. Beyond this,
there are questions about data privacy, autonomous weapons, the ‘echo chamber’ problem of personalised news, the impact on society as increasing numbers of jobs can be automated and so on.
Despite these challenges, I am incredibly excited about the future of technology, and AI is right at the heart of the ‘revolution’. I think over the next
five to ten years AI will make us more productive at work, make us more
healthy and happy at home, and generally change the world for the better.
To exploit these opportunities to the full, businesses need people who
understand these emerging technologies and can navigate around the challenges. This book is essential reading if you want to understand this transformational technology and how it will impact your business.
John Sullivan, CIO and Innovation at Virgin Trains

Acknowledgements

I would like to thank the following people for providing valuable input, content and inspiration for this book:
Andrew Anderson, Celaton
Richard Benjamins, Axa
Matt Buskell, Rainbird
Ed Challis, Re:infer
Karl Chapman, Riverview Law
Tara Chittenden, The Law Society
Sarah Clayton, Kisaco Research
Dana Cuffe, Aldermore
Rob Divall, Aldermore
Gerard Frith, Matter
Chris Gayner, Genfour
Katie Gibbs, Aigen
Daniel Hulme, Satalia
Prof. Mary Lacity, University of Missouri-St Louis
Prof. Ilan Oshri, Loughborough University
Stephen Partridge, Palgrave
Mike Peters, Samara
Chris Popple, Lloyds Bank
John Sullivan, Virgin Trains
Cathy Tornbaum, Gartner

ix

x

Acknowledgements

Vasilis Tsolis, Congnitiv+
Will Venters, LSE
Kim Vigilia, Conde Naste
Prof. Leslie Willcocks, LSE
Everyone at Symphony Ventures

Contents

1	Don’t Believe the Hype   1
2	Why Now?  11
3	AI Capabilities Framework  29
4	Associated Technologies   55
5	AI in Action   73
6	Starting an AI Journey   91
7	AI Prototyping   117
8	What Could Possibly Go Wrong?   129
9	Industrialising AI  147
10	Where Next for AI?  165
Index  177

xi

List of Figures

Fig. 2.1
Fig. 2.2
Fig. 2.3
Fig. 3.1
Fig. 3.2
Fig. 3.3
Fig. 4.1
Fig. 4.2
Fig. 4.3
Fig. 6.1
Fig. 6.2
Fig. 6.3
Fig. 6.4
Fig. 9.1

Basic neural network
Training a neural network
A trained neural network
AI objectives
Knowledge map
The AI framework
Human in the loop
Training through a human in the loop
Crowd-sourced data training
Aligning with the business strategy
AI maturity matrix
AI heat map first pass
AI heat map
AI Eco-System

21
21
22
30
43
51
68
69
69
93
100
102
104
148

xiii

1
Don’t Believe the Hype

Introduction
Read any current affairs newspaper, magazine or journal, and you are likely to
find an article on artificial intelligence (AI), usually decrying the way the
‘robots are taking over’ and how this mysterious technology is the biggest risk
to humanity since the nuclear bomb was invented. Meanwhile the companies
actually creating AI applications make grand claims for their technology,
explaining how it will change peoples’ lives whilst obfuscating any real value
in a mist of marketing hyperbole. And then there is the actual technology
itself—a chimera of mathematics, data and computers—that appears to be a
black art to anyone outside of the developer world. No wonder that business
executives are confused about what AI can do for their business. What exactly
is AI? What does it do? How will it benefit my business? Where do I start? All
of these are valid questions that have been, to date, unanswered, and which
this book seeks to directly address.
Artificial Intelligence, in its broadest sense, will have a fundamental impact
on the way that we do business. Of that there is no doubt. It will change the
way that we make decisions, it will enable completely new business models to
be created and it will allow us to do things that we never before thought possible. But it will also replace the work currently being done by many knowledge workers, and will disproportionally reward those who adopt AI early and
effectively. It is both a huge opportunity and an ominous threat wrapped up
in a bewildering bundle of algorithms and jargon.
But this technological revolution is not something that is going to happen in
the future; this is not some theoretical exercise that will concern a few businesses.
© The Author(s) 2018
A. Burgess, The Executive Guide to Artificial Intelligence,
https://doi.org/10.1007/978-3-319-63820-1_1

1

2

1 Don’t Believe the Hype

Artificial Intelligence is being used today in businesses to augment, improve and
change the way that they work. Enlightened executives are already working out
how AI can add value to their businesses, seeking to understand all the different
types of AI and working out how to mitigate the risks that it inevitably brings.
Many of those efforts are hidden or kept secret by their instigators, either because
they don’t want the use of AI in their products or services to be widely known,
or because they don’t want to give away the competitive advantage that it
bestows. A persistent challenge for executives that want to get to grips with AI is
where to find all the relevant information without resorting to fanciful articles,
listening to vendor hyperbole or trying to understand algorithms. AI is firmly in
the arena of ‘conscious unknowns’—we know that we don’t know enough.
People generally experience AI first as consumers. All our smartphones
have access to sophisticated AI, whether that is Siri, Cortana or Google’s
Assistant. Our homes are now AI enabled through Amazon’s Alexa and Google
Home. All of these supposedly make our lives easier to organise, and generally
they do a pretty good job of it. But their use of AI is actually pretty limited.
Most of them rely on the ability to turn your speech into words, and then
those words into meaning. Once the intent has been established, the rest of
the task is pretty standard automation; find out the weather forecast, get train
times, play a song. And, although the speech recognition and natural language understanding (NLU) capabilities are very clever in what they achieve,
AI is so much more than that, especially in the world of business.
Artificial Intelligence can read thousands of legal contracts in minutes and
extract all the useful information out of them; it can identify cancerous
tumours with greater accuracy than human radiologists; it can identify fraudulent credit card behaviour before it happens; it can drive cars without drivers;
it can run data centres more efficiently than humans; it can predict when
customers (and employees) are going to desert you and, most importantly, it
can learn and evolve based on its own experiences.
But, until business executives understand what AI is, in simple-enough
terms, and how it can help their business, it will never reach its full potential.
Those with the foresight to use and exploit AI technologies are the ones that
need to know what it can do, and understand what they need to do to get
things going. That is the mission of this book. I will, over the course of the ten
chapters, set out a framework to help the reader get to grips with the eight
core capabilities of AI, and relate real business examples to each of these. I will
provide approaches, methodologies and tools so that you can start your AI
journey in the most efficient and effective way. I will also draw upon interviews and case studies from business leaders who are already implementing
AI, from established AI vendors, and from academics whose work focuses on
the practical application of AI.

Introducing the AI Framework

3

Introducing the AI Framework
My AI Framework was developed over the past few years through a need to be
able to make sense of the plethora of information, misinformation and
marketing-­speak that is written and talked about in AI. I am not a computer
coder or an AI developer, so I needed to put the world of AI into a language
that business people like myself could understand. I was continually frustrated by the laziness in the use of quite specific terminology in articles that
were actually meant to help explain AI, and which only made people more
confused than they were before. Terms like Artificial Intelligence, Cognitive
Automation and Machine Learning were being used interchangeably, despite
them being quite different things.
Through my work as a management consultant creating automation strategies for businesses, through reading many papers on the subject and speaking
to other practitioners and experts, I managed to boil all the available information down into eight core capabilities for AI: Image Recognition, Speech
Recognition, Search, Clustering, NLU, Optimisation, Prediction and
Understanding. In theory, any AI application can be associated with one or
more of these capabilities.
The first four of these are all to do with capturing information—getting
structured data out of unstructured, or big, data. These Capture categories are
the most mature today. There are many examples of each of these in use today:
we encounter Speech Recognition when we call up automated response lines;
we have Image Recognition automatically categorising our photographs; we
have a Search capability read and categorise the emails we send complaining
about our train being late and we are categorised into like-minded groups
every time we buy something from an online retailer. AI efficiently captures
all this unstructured and big data that we give it and turns it into something
useful (or intrusive, depending on your point of view, but that’s a topic to be
discussed in more detail later in the book).
The second group of NLU, Optimisation and Prediction are all trying to
work out, usually using that useful information that has just been captured,
what is happening. They are slightly less mature but all still have applications
in our daily lives. NLU turns that speech recognition data into something
useful—that is, what do all those individual words actually mean when they
are put together in a sentence? The Optimisation capability (which includes
problem solving and planning as core elements) covers a wide range of uses,
including working out what the best route is between your home and work.
And then the Prediction capability tries to work out what will happen next—
if we bought that book on early Japanese cinema then we are likely to want to
buy this other book on Akira Kurosawa.

4

1 Don’t Believe the Hype

Once we get to Understanding, it’s a different picture all together.
Understanding why something is happening really requires cognition; it
requires many inputs, the ability to draw on many experiences, and to conceptualise these into models that can be applied to different scenarios and
uses, which is something that the human brain is extremely good at, but AI,
to date, simply can’t do. All of the previous examples of AI capabilities have
been very specific (these are usually termed Narrow AI) but Understanding
requires general AI, and this simply doesn’t exist yet outside of our brains.
Artificial General Intelligence, as it is known, is the holy grail of AI researchers
but it is still very theoretical at this stage. I will discuss the future of AI in the
concluding chapter, but this book, as a practical guide to AI in business today,
will inherently focus on those Narrow AI capabilities that can be implemented
now.
You will already be starting to realise from some of the examples I have
given already that when AI is used in business it is usually implemented as a
combination of these individual capabilities strung together. Once the individual capabilities are understood, they can be combined to create meaningful
solutions to business problems and challenges. For example, I could ring up a
bank to ask for a loan: I could end up speaking to a machine rather than a
human, in which case AI will first be turning my voice into individual words
(Speech Recognition), working out what it is I want (NLU), deciding whether
I can get the loan (Optimisation) and then asking me whether I wanted to
know more about car insurance because people like me tend to need loans to
buy cars (Clustering and Prediction). That’s a fairly involved process that
draws on key AI capabilities, and one that doesn’t have to involve a human
being at all. The customer gets great service (the service is available day and
night, the phone is answered straight away and they get an immediate response
to their query), the process is efficient and effective for the business (operating
costs are low, the decision making is consistent) and revenue is potentially
increased (cross-selling additional products). So, the combining of the individual capabilities will be key to extracting the maximum value from AI.
The AI Framework therefore gives us a foundation to help understand what
AI can do (and to cut through that marketing hype), but also to help us apply
it to real business challenges. With this knowledge, we will be able to answer
questions such as; How will AI help me enhance customer service? How will
it make my business processes more efficient? And, how will it help me make
better decisions? All of these are valid questions that AI can help answer, and
ones that I will explore in detail in the course of this book.

The Impact of AI on Jobs

5

Defining AI
It’s interesting that in most of the examples I have given so far people often
don’t even realise they are actually dealing with AI. Some of the uses today,
such as planning a route in our satnav or getting a phrase translated in our
browser, are so ubiquitous that we forget that there is actually some really
clever stuff happening in the background. This has given rise to some tongue-­
in-­cheek definitions of what AI is: some say it is anything that will happen in
20 years’ time, others that it is only AI when it looks like it does in the movies.
But, for a book on AI, we do need a concise definition to work from.
The most useful definition of AI I have found is, unsurprisingly, from the
Oxford English Dictionary, which states that AI is “the theory and development of computer systems able to perform tasks normally requiring human
intelligence”. This definition is a little bit circular since it includes the word
‘intelligence’, and that just raises the question of what is intelligence, but we
won’t be going into that philosophical debate here.
Another definition of AI which can be quite useful is from Andrew Ng, who
was most recently the head of AI at the Chinese social media firm, Baidu, and
a bit of a rock star in the world of AI. He reckons that any cognitive process
that takes a human under one second to process is a potential candidate for
AI. Now, as the technologies get better and better this number may increase
over time, but for now it gives us a useful benchmark for the capabilities of AI.
Another way to look at AI goes back to the very beginnings of the technology and a very fundamental question: should these very clever technologies
seek to replace the work that human beings are doing or should they augment
it? There is a famous story of the two ‘founders’ of AI, both of whom were at
MIT: Marvin Minsky and Douglas Engelbart. Minsky declared “We’re going
to make machines intelligent. We are going to make them conscious!” To
which Engelbart reportedly replied: “You’re going to do all that for the
machines? What are you going to do for the people?” This debate is still raging
on today, and is responsible for some of those ‘robots will take over the world’
headlines that I discussed at the top of this chapter.

The Impact of AI on Jobs
It is clear that AI, as part of the wider automation movement, will have a
severe impact on jobs. There are AI applications, such as chatbots, which can
be seen as direct replacements for call centre workers. The ability to read thousands of documents in seconds and extract all the meaningful information

6

1 Don’t Believe the Hype

from them will hollow out a large part of the work done by accountants and
junior lawyers. But equally, AI can augment the work done by these groups as
well. In the call centre, cognitive reasoning systems can provide instant and
intuitive access to all of the knowledge that they require to do their jobs, even
if it is their first day on the job—the human agent can focus on dealing with
the customer on an emotional level whilst the required knowledge is provided
by the AI. The accountants and junior lawyers will now have the time to properly analyse the information that the AI has delivered to them rather than
spend hours and hours collating data and researching cases.
Whether the net impact on work will be positive or negative, that is, will
automation create more jobs than it destroys, is a matter of some debate.
When we look back at the ‘computer revolution’ of the late twentieth century
that was meant to herald massive increases in productivity and associated job
losses, we now know that the productivity benefits weren’t as great as people
predicted (PCs were harder to use than first imagined) and the computers
actually generated whole new industries themselves, from computer games to
movie streaming. And, just like the robots of today, computers still need to be
designed, manufactured, marketed, sold, maintained, regulated, fixed, fuelled,
upgraded and disposed of.
The big question, of course, is whether the gains from associated activities
plus any new activities created from automation will outweigh the loss of jobs
that have been replaced. I’m an optimist at heart, and my own view is that we
will be able to adapt to this new work eventually, but not before going through
a painful transition period (which is where a Universal Basic Income may
become a useful solution). The key factor therefore is the pace of change, with
all the indicators at the moment suggesting that the rate will only get faster in
the coming years. It’s clear that automation in general, and AI in particular,
are going to be huge disruptors to all aspects of our lives—most of it will be
good but there will be stuff that really challenges our morals and ethics. As
this book is a practical guide to implementing AI now, I’ll be exploring these
questions in a little more detail toward the end of the book, but the main
focus is very much on the benefits and challenges of implementing AI today.

A Technology Overview
The technology behind AI is fiendishly clever. At its heart, there are algorithms: an algorithm is just a sequence of instructions or a set of rules that are
followed to complete a task, so it could simply be a recipe or a railway timetable. The algorithms that power AI are essentially very complicated statistical

A Technology Overview

7

models—these models use probability to help find the best output from a
range of inputs, sometimes with a specific goal attached (‘if a customer has
watched these films, what other films would they also probably like to watch?’).
This book is certainly not about explaining the underlying AI technology; in
fact, it is deliberately void of technology jargon, but it is worth explaining
some of the principles that underpin the technology.
One of the ways that AI technologies are categorised is between ‘supervised’
and ‘unsupervised’ learning. Supervised learning is the most common
approach out of the two and refers to situations where the AI system is trained
using large amounts of data. For example, if you wanted to have an AI that
could identify pictures of dogs then you would show it thousands of pictures,
some of which had dogs and some of which didn’t. Crucially, all the pictures
would have been labelled as ‘a dog picture’ or ‘not a dog picture’. Using
machine learning (an AI technique which I’ll come on to later) and all the
training data the system learns the inherent characteristics of what a dog looks
like. It can then be tested on another set of similar data which has also been
tagged but this time the tags haven’t been revealed to the system. If it has been
trained well enough, the system will be able to identify the dogs in the pictures, and also correctly identify pictures where there is no dog. It can then be
let loose on real examples. And, if the people using your new ‘Is There a Dog
in My Picture?’ app are able to feed back when it gets it right or not, then the
system will continue to learn as it is being used. Supervised learning is generally used where the input data is unstructured or semi-structured, such as
images, sounds and the written word (Image Recognition, Speech Recognition
and Search capabilities in my AI Framework).
With Unsupervised Learning, the system starts with just a very large data
set that will mean nothing to it. What the AI is able to do though is to spot
clusters of similar points within the data. The AI is blissfully naive about what
the data means; all it does is look for patterns in vast quantities of numbers.
The great thing about this approach is that the user can also be naive—they
don’t need to know what they are looking for or what the connections are—all
that work is done by the AI. Once the clusters have been identified then predictions can be made for new inputs.
So, as an example, we may want to be able to work out the value of a house
in a particular neighbourhood. The price of a house is dependent on many
variables such as location, number of rooms, number of bathrooms, age, size
of garden and so on, all of which make it difficult to predict its value. But,
surely there must be some complicated connection between all of these variables, if only we could work it out? And that’s exactly what the AI does. If it
is fed enough base data, with each of those variables as well as the actual price,

8

1 Don’t Believe the Hype

then it uses statistical analysis to find all the connections—some variables may
be very strong influencers on price whilst others may be completely irrelevant.
You can then input the same variables for a house where the price is unknown
and it will be able to predict that value. The data that is input is structured
data this time, but the model that is created is really a black box. This apparent lack of transparency is one of AI’s Achilles’s heels, but one that can be
managed, and which I’ll discuss later in the book.
As well as the above two types of training, there are various other terms
associated with AI, and which I’ll cover briefly here, although for business
executives they only need to be understood at a superficial level. ‘Neural
Networks’ is the term used to describe the general approach where AI is mimicking the way that the brain processes information—many ‘neurons’ (100
billion in the case of the brain) are connected to each other in various degrees
of strength, and the strength of the connection can vary as the brain/machine
learns.
To give an over-simplified example, in the dog picture app above, the ‘black
nose’ neuron will have a strong influence on the ‘dog’ neuron, whereas a ‘horn’
neuron will not. All of these artificial neurons are connected together in layers, where each layer extracts an increasing level of complexity. This gives rise
to the term Deep Neural Networks. Machine Learning, where the machine
creates the model itself rather than a human creating the code (as in the examples I have given above), uses DNNs. So, think of these terms as concentric
circles: AI is the over-arching technology, of which Machine Learning is a core
approach that is enabled by DNNs.
There are obviously many more terms that are in common use in the AI
world, including decision tree learning, inductive logic programming, reinforcement learning and Bayesian networks, but I will cover these only when
absolutely necessary. The focus of this book, as you will now hopefully
appreciate, is on the business application of AI rather than its underlying
technologies.

About This Book
My working experience has been as a management consultant, helping organisations cope with the challenges of the time, from productivity improvement,
through change management and transformation to outsourcing and robotic
process automation, and now AI. I first came across AI properly in my work
in 2001. I was working as Chief Technology Officer in the Corporate Venturing
unit of a global insurance provider—my role was to identify new technologies

About This Book

9

that we could invest in and bring into the firm as a foundation client (it was
what we used to call the ‘incubator’ model). One of the technologies we
invested in was based around the idea of ‘smart agents’ that could be used as
an optimisation engine—each agent would have a specific goal and would
‘negotiate’ with the other agents to determine the ideal collective outcome. So,
for example, the system could determine the most effective way for trucks to
pass through a port, or the best way to generate the most revenue from the size
and arrangement of advertisements in newspapers. Although we didn’t call it
AI at the time, this is effectively what it was—using computer algorithms to
find optimal solutions to real problems.
Fast forward to 2017, and my work is focused almost exclusively on AI. I
work with enterprises to help create their AI strategy—identifying opportunities for AI, finding the right solution or vendor and creating the roadmap for
implementation. I don’t do this as a technologist, but as someone that understands the capabilities of AI and how those capabilities might address business
challenges and opportunities. There are plenty of people much cleverer than
me that can create the algorithms and design the actual solutions, but those
same people rarely understand the commerciality of business. I see myself as a
‘translator’ between the technologists and the business. And with AI, the challenge of translating the technology is orders of magnitude greater than with
traditional IT. Which is why I wanted to write this book—to bring that
understanding to where it can be used the best: on the front line of business.
So, this is not a book about the theoretical impact of AI and robots in 10
or 20 years’ time and it is certainly not a book about how to develop AI algorithms. This is a book for practitioners of AI, people who want to use AI to
make their businesses more competitive, more innovative and more future-­
proof. That will happen only if the business leaders and executives understand
what the capabilities of the technology are, and how it can be applied in a
practical way. That is the mission of this book: be as informed as possible
about AI, but without getting dragged down by the technology, so that you
can make the best decisions for your business. And it is also a heartfelt appeal:
whatever you read or hear about AI, don’t believe the hype.

2
Why Now?

A Brief History of AI
For anyone approaching AI now, the technology would seem like a relatively
new development, coming off the back of the internet and ‘big data’. But the
history of AI goes back over 50 years, and includes periods of stagnation
(often referred to as ‘AI Winters’) as well as acceleration. It is worth providing
a short history of AI so that the developments of today, which really is AI’s big
moment in the sun, can be put into context.
I’ve already mentioned in the previous chapter the two people that are considered some of the key founders of AI, Marvin Minsky and Douglas
Engelbert, who both originally worked at the Massachusetts Institute of
Technology in Boston, USA. But the person who coined the phrase ‘artificial
intelligence’ was John McCarthy, a professor at Stanford University in
California. McCarthy created the Stanford Artificial Intelligence Lab (SAIL),
which became a key focus area for AI on the west coast of America. The technology driving AI at that time would seem rudimentary when compared to
the neural networks of today, and certainly wouldn’t pass as AI to anyone with
a basic understanding of the technology, but it did satisfy our earlier definition of “the theory and development of computer systems able to perform
tasks normally requiring human intelligence”, at least to a very basic level.
Much of the work in the early development of AI was based around
‘expert systems’. Not wanting to disparage these approaches at all (they are
still being used today, many under the guise of AI), they were really no
more than ‘if this then that’ workflows. The programmer would lay out the
knowledge of the area that was being modelled in a series of branches and
© The Author(s) 2018
A. Burgess, The Executive Guide to Artificial Intelligence,
https://doi.org/10.1007/978-3-319-63820-1_2

11

12

2

Why Now?

loops, with each branch depending on an input from the user or a rule. So,
in a system designed to model a bank account recommendation process, the
user would be asked a series of questions (employment status, earnings, savings, etc.) with each answer sending the process down different branches
until it came to a conclusion. And, because that was essentially performing
a task ‘normally requiring human intelligence’ it was, at the time, considered AI. Today it wouldn’t really pass muster because it doesn’t have any
self-learning capability, which is a key facet of how AI is perceived to be
defined.
Interestingly, even now, this same approach is being used in many of the
chatbots that have proliferated across the internet. Most of these claim to use
AI, and some do, but most are passive decision trees. There are a number of
online chatbot platforms (most are free to use) that allow you to create your
own bots using this approach and they do a reasonable job for simple processes (I actually built my own a while back—it took about half a day and was
very basic—but it proved to me that it could be done by a non-technical
person and that there was very little, if any, AI actually involved).
There were a couple of ‘AI Winters’, when advances in AI stagnated for a
good number of years. Both of these were the result of over-inflated expectations and the withdrawal of funding. The first occurred between 1974 and
1980, triggered by three events: a 1973 report by Sir James Lighthill for the UK
Government which criticised the ‘grandiose objectives’ of the AI community
and its failure to get anywhere near reaching those objectives; and in the United
States where the Mansfield Amendment required the Advanced Research
Projects Agency (ARPA, now known as DARPA) to only fund projects with
clear missions and objectives, particularly relating to defence, which AI at the
time couldn’t satisfy; and the perceived failure of a key project for ARPA that
would have allowed fighter pilots to talk to their planes. These events meant
that much of the funding was withdrawn and AI became unfashionable.
The second AI Winter lasted from 1987 to 1993 and was chiefly due to the
failure of ‘expert systems’ to meet their over-inflated expectations in 1985,
when billions of dollars were being spent by corporations on the technology.
As with my own chatbot experience that I described above, expert systems
ultimately proved difficult to build and run. This made them overly expensive
and they quickly fell out of favour in the early 1990s, precipitated by the
simultaneous collapse of the associated hardware market (called Lisp
machines). In Japan, a 1981 program costing $850m to develop a ‘fifth generation computer’ that would be able to ‘carry on conversations, translate languages, interpret pictures, and reason like human beings’ failed to meet any of
its objectives by 1991 (some remain unmet in 2017). And, although DARPA
had started to fund AI projects in the United States in 1983 as a response to

The Role of Big Data

13

the Japanese efforts, this was withdrawn again in 1987 when new leadership
at the Information Processing Technology Office, which was directing the
efforts and funds of the AI, supercomputing and microelectronics projects,
concluded that AI was ‘not the next wave’. It dismissed expert systems as simply ‘clever programming’ (which, with hindsight, they were pretty close on).
I talk about these AI Winters in a little detail because there is the obvious
question of whether the current boom in AI is just another case of over-­
inflated expectations that will lead to a third spell of the technology being left
out in the cold. As we have seen in the previous chapter, the marketing
machines and industry analysts are in a complete froth about AI and what it
will be capable of. Expectations are extremely high and businesses are in danger of being catastrophically let down if they start to believe everything that is
being said and written. So, we need to understand what is driving this current
wave and why things might be different this time.
From a technology point of view, the only two words you really need to
know for now are ‘machine learning’. This is the twenty-first century’s version
of expert systems—the core approach that is driving all the developments and
applications (and funding). But, before I describe what machine learning is
(in non-technical language, of course), we need to understand what all the
other forces are that are contributing to this perfect storm, and why this time
things could be different for AI. There are, in my mind, four key drivers.

The Role of Big Data
The first driver for the explosion of interest and activity in AI is the sheer volume of data that is now available. The numbers vary, but it is generally agreed
that the amount of data generated across the globe is doubling in size every
two years—that means that by 2020 there will be 44 zettabytes (or 44 trillion
gigabytes) of data created or copied every year. This is important to us because
the majority of AI feeds off data—without data this AI would be worthless,
just like a power station without the fuel to run it.
To train an AI system (like a neural network) with any degree of accuracy
you would generally need millions of examples—the more complex the model,
the more examples are needed. This is why the big internet and social media
companies like Google and Facebook are so active in the AI space—they simply
have lots of data to work with. Every search that you make with Google—there
are around 3.5 billion searches made per day—and every time you post or like
something on Facebook—every day 421 billion statuses are updated, 350 million photos are uploaded and nearly 6 trillion Likes are made—more AI fuel is
generated. Facebook alone generates 4 million gigabytes of data every 24 hours.

14

2 Why Now?

This vast volume of data is then consumed by the AI to create value. To
pick up again on the simple example I used in the previous chapter, to train a
DNN (essentially a machine learning AI) to recognise pictures of dogs you
would need to have many sample images of dogs, all labelled as ‘Dog’, as well
as lots of other images that didn’t contain dogs, all labelled ‘No Dog’. Once
the system had been trained to recognise dogs using this set of data (it could
also go through a validation stage where the algorithm is tuned using a sub-set
of the training data), then it will need to be tested on a ‘clean’, that is, unlabelled, set of images. There are no strict guidelines for how much testing data
is required but, as a rule of thumb, this could represent around 30% of the
total data set.
These huge amounts of data that we create are being exploited every minute
of the day, most of the time without our knowledge (but implicit acceptance).
Take, for example, your Google searches. Occasionally, as you type in your
search term you may spell a word incorrectly. Google usually offers you results
based on a corrected or more common spelling of that word (so for example
if I search for ‘Andrew Durgess’ it shows me results for Andrew Burgess), or
you can choose to actually search for the uncommonly spelled version. What
this means is that Google is constantly collecting data on commonly misspelt
versions of words and, most importantly, which corrections that it suggests are
acceptable or not. All the data is then used to continually tune their AI-powered
spell checker. If, in my example, there was actually someone called Andrew
Durgess who suddenly became famous tomorrow so that many people were
searching for his name, then the correction that Google used on my name
would be quickly phased out as less and less people accepted it and instead
clicked on ‘Search instead for Andrew Durgess’.
But it’s not just social media and search engines that have seen exponential
increases in data. As more and more of our commercial activities are done
online, or processed through enterprise systems, more data on those activities
will be created. In the retail sector our purchases don’t have to be made online
to create data. Where every purchase, whilst not necessarily connecting it to
an identified buyer, is recorded, the retailers can use all that data to predict
trends and patterns that will help them optimise their supply chain. And
when those purchases can be connected to an individual customer, through a
loyalty card or an online account for example, then the data just gets richer
and more valuable. Now the retailer is able to predict what products or services you might also want to buy from them, and proactively market those to
you. If you are shopping online, it is not only the purchase data that is
recorded—every page that you visit, how long you spend on each one and
what products you view are all tracked, adding to the volume, and the value,
of data that can be exploited by AI.

The Role of Big Data

15

After the purchase has been made, businesses will continue to create, harvest
and extract value from your data. Every time you interact with them through
their website or contact centre, or provide feedback through a third-­party recommender site or social media, more data is created that is useful to them.
Even just using their product or service, if it is connected online, will create
data. As an example, telecoms companies will use data from your usage and
interactions to try and predict, using AI, whether you will leave them soon for
a competitor. Their ‘training data’ comes from the customers that have actually
cancelled their contracts—the AI has used it to identify all of the different
characteristics that make up a ‘churn customer’ which it can then apply to the
usage and behaviours of all of the other customers. In a similar way, banks can
identify fraudulent transactions in your credit card account simply because
they have so much data available of genuine and not-so-­genuine transactions
(there are roughly 300 million credit card transactions made every day).
Other sources of ‘big data’ are all the text-based documents that are being
created (newspapers, books, technical papers, blog posts, emails, etc.), genome
data, biomedical data (X-rays, PET, MRI, ultrasound, etc.) and climate data
(temperature, salinity, pressure, wind, oxygen content, etc.).
And where data doesn’t exist, it is being created deliberately. For the most
common, or hottest, areas of AI complete training sets of data have been
developed. For example, in order to be able to recognise handwritten numbers, a database of 60,000 samples of handwritten digits, as well as 10,000 test
samples, has been created by the National Institute of Standards (the database
is called MNIST). There are similar databases for face recognition, aerial
images, news articles, speech recognition, motion-tracking, biological data
and so on. These lists actually become a bellwether for where the most valuable applications for machine learning are right now.
The other interesting aspect about the explosion and exploitation of data is
that it is turning business models on their heads. Although Google and
Facebook didn’t originally set out to be data and AI companies, that is what
they have ended up as. But what that means now is that companies are being
created to harvest data whilst using a (usually free) different service as the
means to do it. An example of this, and one which uses data for a very good
cause, is Sea Hero Quest. At first sight this looks like a mobile phone game,
but what it actually does is use the data from how people play the game to
better understand dementia, and specifically how spatial navigation varies
between ages, genders and countries. At the time of writing 2.7 million people have played the game which means it has become the largest dementia
study in history. Commercial businesses will use the same approach of a ‘window’ product or service that only really exists to gather valuable data that can
be exploited elsewhere.

16

2

Why Now?

The Role of Cheap Storage
All the data that is being created needs to be stored somewhere. Which brings
me on to the second driver in favour of AI: the rapidly diminishing cost of
storage, coupled with the speed that data can be accessed and the size of the
machines that store it all.
In 1980 one gigabyte of storage would cost on average $437,500. Five years
later that had dropped to around a quarter, and by 1990 it was a fortieth of
the 1980 price, at $11,200. But that was nothing compared to the subsequent
reductions. At the turn of the century it was $11.00, in 2005 it was $1.24 and
in 2010 it was 9 cents. In 2016, the cost stood at just under 2 cents per gigabyte ($0.019).
Because of all that data that is generated through Facebook that I described
above, their data warehouse has 300 petabytes (300 million gigabytes) of data
(the amount of data actually stored is compressed down from that which is
originally generated). Accurate numbers are tricky to come by, but Amazon’s
Web Services (its commercial cloud offering) probably has more storage
capacity than Facebook. It is this sort of scale that results in a sub–2 cent
gigabyte price.
It’s not just costs that have shrunk; size has too. There is a photograph from
1956 that I use in some of the talks I give of an IBM hard drive being loaded
onto a plane with a fork-lift truck. The hard drive is the size of a large shed,
and only has a capacity of 5 Mb. Today that would just about be enough to
store one MP3 song. Amazon now has a fleet of trucks that are essentially
huge mobile hard drives, each capable of storing 100 petabytes (the whole of
the internet is around 18.5 petabytes). At the time of writing IBM have just
announced that they have been able to store information on a single atom. If
this approach can be industrialised it would mean that the entire iTunes
library of 35 million songs could be stored on a device the size of a credit card.

The Role of Faster Processors
All this cheap storage for our massive data sets is great news for AI, which, as
I have hopefully made very clear, lives off large amounts of data. But we also
have to be able to process that data too. So, the third driver in AI’s perfect
storm is faster processor speeds.
This is where the often-cited Moore’s Law comes in. The founder of Intel,
Gordon Moore, predicted in 1965 that the number of transistors that could

The Role of Ubiquitous Connectivity

17

fit onto a transistor would double every year. In 1975, he revised this to a
doubling every two years. It was actually David House, an Intel executive,
who came up with the most commonly used version which is that chip performance (as an outcome of more and faster transistors) would double every
18 months. Although there have been deviations from this trend, particularly
in the past few years, it means that the processors being used today are orders
of magnitude faster than during the last AI Winter.
One of the curious things to come out of technology for AI is that traditional computer chips (central processing units, or CPUs) are not ideally
suited to the crunching of large data sets, whereas GPUs (graphical processing
units), which had been developed to run demanding computer visualisations
(such as for computer games), were perfect for the job. Therefore NVidia, a
maker of GPUs, has taken most of the market for computer chips in the
world of AI.
So, faster AI-friendly processors mean that more complex problems, that
use more data, can be solved. This is important because managing and processing all that data does take time—the systems are fine at the second part of
the process, that of scoring and making a decision based on the learnt data,
but the training part can be a slog. Even relatively simple training sessions
need to run overnight, and more complex ones can take days. So any improvement in processor speed will help enormously in the usefulness of AI systems
both in the original development and design of the models but also on a day-­
to-­day basis so that the systems can use the most up-to-date models. Being
able to provide real-time training as well as real-time decision making is one
of the last frontiers of really useful AI.

The Role of Ubiquitous Connectivity
The final driver working in AI’s favour is connectivity. Clearly the internet has
had a huge enabling effect on the exploitation of data, but it is only in the past
few years that the networks (both broadband and 4G) have become fast
enough to allow large amounts of data to be distributed between servers and
devices. For AI this means that the bulk of the intensive real-time processing
of data can be carried out on servers in data centres, with the user devices
merely acting as a front end. Both Apple’s Siri (on the iPhone) and Amazon’s
Alexa (on the Echo) are prime examples of very sophisticated AI applications
that exploit processing power in data centres to carry out the bulk of the hard
work. This means there is less reliance on the device’s processor but it does put
a burden on having the network available and effective enough.

18

2 Why Now?

It’s not only the real-time processing where the internet can provide benefits. The sheer effort in training an AI model, with each training run taking
days or weeks on ‘standard’ hardware, can be significantly sped up using a
cloud-based solution that has specialised hardware available.
Better communication networks can also help AI systems in other ways.
The data sets that I mentioned in the previous section are huge but are made
available, usually publicly, to many users to help train their systems—that
wouldn’t have been nearly as easy previously.
Also, AI systems can be connected to each other using the internet so that
they can share learnings. A program run by a consortium of Stanford
University, University of California at Berkeley, Brown University and Cornell
University, called Robobrain, uses publicly available data (text, sounds,
images, film) to train AI systems which can be accessed by other AI systems.
And, of course, like any good AI system, the ‘recipient’ systems will feed back
everything that they learn back into Robobrain. The challenge for Robobrain,
which mirrors the overall challenge of AI, is that systems tend to be very narrow in their focus whereas Robobrain wants to be all things to all robots (or
‘multi-modal’ to use the lingo).

About Cloud AI
Nowhere has all four of these drivers come together more effectively than in
the concept of Cloud AI. The idea of AI-as-a-service, where the heavy-lifting
of the AI is done on the cloud and on-demand, is now a key catalyst for the
democratisation of AI. Many of the large technology firms, such as Google,
IBM and Amazon, all have cloud solutions for AI that offer easily accessible
APIs (Application Programming Interfaces, which are essentially standardised
access points to programming capabilities) for developers to create AI ‘front-­
ends’ out of them. IBM’s Watson—its heavily branded and lauded AI capability—is ‘just’ a series of APIs, with each one carrying out a specific function,
such as speech recognition or Q&A. Google’s TensorFlow is an open-source
AI platform that offers similar capabilities, as well as additional features such
as pre-trained models.
What this means for business, and specifically any entrepreneur wanting to
start an AI business, is that the value of the AI is not going to be where everyone assumes it be: in the algorithm. If every new customer-service focused AI
business used the open-source speech recognition algorithm from, say,
Amazon, then the competitive advantage would have to lie with the quality of
the training data, the way that the algorithm is trained or how easily the

What Is Machine Learning?

19

resulting system is to use. When you connect to an IBM Watson API, for
example, there is still a lot of work to do training it before you can derive any
real value.
Clearly some AI businesses will create competitive advantage from their
algorithm, but they must compete with the other firms who use an off-the-­
shelf one. It’s fairly straightforward, as I have done myself, to download a free,
open-source Named Entity Recognition algorithm (which is used to extract
names such as people, places, dates, etc., from a body of text) from a university website (in my case it was Stanford’s), feed it some sample text and get it
to return a reasonable answer. But that is not yet a viable AI solution, let alone
a business. To make it commercial I need to train and tune the algorithm
using as much data as I can get my hands on and create a user-friendly interface for it, and that is where the real skills—in data science, in optimisation,
in user experience—come into play. Put all that together and you may have
the seed of an AI business.
For the executive looking to use AI in their business (and the fact that you
are reading this book would suggest that is the case) the provenance of the
value in an AI company is important to understand. Why should you choose
one vendor over another, if they say they can do exactly the same thing?
Getting to the heart of the differences—is it down to the algorithm, the data,
the ease of implementation, the ease of training, the ease of use and so on?—
will allow you to make the right choice for your business. At the moment,
there are far too many smoke-and-mirrors AI businesses that are based purely
on open-source algorithms and built by very clever but inexperienced
20-somethings. They will be riding the hype of AI but will offer little in the
way of long-term value. My point being that you can create a successful AI
business based on great open-source algorithms, but you need all the other
aspects to be great too. As I said right at the start of this book: don’t believe
the hype.
I’ll be covering these selection considerations, as well as the buy-versus-­
build question, in more detail later in the book.

What Is Machine Learning?
These four developments then—big data, cheap storage, faster processors and
ubiquitous networks—are what is driving the acceleration and adoption of AI
in business today. If you were to take any one of those away, AI would struggle
to perform, and we’d probably still be in an AI Winter. Each one has also been
helping drive development in the others—if we didn’t have such cheap storage

20

2

Why Now?

then we couldn’t keep all that data which means there would be no need for
faster processors, for example. But, at the centre of all this activity, feeding off
each of the drivers, is machine learning, one of the most popular AI approaches.
It’s important to have an appreciation of what machine learning is and
what it does, even if you don’t need to understand how it actually works. As
the name suggests, it is the machine that does all of the hard computational
work in learning how to solve a problem—the machine, rather than a human
being, essentially writes the ‘code’. The human developer plants the seed by
defining the algorithm or algorithms that are to be used, and then the machine
uses the data to create a solution specific to it. We have already seen that the
problem could be undefined initially (as in unsupervised learning, where patterns or clusters are found in data) or defined (and therefore trained with large
data sets to answer a particular question, which we know as supervised learning). For machine learning it is usually easier to think about a supervised
learning example, such as my go-to case of being able to tell the difference
between pictures of dogs and pictures of cats.
In the previous chapter, I introduced the idea of DNNs, which is the ‘architecture’ that machine learning uses. A DNN will consist of a number of different layers; the more complex the problem being solved, the more layers are
needed. (More layers also mean that the model is much more complicated,
will require more computing power and will take much longer to resolve
itself.) The Input Layer takes the data in and starts to encode it. The Output
Layer is where the answer is presented—it will have as many nodes as there are
classes (types) of answer. So, in my dog versus cat example, there would be
two output nodes, one for dog and one for cat (although we could define
three if we had pictures that had neither a dog nor a cat in) (Fig. 2.1).
In between the Input Layer and the Output Layer are the Hidden Layers.
It is here that all the hard work happens. Each Hidden Layer will be looking
for different features within the data with increasing complexity, so with
image recognition different layers will look for outlines, shadows, shapes,
colours and so on. In these Hidden Layers, there will be more nodes (­ ‘neurons’)
than the Input or Output Layers, and these nodes will be connected to each
other across each layer. Each of these connections will have a certain ‘weight’
or strength which determines how much of the information in one node is
taken into the next layer: a strong link, which had, through training, led to a
‘right’ answer at the Output Layer, will mean that information is propagated
down through to the next layer. A weak link with a low weighting, which had
led to a ‘wrong’ answer in training, will not pass as much information down
(Fig. 2.2).

What Is Machine Learning?

Input
Layer

Hidden
Layer

Hidden
Layer

Output
Layer

Hidden
Layer 2

Output
Layer

21

Fig. 2.1 Basic neural network

Input
Layer

Hidden
Layer 1

Fig. 2.2 Training a neural network

As the model is being trained with more and more data, the weights are
continually adjusted (this is the machine learning) until it reaches an optimal
solution. The more data we have fed into the model, the more chances the
machine has to refine the weightings (but the harder it has to work), and
therefore the more accurate the solution will be. The ‘matching function’ (the

22

2

Why Now?

DOG

Input
Layer

Hidden
Layer

Hidden
Layer

Output
Layer

Fig. 2.3 A trained neural network

final version of the model) can then be used to solve for a new piece of data;
for example, I give it a picture of a dog that it hasn’t seen before and it should
then be able to correctly identify it as a dog (Fig. 2.3).
From this brief description of machine learning you should be able to see
that it is wholly dependent on the four drivers I have described—we need lots
and lots of data to train the hidden layers to come up with the right weightings, but that means we need to be able to store our data cheaply and process
our models as quickly as possible, as well as needing access to data sets from
as many sources as possible. Without any one of these, machine learning just
wouldn’t be viable, either because it wouldn’t be accurate enough or it wouldn’t
be easy enough to design and implement (and it’s already difficult enough in
the first place).
It’s not just machine learning that benefits from this ‘perfect storm’ of technology. Other types of AI, the main one being ‘symbolic AI’, have also accelerated and found new life through faster, cheaper and connected computers.

The Barriers to AI
So, with all this data, cheap storage, whizzy processors and inter-connectivity,
we are certain to never see another AI Winter, right? Well, there are a few
things that could spoil the AI party.

The Barriers to AI

23

The biggest barrier to AI achieving escape velocity in my opinion is the
over-inflation of expectations. Far too much is being written and said about
what AI might be capable of, especially when Artificial General Intelligence is
considered and not the Narrow AI that we have today. Both the previous AI
Winters have been precipitated by excessive and unrealistic expectations, and
therefore they need to be managed carefully. A key motivation for me to write
this book was to ‘bring AI down to earth’ so that it could be assessed for its
real value today and tomorrow, and not in 10 or 20 years’ time.
Another factor that will affect the uptake of AI, which has been fuelled by
the hype, is a general fear of what changes AI might bring, especially when it
could mean fundamentally changing the way that people work. There have
been many other seismic shifts in ways of working, including the introduction
of personal computers and outsourcing, but AI has the potential to do this on
a much greater scale: speculative reports by academics and institutions repeatedly report the decimation of ‘white collar’ jobs and the ‘hollowing out’ of the
middle classes. All of this may or may to be true, but the fear that it generates
leads to a natural defensiveness from those in businesses that may be affected
by the implementation of AI.
A third aspect, which is connected to the first two, is ignorance. How can
people expect to get real value from implementing AI if they don’t have a good
enough understanding of what it is? The hype doesn’t help, but AI is inherently a complex subject, and it’s not an easy task for a non-technical business
person to understand it. The maxim that a little knowledge is a dangerous
thing is certainly true for AI, but also having all the knowledge is not useful
unless you want to be a data scientist or AI developer. There is a ‘Goldilocks’
level of understanding AI that this book aims to provide—it should be just
enough that you can make the most out of AI but not too much that you
become overwhelmed with technical detail and jargon.
The final thing that could put the brakes on AI is regulation. As I’ve mentioned already, much of the computational work that is done by an AI system
is hidden from the developer or user—there is (generally) no audit trail that
details how it came to a certain decision. If I fed a complex credit decisioning
AI system with lots of loan applications as training data and told it which had
been approved and which hadn’t, it could then make a recommendation of
whether to approve or reject a new loan. But no one would really know why,
and that causes a problem for regulators who need to see that decision process
written down. There are some approaches which can mitigate this but the
opaqueness of AI can be a real challenge to its usefulness.
I’ll cover these aspects in more detail in Chap. 8, but I think it’s important
that you have an early appreciation of them at this point.

24

2 Why Now?

Some AI Case Studies
Although it may be fuelled by perhaps too much hype, there is certainly
momentum in the AI market right now. But who is benefiting from all these
advancements? Is it just research labs and some start-ups that have had success
playing computer games, or is there something more? Is AI providing value to
businesses today?
In my work, I see plenty of examples of AI being used in businesses to provide insights and efficiencies, adding value that would be beyond the means
of humans to create.
One of the sweet spots for AI at the moment is in customer services. At a
train operating company in the United Kingdom they use AI to categorise
emails that come in from their customers. The challenge, of course, is for the
AI to be able to ‘read’ the email, which has been written freeform, and determine what the customer actually wants. For example, if I couldn’t get a seat
on the 08:06 from Euston Station in London to Glasgow Central, for example, I would email the train company, saying something like “I can’t believe
it – I paid all that money for a ticket on the 08:06 from London to Glasgow
and I couldn’t even get a seat. You’ve really excelled yourselves this time!” All
the salient information is extracted from my unstructured text and validated
(Was there a train at 08:06? Have other people complained about the same
service? Is the customer a regular complainer? etc.) so that the AI can immediately route my query to the relevant person in the customer service organisation. Of course, my wording in the second sentence is sarcastic in tone, so
the AI needs to understand the difference between my complaint and an
actual compliment. What it all means is that when the human agent logs in,
they are the right person to deal with my query and they have everything at
hand to respond.
In a similar vein, a UK-based insurance firm that provides a claims processing service to their own insurance customers has automated the input of
unstructured and semi-structured data (incoming claims, correspondence,
complaints, underwriters’ reports, cheques and all other documents relating to
insurance claims) so that it is directed into the right systems and queues. Using
an AI solution, a team of four people process around 3,000 claims documents
per day, 25% of which are paper. The AI automation tool processes the scanned
and electronic documents automatically, identifies claim information and
other metadata, and deposits the results in databases and document stores
ready for processing by the claims handlers and systems. It also adds service
metadata so performance of the process can be measured end to end. Some

Conclusion

25

documents can be processed without any human intervention, and others
need a glance from the human team to validate the AI’s decisions or fill in
missing details.
A good example of AI in the legal sector is ROSS, a system built around
IBM’s Watson. ROSS actually uses a number of AI capabilities including
NLU, Search and Optimisation. So, when a lawyer has a matter that requires
some specific research (e.g. what is the precedent when an employee is fired
for gross misconduct within a few days of the end of their probationary
period?), she can turn to ROSS and type in that query in plain English. ROSS
will then interrogate the whole corpus of employment law and provide
answers ranked by relevancy back to the lawyer within seconds. The alternative would be hours of research by the lawyer, a junior lawyer and/or a paralegal, and likely without being able to examine every relevant document
available. Like all good AI systems, ROSS is self-learning—the lawyer can
evaluate the quality of the responses that ROSS returns and therefore allow it
to provide better answers in the future.
Some companies have used AI to create whole new lines of business. Charles
Schwab, the US bank and brokerage firm, has created an investment vehicle,
called Schwab Intelligent Portfolios, which uses AI to manage its clients’ portfolios (these are sometimes called ‘robo-advisors’). It focuses on low-cost
exchange-traded funds and has no advisory fees, account service fees or commission fees. Since it launched in 2015, a few other firms have appeared with
a similar model, including Betterment, Wealthfront and FutureAdvisor
(although these charge a very small admin fee). The attractiveness of the low
or zero fees and the simplicity of dealing with an AI rather than a person have
meant that these sorts of services have proved popular amongst beginner
investors, thus providing new clients for the banks that they normally wouldn’t
have been able to access.

Conclusion
So, AI is being used successfully by businesses around the world. The perfect
storm of big data, cheap storage, faster processors and ubiquitous connectivity
has allowed researchers to exploit the near-magic that is machine learning.
But AI has managed to escape the lab and is being commercialised by both
start-ups and internet giants. Real opportunities exist for executives to use AI’s
capabilities to deliver new sources of value into their business as well as to
challenge and disrupt their existing business models. In order to do that they
need to understand what those capabilities are and how they can be used.

26

2

Why Now?

The Academic’s View
This is an extract from an interview with Dr Will Venters, Assistant Professor
in Information Systems at the London School of Economics.
AB What’s the basis of all the fuss around AI right now?
WV	There are a few trajectories that are particularly important. The underlying algorithms have become much smarter and are used more effectively,
but that’s not actually that significant. What has really made the difference is the processing capacity of the computer chips as well as the ability
to handle much bigger volumes of data at scale.
Previously we saw IBM’s Big Blue beat Kasparov at chess, then Watson
won at Jeopardy!, but essentially these are just systems that can search
massive databases very quickly. But what AlphaGo is doing (beating the
world’s number one player in the Chinese game of Go) is amazing—it
watched, then played millions of games of Go against itself. The complexity and the sheer quantity of data that it was able to analyse is phenomenal—
I’m sure the future of AI will be around this sort of capability.
It’s interesting if you look at data complexity versus task complexity
for processes that computers want to take on: we are now able to manage
situations that are complex in both of these aspects. Managing elevators
is a complex task, but the data is relatively simple. Self-driving cars are
not doing complex things (just driving and turning and stopping) but the
data they need to do that is vast. But, we are now able to manage situations that are complex in both data and task aspects. Suddenly we have
the mechanisms to do it.
AB You also mentioned processing power as being a key factor?
WV	Yes, GPUs [Graphical Processing Units] have made a significant contribution to this. But also the widespread use of Cloud Computing, which
allows businesses to manage big data without the associated infrastructure investments and risks.
Lots of companies have invested in data infrastructure, either their
own or through the cloud, and AI allows them to utilise this to the full.
AB Do you think the hype helps or hinders AI?
WV	AI is certainly hyped, but it is also productive. It leads to new innovations
and forces businesses to discuss big questions. Hype shouldn’t only be
seen in the negative.
AB What do you see as the most valuable use cases for AI?
WV	AI can offer the opportunity to deal with ‘messy’ data – data that is fuzzier,
such as photographs, natural language speech and unstructured documents.

The Academic’s View

27

Big Data promised to deliver all of this but ultimately it didn’t because it
required armies of statisticians. Robotic Process Automation is great for
dealing with processes that are clean and structured – as soon as it becomes
messier then AI can step in and overcome that.
AB What do you think businesses need to do if they are to exploit AI to the full?
WV	They first need to look at the business case, and not the technology. But
they also need to consider the governance and to understand how they
manage the inherent risks. Humans have an ethical compass which, of
course, AI doesn’t. It will need some level of oversight because of the lack
of transparency with the AI models. They must also make sure the system
is not biased – it’s very easy to bake the bias into the company’s business.
And these risks scale very quickly because the underlying data is so much
bigger.
AB What do you see for the future for AI?
WV	I think we will see even better algorithms to deal with the bigger volumes
of data. But, of course, AI is a victim of its own success – there is an
inherent disillusionment as each exciting step forward then becomes
common place. You just need to look at advances like [Apple’s] Siri and
[Amazon’s] Alexa. The term AI will eventually become redundant and
people will cease to be questioning about it.
But until that day companies will invest in AI wisely as well as inappropriately. There is so much value to be gained that companies should
think deeply about their approach, take expert advice and make sure they
are not left behind.

3
AI Capabilities Framework

Introduction
It’s impossible to get value out of something if it is not understood, unless it’s
by some happy accident. In the world of AI there are no happy accidents—
everything is designed with meticulous detail and with specific goals in mind.
Therefore, the only way to truly make the most of AI in business is by having
a reasonable understanding of it. The challenge, of course, is that is fiendishly
complicated and full of complex mathematics, and is certainly not within the
bounds of what an ‘ordinary’ business person would be expected to grasp.
My approach in this book is to understand AI from the perspective of what
it can do, that is, what are its capabilities in the context of real world problems
and opportunities? In order to do this, I have developed a framework that
‘boils down’ all of the complexity into eight capabilities. In theory, any AI
application should fit into one of these, which allows the AI layman to quickly
assess and apply that application into something that relates to their business.
Conversely, if you have a specific business challenge, the AI framework can be
used to identify the most appropriate AI capability (or capabilities) that could
address that need. Having said that, I’m sure any AI scientist or researcher
would be able to pick some holes in the framework. It’s certainly not bomb-­
proof: it is intended to be a useful tool for business executives to help them get
the most value possible out of AI.
There are many ways to look at AI, some of which I have already discussed in
earlier chapters: Supervised versus Unsupervised Learning; Machine Learning
versus Symbolic Learning; Structured Data versus Unstructured Data;
Augmentation versus Replacement. All of these are valid perspectives, and all can
be understood to a reasonable enough degree within the capability framework.
© The Author(s) 2018
A. Burgess, The Executive Guide to Artificial Intelligence,
https://doi.org/10.1007/978-3-319-63820-1_3

29

30

3 AI Capabilities Framework

So, with each of the capabilities I will explain whether it will require
supervised or unsupervised learning (or both), the type of AI that is usually
used to deliver the capability (e.g. machine learning), whether it processes
structured or unstructured data and, most importantly, how it benefits businesses. In subsequent chapters I will draw out use cases for each, grouped into
the key themes of enhancing customer service, business process optimisation
and enhanced decision making.
Another perspective for AI, which I introduced in Chap. 1, is really the
meta-view, and probably the most important to understand first: these are the
three basic things that AI tries to achieve: capturing information; determining
what is happening and understanding why it is happening. Each of my eight
capabilities fits into one of these ‘AI Objectives’ (Fig. 3.1):
Capturing information is something that our brain does very well but
machines have historically struggled with. For example, the ability to recognise
a face is something that has evolved in humans since our earliest existence—it
is a skill that allows us to avoid danger and create relationships, and therefore
a great deal of brain power and capacity is used to work on this problem. For
machines, the process is very complex, requiring huge volumes of training
data and fast computer processors, but today we have this capability on our
desktop computers and even mobile phones. It may not be very fast or that
accurate yet but AI has been fundamental in achieving this.
Most of the examples of capturing information are where unstructured
data (such as a picture of a face) is turned into structured data (the person’s
name). Capturing information is also relevant to structured data, and AI
comes into its own when that data is big. Again, the human brain is very good
at identifying patterns in data (the impact of a manager on the success of a
football team, for example) but when there are hundreds of variables and millions of data points we lose sight of the forest and can only see the trees. AI is
able to find patterns, or clusters, of data that would be invisible to a human.
These clusters have the ability to provide insights within the data that have
real value to a business. For example, AI can find patterns between purchases
and customer demographics that a human would either take years to unearth,
or never have thought of in the first place. (This associated idea of the AI
being ‘naive’ to the data is something that I will return to later, but is an
important concept to bear in mind.)

CAPTURE INFORMATION

Fig. 3.1 AI objectives

WHAT IS HAPPENING?

WHY IS IT HAPPENING?

Introduction

31

The next AI Purpose is where it tries to determine what is happening, and
is usually a consequence of where information has already been captured by
AI. For example, we could have used Speech Recognition to extract (from a
sound file or a live conversation) the words that someone was speaking, but at
that point we would only know what all the individual words were rather than
what the person was actually trying to say. This is where Natural Language
Understanding (NLU) comes in—it takes the words and tries to determine
the meaning or intent of the complete sentences. So, we have gone from a
digital stream of sound to a set of words (e.g. ‘I’, ‘want’, ‘to’, ‘cancel’, ‘my’,
‘direct’, ‘debit’, ‘for’, ‘mortgage’, ‘protection’) to working out that this person
‘wants to cancel her Direct Debit for her mortgage protection insurance’.
We can then use other capabilities in this category to take that request further. For example, we could use an Optimisation approach to help the customer understand that if they were to cancel that Direct Debit they would
also probably need to cancel the insurance policy that it is associated with.
And then we could use a Prediction capability to work out that this customer
may be about to leave the bank and go to a competitor (because the AI has
determined from lots of other similar interactions that cancelling Direct
Debits is an indicator of potential customer churn).
So, from a simple interaction with a customer we are able to use different
AI ‘capturing information’ and ‘what is happening’ capabilities to build up a
useful picture as well as satisfying the customer’s request. But, in the example
above, although the AI has been able to identify this customer as a ‘churn risk’
it really doesn’t understand what that means. All it has done is correlate one
set of data (customer requests) with another (customers who leave) and then
apply that to a new data point (our customer’s Direct Debit cancellation
request). To the AI system the data might as well be about ice cream flavours
and the weather. It has no idea of the concept of Direct Debits just as much as
it has no idea of the concepts of banks or customers. The AI capabilities that
we have today and will have in the near future (and maybe even ever) do not
have the ability to understand. It is really important to be able to differentiate
between the very narrow things that different AIs can do (usually better than
humans) and the general intelligence that comes with understanding and
relating different concepts—something our brain is brilliant at doing.
Now that we have the three AI purposes defined, and we understand that
only the first two are relevant to us today, we can start to look in more detail
at each of the eight AI capabilities in the Framework.

32

3 AI Capabilities Framework

Image Recognition
One of the most active areas of research in AI at the moment is that of Image
Recognition. This is a prime example of where the four key drivers that I
described in Chap. 2 have aligned to catalyse the technology. Image
Recognition is based on Machine Learning and requires thousands or millions of sample images for training; therefore, it also needs lots of storage for
all the data, and fast computers to process it all. Connectivity is also important to enable access to the data sets, many of which are publicly available (in
Chap. 2 I mentioned image data sets for handwritten numbers and faces, but
there are many more including aerial images, cityscapes and animals).
Images, of course, fall under the category of unstructured data. So, what
sort of applications would you use Image Recognition for? There are three
main types:
Probably the most popular application is identifying what is in the image,
sometimes referred to as tagging. I’ve used this example a few times already—
find out if the picture contains a dog or a cat. It is often used to moderate
photographs by trying to identify pornographic or abusive images, but can
also be used to group photos with similar tags (‘these are all pictures taken at
the beach’). This photo tagging is a prime example of supervised learning—
the AI is trained on thousands, or millions, of tagged photographs—which is
why companies that have access to large volumes of images, such as Google
and Facebook, have some of the most advanced systems in this area.
Another use of Image Recognition is to find images that are similar to other
images. Google’s Reverse Image Search is a popular use of this approach; you
simply upload a picture and it will search for pictures that look similar to your
original (this technique is often used to identify the use of news photos that
have deliberately been used out of context). Unlike photo tagging, this is
mainly an example of unsupervised learning; the AI doesn’t need to know
what is in the picture, just that it looks like another picture. (A simple way to
think about this is to understand that the AI converts the image file into a
long series of numbers—it then looks for other images that have a similar
series of numbers.)
The final application for Image Recognition is to find differences between
images. The most common, and beneficial, use of this is in medical imaging.
AI systems are used to look at scans of parts of the body and identify any
anomalies, such as cancerous cells. IBM’s Watson has been a pioneer in this
space and is used to support radiologists in their work. The approach uses
supervised learning to tag, for example, X-rays as healthy or unhealthy. Based

Speech Recognition

33

on the algorithmic model that is built up, new images can be evaluated by the
AI to determine whether there is a risk to the patient. Interestingly, it has been
reported that Watson is more accurate at spotting melanomas than when it is
done manually (Watson has 95% accuracy compared to an average of between
75% and 84% for humans).
Image Recognition, probably out of all the capabilities I describe in this
chapter, is the most data hungry. Images are inherently unstructured and very
variable and therefore require very large amounts of data to train them effectively. Pinterest, the website that allows users to create ‘boards’ of their favourite images, has used all of the data from their hundreds of millions of users to
help develop their systems further (it does a good job of finding similar images
to ones that you have posted, even if they are not tagged and are pretty
abstract) but also to develop new AI applications such as Pinterest Lens which
enables you to point your phone camera at an object and it will return pictures of objects that are visually similar, have related ideas or are the same
object in other environments.
Other applications of Image Recognition are not so altruistic. A website in
Russia, where privacy laws are more lax than most Western countries, enables
its users to identify people in the street using their phone camera. Find Face,
as the website is called, exploits the fact that all 410 million profile pictures on
the country’s most popular social media site, VK, are made public by default.
Therefore, it is able to match, to an apparent accuracy of 70%, a face that you
point your camera at with a VK profile. The intended, and rather bawdy, uses
of this application are clear from the pictures of women on the website’s
homepage, but it does demonstrate how well image recognition technology
can work if it has a large enough training set to work from.
Image recognition is still at a relatively immature stage—many useful
things are possible now, but the potential is much greater. The use of still and
moving images in our daily lives and in the business world is increasing exponentially and therefore the ability to index and extract meaningful data from
these will be an increasingly important application.

Speech Recognition
Speech recognition, sometimes referred to as speech-to-text, is usually the first
stage in a string of AI capabilities where the user is providing instructions by
voice. It takes the sounds, whether live or recorded, and encodes them into
text words and sentences. At this point other AI capabilities such as NLU
would be needed to determine the meaning of the encoded sentences.

34

3 AI Capabilities Framework

Speech recognition has benefited enormously from the development of
DNNs although some of the more ‘traditional’ AI approaches (usually using
something called a Hidden Markov Model or HMM) are still widely used
today, mainly because they are much more efficient at modelling longer sections of speech. Just like for the images, the input data is unstructured, with
the technology using supervised learning to match the encoded words with
tagged training data (as such there are a number of publicly available speech-­
related training sets).
There are many challenges to an efficient and accurate speech recognition
system, as most readers will have experienced themselves whilst trying to get
their smartphone to understand a voice command from them. One of the
main challenges is the quality of the input—this could be because of a noisy
environment or because the voice is on the other end of a telephone line—
accuracy on the phone drops from a best-in-class Word Error Rate (WER) of
around 7% to more than 16%. (Human accuracy is estimated to be around
4% WER.)
Other challenges include obvious things like different languages and
regional accents but a key consideration is the size of the vocabulary that will
be required. For a very specific task, say checking your bank balance, the
vocabulary will be very small, perhaps ten or so words. For systems that are
expected to answer a wide range of questions, such as Amazon’s Alexa, then
the vocabulary will be much larger, and therefore present a greater challenge
for the AI.
A wider vocabulary requirement also means that context becomes more
challenging to determine. Context is important in speech recognition because
it provides clues to what words are expected to be spoken. For example, if we
hear the word ‘environment’ we would need to understand its context to
determine whether the speaker meant ‘our immediate surroundings’ or ‘the
natural world’. DNNs, and a particular type of DNN called a Recurrent
Neural Network, are very good at looking backwards and forwards through
the sentence to constantly refine the probability that it is one meaning or
another.
It is worth pointing out that speech recognition is a slightly different concept to voice recognition. Voice recognition is used to identify people from
their voices and not to necessarily recognise the words they are saying. But
many of the concepts used in the two applications are similar.
Speech recognition, and its cousin NLU, are seeing plenty of development
activity as people become more used to the technology through their smartphones, and therefore more comfortable and confident with talking to
machines. As a ‘user interface’, speech recognition will likely become the
dominant input method for many automated processes.

Search

35

Search
I use the word ‘Search’ in a specific sense here (another common term for this
is ‘Information Extraction’). What this AI capability is all about is the extraction of structured data from unstructured text. So, just as the Image
Recognition and Speech Recognition capabilities worked their ‘magic’ on pictures and sounds, so the Search capability does with documents.
There are strong elements of Natural Language Processing (NLP) used in
this capability to extract the words and sentences from the passages of text, but
I prefer to use the term ‘Search’ because it better describes the overall capability. Later on in this chapter you will hear about Natural Language Understanding,
which is generally described as a sub-set of NLP, but, in my mind, is a capability in itself which complements the outputs of Speech Recognition and Search.
AI Search, which is almost exclusively a supervised learning approach,
works on both unstructured and semi-structured data. By unstructured data
I mean something like a free-form email or a report. Semi-structured documents have some level of consistency between different instances but will
have enough variability to make it extremely difficult for a logic-based system
(such as a Robotic Process Automation [RPA] robot) to process.
A good example of a semi-structured document is an invoice. An invoice
will generally have the same information as another invoice, such as the supplier’s name, the date and the total value. But one may have a VAT amount on
it, another may have a sub-total and a total, another may have the address
written in the top right-hand corner rather than the top left, another may
have the supplier’s name written slightly different and another may have the
date written in a different format.
The traditional way of getting the information off the invoice and into the
appropriate ‘system of record’ would be to use an Optical Character
Recognition (OCR) system and a template for each different version of the
invoices (which could run into the hundreds) so that it knew where to find
each piece of data. An AI system, once trained on a sample of invoices, is able
to cope with all of the variabilities I have described. If the address is in a different place it will be found; if there is usually a VAT line and it’s not there
then that doesn’t matter; if the date is in a different format it will be recognised (and converted into the standard format).
Interestingly, the AI works in the opposite way to the template approach—as
more and more ‘versions’ of the document are found, rather than becoming
more uncertain and having to create more and more templates, the AI become
more confident and is able to cope with even more variability. This is because it
is essentially matching the ‘patterns’ of one document with the learned patterns
of the documents used in training.

36

3 AI Capabilities Framework

In the case of unstructured text such as a free-form email, the AI can do two
things. The first is to categorise the text by matching the patterns of words
with what it has already learnt. For example, if you gave the AI a random news
article it could work out whether that article was about politics, business,
sport and so on, as long as it had been trained on many other tagged news
articles. It would do more than simply look for the word ‘football’ to identify
a sports article (there are many business articles about football, for example)
but would instead look at the article holistically, creating an algorithmic
model that represents ‘sports articles’. Any new article with a similar model,
or pattern, would likely be about sports.
The second main thing that an AI search capability can do is to extract
‘named entities’. A named entity could be a Proper Noun—a place or a person, for example—or even a date or quantity. So, in the passage of text,
“Andrew Burgess, who lives in London, wrote his second book in 2017 which
was published by Palgrave Macmillan”, the named entities would be ‘Andrew
Burgess’, ‘London’, ‘second’, ‘2017’ and ‘Palgrave Macmillan’.
There are specific Named Entity Recognition (NER) algorithms that carry
out this task, but they all need to be trained and tuned to make them as accurate as possible. The best NER systems for English currently produce near-­
human performance. Accuracy for NER can be improved by training them on
specific domains of information, so one could be trained on legal documents
and another on medical documents, for example.
The categorisation and entity extraction tasks can be combined to enable
free-form, that is, unstructured, text to be ‘read’ and categorised, and then
have all the meta-data extracted. This would mean that, for example, a customer who emailed a request to a company could have that email categorised
so that it can be automatically forwarded to the right person in the organisation along with all the relevant meta-data that has been extracted. This meta-­
data can be automatically entered into the company’s Case Management
System so that the customer service agent has all the information available
when the case is received.
Search, or Information Extraction, is probably one of the most developed
capabilities in the Framework. There are established software vendors with
relatively mature products, as well as many start-ups all busy building
­applications in this space. As you will find in subsequent chapters, the principal attraction of this capability right now is that it provides a useful complement to RPA; the robots need structured data as their inputs and AI Search
can turn unstructured text into structured data, thus opening up many more
processes that can be automated through RPA.

Clustering

37

Clustering
All the capabilities I have described so far work on transforming unstructured
data (images, sounds, text) into structured data. In contrast, Clustering, our
fourth capability, works on structured data, and looks for patterns and clusters
of similar data, within that data; that is, it is a ‘classifier’. The other aspects of
this capability that are different from the others is that it can (but by no means
has to) learn unsupervised, and it takes a very statistical approach. But just
like the others, it still requires large amounts of data to be truly valuable.
Clustering is usually the first part of a series of stages that usually end with
a prediction, for example extracting insights from new data based on alignment with the original patterns, or identifying anomalous new data where it
doesn’t match the expected patterns. In order to be able to make those predictions or identify the anomalies, the patterns/clusters must first be discovered.
In its simplest form, this type of AI uses statistical methods to find the ‘line
of best fit’ for all the data (in mathematical terms it is minimising the square
of the distance from all the points to the line). It is then just a case of making
those statistical methods more and more complex in order to cope with more
and more features of the data. If there isn’t enough data, then the solution can
suffer from what is called ‘over-fitting’—this is where a theoretical line of best
fit is calculated but it bears little resemblance to any real trends in the data.
The more data you have, the more confident you can be in the patterns that
are found.
A good example of the use of clustering is to be able to identify similar
groups of consumers within customer buying-behaviour data. Humans will
usually be able to identify patterns within small data sets and will often use
past experience to help them shape those patterns. But where there are thousands or millions of data points with multiple characteristics and features,
humans will find it impossible to process all that information, and the AI will
come into its own.
The benefit of the approach, apart from the ability to use sheer computing
power to analyse data quickly, is that the AI is effectively naive to the data. It
is only looking for patterns in numbers, and those numbers could relate to
anything, including a person’s height, their salary, their eye colour, their postcode, their gender, likes, dislikes and previous buying history. If you were in
charge of the loyalty card data for a retailer, you may not have realised that
there is a correlation between eye colour and propensity to buy yoghurt (I’m
making this up) but, if it exists, the AI will find it. No one would have thought
of trying to match these two features and so, if they were using a traditional

38

3 AI Capabilities Framework

Business Information tool, wouldn’t have thought to ask the question in the
first place. It is this insight that AI can bring that provides much of its value.
Just like the other capabilities in the ‘Capturing Information’ group, the
Clustering approaches are reasonably mature and are being used in business—
it is the basis of the field of Predictive Analytics. You only have to buy anything online to see how you have been pattern-matched against other
consumers and then presented with an offer to buy something that they have
bought in the past. Or you may have received a special offer from your mobile
phone provider because some of your usage patterns and behaviours have
indicated you may be a ‘flight risk’ and they want to keep you as a customer.
Or you may have received a call from your bank because of some of your
spending behaviour doesn’t fit into the normal patterns raising the possibility
that you may have had your credit card skimmed.
It is this Clustering capability that benefits more than any of the others
from the sheer propensity of data in the world today.

Natural Language Understanding
Now we are ready to move on to the next type of AI objective—‘determining
what is happening’—and its first capability, Natural Language Understanding,
or NLU.
NLU is an important part of the AI world because it provides some meaning to all the text we are bombarded with, without having to resort to humans
to read it all. It acts as a ‘translator’ between humans and machines, with the
machine having to do the hard work. NLU is closely related to the Search
capability I discussed earlier: some people would actually group Search under
the wider banner of NLP, but I prefer to split them out as they are generally
used for different objectives in business. The same can be said of Speech
Recognition, but again, using these capabilities in the context of the business
setting, they provide different objectives. NLU is still, to a certain degree,
turning the unstructured data of a sentence into the structured data of an
intent, but its primary purposes are to work out what is the right structure
(syntactic analysis) and what is the meaning (semantic analysis) of the words
and sentences.
NLU has a special place in the history of AI because it forms the basis of
the Turing Test. This is the test that English polymath Alan Turing devised in
the 1950s to define whether a machine could be classed as artificially intelligent or not. In the test, an evaluator would have a typed conversation with a
computer and a human, both hidden behind screens—the computer would

Natural Language Understanding

39

pass the test if the evaluator could not tell which of the conversations was with
a computer or a human. The Turing Test is now an object of some debate—it
is not only highly influential, particularly when discussing the philosophy of
AI, but also highly criticised. A number of systems, most notably ‘Eugene
Goostman’, a Russian chatbot, have claimed to have passed the test, but the
real question, now that AI has developed so much further, is whether it is a
valid test of intelligence in the first place.
NLU uses supervised learning with machine learning in order to create a
model of the input text. This model is probabilistic, meaning that it can make
‘softer’ decisions on what the words mean. NLU is a very complex research
field and I won’t begin to go into the details here, suffice it to say that there are
many challenges, such as being able to cope with synonymy (different words
that have similar meanings) and polysemy (words that have several meanings).
The phrase ‘Time flies like an arrow. Fruit flies like a banana.’ perfectly sums
up the challenge that NLU researchers face.
We all see NLU in action most days. Siri or Cortana or Alexa all have
usable natural language interfaces. The majority of the time they will understand the words you are saying (using Speech Recognition) and convert that
into an intent. When these systems don’t perform well it is either because the
words weren’t heard correctly in the first place (this is the most common reason) or the question that has been asked doesn’t make sense. They will be able
to understand different versions of the same question (“What was the football
score?”, “Who won the football match?”, “Please can you tell me the result of
the football?”, etc.) but if you ask “Score?” they will struggle to come up with
a relevant answer. Most systems will have default answers if they can’t work
out what you want.
Chatbots, where the questions and answers are typed rather than spoken,
use NLU but without the ‘risk’ of Speech Recognition getting the word interpretation wrong. They therefore tend to be the simpler and preferred choice
when businesses are looking to create NLU interfaces with their customers,
but the use of speech recognition shouldn’t be discounted if the environment
and benefits are good enough.
And, of course, these personal assistants can talk or type back to you, but
the answers tend to be stock phrases. For more bespoke responses a specific
sub-set of NLP called Natural Language Generation (NLG) is required. NLG,
which can be considered as the ‘opposite’ of NLU, is probably the hardest part
of NLP, which is why it has only recently become commercially available.
Rather than the limited set of response phrases used in the majority of chatbots, creating whole new phrases and even articles is possible with
NLG. Applications include creating hyper-local weather forecasts from

40

3 AI Capabilities Framework

weather data, or financial reports from company’s earnings data or stock market data. Each of these provides a short narrative in a natural language that
can be hard to differentiate from one generated by a person.
NLU is also used to try and understand the emotion behind the sentences,
an area known as sentiment analysis. In its simplest form sentiment analysis
looks for ‘polarity’, that is, is the text positive, negative or neutral. Beyond
that it will look for the type of emotion that is being expressed—is the person
happy, sad, calm or angry, for example. The AI will, of course, look for specific
words within the text but it will try and put these in context, as well as try to
identify sarcasm and other tricky idiosyncrasies of language.
Sentiment analysis is already being used widely to assess online text from
customers, such as through Twitter and TripAdvisor. This allows businesses to
evaluate how their products or services are being perceived in the market
without the expense of focus groups or surveys, and allows them to respond
to potential issues in real time. Specific models, with the associated terminology and jargon, are usually generated for particular needs; understanding how
customers perceive a hotel room will be quite different to how they feel about
their mobile phone.
Another common application for natural language technologies is machine
translation. This is where a computer will translate a phrase from one language to another. It is a complex problem to solve but recent developments in
deep learning have made this much more reliable and usable. The key to success, as with most AI challenges, is the availability of data, and companies like
Google have used the transcriptions from the European Parliament, where the
proceedings are translated by humans between 24 different languages, as its
training set. That means that translating from, say, English to French is now
highly accurate, but for less common languages, such as Khmer, the official
language of Cambodia, translations are generally done through an intermediary stage (usually the English language). Chinese is notoriously difficult for
machines to translate, mainly because it is more of a challenge to edit sentences as a whole—one must sometimes edit arbitrary sets of characters, leading to incorrect outcomes.
The natural evolution of machine translation is real-time translation of
speech (epitomised by the Babel Fish in The Hitchhiker’s Guide to the Galaxy).
Already smartphone apps are able to read and translate signs (using Image
Recognition, NLU and machine translation) and can act as interpreters
between two people talking to each other in different languages. Currently
these capabilities are less reliable than typed translations and only work for
common language pairs, but the potential uses for this technology in a business
environment are huge.

Optimisation

41

NLU is leading the charge for AI in business: it allows us to communicate
with computers in the way we feel most comfortable with—we don’t all have
to be computer wizards to be able to work them effectively. But NLU is inherently a difficult technical challenge and will always be compared with, and
evaluated against, humans (unlike other AI capabilities such as Clustering
which can easily surpass human performance). Therefore, until it can be said
to be indistinguishable from us (i.e. truly pass the Turing Test) NLU will
always be open to criticism. Applied in the right environment and with the
right processes though, it can deliver substantial benefits to businesses.

Optimisation
In all the capabilities I have discussed so far, we have been manipulating data
to transform it from one form to another (images to descriptions, sounds to
words, words to meanings, text to information and big data to meta-data).
But even though the transformed data is much more useful to us than the
source data was, we haven’t yet done anything really meaningful with it. That’s
where the AI capability of Optimisation comes in.
Optimisation is at the heart of what people generally think AI does. It is the
closest analogy we get to mimicking human thought processes without having
to call on true cognitive Understanding.
I use the title ‘Optimisation’ for brevity, but it actually includes problem
solving and planning, which makes it quite a broad subject that has a lot of
science behind it all. A broad definition of this capability would be that if you
know a set of possible initial states as well as your desired goal and a description of all the possible actions to get you there, then the AI can define a solution that will achieve the goal using an optimum sequence of actions from any
of the initial states.
Historically, as I discussed at the start of Chap. 2, optimisation and problem solving were achieved through ‘expert systems’, which were really nothing
more than decision trees which had to be designed and configured up-front
by humans. With the advent of machine learning, much of that design and
configuration is carried out by the AI itself using an extreme version of trial-­
and-­error. (There is still, by the way, a place for knowledge-based AI systems
in the world today, as I shall discuss later.)
A useful way to understand the Optimisation capability is by looking at where
AIs are taught to play computer games. Scientists and researchers use computer
games as a kind of test and benchmark for how clever their systems are: can this

42

3 AI Capabilities Framework

particular AI beat the best human at playing that particular game? But it also
provides a very useful way of explaining what the different flavours are.
So, for example, to play chess and beat Gary Kasparov required brute
force—these systems ‘simply’ analysed every possible move much further
ahead than a human can do. That is more logic than it is AI. To learn and win
at Space Invaders or Breakout, however, needs DNNs and what is called ‘reinforcement learning’ (more of which later). From only being given the objective to maximise the score, and no other information on how to play, these AIs
perceive the state of the screen, trying lots of different approaches until the
score increases (shooting aliens, not getting bombed, etc.). Then, through
reinforcement, they start to develop strategies (such as getting the ball behind
the wall in Breakout) that get higher scores. It’s a great example of self-­learning,
but in a relatively simple environment.
In order to learn the Chinese game of Go, things are a lot trickier. Until
very recently, a computer being able to play Go was a Holy Grail for AI
researchers, mainly because the game relies more on ‘intuition’ than logic, but
also because the number of possible move combinations is trillions of times
greater than a game of chess (there are more possible Go moves than there are
atoms in the universe). In 2016, an AI designed by DeepMind, a British company owned by Google, beat the best Go player in Europe four games to one.
AlphaGo, as the system is called, uses the same concept of reinforcement
learning, this time based on studying 30 million moves from human Go
matches. It then plays different versions of itself thousands of times to work
out the best strategies, which it uses to create long-range plans for the actual
games. Apparently, during the second match, the human player, Lee Sedol,
‘never felt once in control’ (a general fear of AI that is shared by many people,
it must be admitted). Since then, AlphaGo has beaten the world’s best Go
player, Ke Jie, three games to zero.
So, the key characteristic of the Optimisation capability is that there is a
goal to be achieved—an idea that needs to be reasoned, a problem to be
solved or a plan to be made. At its simplest form this is done through iterative
trial-and error—small changes are made to the environment or small actions
are taken, after which the situation is evaluated to see whether it is closer to
the goal or not. If it is closer it carries on and makes further changes, if not it
tries something different.
As I’ve hinted with the gaming examples above, there are some subtleties to
this overly simplified description, which all come together to provide the AI
Optimisation capability. I’ve described a number of the most common
approaches below, most of which are inherently inter-connected.

Optimisation

43

Cognitive reasoning systems are the modern equivalent of the old Expert
Systems (which is why some people are reluctant to describe them as true AI
today). They work by creating a ‘knowledge map’ of the specific domain that
is being modelled, which means that they do not require any data to be trained
on, just access to human subject matter experts. These knowledge maps connect concepts (e.g. food and person) with instances (e.g. chips and Andrew
Burgess) and relationships (e.g. favourite food). Different relationships can
have different weights or probabilities, depending on their likelihood, which
means that the system can be interrogated to make recommendations. Unlike
decision trees, the system can start from any point on the map as long as a goal
has been defined (‘what is Andrew’s favourite food?’, ‘which person likes
chips?’) and can handle much greater complexity: its ‘clever’ bit is in the way
it comes to a recommendation using the shortest route through the map
rather than asking a linear series of questions. Compared to machine learning
approaches, cognitive reasoning systems have the major advantage that the
recommendation that has been made is fully traceable—there is none of the
‘black box’ characteristics that machine learning systems suffer from. This
makes these sorts of systems ideal for regulated industries that ‘need to show
their working’ (Fig. 3.2).
Beyond Cognitive Reasoning, the majority of AI optimisation approaches
are based around algorithms. A key AI concept involves the idea of breaking
a big problem down into much smaller problems and then remembering the
solutions for each one. This approach, known as Dynamic Programming, is
Place
located in

studied at
Establishment

Person
awarded

studied
specialises in

Qualification
Subject

Fig. 3.2 Knowledge map

type

44

3 AI Capabilities Framework

epitomised in the Coin Change Problem, which asks: how can a given amount
of money be made with the least number of coins of given denominations? If
my coins are in units of 1,4,5,15,20 and the total amount I need to make is
23, then a system which looked at it as a series of independent, ‘blind’ decisions would start with the biggest coin first and then add smaller coins until
the total was reached: this would be 20+1+1+1, so four coins. A dynamic
system though would have broken the problem down into smaller problems,
and stored each of the optimal solutions. It would have come up with an
answer of 15+4+4, so just three coins.
The system works by looking ahead at lots of different ways that the end
objective could turn out to be based on, analysing a large sample of the different individual steps that can be taken. Those sample actions that lead to, or
get close to, the desired goal are remembered as favourable and therefore more
likely to be chosen.
A common AI approach that exploits this method is called the Monte
Carlo Tree Search (MCTS). When, for example, playing a game it can ‘play
out’ lots of different versions of potential moves of both players, with each
move creating a branching tree of child moves. Using back propagation (a
form of feedback loop), those moves that achieve the desired goal have the
connections between each of the nodes that led to that result strengthened so
that they are more likely to be played.
These MCTS-like approaches do have a couple of drawbacks: they need to
balance the efficiency of a small sample size of moves with the breadth of covering as many moves as possible (this leads to some randomisation being built
into the choice of child nodes), and they can be slow to converge on to the
optimal solution.
Reinforcement Learning, an area of AI that is seeing a great deal of activity
at the moment, helps to alleviate some of these inefficiencies. Reinforcement
learning is really an extension of dynamic programming (it is sometimes
referred to as ‘approximate dynamic programming’) and is used where the
problems are much more complex than, for example, the coin change problem
or simple board games, and where there are more unknowns about the states
of each move that is taken. Reinforcement learning uses extreme ­trial-­and-­error
to update its ‘experience’ and can then use that machine-learned experience to
determine the most optimal step to take next, that is, the one that will get it
closer to achieving the goal. This approach is slightly different to the supervised learning approaches I described in Chap. 1 in that the ‘right answer’ is
never given in the training, nor are errors explicitly corrected: the machine
learns all of this itself through the trial-and-error iterations.

Optimisation

45

A tactic that is usually employed in reinforcement learning is to have the AI
systems play against each other. DeepMind’s AlphaGo, once it had been
trained on human games, then played itself (or, to be exact, two slightly different versions of AlphaGo played each other) over thousands of games in
order to refine its game even further. If it had only learnt from human games
it could only ever be as good as humans, whereas if it plays itself, it is possible
for it to surpass human skill levels. This combative approach, known as
Generative Adversarial Networks, as well as allowing the algorithms to tune
themselves, is a great way to run simulations to test the viability of AI systems
and to run different war-gaming scenarios.
Another ‘weakness’ in AI Optimisation approaches has been that the systems tend to look at the short-term gains at the expense of the longer-term
strategy. Recent advances in AI have seen a number of different systems combined to provide both these perspectives, especially where the domain is particularly complex: a ‘policy’ algorithm will look at the next best move, whilst
a ‘value’ algorithm will look at how the problem, or game, might finish up.
The two algorithms can then work together to provide the best outcome.
Facebook has already been able to train chatbot agents to negotiate for
simple items, in some cases doing it as well as humans. Facebook Artificial
Intelligence Research first used supervised learning to train the agents on a
large number of human-to-human negotiation scripts. This stage helped them
imitate the actions of the humans (mapping between language and meaning),
but didn’t explicitly help them achieve their objectives. For this they used
reinforcement learning, where two AI agents would ‘practise’ negotiating with
each other (interestingly they had to fix the language model in place at this
stage because they found that, if the agents were allowed to continue learning
the language elements, they would start to create their own private language
between them). At the end of each negotiation, the agent would be ‘rewarded’
based on the deal it had managed to achieve. This was then fed back through
the model to capture the learning, and make the agent a better negotiator.
Following training, it could then negotiate for the same types of items with
humans, matching the capabilities of other human negotiators.
Similar ideas of combining a number of AI techniques were used when an
AI system, called Libratus, was able to beat experienced players of poker—this
system used three different types of AI: the first relied on reinforcement learning to teach itself the game of poker from scratch; a second system focused on
the end game, leaving the first system to concentrate on the immediate next
moves; and, because some of the human players were able to detect trends in
how the machine bet, a third system looked for these patterns overnight and
introduced additional randomness to hide its tracks.

46

3 AI Capabilities Framework

Optimisation AI can be applied to many situations where there is a specific
goal to be achieved, such as winning a hand of poker or negotiating for items.
It can, as I mentioned earlier, provide performance that exceeds human capabilities. Other typical uses for the Optimisation capability include route planning, designing rota for shift workers (such as nurses) and making
recommendations.
All the approaches described in this section are just a sample of how AI can
attempt to solve problems, but should give you a feel for the general strategies
that can be employed. At the heart of it is the idea that big decisions can be
broken down into many small decisions, and that these can then be optimised
using trial-and-error so that a defined goal can be achieved, or a specific
reward maximised.

Prediction
Prediction employs one of the core ideas of AI in that it uses lots of historical
data in order to match a new piece of data to an identified group. Thus, prediction generally follows on from the Clustering capability described earlier in
the chapter.
I’ve already mentioned one of the more common uses of prediction, that of
recommending related online purchases (‘you bought this book therefore you
will probably like this other book’). So, in this case, what the retailer calls a
‘recommendation’ is actually a prediction they make in order to sell you more
stuff.
Some decisions can be described as predictions—if you apply for a loan
that is assessed by a machine, then the AI will try and predict whether you will
default on the loan or not. It will try and match your profile (age, salary, regular spending, other loans, etc.) with other customers that have similar profiles.
If those other matching customers generally default on their loans then you
will likely be refused the credit line.
Prediction can also follow on from the Search capability. Because Search
can look for patterns within text, it can match these with patterns from some
pre-defined requirements. For example, CVs (aka résumés) can be matched
with job descriptions in order to predict good candidates for the role.
The prediction capability differs from Optimisation in that it doesn’t have
a specific goal to achieve. There are no steps to determine to achieve a specific
objective—we are ‘just’ matching a new data point with our historical data.
At a simple level, predictions can be made from just a few ‘features’—these are
the specific characteristics that are being measured, such as the number of bedrooms

Prediction

47

in a house and the size of the garden. If you had a table of this data and the actual
price of each of the houses, then you could use this capability to predict the price of
another house if you knew the number of bedrooms and size of the garden. And
you wouldn’t necessarily need AI to do this, or even a computer.
But there are generally more factors (or features) which determine the price
of a house than just these two things. You might also consider how many stories
it has, whether it is detached, semi-detached or terraced, whether it has parking,
a utility room, a swimming pool, its postcode and so on. All this makes it much
more difficult to predict the house value without using some AI ‘magic’.
Clearly, the more features that are considered, the more training data needs
to be used, so that as many ‘variations’ of all the features can be captured and
built into the model. This model will have considered all the different influences (or ‘weights’) that each of the features has on the house price. By entering in the values of a selection of features of a different house, the model fits
(using Regression Analysis in this case) those features as closely as possible and
can therefore predict the house price. It is normal for the system to give a
figure for its level of confidence (i.e. the probability that it is correct).
Once the number of features is in hundreds or thousands then things become
more complex—this just means that more algorithms, more training data and
more computing power are needed. At the extreme end of this, the systems that
are used to predict our weather are some of the most powerful in the world.
An important aspect of the Prediction capability is to remember that it is
essentially naive. By this I mean that it is only manipulating numbers and has
no underlying understanding of what that data actually means. Those house
features could be car features or weather features or people features—to the
computer they are just numbers.
This naivety, whilst seeming like a virtue, is also one of Prediction’s biggest
challenges: that of unintended bias. Normally AI is lauded because it is not
influenced by the prejudices that humans inherently suffer from (there are
many studies that show that, although people may claim not to be biased in,
say, recruiting staff, there is usually some sort of unconscious bias involved).
But the unbiasedness of AI is only as good as the data that is used to train it—if
the training set contains CVs and hiring decisions that have been made by
humans, the biases that were in the original data will be ‘trained into’ the AI.
Another key issue for AI Prediction is the opaqueness of the decision-­making
process. When an AI makes a prediction, for example that a person is more than
likely to default on a loan and therefore their application should be rejected, that
prediction is based on all the training data. The training data has been used by the
machine to build an algorithmic model that it then uses to predict new cases from.
For the more complex algorithmic models this is just a matrix of numbers that

48

3 AI Capabilities Framework

makes no sense to a human trying to read it, and therefore the reason (or, more
likely, reasons) that the person had their loan refused is not easily available to scrutinise. (Simpler algorithmic models, such as Classification and Regression Trees,
do provide some transparency and are therefore more popular). Of course, we
don’t need to understand the workings of every prediction that is made (it would
be interesting to know why the house price predictor came up with that valuation,
but it is the actual house price we are most concerned with), but some industries,
especially those that are regulated, will require it. Also, if you are the person that
has been refused the loan, you should have the right to know why.
AI is being used to predict many things, including, in some US courts, a
defendant’s risk profile. When the consequences of those predictions could
have an influence on whether someone is found innocent or guilty, then the
challenges I have described above become very serious indeed. As Melvin
Kranzberg’s first law of technology succinctly states: technology is neither
good nor bad; nor is it neutral. These challenges of ‘algorithmic transparency’,
naivety and unintended bias will be discussed in more detail in Chap. 8.
AI Prediction is one of the most active areas in the field at the moment.
Where there are lots of good data, predictions can generally be made from
that data. That doesn’t necessarily mean that predictions need to be made or
that they will be useful in any way, but there are lots of cases where this can be
very beneficial indeed, including predicting valuations, yields, customer
churn, preventative maintenance requirements and demand for a product.

Understanding
I include this section on AI Understanding in the book only as a way to
describe what is not currently available in the AI world to businesses, or to
anyone outside of a research lab. It should be seen as a countermeasure to all
of the hype that can be put out by over-excited marketing departments.
By ‘understanding’, I am generally referring to the ability of a machine to
have conscious awareness of what it is doing or thinking (or to act like it
does—see next paragraph). This implies it can understand the intent and
motivations of people, rather than just blindly crunching numbers about
them. It is usually described by the concept of Artificial General Intelligence
(AGI), where the AI is able to mimic all of the capabilities of the human
brain, not just the very narrow ones we have discussed so far.
There is an interesting subtlety to the description of AGI. John Searle, a
philosopher, differentiated ‘strong AI’ and ‘weak AI’. A strong AI system can
think and have a mind, whereas a weak AI system can (only) act like it thinks

Understanding

49

and has a mind. The former assumes that there is something special about the
machine that goes beyond the abilities that we can test for. Ray Kurzweil, the
futurologist, simply describes strong AI as when the computer acts as though
it has a mind, irrespective of whether it actually does. For our purposes, we
will stick to this definition as we are dealing with practicalities rather than
philosophical debate.
So, what sort of tests would we use to claim strong AI? I have already mentioned the Turing Test as a (limited) test of AI. Other tests that people have
proposed are:
• The Coffee Test (Wozniak)—A machine is given the task of going into an
average home and figuring out how to make coffee. It has to find the coffee
machine, find the coffee, add water, find a mug and brew the coffee by
pushing the proper buttons.
• The Robot College Student Test (Goertzel)—A machine is given the task
of enrolling in a university, taking and passing the same classes that humans
would, and obtaining a degree.
• The Employment Test (Nilsson)—A machine is given the task of working
an economically important job, and must perform as well or better than
the level that humans perform at in the same job.
No AI systems are anywhere near passing these tests, and any systems that
are even close will be a combination of many different types of AI. As I’ve
described in the previous sections of this chapter, there are a number of different types of AI capabilities, each one very specialised in what it does. This
means, for example, an AI capability that is being used to recognise images
will be useless at processing language. This is the concept of Artificial Narrow
Intelligence (ANI). Even within the capability groups I have described there
is little or no crossover between specific uses—if I have a system that extracts
data from invoices, it will not be able to do the same for remittance advices,
without training the system from scratch. The same could even be true if I
wanted to take a trained invoice extractor from one business to the another—
there may be enough variation between businesses to mean that retraining
would be required.
Where our brain is so much cleverer than AI is where it is able to use different cognitive approaches and techniques in different situations and, importantly, take learnings from one situation and apply them to a completely
different one. For example, I may know that the value of a house generally
increases with the number of bedrooms it has, but I can then apply the same
concept to other objects (computers with more hard drive capacity cost more)

50

3 AI Capabilities Framework

but also know that this shouldn’t be applied to everything (cars with more
wheels are generally not more expensive). AI is not able to do this right now.
There is work being done to try and create AGI. At a national scale, there
is the Blue Brain Project in Switzerland (which aims to create a digital reconstruction of the brain by reverse-engineering mammalian brain circuitry), and
the BRAIN Project in the United States, which is also looking to model real
brains. Organisations such as OpenCog (an open-source AGI research platform), the Redwood Center for Theoretical Neuroscience and the Machine
Intelligence Research Institute are all at the forefront of researching various
aspects of AGI.
Interestingly, there has been some progress in getting neural networks to
remember what they have previously been taught. This means they could, in
theory, be able to use learnings from one task and apply it to a second task.
Although that may sound like a simple thing for a human to do, ‘catastrophic
forgetting’ (when new tasks are introduced, new adaptations overwrite the
knowledge that the system had previously acquired) is an inherent flaw in neural networks. DeepMind, the UK AI company owned by Google, is developing
an approach it calls Elastic Weight Consolidation which allows the algorithm to
learn a new task whilst retaining some of the knowledge from learning a previous task (they actually use different Atari computer games as their test tasks).
This is showing promise but is a long way from being put to practical use.
Despite the research and small steps that are being taken, the ability for a
computer to fundamentally understand what it is doing is still a long way off
(some argue that it could never be possible). Even using Searle’s definition of
weak AI, there are major hurdles to overcome. Some of these are very technical (such as the challenge of catastrophic forgetting) and some are simply
down to the fact that the necessary computing power doesn’t exist at the
moment. But, as described throughout this book, there are huge advancements being made in ANI that provide real benefits to people and businesses.
Some of these make our lives simpler whilst others are able to far exceed our
own competencies in specific tasks. By understanding each of these capabilities, and their respective limitations, we are able to benefit from AI technologies today and tomorrow.

Using the AI Framework
The AI Capability Framework is my attempt to bring some clarity and order
to the diverse, complex and often confusing field of AI. By ‘boiling it down’
into a set of discrete capabilities, AI hopefully becomes accessible to those

Using the AI Framework

51

who want to benefit from the technology but do not have the skills, ­experience
(or desire) to understand the technical aspects beyond the high-level appreciation I have provided in this chapter.
I have tried to differentiate each of the capabilities as much as possible, but
there are inevitably some overlaps between each of them, for example between
speech recognition and NLU, and clustering and prediction. The boundaries
between these are blurry; some people would describe both speech recognition and NLU as a sub-category of NLP but I think they sit better as discrete
capabilities; and some people might separate out planning and optimisation,
but I think they are close enough aligned to keep them as one. So, please don’t
get too hung up on some of the nuances—AI is a complex subject, with many
different viewpoints and opinions, and is constantly changing. The framework should be treated as your guide rather than a technical manual (Fig. 3.3).
So, the knowledge you now have should enable you to do three things:
• Identify the right AI capabilities for your business need. What are your
business objectives and could AI provide all or part of the solution? Are you
looking to capture information or understand what is happening, or both?
Do you want to replace existing capability (computer or human) with AI,
or do you want to augment their capabilities further? Which specific capabilities will you need to create a solution? Will that require a supervised
learning approach, and what data do I have available?

CAPTURE INFORMATION

structured
data

unstructured
data

Image Recognition

WHAT IS HAPPENING?

WHY IS IT HAPPENING?

Natural Language
Understanding

Speech Recognition

Search

Optimisation

Data Analysis /
Clustering

Prediction

Fig. 3.3 The AI framework

Understanding

52

3 AI Capabilities Framework

• Look beyond the hype. What capabilities do the AI vendors actually have,
rather than those they claim they have? How do these capabilities match
my requirements? Do their assertions about their product make sense? Are
there gaps that will need to be filled with different AI solutions or traditional technology? Will it be as easy to implement as they claim?
• Be realistic. What are the limitations of the AI capabilities that I might need?
Is AI the most appropriate solution for this or are there simpler or more effective solutions? Do I have the necessary data available to train the system
adequately? Will I need to bring in external support to help me get going?
Chapter 5 provides real examples of the different AI capabilities being used
in businesses today. They are focused on common strategic objectives (enhancing customer service, optimising business processes and generating insights)
and combine different capabilities to achieve these. Each of the examples references the capabilities that are being used and, where appropriate, what challenges and limitations were experienced by those responsible for delivering
the solutions.
Some of the approaches and examples that will be described required additional technologies (such as cloud computing and RPA) to enable them to
extract their full value. The following chapter outlines what these associated
technologies are and why they are such a good fit with AI.

The AI Start-Up’s View

53

The AI Start-Up’s View
This is an extract from an interview with Vasilis Tsolis, founder and CEO of
Cognitiv+, an AI start-up in the legal sector.
AB What got you into the world of AI in the first place?
VT	My personal background is not an obvious path for a co-founder of an AI
start-up. I am a chartered civil engineer who switched my career by
undertaking a law degree. After graduation, my career was moving
between law, engineering and commercial management in various sectors
including infrastructure, construction and energy. After a few years,
something became very clear to me: people spend substantial amounts of
time reading contracts, and most of the time there is a tremendous repetition of those tasks. And that is where the problem lies: rather than focus
on advanced solution discovery, professionals consume their daily lives
with data gathering. The solution was an obvious one; using AI will free
the professionals to focus on what they are good at. That is why we started
Cognitiv+ in 2015, which has been a fantastic trip so far.
AB What is it like building an AI start up at the moment?
VT	This is a period where people apply AI or attempt to use AI in almost all
aspects of our lives, where creativity sometimes trumps technology. There
are major projects in process, such as autonomous cars and it very clear
now to everyone that AI is here to stay. It is a very exciting period, but
there are quite a lot of challenges towards the road of maturity.
Working with a number of professional clients, we see numerous
opportunities and dozens of use cases, some possible now, some reachable
within the mid-term future, some look close to science fiction.
Technology advancements advertise extraordinary abilities and this goes
throughout the food chain, where clients are constantly bombarded with
innovative ideas where the success rate is variable. This could be because they
are early in their roadmap and suffer teething problems, or because of a lack
of technical or subject matter depth which is sometimes not an easy fix.
As with any early adoption, it is the clients and investors that get the
pressure to isolate the true technologists and innovators with clear business skills, and some in-house knowledge on how technology works will
definitely bring certainty and a quick return on investment.
AB So what value do your customers get out of your software?
VT	Reading text takes time, even Romans were complaining about this and
AI can help that. Cognitiv+ targets legal text, contracts and the regulation
corpus.
Using AI should be seen as an extrapolation of automation, a technology that improves our lives, makes things faster, simpler and enables us to
explore tasks in a way that was not possible before. For example, clients
can use our software to analyse contractual risk on their whole portfolio

54

3 AI Capabilities Framework

within hours in a way that was never possible before. Time-consuming
tasks can be delegated and free up professionals’ schedules, where technology can provide them with holistic views of their third-party risk to
levels that have been unprecedented.
AB	What do customers or potential customers need to focus on if they are
going to get the maximum value from AI?
VT	While many companies hit the ground by starting to write code, there is a
lot of preparatory work that projects need to undertake and ensure success.
The first focus will be the data. Data scientists require massive data sets
and the new techniques, supported by hardware innovation, command
new levels of data quantities. The good news is that we all produce more
data every day, but this is not necessarily for all the aspects of professional
services. Some data custodians can move faster to this gold rush because
they have better access to data from others.
Quantity is not enough though—it is also the quality that will determine if the data can be used. But this is a synthetic challenge that can be
tackled only if it is isolated and understood on a case-by-case basis.
Second, teams need to make an early decision on what success would
look like, making business and technology targets that are reachable.
The fact that algorithms provide answers that are probabilistic and not
deterministic makes the way we manage AI projects very different from
any previous IT transformation projects companies have participated in.
Thirdly, an important focus would be the talent and skills diversity that
a team has. A sensible balance between business subject matter experts,
data scientists and coders would determine how successful a project will be.
AB How do you see the market developing over the next 5 years or so?
VT	While I can predict where we are heading in the next 5 months, it is practically impossible to hit anything close to a 5 year prediction. The reason
is simple—all the parameters are rapidly changing: the type of data we are
surrounded with, the size of data sets, the maturity of the algorithms,
hardware improvement, the list goes on.
But here is one prediction: as we have started using various NLP, NLU
and machine learning techniques to interpret text and documents, it
seems that certain types of text, certain sources, can be analysed better
than others. People will pick this up and will write in a way that can be
processed and summarised by machines. Why do that? It is for the same
reason where we think twice what tags we are using for our blog articles
as this will enable Search Engine Optimisation (SEO) bots to categorise
our text accordingly and disseminate it to the right channels.
Typically, we write in a way that our audience will understand us—my
take is that we will continue to do that—the difference is that in the
crowd or readers will also be some bots. What might this look like?
Perhaps, simpler and shorter phrases where the object and subjects are
clear and NLP algorithms can consume them much better?

4
Associated Technologies

Introduction
Artificial Intelligence can do a lot of things, but it can’t do everything. Quite
a lot of the time implementing a stand-alone AI solution will satisfy whatever
objectives you are looking to achieve. Sometimes AI will rely on other technologies to make it work well, and other times it will complement them so
that they both work better. This chapter covers some of the associated technologies that anybody looking to implement AI will need to consider.
Some of these technologies are software-based, such as RPA. This is a relatively new technology that automates rules-based processes but struggles to
handle unstructured data and any decision making. Cloud Computing, a
combination of software and hardware capabilities, is a key enabler of AI, and
many of the AI applications that we are seeing now would not be possible
without it.
There are also some hardware technologies that enable, and can be exploited
by, AI. Physical robots (as opposed to the software-based RPA) can be made
more intelligent with the application of AI, and the Internet of Things (IoT)
can provide very useful data sources for AI systems. One of the technologies
that I have included in this chapter isn’t really a technology at all, but is powered by human beings: Crowd Sourcing enables data to be tagged and cleaned
in an efficient and flexible way, and therefore is extremely useful to AI
developers.
I have not included more common enterprise systems, such as Enterprise
Resource Planning (ERP) and Customer Relationship Management (CRM)
systems, in this summary. All could be considered sources of data, and some
© The Author(s) 2018
A. Burgess, The Executive Guide to Artificial Intelligence,
https://doi.org/10.1007/978-3-319-63820-1_4

55

56

4

Associated Technologies

claim to have AI capabilities built into them (e.g. email systems can be a good
data source and also use AI to identify spam messages), but overall, they have
only general associations with AI.

AI and Cloud
Cloud Computing is where a network of remote servers, hosted on the
Internet, are used to store, manage and process data, rather than the ‘traditional’ approach of using a local (on-premises) server or a personal computer
to do those tasks.
Because of these abilities to store, manage and process data away from the
user’s device (PC, mobile ‘phone, etc.) and instead on high-performance,
high-capacity specialised servers, Cloud Computing has become almost an
integral part of how AI systems operate today. As cloud technology itself
matures, the two will become inextricably linked—many people already talk
of ‘Cloud AI’ as the Next Big Thing.
One of the more straightforward applications of cloud combined with AI
is the making available of large, public data sets. Most AI developers, unless
they are working for large corporates, don’t have their own data sets to train
their systems, but instead rely on these public databases. As I mentioned in
the Big Data section of Chap. 2, there exist quite a few of these datasets, covering subjects such as Computer Vision, Natural Language, Speech,
Recommendation Systems, Networks and Graphs, and Geospatial Data.
But the cloud provides more than just access to data—more often it will
actually process the data as well. We can experience this as consumers every
time we use a service like Amazon’s Echo. Although there is a small amount of
processing power on the device itself (mainly to recognise the ‘wake word’),
the processing of the words into meaning (using Speech Recognition and
NLU) is done by software on Amazon’s own servers. And, of course, executing
the actual command is also done on the cloud, perhaps sending the instruction back to your home to, say, turn the kitchen light on.
On an enterprise scale, even more of the processing work can be done on the
cloud, on huge server farms full of powerful machines. The biggest challenge
to this model though is that the data generally needs to sit on the cloud as well.
For enterprises with huge amounts of data (perhaps petabytes’ worth) then
transferring all of this to the cloud can be impractical. As a solution to very
long upload times, Amazon actually offers a huge truck (called the Amazon
Snowmobile) that contains 100 petabytes of computer storage—the truck is
driven to your data centre and plugged in so that the data can be uploaded to

AI and Cloud

57

the truck. It then drives back to the Amazon Data Centre and downloads it all.
But, as network speeds improve, there will be less and less need for these sort
of physical solutions.
Another challenge to Cloud AI is that it can be perceived as risky to store
data off-premises, especially if the data is confidential, such as a bank’s customers’ details. The security aspects can be covered by the type of cloud service that is procured—the best suppliers can now offer security provisions that
are as good as, or better than, hosted or on-premises solutions. The UK
Government in fact released guidance at the start of 2017 that states it is possible for public sector organisations to safely put highly personal and sensitive
data into the public cloud.
The cost, though, of setting up an operation that can store, manage and
process AI data effectively, securely and economically means that the market
for Cloud AI is currently dominated by a small number of suppliers, namely
Amazon, Google and Microsoft. These companies actual offer a complete set
of AI services, including access to ready-to-use algorithms.
Generally, the AI Cloud offerings are made up of four main areas (I’ve used
Amazon’s model as a basis for this description):
• Infrastructure—this consists of all the virtual servers and GPUs (the processor chips) that are required to house the applications that train and run
the AI systems
• Frameworks—these are the AI development services that are used to build
bespoke AI systems, and tend to be used by researchers and data scientists.
They could include pre-installed and configured frameworks such as
ApacheMXNet, TensorFlow and Caffe.
• Platforms—these would be used by AI developers who have their own data
sets but do not have access to algorithms. They would need to be able to
deploy and manage the AI training as well as host the models.
• Services—for those who don’t have access to data or algorithms, the AI
Services offer pre-trained AI algorithms. This is the simplest approach for
accessing specific AI capabilities with a minimum knowledge of how they
work technically.
The ability to tap into relatively complex, pre-trained algorithms is a boon
for anyone wanting to build AI capability into their applications. For example, if you want to build an AI application that has some NLU capability
(a chatbot, for example) then you could use Amazon’s Lex algorithm,
Microsoft’s Linguistic Analysis or Google’s (wonderfully named) Parsey
McParseFace. Each of these is a simple Application Programming Interface

58

4 Associated Technologies

(API), which means that they can be ‘called’ by sending them specific data.
They will then return a result to you, which can be read by your application.
One of the interesting things about these services is that they are either free
or cheap. Microsoft’s Language Understanding Intelligent Service (LUIS) offering is currently based on a threshold of API calls per month—below this
threshold it is free to use, and every thousand calls above that is charged at just
a few cents. Other algorithms are charged as a monthly subscription.
Many businesses today are taking advantage of Cloud AI rather than building their own capabilities. For example: a brewer in Oregon, USA, is using
Cloud AI to control their brewing processes; a public TV company uses it to
identify and tag people that appear on its programs; schools are using it to
predict student churn and an FMCG company uses it to analyse job
applications.
As I’ve discussed in Chap. 2, one of the motivations for these companies to
offer their AI technologies for virtually nothing is that they get access to more
and more data, which in many ways is the currency of AI. But despite the
‘greedy corporate’ overtones, Cloud AI does have the feeling of democratisation about it, where more and more people have simple and cheap access to
these very clever technologies.

AI and Robotic Process Automation
Robotic Process Automation (or RPA) describes a relatively new type of software that replicates the transactional, rules-based work that a human being
might do.
For clarity, it is important to differentiate between RPA and the traditional
IT systems. RPA—at its most basic level—utilises technology to replace a
series of human actions (which is where the ‘robot’ terminology comes in).
Correspondingly, not all technologies provide automation, and replacing a
single human action with technology (e.g. a mathematical equation in a
spreadsheet) is not considered RPA. Similarly, much automation is already
embedded into software systems (e.g. linking customer information across
finance and procurement functions), but since it is part of the normal feature
functionality of a system, it is generally not considered as RPA, but simply a
more powerful system(s).
In an ideal world, all transactional work processes would be carried out
by large, all-encompassing IT systems, without a single human being
involved. In reality, whilst many systems can automate a large part of specific processes and functions, they tend to be siloed or only deal with one

AI and Robotic Process Automation

59

part of the ­end-to-­end process (think of an online loan application process
that will have to get data from, and input data to, a web browser, a CRM
system, a credit checking system, a finance system, a KYC system, an address
look-up system and probably one or two spreadsheets). In addition, many
businesses now have multiple systems that have been acquired as point solutions or simply through numerous mergers and acquisitions. The default
‘integration’ across all of these systems, tying the whole end-to-end process
together, has traditionally been the human being. More often than not,
these human beings are part of an outsourced service.
Robotic Process Automation can replace nearly all the transactional work
that the human does, at a much lower cost (as much as 50% lower). The RPA
systems replicate (using simple process mapping tools) the rules-based work
that the human does (i.e. interfacing at the ‘presentation layer’) which means
that there is no need to change any of the underlying systems. A single process
could be automated and be delivering value within a matter of weeks. The
‘robot’ can then continue processing that activity 24 hours a day, 7 days a
week, 52 weeks a year if required, with every action being completely auditable. If the process changes, the robot only needs to be retrained once, rather
than having to retrain a complete team of people.
So, to give a simple example, if a law firm is managing a property portfolio
on behalf of a client, they would be expected to carry out Land Registry
checks at some point. This is commonly a paralegal role that might involve
the person getting a request from a lawyer, or from the client directly, probably via a template, an email or a workflow system. The person would read the
relevant information off the form, log into the Land Registry site, enter the
information into the site and read the results that came back from the search.
They would then transpose that information back onto the form and respond
to the initial request. It is possible for this whole process to be handled by a
software ‘robot’ without the need for any human intervention.
Although this is a very simple example, it does demonstrate some of the
benefits of RPA:
• the cost of the robot is a fraction of the cost of the human (between a third
and a tenth);
• the robot works in exactly the same way as the human would, so no IT or
process changes are required;
• once trained, the robot will do the process exactly the same way 100% of
the time;
• every step that the robot takes is logged, providing full auditability;

60

4 Associated Technologies

• the robot can carry out the process in the middle of the night or over a
weekend if necessary and
• the robot will never be sick, need a holiday or ask for a pay rise.
What this means is that wherever there are processes that are rules-based,
repeatable and use (or could use) IT systems, the person doing that process
can be replaced by a software robot. Some further examples of processes that
can be automated are:
•
•
•
•
•
•

Employee on-boarding
Invoice processing
Payments
Conveyancing processing
Benefit entitlement checks
IT Service Desk requests

These are just a small sample of the sorts of processes that can be automated
through RPA. Any comprehensive review of the processes carried out in a
back office environment can identify a large number of candidates for automation. Some examples of where RPA has demonstrated particular benefits
include:
• O2 replacing 45 offshore employees, costing a total of $1.35m a year, with
ten software robots, costing $100,000. Example processes included the
provisioning of new SIM cards. The telecoms firm then spent its savings of
$1.25m on hiring 12 new people to do more innovative work locally at its
headquarters.
• Barclays Bank have seen a £175 million per annum reduction in bad debt
provision and over 120 Full Time Equivalents (FTE) saved. Example processes include:
–– Automated Fraudulent Account Closure Process—Rapid closure of
compromised accounts
–– Automated Branch Risk Monitoring Process—Collation and monitoring of branch network operational risk indicators
–– Personal Loan Application Opening—Automation of processes for
new loan applications.
• Co-operative Banking Group has automated over 130 processes with
robotic automation including complex CHAPs processing, VISA
­charge-­back processing and many back office processes to support sales and
general administration.

AI and Robotic Process Automation

61

As well as delivering cost savings, RPA is having a huge impact on the way
companies are organising their resources: shared service centres are ripe for
large-scale automation, and outsourced processes are being brought back in-­
house (on-shore) because the costs and risks are much more favourable after
automation. (This, by the way, is creating a huge threat to the viability of
Business Process Outsourcing (BPO) providers.
In implementing RPA, there are a number of aspects that need to be considered. Firstly, there is a need to understand whether the robots will be running ‘assisted’ or ‘unassisted’. An assisted robot generally works on parts of
processes and will be triggered to run by a human. In a contact centre, for
example, a customer service agent could take a call from a customer who
wishes to change their address. Once the call is complete, the agent can trigger
the robot to carry out the changes, which may be required across a number of
different systems. Meanwhile the human agent can get on with taking another
call. Unassisted robots work autonomously, triggered by a specific schedule
(e.g. every Monday morning at 8) or an alert (a backlog queue is over a certain
threshold). They will generally cover whole processes and are therefore more
efficient than unassisted robots. Different RPA software packages are suited to
these two different scenarios in different ways.
Once the type of robot has been selected, the candidate processes for automation can be considered. There are a number of characteristics that make for
a good candidate process:
• Rules-based, predictable, replicable—the process needs to be mapped and
configured in the RPA software; therefore, it has to be definable down to
key-stroke level.
• High volume, scalable—in most cases a process that is high volume (e.g.
happens many time a day) will be preferable because it will deliver a better
return on the investment.
• Relies on multiple systems—RPA comes into its own when the process has
to access a number of systems, as this is where people are generally employed
to integrate and move data around between them.
• Where poor quality results in high risk or cost—the exception to the high-­
volume criteria is where a low-volume process might have a large risk associated with it, such as in making payments, and where compliance and
accuracy are the main concerns.
These criteria provide plenty of opportunities for RPA in most large businesses. But they do have some limitations. The reason that RPA software is of
particular interest to exponents of AI is that, although the RPA software is

62

4 Associated Technologies

really clever in the way that it manages processes, the robots are effectively
‘dumb’; they will do exactly what they are told to do, with unwavering compliance. In many situations that is a good thing, but there are situations where
there is ambiguity, either in the information coming in, or in the need to
make a judgement. This is where AI comes to the fore.
One of the biggest constraints of automating processes through RPA is that
the robots require structured data as an input. This could be, for example, a
spreadsheet, a web form or a database. The robot needs to know precisely
where the required data is, and if it’s not in the expected place then the process
will come abruptly to a halt. Artificial Intelligence, and particularly the Search
capability, provides the ability to have unstructured, or semi-structured,
source data transformed into structured data that the robots can then
process.
Examples of semi-structured data would include invoices or remittance
advices—the information on the document is generally the same (supplier
name, date, address, value, VAT amount, etc.) but can vary considerably in
format and position on the page. As described in the previous chapter, AI
Search is able to extract the meta-data from the document and paste it into
the system-of-record even though every version of it may look slightly different. Once in the main system, the robots can use the data for subsequent
automated processing.
The robots can even use the AI output as trigger for them to run. For
example, a legal contract can be considered a semi-structured document (it
has some common information such as name of parties, termination date,
limits of liability, etc.). An AI Search capability can extract this meta-data for
all of a business’s contracts so that they can manage their total risk portfolio.
RPA robots could be triggered if there was a change in regulation and all contracts of a specific type (e.g. all those under English and Welsh Law) needed
to be updated.
Another area where RPA falls down is where judgement is required as part
of the process. For example, in processing a loan application, much of the
initial stage (such as taking the information from the applicant’s web form
submission and populating the CRM system, the loan system and carrying
out a credit check) can be automated through RPA. But at some point there
needs to be a decision made whether to approve the loan or not. If the decision is relatively simple this can still be handled by RPA—it would be a case
of applying scores to specific criteria with certain weightings, and then
­checking whether the total score was below or above a threshold. But for more
complex decisions, or where it seems like ‘judgement’ is required, then the AI
prediction capability can be used. (This could either be through a Cognitive
Reasoning engine or using a Machine Learning approach.)

AI and Robotic Process Automation

63

Thus, using a combination of RPA and AI allows many processes to be
automated end to end, which inherently provides greater efficiency benefits
than would partly automated ones.
Conversely, RPA can also help AI automation efforts. As mentioned above,
robots are very good at extracting and collating data from many different
sources. Therefore, RPA can be used as a ‘data supplier’ for AI systems. This
could also include manipulation of the data (e.g. remapping fields) as well as
identifying any unusable (dirty) data.
So far I have mainly focused on automating business processes, but RPA
can also be used to automate IT processes as well. The concept for automating
IT is exactly the same as for business processes: replace human staff doing
rules-based work with software agents. Many of the tasks carried out by an IT
Service Desk can be automated; common examples include password resets
and provisioning of additional software on a user’s desktop. One user of RPA
reported a reduction from 6 minutes to 50 seconds for the average incident
execution time for their Service Desk.
RPA can also work autonomously on infrastructure components, triggered
by system alerts: a robot could, for example, reboot a server as soon as it
received an alert that the server was no longer responding. Just like business
process automation, the robots can access almost any other system, and there
will be no disruption or changes required to those underlying systems. RPA in
this scenario can be considered a ‘meta-manager’ that sits across all the monitoring and management systems.
For IT automation, AI can enhance the capabilities of RPA. By using
Optimisation and Prediction capabilities they can be trained from run-books and
other sources, and will continue to learn from ‘watching’ human engineers. The
AI systems can also be used to proactively monitor the IT environment and its
current state in order to identify trends as well as any changes to the environment
(a new virtual server, for example) and then adjust their plans accordingly.
Companies that have implemented a combination of RPA and AI solutions
have seen, for example, 56% of incidents being resolved without any human
intervention, and a 60% reduction in the meantime to resolution.
So, RPA provides a useful complementary technology to AI. In addition to
AI enabling more processes, and more of the process, to be automated, RPA also
aids the data collation efforts that AI often needs. The two technologies also
work well together when there are larger transformational objectives in mind.
Enabling a self-service capability in a business can be a good way to improve
customer service at the same time as reducing costs. This could be achieved
through a combination of AI systems managing the front-end customer engagement (e.g. through chatbots) and RPA managing the back-end processes.

64

4

Associated Technologies

AI and Robotics
One of the first practical applications of AI was in a physical robot called
SHAKEY, which was designed and built by Stanford Research Institute between
1966 and 1972. SHAKEY used computer vision and NLP to receive instructions which it could then break down into discrete tasks and move around a
room to complete the objective it had been set. Although we would now see it
as very rudimentary, it was, at the time, at the forefront of AI research.
Today we have robot vacuum cleaners costing a few hundred pounds that
can clean your house autonomously, and, although they can’t respond to voice
commands, this would be relatively easy to implement. Other examples of
physical robots that use AI are
• Autonomous vehicles—driverless cars and trucks use AI to interpret all the
information coming in from the vehicle’s sensors (such as using computer
vision to interpret the incoming LIDAR data, which is like RADAR but for
light) and then plan the appropriate actions to take. All of this needs to be
done in real time to ensure that the vehicle reacts quickly enough.
• Manufacturing robots—modern robots are much safer and easier to train
because they have AI embedded in them. Baxter, a robot designed by
Rodney Brooks’s ReThink Robotics, can work on a production line without
a cage because it has the ability to immediately stop if it is going to hit someone or something. It can be trained by simply moving the arms and body in
the required series of actions, rather than having to program each one.
• Care-bots—the use of robots to supplement or replace the human care of
sick or vulnerable people is a controversial one, but they can have some positive benefits. There are robots that help care for the elderly, either by providing ‘company’ (through Speech Recognition and NLU) or support through,
for example, helping them remember things (Optimisation). Other
AI-powered robots that are used in the medical field include the use of telepresence robots—these are mobile pods that can travel around hospitals but
are connected via video and sound to human doctors at a remote location.
• Service robots—Some retailers are starting to use mobile robots to greet
and serve customers. As with the hospital telepresence robots described
above, these are mobile units with computer vision, speech recognition,
NLU and optimisation capabilities built in. As well as being able to serve
customers in a shop, different versions of the robots can take the role of
waiters in restaurants or concierges in hotels. There are also interesting
examples where robots are learning skills through reading material on the

AI and the IoT

65

internet—there is one case where a robot has learnt to make pancakes by
reading articles on WikiHow.
• Swarm robots—these are a specific field of robotics where many small
machines work collaboratively together. They rely heavily on the AI
Optimisation capability, constantly evaluating the next best move for each
of the swarm robots so that they can achieve the shared goal. Generally
ground-based, but can be aerial or water-based, they are usually used in
environments that are difficult for humans to work in, such as disaster rescue missions or, more controversially, in warfare (think autonomous
armies). Autonomous vehicles will also be able to exploit swarm
intelligence.
AI is also being used in human-like ways to help robots learn. Researchers
have developed an approach where a humanoid robot learns to stand itself up
by ‘imagining’ what it is like to stand up. Essentially it will run a series of
simulations using a DNN, and then use a second system to analyse the feedback from the various sensors as it actually tries to stand up.
Cognitive Robotics, therefore, can be considered the physical embodiment
of AI. Using input data from many different types of sensors the robot uses a
combination of the Speech Recognition, Image Recognition, NLU and
Optimisation capabilities to determine the most appropriate response or
action. And, because it is AI, the system can self-learn, becoming more effective the more it is used.
Of course, physical cognitive robots also stir in us the fear that they will eventually be able to ‘take over’. Software-based AI is not going to be a risk to the
human race if it is stuck on a server; we can always pull the plug out. But physical
robots could potentially have the ability to overpower us, especially if they are
able to build better versions of themselves. There are already examples of cognitive robots being able to do this, but I refer you back to my earlier comments in
the section on AI Understanding about the distinction between Artificial Narrow
Intelligence and Artificial General Intelligence. There is, of course, an existential
risk there; it’s just that it won’t need to concern us for a very long time.

AI and the IoT
The Internet of Things, or IoT, refers to simple physical devices that are connected to the internet. IoT devices include webcams, smart lights, smart thermostats, wearables and environmental sensors. Many people believe that IoT
is, along with AI, one of the major technology trends of the decade.

66

4

Associated Technologies

There are billions of IoT devices in the world today, each one generating
data or reacting to data, effectively creating and consuming big data on a
grand scale, which is why AI has such a symbiotic relationship with the IoT.
IoT devices are being used in business today to
• manage preventative maintenance programs by analysing data from sensors
embedded into the assets (e.g. escalators, elevators and lighting)
• manage supply chains by monitoring movement of products
• conserve energy and water usage of machines by analysing and predicting
demand (smart meters, which monitor energy usage every 15 minutes, are
already common in many homes)
• improve customer experience by providing personalised content based on
IoT data
• increase crop yields by analysing field sensors to provide precise feeding
and watering programs
• alleviate parking problems by matching empty spaces with cars and their
drivers
• keep us fitter by monitoring and analysing our steps and exercise patterns
through wearable devices
One of the headline successes of ‘IoT plus AI’ was when Google announced
that they had been able to reduce the energy used for cooling in one of their
data centres by 40%. By using DeepMind’s AI technology, they were able to
analyse data from many different types of sensors (such as thermometers,
server fan speeds, and even whether windows were open or not), precisely
predicting demand so that they could instruct the various machines to work
at optimum levels.
Another area where IoT and AI are key enablers is in the development of
Smart Cities. IoT devices are used to track and extract data for transport,
waste management, law enforcement and energy usage. This data is analysed
and turned into useful information by AI Clustering, Optimisation and
Prediction systems, and then made available to the relevant authorities, the
city’s citizens and other machines. For example, sensors in street lights and
urban furniture are able to measure footfall, noise levels and air pollution—
this data is then used to prioritise the delivery of other services.
The biggest challenge to the proliferation of the IoT is the poor security
that has been associated with it. Many devices do not have the ability to
change passwords, and have only basic authentication. There have been recent
cases of baby monitors allowing strangers to monitor their camera feed,
internet-­connected cars’ entertainment systems and central locking systems

AI and Crowd Sourcing

67

being taken over by hackers, and, probably most worrying of all, medical
devices being compromised so that a hacker could send fatal doses of medicines to drug infusion pumps. The good news is that these high-profile cases
has meant that IoT security is very high on many technology company’s agendas and there is now plenty of activity to fix all the issues. But if you are thinking of implementing an IoT strategy, do keep security at the top of your
considerations.
As more and more IoT devices are used (it is estimated that there will be 50
billion devices by 2020), there will be more need for AI to make sense of all
the data that is generated by them. The analysis that can be carried out can
then be actioned by other IoT devices such as actuators and lights. As in the
Smart Cities example, the real value will be the sharing and collaboration of
this data across organisations, people and other machines.

AI and Crowd Sourcing
Sometimes AI simply isn’t up to the job. Sometimes you will need to pull
humans into the loop to help complete the process.
A classic example of when this sort of situation can arise is when the source
documents for an automated process are handwritten. We already know that AI
is very capable of extracting structured data out of unstructured documents but
this really applies only when the unstructured documents are in an electronic
format in the first place. Optical Character Recognition is able to take typed
documents such as PDFs and convert them into an electronic format, but if the
source is handwritten the challenge becomes a whole lot more difficult.
One solution to this is to use a Crowd Sourcing service. Crowd sourcing is
where large numbers of people are engaged to carry out small parts of a process (‘micro-tasks’). The engagement model is usually via the Internet, with
each person paid a specific rate for every time they complete one of the
micro-tasks.
In the example above, one document can be split into many different tasks
so that one person is sent the First Name to read, another is sent the Last
Name and a third the Social Security Number. Each person will look at the
image of the handwriting and respond with the text that it represents. Because
each person only sees a small part of the information, confidentiality is maintained. To increase accuracy, a single image can be sent to a number of people
so that only the most common answer is selected. Specialised software is used
to manage the interface between the customer and the crowd, including the
splitting of the documents into smaller parts.

68

4

Associated Technologies

A second role for crowd sourcing in the AI space is to provide additional
capability when the AI is not confident enough of its answers. This might be
because the problem is too complex, or it hasn’t seen that particular problem
before (Fig. 4.1).
In this case, the AI will send the problem to a person to answer. For example, some AI systems are used to moderate offensive images on social media
sites. If the AI is unsure whether a particular image is offensive or not, it will
ask a human for their opinion. This approach is usually called ‘Human In The
Loop’, or HITL.
An enhanced version of HITL is where the input from the human is used
to actively train the AI system. So, in the offensive image example, the human’s
choice (offensive or inoffensive) would be sent back to the AI so that it can
improve its learning and perform better in the future (Fig. 4.2).
A third use of crowd sourcing with AI is in the development of the training
data. As I discussed in the Technology Overview section of Chap. 1, supervised learning approaches require data sets that are appropriately tagged (dog
pictures tagged as dogs, cat pictures tagged as cats, etc.). Because of the large
data sets that are required, the job of tagging all these data points is very laborious. Crowd sourcing can be used to satisfy that requirement. Google itself
pays for tens of millions of man-hours collecting and labelling data that they
feed into their AI algorithms (Fig. 4.3).
The most popular general crowd sourcing site is Mechanical Turk (owned
by Amazon) but others, such as Crowd Flower, that focus on supporting the
AI community, are also available. Some also have an Impact Sourcing angle as

Artificial
Intelligence

Confident?

No

Human
Intervention

Fig. 4.1 Human in the loop

Yes

Output

AI and Crowd Sourcing

Artificial
Intelligence

Confident?

Yes

69

Output

No

Active Learning

Human
Intervention

Fig. 4.2 Training through a human in the loop

Tagging By
Humans

Training

Artificial
Intelligence

Fig. 4.3 Crowd-sourced data training

well, where the people doing the work are disadvantaged in some way and this
work gives them new opportunities.
So, despite the amazing things that AI can do, at some point it will probably still rely on people to do part of the job, whether that is training it or
helping it make decisions.

70

4

Associated Technologies

The Established AI Vendor’s View
This is an extract from an interview with Andrew Anderson, CEO of Celaton,
a leading AI software vendor.
AB What got you into the world of AI in the first place?
AA	The story began in 2002. I sold my previous software company to a much
larger company that pioneered the delivery of software as a service, at that
time they were described as an ‘application service provider’. As the VP of
product development, we went on to acquire, license and develop lots of
software applications that solved many different challenges faced by organisations, all delivered as a service. However, despite access to all this technology we still couldn’t solve the challenge of dealing with unstructured,
varied (human created) data that flowed into organisations every day. It
remained a manual, labour-­intensive task.
I saw an opportunity to create a solution to the challenge and so in
2004 I bought my company back, along with my development team and
set about trying to build an ‘as a service’ platform that could automate the
processing of unstructured, varied data. My optimistic view then hoped
that we would have a solution built within a year. The reality was that it
took over six years to create a platform that had the ability to learn, and
therefore enable us to describe what we had created as ‘artificially intelligent’. In fact, it wasn’t until we demonstrated it to analysts that they told
us what we had created.
We spent a few years (2011–2013) trying different approaches until we
realised that the greatest value we deliver was to large, ambitious organisations who deal with demanding consumers.
AB Why do you think AI is being so talked about right now?
AA	AI is being talked about more than ever because it promises so much.
Whether those promises are realised is yet to be seen. It’s like a wonder
drug that has huge potential but the enthusiasm needs to be tempered
with a little reality.
What is emerging now are specialist AIs that are good at specific tasks
rather than general-purpose AIs that are good at everything. Whilst there
are many case studies where specific or narrow AI is delivering real benefits this is being mixed up with noise about general AI.
There’s sanity in numbers and it takes time and effort to educate and
convince enough organisations that they should try something but each
new client means a new case study and each case study leads to new customers. These are the realities behind the hype, but the media often prefer
to hear the more exciting stories of potential rather than reality.

The Established AI Vendor’s View

Automation has been around since the wheel was invented and the
world has continued to invent and innovate and that’s why we see this
continuous improvement in technology. The increased power of technology enables it to achieve outcomes that were previously out of reach.
Automation of any kind is also accelerating and with the emerging of
AI it must seem that everything can be automated. What was previously
considered the domain of human effort is now being eroded and this creates negative media.
In summary, I think it’s being talked about more for its potential than
its reality, although the reality (i.e. the case studies) are helping to fuel the
future potential because people can actually see the future emerging.
AB What value do your customers get out of your software?
AA	There’s a different answer depending on the customer challenge. By
understanding what customers are saying, then reacting and responding
to them enables our customers to deliver better service faster, with fewer
people. In summary, it enables them to achieve competitive advantage,
compliance and improve their financial performance.
AB	What do customers or potential customers need to focus on if they are
going to get the maximum value from AI?
AA	I think the key is not to focus on the technology but to understand the
problem you want to solve. Many people (within organisations) have
trouble understanding their problem and therefore they struggle to identify the technology that might help to solve it.
This is where experienced consultancy is important. Understand the
problem, select and apply the most appropriate technology and then share
the story.
AI is often considered to be the solution to all problems like a magic
drug. The reality is there is no magic drug but there are lots of drugs that
solve different problems. The key is to talk with someone who can help to
understand the problem and prescribe the right ‘drug’.
AB Do you think the current hype is sustainable?
AA	I don’t think that we are witnessing anything different with this hype. The
industry does tend to get ahead of itself because hype serves it well, creating awareness that helps to capture the interest of customers. There’s sanity in numbers.
The difference this time is that technology innovation is moving
quicker and it’s not so long after the thought has been considered that the
invention can be demonstrated.

71

72

4 Associated Technologies

Regardless of how fast things move, it’s the human that is the limiting
factor. The bigger the problem that it solves, the more likely it is to be
adopted by humans and therefore be successful.
AB How do you see the market developing over the next 5 years or so?
AA	Hype will lead to reality, but by solving problems that really exist. There
are a couple of areas that are particularly relevant:
• 	Natural interfaces. We’re seeing this already with the likes of Amazon
Echo and IoT appliances. There is some way to go but they will become
more natural and to such an extent that humans will have to think less
about how they communicate with technology.
•	Broader AI. We’re currently seeing specialist AIs that are good at doing
specific things – these technologies will consolidate into broader solutions that are able to do lots of things well because they’re made up of
lots of vertical AIs.
And I particularly see some of the greatest and most profound use of
AI in medicine, energy and, subject to world events, in the military too.

5
AI in Action

Introduction
Up to this point in the book I have been pretty theoretical, explaining the capabilities of AI and the types of technologies it requires to work. In this chapter I
change focus to look at how AI is actually being used in businesses—real use
cases of AI adding value and changing the way companies do business.
I have split the chapter into different themes which look at specific aspects
of business and the ways that AI adds value: enhancing customer service,
optimising processes and generating insights. There are overlaps between each
of these areas (there are, for example, insights you can generate about your
customers) but it gives a general, useful framework to describe how AI is
being used in business.
In the chapter following this I describe the different types of benefits that
should be considered in an AI Business Case. These can be broadly mapped
against the themes:
• Enhancing customer service leads to revenue generation and customer
satisfaction
• Optimising processes leads to cost reduction, cost avoidance and
compliance
• Generating insights leads to risk mitigation, loss mitigation and revenue
leakage mitigation
Each of these themes will also draw upon different capabilities from the AI
Framework. All will use at least one of these capabilities, whilst others will
exploit a few.
© The Author(s) 2018
A. Burgess, The Executive Guide to Artificial Intelligence,
https://doi.org/10.1007/978-3-319-63820-1_5

73

74

5 AI in Action

I have tried to take examples from a wide range of industries, and have
deliberately not grouped the case studies into specific sectors. This is so you
are not tempted to skip over sectors that are not immediately relevant to
you—I am a firm believer that very different industry sectors can still learn
from each other. Even if you work in the retail sector, for example, you may
still get a spark of an idea from something in the utilities sector.

How AI Is Enhancing Customer Service
The ‘front office’ is one of the more active areas when it comes to implementing AI. This is because there is generally plenty of customer data available to
work off, but it is also down to the proliferation of chatbots.
Chatbots come in all shapes and sizes, which is a rather polite way of saying that there are really good chatbots but also very bad ones. Chatbots, which
aim to have a natural conversation with a customer through a typed interface,
use NLU as their key AI capability. This means that, in theory, a customer
who wants to change their address on the system (this is their ‘intent’ in
AI-speak) can ask the chatbot in any way that they want (‘I’m moving to a
new house’, ‘I’ve got a new address’, ‘I’m not living at the same place any
more’, ‘My postcode has changed’, etc.) and it will still understand the underlying intent.
In reality, this is a big challenge for NLU, and most chatbots use a ‘dictionary’ of equivalent phrases to refer to (e.g. if the input is ‘I’ve got a new address’
or ‘I’m moving to a new house’, then the intent is ‘Customer wants to change
their address’). This, of course adds to the complexity of designing the chatbot, as every alternative phrase has to be identified and entered.
Some chatbots make extensive use of multiple-choice questions. So, instead
of relying on understanding what the customer has typed, the chatbot will ask
a question with a limited number of answers (e.g. Yes/No, Ask for a Balance/
Make a Payment/Change Your Address, etc.). This can be more efficient and
accurate but doesn’t make for a natural conversation.
Another challenge for chatbots is how they move through the conversation.
The simplest and most common approach is a decision tree, where each question will branch off to a new question depending on the answer. If the process
being replicated is complex, this can lead to an over-complicated and burdensome decision tree. A better approach here is to have a cognitive reasoning
engine do all of the ‘thinking’ whilst the chatbot gets on with the conversation. This provides much more flexibility in how the chatbot deals with the
flow of the conversation. Cognitive reasoning systems were described in the
Optimisation section of Chap. 3.

How AI Is Enhancing Customer Service

75

The purest AI approach to chatbots is to train them on thousands of
human-to-human chat conversations, where each interaction has been tagged
with the intent and whether it was productive or not. These chatbot systems
effectively learn the knowledge map (a.k.a. ontology) through those historical
interactions. The challenge with these systems is that there needs to be plenty
of training data available, and they tend to very expensive to implement.
With chatbots, it is very much ‘horses for courses’. A simple, ‘free’ chatbot
system will be fine for very simple and non-critical interactions with customers, but because of the reputational risk involved, implementing chatbots is
worth doing only if you do it right. The better the chatbot system, the better
it will cope with the challenge.
In 2017 Royal Bank of Scotland introduced a chatbot to help answer a
limited number of questions from customers. The chatbot, called ‘Luvo’, uses
IBM Watson’s Conversation capability to interact with customers who are
using the bank’s website or app. It was trialled for nearly a year with internal
staff who manage relationships with SMEs before being released to a small
number of external customers. At the time of writing, it is able to answer just
ten defined questions, such as ‘I’ve lost my bank card’, ‘I’ve locked my PIN’
and ‘I’d like to order a card-reader’.
Luvo demonstrates the cautious approach that enterprises are taking to
implementing chatbots. Over time though, as the system learns from each
interaction it has, the bank will allow it to handle more complex questions,
build in greater personalisation and use predictive analytics to identify issues
for customers and then to recommend the most appropriate action. The key
objective is to free up more time for the human customer service agents to
handle the trickier questions that customers may have.
In a similar two-step implementation, SEB, one of Sweden’s largest banks,
deployed IPsoft’s Amelia software first in their internal IT Service Desk, and
then to their one million customers. During the first three weeks of the trial,
the chatbot-cum-avatar helped to answer the staff’s IT issues, resolving around
50% of the 4000 queries without human interaction. During 2017, it was
rolled out to help manage interactions with the bank’s customers (and named
‘Aida’). Three processes were initially chosen as candidates: providing information on how to become a customer, ordering an electronic ID and explaining how to do cross-border payments.
Although Amelia learns from historical interactions and uses sentiment analysis, it is also able to follow defined workflow paths to ensure compliance with
banking regulations. Whereas most chatbots can be described as probabilistic,
providing likely results, Amelia can be considered a deterministic system because,
once it has worked out the intent, it can, where possible, carry out the required

76

5 AI in Action

actions on the enterprise systems (in a similar way that RPA tools would). IPsoft
is a US-based company, and this was their first non-­English deployment of their
solution.
From an infrastructure point of view, SEB decided to install the Amelia technology on their own servers rather than as a cloud solution. This was due to
concerns over the compliance and legal concerns that a cloud deployment raises.
For IPsoft’s financial services customers, this is the most common approach.
The deployment of Amelia is very much aligned with the bank’s overall
strategy, which includes the line “SEB will focus on providing a leading customer experience, investing in digital interfaces and automated processes”.
The US flower delivery service 1-800-Flowers deployed a simple chatbot
that allowed customers to place orders through Facebook Messenger. It is a
linear, decision tree system with an NLU chat interface that took around three
months to develop and test. This makes it rather constrained in what it can
do, but additional functionality and complexity is being rolled out. Following
the initial two months of operation, 70% of the orders being placed through
the chatbot were from new customers, and were dominated by ‘millennials’
who tend to already be heavy users of Facebook Messenger.
As well as taking orders, the chatbot can direct a customer to a human
agent, of which there may be up to 3,500 at any one time. They have also
implemented an integration with Amazon Alexa, and a concierge service that
is powered by IBM Watson. Together, these digital initiatives have attracted
‘tens of thousands’ of new customers to the brand. They also provide up-to-­
the-minute behaviour data for the company which can then influence real-­
time marketing activity such as promotions.
Another company that has used IBM Watson to enhance customer engagement is Staples, the stationery retailer. They have implemented a range of
different ways for people to buy their products as easily as possible, including
email, Slack, mobile app and (honestly) a big red button. The button is similar to Amazon’s Alexa in that it can understand voice commands (although it
does need a physical press to activate and deactivate it). The mobile app can
also understand voice commands as well as being able to identify products
from photographs taken. All these channels make the buying process as
­frictionless as possible for the customer, and therefore has a direct and positive
impact on revenue for the retailer.
As well as chatbots, recommendation engines are another common AI technology that is used to enhance customer service (and, of course, drive revenue).
Amazon and Netflix have the most well-known recommendation engines, and
these are deeply embedded in the normal workflow of how the customer engages
with the companies. All the required data—the individual customer’s buying

How AI Is Enhancing Customer Service

77

and browsing behaviours and the historical data of all the other customers’
behaviours—are available through those normal interactions; that is, there is no
additional work that the customer has to do to enrich the data set.
In some cases, the recommendation engines will require additional information from the customer and/or the business in order to work effectively. North
Face, a clothing retailer, has implemented a recommendation engine for their
customers who want to buy jackets. Based on IBM Watson this solution, called
XPS, uses a chatbot interface to ask a series of refining questions so that it can
match the customers’ requirements with the product line. According to North
Face, 60% of the users clicked through to the recommended product.
Another clothing retailer, Stitch Fix, uses a slightly different approach by
deliberately including humans in the loop. Its business model involves recommending new clothes to their customers based on a selection of information
and data that the customer provides (measurements, style survey results,
Pinterest boards, etc.). All this structured and unstructured data is digested,
interpreted and collated by the AI solution, which sends the summary, plus
anything that is more nuanced (such as free-form notes written by the customer) to one of the company’s 2,800 work-from-home specialist human
agents, who then select five pieces of clothing for the customer to try.
This is a good example of where the AI is augmenting the skills and experience of the human staff, making them better at their jobs as well as being
more efficient. Having humans in the loop (HITL is the acronym for this)
also makes experimenting easier, as any errors can quickly be corrected by the
staff. To test for bias, the system varies the amount and type of data that it
shows to a stylist—it can then determine how much influence a particular
feature, say a picture of the customer or their address, can make on the stylist’s
decisions. On top of all this, the data that they gather about all of their customers can also be used to predict (and influence?) general fashion trends.
Other customer service AI solutions don’t rely on chatbots or recommendation engines to provide benefits. Clydesdale and Yorkshire Banking
Group (CYBG) is a medium-size bank in the United Kingdom, having to
compete with the ‘Big Four’ of Barclays, HSBC, Lloyds and RBS. Their digital strategy includes a new current account, savings account and app package,
called ‘B’. It uses AI to help manage the customer’s money: it allows you to
open account, and, once opened, will learn the patterns of usage so that it can
predict if you might run out of funds in your accounts, and suggest ways to
avoid unnecessary bank charges. The bank claims that an account can be
opened in 11 minutes. Clearly, customers avoiding bank charges will result in
lower revenue, but the hope is that it will attract enough new customers that
this revenue loss will easily be outweighed by the benefits of having the additional funds.

78

5 AI in Action

Virgin Trains ‘delay/repay’ process has been automated through the application of AI. After implementing Celaton’s inSTREAM AI software to categorise inbound emails (see next section) the train operating company used the
same software to provide a human-free interface for customers to automatically claim refunds for delayed trains.
The Asian life insurance provider AIA has implemented a wide range of AI
initiatives that all impact how they engage with their customers, including
insight generation from prospective customers, enhanced assessment of customer needs, 24×7 chatbots for online enquiries, inbound call handling by
NLU-based systems, enhanced compliance of sales, personalised pricing,
dynamic underwriting and an augmented advice and recommendation engine.
Some companies are building AI into the core of their customer applications. Under Armour, a sports clothing company that has a portfolio of fitness apps, uses AI in one of those apps to provide training programs and
recommendations to the users. The AI takes data from a variety of sources,
including the users’ other apps, nutritional databases, physiological data,
behavioural data and results from other users with similar profiles and objectives. It then provides personalised nutrition and training advice that also
takes into account the time of day and the weather.
Other examples of where AI can enhance customer service are: optimising
the pricing of time-critical products and services such as event ticketing; optimising the scheduling of real-time services such as delivery and departure
times; creating personalised loyalty programs, promotional offers and financial products; identifying candidate patients for drug trials and predicting
health prognoses and recommending treatments for patients.
There is clearly a balance between using AI to serve customers better, and using
AI to gather information about customers. In some of the cases I have described,
the same AI capability will do both. As I have reiterated a number of times, for
anyone looking to implement a ‘front office’ AI solution, it is important to first
understand your objectives. The danger of trying to ask your AI to do too much
is that you compromise on one objective or the other. I discuss approaches and
examples of how to generate customer insights later in this chapter.

How AI Is Optimising Processes
If the previous section was all about enhancing the front office, then this section focuses on what happens behind the scenes in the back office operations.
The benefits here are centred around, although not exclusively, reducing cost
(directly or through avoidance) and improving compliance.

How AI Is Optimising Processes

79

One of the key capabilities exploited in the back office is being able to
transform unstructured data into structured data, that is, from the Image
Recognition, Voice Recognition and Search capabilities.
Tesco have implemented a number of AI-powered solutions to improve the
productivity of their stores. They are using image recognition systems to identify empty shelves on stores, called ‘gap scanning’. They are experimenting
with physical robots that travel down the aisles during quiet periods to film
the shelves so that they can measure stock availability and inform staff to
replenish them where necessary. Not only does this save the time of the staff
(who would normally have to use a diagram on a piece of card as the template)
but it also reduces revenue loss due to stock unavailability.
For their home delivery service Tesco have implemented optimisation systems that minimise the distance the picker walks around the store collecting
the goods, and also a system that maximises the productivity of the delivery
vans through effective route planning and scheduling. Interestingly, in a meta-­
example, they have used AI to help place the many products into the necessary taxonomy that AI systems use for classification.
AIA, the Asian insurance provider I mentioned in the previous section,
have also implemented initiatives in their servicing and claim functions,
including service augmentation, investment management, medical second
opinions, risk management, fraud management, claims adjudication and
health and wellness advice to their customers. AIA have been very clear that
transforming their business requires end-to-end transformation with AI at
the core.
Insurance claims processing is a prime candidate for enhancement through
AI and RPA. Davies Group, a UK-based insurance firm that provides a claims
processing service to their own insurance customers, has automated the input
of unstructured and semi-structured data (incoming claims, correspondence,
complaints, underwriters’ reports, cheques and all other documents relating
to insurance claims) so that it is directed into the right systems and queues.
Using Celaton’s inSTREAM solution, a team of four people process around
3,000 claims documents per day, 25% of which are paper. The tool processes
the scanned and electronic documents automatically, identifies claim information and other meta-data, and pastes the output into databases and document stores ready for processing by the claims handlers and systems (which
could, of course, be either humans or software robots). It also adds service
meta-data so the performance of the process can be measured end to end.
Some documents can be processed without any human intervention, and others need a glance from the human team to validate the AI’s decisions or fill in
missing details.

80

5

AI in Action

AI can also be used in the claims process to identify fraud (by looking for
anomalous behaviours), to help agents decide whether to accept a claim or
not (using cognitive reasoning engines) and by using image recognition to
assess damage. Ageas, a Belgian non-life insurance firm, has deployed
Tractable, an AI image recognition software, to help it assess motor claims. By
scanning images of damage to cars, the software is able to assess the level of
damage, from repairable to being a write-off. Through the process, it also
helps identify fraudulent claims.
Image recognition has also been used in other industry sectors. Axon,
which used to be known as Taser, the manufacturer of police equipment, are
now using AI to tag the thousands of hours of video that is recorded from the
body cameras that they manufacture. The video recordings were previously
available to police forces on subscription but, due to the sheer volume of
recordings, they were rarely used as evidence. Now that the film is being
tagged, the crime agencies will be able to automatically redact faces to protect
privacy, extract relevant information, identify objects and detect emotions on
people’s faces. The time saved, from having to manually write reports and tag
videos, will be huge, but the ability to make better use of the available evidence will likely have a bigger impact. For Axon, it will encourage more police
forces to buy their body cams.
Data from video feeds has been used as the core of the business model of
Nexar, an Israeli-based technology company. They are giving away for free
their dash cam app which helps drivers anticipate accidents and problems in
the road ahead. By using the data provided by the apps from their user base
(which is mandatory if you want to use the app) the system predicts the best
course of action, for example to brake hard if there is an accident up ahead.
This model relies heavily on having large amounts of data available, hence the
up-front giveaway. Rather controversially, the company can use the data for
‘any purpose’, including selling it to insurance companies and giving it to the
government. Whether this is a sustainable business model remains to be seen,
but it does demonstrate how business models are transforming to focus more
on the data than the service being provided to the customers.
A particularly worthwhile use of image recognition is in the medical
field. Probably the most widely reported has been the use of IBM’s Watson
platform to analyse medical images, such as X-rays and MRI scans, so that
it can help identify and diagnose cancer cells. The real benefit of this system is that it can cross-reference what it sees with the patient’s (unstructured) medical records and thousands of other images that it has learnt
from. This approach is already spotting characteristics of tumours that
medical staff hadn’t seen—they are finding that peripheral aspects of the

How AI Is Optimising Processes

81

tumour (e.g. those distributed around, rather than in, the tumour) have a
much greater influence on predicting malignancy than at first thought.
These systems are currently working in two hospitals: UC San Diego
Health and Baptist Health South Florida. (See the section on Generating
Insights for more examples of how AI is helping fight cancer.)
Deutsche Bank are using the speech recognition technology to listen to
recordings of their staff dealing with clients, in order to improve efficiency,
but, more importantly, ensure compliance to regulations. These conversations
are transcribed by the AI and can be searched for specific contextual terms.
This used to be the job of the bank’s auditors, and would require listening to
hours of taped recordings, but the AI system can do the same thing in minutes. The bank is also using other AI capabilities to help it identify potential
customers based on the large amount of public information that is available
on people.
The legal sector is currently going through a big transition period as it starts
to adopt new technologies, including AI. Inherently risk averse, and never an
industry to change quickly, the legal sector is starting to feel the benefit of AI
particularly through implementing Search and NLU technologies to help
make sense of contracts.
Legal contracts can be described as semi-structured documents—although
they follow some general rules (they usually include parties to the contract,
start and end dates, a value, termination clauses, etc.), they are verbose and
variable. There are now a number of software solutions on the market which,
together, work across the full contract lifecycle—some are able to search document repositories to identify where contracts are; others can ‘read through’ the
contracts, identify specific clauses and extract the relevant meta-­data (such as
the termination date and the limits of liability). Some have the capability to
compare the real contracts with a model precedent and show all instances
where there is a material difference between the two. All of this allows large
enterprises and law firms to manage their contracts much more effectively,
especially from a risk point of view.
McKesson, a $179 billion healthcare services and information technology
organisation, uses Seal’s Discovery and Analytics platform to identify all of
their procurement and sourcing contracts across the business (they employ
70,000 people) and store the documents in a specific contract repository.
When in the repository, the contracts could be searched easily and quickly,
saving some staff hours per day. The meta-data from the contract analysis
allowed them to identify potential obligation risks as well as revenue opportunities and savings from otherwise hidden unfavourable payment terms and
renewal terms.

82

5 AI in Action

A Magic Circle law firm, Slaughter & May, adopted Luminance AI software to help manage the hundreds of Merger and Acquisition (M&A) transactions they do a year. This area is a particular challenge because of the complexity
involved (thousands of documents, many jurisdictions) and the intensity of
the work—the law firm was concerned that some of the junior lawyers who
were tasked with managing the M&A ‘data room’ (the contract repository created specifically for the deal) would eventually burn out. Luminance is used to
cluster, sort and rank all of the documents in the data room, with each document assigned an anomaly score to show how it differs from the ideal model.
Whereas in normal due diligence exercises only around 10% of the documents
are analysed, due to the sheer volume, Luminance is able to analyse everything.
It takes about an hour for it to process 34,000 documents. Overall, it halves the
time taken for the whole document review process.
One of the benefits of having the AI systems carry out ‘knowledge based’
works such as this is that it allows lawyers to focus on complex, higher-value
work. Pinsent Masons, a London-based law firm, designed and built in-­
house a system, called TermFrame, which emulates the legal decision-making
workflow. The system provides guidance to lawyers as they work through different types of matters, and connects them to relevant templates, documents
and precedents at the right time. Taking away much of the lower-level thinking that this requires means that the lawyers can spend more time on the
value-add tasks.
As well as calling on Search and NLU capabilities, companies are also using
NLG solutions to optimise some of their processes. Using Arria’s NLG solution, the UK Meteorological Office can provide narratives that explain a
weather system and its development, with each one configured with different
audiences in mind. The Associated Press (AP), a news organisation, uses
Wordsmith to create publishable stories, written in the AP in-house style,
from companies’ earnings data. AP can produce 3,700 quarterly earnings stories which represents a 12-fold increase over its manual efforts.
Artificial intelligence can also be used to enhance the IT department as well
(‘Physician, heal thyself ’). There are a number of AI capabilities that can be
deployed, both in the support function and for infrastructure management.
Generally, AI brings a number of important aspects to the IT function: the AI
systems can be trained from run-books and other sources; they will continue
to learn from ‘watching’ human engineers; they can proactively monitor the
environment and the current state; they can identify trends in system failures
and they are able to cope with the inherent variability of IT issues.
We have already seen how chatbots can support customer engagement, but
they can also be used with staff as well, particularly where there are a large

How AI Is Optimising Processes

83

number of employees, or a large number of queries. Although chatbots can be
deployed on any type of service desk, including HR, they are most commonly
found in IT.
As in the SEB example discussed in the previous section, AI Service Desk
agents are able to receive customer enquiries and provide answers by understanding what a customer is looking for and asking the necessary questions to
clarify the issue. The more advanced systems will, if they cannot help the
customer themselves, raise the issue to a human agent, learning how to solve
the problem itself for future situations.
AI can also be used to manage a company’s IT infrastructure environment.
The AI systems tend to work as meta-managers, sitting above and connecting
all the various monitoring systems for the networks, servers, switches and so
on. They include both proactive and predictive monitoring (using AI) and
execution of the necessary fixes (using software robots). Key performance
indicators such as downtime can be significantly improved using these systems. And, by tackling many of the mundane tasks, the systems free up the IT
engineers to focus on higher-value, innovative work.
TeliaSonera, a European telecoms operator, implemented IPCentre to
help manage its infrastructure of 20 million managed objects, including
12,000 servers, and have subsequently seen cost savings of 30%. A New York-­
based investment firm used the same solution to help it fix failed fixed-­
income securities trades due to system issues. Eighty per cent of the failed
trades are now fixed without human intervention, and the average resolution
and fix time has been reduced by 93% (from 47 minutes to 4 minutes). This
has resulted in a staff reduction of 35%.
Google, a company with many very large data centres, turned to its AI
subsidiary, DeepMind, to try and reduce the cost of powering these facilities.
The AI worked out the most efficient methods of cooling by analysing data
from sensors amongst the server racks, including information on aspects such
as temperatures and pump speeds. From the application of the AI algorithms,
Google was able to reduce cooling energy demand by 40%, and to reduce the
overall energy cost of the initial data centre by 15%. The system now controls
around 120 variables in the data centres including fans, cooling systems as
well as windows.
Another example of where AI can optimise processes includes the ability to
model real-world scenarios on computers, with the most obvious example being
the prediction of long-term weather conditions. But this modelling approach
has been used to test physical robotic systems so that the designers can make
adjustments to their robots virtually without the fear of them falling over and
breaking. Facebook have created a specific version of Minecraft, a virtual-world

84

5 AI in Action

game, for this sort of purpose. AI designers can also use the environment to help
teach their AI algorithms how to navigate and interact with other agents.
In this section, we have seen examples of how AI can optimise the back office
processes of retailers, banks, insurance firms, law firms and telecoms companies,
across a wide range of functions, including claims management, compliance and
IT. Other use cases for AI in optimising processes include: optimising logistics
and distribution of inventory in warehouses and stores; real-time allocation of
resources to manufacturing processes; real-time routing of aircraft, trucks and so
on; optimising the blend and timing of raw materials in refining and optimising
labour staffing and resource allocation in order to reduce bottlenecks.

How AI Is Generating Insights
In the previous two sections I looked at how AI can enhance customer service
and optimise processes. But, in my mind, the biggest benefits from AI are
when it can deliver insight. This is where new sources of value are created from
the data that already exists, enabling better, more consistent and faster decisions to be made. AI can therefore enable a company to mitigate risks, reduce
unnecessary losses and minimise leakage of revenue.
One of the most effective uses of AI to date has been in the identification
of fraudulent activity in financial services. The benefit here is that there is
plenty of data to work from, especially in retail banking. PayPal process $235
billion in payments a year from four billion transactions by its more than 170
million customers. They monitor the customer transactions in real time (i.e.
less than a second) to identify any potentially fraudulent patterns—these ‘features’ (in AI-speak) have been identified from known patterns of previous
fraudulent transactions. According to PayPal, their fraud rate is just 0.32%,
compared to an industry average of 1.32%.
The American insurance provider USAA is similarly using AI to spot identity theft from its customers. It is able to identify patterns in behaviour that
don’t match the norm, even if it is happening for the first time. Axa, another
insurance firm, has used Google’s TensorFlow to predict which of its customers are likely to cause a ‘large-loss’ car accident (one that requires a pay-out
greater than $1,000). Normally up to 10% of Axa’s customers cause a car
accident every year, but about 1% are ‘large-loss’. Knowing which customers
are more likely to be in that 1% means that the company can optimise the
pricing of those policies.
The Axa R&D team initially tried a Decision Tree approach (called Random
Forest) to model the data but were able to achieve only 40% prediction
accuracy. Applying deep learning methods on 70 different input variables,

How AI Is Generating Insights

85

they were able to increase this to 78%. At the time of writing, this project is
still in its proof-of-concept (PoC) phase but Axa are hoping to expand the
scope to include real-time pricing at point of sale and to make the system
more transparent to scrutiny.
Another company that is using AI to optimise pricing in real time is Rue
La La, an online fashion sample sales company who offer extremely limited-­
time discounts on designer clothes and accessories. They have used machine
learning to model historical lost sales in order to determine pricing and predict demand for products that it has never sold before. Interestingly, from the
analysis they discovered that sales do not decrease when prices are increased
for medium and high price point products. They estimated that there was a
revenue increase of the test group of almost 10%.
Otto, a German online retailer, is using AI to minimise the number of
returns they get, which can cost the firm millions of euros a year. Their particular challenge was the fact that they knew that if they got the orders to their
customers within two days then they would be less likely to return them
(because it would be less probable for them to see the same product in another
store at a lower price). But, they also knew that their customers liked to get all
their purchases in one shipment, and, because Otto sources the clothes from
other brands this was not always easy to achieve.
Using a deep learning algorithm, Otto analysed around three billion historical transactions with 20 variables so that it could predict what customers
would likely buy at least a week in advance. Otto claims that it can predict
what will be sold within 30 days to an accuracy of 90%. This means it can
automate much of the purchasing by empowering the system to auto-order
around 200,000 items a month. Their surplus stock has reduced by a fifth and
returns have gone down by more than two million items a year.
Goldman Sachs, the global investment bank, has implemented a range of
automation technologies to both improve decision-making processes and also
to reduce headcount. They started by automating some of the simpler trades
that were done at the bank—in the year 2000 there were 600 equity traders
working in the New York headquarters; at the start of 2017 there were just
two. The average salary of an equity trader in the top 12 banks is $500,000,
so the savings have been significant. The majority of the work is carried out by
automated trading systems that are supported by 200 engineers (one third, or
9,000, of Goldman Sachs employees are computer engineers).
To automate more complex trades such as currency and credit, cleverer
algorithms are required. Where Goldman Sachs has automated currency trading, they have found that one computer engineer can replace four traders.
They are also looking at building a fully automated consumer lending platform that will consolidate credit card balances. This is an innovation that has

86

5 AI in Action

been incubated internally at the bank’s New York office by small ‘bubble’
teams. As Marty Chavez, the company’s deputy chief financial officer, says:
there is plenty of empty office space to use up.
The UK-headquartered bank HSBC has been running five proof-of-­
concepts (PoCs) that will utilise Google’s AI capabilities. Because their 30
million or so customers are engaging online much more, their data store has
increased from 56 petabytes in 2014 to more than 100 petabytes in 2017.
This means they are now able to mine much more value and insight from it.
One of the PoCs is to detect possible money laundering activity. Just as
with PayPal and USAA, they are looking for anomalous patterns that they can
then investigate with the relevant agencies to track down the culprits. As well
as improving detection rates, using the AI software (from Ayasdi) means that
there are less false-positive cases that have to be investigated, which saves the
time of expensive resources. In the case of HSBC, they managed to reduce the
number of investigations by 20% without reducing the number of cases
referred for further scrutiny. HSBC are also carrying out risk assessments
using Monte Carlo simulations (these were described in the Optimisation
section in Chap. 3). From these they will be better able to understand their
trading positions and the associated risks.
Some of the more contentious uses of AI in business are in recruitment and
policing. The main contention is how bias can be (or even perceived to be)
built into the training data set, which is then propagated into the
­decision-­making process. In Chap. 8, I will discuss this challenge and the
potential mitigating approaches.
The Police Force of Durham, a county in the North East of England, have
started using an AI system to predict whether suspects should be kept in custody or not (in case they might reoffend). In this case, there are lots of data
available, from previous cases of reoffenders and non-reoffenders over five years,
that can be used to predict future reoffenders. At the time of writing it would
seem that there still a number of issues with this approach, the biggest of which
is probably with the validity of the data—this comes only from the region’s own
records, so excludes any offending that the suspect may have done outside
Durham. Despite the data challenges, the system accurately forecast that a suspect was a low risk for recidivism 98% of the time, and 88% for whether they
were a high risk. This shows how the system deliberately errs on the side of
caution so that it doesn’t release suspects if its confidence is not high enough.
In recruitment, a number of firms are taking the plunge into using AI to
help short-list candidates (although there are only a few who will currently
admit to it). Alexander Mann, a Human Resource Outsourcing Provider,
initially automated some manual tasks such as interview scheduling and

How AI Is Generating Insights

87

authorising job offers. They have recently introduced some AI software
(Joberate) to help them find candidates that are good two-way matches with
the job requirement. The software analyses both the candidate’s CV and publicly available social media feeds to create profiles of the candidates.
But AI can mean more than just mitigating business risks. Insights from big
data analysis can also be used to help fight diseases, especially cancer. AI has
been applied to the molecular analysis of gene mutations across the whole
body—usually cancer treatment research relates to specific organs—which
means that treatments that have been developed for, say, breast cancer could be
used to treat colorectal cancer. Through this approach personalised treatments
become possible—statistically significant meta-studies have looked at how
much benefit there was in matching the molecular characteristics of the tumour
of a patient with their treatment. This matching resulted in tumours shrinking
by an average of 31% compared to 5% from a non-personalised approach.
AI is also being used to develop cancer drugs. A biotech firm, Berg, fed as
much data as it could gather on the biochemistry of cells into a supercomputer so that the AI could suggest a way of switching a cancerous cell back to
a healthy one. The results have so far been promising, leading to the development of a new drug, and, of course, those results have been fed back into the
AI to further refine the model.
The quality of healthcare can also be monitored using AI. The Care Quality
Commission (CQC), the organisation that oversees the quality of healthcare
delivered across the United Kingdom, implemented a system to process large
numbers of textual documents and understand the opinions and emotions
expressed within them. The CQC is now able to manage the influx of reports
with fewer people and, most importantly, can apply a consistent scoring
methodology. The use of sentiment analysis (which was described in the NLU
section of Chap. 3) is also used in business-to-consumer companies, such as
Farfetch, a UK online retailer, to understand in near real time what their
customers think of their products and services. The companies are then able
to respond quickly and get a better understanding of their customers’ needs.
Other use cases for AI in optimising processes include: predicting failure
for jet engines; predicting the risk of churn for individual customers; predicting localised sales and demand trends; assessing the credit risk of loan applications; predicting farming yields (using data from IoT sensors) and predicting
localised demand for energy.
From the above examples, it should be clear that AI has huge potential to
unearth hidden value from a company’s data, helping manage risks and make
better, more informed, decisions. There are certainly challenges around the
lack of transparency and unintended bias, but, managed correctly the data can
provide insights that would have been impossible for a human to discover.

88

5

AI in Action

The Established AI User’s View
This is an extract from an interview with John Sullivan, CIO of Virgin Trains
West Coast, one of the major train operating companies in the United
Kingdom.
AB Tell me how you first came across AI.
JS	Well, I actually studied artificial intelligence at college. At that time, it was
obvious to me how useful AI would be but it wasn’t yet very practical,
particularly in a business setting. But I did get a good understanding of
what it could do and why it is different from traditional systems.
AB And when did you start looking for AI opportunities at Virgin Trains?
JS	In my current role as CIO, I was interested in looking for applications for
AI that could solve some real issues. We had a customer relationship challenge in that we were getting lots of queries by email that took a long time
for people to deal with. Some of the same questions were coming up again
and again, day in, day out.
So we looked at implementing Celaton’s inSTREAM solution to
improve the service we were giving to the customers. By using AI we were
able to respond to the majority of these queries quicker and more consistently. It also made it more efficient for us as a business and the work
became more interesting for the staff.
AB What does the AI software do?
JS	Basically, it reads all of the incoming emails, which of course are all written
in free-form, and works out what the customer wants – it understands
whether their email is a question, a complaint or a compliment and sends
it to the right agent within the organisation to deal with. It also does a lot
of validation work at the same time – so if a customer is writing about a
specific train, the system will check that that train actually ran, and will
also see if there are any similar queries relating to that journey – that all
helps us prioritise the email.
We’ve managed to reduce the amount of effort required at that initial
stage by 85%, which far exceeded our expectations. All of the people that
were doing that very mundane work are now doing much more interesting
stuff dealing with the more challenging questions we get.
AB So, do humans still deal with the remaining 15%?
JS	That’s right. When we started off we tried to identify which types of queries the AI could respond to and which ones a human would need to do.
Because the system learns as it goes along, more and more of the work can
now be done by the AI. So the agents only work on the really tricky ones,

The Established AI User’s View

89

as well as over-seeing the work done by the AI. This oversight task is
important for Virgin as we want to get the tone of the responses just right
so that they align with the brand. We call it ‘bringing out the Richard
Branson’.
AB	So apart from improving customer service, what other benefits have you
seen?
JS	The really important thing for us is that it has allowed the business to be
scalable. We now have a core team of human agents that is stable, so that
any increase in volume can be handled by the AI. Email volumes are now
not an issue for us.
AB How did you approach the implementation?
JS	We found Celaton quite quickly and immediately got on well with them.
They came in and we did a prototype with the Customer Relationship
team. Doing ‘small speedboat’ projects always works best for me for trials.
It’s always better than trying to build an ocean liner, which we know takes
too long, and are difficult to stop!
AB What challenges were there in implementing the system?
JS	Change management is always going to be a challenge in projects like this.
The CS team, though, got on board really quickly because they actually
wanted it. Remember – this was a system that was going to take away all
of those mundane, repetitive queries for them. They couldn’t wait!
The other thing that can be tricky with an AI project is getting the data
right. We definitely needed to involve the IT team, but, because we had a
good vendor in, Celaton knew what the challenges would be so we could
prepare as much as possible. We relied on their resources quite a bit, as we
didn’t really have any AI capability in-house at the time, but they were
good people who could explain everything in a simple manner to our
people—they weren’t just a ‘technical’ supplier.
AB	Thanks John. And finally, what advice would you give anyone just starting
out on their AI journey?
JS	Well, the communications, which I just mentioned, are a really important
aspect. You need to be able to articulate what AI is and how it works –
don’t assume that even the CIO knows these things. If I were to do this
whole thing again I’d probably bring somebody in that could focus on
this. Almost like doing internal marketing for the project.
I also think it is vital to open your mind up to what AI can do—again
external input can be useful here. We have Innovation Days to look at
developments we are doing with the trains, and we really should try and
do the same sort of thing for AI as well. It’s all about trying to understand
the art of the possible.

6
Starting an AI Journey

Introduction
I am often asked by executives, “I need AI in my business—how can I implement it?”. This is, of course, the wrong question to ask. The much more apposite question is, “I have some big business objectives/challenges—how can AI
help me deliver or address them?” Ideally, executives should be pulling AI in
to where it is needed the most and will provide the greatest value. In reality,
though, it is usually a bit of push and pull—AI is such a new subject and has
such huge potential to disrupt whole business models that it would be foolish
to consider it only in the context of your existing business strategy.
So, if businesses are already implementing AI and deriving real value from
it, as we have seen in the case studies presented in this book, how did they
actually start their journey? How did they discover, plan and implement the
solutions? This chapter will cover the approach to creating a meaningful and
workable AI strategy.
I have deliberately blurred the lines between an ‘AI Strategy’ and an
‘Automation Strategy’. As is hopefully clear from Chap. 4 when I discussed
the associated technologies, AI is rarely the only answer. Usually other automation technologies are required, such as RPA, cloud and IoT. But with this
book’s focus on AI, I will be looking at the automation strategy very much
through the lens of AI, mentioning the other technologies only where I think
it is necessary for clarity.
If I could summarise the best approach to maximising value from AI it
would be: think first, try something out, then go large. It’s very tempting to
jump straight in and create a PoC or bring in some AI software, but creating
© The Author(s) 2018
A. Burgess, The Executive Guide to Artificial Intelligence,
https://doi.org/10.1007/978-3-319-63820-1_6

91

92

6 Starting an AI Journey

an automation or AI strategy, one that matches your ambitions, is based on
your business strategy and has explored the business at a high level for the
prime opportunities, will give you, by far, the best foundation to maximise
value.
Once you have the AI strategy sorted, then you can start trying things out.
This isn’t mandatory, but it usually helps to build knowledge, trust and
momentum within the business. Your initial steps could be through a pilot, a
PoC, a hard prototype or a purchased piece of software. I will explain each of
these and discuss their pros and cons in this chapter.
But once the initial efforts are in and proven, then you will need to look at
pushing on quickly. There can be, at this stage, what I call the AI Bermuda
Triangle—lots of really good ideas that were heading in the right direction
mysteriously disappear from the radar. This happens with other technology
programs as well—once the effort of getting that first piece of work off the
ground has been done, everyone sits back and becomes distracted by other
things, motivations dissipate and momentum is lost. Once this happens it is
very difficult to get it back on track again. Therefore, the time to really push
forward is as soon as that pilot or PoC or prototype is showing promise and
demonstrating value.
And that plan really needs to be big and bold if it is to escape from the AI
Bermuda Triangle. Creating an AI roadmap is a key part of the AI strategy—
after all, how are you going to go on an AI journey without a map to guide
you? But it’s more than just a list of projects—it needs to describe how AI will
be industrialised within the business, and I’ll cover those aspects in more
detail in the penultimate chapter.
So, armed with the knowledge of what AI is capable of and how other businesses are using it, it is now time to start your own AI journey.

Aligning with Business Strategy
In my work as a management consultant and AI specialist, I know that the
one single activity that will ensure that maximum value is extracted from any
AI program is to first create an automation strategy that aligns with the business strategy, but also one that challenges it.
To align the automation strategy with the business strategy it is necessary
to understand the benefits and value that will be derived from that overall
strategy. Business strategies, if they are written succinctly, will have a handful (at most) of strategic objectives—these might be things like ‘Reduce the

93

Aligning with Business Strategy

cost base’, ‘Reduce the exposure to internal risk’ or ‘Improve customer service CSAT score’.
Each of these strategic objectives would deliver benefits to the business:
‘Reducing the cost base’ would deliver lower costs through, for example, not
hiring any new staff, not taking on any additional office space or minimising
travel. ‘Reducing the exposure to risk’ might be delivered through reducing
the number of unnecessary errors made by Customer Service staff, or improving reporting and compliance. And ‘Improving Customer Service’ could be
fulfilled by improving the average handling time (AHT) of inbound queries,
reducing the unnecessary errors (again) and enabling 24×7 servicing.
The important thing from our perspective in creating an automation strategy is to understand its role in enabling some or all of these benefits.
In a hypothetical example: AI Search could be used to reduce AHT by
reading incoming documents; Optimisation tools could be deployed to provide knowledge support to the customer service agents; self-service could be
enabled through a number of AI and RPA tools so as to reduce the demand
on bringing new staff and renting more office space; Clustering and Search
capabilities could be used to offer deep reporting insights into the business
information; Search could be used to identify areas of non-compliance against
a constantly updating regulatory database and chatbots (using NLU) could be
used to provide a 24×7 service desk. RPA could also be brought in to eliminate errors. Each of these tools then becomes an enabler to the business strategy benefits (Fig. 6.1).
These strategic
objectives…

…need to deliver these
benefits…

No new hires

Back Office Automation

No new facilities

“Improve customer
experience”

Analytics & Reporting

Improve AHT

Customer Centricity

24x7 servicing
Reduction in errors

Fig. 6.1 Aligning with the business strategy

Plan

3

Improved reporting
Improve compliance

3rd Parties
ties

“Reduce risk
exposure”

Self Service

Improve AHT

Tech3

2

Minimise travel

Tech2

“Reduce cost base”

Tech1

1

…enabled by
automation:

94

6 Starting an AI Journey

So, before we have even started creating the automation strategy, we must
know what it should achieve. Then it is a case of working backwards to determine the capabilities, technologies, people, plan and so on that will deliver
those objectives.
Of course, there is nothing to stop you from implementing AI simply
because you want to implement AI. In fact, demonstrating to shareholders or
customers that you are an innovative, forward-thinking company because you
are using AI can be of great benefit, and much PR can be made of the fact.
But, generally, aligning your Automation Strategy firmly with the Business
Strategy will ensure that the benefits you are hoping to achieve will deliver
long-term, sustainable value to the business.

Understanding Your AI Ambitions
The second major aspect to consider before starting your AI journey is to
understand, as much as possible, where you want to end up. That may sound
obvious, or you may think why should you bother thinking about that now
when you haven’t even started anything yet. But, as with any journey, it is
crucially important to know your destination. In AI terms, you won’t be able
to know exactly where you will end up (it’s a voyage of discovery, after all) but
you should at least understand your initial AI ambitions.
These ambitions can range from anywhere between ‘we just want to say
we’ve got some AI’ to creating a completely new business model from it. There
are no right and wrong answers here, and your ambition may change along
the way, but knowing your overall aspirations now will mean that you can
certainly start on the right foot and be going in the right direction.
I would consider four degrees, or levels, of ambition, which I have called:
ticking the AI box, improving processes, transforming the business and creating a new business.
The first of these, ‘ticking the AI box’, is for those who just want to be able
to claim in their marketing material and to their customers that their business, service or product has AI ‘built in’. This approach is surprisingly common, but I’m not going to focus too much on this approach because it can be
covered by selecting the most appropriate elements from the other types of
approaches. To be honest, you could justifiably say that your business uses AI
now because you filter your emails for spam, or that you use Google Translate
occasionally. Many businesses that have deployed simple chatbots claim to be
‘powered by AI’, which is factually true but a little disingenuous.

Understanding Your AI Ambitions

95

The first serious step to extracting value from AI will be in improving the
processes that you already have, without necessarily changing the way the
function or the business works. AI is being used to make existing processes
more efficient and/or faster and/or more accurate. These are the most common uses of AI in business right now, and inherently the least risky; process
improvement is the approach that most executives will consider first when
deciding on their initial steps into the world of AI.
I’ve discussed case studies where processes can be streamlined through the
application of AI Search (extracting meta-data from unstructured documents,
for example) or by using big data to carry out predictive analytics (more
­accurate preventative maintenance schedules, for example). Other examples
of process improvements are where AI has helped to filter CVs for job candidates, or helping make more efficient and more accurate credit decisions.
Whilst making existing processes faster, better and cheaper can provide a
great deal of value to a business, there is arguably even more value available
through transforming the processes or the function. By transformation I
mean using AI to do things in a materially different way, or in a way that
wasn’t even possible before. Some of the examples of transformation that I
have discussed earlier include: analysing the customer sentiment from hundreds of thousands of interactions (NLU); predicting when a customer is
going to cancel their contract (Prediction); recommending relevant products
and services to customers (Clustering and Prediction); predicting demand for
your service (Clustering, Optimisation and Prediction); or modelling different risk scenarios (Optimisation).
One transformation that AI is particularly suited for is the enablement of
customer or employee self-service. The benefits of a self-service option are that
it can be made available twenty-four hours a day, seven days a week, and that
it is generally cheaper to run. It also gives the customers or employees a sense
of empowerment and control. AI can be used for the direct engagement
aspects of the process, employing chatbots and/or speech recognition capabilities to communicate with the person, and also for any decisions that need
to be made based on that communication (e.g. should this request for credit
be approved?) by using prediction or reasoning tools. (Also useful when creating a self-service capability is RPA, which is able to handle all of the rules-­
based processing and connect all of the necessary systems and data sources
together in a non-disruptive manner, as discussed in Chap. 4.)
The biggest impact that AI can make on a company is when a whole new
product, service or business can be created using the technology at its core.
Probably the most famous example of this is Uber which uses a number of

96

6 Starting an AI Journey

different AI technologies to deliver its ride-hailing service. For example, it
uses AI to provide some suggestions as to where you might want to go to,
based on your trip history and current location (e.g. I usually want to go
straight home after I am picked up from the pub). It will also help predict
how long it is going to take until the driver arrives to pick you up. Uber have
also analysed the time it takes to actually pick people up (i.e. from the car
arriving at the site to when it leaves again) to enable the app to suggest the
most efficient pickup spots. (After using third-party navigation apps, Uber
eventually developed its own navigation capability, although it still relies on
data from third parties.)
Other types of businesses have been created from a foundation of AI. (I’m
not considering AI software vendors or AI consultancies here as they will
inherently have AI at their core.) Powerful recommendation engines have
been used by companies such as Netflix and Pandora to transform their businesses, and Nest, a ‘smart thermostat’ that uses predictive AI to manage the
temperature of your house; without AI Nest would just be another thermostat. Other businesses that have initially used AI as a foundation for their core
business are exploiting the technology to create new revenue. Pinterest, the
website where people post interesting images from other sites, is a good example of this. They developed a very strong image recognition system so that
users could find, or be recommended, similar images. They are now developing apps based on the same technology which can automatically detect multiple objects in images and then find similar images of the objects on the
internet, including links to buy those objects (from which they are bound to
take some commission).
So, understanding your AI ambitions is an important early step in developing an AI strategy and subsequent AI capability. Those ambitions will guide
the first steps that you take and help steer you on the overall journey. But do
bear in mind that not every part of your business will want or need to go on
the same journey or at the same speed. Your front office might want to adopt
AI wholesale whilst the back office is happy to take things slowly. And each
department or function might be starting with different levels of automation
maturity. If there are functions that have already adopted some form of automation or have made efforts to organise their data, these will provide stronger
platforms to introduce AI than somewhere that still operates very manually or
doesn’t have a strong data governance.
The next section introduces the AI Maturity Matrix that can help you assess
both the current automation maturity as well as your ambitions for each significant area of your business.

Assessing Your AI Maturity

97

Assessing Your AI Maturity
A Maturity Matrix is a relatively simple concept that is a very useful on a
number of fronts. Not only does it encourage discussion and reflection during
the creation process but, once completed, it can also be used as a communication tool.
Maturity Matrices were originally developed by Carnegie Mellon University
to assess the maturity of IT development functions. They generally describe
five levels of maturity, from very immature (Level 1, or ‘Initial’) to world class
(Level 5, or ‘Optimising’). Each maturity level consists of related practices for
a predefined set of process areas that improve an organisation’s overall performance. Thus, an organisation achieves a new level of maturity when a system
of practices has been established or transformed to provide capabilities the
organisation did not have at the previous level. The method of transformation
will differ at each level, and requires capabilities established at earlier levels.
Consequently, each maturity level provides a foundation on which practices
at subsequent maturity levels can be built.
Capability Maturity Model (CMM) levels in the IT development world
can be formally assessed by approved consultancies, and many big IT companies wear their CMM Level 5 badges with pride. But not everyone should
want or need to reach Level 5—in many cases Level 3 (‘Defined’), where
processes are documented, standardised and integrated into an organisation,
is perfectly adequate for most businesses.
This idea of evaluating what is ‘good enough’ for your business, and each of
the major functions in it, is at the heart of developing the AI Maturity Matrix.
The methodology I describe in this section takes the concept of the Maturity
Matrix and applies it specifically to automation. As I mentioned when discussing the Automation Strategy generally, it is best to try and think holistically about automation but with a strong focus and lens on AI, just so that
related opportunities, or dependencies, are not missed along the way. So, in
the matrix we are considering ‘automation’ that includes the full range of AI
tools (chatbots, search, data analytics, optimisation engines, image classification, voice recognition, etc.) as well as RPA, robotics, IoT and Crowd Sourcing
(which are described in detail in Chap. 4).
The AI Maturity Matrix has six levels rather than the usual five—this is to
introduce a Level 0 where there is no automatson at all. Each of the six levels
is described thus:
Level 0: Manual Processing There is very little evidence of any IT automation
in the organisation. Only basic IT systems such as email and ‘office’ applications
are deployed. There are large numbers of people processing transactional work,

98

6 Starting an AI Journey

either in-house or through an outsource provider. Data is not considered an
asset and there is no formal governance in place to manage it. There are no significant projects looking at, or delivering, automation.
Level 1: Traditional IT-Enabled Automation The organisation has implemented task-specific IT applications for particular processes (e.g. an invoice
processing application to process invoices). There is no evidence of automation tools, and specifically AI nor RPA, having been deployed. Data is
­managed only to the extent that it needs to be to ensure the smooth running
of the organisation. There are still large numbers of people carrying out transactional and cross-system processing work.
Level 2: Isolated, Basic Automation Attempts Some individual teams have
used scripting or macros to automate some tasks within a process or application in isolated areas of the business. Minimal benefits are evidenced from
these efforts. There are still large numbers of people carrying out transactional
and cross-system processing work. Data is still managed on a nominal basis.
Level 3: Tactical Deployment of Individual Automation Tools Some functions have deployed individual automation tools such as AI and RPA to automate various processes. Some useful benefits have been identified. No
dedicated automation resources are used or have been set up. Some data is
managed so as to make it valuable for the automation efforts, but an
organisation-­wide data governance framework does not exist. In areas that
have experienced automation, the staff are working in different ways. A case
study or two from the organisation is known within the industry.
Level 4: Tactical Deployment of a Range of Automation Tools Functions or
divisions within the organisation have deployed a number of different automation tools, including AI, across a range of processes. A strong business case
exists that identifies sizeable benefits. Data is being managed proactively, with
some areas having introduced data management and retention policies. Staff
may have been redeployed as a result of the automation efforts, or they are
working in materially different ways. Some dedicated resources have been
organised into an automation operations team. The organisation has a reputation for implementing new technologies and being innovative.
Level 5: End-to-End Strategic Automation The organisation has implemented a strategic end-to-end process automation program with a range of
automation tools including AI. Significant benefits in cost, risk mitigation

Assessing Your AI Maturity

99

and customer service have been delivered. Data is treated as a valuable asset
and is managed through an organisation-wide data governance framework.
Many of the organisation’s staff are working in materially different ways than
before. An Automation Centre of Excellence has been established. The
­organisation is renowned for its innovative and forward-thinking culture.
In the above descriptions, I have used the word ‘organisation’ to describe
the scope of the work being evaluated—this could indeed be the whole organisation but could equally be applied to parts of it, such as the Finance
Department, the Customer Services Department or a Shared Service Centre.
When choosing the level of granularity to assess, it should be at the highest
point (i.e. the biggest scope) that has a significant difference in automation
approach and needs. Generally, this will be by function or division, but could
also be by geography.
The evaluation of maturity levels is a subjective one, assessed through interviews and the review of evidence. The approach should be one that is satisfactory to all people who are involved and must be as consistent as possible across
the different areas. Many people use a third-party consultancy to carry out the
evaluation to ensure consistency but also provide independence from any
internal politics there may be.
There can be some overlap between the different levels. For example, a
department could have tactically deployed a range of automation tools (Level
4) but also created an Automation Centre of Excellence (a Level 5 criteria). It
will be up to the person or team evaluating the levels to decide the most
appropriate level to assign it to. The important thing is to be consistent.
As well as assessing the current automation maturity, the ‘automation
ambition’ should also be agreed upon. As mentioned earlier, the target level
doesn’t necessarily have to be Level 5, and could be different for different areas
that are being assessed. There are many reasons why not all areas should strive
for Level 5, including the cost of implementation, there not being enough
relevant data, it not being aligned with the strategic objectives and it simply
not being appropriate (a human touch might be the best one in some cases).
A completed AI Maturity Matrix would therefore look something like the
following (Fig. 6.2):
It shows both the ‘as is’ level of maturity (in grey) and the agreed ‘automation ambition’ for each area (in black). Although it is a relatively simple chart,
it lays the foundations for the Automation Strategy and Roadmap. It also
provides a useful communication tool to describe, at a high level, what the
business is hoping to do with automation and AI.

100
Maturity Level
Process Area

6 Starting an AI Journey
0

1

2

3

4

5

Manual processing

Traditional ITenabled automation

Isolated, basic
automation attempts

Tactical deployment
of individual
automation tools

Tactical deployment
of a range of
automation tools

End-to-end strategic
automation

Customer Service
Risk Assessment
Operations
Finance
HR
ITSM

Fig. 6.2 AI maturity matrix

Levels 0, 1 and 2 do not involve the sort of automation that I am considering in this book. Level 3 is where AI (and RPA) are starting to be introduced,
generally focusing on one or two individual processes. This equates to the
‘ticking the AI box’ ambition discussed in the previous section. Level 4 applies
AI to a wide range of processes, applying it across a number of areas, although
still tactically. This level equates to the ‘process improvement’ ambition. Level
5, where AI is applied strategically across the business, is equivalent to the
‘transformation’ ambition.
The ambition I described of creating a new service line or business by
exploiting AI goes beyond the maturity matrix structure as it is not based on
an existing function or department. That level of ambition means starting
with a blank sheet of paper.
The AI Maturity Matrix then is a useful tool to start your AI journey. It
provides an opportunity to openly discuss the role and opportunities for AI
and automation across the business and provides a platform to develop a ‘‘heat
map’ of what those opportunities might be. The approach to creating an AI
heat map is described in the next section.

Creating Your AI Heat Map
Following the development of the AI Maturity Matrix, the next step in creating an AI Strategy is to build a ‘Heat Map’ of where the opportunities for
automation might be. Driven by the Business Strategy and the benefits that it
is hoping to achieve, the Heat Map provides a top-down perspective on areas
where AI is desirable, economically viable and/or technically feasible. It starts
to identify the types of AI capabilities that could be applied in each area in
order to realise the automation ambitions (and therefore contribute to delivering the business’s Strategic Objectives).

Creating Your AI Heat Map

101

The approach to creating an AI Heat Map is not too complicated as long as
you understand your business well enough, and, of course, have a sound grasp
of the AI Framework. It is intended as a starting point, a way to bring some
focus and logic to prioritise your initial AI efforts.
Firstly, you must decide on the overall scope of the Heat Map. It usually
makes sense to keep this consistent with the Maturity Matrix. So, if you originally evaluated the whole business split down into, say, five different business
areas, then use the same structure for the heat map.
Each area, then, is assessed in turn, identifying the opportunities and the
relevant AI capabilities required in each one. This is best done in two passes,
the first one identifying all opportunities without any judgement being
applied to them, rather like the brainstorming of ideas: all potential opportunities should be logged, dismissing none of them at this stage.
The opportunities are identified by considering a number of different criteria, each of which is described below. The opportunities are surfaced through
interviews with relevant managers in the areas being assessed, and should be
done by people who have a sound understanding of the AI capability framework as well as the AI technology market. Organisations without the appropriate resources in-house will tend to use third parties for this exercise.
• Alignment with strategic objectives—unless you are looking to implement AI for the sake of it, it is important to identify AI opportunities that
contribute in some way to achieving the Strategic Objectives. For example,
if your only strategic objective is to improve customer satisfaction, then the
focus for this exercise should be on those processes that can impact customer service, rather than, say, cost reduction.
• Addresses existing challenges—there may be opportunities for AI to solve
existing problems, such as inadequate management information, poor
compliance or high customer churn. These opportunities may or may not
be aligned with the strategic objectives but they should still be considered
for inclusion as they could certainly add value.
• Available data sources—because AI, in most of its forms, requires large
amounts of data to be most effective, a key consideration is to identify relevant data sources. Where there are very large data sets, then it is possible
that there is an opportunity for AI to extract some value from them.
Conversely where there is very little, or no data, it may preclude using AI.
(Some AI technologies, such as chatbots and cognitive reasoning engines,
both discussed in Chap. 3, do not require large data inputs, just captured
knowledge, to work effectively, so don’t simply equate low data volume to
no AI opportunities.)

102

6 Starting an AI Journey

• Available technology—understanding the AI capabilities and the relevant
tools that can deliver them is also important in order to identify AI opportunities. Whilst it is always best to ‘pull’ ideas in based on what the business
requires, with AI at the forefront of technology it is also valid to ‘push’ ideas
based on the technology that is available.

Customer Service
Risk Assessment
Operations
Finance
HR
IT
Fig. 6.3 AI heat map first pass

Crowd

IT automation

Image

Voice

Fraud

Risk

Process

Analytics

Chatbots

Function

Search

Automation Type

RPA

By this point you should have a range of different opportunities identified
in each of the respective areas you are investigating. For example, in Customer
Services, you could have identified the opportunity to provide an online self-­
service capability for the purchase of tickets for events. This could help deliver
your strategic objectives to improve the customer experience and to increase
ticket sales. In terms of the AI capabilities it could require Speech Recognition
(if you want to offer telephone access), NLU (for the chatbots) and
Optimisation (to guide the user through the purchasing process). It could also
require some RPA capability to actually execute the purchase. (I would also
include Crowd Sourcing, discussed in Chap. 4, as another capability that
should be considered in order to support any of the AI opportunities.) With
regard to data sources, there are pre-trained voice services available, and there
is a good source of knowledge available from the human customer service
agents that currently process these transactions (Fig. 6.3).
For each of these ideas, you should therefore understand what the opportunity is, how it links to strategic objectives (and/or addresses current issues) and
which AI capabilities you will require to realise it. Building each of these on top
of each other then starts to create a heat map of AI benefits and requirements.

Creating Your AI Heat Map

103

(I tend to use colours to identify the ‘hottest’ areas, but you can use numbers
or whatever approach you are used to.) Areas that are delivering the most benefit are clearly visible, as well as those AI capabilities which are going to be most
important to you (Fig. 6.4).
The second pass across the various opportunities is to filter them down to
those that are desirable, technically feasible and economically viable. If the
idea can’t satisfy all of these criteria then it is unlikely to be successful.
Desirability is a measure of how much the business wants and needs this
new idea, so it relates to how aligned it is with the strategic objectives and how
it addresses existing challenges. But it should also consider the opportunity
from the customer’s perspective (if it impacts them) and, culturally, how
acceptable the idea will be to the department or function that will be implementing it. In some cases, you will have to consider the personality of the
managers and staff of those areas as well and understand how supportive (or
defensive) they may be. Clearly, if an idea does not pass the desirability test (to
whatever threshold you think is appropriate for your organisation), then it
probably shouldn’t be progressed any further for now.
Technical feasibility will already have been assessed to a certain extent
during the first pass when the data sources and the availability of technology
were taken in to account. For the second pass this is done in more detail,
considering aspects such as the quality of the data, the processing power or
bandwidth that may be required, the maturity of the required technology and
the technical skills that will be required internally and externally to implement it. Other aspects such as regulatory restrictions, especially about the use
of data, may also have to be considered.
Testing the economic viability is the first look into what the business case
might be like. A comprehensive business case will be developed in time, but
at this stage the economic viability can be assessed at a relatively high level. It
should start to consider the financial benefits to the business which could be
achieved through cost reduction, risk mitigation, improved debt recovery,
new revenue generation or reduced revenue leakage. Estimates can be made as
to how the opportunity might improve Customer Satisfaction (CSAT) scores,
if relevant. Costs should also be assessed where they can be at this stage. These
could include license fees, IT infrastructure and professional services fees. It
may be difficult to assess the costs at this early stage; therefore, it could be
appropriate to score, out of 10, say, the magnitude of the costs, with anything
scoring above a certain threshold rejected (this obviously needs to be balanced
against the benefits that will be gained).
Each idea should be able to pass all three tests if it is to survive into the final
Heat Map. You can, of course, make exceptions, but be clear on why you are

Opportunity 1

Opportunity 2

Opportunity 1

Fig. 6.4 AI heat map

Finance

Operations
Opportunity 3

Opportunity 2

2

1

1

Customer Services
Opportunity 1

1

Organisation

AI Heat
Map
Maturity

Existing
Challenges

Strategic
Objectives

Strategic Objective 1

Challenge 3

Strategic Objective 3

Strategic Objective 2

Challenge 1
Challenge 2

Cost Reduction
Customer Service

Benefits

Compliance
Risk Mitigation
Loss Mitigation
Revenue Generation
Leakage Mitigation
Image
Voice

AI Capabilities

Search
NLP
Planning
Prediction
RPA
Crowd

104

6 Starting an AI Journey

Developing the AI Business Case

105

making those exceptions, and keep them in mind as you progress forward. Of
the rejected ideas, do keep them safe. Things change, especially around the
technological viability, so something that may not be suitable now may be just
right in the future.
So now you have an AI Heat Map which shows the main areas you will be
focusing on, the benefits they can bring and the capabilities required to
achieve them. Each opportunity, as well as the high-level view presented in
the Heat Map, should also have a corresponding passage of text (or a separate
slide) which provides more of the necessary details about what the opportunity actually entails and some explanation of the numbers.
As a summary, and for presentation purposes, the AI Heat Map can be
rolled up into an overview across the main areas and the organisation as a
whole.
Now that we have a good idea of the opportunities that form the basis of
the Automation Strategy, the next stage is to develop those Heat Map opportunities into a Business Case.

Developing the AI Business Case
Creating a Business Case for an AI project is, in many ways, the same as for
any technology project—there are benefits to be achieved and costs to incur.
But with AI the task is made more challenging because there tends to be more
unknowns to deal with. These unknowns can make calculating an ROI
(Return on Investment) an almost-futile exercise, and the organisation must
rely on an element of intuition as well as the numbers. This is especially true
when new approaches and new services are being developed through AI.
Luckily, the work done creating the AI Maturity Matrix and Heat Map
provides a good methodology for evaluating which are the AI opportunities
that will deliver the most value. Creating a meaningful business case for AI for
each of these is certainly feasible with a little thought.
We already have, from the AI Heat Map, a list of opportunities and a high-­
level assessment of which are the most promising. In my example in the previous section the opportunities which are showing as the ‘‘hottest’ (dark grey)
are the ones to prioritise, especially if they are strong across a number of different aspects (strategic alignment, solving current challenges, benefits types).
At this stage, you can easily apply some scoring mechanisms to make the
prioritisation process easier. For example, you could replace the shades or
colours with scores out of 3, and give weightings to each of the criteria, and
then add all these together to give a total score for each opportunity.

106

6

Starting an AI Journey

If you are feeling confident, or have a third-party supporting you in this
exercise, you could also introduce additional criteria under the broad heading
of ‘Ease of Implementation’. This would be influenced by the maturity of the
function (from the Maturity Matrix) but also considers the technical feasibility and desirability that we looked at when filtering the opportunities during
the second pass of creating the Heat Map. Scoring each opportunity in this
way, again with appropriate weightings, will give additional depth to your
decision-making process.
From this work, we now have a prioritised list of AI opportunities to consider: it’s time to start getting serious with a few of these. Whether you take
just the top scoring idea, or the top 3 or top 10 is entirely up to you, and will
be based on your overall AI ambitions and the time and resources you have
available. However many you take, it is likely that you will need to draft in
some specialist resources at this stage, either from your own organisation or
from third parties.
Different organisations have different ways of calculating business cases.
Some will focus on ROI, some Net Present Value, some internal rate of return
and others payback. I’m not going to explain each of these here as they are
pretty standard approaches, and I am assuming that you are familiar with the
one that is favoured by your own organisation. But all of these do require
calculating the benefits and the costs over time, and so I will provide an aide
memoire of what sorts of things you might need to include in each of these
areas.
The benefits that AI can deliver are significant and diverse, and can be split
into those that are ‘hard’, that is, can easily be equated to monetary value, and
those that are ‘soft’, that is, those that are more intangible and difficult to
quantify. It should also be noted that a single AI implementation may deliver
a range of different benefits, not just one specific type. This is an important
point—whilst you may implement an AI system to, say, reduce costs, there
could well be associated benefits in compliance or risk mitigation. For each
opportunity identified in the AI Heat Map, consider whether each of the
benefit types described below could be applicable.
The Hard Benefits can be categorised as follows:
• Cost reduction: this is the simplest benefit to quantify as there is usually a
base cost that can be compared to a future, lower cost. For AI, these are
generally the situations where the new system will replace the humans
doing a role or activities (remember the discussion in Chap. 1 around
replacement versus augmentation?). The AI search capability will replace
the activities of reading and extracting meta-data from documents—it will

Developing the AI Business Case

•

•

•

•

107

do it faster and more accurately than humans which means the benefits will
be magnified. Chatbots (using NLU) can also replace some of the work
done by human call centre agents, as can speech recognition when combined with a cognitive reasoning engine. NLG (a sub-set of NLP) can
replace business analysts in creating financial reports. More efficient route
planning for workers and/or vehicles reduces time and cost.
Cost avoidance: for businesses that are growing, cost avoidance is the more
acceptable face of cost reduction. Rather than recruiting additional people
to meet demand, AI (as well as other automation technologies such as
RPA) can be implemented as an alternative. The AI solutions will be similar to those for cost reduction.
Customer satisfaction: this is usually measured through a survey-based
satisfaction index such as CSAT or Net Promoter Score (the difference
between positive and negative customer feedback). Some businesses have
connected this to a monetary value, and others relate it directly to a manager’s own evaluation and bonus. For many it is a key performance indicator. AI has the opportunity of improving customer satisfaction through
being more responsive to queries, being more accurate in the responses,
providing richer information in the responses, reducing the friction in customer engagements and in making relevant recommendations. (Of course,
AI can also be used to automatically measure customer satisfaction itself
through sentiment analysis.)
Compliance: some people might challenge the fact that a self-learning system can improve compliance which is based on hard and fast rules, but
actually AI is very good at identifying non-compliance. Using NLU and
Search it can match policies and procedures against regulations and rules
and highlight the variances. Compliance benefits can be measured through
the potential cost of not being compliant, such as through fines, or loss of
business in specific markets. (RPA, by the way, is a strong driver of compliance as each automated process will be completed in exactly the same way
every time, with each step taken logged).
Risk mitigation: AI can monitor and identify areas of risk in areas where
it would be impossible, or very expensive, for humans to do it. The classic
example for AI here is fraud detection in high-volume transactions, such as
credit card payments. AI can, in some circumstances, make better risk-­
based decisions than humans, such as for credit approvals. AI can also contribute to the Know Your Customer (KYC) process in being able to validate
documents and data sources and help with the credit checking process
(RPA also helps here to access the necessary systems and run the process).

108

6

Starting an AI Journey

As with compliance, risk mitigation can be measured by the costs that have
been avoided, such as the loss through fraud or poor credit decisions.
• Loss mitigation: loss mitigation can be achieved through improved debt
recovery. Most of the heavy lifting here can be done through RPA by managing the timing and activities in the process, but AI can play a role, for
example, in generating appropriate letters or engaging with the debtors.
Loss mitigation can be measured through the increase in cash being
­recovered through the use of automation.
• Revenue leakage mitigation: this is usually achieved through reducing the
situations where revenue generating opportunities are lost, such as when
customers no longer use your services. Reducing customer churn through
the early identification (i.e. clustering) of the tell-tale behaviours can be
measured by calculating the revenue value of a customer. Other AI capabilities, such as NLU and prediction, can also contribute by keeping the
customer more engaged with your business.
• Revenue generation: this is the AI benefit that can probably deliver the
most in terms of monetary value. AI can help enable self-service capabilities (through Speech Recognition, NLU, Optimisation and RPA, for
example), identify cross-selling and up-selling opportunities (through
Clustering) and create new revenue generating opportunities either from
your existing business or through completely new services or products.
Revenue generation is easy to measure after the fact, although isolating the
specific contribution of the AI may be challenging and assumptions will
need to be made. Predicting revenue generation will need to use modelling
techniques, most of which should already be available in your business.
One thing to bear in mind with measuring automation benefits is that, in
some cases, staff productivity could actually be seen to decrease after implementation. This is not a bad thing, but just a reflection that the staff are now
handling the more complex cases whilst the technology is handling the simpler ones. Overall, the productivity will have improved, but if you just look at
the human staff they may well be taking longer to process the (more complicated) cases.
The softer benefits are inherently more difficult to put a monetary value to,
but can still be evaluated, even if the accuracy is less certain:
• Culture change: this is probably the most difficult thing to achieve and the
most difficult to measure, but can deliver significant value to an organisation. Of course, it depends on the type of culture you have already and

Developing the AI Business Case

•

•

•

•

109

what you would like to change it to, but AI can help embed a culture of
innovation as well as deliver customer centricity into the business. Actual
benefits will usually be associated with a wider change program and it will
be challenging to isolate these that are specifically due to AI, but, because
of the potential value, these should not be ignored.
Competitive advantage: this can deliver significant benefits if AI provides
a first-mover advantage, a new service line or access to new markets or customers. The benefits will be closely associated with (and usually accounted
for in) some of the ‘hard’ ones such as cost reduction, customer satisfaction
and revenue generation, but competitive advantage should always be
sought as a way to deliver a step-change in value.
Halo effect: AI, at least in the current decade, is a highly marketable aspect
to have within a business. The objective of being able to ‘tick the AI box’
can provide the opportunity to claim you are an ‘AI-enabled’ business
which could attract new customers to your product or service. Any amount
of AI implementation, if marketed correctly, can also have a positive impact
on shareholder value if it results in an increase in the share price of the
company.
Enabling other benefits: as well as providing direct benefits, AI can also
enable other, indirect benefits to be delivered. For example, implementing
AI can free up staff that can be deployed on higher-value activities, or data
that is generated from an automated process can be used to provide additional insights that can be exploited in another part of the business.
Implementing AI may, through eliminating much of the mundane work in
a department, also reduce employee churn.
Enabling digital transformation: transforming businesses to be more
‘digital’ is a common strategic objective for many businesses right now. AI
is clearly part of that transformation, and many of the benefits of a digital
business can be directly delivered, or indirectly enabled, through the implementation of AI.

An AI business case will include at least one of the above hard and soft
benefits I have listed and often quite a few of them. The individual classes of
benefits should be used as a starting point for your AI business case. For each
AI opportunity identified in the heat map, consider each type of benefit carefully so that all can be identified and none are forgotten or left out.
Once all the relevant calculations have been done, the benefits for each
opportunity can be summed across the departments and the organisation to
give a full picture of the value that AI can bring. Also, consider whether there

110

6

Starting an AI Journey

are any synergies between the different opportunities—for example, does having a number of AI initiatives going on enable a favourable cultural change in
that department?
The heat map, bolstered with the business case data, now provides a comprehensive prioritised list of AI opportunities within your business. But,
before you take the first step into prototyping and implementing the most
favourable candidates, you will need to understand what your ultimate roadmap looks like and how you will handle all the change management aspects
that go along with it. These two pieces of the AI strategy jigsaw are covered in
the next two sections.

Understanding Change Management
Even the simplest IT projects require change management. With automation
and AI, the challenges are magnified—not only are you going to fundamentally change the way people work, but you may also have to let people go. And
even if this is not your intention, the reputation of automation projects means
that staff will likely think it is the case.
You may, as part of your AI strategy, be creating a new product or service
that will be competing with your legacy products or services. That ‘healthy
competition’ may be good for business but is not necessarily good for staff
morale. And building a new capability will mean transforming your organisation—new jobs will be created and old jobs will disappear. Some are going to
be able to cope with the change, and some are not.
Right at the beginning of the book I talked about the differences between
AI replacing humans and AI augmenting them. Clearly an augmentation
approach will have less change management issues than a replacement one.
But anything that can demonstrate how AI will enrich the work that people
are doing (even if they have to do it differently) will help. Having greater
insights into their work, or customers, or suppliers, through AI analysing and
interpreting data, will help people become more successful in their jobs.
If automation is replacing tasks, it is often for the more transactional, tedious
ones. There is a useful phrase coined by Professors Willcocks and Lacity, which is
that “automation takes the robot out of the person” (Willcocks and Lacity 2016).
So, if your AI projects in any way relieve people of the mundane, repetitive tasks
that they have been doing for years, then this should be celebrated.
Despite the amount of change that AI inherently throws up, many of the
common change management practices are relevant to AI projects: communicate early and often; involve and empower employees as much as possible;

Understanding Change Management

111

look for short-term wins; consolidate gains to generate further momentum;
incentivise employees where appropriate; pace the change appropriately and
have honest reviews before the next one.
A challenge that automation projects in particular face is a post-prototype
slump. The initial build can become a victim of its own success—the relief of
actually creating something that works, even if it is not a complete solution,
can be so overwhelming that everyone then takes a step back, resulting in a
loss of focus and momentum.
To overcome this, it is first necessary to be aware of it, and then to prepare
for it. Schedule in review sessions for as soon as the prototype is due to be
completed. Have a Steering Group Meeting soon after, or get the CEO to
make a celebratory announcement about how this is ‘the end of the beginning’. If possible, have the next phase of the project, or another prototype,
start before the first project finishes, so that focus can quickly shift to that.
The worst thing you can do at this point is to pause for a few weeks or months.
Generally, a ‘big and bold’ approach to an AI program will deliver the most
success, as it avoids any potential slump periods, and helps maintain that all-­
important momentum. However, all organisations are different and an ‘under
the radar ‘approach may be the most suitable for you—just make sure that
your business case and roadmap (covered in the next section) reflect that.
As well as managing all the change within your organisation, you may have
to manage change for your customers as well. If your AI project is one that
impacts the way customers engage and interact with you, then you will need
to fully prepare them for that change. This will involve plenty of communication, supported by marketing efforts and relevant collateral such as emails,
website articles and frequently asked questions (FAQs).
If your AI project changes the way that you use your customers’ data, then
you will need to make it clear to them what the impact of this is. Any sensitivities around data privacy, in particular, will need be made explicit. You would be
well advised to seek legal advice for any changes to your customers’ data usage.
This isn’t a book about how to manage change (there are plenty of those
about already), and you will likely already have your own preferred methodologies to tackle this. This is only to say that you should not neglect change
management in any way—implementing AI brings about its own unique
challenges, especially when some of the changes in an organisation can be
seismic in nature. These challenges need to be understood and carefully managed, and built into your AI roadmap.

112

6 Starting an AI Journey

Developing Your AI Roadmap
The AI Roadmap provides a medium- to long-term plan to realise your AI
Strategy. It can be a relatively simple affair that draws heavily from the AI
Maturity Assessment and the AI Heat Map.
The Maturity Assessment will provide you with your starting point. Those
areas that are immature when it comes to automation will require more effort
and time to complete. Your Change Management Assessment will also direct
you to those areas that might need additional support along the way.
At the core of the Roadmap will be the opportunities identified through
the AI Heat Map and subsequent Business Case. But rather than having a
series of individual opportunities being worked through, it is good practice to
group them into common themes. These themes can represent a number of
different projects, or work streams, ideally centred around common AI tools.
For example, a process excellence stream might focus on improving the accuracy of processes and reducing the AHT, whilst an analytics and reporting
stream would focus on delivering compliance and improved reporting. Each
stream, using a range of different technologies, relates specifically back to a
number of benefits that are outcomes of the business strategy.
The AI Roadmap should be kept at a relatively high level, identifying the
specific project work streams and associated activities (change management,
program management, governance, etc.). For each work stream there should
then be a much more detailed project plan which identifies dependencies and
responsibilities (Fig. 6.1).
It is likely that you will want to build one or two prototypes of your most
favourable candidate opportunities, and these should be built into the roadmap, ideally with their own detailed project plans. More details on the
approach to prototyping are provided in the next chapter. The prototype
builds will test and validate assumptions but also provide important momentum for your AI program.
The key considerations when building a roadmap include: your ultimate
ambition with regard to automation (is this a game changer for your business or just something you want to try out in certain areas); whether you
want to evaluate the whole business first and then implement AI or try and
get some quick momentum by implementing in some specific areas first;
and how much change your organisation can bear at any one time (see previous section).

 

Creating Your AI Strategy

113

It is important to remember that your AI Roadmap will be an important
communication tool. It should set out clearly what you are going to do and
when, so that you can use it to explain how it will deliver your AI strategy, and
ultimately contribute to your overall Business Strategy.

Creating Your AI Strategy
Your AI Strategy is now effectively complete. You will have looked at your
Business Strategy and worked out what aspects of that could be satisfied by
the application of AI technologies. You will have determined how ambitious
you want to be with your AI program—whether it is just to be able to say
you have some AI, or to improve some processes, or to transform parts of
your business. You may have even decided to create a new business by
exploiting AI.
You will then have looked at your organisation and assessed its level of AI
maturity. Some functions or departments could still be working manually,
whilst others could have started some automation projects and others might
even have some AI projects in place already. For each of these areas you will
have worked out their AI ambition, and thus determined the size of the gap
that you have to work with.
The AI Heat Map is the tool that helps identify how you are going to bridge
that gap. Based on your knowledge of the AI Capabilities Framework you will
have identified a number of different AI opportunities in each of the areas,
assessing and prioritising their potential benefits, challenges and their alignment with the Business Strategy.
From the information in the AI Heat Map you will have developed a high-­
level business case that sets out both the hard and soft benefits that you expect
to achieve.
And finally, you will have considered the change management activities
that are necessary to ensure the success of the AI program, and built them into
an AI roadmap that sets out your medium- and long-term plans of how you
will realise the AI strategy.
It is now time to start building something.

114

6 Starting an AI Journey

The Cognitive Reasoning Vendor’s View
This is an extract from an interview with Matthew Buskell, Head of Customer
and Partner Engagement at Rainbird, a leading vendor of Cognitive Reasoning
software.
AB: Why do you think AI is being so talked about right now?
MB:	I think it’s a combination of forces coming together and driving the interest. Specifically Outsourcing, Cloud Computing, Big Data and, of course,
Hollywood.
The major driving force for outsourcing, realistically, is economic.
However, regulation has increased and margins have continued to reduce.
So, when outsourcing does not give a deep enough return we need to look
elsewhere. In this case AI is attractive because it allows us to amplify the
productivity of the staff we already have in a way that is not possible by
outsourcing alone.
Cloud computing is a big part of the reason AI has become economically viable. That and the algorithms that have been optimised and refined
over the past 30 years. To illustrate this let me tell you about Ian. In 1994
Ian was doing undergraduate studies in software engineering and AI with
me at Birmingham. He decided to code a PhD thesis on linguistics.
When he ran his program the processing power required was so great that
it ground the entire Sun Hyperspark network of servers and workstations
to a halt for the entire campus. So he was banned from using the computers outside of the hours of 10 pm–6 am. As a result, we did not see Ian for
nearly a year. Today we call his field of research Natural Language
Understanding and you can buy it for a fraction of a cent per hit from
several cloud providers.
Big Data has made AI interesting. For some types of AI like Machine
Learning to work well they love large amounts of data. Whilst it’s true
most businesses were collecting data 20 years ago, until recently they had
not put the big data platforms in place that would allow AI systems to
access them. Now that they do, AI is well placed to take advantage of it.
Hollywood might seem a weird one to add, but that fact is as humans
we have been fascinated by the idea computers could think since Ada
Lovelace described the first computer in the eighteenth century. TV programs like Channel 4’s Humans in the UK and movies like Ex-Machina
fuel that excitement. It also sells news articles and ad-words on blogs.
There is a famous, but anonymous quote that sums up Hollywood’s influence on AI: “Artificial Intelligence is the study of how to make real computers act like the ones in the movies.”

The Cognitive Reasoning Vendor’s View

115

This also poses a big challenge for the AI industry today. The fact is we
are way off simulating a human mind, so the expectation of AI is so high
it is bound to lead to some disappointments along the way.
AB: What value do your customers get out of your software?
MB:	The Rainbird software is about mimicking human decisioning, judgement and recommendations. It’s different to other AI in that its ‘people
down’ rather than ‘data up’. Meaning we take a subject matter expert and
sit them down with Rainbird so they can teach it.
When you stop for a moment and think about what you can do if you
mimic human expertise and decisioning the list is endless. We have been
involved with projects from Abbey Road studios trying to mimic the
expertise of their (rather ageing) sound engineers, through to large banks
and tax consultancies trying to reduce headcount by encoding the best
employees’ expertise into Rainbird.
On the whole, however, we have noticed two areas where we see significant value:
1. If you can take expertise and create a solution that allows customers to self-­serve you can drive some staggering efficiencies. At the
moment, this is being manifested as a ‘chatbot’. However, long
term, we believe companies will have a ‘cognitive platform’ and the
UI may or may not be a chatbot.
2. If you can take knowledge that was in a process and move it further upstream the impact to the downstream processes is dramatic.
In one instance, we moved the liability resolution in a motor claim
so that it could be handled by the call centre agent. This removed
the need for over 15 downstream processes.
AB:	What do customers or potential customers need to focus on if they are
going to get the maximum value from AI?
MB:	I have noticed a lot of companies rushing into AI proof-of-concepts without spending enough time working though the customer journey or the
business case.
There are some customers that are approaching this correctly. For
example, I was with a client the other day in Scotland that was taking a
much more Design Thinking approach to the development of a PoC.
When we started the process we all had the idea we could get up and running quickly with a simple bot that could answer FAQs. When we actually looked at the questions and followed the conversation with a real
agent it quickly became clear that the general advice would only last a few
seconds before it became specific. At that point, you needed a human in

116

6 Starting an AI Journey

the loop, so the result of putting AI in place is that you would have paid
twice—once for the virtual agent and secondly for the human. This is a
Bad Idea. We were actually able to find the value by going deeper with the
virtual agent and make it capable of completing a full interaction with a
smaller number of questions.
AB: Do you think the current hype around AI is sustainable?
MB:	Personally, I don’t think the hype is helpful, so I welcome it slowing down
or becoming more grounded. However, I think unfortunately it will continue, mainly because we need AI to work. Without it we have to face
some pretty stark economic realities.
AB: How do you see the market developing over the next 5 years or so?
MB:	The market is very fragmented at the moment and the use cases for the
technology are very broad. So, I think it will do what software innovation
has always done: all these companies will consolidate so they can offer an
end-to-end solution and the use cases that work well will survive and the
rest will die out.
The final thought I would leave you with is from John McCarthy, who
is considered the father of AI: “As soon as it works it’s no longer called AI
anymore.”

Reference
Willcocks LP, Lacity MC (2016) Service automation: robots and the future of work.
Steve Brookes Publishing Warwickshire, UK.

7
AI Prototyping

Introduction
Creating your first AI build, however small, is a key milestone for any AI program. After all of the hard work developing the AI Strategy, this is the point
at which things really start happening, and people in your organisation can
actually see tangible outputs. Getting this stage right, therefore, is vitally
important.
There are a number of different approaches to this, which I cover in some
detail in the ‘Creating Your First Build’ section of this chapter, and they each
have different names and acronyms. For simplicity, I have used the word ‘prototype’ as the general term for all of these.
Prototyping can be done once the AI Strategy is complete, or can be done
to validate some of the assumptions during the development of the strategy.
This may be influenced by how important those assumptions are (if they are
fundamental to the success of the program, then test them early) and how
much stakeholder buy-in and momentum you need (a successful prototype is
a great way to engage with stakeholders).
Some of the decisions you make at the prototyping stage may have an
impact on the rest of the program. The most important of these is your technology strategy—do you build your AI solutions yourself, buy them from a
vendor or build them on top of an established AI platform, or use a mixture
of all of these? It is this question which I will address first.

© The Author(s) 2018
A. Burgess, The Executive Guide to Artificial Intelligence,
https://doi.org/10.1007/978-3-319-63820-1_7

117

118

7 AI Prototyping

Build Versus Buy Versus Platform
Some of the business case costs that were discussed in the previous chapter
will have required some assumptions on how you intend to create your AI
solutions. Generally, there are three main approaches to take: off-the-shelf AI
software, an AI platform or a bespoke AI build. Your final solution may
involve any of these approaches, but the core functionality is likely to initially
focus on one of these:
• Off-the-shelf AI Software is generally the simplest approach, since most of
the hard work in designing the system has already been done by the vendor.
As I’ve described in the earlier chapters, there is a plethora of AI vendors,
each of which can provide a specific capability that can be used as a stand-­
alone application or as part of a wider solution.
As well as all the design, testing, debugging and so on having already
been done, implementing off-the-shelf software means that the product
will be supported by the vendor, and they are likely to have either their own
implementation resources, or partners trained up to support implementation. You will still need to exert effort to identify, clean and make available
the required data, as well as providing subject matter expertise regarding
your own processes (so ‘off-the-shelf ’ is a bit of a misnomer), but generally
this is the ‘path of least resistance’ for implementing an AI capability.
The biggest disadvantage of using a software ‘package’ is that its capabilities may not align well enough with your objectives and required functionality. You may want to implement, for example, a system to analyse the
sentiment of your customers’ feedback. One vendor may provide a system
that analyses the data exactly as you would have expected but doesn’t provide any useful reporting tools. Another may be less effective at the analysis
but has a best-in-class reporting suite.
• There are also a number of risks in relying on a software vendor to provide
your AI capability. Because of the excessive hype around AI right now
(remember Chap. 1?) there are quite a few vendors that are little more than
an idea and a website. The challenge is that many of the AI software vendors are start-ups or young companies, so it can be a difficult task to identify those with viable and stable products. Your due diligence should be
extremely thorough so that it filters out those who do not yet have enough
experience or capability. Using AI experts at this selection stage can save
considerable pain later on in the project. (You may, though, wish to deliberately go with a software vendor that does not yet have a fully ­commercialised
system if their technology could potentially provide you with a distinct

Build Versus Buy Versus Platform

119

competitive advantage, or where you want to help them shape their offering, so that it delivers some specific advantage to you, whether that be
functionality, exclusivity or price.)
The commercial models for AI software can be quite varied. The most
common, and simplest, is an annual subscription that includes the necessary licenses as well as the support and maintenance of the software. It will
usually include hosting of the software by the vendor (or an agent of the
vendor) if it is a software-as-a-service model. These AI software packages
are fully formed, with a user interface for training and using the product.
For other types of AI software where the software is accessed through an
API (e.g. to determine the sentiment of a particular passage of text) a ‘payas-­you-go’ model is used—the user pays per API call made. There are usually bands of pricing depending on the volume of calls made.
• AI Platforms are the middle ground between an off-the-shelf-package and
a bespoke build. They are provided by the large tech companies such as
IBM, Google, Microsoft and Amazon, some large outsourcing providers,
such as Infosys and Wipro, as well as specific platform vendors such as
H20, Dataiku and RapidMiner.
In the section on ‘Cloud Computing’ in Chap. 4 I described Amazon’s
‘stack’ of AI services, which is typical of a platform offering. It provides
ready-made, ready-trained algorithms for those with simple and well-­
defined requirements (such as generic text-to-speech translation); untrained
but ready-made algorithms for those that need to add their own specific
practices, nuances and data into the model; and a set of AI tools for
researchers and data scientists to build their algorithms from the ground
up. (The platform provider can also usually provide the infrastructure facilities as well.) Of course, an organisation may want to use a range of these
services for the different capabilities that their solution requires.
The platform approach may make sense if your organisation already has
a relationship with one of the tech giants, especially if there is already a
strong connection with cloud services. You may find, though, that you
have to make some compromises on certain aspects where a platform provider’s specific AI capability may not be as strong as a stand-alone one. Of
course, you can always mix and match between the two types, depending
on how rigid your partnership model is.
Outsourcing providers also provide AI platforms for their clients—for
example Infosys have Nia and Wipro have Holmes. They usually combine
a data platform (e.g. for data analytics), a knowledge platform (to represent
and process knowledge) and an automation platform which brings these
together with RPA. If your outsourcing provider does have a robust AI

120

7 AI Prototyping

platform then it is certainly worth looking at it—you will probably have to
use their resources for the implementation, but it does provide a relatively
simple route assuming that their capabilities match your requirements.
As well as the platforms provided by the tech giants and outsourcing
providers, there are also development platforms available from AI-only
vendors, such as H20, Dataiku and RapidMiner (these are sometimes
referred to as Data Science Platforms). These provide a solid foundation for
developing AI applications with your own resources or using third parties.
They consist of a number of different tools and application building blocks
that can be integrated with each other whilst providing a consistent look
and feel. This approach will give you the greatest flexibility and control out
of all the platform models—the tools are generally more advanced than
those provided by the bigger platforms. You will need to take a more handson approach to the design and implementation though—the main users of
these systems will be the data scientists.
The pricing models vary between the different providers. For commercial developments, some offer a revenue share model where they will charge
an ongoing percentage (usually around 30%) of any subsequent revenue
generated from the application built on their platform. The most common
approach is to charge per API call—this means that every time the application that you have developed uses a specific piece of functionality of the
platform (e.g. converting a piece of text to speech) there is a small charge.
Commonly there is a threshold below which the API calls are free or at a
fixed price.
In the platforms space, there is some overlap between each of the categories I have defined—IBM, for example, could easily fit into all three—but
this does provide a useful framework for selecting which one (or ones) may
be most appropriate for your AI strategy.
• Building bespoke AI applications will provide you with the greatest level
of flexibility and control; however, it is unlikely that you will want to take
this approach across all the opportunities that you have identified. Bespoke
AI development, just as for any bespoke IT development, can be great for
providing you with exactly what you need but can create related issues
around change management and support. Therefore, for the majority of
enterprises, bespoke AI development should be used only when absolutely
necessary, that is, for complex, very large data problems, or when creating
a completely new product or service that requires technological competitive advantage. (It may make sense to use bespoke development for creating
the initial builds—see next section for details—but subsequently moving
onto a vendor or platform approach.)

Creating Your First Build

121

A bespoke build will also require effort to be put into designing and
building the user interface for the AI system. This can be good in that you
can develop something that suits your users perfectly, but can be a challenge to get it just right. Many excellent and clever AI systems have failed
because the user interface simply wasn’t good enough.
For a bespoke development, you will need your own highly capable data
scientists and developers or (more likely) to bring in specialist AI consultancies that already have these resources. The commercial model is usually
based around Time & Materials, although Fixed Price and Risk/Reward
can also be used when the requirements are well defined.
You may want to use a combination of all the above approaches. If your
AI strategy is ambitious and will impact many areas of the business, then
you will probably want to consider the AI platforms as your prime approach,
making use of off-the-shelf packages and bespoke build where appropriate.
If you only want to target one or two AI opportunities, then it may make
more sense to consider just the off-the-shelf solutions. A strategic program
to create a brand-new product or service may require bespoke development
to ensure that it delivers enough competitive advantage.
So, now that you have an idea of the development approach, or approaches,
that you are going to use, it is time to actually build something.

Creating Your First Build
At this point in your AI journey you will have a list of prioritised opportunities, each with an estimate of the value that it could deliver, as well as an idea
of how they could be implemented (build, buy or platform-based). For the
next stage, we need to test some of our assumptions and build some momentum for the program, which means we have to start building something.
The scale of this initial build stage will depend very much on your intent.
There are five generally accepted approaches to this, each of which focuses on
satisfying slightly different objectives. You won’t need to understand the
detailed intricacies of each of these approaches (your internal development
team or external partner will have the necessary knowledge) but it is important to know which is the most appropriate approach for your project so that
you don’t waste valuable resources and time.
There are some overlaps between the different approaches, and you could
legitimately choose specific aspects from each of them. You may also want to
use a number of these approaches in turn, as they each focus on different
objectives to be achieved.

122

7 AI Prototyping

• Proof of Concept (PoC): a PoC is a software build that will be discarded
once it has been able to demonstrate that the concept that is being tested
has been proven. It can be of any scale, but generally focuses on the barest
functionality whilst ignoring any complexities such as security or process
exceptions (this is called ‘keeping to the happy path’). PoCs will not go into
live production, and will use non-live ‘dummy data’. PoCs are usually carried out to test the core assumption (e.g. we can automate this process
using this data set) but are also useful to manage stakeholder expectations—seeing a system in the flesh early on often provides an extra boost of
momentum for the project as a whole.
• Prototyping: this is a broad term that covers the building of specific capability in order to test the viability of that capability. Prototypes may focus
on the user interface (usually called horizontal prototypes) or specific functionality or system requirements (vertical prototypes). One project may
have a number of different prototypes, each testing for a different thing. It
is rare for a prototype to resemble the final product. As with PoCs, prototypes are generally discarded once they have done their job, but some can
be incorporated into the final product (this is called ‘evolutionary prototyping’)—it is important to know at the start whether the prototype will be
retained as it will need to be built more robustly than a throwaway one.
• Minimum Viable Product (MVP): the proper definition of an MVP is
one where the build, which contains the minimum level of functionality
for it to be viable, is released to its customers or users. The feedback from
these early adopters is then used to shape and enhance the subsequent
builds. The term MVP has, though, become a catch-all description for any
sort of early build and therefore tends to lose focus as to what it might be
looking to achieve. For a build that is not intended to be released, then a
RAT (see below) is generally the better approach. For a build that will go
live, then a Pilot (again, below) is usually preferred. MVPs can have their
place in the development cycle, but only once their objectives have been
well defined.
• Riskiest Assumption Test (RAT): RATs are a relatively new approach that
focus on testing only the riskiest assumption that requires validation. This
is in contrast to the MVP which is less specific about what it is trying to
achieve. Also unlike MVPs, RATs are not intended to go into production.
A RAT will seek to build the smallest experiment possible to test the riskiest assumption that you have. As with prototypes (RATs are really a specific
type of prototype) a number of different tests are carried out to validate
different assumptions—once the riskiest assumption has been tested and
validated, then the next RAT will focus on the next riskiest assumption and

Understanding Data Training

123

so on. Understanding what the riskiest assumption actually is requires
some detailed thought—you will need to identify all of the assumptions
that you have made so far that need to be true for the AI opportunity to
exist. Ask yourself if these assumptions are based on other assumptions,
then try to identify the root assumptions and then find the riskiest one of
those. Then it is a case of working out how you can test that assumption
with the minimum amount of code. RATs are not the easiest approach to
take but can deliver the best long-term value for your project.
• Pilot: a Pilot, like an MVP, will be used in a production environment.
Unlike an MVP, it should be fully formed and complete. It is considered as
a test because it focuses on automating a process or activity that is relatively
small and low risk to the organisation. Pilots take longer to build and are
more expensive than any of the options above, but have the advantage of
not ‘wasting’ time and effort (although focused prototyping is rarely a
waste of time). The Pilot needs to achieve all of things that are expected of
a PoC and MVP as well as being fully functional; therefore, the candidate
process needs to be chosen well—if it is too complex (e.g. it involves lots of
systems or has many exceptions) then it will take too long (and cost too
much money) to build, and crucial momentum will be lost. A successful
Pilot will provide a huge confidence boost to the project as well as deliver
tangible value.
As with the wider automation strategy, when selecting the most appropriate approach for that initial build always bear in mind what your end goal
is—ask yourself whether you need to build any sort of prototype at all (the
answer is probably yes) but also which approach is going to provide you with
the highest chances of success. Speed of implementation and cost will also be
big factors in the choice of your approach. As a default answer, the RAT is
likely to be the best approach for the majority of cases, but do consider all
possibilities.

Understanding Data Training
One aspect where AI projects are generally trickier than ‘normal’ IT projects
is with the dependency on data, and this challenge is particularly acute during
the prototyping stage.
Although it is possible to cut down on the functionality and usability of the
system being built in order to test assumptions, it is rarely possible to do the
same with the training data. Of course, not all AI systems rely on large data

124

7

AI Prototyping

sets—cognitive reasoning engines and expert systems require knowledge
rather than data, but for machine learning applications in particular, data
availability is paramount.
There will be a minimum amount of data that will produce a viable outcome, and this can sometimes mean that the prototype will have to rely on
nearly all the data available, rather than a small sub-set. If your complete data
set contained, say, 10 million records, then a sample size of 10% may still be
workable, but if your complete data set was, say, 100,000 records then
(depending on what you intend to do with it, of course) a sub-set may not
give you any meaningful results.
The more complex the model you want to create (i.e. one with many different parameters), the more training data you will need. If the data is insufficient, your AI model could suffer from what data scientists call
over-fitting—this is where the algorithm models the fit of the data too well
and doesn’t generalise enough (it ‘memorises’ the training data rather than
‘learning’ to generalise from the trend). To understand whether your model is
exhibiting over-fitting and correct for it there are various statistical methods,
such as regularisation and cross-validation, that can be exploited, but usually
at the expense of efficacy.
A good portion of your training data must also be used for testing. This
validation set should, like the training set, be ‘known’; that is, you should
know what output you are expecting from it (‘this is a picture of a cat’) but
you don’t reveal this to the algorithm. Because AI outputs are probabilistic the
answers won’t always be correct, so you need to understand and agree what
level of confidence you are comfortable with. There is no hard and fast rule
about how much of your training data should be retained to do this testing
but 30% is a good starting point. Factors such as the size of the complete set,
the richness of the data and risks of getting the wrong answer will impact this
figure.
You may not be able to get around the fact that, in order to train and test
your prototype system effectively, you need to use nearly all of the data you
have available. If data availability is going to be a challenge then one or more
RATs, as described in the previous section, may be the best way forward at this
point.
Once you have a good understanding of how much data you will need for
the prototypes, you will need to acquire the data, (probably) clean it and
(maybe) tag it. I can’t stress enough how important the quality of the data is
to the value that the AI system will deliver.
The most likely scenario is that the data will be sourced from within your
own organisation—it may be customer records, transaction logs, image libraries

Understanding Data Training

125

and so on. You may also want to acquire data from external sources such as the
open-source libraries I discussed in the section on big data in Chap. 2. This
data will come ready-tagged, but there may also be publicly available data sets
(such as traffic movements) that will require tagging before they become useful
for training your algorithm.
The cleanliness, or accuracy, of the data will clearly have an impact on the
efficacy of your AI system. The old adage of ‘garbage in, garbage out’ is particularly relevant here. Common sense and system testing will determine
whether your data is of sufficient quality to provide useful results.
For training data that needs tagging, such as images or video, the most
efficient way is through crowd sourcing (see Chap. 4 for how crowd sourcing
works). Crowd sourcing firms, such as CrowdFlower, have many human
agents at hand to complete micro-tasks—each are sent items of the data and
simply tag them with the required information (this is a picture of a cat, this
is a picture of a dog, etc.). The fully tagged data set is then returned to you so
that you can train the model.
One risk with crowd sourcing is that any biases in the humans doing the
tagging will eventually be reflected in the AI model. As a rather silly example,
if the human agents doing the tagging were all very short, and they were asked
to tag pictures of people as to whether they were tall or not, then it may be
that a higher proportion than expected would be tagged as tall. That ‘smallness bias’ would be carried through into the model and skew the results of the
new, unseen data. Most good crowd sourcing firms though will ensure that
bias doesn’t enter the data in the first place, but can also correct for any bias
that may exist.
Another, rather innovative, approach to data acquisition and training is to
use virtual worlds or computer games instead of ‘real life’. I mentioned previously that Microsoft have created a specific version of Minecraft which can be
used to train AI agents, and some software companies are using games like
Grand Theft Auto to help train their image recognition software that is used
in driverless cars. So, rather than having to drive around many different environments for hours and hours filming and then tagging objects, all of it was
done in a fraction of the time in the virtual world of the computer game.
So, with your AI Strategy complete, and your first build underway (whichever way you have decided to approach it), it is time to think about the longer
term, and how you can start to industrialise your new-found AI capability.
But not before getting a good understanding of all the risks that AI can raise,
and how you might mitigate these, and that is the subject of the next
chapter.

126

7

AI Prototyping

The AI Consultant’s View
This is an extract from an interview with Gerard Frith, co-founder of Matter
and Sensai.ai, AI Strategy and Prototyping consultancies.
AB: What got you into the world of AI in the first place?
GF:	I have been fascinated with the mind since I was small child. I read Jung
and Freud in my early teens, Godel Escher Bach when I was 18. I initially
studied Psychology at university, but in my first year it was my Cognitive
Science module that grabbed my attention most. I became obsessed! At
the end of my first year, I decided to start over, and signed up for an AI
degree instead. That was 1991.
Once my degree was done, I found out that the rest of the world was
rather less interested in AI than I was. During the AI winter of the 1990s, I
mostly had to pretend my degree was just an exotic computer science degree.
In 2013, the rest of the world was finally catching on, and I established
Matter AI, one of the first AI consultancies anywhere. I sold Matter in
late 2016.
My major focus at the moment is on the use of AI to create better and
deeper customer relationships.
AB: Why do you think AI is being so talked about right now?
GF:	There are two key reasons: firstly, AI tech has recently become practically
useful. That’s mostly due to the vast amounts of data that the digital age
has exponentially spewed, and to the massive advances in available processing power. The theoretical advances in AI that underpin much of
what is being used in business today were developed years, and sometimes decades, ago.
The second reason is that AI has always been exciting, scary, troubling
and challenging in a way that no other technology is or could be. If you take
away the breathless discussions of how AI will take all our jobs, then you are
left with a pretty appropriate level of coverage of a disruptive technology.
The creation of genuine AI could easily claim to be the most significant scientific achievement in history. It has the potential too to be one of
the most significant social and spiritual achievements.
AB: What value do your customers get out of your services?
GF: My customers say they get three main things from working with us:
• Expert guidance. I’ve been a developer, a CEO, and a management
consultant, so I know the technology, I know the market and I
understand what a company needs to be successful.
• I’m a disruptor. I like to challenge the current equilibrium and bring
new thinking to situations that allows organisations to reinvent
their value propositions and the way they deliver them to market.

The AI Consultant’s View

127

• Because of my breadth of experience, I can provide the glue that
brings together tech and business strategy, and then how to execute
those strategies in effective ways.
AB:	What do customers or potential customers need to focus on if they are
going to get the maximum value from AI?
GF:	The same things they should always focus on when assessing new technology—how can I create greater value for my customers using this technology? What competitive advantage could this give me? The critical
thing to do is to move fast. It’s getting easier for disruptors to change
industry dynamics quickly, and with less capital. But do treat AI as an
early stage technology too. Invest seed capital in experiments, but get
those experiments live quickly and then iterate.
AB: Do you think the current hype is sustainable?
GF:	It’s inherent in human systems that hype is never sustainable. Our biology seeks novelty and in doing so makes cycles of hype and deflation
inevitable. That said, I do believe that AI is worthy of tags like the ‘4th
Industrial Revolution’ in ways that, say, the 2nd and 3rd ones weren’t. In
this sense it’s just a continuation of the impact of technology development in general, but its impact will be quicker, but only in the same way
that the internet’s impact was quicker than that of TV, and the impact of
roads occurred faster than the impact of the printing press.
Beyond its vast (but normal) economic impact though, I think it will
actually change everything for humans. It is not like any other technology
in terms of how it affects how humans will think about themselves.
Currently even professors of neuroscience who argue that the mind is
nothing greater than a complicated computer hold on privately to the
idea that they have free will and a ‘soul’. AI will challenge those ideas in
quite new and scary ways.
AB: How do you see the market developing over the next 5 years or so?
GF:	The next five years will be very exploratory. AI is a general-purpose technology and will thrust out in multiple directions at once. We are seeing
new products appear weekly covering a plethora of use cases, with even
the biggest tech players only covering relatively small areas. Companies
will need to look at gluing together a range of best of breed providers,
often with a strong vertical industry focus.
We’ll undoubtedly see a lot of acquisitions, but only slow consolidation as tech companies focus on land grab. We’ll also begin to see some
industries such as Law begin to be taken over by software. In general,
industrial boundaries will become even more porous as data drives the
discovery of new horizontal processes that can cross the traditional
verticals.

8
What Could Possibly Go Wrong?

Introduction
It is important to provide a very balanced view on the benefits and risks of
AI. Throughout the book, I’ve talked about some of the unique challenges
that need to be faced when implementing AI, but it is worth drawing these
out into a separate chapter, as I have done here, to make sure they get the
attention they need.
In the chapter I’ve tried to cover the ‘localised’ challenges—those that any
company will have to consider when implementing AI—but also the more
general ones that will affect the way that we work and how we will live our lives.

The Challenge of Poor Data
Data quality is traditionally measured by the ‘4 Cs’: Clean, Correct, Consistent
and Complete. But in the world of AI and Big Data we need different rules to
guide us. Data quality can be thought of as a combination of data veracity (its
accuracy) and data fidelity (its quality in the context of how it is being used).
Data veracity is less important in the world of Big Data because any small
errors will be drowned out by the sheer volume of correct data. The model
that is created from the algorithm will look at overall trends and fitness, so a
few outliers will not significantly impact the outcome. (If there are fundamental errors in the data, such as the decimal point being one place out on every
data point, then, of course, it will have a material impact, but in this section
I am only considering small issues in data quality.)
© The Author(s) 2018
A. Burgess, The Executive Guide to Artificial Intelligence,
https://doi.org/10.1007/978-3-319-63820-1_8

129

130

8 What Could Possibly Go Wrong?

With data veracity, whether the data is being used for AI or not, there is
always a balance to be made between accuracy and cost. The trick is to understand what level of accuracy is ‘good enough’. With AI applications, that
threshold is generally lower than for more traditional computing applications
because of the generalised modelling effect.
Data fidelity, on the other hand, can have a serious impact on the results.
This is where the data is inappropriate for the task for which the AI has been
set. In Chap. 5 I wrote about one example of AI being used to predict recidivism of people who have been charged with criminal offences by the Durham
Police Force. I noted that the data used for the predictions was sourced only
from the region itself, and therefore excluded any relevant activity the subject
might have taken part in outside of Durham County. So, a career criminal
from Leeds (just down the A1 main road from Durham) who had just committed his first offence in Durham may have been allowed to walk out of the
police station because he was considered a low risk. In this case, the data that
the AI system is using is correct but not wholly appropriate.
Another, slightly different, reason that the source data would not be appropriate is if it contained unintended bias. This is a specific problem with AI
systems and is covered in detail in a later section in this chapter.
Related to data quality is the question of whether more data always means
better results. Generally, with machine learning applications this is true; however, there are some subtleties to consider. There are two, almost conflicting,
perspectives.
The first, called ‘high variance’, is where the model is too complex for the
amount of data, for example where there are many features compared to the
size of the data. This can lead to what is called over-fitting and would cause
spurious results to be returned. In this case, the data scientist could reduce the
number of features, but the best solution, and the one that seems most like
common sense, is to give the model more data.
But for the opposite case, where the model is too simple to provide meaningful results, called ‘high bias’, adding more data to the training set will not
deliver any improvement to the outcome. In this case adding more features,
or making the available data a higher quality, both in veracity and in fidelity
(such as data cleansing and outlier removal) will be the most effective
approaches to improving the outcome.
Generally, it is the role of data scientists to ensure that the data is big
enough, relevant and appropriate for the problem at hand. Although my
police example above is relatively simple, some aspects of data science can
seem more like an art form than a science. In most cases there is no clear
‘right’ and ‘wrong’ answer—the data scientist must use creativity and ­inference

The Challenge of Poor Data

131

to identify the best sources and mix of data. This particular skill set is one of
the reasons that the best data scientists can demand such high fees at the
moment.
Once poor data quality has been identified as an issue then there are a number of approaches to try and fix (or enhance) it. Some of these could be the
traditional methods of data cleansing; crunching through the data to identify
anomalies to constraints and then correcting them using a workflow system.
Crowd Sourcing, which I described in Chap. 4, can also be used as a cost-­
efficient way to cleanse or enhance data sets, particularly with regard to ensuring labels are correct in training data sets. Tasks, usually split down into small
units, are assigned to a distributed team of humans, who will process each one
in turn.
RPA can also be a useful tool to help cleanse data where large enterprise
workflow systems are not available. The robots will need to be configured to
identify all the constraints they need to look for. Once they have found the
data errors they can then validate those with other systems to determine what
the correct answer should be, and then make that correction. RPA can be a
good solution for ‘one off’ cleansing exercises because it is relatively easy to
configure. It can, of course, be used in combination with crowd sourcing.
And, with only a slight element of irony, AI systems themselves can be used
to cleanse data for AI systems. Some vendors now have solutions which can
carry out Field Standardisation (e.g. making two different data sets consistent), Ontology Mapping (e.g. extracting product characteristics),
De-duplication (e.g. removing entries with the same content) and Content
Consistency (e.g. identify and matching abbreviation forms). Generally, they
work by using machine learning to analyse the structure of the data model to
then determine the kind of errors such a model is likely to generate.
Another area where AI could be used to improve the quality of data, and
even generate new data, is in the use of Generative Adversarial Networks
(GANs). This is where one AI system judges the outputs from another system.
For example, the originating AI could generate a picture of what it thinks of
as a cat. The second system then tries to work out what the picture is of. The
assessment, i.e. how close is this to looking like a picture of a cat, is then used
to guide the first system to try again and create a better picture. GANs are
relatively new, and this approach is still at the experimental stage, so I will save
a detailed description of how this works to the final chapter.
As should be clear from much of this book, AI (in most cases) feeds off
data—without the data then there is little value that AI can add. But, as we’ve
seen in this section, more data does not always mean better results (especially
in high bias model scenarios). And it’s not just about volume of data but the

132

8 What Could Possibly Go Wrong?

quality. The veracity and fidelity of the data can have a huge impact on the
performance and effectiveness of the outcomes. Any data-dependent AI project must therefore involve careful planning, ensuring the data is the appropriate size and complexity for the problem that is being addressed.

Understanding the Lack of Transparency
The reason Machine Learning is called Machine Learning is, rather obviously,
that it is the machine, or computer, that does the learning. The computer does
all the hard work in ‘programming’ the model that is going to make the predictions. All it needs is the data to be trained on: feed it the data and it will
come up with the model. The human (the data scientist) has to choose the
right algorithms in the first place and make sure the data is appropriate and
clean (see previous section), but the model that is created is essentially the
work of the machine.
I labour this point somewhat because it is an important, inherent aspect of
machine learning that is responsible for one of its biggest drawbacks—the
lack of transparency. I can ask a trained AI system to, for example, approve a
credit lending decision, or to recommend whether a candidate gets short-­
listed for a job, but I can’t easily ask it how it came to that decision. The model
has looked at the many different features of that customer or that candidate
and, based on the many different features of the training set, has determined
the answer to be, in these examples, yes or no. But which features were influential in that decision, and which ones were irrelevant? We will never really
know. It’s a bit like trying to recreate an egg from an omelette.
Which, of course, can be a problem if you need to explain to that candidate
why they weren’t short-listed, or why you didn’t give that customer the loan.
And if you are in a regulated industry, then you will be expected to be able to
provide those answers as a matter of course.
There are three general approaches to tackling AI opaqueness. The first one
is to not use machine learning in the first place. In Chap. 3, in the section
concerning the optimisation capability, I described Expert Systems and
Cognitive Reasoning Systems, which are particular flavours of AI that work
on the principle of a human subject matter expert creating the original model
rather than it being created using data. (For this reason, some AI experts claim
that expert systems should no longer be described as AI.) The ‘knowledge
map’ that is created from these approaches can be interrogated, by a person or
via a chatbot, to access the information. In some of the more advanced
­systems, the queries can be automatically routed through the map using the
specified weightings that have been applied to each node and connector.

Understanding the Lack of Transparency

133

This all means that the decision that comes out at the end can be traced
back through the map, making it fully auditable. So, using one of my examples above, the failed credit application could be traced to, say, 75% of the
decision due to salary, 40% due to postcode, 23% due to age and so on. And,
once it has been designed, assuming the processes don’t change, it will remain
consistent in its decisions.
The challenge in using expert systems is that they can get very complex very
quickly. The more features that the system has, the more complicated the
model becomes to define. If the processes do change, the model then has to
be changed as well to reflect them, and that can be a laborious task.
Another approach for simple problems is to use self-learning decision trees,
usually called Classification and Regression Trees. From a set of input data,
these CART algorithms create a decision tree that can then be interrogated to
understand which features have been the most influential. CART alogorithms
generally provide the best balance between effectiveness and transparency.
For complex systems with lots of data and many features, an ‘ensemble’
approach that calls on many different algorithms to find the most popular
answer (Random Forest is the most common ensemble approach), will prove
the better choice, but it will have that challenge around transparency. The
most common approach to tackling this is to try and reverse-engineer the
decision by changing just one variable at a time. Returning to my customer
who has had his loan rejected, we could repeat his case but changing each of
the features that he has fed into the system (his salary, his postcode, his age,
etc.). This trial-and-error approach will then be able to indicate (but not necessarily isolate) which of the features had the most influence on the decision.
The problem with this approach is that, with many features, this can be a
long process. Carrying out the analysis on a number of test cases and publishing these first is a good way to avoid analysing every rejected case. The ‘model
cases’ could demonstrate, for example, how, with everything else being equal,
age impacts the decision-making process. But you might have to do that for lots
of different model cases since the influence of age might be different in different
postcodes or for different salary levels. And, if the system self-learns as it goes
along, then the test model will need to be refreshed at regular intervals.
The level of testing required will be very dependent on the type of problem
that is being solved. The more sensitive the problem (medical imaging, for
example), the more robust the testing will need to be.
A disgruntled customer who has just been rejected for that loan may, then,
be satisfied with this demonstration of the ‘logic’ used to make that decision.
But would that satisfy the industry regulators? Again, each use case will have
different requirements for testing, but it could be the case that being able to
demonstrate a fully tested system, and the influences that different features

134

8 What Could Possibly Go Wrong?

have, will be enough for some regulators. In the legal sector where e-­Discovery
software is used to automatically process thousands or even millions of documents to identify those that are relevant or privileged in a litigation case, the law
firm using the software must be able to demonstrate a robust use of the solution, with due process applied at every stage. Once the court is satisfied that this
is the case, then the results from the analysis can be used as valid evidence.
Artificial Intelligence is a new and fast developing technology—regulators
are generally slow and considered in their approach, and will need time to catch
up. Putting the risk of fraud to one side, once AI is more generally accepted in
society it may be the case that customers, and therefore regulators, take a more
relaxed approach to the transparency question. Until then, AI-adopting companies need to ensure that they have the answer when the question is raised.

The Challenge of Unintended Bias
A common fallacy about AI systems is that they are inherently unbiased: being
machines they will surely lack the emotional influences that inflict humans in
our decision making that lead to bias, whether intended or not. But that is
just a fallacy. The core of the problem is that the data used to train the AI may
have those biases already baked in. If you use biased data to train an AI, then
that AI will reflect back those same biases. And, because the model can be
opaque to interrogation (as we saw in the previous section), it can be difficult
to spot them.
Here’s a simplified example to demonstrate the problem. In the world of
recruitment, AI can be used to short-list candidates based on their CV (curriculum vitae, or résumé). The AI system would be trained by feeding in many
CVs, each one labelled as either ‘successful’ or ‘unsuccessful’ based on whether
they got the job they were applying for at the time. The AI can then build up
a generalised picture of what a successful CV looks like, which it can then use
to short-list any new candidate CV that is submitted—a close enough match
between the candidate CV and the ‘successful’ model would mean the CV
gets through to the next round.
But how did those candidates in the training set get their jobs? By being
assessed by humans, with all our conscious and subconscious biases. If the
human recruiters had a propensity to reject older candidates (even if they
didn’t realise they were doing it), then that bias would flow through into the
AI model.
So, how to ensure there is no bias in the training set? The first answer is to
try and use as wide a sample as possible. In my example above, you would try
to take training CVs from many different recruiters, ideally those with as

The Challenge of Unintended Bias

135

neutral a mix of gender, race, age and so on as possible (just as polling firms
try to use a representative sample of the population).
This may not always be possible, especially if the data is from your own
company and is all you have. But even publicly available data sets can be
prone to bias. Some of the data sets used for facial recognition training do not
have enough representative samples from people of colour. This means that
any AI trained using that data could struggle to identify non-white faces. As
these biases in public data sets are unearthed, they are being slowly fixed (there
is even an Algorithmic Justice League which highlights bad practices like
these), but AI developers should be particularly cognisant of the issues with
any public data that they use.
In any of the above cases where there could be bias in the data, it will be
necessary to try and test for it. Just as in the previous section, where we looked
at testing for different influences in decision making, we can similarly test for
the influence on any particular feature for bias, as long as we know what the
‘right’, that is, unbiased, answer should be.
In the recruitment example, we would expect an unbiased response to show
that age does not influence the short-listing decision. If we are able to isolate
the age feature in our data set then we can test, by varying only the age in our
model case, whether it is influencing the outcome at all. As with the transparency test, this would need to be done across a range of samples (e.g. just men,
just women) to ensure that age was not influential for any particular group.
Assuming some level of bias has been detected, the next question is what to
do about it. Finding the type of bias (e.g. age, gender) will immediately help
identify any groups of data in the training set that may be causing it (there
may have been a sub-set of the data that was from a particularly young set of
recruiters, for example). This input data can then be excluded or altered to
eliminate the source of that bias.
The model can also be tweaked by changing the relative weightings between
different terms. If the training CVs in my example showed a strong correlation between Manager (role) and Male (gender), the mathematical relationship between gender-neutral words like ‘Manager’ and gendered words such
as ‘Male’ can be adjusted. This process, called de-biasing, is not insubstantial,
and usually involves humans to identify the appropriate and inappropriate
words, but it can be done.
One important thing to bear in mind is that not all bias is necessarily bad.
There may be instances where the inherent bias is important and where the
consequences justify the means, such as in identifying fraudsters. There is
clearly a fine balance between representing the truth, biases and all, and altering data to represent the accepted social position. Also, on a much simpler

136

8 What Could Possibly Go Wrong?

level, the AI will need to understand specific definitions in areas that could be
biased, such as in the difference between a king and a queen, if it is to work
properly.
Eliminating unintended bias from AI is, largely, still a work in progress and
must be considered carefully where there are specific consequences. Using
publicly available data sets does not, currently, ensure neutrality. Using data
from your own organisation will make de-biasing even more important, and
therefore must be built into the AI development roadmap.

Understanding AI’s Naivety
It was the French philosopher Voltaire who famously said that you should
judge a person by the questions they ask rather than the answers they give.
These may be very wise words when it refers to humans, but for AI the situation is even simpler: the machine doesn’t need to know what the question is
in the first place in order to give a respectable answer.
This is the whole idea behind Clustering, which I described in Chap. 3. You
can present a large volume of data to a suitable algorithm and it will find clusters of similar data points. These clusters may depend on many different features, not just a couple of things like salary and propensity to buy quinoa, but,
in some cases, many hundreds. The AI is providing mathematical muscle
beyond the capability of a human brain to find these clusters.
But these clusters are not (or, more accurately, don’t have to be) based on
any pre-determined ideas or questions. The algorithm will just treat the information as lots of numbers to crunch, without a care whether it represents data
about cars, houses, animal or people. But, whilst this naivety of the data is one
of AI’s strengths, it can also be considered a flaw.
For big data clustering solutions, the algorithm may find patterns in data
that correlate but are not causal. In the section on Clustering in Chap. 3 I
gave the rather whimsical example of an AI system finding a correlation
between eye colour and propensity to buy yoghurt. It takes a human to work
out that this is very unlikely to be a meaningful correlation, but the machine
would be naive to that level of insight.
The AI may also find patterns that do not align with social norms or expectations—these usually centre around issues such as race and gender. I’ve
already written in the previous section on the challenges of unintended bias,
but in this case an awkward correlation of purely factual data may naively be
exposed by the algorithm. The challenge for those responsible for that algorithm is whether this is a coincidence or there is actually a causality that has

Understanding AI’s Naivety

137

to be faced up to. How that is handled will have to be judged on a case-by-case
basis, and with plenty of sensitivity.
There is also the infamous example of the Microsoft tweetbot (automated
Twitter account) that turned into a porn-loving racist. It was originally
intended that Tay, as they called the bot, would act through tweets as a ‘carefree teenager’ learning how to behave through interactions with other Twitter
users. But it quickly turned nasty as the human users fed it racist and pornographic lines which it then learned from, and duly repeated back to other
users. Tay, as a naive AI, simply assumed that this was ‘normal’ behaviour. It
only took a few hours of interaction before Microsoft were forced to take the
embarrassing tweetbot offline.
One useful way of thinking about the naivety of AI is to consider how dogs
learn. Like all other dogs, my own, Benji, loves going for a walk. I know this
because he gets excited at the first signs that a walk might be imminent. These
include things like me locking the back door and putting my shoes on. Now,
Benji has no idea what the concepts of ‘locking the back door’ or ‘putting my
shoes on’ are, but he does know that when these two events happen in close
succession then there is a high probability of me taking him for a walk. In
other words, he is completely naive to what the preceding events mean—they
are just data points to him—but he can correlate them into a probable
outcome.
(This dog/AI analogy is quite useful and can be extended further: my dog
is quite lazy, so if he sees me lock the back door but then put my running shoes
on, he goes and hides to make sure I don’t take him with me. In this scenario,
he is using increased granularity to calculate the outcome this time—it’s not
just ‘shoes’ but ‘type of shoes’. Of course, he doesn’t know that my running
shoes are specially designed for running, just that they are different enough
from my walking shoes. It may be the different colour/shade, a different smell,
the different place where they are kept and so on. This demonstrates the
opaqueness issue I discuss in a previous section: I have no real idea (unless I
do some pretty thorough controlled testing) what aspect of the shoes switches
the outcome from ‘Excellent, I’m going for a walk’ to ‘Hide, he’s going for a
run’, but it clearly does have a binary impact. I should also point out that the
dog/AI analogy also has its limitations: Benji has lots of other basic cognitive
skills, such as knowing when it is time for his dinner without being able to tell
the time, but because AIs are currently very specialised in their capabilities, an
AI that predicted walks would not be able to predict dinner time.)
So, the naivety of AI systems can be a real headache for its users. Suffice it
to say that the outcomes from the clustering must be used carefully and wisely
if they are to yield their full value. Data scientists and AI developers must be

138

8 What Could Possibly Go Wrong?

aware of the consequences of their creations, and must apply heaps of common sense to the outputs to make sure that they make sense in the context for
which they were intended.

Becoming Over-Dependent on AI
As you will have seen from the many examples provided throughout this
book, AI can achieve some remarkable things, particularly when it is carrying
out tasks that would be impossible for humans to do, such as detecting clusters of like-minded customers (or fraudsters) in databases containing many
millions of data points.
The problem may come when companies become overly dependent on
these systems to run their business. If the only way to identify your best customers or people trying to defraud you is through very complex AI algorithms,
then there will always be a risk that these systems actually stop working effectively or, probably worse, stop working effectively without you realising it.
This all comes down to the complexity of the problems that are being
solved and therefore the minimal understanding of how they are really working. I have already covered the lack of transparency of how an AI may come
to a decision, but for very complex algorithms (and one AI solution usually
contains a number of different types of algorithms all strung together) there
will only be a few people—the AI developers and data scientists—that understand how these have been designed and built in the first place. Being dependent on a few (very in-demand) people is clearly a risk for any business. It’s
rather like the world of traditional IT where a core legacy system has been
hand-coded in an obscure language by one developer who could leave the
company, or get run over by the proverbial bus, at any time. The risk can be
compounded further if the AI system is being built by a third party who will
not necessarily have that long-term commitment to their client that an
employee might.
There are mitigating actions though. Most of them are similar to the
approach one would take with the bespoke legacy system coder—make sure
that they have documented everything that they do, and are incentivised to
stick around to help support the system. If they won’t be around for long,
make sure there is a robust succession plan in place.
For a third party, there can be more pressure applied, through contractual
obligations, to get the solution fully documented, but the most important
thing will be to select the right provider in the first place, which I discuss in
more detail in Chap. 9. In that chapter I also write about setting up a Centre

Choosing the Wrong Technology

139

of Excellence (CoE) that will be the guardian of the necessary skills and documentation to maintain and improve the AI solutions in the future.
Another factor to bear in mind when trying to solve complex problems is
that it is difficult to know whether the system is providing a correct, or reasonable, answer. For some AI solutions, such as measuring sentiment analysis, we
can test the output of the machine with what a human would reasonably say
was the correct answer; for example, the machine said the sentence ‘I was very
dissatisfied with the service I received’ would be predominantly ‘negative’ and
a human would agree with this. If there was a high volume of sentences to
assess (as would usually be the case) then a sample can be tested. But for complex problems, such as trading financial instruments or designing drugs, it is
nearly impossible to tell if the machine is making a correct decision. We can
look at the final result (the end-of-day trading position, or the efficacy of the
drug) but we will not be able to determine if this was the best outcome of
all—perhaps more profit could have been made or a better drug could have
been designed.
There is also a more philosophical risk from the over-dependency on AI: as
AI becomes more and more commonplace in our lives, we will eventually lose
the ability to do even the simplest cognitive tasks because we are no longer
practising those skills. Our ability to remember people’s names and ‘phone
numbers, and our ability to read maps are already being eroded by our dependency on smartphones and satnavs.
As AI capabilities become greater, they are bound to impact more of our
cognitive skills. Some people may argue that this is not necessarily a bad thing,
but we have developed all those skills over many millennia for particular reasons and, absent of the technology to help us, we quickly become exposed and
vulnerable.
But, coming back to a more practical basis, AI skills will, over time, become
more commoditised (just as HTML development skills have in the last
20 years) and the problem of dependency on highly skilled AI developers and
data scientists will gradually recede. But for now, the risk of over-dependency
should be built into the AI strategy as soon as it is clear that the solutions will
go beyond an off-the-shelf or simple platform approach.

Choosing the Wrong Technology
The field of AI is fast moving—what was not possible in one year can be
solved in the next. All the drivers for AI that I described in Chap. 2 are all
continuing to get better and more influential—big data is getting bigger,

140

8 What Could Possibly Go Wrong?

storage is getting cheaper, processors are getting faster and connectivity
between devices is now almost infinite. So, how do you choose which
approach and technology to take today, if tomorrow something better could
come along? (Users of Apple products, by the way, face this dilemma every
time they think they want to upgrade anything.)
One of AI’s constraints, that each application can only do one thing well,
is actually an advantage here. If an AI solution has been built up from a number of different capabilities (as per the AI Framework) then each of those
individual capabilities could be swapped out with a newer, better approach if
one comes along. Replacing one chatbot solution with another will not necessarily impact any Optimisation capability you have already built. You will still
have to train the new piece of AI but you won’t necessarily have to throw away
the whole solution.
Say, for instance, that you are using one of the AI platforms exclusively to
provide all your required AI capabilities. If one of the other platform providers brings out a better AI capability for, say, text-to-speech, then it is not too
big a challenge to switch out the current API connection for the new one.
(Some providers will, of course, try to lock you into their platform through
contractual means, or even just canny sales techniques, so you will need to
watch this as you select your approach and provider.)
Bigger investments will, of course, require greater levels of due diligence.
Putting in a multi-million-pound vendor-based AI solution will certainly
commit you to a roadmap that may be difficult to step off. But that is true for
any large IT investment.
It does become more challenging if there is a fundamentally new approach
available. Established users of Expert Systems will have looked at Machine
Learning when it first came along with some envy. But if we assume, as is
likely, that machine learning (and all its associated approaches like DNNs)
will be the fundamental core technology of AI for a long while, then it should
be a relatively safe bet to base a strategy around. (The only technology that
may have a material impact on machine learning is Quantum Computing,
but this is still very much still in the labs and will take decades to deliver practical day-to-day uses.)
Probably the bigger issue with committing to one particular AI technology
is the skills that you will need to retain to develop and support it. Generally,
individual developers will be specialists in a particular tool or platform; therefore, changing the tools may mean changing the developers. If these resources
are being bought in through a consultancy or implementation partner, then
you will just need to make sure that that firm has capabilities across all the
relevant tools and platforms that you may require (I cover this in more detail

Preparing for Malicious Acts

141

in the Vendor Selection section of Chap. 9). If you have started to build up a
Centre of Excellence based around specific technologies though, you may
have to think hard about whether the change is worth it.

Preparing for Malicious Acts
With great power (as Spiderman’s Uncle Ben famously said) comes great
responsibility. All the capabilities that I have discussed throughout this book
have shown the huge benefits that AI can bring. But that power could also be
used for nefarious means as well.
The ability for AI Clustering to, for example, identify customers that may
buy a certain product, could also be used to identify people who are ideal fraud
targets, especially if it is triangulated with other data sources. Criminal AI systems would usually have access only to publicly available data (which is always
more detailed than you think) but, as has been seen with the numerous largescale hacking attacks, other, more private data may also be available to them.
In early 2017, some banks introduced voice passwords for their online services—all the customer had to do was train the system on their voice and they
could then login by simply saying ‘My voice is my password’. Within months
it could be proved that this approach can be fooled using facsimiles of the
user’s voice—the BBC showed in one example how a presenter’s non-identical
twin could log in to his account—and the worry now is that AI-powered
voice cloning technology will be able to do the same job. The AI will need
voice samples to be trained on, but, again, these are surprisingly common
across social media, and if you are a famous person, then widely available.
There is already a commercial mobile app, CandyVoice, which claims to be
able to clone voices.
It won’t just be banking systems that could be fooled by voice cloning technologies; you and I might think we are listening to our mother on the
­telephone saying she has forgotten the password on her bank account, or our
boss asking us to send specific emails to suppliers or to make certain
payments.
Similarly, Image Recognition can be used to subvert Captchas—these are
the small images that appear when you try and make a payment on certain
websites. You can proceed only if you can type the numbers that are in the
photos, or identify only those pictures with, say, cars in. These are meant to
avoid software robots making unauthorised purchases, but are now being
bypassed by armies of people, or clever AI systems, entering the correct
answers.

142

8

What Could Possibly Go Wrong?

The difference between the type of online criminal behaviour we have been
used to in the past and what we see now is the scalability of the AI. Voice and
image recognition can be done at high volume and at low cost—if just a very
small fraction get through then the criminals will be successful. It’s the same
numbers game that spammers exploit.
AI can also be used to socially engineer people’s behaviours. During recent
political campaigns (Brexit in the United Kingdom is one that comes to mind)
software ‘bots’ were used to influence people’s opinions through targeting their
social media accounts with pertinent messages. Social media companies already
use AI to change the order of the articles in a person’s social media feed so that
the most ‘relevant’ are closest to the top. External companies can use a person’s
online behaviour to predict which way they might vote and then try to reinforce or change that using targeted posts and tweets. Extending this to include
more nefarious acts than voting is within the realms of possibility, especially if
the accounts that are being posted from are pretending to be someone else.
Artificial Intelligence systems can also aid malicious behaviour by being
trained to do the wrong thing. ‘Dirty data’ could secretly be fed into a training
set to change the learnt behaviours of the AI. The large search engines are
constantly having to defend themselves against malicious companies who create fake sites to increase their ranking in the search results.
In the earlier section on the issue of naivety in AI systems, I wrote about
Microsoft’s tweetbot, Tay. Tay didn’t understand the nastiness of the ‘data’ that
it was being fed, but Microsoft clearly didn’t expect the deliberate sabotage of
its creation by mischievous users. Although Tay is a slightly humorous example (unless you work for Microsoft, of course), it does show how AI can very
quickly be influenced by dirty data. Extrapolate this to a situation where the
AI is controlling millions of interactions with customers, or even financial
transactions, and you can start to see the potential risks from this.
The fixes for the types of behaviours I have described above are diverse, and
usually very technical (apart from insuring against loss, which should be done
as a matter of course). Ensuring that your voice recognition application cannot be a victim of voice cloning will be an inherent part of its development.
With Captchas, they need to be updated on a regular basis to ensure that the
latest technology is being deployed. To stop dirty data getting in requires the
usual defensive approaches to stop anyone from infiltrating your systems. The
level of defences you put up will depend on the sensitivity of the data and the
risk of anything going wrong.

 

Conclusion

143

Social engineering is more difficult to ameliorate. It is necessary to swim
against an incoming tide of more and more people who want to use, and generally trust, social media channels. The mitigating actions are awareness and
education. Just as young people (mostly) know that someone at the other end
of a messaging conversation may not be who they claim to be, people also
need to be aware that the chatbot at the other end of their chat conversation
may not be who, or what, it claims to be.
As with all online activities, being one step ahead of the criminals is the best
that can be hoped for. AI has the ability to do very good and useful things but,
as we have seen in this section, also has the potential to do bad things at scale.
Simply being aware of what can go wrong is the first step in combating it.

Conclusion
For readers hoping for a book all about the benefits and value of AI, this chapter may have come as a bit of a shock. But, as I said at the start, it would have
been remiss of me not to include details of how AI can also present risks to
businesses, by being either misappropriated or sabotaged.
I deliberately included some of the worst cases of those risks, as awareness
of what can possibly go wrong will be the most important part of being able
to mitigate them. Some of these defensive steps are part of the normal IT
development cycle, some are unique to AI development and some are more
sociological. For big, complex AI projects, you will be dependent on strategists, developers, security experts, data scientists and sociologists to ensure
that your application is not open to abuse or unnecessary risk.
In Chap. 10, the final one, I will cover some of the more philosophical
debate around the impact of AI, including how it affects jobs, and the big
question of what happens when (if ever) AI gets smarter than us.
But, before that, we need to look at how your AI efforts and projects can be
industrialised and embedded into your business so that they provide sustainable benefits for the long term.

144

8 What Could Possibly Go Wrong?

The AI Ethicist’s View
This is an extract from an interview with Daniel Hulme, CEO and Founder
of Satalia, a London-based AI development firm with a strong ethical
mission.
AB: Daniel, first tell me about Satalia, and how you approach AI.
DH:	In this capitalistic world right now, businesses are naturally looking at AI
to either make more money or to reduce costs. Satalia helps companies do
this by providing full-stack AI consultancy and solutions, but our higher
purpose is to ‘enable everyone to do the work they love’. We think we can
achieve that by building a completely new ‘operating system for society’
that harmonises technology with philosophy and psychology. We are taking the core learnings from the solutions we have developed for our clients and then making them available as blueprints and tools for people to
use to make their businesses and lives better. Satalia is also a global voice
for purposeful AI start-ups, of which there are not enough of in this world
right now.
Within Satalia itself, our staff are free to do what they want—they set
their own salary, working hours and vacation days, and they have no
KPIs. We combine AI and organisational psychology to enable our
employees to work on the projects they want to, unburdening them
from bureaucracy, management and administration. This gives them the
freedom to rapidly innovate in ways that they would rarely find anywhere else.
AB:	Can you talk to me more about how you see the ethics of artificial intelligence playing out?
DH:	There are three perspectives of AI that need to be considered differently if
we are talking about ethics.
The first is where decisions are made through a trained but static
model. The ethical question is who is responsible if that model doesn’t
behave according to our norms—if it is racist or sexist for example. This
is the Unintended Bias problem. This is linked to the big challenge of
building explainable a­lgorithms—how do we have transparency from
what is essentially a black box model? Now, if governments legislate for
these things, will that stifle innovation? Or to stay competitive should
companies be allowed to ‘move fast and break things’ as the current trend
goes? AI developers have a real responsibility in the middle of all of this,
and breaking things without careful consideration of the impact is not
the best way forward.

The AI Ethicist’s View

145

Secondly, there are the more general AI solutions where multiple types
of AI are combined into an ever-adapting system that changes its model
of the world when in a production environment. The most famous example of this is the Trolley Problem that designers of driverless cars are facing. Should the crashing car hit the child or the three adults, if there is a
choice? What if the car adapted its model in an unpredictable way ‘deciding’ to hit as many people as possible? You might actually end up with a
Preferences setting in your car that allows you to set these choices ahead
of time.
Here’s another example: in a burning building, there is a chance to
save either a baby or a suitcase full of £1 million. Most people’s instinctive
reaction is to save the baby, but if you think about it, perhaps the best
action for society would be the suitcase of money, as you could probably
save many more babies’ lives using that money. What is the right decision
for society? We’re at an intriguing time in humanity whereby we are having to start to codify our basic ethics.
Liability is also a big challenge with these adaptive systems. If a CEO
decides that his company should make an AI-powered machine that
administers medication to people, who is liable if that machines makes a
bad medication decision because it has learnt the wrong thing since it left
the factory? It’s almost impossible to predict how these algorithms will
behave. Just look at the Flash Crash that hit the financial market in 2010.
The third type of AI to consider is where the machines become more
intelligent than humans—this is the Singularity. I don’t believe that we
can build ethics into a super-intelligent machine (this is the Control
Problem)—they will have to be learnt. You could, perhaps, exploit game
theory to help it learn some ethics. Most people know of the Prisoner’s
Dilemma—the winning strategy is always one of ‘tit for tat’, that is, follow
what your opponent did previously. This is the sort of ethical approach
that a super-intelligent machine could learn, that has been embedded
into humanity over thousands of years: “do unto others as you would
have them do unto you.”
Obviously, the big problem with super-intelligent machines is that
they may not be concerned with the impact on humans for whatever
plans they have, and this is perhaps one of our greatest existential risks.
One thought I’ve had is that we should perhaps help them decide to
depart our planet and leave us in peace, a bit like how we might have
made an offering to the gods in ancient times in the hope they’d have
mercy on us.

146

8 What Could Possibly Go Wrong?

AB:	That’s a lot of different ethical issues to consider, especially if the future of
the planet is at stake.
DH:	We may not see the Singularity come along anytime soon but we absolutely have to think about the impact of an AI-driven capitalistic society,
and to work out—both individually and collectively—how we can create
a radically better society today.

9
Industrialising AI

Introduction
This book is predominately about how to introduce AI into your business, how
to take those first steps in understanding and then implementing prototype AI
capabilities. But at some point, you are going to want to move beyond that,
especially if you think that AI has the potential to transform your business.
So far you should have a good grasp of what the different AI capabilities
are, where they are being used effectively in other businesses, the best way to
create an AI strategy and start building AI prototypes, and being aware of
some of the risks involved.
This chapter then is about how to industrialise those AI learnings and capabilities within your business. It focuses on how you move from a few projects
to a mature capability that can implement most, if not all, of the AI requirements that your business will need in the future.
The AI strategy that was discussed in some detail in Chap. 6 will have given
you an excellent starting point. It should include your AI Ambitions which
will determine how far down the industrialisation road you want and need to
go. I have assumed in this chapter that you will want to at least create a permanent AI capability within your organisation, one that is able to be a focal
point for AI activity, and act as a catalyst for ideas and opportunities. If your
ambitions are more or less than this, then just dial up or down my recommendations appropriately.
You will also recall from the Change Management section in Chap. 6 that
there is often a ‘slump’ after the initial success of a prototype. Now is the time
to ‘go large’ and ensure that the momentum from those early wins is t­ ranslated
© The Author(s) 2018
A. Burgess, The Executive Guide to Artificial Intelligence,
https://doi.org/10.1007/978-3-319-63820-1_9

147

148

9

Industrialising AI

into continued success. Having a robust plan for how that will happen will go
a long way to achieving that. This chapter will give you the foundations for
that plan.

Building an AI Eco-System
You may already have been working with a number of third parties to help
you build a strategy and the initial AI project builds. Some of them you may
want to continue working with on a longer-term basis; others may have just
been useful for a specific task and are no longer required. If you are going to
build a long-term AI capability though, you will need to start thinking about
how much of it will be supported by third parties, and which pieces you
intend to build up internally.
A useful way to think of this is as an ‘Automation Eco-System’. This may
include software vendors, strategists, implementation partners, change management experts and support teams, some of whom will be internal and some
of whom will be external. Some may have a bigger part to play in the early
days, whilst others will become more important once some of the projects
have been firmly established.
If you are planning to use a third party to provide any of the Eco-System
capabilities, be aware that some may be able to provide more than one; an
implementation partner, for example, may also have change management
capabilities. Whether you use both those capabilities from that provider will,
of course, be up to you. I cover some of the subtleties of this in the section on
Vendor Selection later in this chapter.
An Automation Eco-System may end up looking something like this
(Fig. 9.1):
Strategy
Prototyping

Implementation
AI Vendor
AI Vendor

RPA Vendor
Support

AI Vendor

Assoc. Tech. Vendor

AI Vendor

Assoc. Tech. Vendor

Change Management
Fig. 9.1 AI Eco-System

Building an AI Eco-System

149

Each of the capabilities can be provided internally or externally. Apart from
the software vendors, the roles can cover AI specifically, or Automation
(including, say, RPA) generally. I will cover each of these roles in turn.
Strategy This area concerns the sorts of things you are reading about in this
book, particularly those in Chap. 6. These are the activities that I, as an automation advisor, carry out for my clients. It involves building the Automation
Strategy, which will include the Maturity Matrix, the Heat Map, the Business
Case and the Roadmap. It will likely involve the technical approach (vendor
versus platform versus bespoke) and, connected to that, the approach to this
Eco-System. It can also include supporting the selection of the software vendors and the other service providers. Most importantly, it should provide
thought leadership to the organisation.
As the AI market is very complex but relatively young there are very few
advisors that are able to cover the business strategy aspects as well as the technical approaches. Some of these (including myself ) are able to provide prototyping services as well. The connection between these two capabilities is pretty
tight; therefore, it may make sense to cover both of these with one provider.
Although the bulk of the strategy work is carried out up-front, it can be
beneficial to retain the strategy advisor longer term to ensure that the benefits
are realised and the internal capability can be realised. Most clients who wish
to pursue a predominantly internal capability (with little dependency on
external parties) keep a light-touch external strategic involvement to provide
appropriate checks and balances.
Prototyping The building of the prototypes, pilots and PoCs requires a flexible, agile approach, and follows on closely from the output of the automation
strategy (and, in some cases, can form part of the automation strategy). There
are not (at the time of writing) many consultancies specifically focused on
prototyping—this capability is generally provided by either a few of the strategy advisors, or by the implementation providers.
The activities are those that are largely covered in Chap. 7 of this book,
including the build of PoCs, RATs, MVPs and Pilots. If your company is just
starting its AI journey then it makes sense to buy in this capability until you
have enough experienced and skilled resources in-house.
AI Vendor The AI Vendor could include a packaged software provider, a platform provider or a vendor of some of the various development tools that are used
to create bespoke AI builds, depending on the technical approach being taken.

150

9 Industrialising AI

For some of the platform providers, and maybe some of the development tools,
you may have existing relationships in place. No matter what arrangement
though, the software obviously sits at the core of your Automation Eco-System,
and many of the other members of that eco-system will be dependent on those
choices.
It should also be noted that some of the packaged software vendors can
provide their own implementation resources. This is usually the case for those
that have yet developed their software to be ‘implementation partner ready’.
One or two of the platform providers, such as IBM, also have large implementation service capabilities, although other third-party providers can also be
used with their platforms.
It will be likely that you will end up with a number of different software
vendor relationships, irrespective of which technical approach you take. Even
with a platform-based strategy, it is probable that you will need to bring in
other vendors to supplement any gaps in the platform capabilities.
Implementation The Implementation Partners are usually brought in following successful prototyping (unless they are carrying out those activities as
well). They are generally concerned with the industrialisation of the AI capability and can provide a full range of development resources. The Big Four
global consultancies can provide these sorts of services, as well as a number of
their smaller competitors.
Whether you need an Implementation Partner at all will depend on how
keen you are to build an internal capability, and how quickly you want to get
there. A Strategy Advisor and a Prototyping Partner can get you a long way
towards self-sufficiency, but if you want to move fast (or if you are not interested in building an internal capability at all) and have deep pockets then an
Implementation Partner makes sense.
An Implementation Partner will be able to continue the work started during the Strategy and Prototyping stages, and help build and ‘bed in’ those
applications. Depending on their capabilities they may be able to build further applications as well as deliver some or all the Change Management
requirements.
RPA and Associated Technologies Vendor As I detailed in Chap. 4, AI is
rarely the only technology you will need to deploy to realise the full benefits
available. You may also need to bring in an RPA vendor, a Cloud vendor, a
Workflow vendor, and a Crowd Sourcing vendor, depending on the complexity

Building an AI Eco-System

151

of your solution and what existing relationships you already have (many
enterprises looking at AI will already have a Cloud capability, for example).
Some vendors will work better together, either technically or culturally, so
do consider this if you have a choice. Build these eco-system members around
your AI vendors, looking at who they have worked well with before and
whether there may be any particular integration challenges or not.
Your RPA vendor, should you require one, will be an important choice.
Some AI vendors have strategic relationships with RPA vendors, and some
Implementation Partners may also have ‘preferred’ RPA vendors that they
work with. In a very few examples (to date) there are software vendors that
can provide both RPA and AI capabilities, with some of these also offering
Crowd Sourcing as well. Generally though, most enterprises bring in separate
vendors for these crucial capabilities.
Change Management This is an important capability for your eco-system
and should be considered early on in the lifecycle. Many large organisations
will be able to provide a general change management capability internally, but
the specific characteristics of an AI project should be borne in mind to determine whether this will be the most appropriate approach.
The capability can be provided by a generalist third-party provider but
those that can claim to have experience in the unique characteristics of an AI
implementation will usually be preferable (the exceptions being when you
have an existing, trusted provider or you need one with experience in your
specific industry sector).
Support With this I mean the ongoing management of the developed AI
solutions. That could mean ensuring that the applications are available
through management of the underlying infrastructure and networks, or
ensuring that the application is providing accurate and meaningful results. It
could also mean the debugging of any bespoke or procured software. Each of
these activities will generally be done by different parties and will depend
heavily on the technical approach that you take, any existing IT policies and
the complexity of the developed solutions.
For infrastructure support, you may have passed the responsibility to an
outsourcing provider or a Cloud provider, but if you have decided to keep it
in-house your IT team may have to support some unfamiliar technologies,
including the management of high-powered GPUs and large storage arrays.

152

9 Industrialising AI

Management of the AI applications (apart from actual code errors) could
be carried out in-house if you have built up the necessary capability, probably
in an Automation Centre of Excellence, which I describe later in this chapter.
It may well involve recruiting AI developers and data scientists, neither of
which are readily available or cheap at the moment. There are very few providers, though, that specialise in supporting AI applications—they tend to be
either the Prototyping firms or the Implementation providers. The skills will
need to include the ability to understand the business requirements, the data
and the technology, hence the challenge in finding the perfect provider, if that
is the route you are taking.
If the procured or bespoke-build software is faulty, then it will usually be
down to the software vendor to fix it—this is the classic ‘third line’ support and
should be included in any contracts between your organisation and the vendor.
I have not included Sourcing Advisors in this eco-system. They are likely to
be a short-term requirement at the start of the journey, and therefore would
not form part of an on-going capability. I do talk more about the role of
Sourcing Advisors in the next section though.
Whether you end up creating an eco-system with these elements, or you
break it into smaller chunks, it is important to consider all the aspects I have
mentioned above. For those elements that you plan to buy-in, you will need
to have a robust selection process, which I cover in the next section.

Selecting the Right AI Vendor
To a large extent, selecting AI software vendors and service providers will follow the same rules as with any IT vendor or provider, but there are some key
differences with AI that can present some additional challenges as well as
opportunities.
One of the first questions that will need to be answered is what exactly is
being procured. For software, the answer to this will depend on the technical
approach that has been decided upon—off-the-shelf, platform or bespoke
build (or any combination of these). There is likely to be one piece of software
that is at the centre of most things, particularly in the early stages, and this
should be the initial focus. This could be the off-the-shelf software capability
that is required for the pilot, or the platform that the bulk of the applications
will be built out from.
Your organisation may be constrained in what software it can choose. This
may be due to the IT strategy stipulating particular criteria (e.g. it must operate as-a-service), specific standards it wants to adhere to (e.g. no JavaScript) or

Selecting the Right AI Vendor

153

general procurement rules (e.g. third-party company must be at least two
years old). Beyond these constraints though, it is important to consider the
following when selecting AI software vendors:
• Proof of capability—because there is so much hype in the market many
vendors will exaggerate their AI credentials in order to attract attention to
themselves. Some will have only a minimal AI capability as part of their
overall solution (maybe a bit of simple NLU embedded in it somewhere)
but claim it all to be ‘powered by AI’.
Now, of course, the solution you actually require may not necessarily
need full-on AI, but if you have a specific application in mind then you will
need to make sure you can cut through the hype. The understanding of AI
capabilities that you will have gained from this book will hopefully go a
long way to helping achieve that, but you can also call on external advice as
well if need be.
An important step in being able to understand and prove the capability
of the vendor will be to have a comprehensive demonstration of the system
and, if appropriate, a small test. Tests, as with pilots and PoCs, can be tricky
for AI solutions as they generally need a lot of data and training to get them
to work effectively. If the vendor can do something meaningful here
though, do take up the offer.
• Proof of value—you should already have at least an outline business case
by this stage, but it is useful to understand how the vendor measures the
value of their solution. Some will have an approach that allows you to
quickly assess their value whilst others may expect you to do all the hard
work on this (this may, of course, mean that they will struggle to create any
value from their solution, and it should be seen as a potential warning). If
the vendor does have a model that can validate your own business case
assumptions, then do take them up on that approach. The key thing is for
the value that they can demonstrate to align as much as possible with your
own business case.
• References—taking up client references should be an obvious thing to do,
but many enterprises don’t bother (and live to regret it later). The challenge with an immature market like AI means that client references and
case studies will be thinner on the ground. How far you want to pursue
this will depend on the perceived fit of the vendor’s solution to your
requirements, their uniqueness and your appetite for risk. If you want to
bring in a vendor who has no existing clients, then you may want to see
what opportunities there are for you to be a ‘foundation client’ for them—
this is covered below.

154

9

Industrialising AI

• Pricing model—the commercial aspects are obviously very important when
selecting a vendor, but it’s not all about the headline price. In Chap. 6
I described some of the different pricing models that are being used, and you
should explore which ones might be the most relevant to you, and whether
the vendor can provide them. For variable pricing, model different scenarios, especially if it materially impacts the business case—a price per API call
may sound attractive at first, but if your application is very popular, those
escalating costs may start to outweigh the benefits. Also, seriously consider
gain-share or risk-reward models, but make sure that they are based on real
numbers that you can measure easily.
• Cultural fit—although you are just buying some software, it is always
worth finding a vendor company that has a good cultural fit with yours.
The AI project may be a long, and sometimes stressful, journey for both
parties, so you need to know that there is some common ground and purpose between both of you. Cultural fit can be assessed simply from having
a good feeling about it, but it can also be worth trying to do this more
objectively, especially if you want it to form part of your selection criteria.
• Future proofing—ensuring that the software you buy will still be relevant
and working in five years’ time can be difficult to determine, but it should
definitely be considered carefully. In the previous chapter, I talked about
the risks of selecting the wrong technology and those considerations can be
applied here. But it is probably the more traditional elements of the technology that are the ones that are most at risk of going out of date. Ask what
technologies the software is built on, and what technology standards does
it rely on. External expert advice may also be useful here.
• Foundation client—procuring software that has not been used in anger
with any other clients may seem like an unnecessary risk, but there can be
benefits from taking this approach. All other things being equal (you have
tested their capability, the technology is good and future-proofed and there
is a good cultural fit) then, versus a more established vendor, a new vendor
may be willing to offer you some significant benefits in return for taking
the risk of being their first client. Usually the benefits amount to heavy
discounts, especially if you also agree to promote your use of the software.
But it can also mean that you get the opportunity to shape the development of the solution, and make it more closely match your own requirements. This will also ensure that you get the undivided attention of the
vendor’s best people, and that can count for an awful lot in delivering a
successful implementation.

Selecting the Right AI Service Provider

155

• Technical requirements—it should go without saying that the technology
needs to fit into your current environment and align with your IT standards and strategy. But with AI you will also need to consider the data. Is
your data the right sort of data for the vendor’s solution, is it of good
enough quality and quantity? These are key questions that must be satisfied
early on in the selection process.
• Professional services requirements—the final consideration is how the
vendor’s software will be implemented. What professional services are
required, and who can provide them (the vendor, third parties or your own
organisation)? If it requires a third party, then understand how many providers out there can do it, and which ones are most familiar with the specific vendor’s software. Some vendors may have a partnership program
where they will have certified or approved service providers. If you are
already developing your eco-system, then you will need to determine how
much commonality there is between these. You will also obviously need to
evaluate the cost of the services at this stage. The selection of service providers is covered in the next section.

Selecting the Right AI Service Provider
In the earlier section concerning the Automation Eco-System I also talked
about the type of service providers that may be required, including strategy,
change management and implementation consultancies—these will also
have their unique AI-related challenges and opportunities when procuring
them in.
Each of the services capabilities you require should be considered as separate ‘work packages’, which can be procured individually or in groups. From
the Eco-System considerations, you should have a starting point for which
capabilities need to be bought in, and how they might be grouped together.
For example, you may think that you need to buy-in Strategic Advisory,
Implementation Services and Change Management, but you think that you
will be able to handle the Support Services in-house, and you also think that
a single provider could do Implementation and Change Management. In this
example then you would be looking for two providers: one for the Strategic
Advisory work package and one to handle the Implementation and Change
Management work packages.
During the selection process (or processes) these initial assumptions should
be tested, but they do provide a useful starting point. You should therefore
provide the flexibility in your Request for Proposal (or whichever procurement

156

9 Industrialising AI

method you are using) for providers to bid for each work package separately
and combined. You may find as you go through the procurement process that
one provider is really good at Implementation, while another is better at
Change Management—in this case you can always adjust your approach to
instead procure two separate providers for these work packages.
You should also test your in-house capability (if you have one) against the
external providers if you are not sure whether to ‘buy or build’. You can treat
them as a bidder in the procurement process and even price up the cost of
their services.
Another important consideration when looking to identify potential providers is your incumbent or strategic providers. Many large organisations have
preferred suppliers and you may have to align with these. AI is such a specialised technology, though, that (unless your incumbents are actually AI specialists) you should argue strongly to bring in providers with the necessary skills
and experience in AI.
The strategic advisory role should be your first procurement consideration,
as many of the other decisions will be based on the Automation Strategy that
is developed with your advisor. Robust, demonstrable methodologies, independence and (maybe most crucially) cultural fit will likely be strong factors
in your selection here. For the procurement of the remaining work packages,
I think the following are the key AI-specific considerations, whether the
resources come from an established provider, software vendors or contractors
(I have used the words ‘service provider’ below to cover these):
• Experience—it may be difficult for a lot of providers to demonstrate a real
depth of experience in implementing AI, simply because the market is so
young, so you will need to bear this in mind. You will need to look for
capability in the particular approaches and tools that you have decided
upon and any relevant partnerships with software vendors (see below).
There will be some experience that is technically specific (skills in a certain
tool, for example) and others that is more industry sector specific (e.g. ability to understand customer data for telecoms firms), and there will realistically be a compromise that has to be made between these two (unless you
are lucky enough to find the perfect match). It may be that you have the
data knowledge already in your organisation.
• Partnerships and Independence—for the tools and approaches that you
have selected, you will want to look for a service provider that has a strong
relationship with these. Many vendors have ‘partnerships’ with providers,
which can range from a loose connection to an almost symbiotic one—it is
important to understand the level and history of those partnerships.

Selecting the Right AI Service Provider

•

•

•

•

157

In most cases the service provider will be rewarded by the vendor for
selling their software licences. But you may also want your service provider
to be independent of the software vendors, and a balance has to be struck
here. Probably the ideal arrangement is where the service provider has
strong, independent relationships with a number of vendors and will be
happy to recommend and implement any of them. If you can’t find true
independence in the service providers then you will need to rely on your
strategic advisor to provide this.
Cultural Fit—this is often overlooked in a service provider selection process but, in my experience, is the one factor that causes most relationships
to break down. Because your AI projects could be quite fraught at times,
you will want to work with a provider that shares your values and objectives, and has people that get on with your people. Cultural fit can be formally measured during the selection process or it can be assessed informally,
but it should not be ignored.
Approach—most service providers will have methodologies, and these
should, of course, be evaluated and compared carefully. But you should
also try and look ‘under the bonnet’ (or ‘under the hood’) to see how
they may approach different challenges, such as testing, poor data or
biased data. Their answers should tell you a lot about their underlying AI
capabilities.
Pricing Model—as with the software vendors, pricing is going to be
important but so is the pricing model. With service providers, there is the
opportunity to introduce ‘gain share’ or ‘risk/reward’-type approaches
where the provider will share some of the risks but also some of the gains
of the project. This helps align their objectives with yours. The challenge
with these pricing models is getting practical metrics—they need to be
general enough that they relate to business outcomes but specific enough
that they can be measured. The metrics also need to be timely—service
providers will not agree to a target that can be measured only in two years’
time, for example. Your strategic advisor, as an independent resource,
should be able to help you identify and implement practical KPIs.
Knowledge Transfer—the final consideration is how the service provider
will transfer all the relevant knowledge to your internal team. This assumes
that, at some point in the future, you will not want to be dependent on a
third party for your AI capability—some enterprises want to get to this
point as quickly as possible, whilst others are happy to remain dependent
for as long as it takes. The knowledge that will be transferred from the service provider to your teams will include a range of things, and will depend
on the relationship and contract you have with them. It may involve some

158

9 Industrialising AI

licensing of Intellectual Property (IP) from the provider (such as methodologies and tools), but will generally be in the form of work shadowing and
the hands-on experience gained by your team. As your resources get more
skilled, and you bring new employee resources in, the service provider can
start to back away; as you ramp up, they will ramp down.
There is another potential third party that you may wish to engage in.
Sourcing Advisors are experienced procurement specialists that can help you
with both the vendor and service provider selection processes. They will be
able to help create a sourcing strategy, find appropriate companies to evaluate,
build and manage the selection process, and support negotiations.
The over-riding requirement for any sourcing advisor is their independence. Then you should look for capability in AI and in your specific industry
sector. If they do not understand the AI market well enough (it is, after all,
sprawling and dynamic) then you may want to team them up with your strategic advisor (or find someone that can do both).
With both the vendor and the service provider selection there will need to
be a fine balance between a robust selection process and the flexibility that
dealing in a young, dynamic technology requires. Be prepared to change your
plans if things don’t initially work out, or better options come along. As the
cliché goes, be ready to ‘fail fast’. And this means ensuring that the arrangements you have in place with your vendors and providers are as flexible as
possible, whilst still demonstrating your commitment to them. Work closely
with your procurement department to make sure that they understand the
special nature of AI projects and the demands that it may place on their usual
methodologies.
Now that we have an approach for bringing in the necessary third parties,
we need to look at the internal organisation and the skills and people that
might be needed.

Building an AI Organisation
There are a number of things to consider when building an organisational
capability to manage your AI (and other automation) efforts, all of which will
be based on how bold your AI ambitions are. You will remember from Chap. 6,
as part of building an Automation Strategy, I stressed how important it was to
understand your ultimate AI ambitions: do you want to simply ‘tick the AI
box’, to improve some processes, to transform your function or business, or
even create new businesses and service lines? Assuming that you are some-

Building an AI Organisation

159

where in the middle of this range, that is, you want to improve processes and
transform some areas of your business through AI, then you will need to create a team or even a CoE to ensure that you are successful and can extract the
maximum value from your efforts.
In this section I will describe what an AI CoE could look like. Based on the
skills, capabilities and ambitions of your own organisation you should be able
to assess which elements are relevant for you, and therefore what your own
CoE might look like. I will also describe how a CoE can be integrated into an
overall company organisation.
It is important for an Automation CoE to have a mission. This will describe,
and help others understand, its purpose. You will be able to find the words
that are most relevant to your own requirements and objectives of your company, but generally the CoE’s mission should be focused on driving the introduction and adoption of AI technologies. It should act as a central control
point for assessing AI technologies and monitoring progress of ongoing projects. Perhaps most importantly, it should provide leadership, best practices
and support for projects and teams implementing AI solutions within the
business.
Two of the key inputs to determine the scale and structure of your CoE,
apart from your AI ambition, are your AI Heat Map and Roadmap. These will
list the functions, services and processes that are being targeted for automation, and will give the priority of how they are likely to be rolled out. Particular
aspects to consider will be the different technologies that will need to be
deployed, the complexity of the solutions and the current state of the data
that will be exploited.
The roles that are generally included within an Automation (and particularly an AI) CoE fall into four main functions: the Management of the CoE,
an overall Architecture Management, the Implementation Teams and
Operations. Some of these roles, including, particularly, parts of the
Implementation Teams, can be provided by third-party providers or vendors,
especially in the early days of the CoE. The following is a guide for how some
organisations are structuring their automation CoEs, but can obviously be
adjusted for your own needs and aligned with your company culture.
The Management Team should ideally start out relatively small, including
at least a manager and some project management capability. The manager
should have responsibility for the whole CoE and the people assigned to it, as
well as the communications across the organisation and upwards (communications could be taken by an internal communications specialist later on). For
management and control of projects, the team (which could just be one person at first) should be responsible for planning, project management, resource

160

9

Industrialising AI

tracking and reporting across all the various AI projects in the CoE. Other key
areas that usually come under this team would include the co-ordination of
training and education for both members of the CoE and users of the
systems.
Taking on the formulation of the business case and solutions are the
Architect Team. This team is best headed up by someone who understands
the business functions and processes but also the technical aspects as well. The
team will look at the opportunities for automation in the wider context (so
would have control of the AI Heat Map) and would create business cases for
each of them, carrying out the initial scoping of effort and technologies, all of
which would be done in close collaboration with the Subject Matter Experts
(or Process Owners or Business Analysts) from the Implementation Team.
The Architect Team would be responsible for managing the pipeline of
opportunities, ensuring that these are being proactively sought, and reactively
received, from across the business. They would also include technical architects (unless this role is retained within the IT Department), data scientists (if
required) as well as managing customer experience.
The Implementation Teams are where the bulk of the CoE resources
would reside. It would consist of a number of project teams, each focused on
a specific solution. Depending on the size and complexity of each project they
could include: a project manager, a project architect, developers (who would
have responsibility for importing and training data, creating and configuring
the models), SMEs/Business Analysts/Process Owners (carrying out solution
design and validation) and Quality Assurance. If appropriate, specialist
resources such as customer experience specialists, linguists, web developers
and integration developers could also be assigned to projects.
Agile development approaches are the most suited to building AI solutions.
The specific type of Agile (e.g. Kanban, Scrum) is not that important, just
that the developers and SMEs work closely together and on fast iterations.
Because of this, and the transformational nature of AI, the members of the
Implementation Team (and the CoE generally) should have a strong mix of
both business and technical knowledge. For regulated processes, which throw
up additional control requirements, there are a number of Agile approaches
that try to cater for these, including R-Scrum and SafeScrum.
Support and maintenance of the released applications (fixing bugs and carrying out any enhancements) can be handled in a couple of different ways.
For small implementations, the original project team can retain responsibility
for support, whilst for larger or more complex projects a separate support
team should be created. Therefore, young or small CoEs have the support
responsibility within the Implementation Team, whereas larger, more mature

Building an AI Organisation

161

CoEs will generally create a new team for this purpose, or have it embedded
as part of the Operations Team.
The final group is the Operations Team. This team will have responsibility
for the deployment, testing, updating and upgrading of any of the live systems. They will also have responsibility for the technical integration of the AI
solutions with any other systems. For organisations used to taking a DevOps
approach (where the Operations and Development resources work as integrated teams), then much of this Operations Team would be subsumed into
the Implementation Team.
As I mentioned earlier, the above groups are only meant as a guide and can
be flexed to match your own company’s requirements and practices. Another
consideration regarding the setting up of a CoE is how it fits into the overall
organisation structure of the company.
Some companies, such as Aviva and Clydesdale and Yorkshire Bank, have
created ‘lab’ or ‘innovation hub’ environments (the Aviva one is actually called
a Digital Garage). These are useful vehicles to generate additional buzz around
the initiatives and to foster a culture of innovation. They usually focus on the
early stage development of ideas. CoEs though tend to be a little more practical and will include most, if not all, of the capabilities to identify, build and
run new technology initiatives.
A key consideration, especially for larger organisations, is whether to have
a single, centralised CoE or to spread the capability across the various divisions or business units whilst retaining a central control function. The
approach that usually works best for AI, at least initially, is to keep everything
as centralised as possible. This is for a couple of reasons.
Firstly, the impetus and momentum for the early AI initiatives usually
come from one area of the business (because of greatest need, biggest opportunities, most enthusiastic management, etc.), and this can form the genesis
of the CoE. When other business units see the benefits that are being generated, then they can tap into the existing capability rather than creating it
themselves from scratch.
Secondly, an ‘independent’ team not involved in the day-to-day running of
the function is much more likely to identify, and be able to implement, transformational change. Left to their own devices, business units rarely consider
the full range of opportunities available and will instead focus on the simpler
process improvements. Both are valid types of opportunities, but AI has much
to offer with regard to transformational change and therefore it should be
promoted at every opportunity.
A final consideration regarding the organisational structure of your industrialised AI capability is whether to appoint a senior executive to oversee it all.

162

9 Industrialising AI

Many large companies, especially in the financial services sector, already have
Chief Data Officers (CDO). These roles take responsibility for the enterprise-­
wide governance and utilisation of information as an asset (including in some
cases a large element of revenue generation). For some companies, such as
Chartis, Allstate and Fidelity, the CDO has a big influence on the overall
strategy of the company.
A relatively new position that is starting to be seen in some businesses is the
Chief Automation Officer (CAO). These can sometimes be called Chief
Robotics Officers or Chief AI Officers (CAIO). This role should try to embed
automation, and AI, at the centre of the business strategy. A CAO, more than
a CDO or a CTO, can look outward and forward. They will tend to have a
greater focus on the business opportunities of automation than, say, a CIO.
(Some analysts believe that a CAO is more suited to a move up to CEO than
a CIO might be.)
CAOs are still relatively rare, and it may be that it is a role that is needed
for a certain period of time, to get automation up and running and firmly
embedded in an organisation, and then have those responsibilities eventually
absorbed back into business-as-usual.
I am not aware, at time of writing, of any organisations with CAIOs yet,
although there is plenty of talk about the role. Again, it might be a short-lived
necessity, and many people are sceptical that this level of focus is required in
the boardroom. It may be that the CDO or CAO (if they exist in a company)
are able to cover AI specifically.
Whether an organisation chooses to bring in a CDO, CAO or CAIO, the
people carrying them out need to have some pretty special capabilities: they
will need at least a reasonable understanding of the technologies and the data
infrastructure; they will need to be able to work cross-functionally, since automation opportunities can exist across the whole business and will require the
cooperation of a number of departments; they will need to behave like
­entrepreneurs within the business and they will need to have the industry
standing and inter-personal skills to attract and retain the best talent. That is
not an insubstantial shopping list.
We have now come to end of the journey to understand, create and industrialise AI in businesses. Throughout the book, I have focused specifically on
what AI can do for organisations today, and the practical approaches that can
be taken to exploit the inherent value in this technology. But now, in the final
chapter, it is time to look to the future: how AI may develop, what new
opportunities and risks it will create, and how we can best plan to get the
most from it.

The Data Scientist’s View

163

The Data Scientist’s View
This is an extract from a dialogue with Richard Benjamins, who, at the time
of the interview, was working as Director, External Positioning & Big Data
for Social Good at LUCA, part of Telefonica, a European Telecoms Company.
He is now Group Chief Data Officer & Head of the Data Innovation Lab,
AXA, a global insurance company.
AB:	In your role at Telefonica’s data-led business, LUCA, you understand the
inherent value, as well as the challenges, in using Big Data for AI applications. Can you tell me more about this?
RB:	Firstly, you have to look at the numbers. Six years ago, McKinsey estimated that Big Data would bring $300bn in value for healthcare and
€250bn for the European Public Sector, but in their most recent update
at the end of 2016, they reckon that the numbers were actually 30%
higher than this.
But none of that tells you how you should get the value from your own
big data and how to measure it, and that’s the biggest challenge for organisations who want to be able to realise that value. I think there are four ways
that you can get at that inherent value.
Firstly, and probably most simply, you can reduce the cost of the IT infrastructure that is being used to manage your big data, using open-source tools
such as Hadoop. This can save significant amounts of money, and is easy to
measure. Secondly, big data can be used to improve the efficiencies of your business, that is, enable you to do more with less. Thirdly, it can be used to generate
new revenue streams from your existing business—this can be a challenge
though as it is difficult to know where to start, and it can be hard to measure the
value. The final source of value is where completely new revenue streams are created from big data, that is, new products where data is at its centre, or by creating
insights from data to help other organisations optimise their own businesses.
I believe that in the next few years, the dominant source of big data value
will be from business optimisation, that is, by transforming companies into
data-driven organisations that take data-driven decisions.
AB:	So how do you start to put a number against business optimisation
through big data?
RB:	This is the big challenge. Of course, you should have a base case and compare the difference before and after, but big data initiatives are rarely the
only reason for doing something, so it’s difficult to isolate the value. There
are two ways to help with this though. The first is to do a segmented
experiment; for example, have one group of customers that are being analysed through big data, and another that are not (and maybe even another

164

9 Industrialising AI

where a different approach is being used). The second is to do something
with big data that has never been done before – that way you can compare
the results directly with your base case.
From my experience, though, mature organisations know that there is
value in big data and will not be overly obsessed with trying to measure
every bit of value. When companies reach the stage where big data has
become business-as-usual, then it changes the way that departments interact
and the value naturally flows from this.
AB:	In order to manage all of this data and value, the role of the Chief Data
Officer (CDO) must be crucial?
RB:	Definitely. We are seeing a big increase in the number of businesses that
have a CDO. According to one survey I have seen, 54% of companies
now have a CDO, up from just 12% in 2012.
Where the CDO sits in the organisation is an open question still. It’s getting closer to the CEO level every year—in Telefonica for example, five
years ago the CDO was originally five levels below CEO but now reports
directly into that position.
The best position for a CDO to be in is where they have responsibilities
across various functions or departments, so that their role, and objectives,
are not defined by a single department. Reporting into the COO is probably a good balance as they will then have good cross-functional visibility.
AB:	Big Data is obviously closely related to Artificial Intelligence. Do people
get confused between these terms?
RB:	The term Artificial Intelligence can be used for many things including Big
Data and Machine Learning. The hype that is around AI can be useful
because it brings an increase in interest and attention but it is important
to keep in mind what we are talking about. It is not just about building
wonderful applications, it is also about asking fundamental questions
related to how people think, how they solve problems and how they
approach new situations. When thinking about AI I think it is important
to consider these three things: AI can solve complex problems which used
to be performed by people only; what we consider today as AI may just
become commodity software; and, AI may help us shed light on how
humans think and solve problems.
And remember, in the big scheme of things, this is only the beginning...

10
Where Next for AI?

Introduction
This book has always been about what executives need to do today to start
their AI journey. Rightly so, but it is also important to understand what is
coming down the line. Organisations need to be prepared as much as possible to exploit any future advancements as well as mitigate any of the developing risks.
This final chapter tries to predict the near future for AI. This is, of course,
a bit of guessing game, but hopefully it is a well-informed one. Firstly, I look
at which of the AI capabilities I have discussed in the book are likely to flourish and thrive the most, in other words which should be watched most closely
for significant developments.
I then have a go at predicting when AI might become ‘business as usual’,
that is, when we are no longer talking about it as a new and exciting thing, in
a similar way that outsourcing has now become simply another tool for the
executive to deploy. This section covers some of the more general predictions
about AI and its use in business.
And finally, I close the book with some words of advice about how to
future-proof your business (hint: it includes AI).

© The Author(s) 2018
A. Burgess, The Executive Guide to Artificial Intelligence,
https://doi.org/10.1007/978-3-319-63820-1_10

165

166

10

Where Next for AI?

The Next Opportunities for AI
In Chap. 3 I described the core capabilities of AI and then in Chap. 5 I gave
numerous examples of how those capabilities are being used in businesses
today. But the technology will inevitably get better, and the speed, accuracy
and value of these capabilities will only increase, probably at an ever-­
accelerating pace.
This section provides my summary of how each of those capabilities might
develop over the next few years—some will accelerate whilst others may hit
some bumps in the road.
• Image recognition has already advanced dramatically over the last five
years and, particularly, over the last couple. Its development over the next
few years will continue at this heady pace as more image sets become available for training purposes, and better algorithms, faster processors and
ubiquitous storage will all contribute to the ability to improve accuracy but
also tag moving images as well. There has already been some progress in this
area but I think that we will see objects in films, clips and YouTube movies
identified and tagged automatically and efficiently. This would mean you
could, for example, search for an image of a DeLorean car across all of the
movies online, and it would return the frames in the films where the car
appeared.
The next development beyond that will be to identify faces in movies,
just as Google Images and Apple’s Photo applications can do today for still
photos. Being an optimist I am hoping that the medical applications of
image recognition, such as in radiology, will really start to take off—they
will have a huge impact on society, and therefore deserve to be promoted.
The technique of having AI systems play (or battle) with each other,
called GANs, will enable systems to create new images from scratch. Put
simply, one system, after an initial period of training, will create an image,
and another will try and work out what it is. The first will keep adjusting
the image until the second system gets the right answer. This technique
could also be used to create scenery for video games or de-blur video images
that are too pixelated.
Image recognition will have to face the challenges around bias, which
are certainly coming to the fore at the time of writing. Ensuring unbiased
image sets should be a mission for all those creating and controlling data,
and it shouldn’t become an issue only when things have clearly gone
wrong.

The Next Opportunities for AI

167

• Speech Recognition, as can be evidenced in domestic devices such as
Amazon’s Alexa, Google’s Home and Apple’s HomePod, is already very
capable, although there are plenty of examples of poor implementations as
well. I predict that this capability will become much more prevalent in a
business-to-business environment, powering, for example, an HR
Helpdesk, or a room booking system.
Real-time voice transcription already works reasonably well in a controlled environment (a nice, quiet office with a microphone on a headset)
but advances in microphone technology in parallel with better algorithms
will make speech recognition viable in most environments.
In the section on malicious acts in Chap. 8 I talked about voice cloning.
As long as this technology can stay out of trouble (or the people wanting to
use it for good can keep one step ahead of the hackers) then this has the
potential to help many people who have voice-crippling diseases.
• Search capability really took off in 2017, especially in specialised sectors
such as legal. There are a number of very capable software vendors in the
market which means that this will be a useful introduction for businesses as
they look to adopt some level of AI capability. The algorithms will inevitably get better and more accurate, but I think that the future for this capability is as part of a wider solution, either by extending its own capabilities
(e.g. to be able to auto-answer emails from customers) or by embedding it
in others (as part of an insurance claims processing system, for example).
• Natural language understanding is closely connected to both Speech
Recognition and Search, and will benefit from the advances there, particularly in areas such as real-time language translation. But I think the hype
and elevated expectations that have surrounded chatbots will have a detrimental effect on their adoption. I predict that there will be a backlash
against chatbots before they start to make a recovery and become widely
adopted and used every day. Those firms that implement chatbots properly
(i.e. for the right use cases, with appropriate training, and capable technology) will see successes, but those that try and take shortcuts will speed the
oncoming backlash.
The sub-set of NLP that focuses on generating natural language (NLG)
will see some big strides, and therefore wider adoption, over the next few
years. It will benefit from broader acceptance of AI-generated reports, particularly where they add real value to people’s lives, such as with personalised, hyper-local weather forecasts. As more data is made available
publicly, then more ideas will surface on how to extract value from it and
to communicate those insights to us using natural language.

168

10

Where Next for AI?

• Clustering and prediction capabilities can, for this purpose, be considered
together. Two things will happen regarding the use of big data: firstly, as
businesses become more ‘data aware’, they will start to make better use of
the data they have, and also exploit publicly available data as well; and,
secondly, less data will be needed to get decent results from AI, as a­ lgorithms
become more efficient and data will become ‘cleaner’ and (hopefully) less
biased.
Financial Services will be the main beneficiary from Prediction, being
used in more and more companies (not just the global banks) to combat
fraud and up-sell products and services. Cross- and up-selling will also
become the norm for retailers.
The challenge for these capabilities will be whether they will be used
productively and ethically. Increasing the number of click-throughs and
delivering targeted adverts may make sense for some businesses, but it is
never going to change the world, and will no doubt frustrate many customers. And, of course, the questions around bias and naivety will remain on
the table and will need to be addressed for every use case.
• Optimisation, out of all the capabilities, has, for me, the greatest promise.
The work that is being done with Reinforcement Learning (that was at the
heart of AlphaGo’s victory) has enormous potential in the worlds of risk
modelling and process modelling, and specific examples where energy bills
can be significantly reduced (as at Google’s data centre) demonstrate that it
should have broad and important applications for us all.
One of the biggest constraints on AI right now is the availability of
properly labelled, high-quality data sets. GANs, which I mentioned earlier,
have the potential to allow AI systems to learn with unlabelled data. They
would, in effect, make assumptions about unknown data, which can then
be checked (or, more accurately, challenged) by the adversarial system.
They could, for example, create ‘fake but realistic’ medical records that
could then be used to train other AI systems, which would also avoid the
tricky problem of patient confidentiality.
But reinforcement learning and GANs are the AI technologies that
could still be described as ‘experimental’, so we may not see these being
used on a day-to-day basis for a few years yet. In the meantime, we have
systems that can make important decisions on our behalf, including
whether we should be short-listed for a job, or approved for a loan. The
main advances here, I think, will be in how they are used, rather than any
major advances in the underlying technologies. Eliminating bias and providing some level of transparency will be the key areas to focus on.

When Will AI Be Business as Usual?

169

• As we are looking into the future, it is worth considering some of the questions around Understanding, which I stated in Chap. 3 was still, as an AI
capability, firmly based in the AI labs, and may be there for a very long
time, if not forever. Having said that, a really interesting area of development to keep an eye on is that of Continual Learning. This is an approach,
currently being developed by DeepMind, that hopes to avoid ‘Catastrophic
Forgetting’, the inherent flaw in neural networks that means that a system
designed to do one thing won’t be able to learn to do another unless it
completely forgets how to do the first.
Because the human brain learns incrementally it is able to apply learnings from one experience and apply it to another, but neural networks are
extremely specialised and can learn only one task. DeepMind’s approach,
using an algorithm they call Elastic Weight Consolidation (EWC), allows
the neural network to attach weight to certain connections that will be useful to learn new tasks. So far, they have demonstrated that the approach
works with Atari video games, where the system was able to play a game
better once it had been trained on a different game.
Although there are a number of labs looking at tackling Artificial General
Intelligence, EWC is probably the most exciting opportunity we have for
getting closer to an answer to this most intriguing of challenges.

When Will AI Be Business as Usual?
In the previous section I talked about the developments I expect to see in the
next few years for each of the AI capabilities. In this section I want to be able
to generalise those predictions and look at the wider context around
AI. Specifically, I will look at the question of when (if ever) AI will become
‘business as usual’.
The traditional way of understanding the maturing of any technology is to
see where it is on the Gartner Hype Cycle. This is a well-used graph which
plots the life of a technology from its initial ‘technology trigger’, through its
rise in popularity to a ‘peak of inflated expectations’ that is then followed by a
decline into the ‘trough of disillusionment’. After that it recovers through the
‘slope of enlightenment’ and finally reaches the ‘plateau of productivity’. It is
this final stage that can be defined as business-as-usual.
The challenge with plotting AI on this curve is that AI is many different
things and can be defined in many different ways. We could plot some of the
individual capabilities, such as NLU (heading into the trough of disillusionment) or some of the technologies, such as machine learning (right at the peak

170

10

Where Next for AI?

of inflated expectations) or GANs (as a technology trigger). To generalise
though, most of the machine-learning-based AI technologies are pretty much
peaking with regard to expectations and are therefore primed to enter the
steep slide into disillusionment.
Being in the trough doesn’t mean that everyone abandons the technology
because it doesn’t work anymore—in fact it’s usually quite the opposite.
Developments will continue apace and great work will be done to create new
applications and use cases. It’s just that the perception of the technology across
businesses and society will shift; there will be less hype, with people taking a
more sober, considered approach to AI. The best thing about being in the
trough of disillusionment is that everyone’s expectations are now much more
realistic. And that surely must be a good thing.
How long it stays in the trough will depend on a number of factors. The
four key drivers I described in Chap. 2 (big data, cheaper storage, faster processors and ubiquitous connectivity) will only continue on their current trajectory. Big data may be the one area that makes or breaks it through.
I think the greater availability of public data will be one of the main reasons
that AI is adopted by the mainstream and reaches the plateau of productivity
quicker. But there are challenges to this, as the tech giants hoard their own
proprietary data sets farmed from their billions of ‘customers’ (who, by the
way, should more accurately be called suppliers, since they provide all the data
that is used by the tech companies to sell advertising space to the advertisers,
their actual customers).
But even their data will become more difficult to farm—there is a complicit
deal between the data provider (you and me) and the user (tech giants) and
this will need to continue despite being constantly eroded by examples of
misuse or abuse or hacking. It’s clear what the tech giants need to do to keep
their supply of data coming in: there is a balance of utility, trust and transparency in the use data, and increased transparency can often compensate for
lower levels of trust.
Once we have more publicly available data and higher levels of transparency from the tech giants we can start to exploit all the connections between
these data. Rather than just having public transport data, for instance, we can
bring in our own personal data that might include our inherent preferences
for how we like to travel. (These could be set manually, of course, but better
to pull that information from actual, real-time behaviours rather than what
we think we liked a year ago.)
This idea of hyper-localised, hyper-personalised information may well
be the key to widespread acceptance of the benefits of AI. Some examples
can already be seen (e.g. hyper-localised next-hour weather forecasting,

When Will AI Be Business as Usual?

171

hyper-­personalised what-to-wear recommendations) but the bringing to
together of all the different sources of data to provide really useful information that is relevant to just you at that specific time in that specific location
will be, to use a cliché, a game changer.
If there is trust and transparency around the data that consumers find useful,
then they are more likely to allow businesses open access to that, therefore
increasing the utility even further. (This is a mind-set change from today where
consumers know their data is being exploited but don’t ask too many questions
because they believe the benefits are usually worth it.) And businesses will benefit from this as well, and can use the same ‘three-dimensional data’ approach
to help run their businesses better internally—salespeople can be much better
informed, as can couriers, caterers, HR managers, executives and so on.
So, there is a huge dependency on the levels of trust in data usage, which
will be driven by the level of transparency that there is in how the data is
farmed and used. Other areas that might delay the date at which AI reaches
the plateau of productivity could be a general backlash against the technology,
especially as the amount and types of jobs begin to be materially affected.
Right now, though, one of the biggest challenges to the widespread adoption of AI is actually a shortage of skilled and experienced people. Data scientists and AI developers will become superstars (with superstar salaries) as the
tech giants and AI organisations, such as OpenAI, fight over these scarce
resources. The pipeline for supplying skilled people, which includes the education systems of the world, will take time to adapt, with quick fixes such as
online training courses taking the immediate strain.
People will also make a difference on the demand side as well. The
Millennials, as we all know, have grown up with many of these technologies
and are both familiar and dependent on them. They will absorb new AI applications without thinking and will have fewer issues with some of the ethical
challenges I have described (although this is clearly a generalisation—you
only have to read the interview with Daniel Hulme in this book to realise that
there are people who care deeply about the dangers of AI and how it can benefit the world).
Customers for AI will therefore have a much bigger impact on the adoption
of the technology. It is fair to say that AI is currently a vendor-driven market—the innovations and ideas in the market are mainly borne out of the fact
that it can be done, rather than it needs to be done. This balance will slowly
shift over the coming years as people became more familiar with the capabilities of AI and there are more examples of its use in society and in the
workplace.

172

10

Where Next for AI?

Things like speech recognition, and particularly voice authentication, have
already become part of some people’s normal lives as they become familiar
with, and then maybe dependent on, their Google Home, Amazon Alexa or
Apple HomePod devices. Once this technology enters the business
­environment, in, say, an HR Helpdesk, we will start to see the beginnings of
AI becoming ‘business as usual’.
The ‘fully connected home’ dream that AI product vendors try and sell currently falls far short of expectations—there are lots of different products and
standards out there which rarely work together and can be difficult to set
up—but one can see signs that the seamless and effortless use of AI to help
run our homes is on the horizon. So when people go to work, they will expect
the same sort of experience, and this will drive further developments, and the
wider acceptance, of AI in business.
One of the most interesting developments I am starting to see is the creation of ‘vertical’ AI solutions, where a number of different AI (and associated
technology) capabilities have been brought together to solve a particular challenge in a particular domain, such as financial services, or HR, or healthcare.
Many of the AI ‘solutions’ being developed right now fall short of being complete solutions—they just look at a very specific capability or service—and
these will quickly become commoditised or embedded into existing enterprise
systems. ‘Full stack’ vertical solutions, however, will have AI at their core, and
will focus on high-level customer requirements that can be better met using
AI, or will be used to discover new requirements that can be satisfied only by
using AI. These solutions will take a much more holistic approach and draw
upon subject matter expertise just as much as they do technical know-how. In
my mind, once we start to see more and more of these types of solutions, the
quicker AI will achieve business-as-usual status.
Another bellwether for AI’s acceptance onto the plateau of productivity
will be how business’s organisational structure changes to accommodate
AI. There are already a number of enterprises that have created Automation
(and AI) Centres of Excellence (I discussed these in Chap. 9) and these will
become much more commonplace in the next few years. Full acceptance
though will come when enterprises no longer need these CoEs—when AI is
seen as the normal way to do business. By this point we won’t have AI specialists—everyone will be a generalist with AI skills. Just as in the 1970s there
were people who had jobs ‘working the computer’ and now everyone just
‘works with computers’, so the same will be true for AI.
There is a slightly tongue-in-cheek definition of AI which says it is ‘anything that is 20 years in the future’, but there is probably some truth in that.
The AI capability that we have now and use every day would have seemed like

Future-Proofing Your Business

173

black magic 20 years ago, yet we don’t really consider it to be AI anymore. We
are always looking forward for the new shiny thing that we can label AI. So it
is almost a self-defeating exercise to try and predict when AI will become business as usual. I have, though, tried in this section to identify those things that
will signal when AI (as we consider it today) has reached the plateau of productivity. But if your business is waiting until then before looking into or
adopting AI then you will be too late. The time to start your AI journey is
now: to future-proof your business so that you can defend against, and prepare for, the new AI-driven business models and opportunities that will inevitably come.

Future-Proofing Your Business
In the very first chapter of this book I implored you to ‘not believe the hype’.
Over the subsequent nine chapters I have hopefully provided you with a practical guide on how to approach AI so that you are as informed and prepared
as you can be. I also hope that you are now as excited as I am about the benefits that AI can bring to business and society, and that you can see how AI
could be beneficial to your company.
But the most important thing I want you to have taken away from this is
that every company, every senior executive, needs to be thinking about AI
now, today. The development of this technology is already moving apace but
it will only get faster. Those who wait will be left behind. Now is the time to
start future-proofing your business to capture the full value from AI.
There are three broad phases to this: understand the art of the possible,
develop an AI strategy and start building stuff.
Understanding the art of the possible is the first stage in future-proofing
your business. By reading this book you will have taken an important step
forward, but you should also be reading many of the very decent books and
resources that are also available, many of them for free on the internet. There
are, of course, significantly more poor resources on the internet than there are
decent ones, but sites such as Wired, Quartz, Aeon, Disruption, CognitionX,
Neurons, MIT Technology Review, The Economist, The Guardian and The
New York Times make very good starting points. You should also consider
attending conferences (there are now a plethora of automation and AI conferences available, some of which you may find me speaking at) and also signing
up for AI masterclasses and/or boot camps.
The ‘art of the possible’ should, of course, be grounded with some level of
realism. Therefore, it is important to understand what other enterprises are

174

10

Where Next for AI?

doing, particularly those in the same sector or with the same challenges.
Advisors can be useful for this as they (we) tend to have further insights than
those that are available publicly. Advisors will also be able to challenge you on
what can be done and to help you ‘open your mind’ to all the available
opportunities.
Some useful thought experiments that can be done at this stage include
imagining what a company like yours might look like in 10 or 20 years’ time,
or to try and imagine how you would build a company with the same business
model as yours from a blank sheet of paper—what are the core things you
would need, what could you discard and what could be delivered through
automation? Holding ‘Innovation Days’ can also work—these would bring
together stakeholders from the business to listen to the art of the possible
(from experienced third parties) and have demonstrations from AI vendors of
their solutions. This should then spark new ideas that can be fed into the AI
Strategy.
Following on from the initial thinking, the next phase will all be about
building the AI Strategy. A good chunk of this book describes this process in
detail, so I won’t repeat myself here. Suffice it say that the strategy should be
built around identifying and addressing existing challenges, as well as identifying new opportunities that can be addressed by AI.
Creating and delivering innovation within any business is never an easy
ride. It takes time and effort, with planned investments and a clear mandate.
For AI, you could create a small working group with that specific focus, or
you could sit on the back of any innovation capability you may already have
in-house. The risk with the latter of these is that you may end up with the
‘lowest common denominator’ and miss out on some of the bigger AI opportunities. The advantage is that you can call upon additional resources and
other technologies that may be instrumental to your solutions.
The important thing to remember about any program to implement innovation, but particularly with AI, is that it will require a cultural shift. People
will need to think very differently, and they will be challenged about many of
their core beliefs of how a business should be run. Be prepared for some heavy
discussions.
The final phase to start future-proofing your business is to build some AI
capability. Based on your AI Strategy, building some prototypes will be the
first real test for AI, and the first time that many people will have seen AI in
action in your business. This act of disruption will have a profound affect and
will be a great catalyst to start new conversations across the business. Ensure
that you make the most of this opportunity to showcase AI and bring the rest
of the business along with you.

Final Thoughts

175

You should always be ready for new developments in the technology. AI is
always getting better—advances that we weren’t expecting for decades are suddenly with us, and you need to be informed and ready to bring those into
your business if they are relevant. Having an AI, or automation, CoE, perhaps
backed up with external advice, will ensure that the market is being constantly
scanned and all potential opportunities are captured as soon as possible.
Some businesses, once they have some prototypes under their belt, take the
disruptive power of AI to the next level and build parallel digital businesses.
These in effect compete with the legacy business, and are therefore not encumbered by existing systems or culture. Even if you do not create a new business
unit, this healthy competitive mind-set can be useful to bring out the full
potential of the technology.

Final Thoughts
This has been a fascinating book to write, and I hope that you have, at least,
found it an interesting and thought-provoking read. I’d like to think that you
are now much better prepared and informed to start your AI journey with
your business.
That journey won’t be an easy one—disrupting any business with new technology has more than its fair share of challenges—but the rewards should
certainly be worth it. Be very aware that all your competitors will be considering how AI can help their businesses, but the advantage will be to those who
actually start doing something about it now. As I said in the previous sections,
if you are waiting for a mature and stable AI market then you will be too late.
Thank you for spending your precious time in reading my book. If you
need any further help or advice, then you know where to find me. But for
now, put down the book, and go start your AI journey.

Index

NUMBERS AND SYMBOLS

B

1-800-Flowers, 76

Baptist Health South Florida, 81
Barclays, 60, 77
Baxter, 64
benefits, 1, 6, 18, 22, 30, 37–9, 41, 50,
51, 59, 60, 63, 64, 70, 73, 77,
78, 80–2, 84, 87, 89, 92–5, 98,
100, 102, 103, 105–9, 112, 113,
129, 141, 143, 149, 150, 154,
161, 167, 170, 171, 173
Berg, 87
Berkeley, 18
bias, 27, 47, 48, 77, 86, 87, 125,
130, 131, 134, 136, 144,
157, 166, 168
big data, 3, 11, 13–15, 19, 25–7, 41,
56, 66, 87, 95, 114, 125, 129,
136, 139, 163, 164, 168, 170
Brown, 18
business case, 27, 73, 98, 103,
105–13, 115, 118, 149,
153, 154, 160

A

Advanced Research Projects Agency
(ARPA), 12
Ageas, 80
AIA, 78, 79
Alexa, 2, 17, 27, 34, 39, 76, 167, 172
Alexander Mann, 86
algorithms, 1, 2, 6, 9, 14, 18–20, 26,
27, 36, 43, 45, 47, 50, 54, 57,
58, 68, 83–5, 114, 119, 124,
125, 129, 132, 136, 138, 144,
145, 166–9
Amazon, 2, 16–18, 27, 34, 56, 57, 68,
72, 76, 119, 167, 172
ambitions, AI, 100, 106, 113, 147,
158, 159
Apple, 17, 27, 140, 166, 167, 172
Associated Press (AP), 82
automation, 2, 3, 5, 6, 24, 53, 58, 71,
85, 91–4, 96–8, 100, 107, 108,
110–13, 119, 123, 149, 158–60,
162, 173–5
Axa, 84, 163

C

Care Quality Commission (CQC), 87
Celaton, 70, 78, 79, 88, 89

© The Author(s) 2018
A. Burgess, The Executive Guide to Artificial Intelligence,
https://doi.org/10.1007/978-3-319-63820-1

177

178

Index

change management, 8, 89, 110,
112, 113, 120, 147, 148,
150, 151, 155, 156
chatbots, 5, 12, 39, 45, 57, 63,
74–8, 82, 83, 93–5, 97, 101,
102, 115, 132, 140, 143, 167
cloud, 16, 18, 19, 26, 52, 55–8, 76,
91, 114, 119, 150, 151
clustering, 3, 4, 37, 38, 41, 46,
51, 66, 93, 95, 108, 136,
137, 141, 168
Clydesdale and Yorkshire Banking
Group (CYBG), 77, 161
Cognitiv+, 53
cognitive reasoning, 6, 43, 62, 74,
80, 101, 107, 114–16, 124
consultancy, AI, 144
Cortana, 2, 39
crowd sourcing, 55, 67–9, 102,
125, 131, 150, 151
Customer Satisfaction (CSAT), 93,
103, 107
customer service, 4, 18, 24, 30, 36,
52, 61, 63, 73–8, 84, 89, 93,
99, 101, 102

E

D

H

DARPA, 12
data, 1, 13, 30, 55, 74, 95, 118,
129–32, 152, 166
data scientists, 23, 54, 57, 119–21,
124, 130–2, 137–9, 143,
152, 160, 163, 171
Davies Group, 79
DeepMind, 42, 45, 50, 66, 83, 169
Deep Neural Networks, 8
Deutsche Bank, 81
developer, AI, 3, 18, 23, 55–7,
135, 137–40, 144,
152, 171
Durham, Police Force, 86, 130

hacking, 141, 170
heat map, AI, 100–6, 112, 113, 159, 160
Hidden Markov Model (HMM), 34
HMM. See Hidden Markov Model
(HMM)
HSBC, 77, 86

eco-system, 148–52, 155
expert systems, 11–13, 41, 43, 124,
132, 133, 140
F

Facebook, 13, 15, 16, 32, 45, 83
Farfetch, 87
framework, AI, 3, 4, 7, 29–54, 57, 73,
101, 113, 120, 140
G

GANs. See Generative Adversarial
Networks (GANs)
general artificial intelligence, 4, 23, 48,
56, 65, 70, 73, 145, 165, 169
Generative Adversarial Networks
(GANs), 45, 131, 166, 168, 170
Goldman Sachs, 85
Google, 2, 13–15, 18, 32, 40, 42, 50,
57, 66, 68, 83, 84, 86, 94, 119,
166–8, 172
graphical processing units (GPUs), 17,
26, 57, 151

I

IBM, 16, 18, 19, 25, 26, 32, 75–7, 80,
119, 120, 150
image recognition, 3, 7, 20, 32, 33, 35,
40, 65, 79, 80, 96, 125, 141,
142, 166

Index
  

infrastructure, 26, 53, 57, 63, 76, 82,
83, 103, 119, 151, 162, 163
Internet of Things (IoT), 55, 65–7,
72, 87, 91, 97
IoT. See Internet of Things (IoT)

179

NLP. See Natural Language Processing
(NLP)
NLU. See Natural Language
Understanding (NLU)
North Face, 77
NVidia, 17

L

London School of Economics, 26
M

machine learning, 3, 7, 8, 13–15,
19–22, 25, 29, 32, 39, 41, 43,
54, 62, 85, 114, 124, 130–2,
140, 164, 169, 170
machine translation, 40
Maturity Matrix, AI, 96, 97, 99,
100, 105
McKesson, 81
Minimum Viable Product (MVP),
122, 123, 149
MIT, 5
MNIST, 15
MVP (see Minimum Viable Product
(MVP))
N

naivety, 47, 48, 136, 142, 168
narrow artificial intelligence, 4, 23, 70
Natural Language Processing (NLP),
35, 51, 54, 64, 107, 167
Natural Language Understanding
(NLU), 2–4, 25, 31, 33–5,
38–41, 51, 54, 56, 57, 64, 65,
74, 76, 78, 81, 82, 87, 114, 167
Netflix, 76, 96
neural networks, 8, 11, 13, 21,
22, 50, 169
neurons, 8, 20
Nexar, 80

O

opaqueness, 23, 47, 132, 137
optimisation, 3, 4, 9, 19, 25, 30,
31, 41–6, 51, 63–6, 74, 79,
86, 93, 95, 97, 102, 108,
132, 140, 163, 168
Otto, 85
P

pilots, 12, 92, 122, 123, 149, 152, 153
Pinsent Masons, 82
planning, 3, 5, 41, 46, 51, 79, 107,
132, 148, 159
platform, AI, 18, 117–21, 140
PoC. See proof of concept (PoC)
prediction, 3, 4, 7, 31, 37, 46–8, 51,
54, 62, 83, 84, 95, 108, 130,
132, 165, 168, 169
processors, 16, 17, 19, 20, 22, 25, 30,
57, 140, 166, 170
proof of concept (PoC), 86, 91, 92,
115, 122, 123, 149, 153
prototypes, 89, 92, 111, 112, 117,
122–4, 147, 149, 174, 175
R

Rainbird, 114, 115
RAT. See Riskiest assumption test
(RAT)
reasoning, 95
reinforcement learning, 8, 42, 44,
45, 168

180

Index

ReThink Robotics, 64
riskiest assumption test (RAT),
122–4, 149
risks, 1, 2, 26, 27, 33, 39, 48, 53, 54,
60–2, 65, 73, 75, 79, 81, 84, 86,
87, 93, 95, 98, 103, 106, 107,
118, 123–5, 129, 130, 134, 138,
139, 142, 143, 145, 147, 153,
154, 157, 162, 165, 168, 174
roadmap, AI, 9, 92, 111–13, 136, 159
robotic process automation (RPA), 8,
27, 35, 36, 52, 55, 58–63, 76,
79, 91, 93, 95, 97, 98, 100, 102,
107, 108, 119, 131, 149, 150
robotics, 60, 64, 65, 83, 97
robust, 152
ROSS, 25
Royal Bank of Scotland (RBS), 75, 77
RPA. See robotic process automation
(RPA)
Rue La La, 85
S

Satalia, 144
search, 3, 7, 13, 14, 25, 26, 32,
35, 36, 38, 46, 59, 62, 79,
81, 82, 93, 95, 97, 106,
107, 142, 166, 167
SEB, 75, 76, 83
selection, 19, 47, 77, 118, 141,
148, 149, 152, 154–8
self-service, 63, 93, 95, 102, 108
Sensai, 126
service providers, 149, 152, 155–8
Siri, 2, 17, 27, 39
Slaughter & May, 82
speech recognition, 2–4, 7, 15, 18,
31, 33–5, 38, 39, 51, 56, 64, 65,
81, 95, 102, 107, 108, 167, 172
Stanford, 19, 64

Stanford Artificial Intelligence Lab
(SAIL), 11
Staples, 76
Stitch Fix, 77
storage, 16, 19, 22, 25, 32, 56, 140,
151, 166, 170
strategy
AI, 9, 91, 92, 96, 100, 110, 112,
113, 117, 120, 121, 125, 126,
139, 147, 148, 174
automation, 91–4, 99, 105, 123,
149, 156, 158
business, 91–4, 100, 112, 113, 127,
149, 162
supervised learning, 7, 20, 32, 34, 35,
39, 44, 45, 51, 68
T

Telefonica, 163, 164
TeliaSonera, 83
training, data, 7, 14, 15, 18, 23, 30,
34, 47, 68, 75, 86, 123–5, 131
transparency, 8, 27, 87, 132, 135, 138,
144, 168, 170, 171
U

UC San Diego Health, 81
UK Meteorological Office, 82
Under Armour, 78
understanding, 3, 4, 9, 11, 23, 29–31,
40, 41, 47–50, 71, 74, 83, 87,
88, 94–6, 101, 102, 110, 111,
123–5, 132–4, 136–8, 147, 153,
162, 169, 173
unintended bias, 47, 48, 130, 134–6,
144
unsupervised learning, 7, 20, 29, 30,
32, 37
USAA, 84, 86

Index
  
V

W

vendors, 2, 9, 19, 36, 52, 70–2, 89, 96,
114, 117–20, 131, 148–50,
152–9, 167, 172, 174
Virgin Trains, 78, 88

Watson, 18, 19, 25, 26, 32,
33, 75–7, 80
Winter, AI, 11–13, 17, 19,
22, 23, 126

181



Source Exif Data:
File Type                       : PDF
File Type Extension             : pdf
MIME Type                       : application/pdf
PDF Version                     : 1.4
Linearized                      : Yes
Language                        : EN
XMP Toolkit                     : Adobe XMP Core 5.2-c001 63.143651, 2012/04/05-09:01:49
Create Date                     : 2017:11:06 15:36:25+05:30
Creator Tool                    : Adobe InDesign CC 2017 (Windows)
Modify Date                     : 2017:11:07 23:01:24+05:30
Metadata Date                   : 2017:11:07 23:01:24+05:30
Producer                        : Adobe PDF Library 15.0
Format                          : application/pdf
Document ID                     : uuid:66aea5e3-3217-43c9-adcc-dc9f4b804af7
Instance ID                     : uuid:bae45cd8-e7f7-4747-809b-1d36a6a9dbdd
Rendition Class                 : default
Version ID                      : 1
History Action                  : converted, converted
History Instance ID             : uuid:564aea4f-253c-4f25-8d4f-594152f31437, uuid:ecd8a48e-bbb9-4733-a0c9-37b482257c58
History Parameters              : converted to PDF/A-2b, converted to PDF/A-2b
History Software Agent          : pdfToolbox, pdfToolbox
History When                    : 2017:11:06 15:41:59+05:30, 2017:11:07 23:01:23+05:30
Part                            : 2
Conformance                     : B
Schemas Namespace URI           : http://ns.adobe.com/pdf/1.3/
Schemas Prefix                  : pdf
Schemas Schema                  : Adobe PDF Schema
Schemas Property Category       : internal
Schemas Property Description    : A name object indicating whether the document has been modified to include trapping information
Schemas Property Name           : Trapped
Schemas Property Value Type     : Text
Page Layout                     : SinglePage
Page Mode                       : UseOutlines
Page Count                      : 187
Creator                         : Adobe InDesign CC 2017 (Windows)
EXIF Metadata provided by EXIF.tools

Navigation menu