Last year, a
friend of mine asked me my thoughts on an HBR article “If Your Company Isn’t
Good at Analytics, It’s Not Ready for AI” (1). It was an interesting one and
caused me to pause. I had started writing a blog post about it but abandoned it
half-way. Today this topic is more relevant than ever. So here goes.
If you read some
of the thousands of articles, or listen to people’s reactions around AI, you
would have thought everything is AI nowadays. But this isn’t the case; why?
Similarly, the
same friend was surprised to learn that some organisations such as tencent has
a huge team of data scientists; so does Zhong An(2) by the way (at least 54% of
their employees are engineers or technicians). That does seem to indicate that
the deeper you get into analytics/ml/ai, the more data scientists/machine
learning engineers… you need, not less.
Why is this so?
Why can’t an organisation “put all the data in one machine and all answers come
out” like what one of my bosses wanted? Can’t everyone just adopt AI?
If you build it, they will come (3)
I love the movie
“field of dreams”. It’s about a whacky farmer who decides he wants to put all
his eggs in one innovation and goes full hog, uprooting his previous business
model and establishing a new one driven by his passion (and a voice in his
head).
To me, the most
memorable line of the movie is “If you build it, they will come”.
This is often
the idea behind transformation projects, where the idea is to transplant analytics
(let alone AI) into an organisation. Whether it is from a bunch of external
consultants, or via hiring a relatively experienced head and build a team
internally, the results are most often the same. There are only ghosts on the
baseball field.
This is why I
consider most Insurance company big data labs a failure; the men and women in
white coats/uniforms are playing amongst themselves, and life goes on as per
normal for ordinary folks, like 2 distinct worlds. So what is the RoI of these
labs? To me, I would call this a failure since I believe the main aim of
analytics/”Data Science” is to generate RoI and benefit all parties – grow the
pie and everyone gets more.
So why the do
many attempts at embedding analytics in organisations end in failure? Why do
85% of data lake projects fail (4)? The technology is there. Sure, there messy
implementations, broken pipelines, choked pipelines, clogged processing
engines, extremely dirty data where ELT works only sporadically or knowledge
disappears with staff attrition…
harder
Well, as the
article (4) says “More than 85 percent of respondents report that their firms
have started programs to create data-driven cultures, but only 37 percent
report success thus far. Big Data technology is not the problem; management
understanding, organizational alignment, and general organizational resistance
are the culprits. If only people were as malleable as data.”
Even if you
build it, they may not come (for longer than the
rah-rah show).
Basically,
production of analytical pieces is ‘easy’. You drop me or any decent analytical
person in a datalake, throw an SMEs and a good data engineer (my claim of
analytics as a team sport (5)) and we are bound to catch some fish for the
client and clean it, and cook it for him/her; but he/she is unlikely to know
how to continuously include fish in his/her diet unless the people from the
client team are ready.
What is most
often missing is the ability to consume the pieces of analytics in a consistent
and on-going manner.
Consumption of
analytical/”data science”/AI output s not as obvious as you may think. And this
is the part that most failed implementations have in common, (and also why I
have consistently refused to join any organisation trying to transform itself
in terms of analytics if the role focuses on the IT production side).
There
can only be one, can’t there?
You could argue that it is
only necessary to have 1 good consumer in the organisation, 1 department adopt
analytics/”data science”, show the benefits and drag all other departments
along. Afterall, once a piece of analytics is successful, each head of
department can choose to adopt analytics and enjoy the benefits at a much lower
risk.
There are 2 flaws in this
argument. Firstly, we are forgetting the ability to consume, wanting to consume
is one thing, but being able to (analytically mature enough) is not a given,
Secondly, departments rarely exist in isolation in an organisation. A simple
example will illustrate this.
A while ago, I was
demonstrating how quickly a selection of customers based on their behavioural similarities
can be made and readied for an experiment. I gave up when the customer informed
me it usually takes 6 months to run a campaign (even a mini-one) and that was the
only way to run experiments. An organisation often moves at the pace of its
slowest department.
This brings us to
organisational analytical maturity.
I will admit that this is a topic that is very close to my heart and mind at the moment (hence the idea to revive the blog from last year). I fundamentally believe that in order for an organisation to fully benefit from the advantages provided by analytics or eventually becoming data-driven, it is critical for all parts of the organisation to be pulling in the same direction and preferably at the same speed.So how do I define analytical maturity?
To me, the easiest way to understand how mature an organisation is, is to understand the kind of questions that the people within the organisation are trying to answer using data.
The range of
questions where analytics can provide a good answer ranges from what has
happened to how can we make this happen. For simplicity the analytical maturity
can be broken into 4 stages.
Descriptive
Stage
The descriptive
stage is the first encounter many organisations have with data. It often takes
the shaped of backward looking reports: what has happened? How many of item X
did I sell last month? This is a stage most organisations will be familiar
with.
Diagnostic
Stage
After getting
the hand of static reports, the next stage is the diagnostic stage, where
hypotheses are formed. Questions are asked around “why” and often require
further slicing and dicing to find potential answers.
Predictive
Stage
The predictive
stage is when the questions move from looking backwards, to looking forwards.
While concerns about the future may have been implicit in the diagnostic stage,
it is in the predictive stage where specific tools, methodologies and
algorithms are employed to uncover what is likely to happen, and often how
likely it is to happen, what are the drivers of the behaviour.
Pre-emptive/Pro-active
stage
At this more
advanced stage, instead of taking certain variables/inputs as given and trying
to predict the outcome, the idea is to influence the variables and thereby
cause a change in the behaviour/status… Nudging, Behavioural Economics, Game
Theory are common strategies and approaches.
A simple example
can illustrate the difference, the “drain the swamp”(6) example:
·
Descriptive Stage: How many
people voted against me?
·
Diagnostic Stage: Why did these
people vote against me?
·
Predictive Stage: Who is that
person likely to vote for?
·
Prescriptive Stage: How do I
get that person to vote for me?
It is too easy
to underestimate how difficult it can be for people to climb through the stages
of analytical maturity, some never get to the pre-emptive/pro-active stage.
I believe that
usually people do not want to make their lives harder than it is, hence the
best way to make people in various parts of an organisation more analytically
mature is by showing them direct benefits to their own selves. It is about
change management.
At eternity’s gate (7)
For
organisations with people in departments who are only used to static reports or
even who are so busy that they don’t look at the reports, making descriptive
analytics visual is a natural step. To anyone who is interested in helping make
reports relevant to people, creating meaningful dashboards and triggering
people to think using numbers, I would recommend books by Stephen Few (8); I
had the opportunity to attend a course by the author a few years ago, and would
like to think I learnt a lot and I try to follow the guidelines as much as I
can.
The great thing
about this book is that the principles can be applied using most software, so
you can start from today itself.
One of the more
logical approaches to (re-)introduce the use of simple reports in an
organisation is to take stock of existing reports, gather business requirements,
and do a gap analysis. In parallel or even prior to that, it would be good to have
special purpose pieces of work answering specific ad hoc business questions.
When immediate needs are met, the focus can switch to future needs, the
discussion can move easier to dashboard design and ability to drill, slice and
dice.
Basically the
idea is to use ad hoc analyses and visualisations to encourage people to think
about data, and to use data to try solve their problems, moving from the descriptive
stage to diagnostic stage.
One of the
important aspects of the diagnostic stage is the culture of experimentation.
Hypotheses can be formed, may be even theoretically tested, but true learning
comes from actual experimentation, and this gets more important in the next
phase.
Back to the future (9)
The move from
backward looking to forward looking is a very important one. Creating
hypotheses (as in diagnostic stage) can still be done without knowledge of
statistics for example, but evaluating them and making inferences requires some
statistical knowledge, so does the evaluation of the results of experiments. This
is even more so when one moves into the realm of predictive analytics.
Why statistics?
Well I believe that having working knowledge of maths and stats allows the
understanding of many techniques used for predictive analytics. And I will, as
usual, place my favourite data science diagram (10):
Advanced
Analytics/”Data Science” is concerned about predictions, and as it can be seen
above, knowledge of stats/maths is an important characteristic of “data
science”.
Once an
organisation is comfortable in the world of creating hypotheses and possibly
testing them, the next step is to use predictions to guide the ‘best’ course of
action. It is important to note that in order to maximise the impact of
predictive analytics, the culture of the organisation must have evolved to one
of experimentation.
Once the culture
of experimentation is established, we have a learning organisation and can
become data driven. Again, it is important that experimentation permeates the
organisation, it is critical to understand some experiments will not get the
expected results, and learning from them is the point, not learning is a
failure.
Minority Report: A Beautiful Mind (11)(12)
Predictive
analytics assumes that the behaviour variables are given;
pre-emptive/pro-active analytics attempts to change the behaviour. This falls
in the realm of behavioural economics, game theory, nudging, precogs(11)… Most organisations
are not there yet, plus there may be some ethical implications (after all the
swamp hasn’t been drained yet, has it?)
In sum,
analytical maturity is critical to ensure the successful adoption of the more
advanced tools of analytics/”Data Science” (to me AI is a tool); to paraphrase
the article quoted earlier (4), people are not ‘malleable’, putty is. So as
long as we are dealing with people, change management, bringing people across
an organisation up the analytical maturity stages is important.
However, that is
not to say that it is not possible for organisation to engage technological
leapfrogging. One of the interesting aspects of technology is that you do not
need to understand it fully to use it to make decisions. As someone said in the
Global Analytics Summit in Bali last year (you can find a piece I presented
there in a previous blog post (13)), “managers
who know how to use data to make decisions will replace managers who don’t”.
Once a
technology gets to the bottom of the through of despair in the hype cycle (14),
what brings is back up via the slope of enlightenment is that it starts getting
applied beyond the purely technical hype, real life applications are what make
technologies reach the plateau of productivity.
In Sum
To me it’s our
job as analytics/“data science” practitioners to help organisations go through
the analytical maturity. What about new technologies to come you would ask? The
answer is that if an organisation is mature enough, has become data-driven, it
will naturally seek to adopt new technologies and be competing with data.
So to answer my
friend, yes, if an organisation is not doing analytics, it can’t simply adopt
AI. However, it is not necessarily take that long to learn and become
analytically mature, as long as there is a framework and commitment through-out
to do so. And I would like to add, I certainly believe in technology
leap-frogging, I am betting on it.
- https://hbr.org/2017/06/if-your-company-isnt-good-at-analytics-its-not-ready-for-ai
- https://asia.nikkei.com/Business/Chinese-online-insurer-leaves-traditional-rivals-in-the-dust
- https://en.wikipedia.org/wiki/Field_of_Dreams
- https://www.techrepublic.com/article/85-of-big-data-projects-fail-but-your-developers-can-help-yours-succeed/
- http://thegatesofbabylon.blogspot.com/2018/08/if-you-dont-have-phd-dont-call-yourself.html
- https://www.tampabay.com/florida-politics/buzz/2018/03/20/and-i-was-in-florida-with-25000-people-going-wild/
- https://www.imdb.com/title/tt6938828/
- https://www.goodreads.com/book/show/336258.Information_Dashboard_Design
- https://www.imdb.com/title/tt0088763/
- http://drewconway.com/zia/2013/3/26/the-data-science-venn-diagram
- https://www.imdb.com/title/tt0181689/
- https://www.imdb.com/title/tt0268978/
- http://thegatesofbabylon.blogspot.com/2018/01/
- https://en.wikipedia.org/wiki/Hype_cycle