28 June 2020

Ecosystem View: One way to think about data


(A guest post from Andrew Brice. Data analytics and visualisation is Andrew's passion, turning data into insights. Enjoy some fascinating insights here: http://www.zyan.co.nz/)

The calls were stacking up. They weren’t meant to be, this was just a normal call centre day. Call agents were racing, management hearts were racing, but no one was quite sure where the finish line was. Or why there was even a race.

Afterward, the vast swathes of data collected by the call system was able to be analysed and it started to become clear as to what had occurred and why. The needle in the haystack was found.

The thing is, it was all preventable. The needle could have been tracked if only they had been analysing their own data, ideally with an ecosystem perspective. If only.

OK, so how would that have worked?

Example dashboard for call flow monitoring
Two parts of the organisation already knew this was not a normal call centre day. It was an end-of-year filing date for a subset of clients (yes, it’s a Government agency). Not a familiar filing date, but a statutory one for these clients. The business knew and so did the call centre. Every year, on the same date, a spike occurred. But the call centre didn’t analyse call patterns at a deep enough granularity so they hadn’t picked this up. They were busy with “real” call volumes… The data was collected but it was never sufficiently deployed through to analytics. Of course, you kind of hope the business might have reminded the call centre. Sigh.

As the calls arrived, the call notes were not being augmented with robust, pre-determined, keywords which would have enabled rapid and clear analysis of that entered text on a near-real-time basis. The same effect could be achieved by simply asking call agents what all the calls were about, but (to be fair), the call agents were rather busy.

That keyword data might have shown that the callers were all identifying as “Directors” and most were asking (in quite elongated conversations), about how to log-in to one of the organisation’s systems. Which, just about then, went splat. That’s a technical IT term which means it stopped working for no particular reason. Now the keyword flow changed from password issues to “the system won’t work” issues.

Luckily, the super-heroes in IT noticed a red-flashing light (I jest) and rebooted the server. The system returned. However, the rest of the morning it kept splatting, being rebooted, … You get the idea.

The data that IT had could have painted a pretty clear picture that this was going to happen. It’s just that they didn’t really care too much about that one little server, they had lots of other more important ones. But their data (for the naughty server) showed annual usage growing at around 4% a year over the last few years and that, at its peak (which oddly only seemed to occur once a year), last year was at 88%. Do the maths, the poor server was over-run. But each month, IT drew beautiful charts showing all the organisations’ servers were paragons of good health. Sometimes, they even shared these charts with the business (not the call centre though).

The final piece of the jigsaw was understanding why logging-in was proving such a problem for this subset of clients and this particular service. Connecting call data to client CRM data quickly showed that this was almost the only interaction these clients had with the organisation each year. The call records then showed that “remembering their password” or “knowing what data to enter” were the key tags to describe the clients’ issues. Eminently common issues and ones that can readily be addressed with better user interface design, education of agents and users, reminder communications, and perhaps adopting external user-validation processes. Of course, the next crisis arrived so these sorts of improvements never actually happened.

So, what lessons do we learn from this? It might be useful to use a simple graphic to show one way of contemplating how to think about getting good data that enables quality analytics.

Data has different owners who, traditionally, impose their own standards, definitions, segmentation, and so on. Adopting co-ordinated governance practices (just like for finance or risk) is a massive enabler of good data.

Data is moving from controlled source systems to all-over-the-place. And there’s so much more of it than there ever used to be. Active and consistent curation is needed to enforce (in a friendly way) an organisation’s data governance model.

Data now lives in many places. Collecting that data together in coherent and consistent ways matters. It’s also quite difficult. This isn’t about data marts, it’s about descriptors and licences and metadata and data packaging. Building confidence in the quality of the data being collected.

Data never seems to arrive in perfect shape. It always needs consolidating or averaging or building into new fields or any one of a myriad number of ways that data gets augmented. It might need to be assembled with other data (IT data and call centre data for example) or shaped a certain way for a particular visual. But there are pretty consistent augmentation requirements and there are pretty consistent ways of standardising and automating such augmentation. Done right, we can join a call with a CRM entry and associate them with a system and then to a server. Now we can see much of the ecosystem. And we can keep rerunning the analysis as we make improvements to see if those improvements really are working.

We also have an endless need to categorise data, to tag data. This is a phone call, it’s a happy person on the line, it’s about this system, it is from that person. These are all ways of adding tags to data. Tags are incredibly useful because that’s how we can start joining things up, clumping them together, showing how processes flow, visualising pathways to outcomes. Tags are exciting. But if we all invent our own tags then value is lost. Governance and curation are how tags are kept moderated.

And, finally, we get to the fun bit. Actually, deploying the data in ways that enable rapid, effective, and (hopefully), elegant visuals that communicate the story inherent in the combined data and which enable audiences to quickly understand and react to complex ecosystems. It might also be deploying into AI systems to read call agents’ notes in real-time and then to tag calls automagically.

Deployment is the real value proposition. But only if you do something about what you’ve learned.

(Andrew Brice works with New Zealand government agencies on the visualisation of business ecosystems using complex, multi-faceted, data.)

No comments:

Post a Comment