Story image

Why 2019 is the year of post-modern analytics

04 Dec 18

Data is the new currency on which the modern economy thrives. Businesses that possess the know-how to effectively translate data into insights hold the key to unlocking unparalleled power and influence, and the best part is that every business today has the potential to become a powerful and influential entity. 

A 2018 Data Literacy Index conducted by Qlik uncovered that large enterprises that have higher corporate data literacy experience US$320-$534 million in higher enterprise value. This shows that there is a clear business case for leaders to commit resources to ensure that employees are well equipped to operate in the new, data-driven environment. The best way to do so is to fundamentally change the way data and analytics are being managed and used by making it accessible to every employee across all levels. Moving into 2019, here are five trends that demonstrate why it will undoubtedly become the year of post-modern analytics. 

1. Artificial Intelligence will make analytics more human, not less

The use of AI will make analytics more human, not less. Designing AI around humans will result in a higher impact for organisations over the next five years, rather than designing to remove humans from the process.

Across the information value chain - from reading the data, to preparing it, to critically analysing it with less bias, to presenting contextual results - AI can and will remove many of the bottle-necks that make users give up. Machine learning and telemetry will also capture the power of the collective, which can be fed back in a virtuous loop, further improving and contextualising the user experience. The paradox is, therefore, that data and analytics will become more human than ever, with AI in the mix.

What is often overlooked is the huge gap between the data created, and the human ability to process it and act upon it. There is also a gap between the availability of today’s analytical tools and their adoption within organisations. Those gaps can and should be closed, and the result will empower humans. 

2. Analytics will be more pervasive, and even re-shape business processes 

Embedding analytics into business processes is not new but is now becoming mainstream. Users want analytics in their workflows as they help make data more actionable and increasingly also accessible in real-time. All of this is being fueled by machine learning and AI, which can provide contextualised insights and suggested actions. 

It’s the foundation of “continuous analytics” in which real-time analytics will be gradually integrated within a business operation or IoT device, processing data to prescribe actions in response to business moments and at the edge. In the next five years, intelligent applications will be ubiquitous.

New technologies, such as robotic process automation and process mining look at digital footprints and from a chain that can further automate or re-shape business processes. For example, when a customer places an order for a product online, it will have the ability to automate and re-shape sub-processes including receiving, fulfilling and invoicing the order.

3. The focus will shift from putting data in one place to attaining one view of the data

The ability to have a single view of all data has never been more important than now. Data is coming from all different directions, speeds, and formats, and being able to control that will be one of the key markers for empowerment and success in the data age. Historically, it has been cumbersome to put all data in one place, hence the emergence of data silos and governance issues. 

There are two major forces that make getting a single view of data while keeping it where it resides a possibility. Firstly, as different vendors come together and standardise data models, what this means is that cloud-based data sources will have more consistent formats. Secondly, and more importantly, is the emergence of enterprise data catalogues. Accessible in a hub, data catalogues make it possible to audit the entire distributed data estate, delivering a shop-for-data marketplace experience. 

4. Performance takes centre-stage as analytics scale

Performance has been a bottle-neck for distributed big data at scale and the reason why many Hadoop projects failed to become much more than cheap storage. Breakthroughs have recently been achieved through indexing, caching and pre-preparing very large and distributed datasets. 

As companies of all sizes start to increase their adoption of hyperscale data centres, performance will rise in the selection criterion. Some organisations have moved their data back through ’re-patriation‘ because they have not been seeing strong enough performance. 

This becomes even more important in an IoT application. Increasingly, more workloads will run locally or at the edge to avoid latency. In short, efficient performance will be a deciding factor for how architectures will look – centralised or distributed.

5. Analytic platforms will evolve into virtuous systems, feeding of participation

BI and analytics are the most effective for organisations when viewed as a system and not simply as a series of artefacts and tools. An important difference is that individuals use tools, but people participate in systems. 

A post-modern system contains a host of people with differing roles, skills or intentions. In this system, humans interact with non-human participants such as digital services, bots, intelligent agents, extensions, algorithms and so on, resulting in exchanges and learning that increases the value of the system, augmenting both the human and machine intelligence within it. 

An open, self-learning system that improves with further participation will define how future platforms look like and enable data democracy and analytics empowerment.

Article by Julian Quinn, Senior Vice President, APAC, Qlik

MulteFire announces industrial IoT network specification
The specification aims to deliver robust wireless network capabilities for Industrial IoT and enterprises.
Google Cloud, Palo Alto Networks extend partnership
Google Cloud and Palo Alto Networks have extended their partnership to include more security features and customer support for all major public clouds.
DigiCert conquers Google's distrust of Symantec certs
“This could have been an extremely disruptive event to online commerce," comments DigiCert CEO John Merrill. 
Schneider Electric's bets for the 2019 data centre industry
From IT and telco merging to the renaissance of liquid cooling, here are the company's top predictions for the year ahead.
$43m tax-break for Google to house data centre in Ohio
There could soon be another hyperscaler on the horizon as the Ohio Tax Credit Authority has laid the bait for Google.
China to usurp Europe in becoming AI research world leader
A new study has found China is outpacing Europe and the US in terms of AI research output and growth.
Google says ‘circular economy’ needed for data centres
Google's Sustainability Officer believes major changes are critical in data centres to emulate the cyclical life of nature.
52mil users affected by Google+’s second data breach
Google+ APIs will be shut down within the next 90 days, and the consumer platform will be disabled in April 2019 instead of August 2019 as originally planned.