A lot of the value in data analytics programmes can only come if you can source a real-time view of your entire organisation’s data.

However, one of the challenges with this is that, in a distributed organisation, branch offices often store data in local siloes. These insights are locked, preventing many from gaining a full view of what’s going on and limiting the value of any insights derived from big data projects.

And of course, service demands on IT at remote and branch offices are no different than at central offices, except there are often fewer IT staff to handle requests and they’re often isolated from HQ. In today’s data driven environment operating in siloes makes little sense. In fact, consolidation and integration is key to reducing cost, complexity, managing regulatory risk and delivering a seamless customer experience.

One crucial opportunity to knock down these siloes is the opportunity to take the content from the local branch office and add it to the wider organisation’s ‘big data’ analytics programs.

Data lakes are key to this, offering a way to assemble information created across a business, including content important from third party sources and services, eliminating or at least significantly reducing the ‘Extract, Transfer, Load’ time necessary at the start of any big data programme. And as we all know, big data analytics is vital for spotting and solving previously unaddressed issues and capitalising on emerging trends.

New technologies are emerging which enable organisations to add local branch office data to its central ‘big data’ lake more efficiently and cost-effectively than ever before, all in a way which enables the deployment of new technology without disrupting users, with the flexibility to roll-back data to a previous state according to any given need.

The benefits are clear, but here are four things to consider before using these solutions to bring your branch office into your data lake architecture:

1. A unified architecture will simplify management and reduce operational costs:

Plan to consolidate to the same scale-out platforms you have within your data centre if possible. This will reduce the burden of operational management and make centralised management more straightforward.

 

2. Set your policies for your regulatory and customer risk profile:

Recovery options within a data lake give you a lot more flexibility to ensure you’re aligned with your particular industry’s demands. For example, if it is important to retain multiple ‘states’ of your key data at different stages during each working day

 

3. Deliver multiprotocol support:

You may end up experimenting and/or building big data analytics applications on a number of different platforms, including Hadoop, Hortonworks, Cloudera etc., so you’ll need your data infrastructure to support them.

 

4. Capitalise on the value of software:

The hardware you’ll need is largely "commodity;" it’s the software and the updates that come down to it over the years that will deliver value. So look to find the features, scale, control you need and build out an enterprise license agreement that covers the lifetime of your project.

 

At a time when understanding the nuances of customers or a particular region means the difference between just surviving or thriving, finding ways to incorporate the hidden value within branch data into wider analytics programmes is crucial. Adding branch data, and even archived cold data in public cloud data stores, into your data lake, could deliver the critical insights your business needs for future success.