How better data management cuts costs and improves efficiency

In the current challenging economy, all businesses are trying to find ways to reduce costs and create efficiencies. IT teams are by no means immune, but not all IT costs are easy to reduce: some, like cloud storage, are seen as a necessary cost of business. Analyst firm Canalys expects worldwide cloud infrastructure services expenditure to have risen 28% year on year in the third calendar quarter 2022, due to a range of factors including inflationary pressures and rising energy prices. Because of this, IT must look at the immediate returns on investment – of both money and time. One overlooked way to achieve rapid ROI is to look at a modern data management strategy.

Reduce data volumes

An organisation will often be paying too much for storage because it holds onto far too much data. And much of these data volumes will be the same data that’s duplicated across employees, business units and sat in siloes. Classifying data and deduplication is often pushed down the action list because it’s complex and the annual costs of storage are often dismissed as a tax of running the business.

The truth is that much value can be derived from classifying and deduplicating data. This means for organizations to classify the data first, and deduplicate only that which is required to be retained by the company, to make this process as cost/time efficient as possible.

Switch off old hardware

Decluttering and streamlining how data is stored and managed can reduce the amount of physical storage required, and simplify how and where data is stored. When the whereabouts of data is fully understood, just one “single source of truth” for every datapoint exists. In this situation it should be possible to begin reducing legacy hardware from the IT environment. There is less storage and server needed and organisations have moved past the point of paralysis over what stays or goes because they now know what data is stored where.

Simplify and automate workloads

As a consequence of those first steps in data de-dupe, tasks which were once complex – such as classification, data aging, tiering (for lowering cost), Ops efficiencies, basic hygiene and security patching – can be simplified, and it may be possible to automate workloads which previously needed manual intervention because of data complexity. All of these outcomes mean that the IT Team time can be freed up for other tasks, and that the parts of the business which use the streamlined systems get what they need faster – and perhaps even with greater assurance of accuracy too.

Streamline storage

Backups should be included in the quest to reduce data volume because modern, cloud-based backup systems can reduce costs significantly. There are also operational benefits, such as allowing for a range of search and restore options that aren’t available with legacy backup infrastructure. A report by Forrester found organisations achieved a 66% reduction of backup and data management costs with modern solutions versus legacy vendors.

Classification of data, mentioned earlier, has a key role to play in streamlining storage, improving cost efficiencies, enabling your sustainability agenda, and giving you predicable technology trends. If you classify your data according to your Relevant Record Strategy, you will be able to make a defensible decision about it. You will know what you need to keep, and where, allowing you to utilise primary, nearline, and offline solutions to maximise efficiency. More importantly though, you will only store that which you need to store, enabling defensible deletion decisions which will quickly control spiralling data growth.

Reduce security and compliance risks

Deduplication and classification will also help to stay compliant with regulation around personally identifiable information, such as HIPAA, GDPR or other national rules[CG1], as well as other regulations, such as DORA, that set out operational resiliency requirements.

What’s more, a better understanding of data means reduced security vulnerabilities. If you do not know where data is or what it is, then you have no idea of its importance or if it’s been compromised. Once you know where your data is, what it is, and how secure it is, you reduce its vulnerability rating, which also allows you to prioritise any patches required for vulnerabilities to a priority order of rollout across the organisation, rather than a basic rollout schedule. Going further, the compression, deduplication, encryption, classification and shorter backup times from modern data management enables much faster recovery at scale and improves the future spend profile, as you now know exactly what you are keeping, where and for how long.

Index old tapes for faster retrieval and reduces compliance risks

Holding legacy data on tape may be important for legal or compliance reasons, but so is the ability to find and restore specific records. Too often, organisations save everything instead of coming up with more sensible archive strategies. Instead indexing and classifying the records needs to be a priority.

Forrester’s report noted that the organisations it interviewed saved hundreds of thousands of dollars on proprietary hardware, maintenance and off-site storage related to tapes, by storing the tape’s contents in the cloud instead. All risk of the loss or damage of the physical tapes themselves is mitigated by moving their data to cloud, data can be indexed making retrieval easier, and as the process of conversion to cloud is provided by the vendor, there are no legacy hardware costs to consider. Classification ensures you only move (and keep) the data that you need to.

Whatever else a business might do with its data, it certainly won’t stop generating more. Implementing measures to manage data better can inevitably improve efficiency and reduce costs. At a time when the economic situation is challenging, this could prove a blessing. Maybe it is time for data management strategies to bubble to the top of the to do list, because getting best practice in place today will streamline how data is collected, stored and used tomorrow.

Mark Molyneux

Mark Molyneux is CTO for EMEA at Cohesity

How E-commerce Marketers Can Win Black Friday

Sue Azari • 11th November 2024

As new global eCommerce players expand their influence across both European and US markets, traditional brands are navigating a rapidly shifting landscape. These fast-growing Asian platforms have gained traction by offering ultra-low prices, rapid product turnarounds, heavy investment in paid user acquisition, and leveraging viral social media trends to create demand almost in real-time. This...

Why microgrids are big news

Craig Tropea • 31st October 2024

As the world continues its march towards a greener future, businesses, communities, and individuals alike are all increasingly turning towards renewable energy sources to power their operations. What is most interesting, though, is how many of them are taking the pro-active position of researching, selecting, and implementing their preferred solutions without the assistance of traditional...

Is automation the silver bullet for customer retention?

Carter Busse • 22nd October 2024

CX innovation has accelerated rapidly since 2020, as business and consumer expectations evolved dramatically during the Covid-19 pandemic. Now, finding the best way to engage and respond to customers has become a top business priority and a key business challenge. Not only do customers expect the highest standard, but companies are prioritising superb CX to...

Automated Testing Tools and Their Impact on Software Quality

Natalia Yanchii • 09th October 2024

Test automation refers to using specialized software tools and frameworks to automate the execution of test cases, thereby reducing the time and effort required for manual testing. This approach ensures that automation tests run quickly and consistently, allowing development teams to identify and resolve defects more effectively. Test automation provides greater accuracy by eliminating human...

Custom Software Development

Natalia Yanchii • 04th October 2024

There is a wide performance gap between industry-leading companies and other market players. What helps these top businesses outperform their competitors? McKinsey & Company researchers are confident that these are digital technologies and custom software solutions. Nearly 70% of the top performers develop their proprietary products to differentiate themselves from competitors and drive growth. As...

The Impact of Test Automation on Software Quality

Natalia Yanchii • 04th October 2024

Software systems have become highly complex now, with multiple interconnected components, diverse user interfaces, and business logic. To ensure quality, QA engineers thoroughly test these systems through either automated or manual testing. At Testlum, we met many software development teams who were pressured to deliver new features and updates at a faster pace. The manual...