Data Architecture

The More You Refine Stored Data, the More Valuable it Becomes

Actian Corporation

April 1, 2020

stored data vector illustration

Having a lot of data available to you is a good thing, right? That depends. Data is a raw material, like a mineral mined from the ground. It contains a potential for creating value, but that potential is only realized through refinement.

Your company produces a lot of data every day (every second really). Merely creating and/or possessing this data doesn’t mean it is generating value for you. Harvesting value from your company data requires a transformation process to convert data into information, actionable insights, to decisions, and then to action.

Data management professionals, IT staff, and business analysts are the people responsible for guiding data transformation. They employ a series of refinement steps to convert the raw materials that your operations generate into meaningful and actionable insights that decision-makers across the company use to direct your staff, processes, and resources. Here is an overview of the refinement steps your data goes through, and the value addition that takes place along the way.

Collection

Data exists in your operations whether you collect it or not. The first step in data refinement is collection. This takes place within your operational systems, embedded sensors, and transactional workflows being executed throughout your company. Some data is collected in real-time through sensors, telemetry, and monitoring, while other data is collected periodically (perhaps hourly or at the end of the day). Data collection is all about measurement. The data management adage goes, “If you don’t measure it, you can’t manage it.” Extending this a step further, if you don’t collect the data, you can’t use it for decision-making.

Aggregation

There are many data sources across your organization, and no single source contains all the information that is needed for effective decision-making. Why is this?… because each data source provides a point of view on your operations. Using a single data source is like walking around a sports arena at night with only the light of a flashlight – you only see a very narrow view of your environment, not the big picture. Data aggregation brings the data from various sources together in one place, like illuminating a bunch of lightbulbs in that sports arena. Some data will overlap, so it can be filtered out, and there will be some gaps and shadows, but aggregation gets you one step closer to seeing the bigger picture of your operations.

Reconciliation

Once you have your data aggregated in one place, the next step in the refinement is reconciling the different data sets together to address gaps, overlaps, and conflicting information. This is also sometimes called data harmonization. A way to image this is considering the days before digital cameras when people took photos on film. To create a panoramic image, you took multiple pictures of adjacent scenes and then (after waiting for them to be developed) aligned the images by overlapping the frames into a panoramic view. Data reconciliation is similar, although considerably more complex. Some of the factors used in data reconciliation are data source, data quality when the data was captured (because you’re not viewing a still image, business data is a moving target). The result of data reconciliation is a unified data set that includes inputs from all your data sources.

Categorization

Categorization (often called cataloging) is the first step in understanding the content of your data. The purpose of categorization is to help you understand “what your data is.” Note, this is different from understanding “what your data means,” which is addressed in the next step. The best way to understand data categorization is a library full of books. Individual books represent different pieces of data. Librarians use a cataloging system (Dewey decimal system, the library of congress, etc.) to sort and organize books according to their content.

In the business world, companies have data metamodels which provide the cataloging structure. Categorization is all about aligning operational data (from whatever source it was collected) to these metamodels, so like concepts (such as customer data) can be analyzed together. This is when data is transformed into information.

Analysis

Data analysis is all about understanding what your information means. Data and business professionals are summarizing, sorting, filtering, correlating, projecting, performing trend analysis to refine categorized information into meaningful and actionable insights about your business. It’s interesting that the data showed a specific process step took 2.385 seconds. It’s informative to know that the process measurement was the time it took to authorize a credit card transaction. But is that number good or bad? Is it relevant? Does it indicate something is wrong? Does someone need to initiate action because of it? Data analysis is the refinement step that converts information into insights about your business.

Presentation

Possessing data, information, and insights does not create value for your organization. Value comes from the decisions you make and the actions that result from interpreting the data. The last step in the data refinement process is taking the insights that you’ve generated and presenting them to decision-makers, system operators, and making them available for automation systems. Just as you aggregated data earlier in the process, this step involves disseminating, publishing, and visualizing the insights for consumption.

The quality of data insights available for presentation is directly related to the effectiveness of your collection, aggregation, reconciliation, categorization, and analysis processes. Actian provides a set of data management capabilities to help your staff in orchestrating the refinement process – enabling not only these necessary steps but also the implementation of robust activities within these steps. Data becomes more valuable, the more you refine it.

With the right tools, you will be able to develop better insights faster. This will lead to better decisions and greater value realization for your company. The Actian Data Platform Data Warehouse includes connectors to hundreds of data sources and functions to refine and transform raw data into information.

You can learn more about Actian’s Real-Time Connected Data Warehouse at https://www.actian.com/solutions/connected-data-warehouse/

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale, streamlining complex data environments and accelerating the delivery of AI-ready data. The Actian data intelligence approach combines data discovery, metadata management, and federated governance to enable smarter data usage and enhance compliance. With intuitive self-service capabilities, business and technical users can find, understand, and trust data assets across cloud, hybrid, and on-premises environments. Actian delivers flexible data management solutions to 42 million users at Fortune 100 companies and other enterprises worldwide, while maintaining a 95% customer satisfaction score.
Data Integration

Five Reasons Why You Need an Enterprise Data Integration Strategy

Actian Corporation

March 31, 2020

data integration strategy

Modern businesses don’t run on a single application – they leverage many IT systems to provide the capabilities to enable operational processes and users to be effective. To ensure complex IT environments run smoothly, companies are placing increased focus on how systems are integrated, and the activities required to manage integrations across the ecosystem. While system integration is a multi-layered challenge, the essential part to get right is your enterprise data integration strategy.

An enterprise data integration strategy is a set of policies, processes, design guidelines, and governance controls that you put in place to ensure data integrations are implemented consistently, controlled centrally, and are fully supported by your IT function. If you think about it, data integrations are applications in themselves.

They are a set of functions that move data from one system to another, monitor the flow of data, enforce security rules, and enable business processes. When you consider how many integrations your company has (typically 3-5x the number of applications), it is clear why you need a holistic enterprise data integration strategy to ensure these critical IT components are well managed.

The following is a list of 5 reasons why you need an enterprise data integration strategy for your company

1. Save Time and Resources in Building Integrations

How many ways are there to connect two systems together? If you don’t have any standardized design patterns for data integration, the answer is infinite. More importantly, how much time will your expensive IT staff spend brainstorming and designing novel ways of doing integrations that result in increased complexity and risk a few years down the line when your IT service management (ITSM) team is trying to support them? An enterprise data integration strategy should include a set of design patterns for integrating IT systems that your developers can choose from. If you have a data integration platform like Actian DataConnect in your environment, you can use the integration templates as a starting point.

2. Ensure Business Continuity With Fewer Disruptions to Business Processing

You have Service Level Agreements (SLAs) for application performance and availability, do you have SLAs for your data integrations? Most companies don’t. When an integration fails, how quickly are you able to respond and repair it? An enterprise data integration strategy should include a set of guidelines for how to monitor the availability and performance of your data integrations to ensure any failures that could impact business services and operational processes are identified quickly and resolved. Most companies have ITSM functions in place, with incident management teams ready to respond to application and hardware failures. With the right data integration strategy, you can enable enterprise support for integrations as well.

3. Centralized Governance and Management Lowers Risk

Your enterprise data integration strategy should include a plan for how you will manage integrations once they have been developed. This includes things like access control (who can see the integrated data), change management processes, rules for extending integrations (re-using them for other purposes), management of system credentials, and data encryption. The strategy must go beyond a set of published guidelines and be supported by a data management system that can help you enforce controls and monitor compliance across the organization. Centralized governance and visibility is a critical part of managing both the risk of business disruption and ensuring efficient use of organizational resources.

4. Faster Response to Threats

Data security threats are a real problem for organizations that have a high dependency on IT systems. Your enterprise data integration strategy provides your company with an extra set of tools for defending against intrusion and responding to data security threats. Managing your integrations in a consistent way (with the governance controls discussed earlier) not only enables you to control the free-flow of information across your organization, but it also means that you have control points to shut-off that flow if a threat is identified. With consistent integration patterns and central control plane, you can also (confidently) shut any backdoors that could enable an intruder to bypass your security. Once a threat is contained, you can then update credentials, secure the environment, and restore business processing to normal.

5. Increased Agility in Times of Change

Recently there has been a lot of discussion about enterprise business agility – the need for companies to adapt their processes, systems, and strategies in response to forces and events (both inside and external to the company like COVID-19). An agile enterprise data integration strategy that is consistently applied across the organization and supported by the right set of data integration tooling can help your company reduce the technical debt (the spaghetti pile of system integrations) that prevent you from safely implementing changes. It can also provide you the tools to change your existing integrations and create new ones faster – meaning you can change your overall IT systems more quickly to respond to evolving business needs and strategic opportunities.

Actian DataConnect is an industry-leading data integration platform, enabling you to connect anything, anytime anywhere. DataConnect can help bring your enterprise data integration strategy to life – providing a centralized set of tools for implementing and managing the data integrations across your company.

To learn more, visit DataConnect.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale, streamlining complex data environments and accelerating the delivery of AI-ready data. The Actian data intelligence approach combines data discovery, metadata management, and federated governance to enable smarter data usage and enhance compliance. With intuitive self-service capabilities, business and technical users can find, understand, and trust data assets across cloud, hybrid, and on-premises environments. Actian delivers flexible data management solutions to 42 million users at Fortune 100 companies and other enterprises worldwide, while maintaining a 95% customer satisfaction score.
Data Analytics

Real-Time Reporting Platform for Operational Systems

Actian Corporation

March 30, 2020

interface showing future computer technology for operational systems

Digital transformation has drastically expanded the use of operational systems, IoT, and other technologies within companies’ business processes. This is a good thing.  It means that end-users, whether they be employees, partners, or customers, are more productive. The introduction of modern operational technologies and integrating them into business processes is only the first step in the digital transformation journey. Next comes a focus on managing operational data and developing real-time reporting platforms that give operators and decision-makers better control over digitally transformed processes.

Digital Transformation has Created More Real-Time Data

A farmer with a tractor is more productive than a farmer with a hoe and rake. That was the story 20 years ago when companies embraced the first wave of IT in things like CRM and ERP. Today there is a new story “A farmer with a drone, a robotic tractor, and data at his disposal can increase crop yield and bigger profits.” The difference here isn’t the technology – technology is continuously improving and improves efficiency/productivity incrementally. The change is how the modern farmer uses data. New farm technology generates real-time data, but it is the reporting platform that enables the farmer to make better-informed decisions, directing his/her time and the farm’s resources in more effective ways.

Data-Driven Decision-Making Drives Value and Profits

Most people reading this article aren’t farmers – this situation holds true for other businesses too. The digital transformation brought with it a technology upgrade and the introduction of new operational systems in all industries. Retailers are using new point-of-sale (POS) technology for tracking each individual item that is sold and RFID on shopping carts to track customer flow throughout their stores. This new technology is excellent, but it is the reporting platform that enables them to do dynamic inventory management, optimize stock levels to maximize inventory turnover and store layouts and product offerings to changing customer preferences.

Hospitals and healthcare companies are using recently installed healthcare management systems to capture patient data, share information between providers, and orchestrate orders for tests and medication. These systems make doctors, nurses, and other staff more effective in directing patient care. What you don’t see as a patient is the additional layer of technology on top of the operational system. A real-time reporting platform is used by clinic managers, scheduling staff, and others (similar to the retail example) to ensure that the provider’s time is optimized, exam rooms, and other facilities are fully utilized, supplies, and medications are stocked and available. The data collected from the operational systems and operational processes are aggregated in a reporting platform to enable real-time management decisions.

Telecommunications companies have implemented monitoring within their networks, in mobile apps, and integrated into customer devices (like cable set-top boxes) to capture data about customer viewing preferences and usage of their services. This data is streamed into a real-time reporting platform where it can be analyzed for patterns and opportunities to increase revenue, calculate costs, and identify system issues impacting user experience. Operators and decision-makers then use these insights to optimize networks for performance and peak utilization, provide personalized content offerings/ads to customers, and direct repair and maintenance efforts.

Real-Time Reporting Platforms Unlock the Value Potential of Operational Systems

If you are like many companies, you’ve spent the past few years doing technology upgrades focused on business processes and end-user experience. Are you harvesting the full value from these operational system investments? Real-time reporting platforms like Actian can help you unlock the unrealized potential of your operational systems. Actian can help you aggregate streaming data from all your data sources into one place, analyze trends/patterns, convert data into meaningful insights, and get those insights into the hands of operators and decision-makers faster.

This isn’t about enabling your users and processes; it’s about optimizing your business. Real-time operational analytics and reporting help you identify opportunities, risks, and issues that need to be addressed faster. By seeing the problem sooner, you can address it sooner. By addressing the problem sooner, you can maximize the opportunities, mitigate the risks, and address the issues to optimize both your costs and revenue. The end result… more profits for your business. This is the goal, and a real-time reporting platform from Actian can help you achieve it.

Learn more about Actian’s real-time connected data warehouse at https://www.actian.com/solutions/connected-data-warehouse/.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale, streamlining complex data environments and accelerating the delivery of AI-ready data. The Actian data intelligence approach combines data discovery, metadata management, and federated governance to enable smarter data usage and enhance compliance. With intuitive self-service capabilities, business and technical users can find, understand, and trust data assets across cloud, hybrid, and on-premises environments. Actian delivers flexible data management solutions to 42 million users at Fortune 100 companies and other enterprises worldwide, while maintaining a 95% customer satisfaction score.
Data Intelligence

Everything You Need to Know About Data Ops

Actian Corporation

March 26, 2020

data ops

“Within the next year, the number of data and analytics experts in business units will grow at three times the rate of experts in IT departments, which will force companies to rethink their organizational models and skill sets” – Gartner, 2020.

Data and Analytics teams are becoming more and more essential in supporting various complex business processes, and many are challenged with scaling the work they do in delivering data to support their use cases. The pressure to deliver faster and with higher quality is causing data & analytics leaders to rethink how their teams are organized.

Where traditional waterfall models were implemented and used in enterprises in the past, these methodologies are now proving to be too long, too siloed, and too overwhelming.

This is where Data Ops steps in: a more agile, collaborative, and change-friendly approach for managing data pipelines.

Data Ops Definition

Gartner defines Data Ops as being a “collaborative data management practice focused on improving the communication, integration and automation of data flows between data managers and data consumers across an organization”. Basically, it makes life easier for data users.

Similar to how DevOps, a set of practices that combines software development (Dev) and information-technology operations (Ops), changed the way we deliver software, DataOps uses the same methodologies for teams building data products.

While both agile frameworks, DataOps requires the coordination of data and anyone that works with data across the entire enterprise.

Specifically, data and analytics leaders should implement these key approaches proven to deliver significant value for organizations:

  • Deployment Frequency Increase: Shifting towards a more rapid and continuous delivery methodology enables organizations to reduce the time to market.
  • Automated Testing: Removing time-consuming, manual testing enables higher quality data deliveries.
  • Metadata Control: Tracking and reporting metadata across all consumers in the data pipeline ensures better change management and avoids errors.
  • Monitoring: tracking data behavior and the usage of the pipeline enables more rapid identification on both flawed – that needs to be corrected – and good quality data for new capabilities.
  • Constant Collaboration: communication between data stakeholders on data is essential for faster data delivery.

Who is Involved in Data Ops?

Given the importance of data and analytics use cases today, the roles involved in successful data project delivery are more numerous and more distributed than ever before. Ranging from data science teams to people outside of IT, a large number of roles are involved:

  • Business analysts
  • Data architects.
  • Data engineers.
  • Data stewards.
  • Data scientists.
  • Data product managers.
  • Machine Learning developers.
  • Database administrators.

As mentioned above, a Data Ops approach requires fluid communication and collaboration across these roles. Each collaborator needs to understand what others expect of them, what others produce, and must have a shared understanding of the goals of the data pipelines they are creating and evolving.

Creating channels through which these roles can work together, such as a collaboration tool, or metadata management solution, is the starting point.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale, streamlining complex data environments and accelerating the delivery of AI-ready data. The Actian data intelligence approach combines data discovery, metadata management, and federated governance to enable smarter data usage and enhance compliance. With intuitive self-service capabilities, business and technical users can find, understand, and trust data assets across cloud, hybrid, and on-premises environments. Actian delivers flexible data management solutions to 42 million users at Fortune 100 companies and other enterprises worldwide, while maintaining a 95% customer satisfaction score.
Data Intelligence

How You’re Going to Fail Your Data Catalog Project (or Not…)

Actian Corporation

March 25, 2020

data catalog difficultés

Many solutions on the data catalog market offer an overview of all enterprise data, all thanks to the efforts conducted by data teams.

However, after a short period of use, due to the approaches undertaken by enterprises and the solutions that were chosen, data catalog projects often fall into disuse.

Here are some of the things that can make a data catalog project fail…or not:

Your Objectives Were Not Defined

Many data catalog projects are launched using a Big Bang approach to document assets, but without truly knowing their objectives.

Fear not! In order to avoid bad project implementation, we advocate a model based on iteration and value generation. Conversely, this approach allows for better risk control and the possibility of a faster return on investment.

The first effects should be observable at the end of each iteration. In other words, the objective must be set to produce concrete value for the company, especially for your data users.

For example, if your goal is data compliance, start documentation focused on these properties and target a particular domain, geographic area, business unit, or business process.

Your Troop’s Motivation Will Wear Off Over Time

While it is possible to gain adherence and support regarding your company’s data inventory efforts in its early stages, it is impossible to maintain this support and commitment over time without automation capabilities.

We believe that descriptive documentation work should be kept to a minimum to keep your teams motivated. The implementation of a data catalog must be a progressive project and will only last if the effort required by each individual is greater than the value they will get in the near future.

You Won’t Have the Critical Mass of Information Needed

For a data catalog to bring value to your organization, it must be richly populated.

In other words, when a user searches for information in a data catalog, they must be able to find it for the most part.

At the start of your data catalog implementation project, the chances that the information requested by a user is not available are quite high.

However, this transition period should be as short as possible so that your users can quickly see the value generated by the data catalog. By choosing a tactical solution, based on its technology and connectivity to information sources, a pre-filled data catalog will be available as soon as it is implemented.

Does Not Reflect Your Operational Reality

In addition to these challenges, data catalogs must have a set of automated features that are useful and effective over time. Surprisingly, many solutions do not have offer these minimum requirements for a viable project, and are unfortunately destined for a slow and painful death.

Connecting data catalogs to your sources will ensure that your data consumers:

  • Reliability as to the information made available in the data catalog for analysis and use in their projects.
  • Fresh information: Are they up to date, in real time?
actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale, streamlining complex data environments and accelerating the delivery of AI-ready data. The Actian data intelligence approach combines data discovery, metadata management, and federated governance to enable smarter data usage and enhance compliance. With intuitive self-service capabilities, business and technical users can find, understand, and trust data assets across cloud, hybrid, and on-premises environments. Actian delivers flexible data management solutions to 42 million users at Fortune 100 companies and other enterprises worldwide, while maintaining a 95% customer satisfaction score.
Data Intelligence

How Spotify Improved Their Data Discovery for Their Data Scientists

Actian Corporation

March 19, 2020

spotify lexikin cover

As the world leader in the music streaming market, it is without question that the huge firm is driven by data.

Spotify has access to the biggest collections of music in the world, along with podcasts and other audio content.

Whether they’re considering a shift in product strategy or deciding which tracks they should add, Spotify says that “data provides a foundation for sound decision-making”.

Spotify in Numbers

Founded in 2006 in Stockholm, Sweden, by Daniel Ek and Martin Lorentzon, the leading music app’s goal was to create a legal music platform in order to fight the challenge of online music piracy in the early 2000s.

Here are some statistics & facts about Spotify in 2020:

  • 248 million active users worldwide.
  • 20,000 songs are added per day on their platform.
  • Spotify has a 40% share of the global music streaming market.
  • 20 billion hours of music were streamed in 2015.

These numbers not only represent Spotify’s success, but also the colossal amounts of data that is generated each year, let alone each day! To enable their employees, or as they call them, Spotifiers, to make faster and smarter decisions, Spotify developed Lexikon.

Lexikon is a library of data and insights that helps employees find and understand the data and knowledge generated by their expert community.

What Were the Data Issues at Spotify?

In their article How We Improved Data Discovery for Data Scientists at Spotify, Spotify explains that they started their data strategy by migrating data to the Google Cloud Platform, and saw an explosion of their datasets. They were also in the process of hiring many data specialists such as data scientists, analyst, etc. However, they explain that datasets lacked clear ownership and had little-to-no documentation, making it difficult for these experts to find them.

The next year, they released Lexikon, as a solution for this problem.

Their first release allowed their Spotifiers to search and browse through available BigQuery tables as well as discover past researches and analysis. However, months after the launch, their data scientists were still reporting data discovery as a major pain point, spending most of their time trying to find their datasets therefore delaying informed decision-making.

Spotify decided then to focus on this specific issue by iterating on Lexikon, with the unique goal to improve data discovery experience for data scientists.

How Does Lexikon Data Discovery Work?

In order for Lexikon to work, Spotify started out by conducting research on their users, their needs as well as their pain points. In doing so, the firm was able to gain a better understanding of their users intent and use this understanding to drive product development.

Low Intent Data Discovery

For example, you’ve been in a foul mood so you’d like to listen to music to lift your spirits. So, you open Spotify, browse through different mood playlists and put on the “Mood Booster” playlist.

Tah-dah! This is an example of low-intent data discovery, meaning your goal was reached without extremely strict demands.

To put this into Spotify’s data scientists context, especially new ones, their low intent data discovery would be:

  • Find popular datasets used widely across the company.
  • Find datasets that are relevant to the work my team is doing.
  • Find datasets that I might not be using, but I should know about.

So in order to satisfy these needs, Lexikon has a customizable homepage to serve personalized recommendations to users. The homepage recommends potentially relevant, automatically generated suggestions for datasets such as:

  • Popular datasets used within the company.
  • Dataset recently used by the user.
  • Datasets widely used by the team the user belongs to.

High Intent Data Discovery

To explain this in simple terms, Spotify uses the example of hearing a song, and researching it over and over in the app until you finally find it, and listen to it on repeat. This is high intent data discovery.

A data scientist at Spotify with high intent has specific goals and is likely to know exactly what they are looking for. For example they might want to:

  • Find a dataset by its name.
  • Find a dataset that contains a specific schema field.
  • Find a dataset related to a particular topic.
  • Find a dataset that a colleague used of which they can’t remember the name.
  • Find the top datasets that a team has used for collaborative purposes.

To fulfill their data scientists needs, Spotify focused first on their search experience.

They built a search ranking algorithm based on popularity. By doing so, data scientists reported that their search results were more relevant, and had more confidence in the datasets they discovered because they were able to see which dataset was more widely-used by the company.

In addition to improving their search rank, they introduced new types of properties (schemas, fields, contact, team, etc.) to Lexikon to better represent their data landscape.

These properties are able to open up new pathways for data discovery. In the example down below, a data scientist is searching for a “track_uri”. They are able to navigate through the “track_uri” schema field page and see the top tables containing this information. Since adding this new feature, it has proven to be a critical pathway for data discovery, with 44% of Lexikon users visiting these types of pages.”

Final Thoughts on Lexikon

Since making these improvements, the use of Lexikon amongst data scientists has increased from 75% to 95%, putting it in the top 5 tools used by data scientists!

Data discovery is thus, no longer a major pain point for their Spotifiers.

Sources:

Spotify Usage and Revenue Statistics (2019): https://www.businessofapps.com/data/spotify-statistics/
How We Improved Data Discovery for Data Scientists at Spotify: https://labs.spotify.com/2020/02/27/how-we-improved-data-discovery-for-data-scientists-at-spotify/
75 Amazing Spotify Statistics and Facts (2020): https://expandedramblings.com/index.php/spotify-statistics/

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale, streamlining complex data environments and accelerating the delivery of AI-ready data. The Actian data intelligence approach combines data discovery, metadata management, and federated governance to enable smarter data usage and enhance compliance. With intuitive self-service capabilities, business and technical users can find, understand, and trust data assets across cloud, hybrid, and on-premises environments. Actian delivers flexible data management solutions to 42 million users at Fortune 100 companies and other enterprises worldwide, while maintaining a 95% customer satisfaction score.
Data Analytics

Why You Should Offload Analytics into a Data Warehouse

Actian Corporation

March 17, 2020

Data warehouse image

Modern businesses are fueled by data. The insights that your data bring are what power decision-making, enable you to optimize business processes and respond to changing market conditions. The organizations that have data and manage it well – excel. Those who lack data or struggle to harvest actionable insights from their data have a tougher time.

One of the most significant challenges IT has as the stewards of company data is striking a balance between high-performance real-time data processing performance of individual business processes and the deep/enterprise-scale analytics required for solving the company’s biggest problems. By offloading analytics from Online Transaction Processing (OLTP) systems into a cloud data warehouse like the Actian Data Platform, your company can achieve both objectives at the same time.

Sustaining High-Performance in Your Business Systems

OLTP systems are your transactional business systems – the tools that your employees, partners, and customers interact with within the course of normal day-to-day business activities. These systems are optimized for real-time data processing (as they should be). Any impact on performance has a direct impact on your process cycle times and employee productivity. With each new business transaction, you create more data.

As the size of your OLTP database grows, the applications that run on it begin to slow down. Adding an analytics load on top of the transactional processing makes the problem even worse.  Sustaining a high-performing business system requires continuous active tuning of the OLTP system to eliminate any non-essential activities. A key technique that IT teams employ is to offload analytics processing into a data warehouse, freeing up compute capacity in the OLTP system, so business software has more system resources from which to draw.

Leveraging Change Data Capture for Real-Time Analytics

Change data capture is an analytics capability available in nearly all databases but is mostly used to populate data warehouses. What this capability does is to monitor for changes in your transactional data that might correspond with business events that represent opportunities or threats for your business. Some changes in business transactions are good… an incremental increase in sales transaction values. Other changes are adverse… such as a sudden drop in the number of users logged in to your website. Change data capture can help you understand when something is awry, so you can assess the impact and determine if any corrective action is required.

Running change data capture operations on OLTP systems can be problematic. The overhead load it places on the system has to be monitored to minimize the performance impact to your business systems. Change data capture is most valuable with larger data sets for analyzing trends. If you have a good log archive management in place, performance overheads can be contained. So, it makes sense to use change data capture to populate your data warehouse system with near-real-time business operational data.

Extend the Useful Life of Your Business Systems

Business systems like Enterprise Resource Planning (ERP), Customer Relationship Management (CRM), Human Resource Management (HRM), IT Service Management (ITSM), and eCommerce systems are costly to install, but disruptive to the business when you need to replace them. If the systems you are using today run on-premises in your data center, upgrades to hardware infrastructure to add compute capacity may require new capital outlays and/or migration to the cloud. Offloading analytics from these systems into a data warehouse can help you keep these systems running longer with existing resources – postponing the impacts of system upgrades.

Over the next few years, new business systems offerings that are cloud-native, integrate Artificial Intelligence (AI) capabilities, and have enhanced support for streaming data are poised to come to market, creating a natural time to upgrade. Extending the useful life of your existing systems gives your company the flexibility to wait for the new features that are “coming soon” and catch the next wave of emerging technology to maximize the return on investment of your upgrade projects.

Offloading analytics from your OLTP system into a data warehouse is a smart IT decision. It helps keep your business systems running faster, gives you the real-time data insights you need for agile decision-making, and extends the useful lifespan of your existing systems, so you capture the next wave of technology innovations that are just over the horizon. Actian Data Platform can help. As a data warehouse solution, the Actian Data Platform can run on-premises, in the cloud, or even as a hybrid, split across different environments giving you the analytics capabilities and scale that you will need to manage your company’s data successfully.

To learn more, visit www.actian.com/data-platform

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale, streamlining complex data environments and accelerating the delivery of AI-ready data. The Actian data intelligence approach combines data discovery, metadata management, and federated governance to enable smarter data usage and enhance compliance. With intuitive self-service capabilities, business and technical users can find, understand, and trust data assets across cloud, hybrid, and on-premises environments. Actian delivers flexible data management solutions to 42 million users at Fortune 100 companies and other enterprises worldwide, while maintaining a 95% customer satisfaction score.