Blog | Data Intelligence | | 6 min read

Data Culture: 5 Steps for Your Enterprise to Acculturate to Data

data-culture-cover

Exploding quantities of data have the potential to fuel innovation and produce more value for organizations. Stimulated by the hopes of satisfying customers, enterprises have, for the past decade or so, invested in technologies and paid handsomely for analytical talent. Yet, for many, data-driven culture remains elusive, and data is rarely used as the basis for decision-making.

The reason is that the challenges of becoming data-driven aren’t technical but rather cultural. Describing how to inject data into decision-making processes is far easier than shifting an entire organization’s mindset. In this article, we describe five ways to help enterprises create and sustain data culture at its core.

By 2023, data literacy will become an explicit and necessary driver of business value, demonstrated by its formal inclusion in over 80% of data and analytics strategies and change management programs.

What is Data Culture

”Data culture” is a relatively new concept that is becoming increasingly important to put in place, especially for organizations developing their digital and data management strategies. Just like organizational or corporate culture, data culture refers to a workplace environment where decisions are made through emphatic and empirical data proof. In other words, executives make decisions based on data evidence and not just on instinct.

Data culture gives organizations more power to organize, operate, predict, and create value with their data.

Here are our five tips for creating and sustaining data culture:

Step 1: Align With Business Objectives

“The fundamental objective of collecting, analyzing, and deploying data is to make better decisions.” (McKinsley)

Trusting your data is one of the most important tips for building data culture, as distrust in data leads to disastrous organizational culture. And to trust in data, it must align with business objectives. To drive strategic and cultural changes, it is important for the enterprise to agree on common business goals, as well as the relevant metrics to measure achievements or failures across the entire organization.

Ask yourself the right questions: How can we not only get ahead of our competitors, but also maintain the lead? What data would we need to decide what our next product offering should be? How is our product performing in the market? By introducing data into your business decision-making process, your enterprise will have already made the first step to building a data culture.

Step 2: Destroy Data Silos

In this case, data silos refer to departments, groups, or individuals who are the guardians of data, who don’t share, or don’t know how to share, data knowledge with other parts of the enterprise. When crucial information is locked away and available to only a few connaisseurs, it prevents your company from developing a cross-departmental data culture. It is also problematic on a technical standpoint: multiple data pipelines are harder to monitor and maintain, which leads to data being stale and obsolete by the time anyone uses it for decision-making.

To break data silos, enterprises must put in place a single source of truth. Empower employees to make data-driven decisions by relying on a centralized solution. A data catalog enables both technical and non-technical users to understand and trust in the enterprise’s data assets.

Step 3: Hire Data-Driven People

When building a data culture, it’s important to hire data-driven people. Enterprises are reorganizing themselves, forcing the creation of new roles to support this organizational change:

Data Stewards

Data Stewards are here to orchestrate an enterprise’s data systems. Often called the “masters of data”, they have the technical and business knowledge of data. Their main mission is to ensure the proper documentation of data and facilitate their availability to their users, such as data scientists or project managers for example.

This profession is on the rise. Their social role allows data stewards to work with both technical and business departments. They are the first point reference for data in the enterprise and serve as the entry point to access data.

Chief Data Officers

Chief Data Officers, or CDOs for short, play a key role in the enterprise’s data strategy. They are in charge of improving the organization’s overall efficiency and the capacity to create value around their data. At first, CDOs had to lead a mission to convince interest organizations to exploit data. The first few years of this mission were often supported by the construction of a data universe adapted to new uses, often in the form of a Data Lake or Data Mart. But with the exponential development of data, the role of the CDO took a new scope. From now on CDOs must reconsider the organization in a cross-functional and globalizing way. They must become the new leaders of Data Democracy.

In order to obtain the support for data initiatives from all employees, they must not only support them in understanding data (original context, production, etc.) but also help them to invest in the production strategy and the exploitation of data.

Step 4: Don’t Neglect Your Metadata

When data is created, so is metadata (its origin, format, type, etc.). However, this type of information is not enough to properly manage data in this expanding digital era; data managers must invest time in making sure this business asset is properly named, tagged, stored, and archived in a taxonomy that is consistent with all of the other assets in the enterprise.

This metadata allows for enterprises to assure greater Data quality & discovery, allowing data teams to better understand their data. Without metadata, enterprise find themselves with datasets without context, and data without context has little value.

Step 5: Respect the Various Data Regulations

If you’re in Europe, this is old news by now. With the GDPR put into place in May 2018 as well as all of the other various regulations slowly seeing the day in the United States, UK, or even Japan, it is important for enterprises to respect and follow the guidelines to conform.

Implementing data governance is a way to ensure that all personal data privacy, data security, and ensure risk management. It is a set of practices, policies, standards, and guides that will supply a solid foundation to ensure that data is properly managed thus, creating value within an organization.

Step 6 (BONUS TIP): Choose the Right Solutions

Metadata management is the new black: it is an emerging discipline, necessary for enterprises wishing to bolster innovation or regulatory compliance initiatives on their data assets. A metadata management solution offers enterprises a centralized platform to empower all data users in their data culture implementation.

For more information on metadata management, contact us.


Two weeks ago, I likened the performance of SQLite to that of a banana slug. Now, some may consider that a bit of hyperbole (and some UC Santa Cruz alumni may feel that I impugned the good name of their mascot, which was not my intent) but the numbers don’t lie. The measurable difference in local processing performance between SQLite and a modern edge data management system like Actian Zen is two to three orders of magnitude. So, the cheetah-to-banana slug comparison is quantitatively accurate.

In fact, strap in—because I’m doubling down on the banana slug analogy—but I’m going to swap out the cheetah (with its top speed of 70 mph) for a peregrine falcon, with a top speed of 240 mph. The reason? In considering modern edge data management for IoT, or any edge-to-cloud environment for that matter, you have to consider performance in terms of distributed data reads and writes—across devices, gateways, on-premises workstations, and servers, and, of course, the cloud. Such distribution poses complications that SQLite simply cannot overcome with anything remotely resembling speed.

Let me give you an example: SQLite only works as a serverless database, which mandates integration with, and therefore transformation of its data into, a client-server database. You’ll often see SQLite paired with Microsoft SQL Server, MySQL, Oracle, or Postgres. Additionally, there are stealth pairings in which SQLite is present but seemingly invisible. You don’t see SQLite paired with MongoDB or Couchbase, for example, but the mobile client version of both these databases is really SQLite. “Sync servers” between the mobile client and the database servers perform the required extract, transform, and load (ETL) functions to move data from SQLite into the main MongoDB or Couchbase databases.

But wait, you say: Isn’t the whole point about modern edge data management that data at the edge is going to be captured, shared, and processed by smarter components at the edge? Moreover, that some of that data is going to be sent from devices at the edge up to servers in the cloud? Why are you picking on SQLite?

So, in order, the response to your objections is yes, yes, and I’ll tell you.

Modern Edge Data Management is Shared and Distributed

Yes, we should all take it as a given that IoT and mobile environments will send data from devices and IoT sensors at the edge up to servers in the cloud. And, yes, new network standards like 5G (and industrial variants of 4G-LTE), coupled with AI running on more edge devices, will lead to more local and peer-to-peer processing. That will bring device and metadata management out of the cloud/data center to edge gateways on on-premises servers. Both scenarios share and distribute massive amounts of data, and, where SQLite is involved, will entail an explosion of ETL because SQLite isn’t going to be running on the larger servers at the edge or in the cloud. That’s where you’re seeing SQL Server, MySQL, Oracle, Postgres, and others (including Actian Zen Edge, enterprise, and cloud editions).

Which brings us to the question of why ETL matters. When you think about the characteristics of the systems that will be sharing and distributing all this data, three key things stand out: performance, integration, and security. We’ve already discussed the actual processing performance characteristics of our banana slug when it comes to local data operations. When we look closely at SQLite in the broader context of data sharing and distribution, it becomes apparent that the use of SQLite can have a profound impact on operational performance and security.

It’s All About the “T” in ETL

From a data management system standpoint, the transform action in ETL is the most critical element of that initialism. Unlike the E and L which aren’t impacted by data management systems as data transfer is a function of the virtual machine, operating system, hardware abstraction layers and of course I/O subsystems, the data management implementations dictate if, when, and how data transformations will occur.. When moving data from one database or file management system format to another, it is always necessary to reformat the data so that the system receiving the data can read it. SQLite touts the consistency of its underlying file system on all platforms, which would suggest that moving data from one platform to another requires no transformation. For an SQLite application operating as a simple data cache in a mobile device or moderately trafficked web sites this may be true. But that’s not what a shared and distributed IoT environment looks like. Today’s modern edge data management environments are fully managed, secure, and built to perform complex data processing and analysis on a variety of systems in a variety of places—on device, at the edge, and in the cloud. These are environments replete with data aggregation, peer-to-peer sharing, and other data management operations that require a transformation from a SQLite format into something else—quite possibly several something else’s.

And You Thought the Banana Slug Was Slow

That’s the second dimension where SQLite simply becomes sluggish. Actian conducted a series of tests comparing the transformative performance of Zen Core and SQLite. One set of tests compared performance of data transfers between SQLite and MS SQL Server to the same data transfer between Zen Core and Zen Enterprise Server. Both the SQLite and Zen Core serverless clients ran on a Raspberry Pi device while SQL Server and Zen Enterprise ran on a Windows server-based system.

The performance results are eye-popping: Taking a block of 25K rows from Zen Core and inserting it into Zen Enterprise took an average of 3 ms. Taking the same block from SQLite and inserting it in Microsoft SQL Server took 73 ms, or roughly 24X more time. Other tests, comparing Indexed and non-indexed updates, reads, and deletes all had similar results. Why? Because of the transformations required. In moving data between SQLite and SQL Server, the data from SQLite had to be transformed into a format that SQL Server, which has a different format and different data model, can read. When moving the data from Zen Core to Zen Enterprise Server, which rely on the same format and data model, no such transformation is necessary.

So Much for Faster, Better, Cheaper

Zen isn’t the only database with a common architecture stretching from client to server. Microsoft SQL Server has such an architecture, but it only runs on Windows-based devices. Actian Zen runs on pretty much everything—from Android-based IoT and mobile devices to Windows-based edge devices, to data center and cloud servers running a wide range of Linux implementations. Zen has a single, secure, modular architecture that allows the serverless version to interact with the Edge, Enterprise and Cloud versions using the same APIs, data formats, and file system, removing any need for transformations.

And that’s really where the distinction between the peregrine falcon and the banana slug becomes palpably real. If SQLite were capable of interacting directly with other elements in the modern edge data management environment, everyone happily using SQLite could avoid data transformations and heavy ETL. But that’s not the world in which we operate. SQLite will always involve heavy ETL, and a banana slug it will remain.

There’s an age-old tradeoff in the world of engineering development that goes like this: We can give you faster, better, or cheaper. Pick two. SQLite promises faster but in practice delivers slower, as the benchmarks above prove. That leaves better and cheaper—except that, as we’ll see, with SQLite we don’t even get better or cheaper. Stay tuned for the next post in this series, where we’ll discuss why SQLite is not better. After that, we’ll take a sharp, falcon-like look at total cost of ownership.

You can learn more about Actian Zen. Or, you can just kick the tires for free with Zen Core which is royalty-free for development and distribution.


In the last blog on Real-Time Decision-Making (RTDM), we discussed exactly what it is and why it matters. Before proceeding with more theoretical discussions, though, it’s probably worthwhile to give you some practical examples of RTDM and how data management, integration, and analytics support, build RTDM strategic capabilities that significantly impact business operations.

Retail: Ground Zero for COVID-19 Business Disruption

Regardless of your role or the industry you work in, we all have at least a rudimentary understanding of the retail industry and are painfully aware of the business disruptions it has been subjected to over the recent months and the market uncertainties it faces in the upcoming years.

While the sub-verticals of this industry may be seeing vastly different impacts, particularly if you’re designated as essential services as are supermarkets or if you’re a major department store chain, typically anchoring a shopping mall. Further differences in fortune are a function of how much your business is online versus brick-and-mortar.

We’ve all seen the rapid change in fortunes for retail: a shift to online, store closures, and social distancing within, and curbside pick-up outside of those allowed to remain open. Online or offline, we’ve all been subjected to acute shortages due to hoarding (Hey Charmin is back at Costco if you get there Friday mornings at 9 AM), and sadly, price gouging and fake merchandise – particularly for PPE.

Customers are Greek Gods

I had the good fortune of spending four years in the mid-90s working in Japan, and the Japanese had a saying: “Kokyaku-sama wa kami desu,” which translates to the Customer is God. Well, in times like these, customers are Greek Gods. You know, the gods that are inscrutable and fickle. Yet, understanding their behavior is critical to know what to sell them, where, and how. Further, things keep rapidly changing. Two months ago, it was which stores to close or which products to ship where. Now, it’s projecting which stores to open and which products to sell over the next quarter. The point is you will need to collect customer behavioral data and reassess your responses continually.

The common thread across all retail sub-verticals is the need to sharpen customer focus, with the key driver being RTDM capabilities. The three key questions retailers must ask themselves in their quest to sharpen customer focus:

  • Which data do I need to generate the right recommendations to support RTDM?
  • Who needs to leverage RTDM, is it actionable within my existing business process?
  • Does RTDM help to make my business more agile such that it reduces cost, mitigates risk, or delivers guidance on customer behaviors that positively impact sales and service?

Making Customer Loyalty Programs a RTDM Strategic Capability

Kiabi, a global retailer with 500 physical outlets in more than 15 countries as well as an online store, needed RTDM capabilities into its customer loyalty program. The data for this program resided in an aging legacy data warehouse that was slow, inflexible, and expensive to scale. They required real-time performance on an inexpensive, scalable platform that would provide them with current and accurate data with which they could agilely adjust their marketing programs, leveraging historical and new data to price and promote what’s in fashion dynamically. They chose Actian to build out their new cloud data warehouse as an offload and bypass from their day-to-day operational applications and enterprise data warehouse. Actian was able to offer them a 200X increase in performance that met the real-time requirements of their business, and integration with both existing and new data sources.

Once a capability like Kiabi’s is built, it can be leveraged to react to circumstances like COVID-19. The only difference here is what additional datasets are needed to adjust supply, price, and promotion during periods of business disruption or market uncertainties.

Let’s say for the sake of argument that instead of Kiabi, we’re talking about a retail clothing chain in the US, and they have 500 stores nationally as well as an online store. Their stores in malls will remain closed for some time in some states, opened in others. They may have stores that are in densely populated urban areas, some in suburbia, and perhaps a couple in more rural areas. For business-as-usual, their customer loyalty program would have transactional customer data through which they can most likely infer gender, age, and other demographic traits. Further, if they leverage clickstream data combined with the customer loyalty program, they should be able to determine a customer’s favorite store if there’s more than one in their area, and information on where they live from online transactions and shipping to their address.

All of these data points are necessary but insufficient to make quick and accurate decisions on which stores to open and what to expect if you do. What you also need is additional external data. For example, if you could get anonymized cell phone data from carriers, you could determine how far people are willing to travel and overlay this against your customer’s residential address and distances to your stores. This would tell you when travel range is returning to a point where you get a threshold of foot traffic from likely buyers.  To further the accuracy of this, you could also look at 3rd party data on the use of public transportation versus private cars (or perhaps which stores have larger parking lots). In suburban areas, shoppers with private transport will be able to maintain social distancing easier than in urban areas with public transit.  This will generate more foot traffic and make curbside pick-up an easier option without relying on your customers to socially distance.  I could go on adding additional data sources that tell you additional factors that make your decision more accurate, but hopefully, this has gotten the point across.

What we’re really talking about here is adding the missing puzzle pieces – or data – necessary to build out a better common operational picture.  The additional data would be used very differently by HQ versus each individual store manager in how they make their decisions.  But, in both cases, decisions would collapse from weeks to days and days to hours as the situation on the ground changes.

Obviously, Retail isn’t the only industry that’s customer-facing and having to change how they amplify their customer focus or accelerate responses to customer behavior. For more on RTDM or how a Real-Time Connected Data Warehouse can provide you with enhanced dynamic pricing, market basket analysis for best offer promotions, or supply chain management, stay tuned.


Healthcare companies are experiencing some trying times, made worse by the COVID-19 crisis.   Competitors are now becoming partners, research institutions, and pharma companies are taking a more active role in monitoring patient care, and governments are closely monitoring statistics on patient diagnosis.  Everyone in the extended healthcare industry is actively seeking to understand, plan, and respond to the evolving COVID-19 pandemic. For them to do that successfully, everyone is in dire need of real-time data.  The challenge for hospitals and care providers (such as nursing homes) is quickly onboarding new customers, new partners, and new carriers to give them the information they need without compromising data security, HIPAA compliance, or disrupting the company’s operations.

Integration Requests Escalate to the Top of the IT Priority Queue

In normal times, deploying integrations with third-party systems were activities that IT departments approached slowly and methodically – often developing custom interfaces that required long testing cycles.  The requests were essential for the organization, but they weren’t urgent and required immediate attention.  COVID-19 has changed that situation, escalating integration requests to the top of the IT priority queue.  The Centers for Medicare and Medicaid Services (CMS) issued an Interoperability and Patient Access final rule that mandates hospitals must be able to send notifications about admissions, transfers, and discharges to a wide range of recipients by September. This data must now include the patient’s primary care physician, service providers, suppliers, and other healthcare practitioners responsible for patient care.

For small and mid-size hospitals, this rule significantly expanded the breadth of data integrations that are required.  IT can no longer take an ad-hoc approach to integration; they need centralized tools, repeatable processes and templates, and standardized data translations to onboard new data-sharing partners quickly, effectively, and safely. Actian’s Integration Platform for healthcare delivers all this capability and more.

Develop Once, Deploy Anywhere

The key to rapid integration is re-use.  The Actian integration platform and robust design environment help healthcare companies accelerate time to implementation with a “design once, deploy anywhere” approach. This approach greatly improves IT productivity and efficiencies that translate into more integrations, better quality, and reduced implementation times.  Connectivity schemes for industry standards like HIPAA and HL7 provide a common taxonomy and interface format to span systems and companies.  With Actian, you have access to all your data, including legacy data, unstructured data, flat files, and messages in real-time.  This is important because data integration is a combination of content and connectivity – the Actian Integration platform addresses both.

Actian Provides the Integration Platform that Enables Rapid Response

The current healthcare environment is requiring IT departments to be more agile than ever – identifying emerging needs/opportunities and responding quickly to deploy new system features and data integrations.  Actian Integration Platform for Healthcare is a robust set of tools to help IT departments provide the responsiveness their organizations demand.

  • Integrate new applications, hosted systems, and legacy data sources with ease.
  • Rapidly adapt to changes in industry standards and regulations to ensure compliance.
  • A Secure and scalable platform for exchanging data with customers, partners, and suppliers.
  • Centralized management of data connections across the organization from a single pane of glass.
  • HIPAA/HL7 compliant message broker.
  • Support for federated deployments, such as health information exchanges (HIE).

Actian’s Healthcare Integration Platform provides an infrastructure for integrating disparate data in a cost-effective, standards-based, and real-time manner. Actian helps policy setters manage the transformation of incoming and outgoing data from hundreds, or even thousands, of trading partners by speeding up onboarding time and reducing effort.  COVID-19 is causing turmoil in many industries, including healthcare.  IT staff need all the help they can to support their companies and communities through this trying time.  With Actian’s Healthcare Integration platform, data integration can be made easy.

Visit DataConnect to learn more about integration solutions for healthcare as well as other industries.


Blog | Data Integration | | 5 min read

Why Data Interoperability is Vital in Healthcare

data interoperability

The COVID-19 pandemic has exposed fatal cracks in the way healthcare organizations share patient data. Actian DataConnect can help – supporting various healthcare industry standards and enabling rapid data interoperability interchange and transformation of electronic health records (EHR) between healthcare entities.  All while supporting data quality and adhering to healthcare privacy and security regulations.

Industry Ambition Becomes a Regulatory Requirement

The controlled exchange of electronic health records is something that the healthcare industry has been working on for the past two decades with companies replacing paper charts with electronic records systems. Those systems connect to other entities (insurance, pharmacies, labs, hospitals, etc.) to enable patient data to be shared with other providers. That is the idea, at least. Unfortunately, many organizations have fallen far short of the goal, leaving healthcare providers with incomplete and outdated patient data to use in healthcare decisions. As hospitals, labs, and individual providers face an overwhelming spike in COVID-19 patients, their IT departments are struggling to resolve gaps in information-sharing capabilities.

In the US, the Centers for Medicare and Medicaid Services issued an Interoperability and Patient Access final rule (CMS) that mandates that by September, hospitals must be able to send notifications about admissions, transfers, and discharges to a wide range of recipients, including the patient’s primary care physician, service providers, suppliers, and other healthcare practitioners responsible for patient care. COVID-19 has taken an industry ambition and converted it into a regulatory requirement.

The Challenge for IT in Meeting the CMS Deadline

For most hospitals, achieving data interoperability compliance will require significant effort from internal IT staff and Healthcare IT vendors. Major EHR software vendors have already developed integration interfaces for their systems; however, the responsibility falls on IT staff to configure, test, and deploy these updates in the hospital environment. Once software updates have been applied, connections must be set up to the various recipient systems that need to receive the mandated notifications. This is where the real challenge lies. IT departments that have a strong competency in system implementation, operations, and support are having to quickly shift focus to system integration projects – something that many IT professionals have not had to do for decades. The tools and best practices have changed a lot since then.

Compliance will be particularly difficult for smaller hospitals with limited IT staff who have already been pulled to enable the remote working of essential employees, provide data to support crisis operations, and who may be facing the impacts of the COVID-19 crisis in their personal life as well. Although these professionals work in IT, they are intensely aware of the importance of their role in helping their organizations meet the expected needs of their communities and their patients. They understand the urgency and importance of data interoperability goes beyond the government mandate, and patient lives are at stake. Solving this problem is critical!

Actian DataConnect Delivers a Scalable and Secure Platform for Integration

Managing a network of data integrations that span organizations and medical facilities can be difficult. Electronic Health Records systems address the challenge for hospitals and doctors that are part of the same company, but the healthcare industry does not run as independent companies. It runs as an industry-wide network of companies. Cross-company data integration is where the challenge is. Actian is your proven partner for data integration/translation, customer data on-ramps, claims processing, and data auditing. Actian helps policy setters manage the transformation of incoming and outgoing data from hundreds, or even thousands, of trading partners by speeding up onboarding time and reducing effort.

Actian DataConnect supplies an integration platform that can manage both the inbound and outbound interfaces of the healthcare organization. DataConnect is a highly flexible platform service that enables companies to connect any data source, use policies to control access and security, and supply service assurance through centralized management.

Actian’s Healthcare Integration platform provides the infrastructure for integrating disparate data in a cost-effective, standards-based, and real-time manner. No other integration provider offers the combination of value, agility, and power that you’ll find in Actian Integration.

  • HIPAA/HL7 message broker, message queuing, and resubmittal.
  • Clinical web services, systems adapters, data mapping, transformation and translation, and partner profile management.
  • Web-based deployment manager and self-service administration in federated deployments, such as a health information exchange (HIE), Actian’s products are “aware” of other remote deployments, creating seamless communication and data transfer. Actian’s standards-based approach provides flexibility and stability, and also facilitates rapid deployment by simplifying the configuration of data communication, messaging and security.

Ease of implementation, rapid deployment of integrations, and the ability for a limited IT staff to centrally manage and monitor all the company’s integrations in one place make DataConnect an ideal choice for addressing the EHR integration and patient access challenge.  To learn more about how DataConnect supports the healthcare industry, visit www.actian.com/data-integration/dataconnect/