Data Intelligence

The Chief Data Officer’s Evolution to a Data Democracy Sponsor

Actian Corporation

July 22, 2019

cdo evolution

Under the pressure of digital transformation, Chief Data Officers (CDO) have appeared within large companies. According to Gartner, 90% of large companies will have a CDO by the end of 2019.

The thousands of CDOs appointed in the course of the past few years were in charge of improving efficiency and capacity to create value for their organization’s information ecosystem. That is to say, they were invited to direct their organization in processing and exploiting information with the same discipline as the other, more traditional, assets.

Companies that valorize their information assets surpass their rivals in using them to reinvent, digitize, or eliminate existing processes or products.

The CDO’s missions can be summarized as exploiting and finding uses for corporate data, as well as being in charge of developing the use of and trust of employees regarding internal enterprise data. As we have seen, these missions often collide with the powerful cultural restraints within organizations.

How Have the Chief Data Officer’s Missions Evolved?

The CDO has many responsibilities. Gartner identified the main responsibilities of a CDO during their 2019 Data & Analytics event in London. These are, among others:

  • Defining a data and analytics strategy in their organization.
  • Supervising operational initiatives in response to the established upstream strategy.
  • Ensuring information made available on the data is trustworthy and valuable.
  • Constructing data governance.
  • Creating business value in data analytics.
  • Managing efforts regarding data science.
  • Operating and maintaining efforts in infrastructure in response to data analysis needs.

We believe that this impressive list of responsibilities is complemented by another, which could serve as a common thread for all the others and facilitate them: promoting Data Democracy and supporting cultural changes.

At first, CDOs had to lead a mission to convince interest organizations to exploit data. The first few years of this mission were often supported by the construction of a data universe adapted to new uses, often in the form of a Data Lake or Data Mart. The investments agreed upon to construct these data universes were significant but often reserved to specialists. In brief, organizations had more so implemented Data Aristocracies rather than Data Democracies.

The CDO Towards a New Role

 With the exponential development of data, the role of the CDO took a new scope. From now on CDOs must reconsider the organization in a cross-functional and globalizing way. They must become the new leaders in Data Democracy within companies and respond to the call of numerous data citizens who have understood that the way in which data is processed must change radically. The new CDOs must break the bonds of data silos.

In order to obtain the support for data initiatives from all employees, they must not only support them in understanding data (original context, production, etc.) but also help them to invest in the production strategy and the exploitation of data.

From now on, the involvement of stakeholders in the exploitation of data must extend to all levels of the enterprise. It is by facilitating understanding, exchanges, and access around data that organizations will become data-driven.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale, streamlining complex data environments and accelerating the delivery of AI-ready data. The Actian data intelligence approach combines data discovery, metadata management, and federated governance to enable smarter data usage and enhance compliance. With intuitive self-service capabilities, business and technical users can find, understand, and trust data assets across cloud, hybrid, and on-premises environments. Actian delivers flexible data management solutions to 42 million users at Fortune 100 companies and other enterprises worldwide, while maintaining a 95% customer satisfaction score.
Data Architecture

Managing Data Across Distributed Data Warehouses

Actian Corporation

July 17, 2019

distributed data warehouse

Data integration, like the digital transformation initiatives it supports, is a journey and not a destination. Every company is somewhere on a path from the past to a desired state of future integration they’d like to achieve. If your company has existed for a number of years, then you likely have multiple databases, data marts and data warehouses, developed for independent business functions, that now must be integrated to provide the holistic perspective that digitally transformed business processes require.

You may have the ambition to merge all of your data into a single data warehouse (a potentially multi-year effort); you might also decide to keep your legacy systems; or you could choose to restructure your data in a new way and distribute it across regional data warehouses. There is a commonality to any of these scenarios – you will be managing distributed queries for quite a while (if not indefinitely).

IT leaders are increasingly accepting the necessity of distributed queries, but, simultaneously, are becoming more concerned about performance implications for both operational systems and analytics that must leverage data from distributed warehouses. Cloud-based integration platforms and hybrid data warehouses are providing an answer to some of these challenges.

Why are Distributed Queries Problematic?

Distributed queries that span multiple data warehouses are a performance challenge because of the latency caused by remote joins, iterative operations and large data sets that (in addition to being processed within a database) must also traverse network infrastructure. As analytics become more complex and the underlying data sets increase (both common situations), the query requires more time to process.

This isn’t a big deal for batch-processing operations, but the use cases for integrated data are increasingly focused on providing real-time insights. This may be part of a transactional process an agent is performing (such as checking order history, stocking or compiling a 360-degree view of the customer, etc.) or it may be real-time monitoring and analytics to drive operational decision-making/process-tuning. Either way, end users have a low tolerance for data-processing delays and expect the data to be correct, robust and current (real-time).

How to Address the Distributed Queries Challenge?

IT teams have a few available options to address the performance issues of distributed queries.

  1. Merge distributed data warehouses into a single instance – While this seems like the most efficient solution, it often is not a cost-effective alternative due to legacy infrastructure investments, migration costs and business disruption.
  2. Separate the queries and perform aggregation processing in the application layer – Many small companies have used this approach, relying on either Web services or client applications to combine data from distributed sources. While it avoids the distributed query problems, application infrastructure typically has less processing capacity and speed than database infrastructure and as a result, desired performance gains are rarely achieved.
  3. Add an operational data warehouse aggregation layer to the solution architecture – This approach is proving to be the most effective method to achieve high-performance data processing at enterprise scale. Leave the data in the source systems and replicate the info you need for queries into a specialized data warehouse layer that is purposely built for aggregation and servicing real-time query requests. These systems often can be implemented with minimal modification to application code and can be scaled using cloud services to support even large enterprise data sets.

Distributed data warehouses are likely to be a part of the IT ecosystem of companies for many years. Making informed decisions about how to manage data across these warehouses and support real-time, distributed-query operations is essential to helping your company move from basic digital transformation towards real-time, data-driven decision-making and enterprise business agility.

Actian cloud-based data management platforms can help. In addition to providing a hybrid cloud-based integration platform based on Actian DataConnect, Actian also enables organizations to deploy data warehouses across cloud platforms and on-premises.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale, streamlining complex data environments and accelerating the delivery of AI-ready data. The Actian data intelligence approach combines data discovery, metadata management, and federated governance to enable smarter data usage and enhance compliance. With intuitive self-service capabilities, business and technical users can find, understand, and trust data assets across cloud, hybrid, and on-premises environments. Actian delivers flexible data management solutions to 42 million users at Fortune 100 companies and other enterprises worldwide, while maintaining a 95% customer satisfaction score.
Data Intelligence

Understanding the Different Data Cultures

Actian Corporation

July 16, 2019

data cultures

Just like corporate or organizational culture, each enterprise that deals with data has its own data culture. We believe that what distinguishes Web Giants isn’t the structure of their governance, but the culture that irrigates and animates this organization.

We believe in putting in place a Data Democracy. It refers to corporate culture, an open model where freedom rhymes with responsibility.

To better understand Data Democracy, it is necessary to compare it to other data cultures. Here are the main data cultures:

Data Anarchy

In this system, operational professionals feel poorly served by their IT departments, and each one develops its clandestine base (shadow IT), which serves their immediate interests while freeing them from all control regulations and conformity to standards. In 2019, this culture brings sizable risks: data leaks, contravention of ethical regulations, service quality degradation, reinforcement of silos, etc.

Data Monarchy

This system translates to a very strong asymmetry in data access depending on the hierarchical position. Data, here, is very strictly controlled; its consolidation level is carefully aligned with the organizational structure, and its distribution is very selective.

This monarchical culture prevailed for a long time in Business Intelligence (BI) projects: data collected in data warehouses was carefully controlled, then consolidated in reports where access was reserved to a few select people who were close to decision-making bodies. This method promotes a “top-down” approach and willingly encourages a defensive strategy, where rules, restrictions, and regulations insulate data. Its main theoretical benefit is the almost infallible control over corporate data, but that translates into very limited access to data, only reserved for certain privileged groups.

Data Aristocracy

A Data Aristocracy is characterized by a more significant degree of freedom than in Data Monarchy, but which is solely reserved to a very select subset of the population, mainly expert profiles such as Data Engineers, Data Analysts, Data Scientists, etc. This aristocratic approach is often the one that brings the most successful data governance projects to the surface.

Such a culture can be favorable to more offensive strategies, as well as to heterogeneous one, combining top-down and bottom-up. However, it deprives the majority of employees access to data and thus, a certain number of possible innovations and valorizations.

Data Democracy

Data Democracy’s main objective is to make a company’s data widely accessible to the greatest number of people, if not to all. In practice, every employee is able to pull data values at any level. This freedom of access offers maximum opportunities to create value for the company; it provides each employee with the ability, at their level, to use all accessible and compatible resources within their needs in order to produce locally, and through a trickle effect, it will benefit the entire company.

This freedom only works if the regulations and the basic tools are implemented, and each employee is responsible for how they use their data. Therefore, the distribution of necessary and sufficient information is required to allow employees to make proper use of it while adhering to regulations.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale, streamlining complex data environments and accelerating the delivery of AI-ready data. The Actian data intelligence approach combines data discovery, metadata management, and federated governance to enable smarter data usage and enhance compliance. With intuitive self-service capabilities, business and technical users can find, understand, and trust data assets across cloud, hybrid, and on-premises environments. Actian delivers flexible data management solutions to 42 million users at Fortune 100 companies and other enterprises worldwide, while maintaining a 95% customer satisfaction score.
Data Integration

Address Integration Challenges With the Proliferation of IoT Devices

Actian Corporation

July 16, 2019

Integration Challenges: IoT devices

During the past few years, IoT devices have evolved from being a consumer novelty to a core part of next-generation IT ecosystems for businesses of all sizes. IoT devices (such as smartwatches, fitbits, home sensors, RFID point-of-sale scanners, heartbeat monitors, etc) have become attractive and useful because of their inexpensive price point, ease of setup/administration and their diverse capabilities. IoT devices are everywhere and used in a variety of industries and use cases.

Now that business users are experiencing the value of IoT devices in their personal lives, they are understandably asking for similar capabilities at work. As companies embark on digital transformation journeys, they expect IoT to be one of the key technologies to serve as a source of new operational insights and performance gains. Small, inexpensive devices connected to a company’s network provide real-time telemetry and monitoring of business processes, operations, delivery logistics, facilities issues and much more.

The characteristics that make IoT so attractive to business users is also leading to a rather large headache for IT staff. In the personal environment, smartphone applications or a commercial digital assistant (such as Siri or Alexa) manage the connections to individual IoT devices, unify the interaction experience and aggregate the data from all of a user’s IoT devices.

In the business context, where there can be thousands (or more) IoT devices deployed and operating across the IT ecosystem, consistently managing connectivity and integration can be challenging. One of the biggest areas of concern for IT staff (and business users) is integrating the data from IoT devices into a company’s data ecosystem and providing real-time insights based on that data.

IoT Data Integration and Digital Transformation

Most companies have embarked on some sort of digital-transformation journey during the past few years – seeking to re-invent business processes to leverage modern technology at a deeper level. IoT has been critical to enabling digital transformation by providing a cost-effective set of solutions for monitoring end-to-end business and operations processes and controlling physical infrastructure.

Applications for IoT have included such diverse scenarios as monitoring manufacturing quality, optimizing power consumption in company facilities and tracking the flow of customers through retail stores. As more companies move towards digital business processes, IoT will be even more crucial to enabling centralized monitoring and control, and with real-time process optimization.

Leveraging IoT effectively within a business environment requires a scalable and effective means of collecting, connecting and delivering data to each of the deployed IoT devices. The issue with managing connectivity with IoT devices is the sheer volume of individual devices that are deployed as well as variety of data that is collected . Where many more sophisticated IT infrastructure devices (such as wireless access points) can be loaded with standardized configurations and administration profiles, IoT devices are often simple and lacking the robust and typical administrative features of IT environments. There is also very little API standardization across IoT vendors and as a result, a centralized service independent of the IoT devices must typically perform connection management and data aggregation.

The Value of an Integration Platform for IoT Solutions

In the consumer space, a smartphone or digital assistant may be all that is needed to manage integration across a handful of IoT devices and a few users. Businesses with hundreds or thousands of IoT devices and thousands of employees and the need to leverage them require an enterprise-scale set of integration capabilities to be successful. That is the reason organizations need a hybrid data-integration platform like Actian DataConnect, that supports both batch as well as real-time data integration capabilities required by IoT use cases. Real time data integration is crucial for IoT applications.

Driven by business use cases, IoT solutions may require data processing to be brought close to the edges — to the “things” where data is generated. So you need a data integration platform that integrates data at rest as well as data in motion. So you need an integration platform that really spans from the core of the network to the edge. They may also sometimes require aggregation of data in IoT platforms and other data stores, where it can be analyzed in bulk. New IoT data must be reconciled with traditional enterprise master data, for example. Interacting with and reusing data that is constantly in motion and available in various locations throughout highly distributed IoT infrastructures, and working with data within IoT operational flows, will push integration work to the edges.

Integration platforms built for business scale have been optimized for the central administration of large numbers of data endpoints. With many IoT devices, saving a couple of minutes of admin time on each device can add-up quickly. Add new devices quickly, monitor device connectivity status, orchestrate the flow of IoT data to data warehouses and analytics tools and manage who is using your IoT data in a central place. With Actian DataConnect, you don’t need to worry about IoT proliferation bloating your administrative costs. Instead, you can focus on deploying the IoT capabilities your business needs and be confident of your IT staff’s ability to support them.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale, streamlining complex data environments and accelerating the delivery of AI-ready data. The Actian data intelligence approach combines data discovery, metadata management, and federated governance to enable smarter data usage and enhance compliance. With intuitive self-service capabilities, business and technical users can find, understand, and trust data assets across cloud, hybrid, and on-premises environments. Actian delivers flexible data management solutions to 42 million users at Fortune 100 companies and other enterprises worldwide, while maintaining a 95% customer satisfaction score.
Data Intelligence

Data Stewardship and Governance: The Data Steward’s Multiple Facets

Actian Corporation

July 11, 2019

data stewardship

Where Stewardship refers to the taking care of and the supervision of a specific property or organization, Data Stewardship refers to data supervision. Initially, the idea was that a domain expert would be in charge of qualifying and documenting data from their professional standpoint. Data Stewards are those who work closest to where the data is collected; they are often those who best understand the different aspects of data and the standards to which they must adhere.

Data Stewardship and Governance: The Responsibilities

In practice, Data Stewardship covers a wide range of responsibilities, depending on the maturity level of the organization. We can organize these responsibilities in four broad categories:

Operational Supervision and Quality

This refers to monitoring and supervising the complete life cycle of a dataset.

More specifically, Data Stewards must define and implement the necessary processes for the acquisition, storage, and distribution of datasets.

They must also ensure that the data produced fulfills the quality criteria that were defined (values, gaps, completeness, freshness, etc.) and that the procedures are put into place to evaluate and correct potential quality problems.

Documentation

A Data Steward is in charge of defining and documenting data and creating a glossary of industry-specific terms.

They must ensure that each element of a dataset possesses a clear definition and a specific use.

The documentation constitutes a collection of technical and functional metadata according to a meta model in common principle.

Conformity and Risk Management

Data protection and the management of regulatory risks or ethics is one of the most challenging aspects of the Data Steward’s role.

The regulatory environment around data is more restrictive and shifting. It’s up to them to ensure that the proliferation of data is framed by a collection of protocols ensuring conformity with the applicable standards – especially regarding privacy protection.

Security and Access Control

Finally, Data Stewards must define the rules governing data access.

These include the different levels of confidentiality and procedures, allowing the authorization of a person or group to access data.

How Does Data Democracy Strengthen Agile Data Governance?

Orchestrated by a Data Management division, implemented by different types of Data Stewards, data governance must be deployed in an organization. To ensure this deployment, several operational models are conceivable in theory – decentralized, federated, centralized, etc. We think what distinguishes organizations is not the structure of their governance but the underlying culture of this organization. This culture has a name: Data Democracy.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale, streamlining complex data environments and accelerating the delivery of AI-ready data. The Actian data intelligence approach combines data discovery, metadata management, and federated governance to enable smarter data usage and enhance compliance. With intuitive self-service capabilities, business and technical users can find, understand, and trust data assets across cloud, hybrid, and on-premises environments. Actian delivers flexible data management solutions to 42 million users at Fortune 100 companies and other enterprises worldwide, while maintaining a 95% customer satisfaction score.
Data Management

Will Data Management in Operational Technology Finally Be Standardized?

Actian Corporation

July 11, 2019

Rows of virtual files in a data catalog, contributing to powerful data management

The proliferation of Operational Technology (OT) within companies is increasing, which has led many to ask whether this will lead to standardization of how operational technology-generated data is managed. The answer is “it’s complicated.”

There is unlikely to be standardization of the data individual operational technology devices generate, but there will be new capabilities for interoperability, data aggregation, and unified analysis.

What is Operational Technology?

Before examining the standardization issue, it is important to understand the definition of “operational technology.” OT is an umbrella term that describes technology components used to support a company’s operations – typically referring to traditional operations activities, such as manufacturing, supply chain, distribution, field service, etc. (Some companies are relying on operational technology to support, for example, marketing, sales and digital delivery of services, but that is the topic of a future article.)

Operational technology includes, for example, embedded sensors within manufacturing equipment, telemetry from operations components deployed in the field (e.g., oil pipelines, traffic signals, windmills, etc.); industrial IoT devices; location-enabled tablets, which field service personnel use; and much more – the list is long. This is important because OT is not a single classification of technology, it is a descriptor of how technology components are used.

The Push for Interoperability

Some efforts are occurring within the industry to drive interoperability amongst IT and OT components. Open Platform 3.0 (OP3) from The Open Group is a good example. What this standard and others like it seek to do is enable components from different manufacturers to co-exist and work better together within a company’s technology ecosystem. They aren’t seeking to standardize the data coming from individual OT systems or how that data is managed. That challenge is being left to individual companies and the data sciences profession to address.

Data science professionals have been working with companies and individual technology providers for many years to determine a scalable and efficient method to aggregate data from diverse data sources. Efforts to standardize data models and interfaces have been largely unsuccessful due to the desire of some large players in the market to develop and defend closed technology ecosystems.

In light of this, most of the recent developments have been centered on the use of data warehouses to aggregate diverse data and then applying machine learning and artificial intelligence to reconcile differences.

Why Operational Technology Data Management May Never be Standardized

The biggest challenge to standardizing OT data management is managing change. It would be entirely possible to design and deploy a standardized solution to manage all the data generated from OT systems today. The problem is that the technology in this space is continuously evolving and the data being generated is changing too.

Neither technology suppliers nor the companies consuming OT have any desire to slow the pace of technological innovation or constrain it through standardization. New OT innovations will be the driving force behind the next generation of business modernization and companies are eager to consume new capabilities as soon as they can be made available.

How Companies Are Integrating Operational Technology Data

Even though companies don’t have a desire to standardize the data coming from various OT source systems, they have a very critical business need to combine data and analyze it as part of an integrated data set. That is where data management tools, such as Actian, come into play.

Actian’s suite of products, including DataConnect, Actian Data Platform and Zen, provide companies with a platform to manage the ingestion of data from all of their OT data sources, reconcile it in real-time using cloud-scale analytics and machine learning, and then apply the robust statistical analysis (e.g., time series and correlation analysis) to translate data into meaningful insights in an operations context.

The operational technology space is poised to be one of the most important sectors of the IT industry during the next few years. New components will enable companies to generate data from almost all facets of their operations and robust data management solutions, such as Actian, will enable them to interpret this data in real-time to generate valuable operational insights.

While standardization is unlikely, component interoperability is improving and emerging technologies, such as AI, are making data analytics easier. To learn more about how Actian can support your OT efforts, visit www.actian.com/zen.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale, streamlining complex data environments and accelerating the delivery of AI-ready data. The Actian data intelligence approach combines data discovery, metadata management, and federated governance to enable smarter data usage and enhance compliance. With intuitive self-service capabilities, business and technical users can find, understand, and trust data assets across cloud, hybrid, and on-premises environments. Actian delivers flexible data management solutions to 42 million users at Fortune 100 companies and other enterprises worldwide, while maintaining a 95% customer satisfaction score.