Data Architecture

Avoid the Potholes Along Your Migration Journey From Netezza

Actian Corporation

August 20, 2019

Data Warehouse Migration

With certain models of Netezza reaching end-of-life, you may be considering your options for a new, modern data warehouse. But don’t make migration an afterthought. Migrations of terabytes of data, thousands of tables and views, specialized code and data types, and other proprietary elements do not happen overnight.

Given the dependencies and complexities involved with data warehouse migrations, it’s no wonder that, according to Gartner’s Adam Ronthal, over 60% of these projects fail to achieve their intended goals.

Here are some common pitfalls to watch out for when you are ready to start on your migration journey.

1. Pressure to ‘Lift and Shift’

One of the key decisions in designing your journey is whether to “lift and shift” the entire data warehouse or incrementally offload project by project— gradually reducing the burden on your legacy system. IBM is promoting “lift and shift” of Netezza to their IBM Cloud. But think back to how prior forklift upgrades have gone—whether involving mainframes, data center consolidation, service-oriented architecture, or any other legacy technical debt—and you can quickly see the high risk of a “lift and shift” migration strategy.

To mitigate risk and realize faster time-to-value, you want to choose an incremental approach that would enable you to gradually transition from Netezza while first offloading your most simple workloads.

2. “Free” Migration Support From IBM

IBM is offering free migration assistance for Netezza Twinfin and Striper customers to move to the IBM Integrated Analytics System (IIAS) based on Db2 technology. But what in life is ever truly free? Moving to IIAS guarantees that you are locked in yet again to the expensive IBM ecosystem, requiring specialized skills and knowledge to operate and manage.

And you will certainly be disappointed by the difference in functionality offered by Db2.

3. Mishandling Proprietary Elements

There are broadly two classes of workloads running on Netezza systems: Business Intelligence-led systems and Line of Business (LoB) applications. BI applications are relatively straightforward to migrate from one platform to another. LoB applications, however, often contain significant amount of complex logic and represent the most challenging workloads to migrate to a new platform. We frequently find that customers have written some of the logic of these systems using stored procedures or user-defined functions, which are the least portable way of building an application.

If your application estate consists of such complex code, your target data warehouse should adhere to SQL, Spark, JDBC/ODBC, and other open standards and have a wealth of partners capable of automatically identifying and converting proprietary Netezza elements to standards-based views, code, data, etc.

4. Rushing the Business Assessment

A thorough assessment of your legacy environment is a critically important step in the migration journey. Your company has most likely invested decades of logic into your Netezza platform. There will be a lot of junk data that got created over that time. Tables that may not have been touched for years.  Countless queries and workloads that are irrelevant to the business. These objects should not be moved during the migration.

You can reduce your migration risk by using an automated tool to analyze the logs from your Netezza data warehouse to gain a complete understanding of your current environment. Based on numerous factors, the tool can identify redundancies that should not be migrated, decide what should, prioritization for migration, and how to work with phased migrations.

5. Being Locked in Without Options

Be sure to choose a solution that provides the flexibility for you to chart any path you desire without compromise. For instance, do you want to move all at once to the cloud as part of a cloud-first strategy, or conduct your data warehouse migration in stages? Businesses with rigorous compliance or privacy demands often prefer to store some data on-premises. Do you want to be locked into a particular cloud platform, or go with AWS now but have the option to move some apps to Azure later? Whatever your situation is, don’t trade in your current vendor lock-in for another.

rethinking webinar

Learn more migration best practices by watching our on-demand webinar, “Top 7 tips for Netezza migration success,” featuring former Director of Tech Services at Netezza.

Migrate With Confidence

Actian Data Platform can help you incrementally migrate or offload from your Netezza data warehouse until it can be retired in a managed fashion—according to your timeframe and your terms. Choose the path that is best for you – cloud, on-premises, or a combination of both, with a seamlessly architected hybrid solution. Prior studies from our customers and prospects have shown typical gap analysis to resolve to over 95% on average for migrations.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale. Actian data intelligence solutions help streamline complex data environments and accelerate the delivery of AI-ready data. Designed to be flexible, Actian solutions integrate seamlessly and perform reliably across on-premises, cloud, and hybrid environments. Learn more about Actian, the data division of HCLSoftware, at actian.com.
Data Management

Does it Make Sense to Push All Your IoT Data to the Cloud?

Desmond Tan

August 19, 2019

Cloud computing networking

IoT devices create plenty of data – much more that you might think. When you multiply this amount of data by the number of devices installed in your company’s IT ecosystem, it is apparent IoT is a truly big data challenge.

Two common questions IT staff and data management professionals have been asking during the past few years are: “Does it make sense to push all your IoT data to the cloud?” and “What are the pros and cons of doing so?”

Why Would You Move IoT Data to the Cloud?

There are three primary reasons why companies move their IoT data to the cloud for processing.

  1. Aggregate data from multiple devices in one place.
  2. Leverage cloud-scale compute to process the data.
  3. Enable business analytics and decision-making.

As you might imagine, these reasons are not entirely independent. Part of what makes IoT such an attractive technological component is its simplicity. IoT devices aren’t highly sophisticated, don’t contain much internal storage and typically aren’t capable of complex data processing.

Consequently, they can be less expensive to acquire and deploy and operate with low power consumption. IoT devices work best when coupled with cloud services of some sort where data from the devices can be collected, transferred somewhere for processing and then response signals are sent back to the device. As their name suggests, IoT devices are designed to be networked and interact with other technological components (they aren’t as powerful on their own as they are as part of a system).

Benefits of Moving Your IoT Data to the Cloud

Yes, it’s great that you can move IoT data to the cloud for processing, but it is never a good strategy to do something for the sake of doing it – there should be a purpose. In the case of IoT data, that purpose is to enable decision-making – either strategic or operational.

IoT devices provide a bridge between the digital world and the physical world by capturing information about the environment, so remote consumers can make decisions and initiate actions. Those actions may be as simple as turning a light on or off or part of a complex operational process (such as a manufacturing system).

IoT devices enable users to interact with physical environments remotely. To do this, the data and capabilities of the IoT device must be made available to the remote user. The best method for this in an enterprise environment is to move the data to the cloud, so whoever needs it may access it. Often, IoT devices work as components of a greater system or workflow. Once data is in the cloud, it can be merged with other data, analyzed for meaning and relevance and, in some cases, used to drive automation. There are many business problems IoT can help companies solve.

Drawbacks to Moving Your IoT Data to the Cloud

Using IoT data to solve business problems can create tremendous value for a company in pursuit of digital transformation or enterprise business agility goals. Unfortunately, not all the data IoT devices produce is useful and valuable – some of it is just noise. The two biggest issues with managing IoT data are:

  1. The large volume of data produced.
  2. Sorting the meaningful information from the data clutter.

These two issues are of importance when evaluating what data to move to the cloud, as you want to avoid adding clutter to a data warehouse or clogging your infrastructure with the transmission and processing of data that you don’t intend to use. A good rule is to focus on determining what data you actually need and plan to use.

Because IoT data is most valuable in real-time, consider what data is needed to support your operational processes or real-time decision-making. Once you understand this, look at the data produced by your IoT devices and see how well it serves your needs. You may find you don’t have the right kinds of IoT devices in your environment or you only need a subset of the data being generated to make meaningful decisions.

Your company’s decision-making and operational needs will determine what IoT data to move to the cloud, how you aggregate data from disparate systems and who uses it within your organization.

The composition of your IT infrastructure, including, for example, security and cost models, may apply some additional constraints to what should be moved to the cloud vs. filtered at the source. If your company uses analytics tools that perform time-series analysis, then you may decide to move all your IoT data to the cloud for analysis and then filter it later.

IoT is positioned to assume an expanded purpose within the IT ecosystems of companies during the next few years, enabling them to move beyond basic digital transformation towards true business agility.

Actian provides a suite of solutions to help you manage the data on IoT devices to operational data warehouses, capabilities for processing large-scale data in near-real time, and the time-series analytics capabilities to understand the value of your IoT data for your company.

desmond tan headshot

About Desmond Tan

Desmond Tan is Senior Director of Product Engineering at Actian, guiding R&D for the Actian Zen Embedded Database team. With 18+ years in database research, development, and product support, Desmond brings a customer-focused approach to advanced embedded solutions. Desmond has authored technical documentation and knowledge base articles on optimized embedded data storage. He frequently mentors upcoming engineers. His blog content covers lightweight database design, performance tuning, and embedded analytics. Explore his posts to enhance your data-driven applications.
Data Intelligence

Data Revolutions: Towards a Business Vision of Data

Actian Corporation

August 19, 2019

data revolutions

The use of massive data by the internet giants in the 2000s was a wake-up call for enterprises: Big Data is a lever for growth and competitiveness that encourages innovation. Today, enterprises are reorganizing themselves around their data to adopt a “data-driven” approach. It’s a story constituting several twists and turns that tends to finally find a solution.

This article discusses the different enterprise data revolutions undertaken in recent years up to now in an attempt to maximize the business value of data.

Siloed Architectures

In the 80s, Information Systems developed immensely. Business applications were created, advanced programming languages emerged, and relational databases appeared. All these applications stayed on their owners’ platforms, isolated from the rest of the IT ecosystem. 

For these historical and technological reasons, an enterprise’s internal data was distributed in various technologies and heterogeneous formats. In addition to organizational problems, we then speak of a tribal effect. Each IT department has its own tools and implicitly manages its data for its own use. We are witnessing a type of data hoarding within organizations. To back these suggestions, we frequently recall Conway’s law: “All architecture reflects the organization that created it.” Thus, this organization, called silos, makes for very complex and onerous cross-referencing of data originating from two different systems.

The search for a centralized and comprehensive vision of an enterprise’s data will lead Information Systems to a new revolution.

The Concept of a Data Warehouse

By the end of the 90s, Business Intelligence was in full swing. For analytical purposes and intending to respond to all strategic questions, the concept of a data warehouse appeared. 

To make this, we will recover the data from mainframes or relational databases and transfer them to an ETL (Extract Transform Loader). Projected in a so-called pivot format, analysts and decision-makers can access data collected and formatted to answer pre-established questions and specific cases of reflection. From the question, we get a data model.

This revolution always comes with some problem. Using ETL tools has a certain cost, not to mention the hardware that comes with it. The elapsed time between the formalization of the need and the receipt of the report is time-consuming. It’s a revolution that is costly for perfectible efficiency.

The New Revolution of a Data Lake

The arrival of data lakes reverses the previous reasoning.  A data lake enables organizations to centralize all useful data storages, regardless of their source or format, for a very low cost. . We stock an enterprise’s data without presuming their usage in the treatment of a future use case. It is only according to a specific use where we will select these raw data and transform them into strategic information.

We are moving from an “a priori” to an “a posteriori” logic. This revolution of a data lake focuses on new skills and knowledge: data scientists and data engineers are capable of launching the treatment of data, producing results much faster than the time spent using data warehouses.

Another advantage of this Promised Land is its’ price. Often offered in an open-source way, data lakes are cheap, including the hardware that comes with them. We often speak of community hardware.

…or Rather, a Data Swamp

Certain advantages are present with the data lake revolution but come along with new challenges. The expertise needed to instantiate and to maintain these data lakes are rare and thus, are costly for enterprises. Additionally, pouring data in a data lake day after day without efficient management or organization brings on the serious risk of rendering the infrastructure unusable. Data is then inevitably lost in the mass.

This data management is accompanied by new issues related to data regulation (GDPR, Cnil, etc.) and data security: already existing topics in the data warehouse world. Finding the right data for the right use is not yet an easy thing to do.

The Settlement: Constructing Data Governance

The internet giants understood that centralizing these data is the first step, however insufficient. The last brick necessary to go towards a “data-driven” approach is to construct data governance. Innovating through data requires greater knowledge of these data. Where are my data stored? Who uses them? With which goal in mind? How are they being used?

To help data professionals chart and visualize the data life cycle, new tools have appeared: we call them, “Data Catalogs.” Located above data infrastructures, they allow you to create a searchable metadata directory. They make it possible to acquire a business vision and data techniques by centralizing all collected information. In the same way that Google doesn’t store web pages but rather, their metadata to reference them, companies must also store their data’s metadata in order to facilitate the exploitation of and discovery of them. Gartner confirms this in their study, “Data Catalog is the New Black”: if your data lake’s data is without metadata management and governance, it will be considered inefficient.

Thanks to these new tools, data becomes an asset for all employees. The easy-to-use interface doesn’t require technical skills, becoming a simple way to know, organize, and manage these data. The data catalog becomes the reference collaborative tool in the enterprise.

Acquiring an all-round view of these data and to start data governance to drive ideations thus becomes possible.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale. Actian data intelligence solutions help streamline complex data environments and accelerate the delivery of AI-ready data. Designed to be flexible, Actian solutions integrate seamlessly and perform reliably across on-premises, cloud, and hybrid environments. Learn more about Actian, the data division of HCLSoftware, at actian.com.
Data Engineering

Cloud Software Development: Planning the Journey

Actian Corporation

August 15, 2019

A river flowing

For software companies developing new solutions in the cloud, the overall mission can be a daunting proposition. Cloud Software Development is an ever-evolving landscape, where one piece of technology may be popular one day, only to be replaced for something better a little while later. Then, there is a slew of tech jargon along with an endless list of acronyms which one must learn and constantly keep up in order to remain technologically adept and to have a chance, at a minimum, at contributing to a discussion with like-minded people.

When software architects and leaders meet to discuss and devise a plan to build something new, it is imperative to recognize the trends which have been adopted by others. This is necessary in order to capitalize from a common pool of knowledge, where the next big project can be the result of a collective wealth of skill and experience. In general, history has shown to us that each technological advancement that we achieve is, in most cases, simply a better version of an existing product.

So, if we are to develop the next great cloud-based software solution, a few questions come to mind:

  • Which cloud providers are going to be used?
  • How will the infrastructure be provisioned?
  • How will the software be built and deployed (CI/CD)?
  • What is our containerization strategy?
  • Should we leverage “Serverless” computing?
  • What is our monitoring strategy?
  • How can we ensure compliance with security and data privacy regulations, e.g. GDPR?
  • Should we choose open-source or proprietary tools, and which of these will provide maximum ROI?
  • How can we best automate the entire Software Development Life Cycle?
  • Do we have a strategy for CI/CD where can implement Agile best practices?
  • How can we allocate our limited resources effectively?
  • Which discipline is best equipped to answer these questions?

The answer lies with a set of software development practices known as “DevOps”. Understanding this discipline will allow us to pick the right tool for the job, incorporate well-known and established processes, and recognize that there are many “flavors” of cloud available where we must be ready to deploy at the behest of our customers.

The following article covers in much greater detail how to best choose the right DevOps tools for the job.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale. Actian data intelligence solutions help streamline complex data environments and accelerate the delivery of AI-ready data. Designed to be flexible, Actian solutions integrate seamlessly and perform reliably across on-premises, cloud, and hybrid environments. Learn more about Actian, the data division of HCLSoftware, at actian.com.
Data Platform

Using a Cloud Data Warehouse to Support Localized Systems

Actian Corporation

August 7, 2019

Cloud Data Warehouse

Over the past few years, we’ve seen an increasing trend of regional governments applying unique restrictions and controls on where data is stored and how it is managed for users and businesses in their jurisdiction. The EU and Japan have recently imposed some strict rules about data export.

A cloud data warehouse can be an effective tool in helping your company remain compliant with regional regulations by keeping data within the region it was created.

Distributed Architecture of the Cloud

Cloud infrastructure is different from on-premises data centers in that it is inherently designed to support distributed systems. This could be either the replication of common capabilities to multiple regions or the localized deployment of specialized capabilities to a specific regional audience. Cloud management platforms include the tools to be able to define regions and effectively direct transactions (and data) to technical resources aligned with your unique business needs.

Separation of Applications and Data Stores

In the past, there was typically a 1:1 relationship between applications that users interact with and the databases where transactional data is stored. If you wanted to keep the data from one region localized in that region, you needed to create a new database and a new version of the application to connect to it. Modern cloud architecture changes that. You can now have a single application that is replicated globally and use routing rules to connect the application to different databases, either based on performance criteria or geopolitical rules.

The Impact of Data Localization Rules on Data Warehouses

One of the unique challenges that companies have faced when seeking conformance to data localization regulations is the need to not only store transactional data within the local region but also to control data being copied and exported outside of the source jurisdiction. This means that if companies want to perform deep analytics and data mining, they may need to do these activities in data warehouses that are in a specific region. When data warehouses were predominantly hosted on-premises, this was a costly proposition – acquiring a data center, standing up a data warehouse, operating it, and performing analytics within a specific region.

Data Warehouses in the Cloud

The cloud makes deploying localized data warehouses easier and cheaper than on-premises alternatives. The same cloud providers that are hosting localized applications can provision a data warehouse in-region with both the compute and storage capacity for performing business analytics and data mining. While the raw data is often export-controlled, there are often fewer restrictions (if any) on exporting the analytics and insights derived from data to decision-makers at a corporate headquarters in a different region. Companies can now distribute their data warehousing capabilities across the globe the same way they are distributing applications and transactional data stores, while at the same time retaining the ability to do high-performance processing and achieve centralized information insights.

Actian is the leader in hybrid data warehousing that span on-premises and multiple cloud platforms. Actian Data Platform provides a full-featured cloud data warehouse solution that can be deployed quickly – either globally or to a targeted region, helping your company achieve superior business value, fast time to market, and sustainable operational costs.  To learn more, visit www.actian.com/data-platform.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale. Actian data intelligence solutions help streamline complex data environments and accelerate the delivery of AI-ready data. Designed to be flexible, Actian solutions integrate seamlessly and perform reliably across on-premises, cloud, and hybrid environments. Learn more about Actian, the data division of HCLSoftware, at actian.com.
Data Integration

Considering Cloud Migration? First, Read This Checklist.

Actian Corporation

August 5, 2019

3d rendering robot learning or machine learning

Most modern companies have fully embraced the cloud as the preferred place to host the applications that their businesses use.  Many new applications are “cloud-native” or are designed specifically for operating in the cloud, while many of the legacy applications that your company uses today may have been designed to run on hardware in your company’s data center, i.e., on-premises.  The good news is that even if you have legacy applications, most software providers are now offering cloud-hosted options for their solutions.

Before you make the move to migrate legacy apps to the cloud, there are a few things you need to consider.  Use the following checklist as a tool to help you understand if cloud migration is right for your app, and if so, how you can increase the success of the migration process.

  1. Does the Cloud App Have All the Functions You Need?
    While most commercially available software packages that were developed for on-premises hosting now have cloud equivalents, the first thing you should look for is functional parity.  To accelerate time to market, some software providers have only implemented a subset of their tool’s functionality in their cloud offerings.  Now is a good time to review not only what features you have in your applications but also how are they being used by your users and business processes.
  1. Can Your Users Access the Cloud Application?
    One of the most significant downsides of cloud apps is that they require connectivity from the end user’s device to the cloud services in order for the app to work.  If your users need to be able to leverage the application when traveling to remote areas where reliable internet connectivity is unavailable, this can be an issue.  Connectivity can also be a challenge in corporate environments with robust network security protocols and firewalls that restrict access to external resources.  Enabling users to access cloud apps may require configuration to your network’s access control configurations.
  1. Where is the App Hosted?
    Many people don’t understand that the cloud is an extensive network of data centers run by large service providers.  In this network, applications may be hosted in a specific data center or region, or they may be replicated to operate in data centers across the globe.  It is important to understand where your users are located and use this information to guide your cloud deployment.  If your users are all in one city or region, it may be sufficient to have your app hosted only in that region.  If you have users in many different regions, more of a global footprint may be needed.  What you want to avoid is having you users in one region and your app hosted in another.  Network latency can cause significant impacts on your end-user’s experience related to app performance.
  1. Where is the Data Stored?
    In on-premises applications, typically the database that the application uses is co-located with the web or app server that brokers user transactions.  With cloud implementations, this is not always the case.  You may have application instances distributed around the globe, but leverage a centralized database to record transactions. Depending on the nature of the application, this configuration could cause performance issues if the network latency between the app and its data store is significant. Some modern architectures afford the ability to store and process data within the application itself (wherever it is deployed).  This is an important issue for your solution architects to investigate.
  1. How Will You Manage Integration?
    Digital business processes often involve the use of multiple applications and many data sources. Integrating your cloud apps with other systems and data stores, both on-premises and in the cloud can be complex and difficult to maintain.  Legacy approaches that leverage point-to-point integrations are often not effective in cloud environments.  Consider the use of a hybrid data integration platform such as Actian DataConnect to help you solve this problem.  By managing your connections in one place, you can enable greater flexibility in your deployment strategies.
  1. Do You Need to Replicate Data to Your Cloud Data Warehouse?
    When looking at your data integration needs for a cloud app, consider what (if any) of the application’s data needs to be replicated into your data warehouse for data mining and detailed analytics. A cloud data warehouse can be a powerful part of your cloud strategy, enabling less data to be retained locally (a key cost driver), improved analytics capabilities, and easier data retention should you decide to replace the cloud app in the future

As you can see from this checklist, there are some basic things that you need to evaluate to determine whether the migration of your application to the cloud will meet your business needs.  There are also some considerations that can help increase business value once your cloud migration is complete.  The choices you make about how your application data is managed are important to both achieving sustainable value for your business as well as enabling agility for the future.

Actian DataConnect is a hybrid data integration platform designed to help companies manage the connections between applications, platforms, and data sources across the enterprise. By managing your connections in one place, you can lower your operational costs, improve security, and enable business agility. To learn more, visit DataConnect.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale. Actian data intelligence solutions help streamline complex data environments and accelerate the delivery of AI-ready data. Designed to be flexible, Actian solutions integrate seamlessly and perform reliably across on-premises, cloud, and hybrid environments. Learn more about Actian, the data division of HCLSoftware, at actian.com.
Actian Life

Austin’s First Salsa and Guacamole Contest

Actian Corporation

August 1, 2019

Actian Culture

I am so thrilled to tell you about all about our first social event which put a little “spice” into everyone’s work week!

We sponsored a best salsa and guacamole contest. We asked everyone to bring in their favorite salsa and guacamole recipe and let their fellow colleagues’ taste and vote for the best one in each category. There were several entries, and everyone got a chance to place their vote anonymously.

We all had multiple “taste testing’s, just to make sure the vote chose was correct. 

We decorated our café area and blew up balloons which came in cactus and avocado shapes; red balloons, mini maracas and a table runner completed the look. 

The winner for best salsa went to Jay Clark and the winner of best guacamole went to Melanie Richards. They each won an Amazon gift card.

Everyone had a good time, sampling and resampling, talking to each other and enjoying the event. Sometimes in life, it’s the little things that bring a smile to someone’s face.

The biggest take away we received from this event; when is our next event? 

Thank you for reading about one of the events that make the Actian Culture phenomenal!

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale. Actian data intelligence solutions help streamline complex data environments and accelerate the delivery of AI-ready data. Designed to be flexible, Actian solutions integrate seamlessly and perform reliably across on-premises, cloud, and hybrid environments. Learn more about Actian, the data division of HCLSoftware, at actian.com.
Data Management

Should You Have Separate Document, Time-Series, NoSQL SQL Databases?

Actian Corporation

July 31, 2019

Server Image for Data Analysis

Managing and analyzing heterogeneous data is a challenge for most companies, which the oncoming wave of edge computing-generated datasets has only exacerbated. This challenge stems from a rather large “data-type mismatch” as well as how and where data has been incorporated into applications and business processes. How did we arrive here?

At one time, data was largely transactional and Online Transactional Processing (OLTP) and Enterprise resource planning (ERP) systems handled it inline, and it was heavily structured. Primarily, Relational DataBase Management Systems (RDBMS) managed the needs of these systems and eventually evolved into data warehouses, storing and administering Online Analytical Processing (OLAP) for historical data analysis from various companies, such as Teradata, IBM, SAP, and Oracle.

As most manual processes utilizing paper moved to digital records management, content management systems emerged as a means to manage all the unstructured documents from knowledge workers or which the expanded functionality within ERP and personal computing systems autogenerated. These systems contain semi-structured and unstructured document data stored in eXtensible Markup Language (XML) and JavaScript Object Notation (JSON) formats.

In parallel, and more so during the last few years with the Internet of Things (IoT) revolution, the third wave of digitization of data is upon us, operating at the edge in sensors, video, and other IoT devices. They are generating the entire range of structured and unstructured data but with two-thirds of it in a time-series format. Neither of these later datasets lends itself to RDBMS systems that underpin data warehouses due to how the data is processed and analyzed, the data formats used and the mushrooming dataset sizes.

Consequently, separate Document Store Databases, such as MongoDB and Couchbase, as well as several time-series databases, including InfluxDB and a multitude of bespoke Historians, emerged to handle these very distinct datasets. Each has a separate Application Programming Interface (API), lumped together as NoSQL – as in everything that’s not Structured Query Language (SQL).

The aftermath of these three waves of data types and database structures is data architects must now implement separate databases for each type of data and use case or try to merge and aggregate all of the different data types into a single database. Until recently, the only significant or enterprise-wide aggregation point for multiple databases and data types was the traditional data warehouse. The legacy data warehouse, however, is lagging as an aggregation point for two reasons.

First, many of them are based on inflexible architectures in terms of their capability to manage JSON and time-series data and the cost to expand them to administer larger datasets or complexity of modern analytics, such as Artificial Intelligence (AI) and Machine Learning (ML). Second, sending all the data to them in a single, centralized location on-premises can be costly and hinders decision-making at the point of action at the edge of the network.

During the era of edge computing and a wholesale flip of the majority of data being created and emanating from the edge instead of from the data center or a virtualized image in the cloud, specialized applications and platforms have an essential purpose in business process enablement. Just as each business process is unique, the data requirements for that technology to support those processes are also unique. While it may seem best-of-breed database technology for document store versus time-series versus traditional, fully structured transactional data may remove constraints on the use of technology within a business, you should be very careful before you go that route.

In general, the more APIs, underlying database architectures, resulting differences in supporting file formats, management, and monitoring systems and changes in which ones you use based on use case simply increase the complexity of your enterprise data architectures. This is particularly the case if you offer or implement multiple products, technologies and integration methodologies with this medley of databases. This complexity tends to have a domino effect into your support lifecycle for any software leveraging these databases – even the procurement of the databases.

Provided you can find a single database with similar performance and addresses all the data types and SQL as well as direct manipulation of the data through a NoSQL API, it makes far more sense to merge and aggregate heterogeneous data into a common database structure, particularly in Edge Computing use cases. For example, if you are looking at video surveillance data, sensor networks, and logs for security, then combinations of these and other disparate data sets must be aggregated for cross-functional analytics.

If you need to analyze, create reports and dashboards based on data of different types and in different source systems, then you will need some sort of capability for normalizing the data, so it can be queried either onsite or remotely from a single data set.

The requirements have changed during the last 30 years and Actian has built a new modular database that is purpose-built for edge-computing technologies and use cases and is capable of handling all datasets through a single NoSQL API, yet provides full SQL compliance. In both SQL and NoSQL functions, our 3rd party benchmark results show far better performance than any of the major Document Store, Time-Series or traditional SQL databases capable of handling Mobile and IoT.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale. Actian data intelligence solutions help streamline complex data environments and accelerate the delivery of AI-ready data. Designed to be flexible, Actian solutions integrate seamlessly and perform reliably across on-premises, cloud, and hybrid environments. Learn more about Actian, the data division of HCLSoftware, at actian.com.
Data Architecture

Rethinking Data Warehouse Modernization

Actian Corporation

July 30, 2019

Data Transformation

Data Warehouse Modernization

Sounds like a no-brainer. Move your Netezza, Teradata or Exadata data warehouse that likely runs on obsolete, proprietary hardware to a shiny new system that runs in the cloud, costs a fraction of what you were paying before and can be turned on and off like a light switch.

Is it That Easy? That Simple? Let’s Discuss…

Data warehouse modernization projects come in many flavors. Some are straightforward database upgrades or tweaks to data models. Others are more ambitious, such as redesigning your company’s primary data platform. Modernization may involve incorporating previously untapped features, such as vectorized processing, real-time columnar in-memory analytics, in-database algorithms or SPARK-based AI/ML analytics. Further, today’s next-gen cloud data warehouses deliver unprecedented economics due to their elastic, pay-as-you-go consumption model. Increasingly, modernization may include all of the above in order to address new data sources such as social, mobile and IoT data.

You can read my full article here.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale. Actian data intelligence solutions help streamline complex data environments and accelerate the delivery of AI-ready data. Designed to be flexible, Actian solutions integrate seamlessly and perform reliably across on-premises, cloud, and hybrid environments. Learn more about Actian, the data division of HCLSoftware, at actian.com.
Data Architecture

Smart IT Managers are Forging Their Data Warehouse Path to the Cloud

Actian Corporation

July 30, 2019

it's the journey not the destination

As Ralph Waldo Emerson once said, It’s not the destination it’s the journey. Unfortunately, this wise saying is not nearly applied as often as it should be to the contemporary topic of data warehouse modernization project design and execution. The result, thousands of data warehouse modernization projects unnecessarily end up in failure. Gartner’s Adam Ronthal, suggests that over 60% of all database migrations fail. That is, a migration is started, then starts slipping, overruns the original budget by multiples, until finally, management pulls the plug. Time is lost, money wasted, and careers wrecked. How do you make sure yours is not added to this dubious and growing list? 

There are Some Key Steps to Designing the Optimal Journey: 

  1. Migrate or Offload: A first step that many organizations take that mitigates risk and accelerates value delivery is the off-load journey. In this situation, several demanding workloads are identified to offload in a phased manner rather than migrate the entire data warehouse. Unless an organization needs to completely get off its existing data warehouse, e.g. your vendor has announced EOL plans as is happening with Netezza today, taking a phased approach with your migration usually makes sense and lets you get started sooner. Great candidates may include workloads with these characteristics: large data sets, ad-hoc query tool diversity or requests for new unsupported tools, hybrid data from multiple diverse data sources and complex queries.
  2. Baby Step or Giant Leap to the Cloud: Some organizations, as part of a cloud-first strategy move all of their data warehouse to the cloud. In other cases, pragmatic organizations often choose to conduct their migrations in stages. For example, moving a data landing and staging area to the cloud provides useful elasticity and agility while reducing the risk of disruption.
  3. Automate, Automate, Automate: Leading solution providers now offer sophisticated migration assessment tools that can identify SQL queries that can be translated to the new target system automatically. Typically, data warehouse systems that fully support the entire current SQL industry standard will do better at supporting the automation of query migration. It is not untypical for industry-standard SQL systems to support 95% and higher levels of automatic conversion from the source to the target system.
  4. Replicate, Augment or Breakthrough: Most organizations take a 3-step approach to the migration: first, replicate the source system report generation, then utilize the upgraded performance of the target system to augment the query base with additional dashboards and other ad-hoc analytics. Finally, in the third step, forward-thinking organizations look to develop new composable analytics applications such as real-time offers and fraud analysis that were prohibited for cost and/or performance reasons from considering before.
  5. Think “Business” with a Big “B”: A successful data migration project is always dependent on understanding and addressing the current and future needs of the business. This requires proactive collaboration with the direct and indirect users of the insight that you hope to deliver with your modernized data warehouse project. Especially important will be to identify and prioritize discovery-oriented analytics including ad-hoc analysis that the business side seeks and values.

How to Get Started? 

Your first step is to seek out and short-list a set of next generation solution vendors that can offer you a truly hybrid data warehouse journey that runs both in the cloud and on-premises with zero changes to your query stream and easily reroutes your ETL connections to your source data warehouse apps and external data sources, supports multiple cloud platforms to eliminate lock-in, and excels at workloads that demand high levels of scale and concurrency.  

 

Solutions such as Actian Data Platform designed to run seamlessly in the cloud and on-premises represent a potential breakthrough worth checking out. As the above diagram conveys an organization can charter its modernization journey through multiple paths with different combinations of cloud and on-premises deployments. A glass of hybrid with your slice of cloud data warehouse computing sir? Drink up – it’s a hybrid world we live in, today and tomorrow. 

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale. Actian data intelligence solutions help streamline complex data environments and accelerate the delivery of AI-ready data. Designed to be flexible, Actian solutions integrate seamlessly and perform reliably across on-premises, cloud, and hybrid environments. Learn more about Actian, the data division of HCLSoftware, at actian.com.
Data Integration

What You Should Know Before Moving to a Multi-Cloud Environment

Actian Corporation

July 30, 2019

Multi-Cloud Environment

Most IT leaders will agree: the cloud is the right place to host and operate many of their applications and systems. What isn’t quite as clear for these leaders is “which cloud” they should use.

The answer is “There is no right answer.” Each company (and application) has its own unique set of technical and operating needs. They will drive the decision about whether it should run in one of the public cloud environments or a company’s private cloud or on a 3rd party’s cloud environment (e.g., hosted and consumed as Software as a Service (SaaS)). This is a decision that should be made on a case-by-case basis and informed by a clearly documented and enforced enterprise-cloud policy.

Most organizations will likely choose and operate a hybrid environment with capabilities stratified across multiple clouds. Sensitive data apps may be on a private cloud, customer-facing systems that require geographic scalability may be on public clouds and SaaS components will likely be on someone else’s cloud. In this configuration, your focus should be on how you manage and connect the data in these systems. You will likely have business processes and data-driven analytics that require data to be integrated from different systems for your company to operate effectively.

Establishing and maintaining data connections in a multi-cloud environment is often more complex than managing connections in a single-cloud environment. Although it may be difficult, there are 3 reasons why data integration across multiple clouds is essential:

  1. Avoiding latency in business processes.
  2. Enabling data aggregation for analytics.
  3. Replicating data to manage business continuity risk.

Cloud environments provide a tremendous amount of cost leverage and scalability potential but require more robust data-connectivity capabilities than your IT department’s likely experience in legacy environments. In multi-cloud setups, it is often necessary to shift your integration patterns from point-to-point integrations between applications to more of a centralized data integration hub architecture, leveraging a platform, such as Actian DataConnect, to broker the connections to all your systems.

In addition to making manageability easier, DataConnect can help you monitor the flow of data between different cloud environments – you’ll make more informed decisions about how to improve end-to-end system performance and how to reduce infrastructure operating costs. Most companies make their initial cloud-selection decisions based on functionality criteria; however, cost savings is often the driver for making the decision to migrate systems from one cloud to another. Since cloud infrastructure is charged on a consumption model, understanding how your cloud applications are being used can help you identify cost-savings opportunities quicker.

Whether you are designing your cloud strategy in preparation to begin migrating capabilities to the cloud or seeking to optimize your current cloud utilization – selecting the right set of tools to help you manage your data connections is an essential step for harvesting maximum value from your cloud investments. Actian is an expert in hybrid data management, with modern solutions to help you manage data throughout the lifecycle of your IT systems. To learn more, visit DataConnect.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale. Actian data intelligence solutions help streamline complex data environments and accelerate the delivery of AI-ready data. Designed to be flexible, Actian solutions integrate seamlessly and perform reliably across on-premises, cloud, and hybrid environments. Learn more about Actian, the data division of HCLSoftware, at actian.com.