Databases

Securing Your Data With Actian Vector, Part 2

Martin Fuerderer

January 13, 2025

Securing Your Data With Actian Vector hero image

Following up on my first blog post about Actian Vector’s database encryption capabilities, the next blog post in the series on data security explains function-based encryption in Actian Vector.

Function-Based Encryption in Actian Vector

The encryption that is completely controlled by a database application is called function-based encryption in Actian Vector. The application uses a pair of built-in SQL functions to encrypt and decrypt individual data values in table columns. Apart from applying these functions on the data as told by the SQL statements, the database server does not handle the encrypted data in any specific way. In particular, the database server is not capable of decrypting the encrypted data on its own.

The application specifies a passphrase with the encrypt function and needs to specify the same passphrase again with the decrypt function. The database server derives a symmetric encryption key from the given passphrase and uses this key to encrypt or decrypt the data. Neither the passphrase nor the derived encryption key are stored anywhere in the database server. The application is always responsible for supplying the correct passphrase to the encrypt and decrypt function.

Typically, the application chooses a passphrase and uses the encrypt function when inserting data values into a table. The data gets encrypted based on the chosen passphrase and then stored. When accessing the stored data, either during select or update, the application uses the decrypt function with the same chosen passphrase to get the data in cleartext. Accessing encrypted data without the decrypt function simply retrieves the encrypted data, which normally is not very useful.

The application can specify different passphrases with individual invocations of the encrypt and decrypt functions. For example, different database users can each use their own passphrase to insert data with encryption into the same table. The users can be sure that no other user can see the cleartext of their own data. Or, an application can encrypt data for different table columns with different passphrases, and provide the different passphrases to different sets of users. That way, one group of users can see the cleartext data only in one column whereas another group of users can see the cleartext of data in the other column.

The encrypt function can be applied multiple times with different passphrases on the same data. To access the data in cleartext, the same passphrases must be used in reverse order to decrypt the data. When updating encrypted data, the decrypt function is applied first, the cleartext value is updated, then the encrypt function is used to newly encrypt the updated data. These are just a few examples of the endless possibilities to combine the use of the encrypt and decrypt functions, all controlled by the application.

All the function-based encryption is done by using the crypto functions in regular SQL statements. This also means that the cleartext passphrases are provided to the crypto functions in the SQL statements. Proper use of function-based encryption therefore should take care that the SQL statements are kept sufficiently secure, e.g. that they are not traced or made visible in some SQL command history.

Also, communication between the application and the database server should be secured to avoid leaking the passphrases. With this regard, the “sql” Terminal Monitor utility is not really a secure example of an application because the typed-in cleartext passphrases in the SQL statements are visibly echoed in the terminal Window.

Examples of Function-Based Encryption:

  • Definition of the table used in the following examples:

CREATE TABLE t1 (name CHAR(20), salary VARCHAR(50)); \g

Even though the salary is a numerical value, the “salary” column is defined as a sufficiently large character column to accommodate the bytes of the encrypted value.

  • Insert data with encryption into a table:

INSERT INTO t1 values ('Joe', AES_ENCRYPT_IV(40000, 'Secret passphrase')); \g
INSERT INTO t1 values ('Adam', AES_ENCRYPT_IV(38000, 'Secret passphrase')); \g
INSERT INTO t1 values ('CEO', AES_ENCRYPT_IV(350000, 'Super passphrase')); \g

The first two statements encrypt the salary value based on the passphrase ‘Secret passphrase’, whereas the third statement uses the different passphrase ‘Super passphrase’.

  • Selecting encrypted data:

    1. Select without using the decrypt function:
SELECT name, salary FROM t1; \g
name          salary
Joe           XnE_msRe1?Uh]!cUUerQn4opT(C[!)O~yTrzE%JS
Adam          lyr*M%o53]O(5d$2`!3a:Bg$zW40wMySO5[-E7+u
CEO           a*#eE!~p(1-euLx?K$w(2)@FHk+dzM`5}Q#=lD@g
(3 rows)

The encrypted data in the “salary” column is shown encrypted and therefore meaningless. The length of the output for the salary column demonstrates that encrypted data usually takes more space than the corresponding cleartext values.

    1. Select using the decrypt function with the first passphrase:
SELECT name, AES_DECRYPT_IV(salary, 'Secret passphrase') FROM t1; \g

name          col2
Joe           40000
Adam          38000
CEO                  \020���x���
(3 rows)

The values encrypted with the first passphrase are correctly decrypted. The value encrypted with a different passphrase is not shown in cleartext.

    1. Select a specific data record using its correct passphrase:
SELECT name, AES_DECRYPT_IV(salary, 'Super passphrase')
  FROM t1 WHERE name = 'CEO'; \g

name		col2
CEO		350000
(1 row)

The encrypted value is shown in cleartext.

When selecting encrypted data, the decrypt function needs the correct passphrase to show the cleartext. When using different passphrases for different data records, it is better to use a WHERE clause that restricts the result to those data records encrypted with the passphrase specified in the SELECT.

  • Updating encrypted data:

When updating encrypted data, the data needs to be decrypted and usually should be encrypted again after the update. For this, both functions, decrypt and encrypt, are used in the same UPDATE statement. At the same time it is also possible to change the passphrase.

UPDATE t1 SET salary = AES_ENCRYPT_IV(
AES_DECRYPT_IV(salary, 'Super passphrase') + 20000,
'Secret passphrase' ) WHERE name = 'CEO'; \g

The statement adds 20000 to the salary of ‘CEO’ and at the same time changes the passphrase to ‘Secret passphrase’. The decrypt function requires the original passphrase and the encrypt function is provided with the new passphrase. As the new passphrase is the same one that was used to encrypt the salary of the other two data records, a SELECT using this passphrase now can show the cleartext for all records:

SELECT name, AES_DECRYPT_IV(salary, 'Secret passphrase') FROM t1; \g

name          col2
Joe           40000
Adam          38000
CEO           370000
(3 rows)

The next blog post in this series will explain how encryption keys for the database encryption are used in Actian Vector.

Martin Fuerderer headshot

About Martin Fuerderer

Martin Fuerderer is a Principal Software Engineer for HCLSoftware, with 25+ years in database server development. His recent focus has been on security features within database environments, ensuring compliance and robust data protection. Martin has contributed to major product releases and frequently collaborates with peers to refine database security standards. On the Actian blog, Martin shares insights on secure database server development and best practices. Check his latest posts for guidance on safeguarding enterprise data.
Data Platform

Why Analyst Reports Matter: Actian Leads the Way in Data Management

Jennifer Jackson

January 8, 2025

Gartner MQ news release blog

In a world where data-driven decisions define success and determine business outcomes, organizations often turn to industry technology analysts for trusted insights when choosing technology. Analyst reports provide a clear picture of emerging trends, competitive landscapes, and the vendors setting the pace.

These insights help businesses across all industries navigate complex markets, make informed IT investments, and strategically align their technology choices with both short- and long-term goals. By leveraging the research and expertise of analysts, companies gain the confidence needed to select the right platform and approach to drive innovation.

Recent analyst reports–several of which highlight Actian products–show how leading technologies play a pivotal role in managing and leveraging data throughout an organization. Here’s why business and IT leaders should pay attention to these reports and why Actian capabilities matter to modern companies.

Enabling Trusted Data Management Across All Environments

Actian has solidified its position as a key player in data management, earning recognition as an Honorable Mention vendor in the 2024 Gartner® Magic Quadrant™ for Cloud Data Management Systems. We are also included in the Forrester Data Management for Analytics Platforms Landscape, 2024. In addition, Actian was named an Innovative provider in ISG Buyers Guide™ for Data Products. These acknowledgments underscore Actian’s leadership in providing comprehensive data management products.

Why managing data matters. Effective data management–from improving data accuracy to integrating disparate sources and enabling actionable insights–is the backbone of modern business success. With the growing reliance on cloud environments, businesses must ensure their data is secure, accessible, and optimized for decision-making. The cloud offers unparalleled scalability and flexibility, but without robust data management, companies risk inefficiencies, compliance challenges, and missed opportunities.

Whether operating in cloud-first, hybrid, or on-premises settings, organizations must have the ability to manage trusted data efficiently. Actian understands that customers have data spread across multiple environments and can help them with their data no matter where it resides.

Supporting a Range of Workloads With Translytical Data Platforms

Forrester defines translytical data platforms as next-generation data solutions built on a single database engine to support transactional, operational, and analytical workloads without impacting data integrity, performance, or real-time analytics. This trend is critical as businesses strive to unlock real-time insights for various workloads. Actian’s inclusion as a Contender in the Forrester Wave™: Translytical Data Platforms, Q4 2024, reflects our leading capabilities to unify transactional and analytical workloads.

Forrester notes platform strengths and offers recommendations for organizations considering implementation:

  • Accelerate use cases through advanced self-service and automation. Look for vendors that provide a multi-persona user interface to facilitate collaboration, automate all translytical functions, offer translytical solutions as a service, and incorporate automation for tiered storage.
  • Enhance data management capabilities to support a wide range of use cases. Look for comprehensive data management support for data integration, curation, vector search, large language model integration, processing, transformation, security, metadata management, and connectivity to diverse data repositories.
  • Enable distributed processing and transactions across diverse data and clouds. Look for a platform that can leverage external data sources while ensuring consistency, such as databases, lakehouses, software as a service, devices, and marketplaces in data processing and transactions across hybrid and multiple clouds.

Why handling multiple workloads matters. Businesses require data platforms that keep up with the demand for real-time insights and operational efficiency. Translytical data platforms are emerging as a modern solution to drive faster insights, reduce complexity, and ensure data integrity. They enable businesses to streamline operations. By performing both transactional and analytical workloads, they allow businesses to streamline operations and support real-time decision-making without the need for separate systems.

Optimizing Industry-Leading Data Platform Software

The 2024 Data Platforms Buyers Guide from Ventana Research/ISG Research offers a comprehensive evaluation of 25 data platform providers, assessing their ability to meet buyers’ diverse needs. Actian’s designation as Exemplary highlights our commitment to delivering a platform that excels in functionality, innovation, and value.

Actian performed exceptionally well in the Manageability category, which evaluates how well products can be managed technologically and by business users, among other criteria. Actian was ranked as a Leader in the category.

With the ability to support both operational and analytic workloads, the Actian platform meets organizations’ need to perform a variety of use cases, including those involving GenAI. As ISG notes, “We assert that through 2027, the development of intelligent applications providing personalized experiences driven by GenAI will increase demand for data platforms capable of supporting hybrid operational and analytic processing.”

With capabilities spanning data integration, quality, and usability, the Actian Data Platform equips businesses to address complex data challenges and capitalize on growth opportunities. The recognition from ISG Research validates Actian’s industry leadership and ability to empower organizations with reliable, high-performing data platform software.

Why understanding the data platform landscape matters. Understanding the data platform ecosystem is crucial for businesses that want to stay competitive. With the right data platform, organizations can unify disparate data sources, improve data governance, and drive actionable insights. Staying informed about the capabilities of top providers helps businesses make strategic investments, ensuring their chosen platform aligns with both current and future needs. By leveraging insights from evaluations like analysts’ buyers guides, companies can confidently select solutions that optimize performance and support their ongoing needs.

Ensuring Data Quality for Diverse Use Cases

Data quality is the foundation of data-driven decision-making and operational excellence. Actian’s recognition as Exemplary in ISG Research’s 2024 Data Quality Buyers Guide reflects our platform’s ability to address business needs with robust data validation, cleansing, and enrichment capabilities. Whether ensuring accurate reporting, improving customer experiences, or supporting GenAI use cases, the Actian platform enables organizations to confidently optimize their data across a variety of use cases.

As ISG noted in its report, data quality is increasingly important for modern tools, such as those supporting machine learning. “Although data quality software has historically provided users with an environment to check and correct data quality issues manually, the use of machine learning (ML) to automate the monitoring of data is being integrated into data quality tools and platforms to ensure that data is complete, valid, and consistent as well as relevant and free from duplication.”

Why data quality matters for analysts and business users. Data quality directly impacts outcomes, including accurate analytics, reliable forecasting, and sound decision-making, reducing the risk of costly mistakes. Poor data quality, on the other hand, leads to inefficiencies, mistrust, and missed opportunities. For data analysts, data quality can be the difference between actionable insights and misleading conclusions, while for business users, it ensures processes run smoothly and decisions deliver value.

Making Data Integration Easy—No IT Skills Required

For data to reach its full potential, it must be integrated on a single platform, then made available to everyone who needs it. As data becomes available from more sources than ever before, connecting data can be tricky. It requires scalability and the ability to integrate systems, applications, and data sources. Requiring IT or specialized skill sets for integration creates bottlenecks and slows down insights.

Actian’s data integration capabilities stand out for their user-friendly design, enabling organizations to connect, transform, and manage data without requiring advanced IT skills. That’s why Actian was recognized as Exemplary in ISG Research’s 2024 Data Integration Buyers Guide. Our platform delivers tools that empower users to take control of their data integration processes. Intuitive interfaces and powerful automation simplify complex workflows, helping businesses unlock insights faster and more efficiently.

Data integration is playing an important role in data intelligence, serving as a critical layer in the tech stack. Data integration is enabling enterprises to extract data from applications, databases, and other sources and combine it for analysis in a data warehouse or data lakehouse with the intention of generating business insights, according to ISG Research. “Without data integration, business data would be trapped in the applications and systems in which it was generated.”

Why simplifying data integration matters. Bringing together data is crucial for organizations striving to keep pace with the rapidly increasing volume and complexity of data. When integration is made easy, teams across the organization benefit by accessing and using data without limitations, fostering innovation and improving collaboration. By reducing the technical barriers to integration, businesses can focus on leveraging their data to create value, improve operations, and respond quickly to market demands.

Proven Value Backed by Analyst Reports

In this era when data is the currency of innovation, organizations must implement industry-recognized technologies that deliver tangible results. Analyst reports serve as guideposts for navigating data products and features. They offer independent validation of vendors’ capabilities and provide a framework for assessing which solutions align with an organization’s specific goals. For businesses evaluating data platforms, these reports help differentiate between hype and genuine leadership.

Actian’s recognition in multiple analyst reports—including those from Gartner, Forrester, and ISG Research—highlights our ability to deliver measurable value across data manageability, quality, and integration, among other benefits. You can read more about our offerings in the IDC Vendor Profile and the Bloor Research InBrief. These accolades reflect not just market trends but the real-world impact of Actian’s product suite.

Now, we have added even more capabilities to optimize data, such as metadata management and data discovery. As organizations grapple with disparate data sources, our ever-expanding range of capabilities enable seamless discovery, understanding, and utilization of data. Have questions about our products? Connect with one of our experts.

Jennifer Jackson headshot

About Jennifer Jackson

Jennifer"JJ" Jackson is CMO of Actian, leading global marketing strategy with a data-driven approach. With 25 years of branding and digital marketing experience and a background in chemical engineering, JJ understands the power of analytics from both a user and marketer perspective. She's spearheaded SaaS transitions, partner ecosystem expansions, and web modernization efforts at companies like Teradata. On the Actian blog, she discusses brand strategy, digital transformation, and customer experience. Explore her recent articles for real-world lessons in data-driven marketing.
AI & ML

The Key to Reaching GenAI’s Full Potential: Data Quality

Actian Corporation

December 31, 2024

The Key to Reaching GenAI’s Full Potential: Data Quality

Generative AI (GenAI) promises to revolutionize industries—from automating business processes and improving decision-making to driving innovation at never-before-seen speeds. However, behind every successful GenAI model lies a foundational truth: GenAI is only as good as the data that fuels it.

When data is incomplete, inconsistent, or inaccurate, even the most advanced GenAI tools will deliver flawed results. As a result, businesses risk poor decisions, operational inefficiencies, compliance failures, and reputational damage.

For organizations looking to optimize and scale GenAI outcomes, ensuring high-quality data isn’t just a technical requirement—it’s a strategic imperative. Like other data-driven use cases, GenAI requires trusted data that gives you full confidence in the results.

Data Quality Matters

Organizations racing to adopt GenAI are realizing that data quality can make or break their investment. High-performing GenAI models rely on large volumes of clean, trusted data to train, predict, and deliver valuable outcomes. Poor data, on the other hand, introduces biases, amplifies errors, and undermines trust.

Consider the following cautionary examples:

These cases demonstrate that flawed inputs—whether incomplete, inconsistent, or inaccurate data—produce unreliable AI outputs. In short, the “garbage in, garbage out” mantra applies to GenAI data.

According to Gartner, 30% of GenAI projects will fail by the end of 2025 due to poor data quality, unclear business value, and inadequate risk controls. These risks are real, but they are also avoidable. Ensuring high-quality data is the single most effective way to unlock the promise of GenAI.

5 Challenges in Ensuring Data Quality

Data professionals face several key challenges when preparing their data for GenAI. They include:

  1. Disparate and Siloed Data Sources
    Many organizations struggle with fragmented data across systems, regions, or departments. Without unified, consistent data, GenAI models lack the full picture to deliver meaningful insights.
  2. Data Volume and Complexity
    Businesses now manage massive, ever-growing data volumes coming from more sources than ever. Managing petabytes of sensor data, transactional records, time-series inputs, and other data requires modern solutions capable of real-time processing.
  3. Outdated or Incomplete Data
    Stale or flawed data leads to inaccurate predictions. For GenAI to remain relevant, data must reflect real-time updates, evolving business needs, and shifting trends.
  4. Lack of Governance and Transparency
    Without proper governance, organizations cannot ensure data accuracy, lineage, and compliance—critical elements for GenAI to meet regulatory and business standards.
  5. Inconsistent Data Quality Processes
    Many organizations lack standardized processes for maintaining data quality across departments and systems. Inconsistencies in data validation, cleansing, and monitoring can result in discrepancies that undermine the accuracy and ultimately the confidence in GenAI outputs. Without a unified approach, organizations struggle to ensure data remains reliable, up-to-date, and aligned with business goals. 

The Solution: Building a Strong Foundation for Data

A unified data platform addresses these challenges, ensuring businesses can deliver clean, trusted, and integrated data to their GenAI models. This type of platform provides a comprehensive solution that supports:

  • Data Integration. The platform should seamlessly integrate data from disparate systems, creating a single, unified source of truth. This eliminates silos and ensures GenAI has access to all relevant inputs.
  • Real-Time Processing. A unified, scalable platform can handle massive data volumes, streaming extremely large data sets to power real-time predictions and insights.
  • Data Quality Management. From data deduplication to error detection, the platform can ensure that data is accurate, complete, and reliable before feeding it into GenAI models.
  • Governance and Compliance. Built-in governance tools in modern data platforms provide transparency and trust, allowing organizations to meet regulatory standards with confidence.
  • Scalability. A platform’s ability to quickly scale supports modern data formats, enabling businesses to integrate all required data to scale their GenAI operations.

With the right platform, businesses can cleanse, connect, and prepare their data for GenAI use cases—whether automating workflows, delivering predictive analytics, optimizing operations, or achieving other business goals. A platform can be a driving force for GenAI success—or limit outcomes. That’s why organizations must understand their needs and implement a platform that meets their current and future requirements.

Achieving GenAI Excellence: The Benefits of Quality Data

When organizations address data quality challenges, they unlock the full potential of GenAI and other data-driven use cases. Ongoing benefits include:

  • Accurate Insights. High-quality inputs result in reliable, actionable outputs, empowering smarter decision-making.
  • Operational Efficiency. Automating tasks with GenAI reduces manual effort, increases productivity, and frees staff time for other tasks.
  • Cost Savings. Clean data minimizes costly errors and accelerates the return on investment (ROI) for GenAI.
  • Innovation. With trusted data, businesses can confidently deploy GenAI for complex use cases, like predictive maintenance, demand forecasting, and customer personalization.
  • Competitive Advantage. Data-driven insights enabled by GenAI allow businesses to move faster, adapt to change quickly, and outperform competitors.

Take the Next Step: Prepare Your Data for GenAI

The journey to successful GenAI adoption starts with data readiness. Organizations that prioritize data quality will not only solve common challenges but also accelerate innovation and deliver measurable business outcomes.

The eBook “Realize the Promise of GenAI Today—and Avoid Common Pitfalls” offers proven strategies to help organizations ensure their data is ready for GenAI. It offers seven steps to achieve data readiness for GenAI and shares strategies to future-proof your data infrastructure for GenAI-driven success. It can equip organizations with the tools and knowledge to make GenAI work for them—powered by clean, trusted, and accurate data.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale. Actian data intelligence solutions help streamline complex data environments and accelerate the delivery of AI-ready data. Designed to be flexible, Actian solutions integrate seamlessly and perform reliably across on-premises, cloud, and hybrid environments. Learn more about Actian, the data division of HCLSoftware, at actian.com.
Best Practices

The 7 Fundamentals That Are Crucial for CDO Success in 2025

Dee Radh

December 24, 2024

CDO fundamentals for success

As data volumes continue to rapidly grow and organizations become increasingly data driven in the AI age, the data landscape of 2025 is poised to be more dynamic and complex than ever before.

For businesses to excel in this fast-evolving environment, chief data officers (CDOs) of the future must move beyond their traditional roles to become strategic transformation leaders. Key priorities will shape their agenda and be a driving force for success in an era of sweeping change.

The eBook “Seven Chief Data Officer (CDO) Priorities for 2025,” explores seven key priorities that will define successful data leadership in 2025. From crafting unified data strategies that feel less like governance manifestos and more like business transformation blueprints, to preparing trusted data for the AI revolution, you will learn:

  1. What tomorrow’s successful CDOs look like.
  2. The seven fundamentals that are crucial for CDO success.
  3. Practical strategies for data management in 2025.

Expanding from Data Custodian to Strategic Visionary

The role of the CDO has undergone a significant change over the last few years—and it’s continuing to be redefined as CDOs prove their value. CDOs are now unlocking competitive advantages by implementing and optimizing comprehensive data initiatives. That’s part of the reason why organizations with a dedicated CDO are better equipped to handle the complexities of modern data ecosystems and maintain a competitive edge than those without this role.

As noted in our eBook “Seven Chief Data Officer (CDO) Priorities for 2025,” this critical position will become even more strategic. The role will highlight a distinct difference between good companies that use data and great companies that rely on data to drive every business decision, accelerate growth, and confidently embrace whatever is next.

The idea for this eBook began with a simple observation: The role of CDO has become a sort of organizational Rorschach test. Ask 10 executives what a CDO should do, and you’ll get 11 different answers, three strategic frameworks, and at least one person insisting it’s all about AI (it’s not).

While researching this piece, a fascinating pattern emerged. Data strategy isn’t just about governance and quality metrics, but about fundamental business transformation. But perhaps most intriguing is the transformation of the CDO role itself. What started as a data custodian and governance guru has morphed into something far more nuanced: Part strategist, part innovator, part ethicist, and increasingly, part business transformer.

The eBook dives deeper into these themes, offering insights and frameworks for navigating this evolution. But more than that, it attempts to capture this moment of transformation–where data leadership is becoming something new and, potentially, revolutionary.

The seven priorities outlined in the eBook aren’t just predictions; they’re emerging patterns. When McKinsey tells us that 72% of organizations struggle with managing data for AI use cases, they’re really telling us something profound about the gap between our technological ambitions and our organizational readiness. We’re all trying to build the plane while flying it–and some of us are still debating whether we need wings.

This eBook is for leaders who find themselves at this fascinating intersection of technology, strategy, and organizational change. Whether you’re a CDO looking to validate your roadmap, or an executive trying to understand why your data initiatives feel like pushing boulders uphill, we hope you’ll find something here that makes you think differently about the journey ahead.

Download the eBook if you’re curious about what data leadership looks like when we stop treating it like a technical function and start seeing it as a strategic imperative.

dee radh headshot

About Dee Radh

As Senior Director of Product Marketing, Dee Radh heads product marketing for Actian. Prior to that, she held senior PMM roles at Talend and Formstack. Dee has spent 100% of her career bringing technology products to market. Her expertise lies in developing strategic narratives and differentiated positioning for GTM effectiveness. In addition to a post-graduate diploma from the University of Toronto, Dee has obtained certifications from Pragmatic Institute, Product Marketing Alliance, and Reforge. Dee is based out of Toronto, Canada.
Data Management

What is Data Stewardship?

Actian Corporation

December 20, 2024

depiction of data stewardship

Data stewardship is a critical business function that oversees and manages an organization’s data assets. It ensures that high-quality data is available to consumers while adhering to data governance policies. In today’s data-driven business landscape, effective data stewardship is essential for organizations that rely on high-quality and efficient data to make informed decisions, drive innovation, and maintain a competitive edge.  

For any organization to implement a successful data program, its members must be aware of their responsibilities. Data stewardship helps ensure stakeholders are given clear responsibilities aligning with overall goals and data governance. This article will explore the concept of data stewardship, its importance in modern businesses, and how it benefits both data consumers and producers. It will also delve into the role of data stewards, best practices for implementing successful data stewardship programs, and background information to help harmoniously integrate data stewardship plans into data governance programs.  

The Role of a Data Steward

To better answer “What is data stewardship?” it helps to understand what a data steward does. Data stewards play a crucial role in ensuring the quality, security, and accessibility of an organization’s data assets. Their responsibilities span various aspects of data management, impacting multiple departments and stakeholders within the business. 

On a daily basis, data stewards: 

  1. Monitor Data Quality: They continuously assess and maintain data quality standards, ensuring that data is accurate, consistent, and reliable across all systems and processes. 
  2. Implement and Enforce Data Governance Policies: Data stewards work closely with the data governance team to implement and enforce data usage, security, and compliance policies. 
  3. Collaborate With Stakeholders: They act as liaisons between different departments, such as IT, business units, and compliance teams, to address data-related issues and requirements. 
  4. Manage Metadata: Data stewards oversee the creation and maintenance of metadata, which helps users understand the context, origin, and quality of data assets. 
  5. Resolve Data Issues: They investigate and resolve data quality issues, discrepancies, and conflicts that arise within the organization. 
  6. Provide Data-Related Training and Support: Data stewards educate users across the organization on data best practices, tools, and processes. 
  7. Participate in Data-Related Projects: They contribute to data integration, migration, and analytics initiatives to ensure data quality and consistency throughout these processes. 

Data stewards commonly work with several departments and stakeholders across the organization. They collaborate closely with IT departments on data infrastructure, security, and integration projects. Compliance and legal teams work hand-in-hand with data stewards to ensure data management practices meet compliance standards. Data analytics teams benefit from the support of data stewards in data preparation and quality assurance for their analytical initiatives. At the executive level, data stewards provide valuable insights on data asset management and its impact on business objectives, helping to shape overall data strategy. 

In a larger organizational sense, data stewards advocate for end users, ensuring that data producers and IT organizations meet the needs of those who rely on high-quality, accessible data. They play a crucial role in supporting regulatory compliance, and by maintaining high-quality data, they enable executives and managers to make informed decisions based on reliable information. Perhaps most importantly, data stewards help organizations recognize and leverage the value of their data assets, promoting data as a strategic asset that can drive business growth and innovation. 

What is Data Stewardship’s Relation to Data Governance?

Data stewardship and data governance work together to ensure the effective management of an organization’s data. While data governance provides the overall framework, policies, and standards for data management, data stewardship focuses on the day-to-day implementation and execution of these policies.  

In terms of organizational structure, data stewardship often operates within the broader data governance framework: 

  • Data Governance Managers: Sets overall strategy and policies. 
  • Data Stewardship Managers: Coordinates data stewardship activities across the organization. 
  • Domain Data Stewards: Manage data within specific business domains or functional areas. 

This hierarchical structure ensures that data stewardship activities align with the organization’s overall data governance goals while allowing for specialized management of data assets within different areas of the business.  

Best Practices for Data Stewardship Programs

Now that you’re familiar with its larger place in an organization, you need to know what data stewardship practices can make your program as successful as possible. Managers of successful data stewardship programs employ several best practices to maximize efficiency and achieve business goals: 

  1. Establish Clear Roles and Responsibilities: Define the roles of data stewards, data owners, and other stakeholders involved in data management. 
  2. Develop Data Quality Standards: Create and maintain detailed standards for data accuracy, completeness, consistency, and timeliness. 
  3. Implement Data Governance Frameworks: Align data stewardship activities with broader data governance initiatives to ensure consistency and effectiveness. 
  4. Leverage Automation and Tools: Utilize data quality tools, metadata management systems, and data catalogs to streamline data stewardship processes. 
  5. Foster Cross-Functional Collaboration: Encourage communication and cooperation between different departments to address data-related challenges. 
  6. Provide Ongoing Training and Education: Offer regular training sessions to keep data stewards and other stakeholders updated on best practices and emerging trends. 
  7. Establish Key Performance Indicators (KPIs): Define and track metrics defined by executives to measure the success of data stewardship initiatives and identify areas for improvement. 
  8. Conduct Regular Audits and Assessments: Perform periodic reviews of data assets, processes, and policies to ensure continued alignment with business objectives and data compliance. 
  9. Implement a Data Catalog: Maintain a centralized metadata repository to help users utilize data assets to their fullest capabilities. 

Benefits of Data Stewardship for Data Consumers

Effective data stewardship provides numerous benefits to data consumers within an organization. By ensuring high-quality, accessible, and reliable data, data stewards enable users to make informed decisions and drive business value. 

Key benefits for data consumers include: 

  • Improved data quality and reliability.
  • Enhanced data discoverability and accessibility.
  • Greater trust in data-driven insights and decisions.
  • Reduced time spent on data preparation and cleansing.
  • Increased compliance with data usage policies and regulations.
  • Streamlined data request and access processes.
  • Reduced risk of using outdated or inaccurate data.
  • Improved collaboration and data sharing between departments.

Benefits of Data Stewardship for Data Producers

Data stewardship not only benefits consumers but also provides significant advantages to data producers. By implementing best practices and standards, data stewards help producers optimize their data management processes and increase the value of their data assets. 

Benefits for data producers include: 

  • Streamlined data production processes.
  • Reduced data redundancy and duplication.
  • Improved data integration capabilities.
  • Increased efficiency in data management workflows.
  • Better alignment between data production and business needs.
  • Improved data lifecycle management.
  • Greater visibility into data usage and value.

Use Cases for Data Stewardship

Data stewardship plays a crucial role across many industries, helping organizations manage their data assets effectively and derive maximum value from their information. Here are some industry-specific use cases that highlight the importance of quality data stewardship programs: 

Healthcare

In the healthcare industry, data stewardship is essential for managing patient information, ensuring regulatory compliance, and supporting medical research: 

  • Electronic Health Records (EHR) Management: Ensuring the accuracy and completeness of patient records, facilitating better care coordination and reducing medical errors. 
  • HIPAA Compliance: Helping maintain patient data privacy and security in accordance with HIPAA regulations. 
  • Clinical Research: Supporting the integration and quality control of data used in medical research and clinical trials. 

Financial Services

The financial sector relies heavily on accurate and timely data for risk management, regulatory reporting, and customer service: 

  • Regulatory Reporting: Ensuring the accuracy and consistency of data used for regulatory reports. 
  • Customer Data Management: Maintaining the quality and integrity of customer information across various systems, supporting Know Your Customer (KYC) initiatives and personalized services. 
  • Fraud Detection: Ensuring the quality and integration of data used in fraud detection algorithms. 

Telecommunications

In the rapidly evolving telecommunications industry, data stewardship is vital for managing customer data, network information, and service quality: 

  • Customer Experience Management: Improving the accuracy and consistency of customer data, enabling personalized services and effective customer support. 
  • Network Performance Monitoring: Managing the quality and reliability of data used to monitor network performance and plan infrastructure improvements. 
  • Regulatory Compliance: Supporting compliance with telecommunications regulations by ensuring data accuracy in regulatory reporting. 

Manufacturing 

In manufacturing, data stewardship is crucial for optimizing production processes, managing supply chains, and ensuring product quality: 

  • Supply Chain Optimization: Maintaining accurate and up-to-date information on suppliers, inventory, and logistics, enabling better supply chain management. 
  • Quality Control: Stabilizing the integrity and consistency of data used in quality control processes, supporting continuous improvement initiatives. 
  • Predictive Maintenance: Managing the data used for predictive maintenance models, helping to reduce downtime and maintenance costs. 

Actian Supports Data Stewardship Best Practices

Actian provides a comprehensive suite of data management solutions that support and enhance data stewardship practices across organizations. The Actian Data Intelligence Platform offers a robust foundation for implementing effective data stewardship programs through its various features and capabilities. Its data integration tools, including DataConnect, help organizations maintain high-quality data by providing robust Extract, Transform, Load (ETL) capabilities and quality checks. And businesses can utilize the Actian Data Intelligence Platform to standardize data storage and usage according to data governance policies, which increases data democratization while making data stewardship simpler and easier to understand for stakeholders. 

Actian’s solutions integrate seamlessly with data governance frameworks, helping organizations align their data stewardship practices with broader governance initiatives. As organizations grow and evolve, the Actian Data Platform is designed to scale with increasing data volumes and user communities, providing ongoing support for data stewardship efforts. More help creating large-scale data stewardship initiatives can be found in the blog “The Importance of Data Stewardship Agreements Across Companies.” 

Key Takeaways

data stewardship

 

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale. Actian data intelligence solutions help streamline complex data environments and accelerate the delivery of AI-ready data. Designed to be flexible, Actian solutions integrate seamlessly and perform reliably across on-premises, cloud, and hybrid environments. Learn more about Actian, the data division of HCLSoftware, at actian.com.
Data Platform

What is a Data Marketplace and Why Should You Use One?

Actian Corporation

December 20, 2024

data marketplace depiction

Collecting data takes time, but the more information you have at your disposal, the easier it is to plan your business’s growth strategies and identify areas of improvement. Though many businesses rely on data they collect in-house, sourcing your own information can often result in more limited data sets that make it difficult to put a full picture together. That’s where data marketplaces can help.

Let’s take a closer look at how these marketplaces work and why businesses should consider leveraging marketplaces when collecting data.

What is a Data Marketplace?

A data marketplace is essentially an online store for data where businesses and individuals alike can buy and sell information as well as exchange data to facilitate growth. These marketplaces allow those who collect data to turn a profit from the information they gather and make it easy for businesses to access data that would otherwise take weeks or months to collect on their own.

Though both marketplaces and data exchanges allow people to trade data sets, they do so in different ways. In marketplaces, individuals and businesses have the opportunity to buy, sell or trade information. The information can be purchased or traded by anyone.

On data exchanges, the system works more as an even trade between approved organizations and individuals. The data provider can determine who is eligible to receive and share data with or restrict it from certain parties as needed.

Both systems can help people access critical data to help them drive decisions, but marketplaces can be more flexible.

What Types of Information Can You Get on a Data Marketplace?

The information available on each marketplace can vary. However, there are a few common types of data that people typically offer for sale or exchange on most marketplaces. This includes but is not limited to the following:

  • Demographic information.
  • Consumer behavior.
  • Social media information.
  • Scientific studies.
  • Public records.
  • Market conditions and trends in industries.
  • Sales data for specific markets.

If you’re looking for something specific, you may need to search several data marketplaces to find the information you’re looking for at a price point that works for your budget. Marketplaces like Kaggle and World Bank Data may help you round out your data sets and give you access to the information you need.

The Benefits of Using a Data Marketplace

Data marketplaces give you access to more data than you can reasonably compile on your own. But that’s far from the only benefit they can provide. Here’s how using a data marketplace can streamline your business:

  • Better Collaboration: Data marketplaces foster an open-source culture of publishing and sharing data with the knowledge that more data means more productive and healthier societies.
  • Opportunity to Monetize Data: You’re able to both purchase and sell data on a marketplace. This means you can make money from the data you already collect through the course of your business, thereby increasing your cashflow and diversifying your income sources.
  • Improved Efficiency: Purchasing data on marketplaces can take just a few minutes compared to the months it may take to collect the data on your own. This improves efficiency and helps your team make informed decisions faster.
  • Enhanced Productivity: Getting the right data from a marketplace and cleansing it against your data quality standards gives you all the information you need to begin analysis. Make sure the dataset meets your data governance protocol to ensure the dataset remains current. This means your team will spend less time formatting and interpreting the data to fit their needs and more time producing quality work.
  • Easy to Integrate: Data marketplaces offer the assets you’re looking for in different formats that can seamlessly integrate with your existing data fabric. Key stakeholders will be able to access the information in a way that works for their needs.

Keep in mind that these are just a few ways that data marketplaces can improve your business and your data acquisition efforts.

How the Enterprise Data Marketplace Can Help Your Business

Though there are many data marketplaces that you can use to gain access to essential information, Actian Data Intelligence Platform’s new data marketplace can take the stress out of choosing a quality service that you can trust.

The data intelligence platform allows you to access data in a familiar format. The interface is designed similarly to an e-commerce site, so you can pick and choose data based on your needs and queries. Even better, the system can integrate with Actian’s DataConnect service to make interpreting, organizing, and managing that newly purchased data a breeze.

Transform Your Information into a Strategic Data Asset

With data marketplaces, you’re able to access new data, transforming the information you rely on into a strategic asset that can benefit your business in multiple ways. Explore the different assets available to you and your team on the Enterprise Data Marketplace today.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale. Actian data intelligence solutions help streamline complex data environments and accelerate the delivery of AI-ready data. Designed to be flexible, Actian solutions integrate seamlessly and perform reliably across on-premises, cloud, and hybrid environments. Learn more about Actian, the data division of HCLSoftware, at actian.com.
Data Governance

What is Data Discovery and Why Does it Matter?

Actian Corporation

December 19, 2024

data discovery blog

No matter what industry your business is in, you rely on complex and growing data volumes to keep your organization running efficiently and sustainably. There are many ways to collect and organize that data, but to put all that information to use, you need to have trusted systems in place to interpret and optimize that information. The data discovery process aims to simplify your efforts. Here’s what you need to know about the process and why it’s so essential for businesses.

What Exactly is Data Discovery?

Data discovery is the process of collecting data that businesses like yours rely on daily, along with using various tools and programs to analyze the collected information. During analysis, you’ll be able to spot trends, identify patterns, and find ways to improve your operations. This form of data analysis relies on visual presentations of information. As a result, it is easy for anyone, not just experienced data scientists and data analysts, to interpret and understand the insights the data is revealing.

Many data discovery tools use Artificial Intelligence (AI) to sort through and interpret complex data sets. This effectively makes it easier for data users at all levels of the business to access information clearly and concisely.

Data Discovery vs. Traditional Data Analysis

Traditional data analysis relies on predefined queries to sort and interpret data. This means users need to know what they’re looking for and how to input that specific query into the system to gain access to relevant information.

Data discovery, on the other hand, offers a more flexible approach. Users can interact with the data and sort the information in a way that meets their needs each time. The query can be adjusted on the fly and doesn’t have to conform to pre-existing rules or benchmarks.

Why Data Discovery is Important

Data discovery can simplify your data management and analysis efforts, making data more accessible and user-friendly at all levels. Here are a few benefits of incorporating this type of data management into your organization:

  • Improved Compliance: Data discovery lets you better interpret and understand your data. This makes it easier to spot compliance violations quickly so you can course correct before damage is done.
  • Better Efficiency: This method of data analytics makes it easier to interpret data and get the information you need quickly. This boosts efficiency and helps key stakeholders make important decisions in less time.
  • Real-Time Monitoring of Information: AI systems commonly used for data discovery methods are capable of monitoring information 24/7. This makes it easier to identify risks and make adjustments to keep those risks from turning into major issues.
  • Improved Workflows: Because data discovery systems rely on programs and data analytics software to interpret data, they automate much of your data team’s workflow. This automation frees up teams for more important tasks while also reducing the risk of human error.
  • Deeper Insights into Customer Behavior: Understanding your customers’ needs and behaviors can help you make better business decisions. Data discovery methods help you evaluate those behaviors and identify key areas of improvement to drive sales.

Implementing a data discovery system can help your business make sense of complex data and streamline your operation so you can grow sustainably and strategically.

How the Data Discovery Process Works

Though the discovery data management method can be customized to fit your business’s unique needs, the process you’ll want to use will largely be the same across different industries. These are the four steps you’ll need to take:

1. Set Clear Goals

As with any form of data management, you’ll need to identify clear goals for your discovery process. This will help ensure that the data you collect and analyze meets your business’s needs. Some goals may include monitoring and identifying customer behavior, tracking performance of certain products or services, and other similar targets.

Figure out how those business goals can be achieved by leveraging the data you collect and determine what types of information you want the database to contain. Remember that your goals can change over time, so you may need to re-evaluate throughout the year.

2. Gather Your Data

Data discovery software lets you collect raw data based on your goals and analyze that information to help you figure out where you need to improve and what you’re doing right. Collect data for each key goal you’re trying to reach. Then, let your system analyze the information for you. Comprehensive data discovery tools will help you take that raw data and turn it into visual representations that everyone on your team can understand.

3. Analyze the Visualizations

The data you collect still needs a human layer of review. This is your opportunity to look at the information and see what it tells you. Take your time and go through each dataset closely. If you find that certain information is lacking, go back through the discovery process and fill in the gaps. Your tools will help you do this based on the parameters you specify.

4. Repeat the Process

Data discovery is an ongoing process that can grow and change with your business. It’s not something you do only once. Use the information you’ve gathered to make the necessary changes and repeat the discovery process to see how behaviors have changed. This information can help you make better-informed business decisions that support your long-term goals.

What to Look for in Your Data Discovery Tools

Sorting through tons of data requires comprehensive data discovery software and tools. Every tool works in slightly different ways, so make sure you’re choosing the best programs for your needs from the beginning. Look for these important characteristics as you explore your options and identify ways to power your data strategy:

  • Easy to Use Solutions: The tools you select should be easy to use and designed with the average data team skill set in mind. You shouldn’t have to hire an experienced data scientist to show you how to use the system or interpret the information.
  • Multiple Visualization Methods and Options: Different types of data can benefit from different forms of visualization. Review the methods and options each tool offers and choose the system that can present data in a way that your team can understand and use.
  • AI-Powered Analytics to Make Interpretations Easier and Faster: You have better things to do than spending hours analyzing and interpreting complex datasets. Look for tools that feature AI-powered analytics systems to help take the strain off of your team while still giving you actionable insights for your brand.
  • Mobile-Friendly Interfaces to Improve Accessibility: There may be times when you need to access information away from your headquarters and your computers. Make sure the tools you choose are designed with mobile devices like smartphones and tablets in mind.

The right tools can make all the difference in your data management and data governance efforts.

Gain the Insights Your Business Needs

Data discovery should help your business improve your processes and make accessing and interpreting the information you collect easier and faster. With the Actian Data Intelligence Platform, you can take control of your data and put it to work for your business. Learn more about our platform and see how it can simplify your business.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale. Actian data intelligence solutions help streamline complex data environments and accelerate the delivery of AI-ready data. Designed to be flexible, Actian solutions integrate seamlessly and perform reliably across on-premises, cloud, and hybrid environments. Learn more about Actian, the data division of HCLSoftware, at actian.com.
Data Management

Why Next-Gen Data Catalogs are a Must-Have

Traci Curran

December 17, 2024

data-catalogs-for-modern-data-management

Next-Gen Data Catalogs

Data management is undergoing a significant transformation with the emergence of next-generation data catalogs. As organizations face an unprecedented surge in data volumes across their operations, sophisticated yet user-friendly tools are becoming increasingly necessary to manage this information overload effectively. Next-generation data catalogs have emerged as a compelling solution, offering an impressive array of advanced features and intuitive functionality that caters to diverse user needs.

Data Catalogs Make Big Moves

The evolution of data catalogs in recent years has been remarkable. While proficient at metadata organization and management, traditional catalog systems were primarily designed with IT professionals in mind, resulting in interfaces and workflows that proved challenging for users with limited technical expertise. As businesses across industries began generating and exchanging increasingly larger volumes of data, a pressing need arose for non-technical users to harness data insights for various business objectives and strategic decision-making processes.


Modern data catalogs have undergone a fundamental shift in their approach, prioritizing user experience and accessibility. These next-generation solutions emphasize intuitive data discovery, comprehensive understanding, and seamless access to critical information regardless of technical proficiency. This transformation aligns perfectly with the contemporary perspective that data represents a valuable organizational asset that should be readily available to all stakeholders who require it for their roles and responsibilities.

Reasons for the Shift

The evolution of data catalogs can be attributed to several compelling factors:

  • More Data Getting Made: The exponential growth in data generation across various sources and platforms necessitates more sophisticated and efficient data discovery, management, and utilization methods.
  • People Want to Crunch Numbers Themselves: Business professionals increasingly seek data autonomy, driving the demand for self-service analytics capabilities through user-friendly catalog interfaces.
  • Moving to the Cloud: The distributed nature of modern data environments, spanning cloud services, on-premises data centers, and third-party applications, creates an urgent need for enhanced visibility and unified data management.
  • Playing by the Rules: Stringent regulatory requirements and compliance standards mandate a comprehensive understanding of data lineage, governance, and metadata management.
  • Smart Tech Gets Smarter: Advanced artificial intelligence and machine learning capabilities continue to evolve, offering increasingly sophisticated automated solutions for data classification, tagging, and discovery processes.

The Eckerson Group recently published a Top Tier Report on Next-Generation Data Catalogs that provides detailed insights into the future trajectory of this technology space. The report emphasizes that next-generation data catalogs need to focus on ease of use, speed of deployment, and affordability to meet evolving business requirements effectively.

In their comprehensive analysis, the report specifically highlighted Actian Data Intelligence Platform, noting that it stands out as unique among data catalog vendors. The platform supports an enterprise data marketplace in which multiple business domains can share data products in a seamless fashion. When business users request access to a data product, the platform routes their requests through a third-party workflow system, such as Jira or ServiceNow. Actian Data Intelligence Platform also integrates with numerous third-party applications, such as Monte Carlo and Soda, to import metadata into the catalog. This innovative approach demonstrates the platform’s commitment to delivering a comprehensive and integrated data management solution.

Ways to Use and Put Next-Gen Data Catalogs into Action

Next-generation data catalogs are revolutionizing how organizations manage and utilize their data assets. By prioritizing ease of use, data accessibility, and improved decision-making, these innovative tools empower businesses to navigate the complexities of modern data landscapes with unprecedented efficiency. The emergence of enterprise data marketplaces, as exemplified by the Actian Data Intelligence Platform, marks a significant leap forward in collaborative data sharing and governance. These marketplaces facilitate seamless data exchange between different business domains while maintaining strict security protocols and access controls.

As data continues to grow exponentially in volume and importance, the role of next-gen catalogs in driving business success cannot be overstated. Their ability to democratize data access while maintaining robust governance frameworks positions them as indispensable assets for forward-thinking organizations. These catalogs enable organizations to implement comprehensive data governance strategies, ensuring compliance with regulatory requirements while fostering a culture of data-driven decision-making. Through automated metadata management and advanced search capabilities, they significantly reduce the time and effort required to locate and utilize relevant data assets.

Curious about harnessing the power of next-gen data catalogs for your business? Explore Actian and discover how it can transform your data management strategy. In this era of data-driven decision-making, the question isn’t whether you can afford to invest in advanced data catalog solutions—it’s whether you can afford not to. With features like automated data discovery, intelligent metadata management, and seamless integration capabilities, Actian provides a comprehensive solution that addresses the evolving needs of modern enterprises while ensuring scalability and future-readiness.

Traci Curran headshot

About Traci Curran

Traci Curran is Director of Product Marketing at Actian, focusing on the Actian Data Platform. With 20+ years in tech marketing, Traci has led launches at startups and established enterprises like CloudBolt Software. She specializes in communicating how digital transformation and cloud technologies drive competitive advantage. Traci's articles on the Actian blog demonstrate how to leverage the Data Platform for agile innovation. Explore her posts to accelerate your data initiatives.
Databases

User-Friendly External Smartblobs Using a Shadow Directory

Mary Schulte

December 16, 2024

smartblobs using a shadow directory

I am very excited about the HCL Informix® 15 external smartblob feature.

If you are not familiar with them, external smartblobs allow the user to store actual Binary Large Object (blob) and Character Large Object (clob) data external to the database. Metadata about that external storage is maintained by the database.

Notes: This article does NOT discuss details of the smartblobs feature itself, but rather proposes a solution to make the functionality more user-friendly. For details on feature behavior, setup, and new functions, see the documentation.

At the writing of this blog, v15.0 does not have the ifx_lo_path function defined, as required below.  This has been reported to engineering.  The workaround is to create it yourself with the following command:

create dba function ifx_lo_path(blob)
  returns lvarchar
  external name '(sq_lo_path)'
  language C;

This article also does not discuss details of client programming required to INSERT blobs and clobs into the database.

The external smartblob feature was built for two main reasons:

1. Backup size

Storing blobs in the database itself can cause the database to become extremely large. As such, performing backups on the database takes an inordinate amount of time, and 0 level backups can be impossible. Offloading the actual blob contents to an external file system can lessen the HCL Informix backup burden by putting the blob data somewhere else. The database still governs the storage of, and access to, the blob, but the physical blob is housed elsewhere/externally.

2. Easy access to blobs

Users would like easy access to blob data, with familiar tools, without having to go through the database. 

Using External Smartblobs in HCL Informix 15

HCL Informix 15 introduces external smartblobs. When you define an external smartblob space, you specify the external directory location (outside the database) where you would like the actual blob data to be stored. Then you assign blob column(s) to that external smartblob space when you CREATE TABLE. When a row is INSERTed, HCL Informix stores the blob data in the defined directory using an internal identifier for the filename.

Here’s an example of a customer forms table: custforms (denormalized and hardcoded for simplicity). My external sbspace directory is /home/informix/blog/resources/esbsp_dir1.

CREATE TABLE custforms(formid SERIAL, company CHAR(20), year INT, lname CHAR(20), 
formname CHAR(50), form CLOB) PUT form IN (esbsp);

Here, I INSERT a 2023 TaxForm123 document from a Java program for a woman named Sanchez, who works for Actian:

try(PreparedStatement p = c.prepareStatement("INSERT INTO custforms 
(company, year, lname, formname, form) values(?,?,?,?,?)");

FileInputStream is = new FileInputStream("file.xml")) {
p.setString(1, "Actian");
p.setString(2, "2023");
p.setString(3, "Sanchez");
p.setString(4, "TaxForm123");
p.setBinaryStream(5, is);
p.executeUpdate();
}

After I INSERT this row, my external directory and file would look like this:

[informix@schma01-rhvm03 resources]$ pwd
/home/informix/blog/resources
[informix@schma01-rhvm03 resources]$ ls -l esbsp*
-rw-rw---- 1 informix informix 10240000 Oct 17 13:22 esbsp_chunk1

esbsp_dir1:
total 0
drwxrwx--- 2 informix informix 41 Oct 17 13:19 IFMXSB0
[informix@schma01-rhvm03 resources]$ ls esbsp_dir1/IFMXSB0
LO[2,2,1(0x102),1729188125]

Where LO[2,2,1(0x102),1729188125]is an actual file that contains the data that I could access directly. The problem is that if I want to directly access this file for Ms. Sanchez, I would first have to figure out that this file belongs to her and is the tax document I want. It’s very cryptic!

A User-Friendly Smartblob Solution

When talking to Informix customers, they love the new external smartblobs feature but wish it could be a little more user-friendly.

As in the above example, instead of putting Sanchez’s 2023 TaxForm123 into a general directory called IFMXSB0 in a file called LO[2,2,1(0x102),1729188125, which together are meaningless to an end-user, wouldn’t it be nice if the file was located in an intuitive place like /home/forms/Actian/2024/TaxForm123/Sanchez.xml or something similar…something meaningful…how YOU want it organized?

Having HCL Informix automatically do this is a little easier said than done, primarily because the database would not intuitively know how any one customer would want to organize their blobs. What exact directory substructure? From what column or columns do I form the file names? What order? All use cases would be different.

Leveraging a User-Friendly Shadow Directory

The following solution shows how you can create your own user-friendly logical locations for your external smartblobs by automatically maintaining a lightweight shadow directory structure to correspond to actual storage locations. The solution uses a very simple system of triggers and stored procedures to do this.

Note: Examples here are shown on Linux, but other UNIX flavors should work also.

How to Set Up in 4 Steps

For each smartblob column in question

STEP 1: Decide how you want to organize access to your files.

Decide what you want the base of your shadow directory to be and create it. In my case for this blog, it is: /home/informix/blog/resources/user-friendly. You could probably implement this solution without a set base directory (as seen in the examples), but that may not be a good idea because users would unknowingly start creating directories everywhere.

STEP 2: Create a create_link stored procedure and corresponding trigger for INSERTs.

This procedure makes sure that the desired data-driven subdirectory structure exists from the base (mkdir -p), then forms a user-friendly logical link to the Informix smartblob file.    You must pass all the columns to this procedure from which you want to form the directory structure and filename from the trigger.

CREATE PROCEDURE

CREATE PROCEDURE create_link (p_formid INT, p_company CHAR(20), p_year INT,
p_lname CHAR(20), p_formname CHAR(50))
DEFINE v_oscommand CHAR(500);
DEFINE v_custlinkname CHAR(500);
DEFINE v_ifmxname CHAR(500);
DEFINE v_basedir CHAR(100);
-- set the base directory
LET v_basedir = '/home/informix/blog/resources/user-friendly';
-- make sure directory tree exists
LET v_oscommand = 'mkdir -p ' || TRIM(v_basedir) || '/' || TRIM(p_company) || '/' || 
TO_CHAR(p_year);
SYSTEM v_oscommand; 

-- form full link name 
LET v_custlinkname = TRIM(v_basedir) || '/' || TRIM(p_company) || '/' || TO_CHAR(p_year) 
|| '/' || TRIM(p_lname) || '.' || TRIM(p_formname) || '.' || TO_CHAR(p_formid);

-- get the actual location 
SELECT IFX_LO_PATH(form::LVARCHAR) INTO v_ifmxname FROM custforms WHERE formid = p_formid; 

-- create the os link 
LET v_oscommand = 'ln -s -f ' || '''' || TRIM(v_ifmxname) || '''' || ' ' || v_custlinkname; 
SYSTEM v_oscommand;

END PROCEDURE

CREATE TRIGGER

CREATE TRIGGER ins_tr INSERT ON custforms REFERENCING new AS post
FOR EACH ROW(EXECUTE PROCEDURE create_link (post.formid, post.company,
post.year, post.lname, post.formname));

STEP 3: Create a delete_link stored procedure and corresponding trigger for DELETEs.

This procedure will delete the shadow directory link if the row is deleted.

CREATE PROCEDURE

CREATE PROCEDURE delete_link (p_formid INT, p_company CHAR(20), p_year INT,
p_lname CHAR(20), p_formname CHAR(50))
DEFINE v_oscommand CHAR(500);
DEFINE v_custlinkname CHAR(500); 
DEFINE v_basedir CHAR(100);
-- set the base directory
LET v_basedir = '/home/informix/blog/resources/user-friendly';
-- form full link name
LET v_custlinkname = TRIM(v_basedir) || '/' ||
TRIM(p_company) || '/' || TO_CHAR(p_year) || '/' || TRIM(p_lname) || '.'
|| TRIM(p_formname) || '.' || TO_CHAR(p_formid);
-- remove the link
LET v_oscommand = 'rm -f -d ' || v_custlinkname;
SYSTEM v_oscommand;

END PROCEDURE

CREATE TRIGGER

CREATE TRIGGER del_tr DELETE ON custforms REFERENCING old AS pre FOR EACH ROW
(EXECUTE PROCEDURE delete_link (pre.formid, pre.company, pre.year, pre.lname, pre.formname));

STEP 4: Create a change_link stored procedure and corresponding trigger for UPDATEs, if desired.   In my example, Ms. Sanchez might get married to Mr. Simon and an UPDATE to her last name in the database occurs. I may then want to change all my user-friendly names from Sanchez to Simon.  This procedure deletes the old link and creates a new one.

Notice the update trigger only must fire on the columns that form your directory structure and filenames.

CREATE PROCEDURE

CREATE PROCEDURE change_link (p_formid INT, p_pre_company CHAR(20), 
p_pre_year INT, p_pre_lname CHAR(20), p_pre_formname CHAR(50), p_post_company CHAR(20), 
p_post_year INT, p_post_lname CHAR(20), p_post_formname CHAR(50))

DEFINE v_oscommand CHAR(500);
DEFINE v_custlinkname CHAR(500);
DEFINE v_ifmxname CHAR(500);
DEFINE v_basedir CHAR(100);
-- set the base directory
LET v_basedir = '/home/informix/blog/resources/user-friendly';

-- get rid of old

-- form old full link name
LET v_custlinkname = TRIM(v_basedir) || '/' || TRIM(p_pre_company) || '/' || 
TO_CHAR(p_pre_year) || '/' || TRIM(p_pre_lname) || '.' || TRIM(p_pre_formname) || '.' 
|| TO_CHAR(p_formid) ;

-- remove the link and empty directories
LET v_oscommand = 'rm -f -d ' || v_custlinkname;
SYSTEM v_oscommand;

-- form the new
-- make sure directory tree exists
LET v_oscommand = 'mkdir -p ' || TRIM(v_basedir) || '/' || TRIM(p_post_company) || '/' || 
TO_CHAR(p_post_year);
SYSTEM v_oscommand;

-- form full link name
LET v_custlinkname = TRIM(v_basedir) || '/' || TRIM(p_post_company) || '/' || 
TO_CHAR(p_post_year) || '/' || TRIM(p_post_lname) || '.' || TRIM(p_post_formname) 
|| '.' || TO_CHAR(p_formid) ;

-- get the actual location
-- this is the same as before as id has not changed
SELECT IFX_LO_PATH(form::LVARCHAR) INTO v_ifmxname FROM custforms WHERE formid = p_formid;

-- create the os link
LET v_oscommand = 'ln -s -f ' || '''' || TRIM(v_ifmxname) || '''' || ' ' || v_custlinkname;
SYSTEM v_oscommand;

END PROCEDURE

CREATE TRIGGER

CREATE TRIGGER upd_tr UPDATE OF formid, company, year, lname, formname ON custforms
REFERENCING OLD AS pre NEW as post

FOR EACH ROW(EXECUTE PROCEDURE change_link (pre.formid, pre.company, pre.year, pre.lname, 
pre.formname, post.company, post.year, post.lname, post.formname));

Results Example

Back to our example.

With this infrastructure in place, now in addition to the Informix-named file being in place, I would have these user-friendly links on my file system that I can easily locate and identify.

INSERT

[informix@schma01-rhvm03 2023]$ pwd
/home/informix/blog/resources/user-friendly/Actian/2023
[informix@schma01-rhvm03 2023]
$ ls Sanchez.TaxForm123.2

If I do an ls -l, you will see that it is a link to the Informix blob file.

[informix@schma01-rhvm03 2023]$ ls -l
total 0
lrwxrwxrwx 1 informix informix 76 Oct 17 14:20 Sanchez.TaxForm123.2 -> 
/home/informix/blog/resources/esbsp_dir1/IFMXSB0/LO[2,2,1(0x102),1729188126]

UPDATE

If I then update her last name with UPDATE custforms SET lname = ‘Simon’ where formid=2,my file system now looks like this:

[informix@schma01-rhvm03 2023]$ ls -l
lrwxrwxrwx 1 informix informix 76 Oct 17 14:25 Simon.TaxForm123.2 -> 
/home/informix/blog/resources/esbsp_dir1/IFMXSB0/LO[2,2,1(0x102),1729188126]

DELETE

If I then go and DELETE this form with DELETE FROM custforms where formid=2, my directory structure looks like this:

[informix@schma01-rhvm03 2023]$ pwd
/home/informix/blog/resources/user-friendly/Actian/2023
[informix@schma01-rhvm03 2023]$ ls
[informix@schma01-rhvm03 2023]$

We Welcome Your Feedback

Please enjoy the new HCL Informix15 external smartblob feature.

I hope this idea can make external smartblobs easier for you to use. If you have any feedback on the idea, especially on enhancements or experience in production, please feel free to contact me at mary.schulte@hcl-software.com. I look forward to hearing from you!

Find out more about the launch of HCL Informix 15.

Notes

1. Shadow directory permissions. In creating this example, I did not explore directory and file permissions, but rather just used general permissions settings on my sandbox server. Likely, you will want to control permissions to avoid some of the anomalies I discuss below.

2. Manual blob file delete. With external smartblobs, if permissions are not controlled, it is possible that a user might somehow delete the physical smartblob file itself from its directory. HCL Informix, itself, cannot control this from happening. In the event it does happen, HCL Informix does NOT delete the corresponding row; the blob file will just be missing. There may be aspects to links that can automatically handle this, but I have not investigated them for this blog.

3. Link deletion in the shadow directory. If permissions are not controlled, it is possible that a user might delete a logical link formed by this infrastructure. This solution does not detect this. If this is an issue, I would suggest a periodic maintenance job that cross references the shadow directory links to blob files to detect missing links. For those blobs with missing links, write a database program to look up the row’s location with the IFX_LO_PATH function, and reform the missing link.

4. Unique identifiers. I highly recommend using unique identifiers in this solution. In this simple example, I used formid. You don’t want to clutter things up, of course, but depending on how you structure your shadow directories and filenames, you may need to include more unique identifiers to avoid directory and link names duplication.

5. Empty directories. I did not investigate if there are options to rm in the delete stored procedure to clean up empty directories that might remain if a last item is deleted.

6. Production overhead. It is known that excessive triggers and stored procedures can add overhead to a production environment. For this blog, it is assumed that OLTP activity on blobs is not excessive, therefore production overhead should not be an issue. This being said, this solution has NOT been tested at scale.

7. NULL values. Make sure to consider the presence and impact of NULL values in columns used in this solution. For simplicity, I did not handle them here.

Informix is a trademark of IBM Corporation in at least one jurisdiction and is used under license.

 

mary schulte headshot

About Mary Schulte

Mary Schulte is Senior Sales Engineer at Actian, drawing upon decades of experience with powerhouse database vendors like Informix and Netezza. She has written thousands of lines of Informix 4GL and ESQL/C for global clients, including American Airlines' groundbreaking datablade implementation. Mary has delivered countless training sessions, helping organizations optimize their database environments. Her posts on the Actian blog center on query performance, analytics databases like Vector, and practical tips for leveraging Informix. Browse her articles for expert guidance.
Data Management

Accelerating Innovation: Data Discovery in Manufacturing

Kasey Nolan

December 9, 2024

data discovery drives competitive manufacturing advantage

The manufacturing industry is in the midst of a digital revolution. You’ve probably heard these buzzwords: Industry 4.0, IoT, AI, and machine learning– all terms that promise to revolutionize everything from assembly lines to customer service. Embracing this digital transformation is key in improving your competitive advantage, but new technology doesn’t come without its own challenges. Each new piece of technology needs one thing to deliver innovation: data.

Data is the fuel powering your tech engines. Without the ability to understand where your data is, whether it’s trustworthy, or who owns the datasets, even the most powerful tools can overcomplicate and confuse the best data teams. That’s where modern data discovery solutions come in. They’re like the backstage crew making sure everything runs smoothly– connecting systems, tidying up the data mess, and making sure everyone has exactly what they need, when they need it. That means faster insights, streamlined operations, and a lower total cost of ownership (TCO). In other words, data access is the key to staying ahead in today’s fast-paced, highly competitive, increasingly sensitive manufacturing market. 

Data Discovery Manufacturing Problems

Data from all aspects of your business is siloed– whether it’s coming from sensors, legacy systems, cloud applications, suppliers or customers– trying to piece it all together is daunting, time-consuming, and just plain hard. Traditional methods are slow, cumbersome, and definitely not built for today’s needs. This fragmented approach not only slows down decision-making, but keeps you from tapping into valuable insights that could drive innovation. And in a market where speed is everything, that’s a recipe for falling behind. 

So the big question is: how can you unlock the true potential of your data?

Data Discovery Manufacturing Solutions

So how do you make data intelligence into a streamlined, efficient process? The answer lies in modern data discovery solutions– the unsung catalyst of a digital transformation motion. Rather than simply integrating data sources, data discovery solutions excel in metadata management, offering complete visibility into your company’s data ecosystem. They enable users– regardless of skill level– to locate where data resides and assess the quality and relevance of the information. By providing this detailed understanding of data context and lineage, organizations can confidently leverage accurate, trustworthy datasets, paving the way for informed decision-making and innovation, 

Key Components

Easy-to-Connect Data Sources for Metadata Management

 One of the biggest hurdles in data integration is connecting to a variety of data sources, including legacy systems, cloud applications, and IoT devices. Modern tools like the Actian Data Intelligence Platform offer easy connectivity, allowing you to extract metadata from various sources seamlessly. This unified view eliminates silos and enables faster, more informed decision-making across the organization.

Advanced Metadata Management

Metadata is the backbone of effective data discovery. Advanced metadata management capabilities ensure that data is well-organized, tagged, and easily searchable. This provides a clear context for data assets, helping you understand the origin, quality, and relevance of your data. This means better data search and discoverability.

Data Discovery Knowledge Graph

A data discovery knowledge graph serves as an intelligent map of your metadata, illustrating the intricate relationship and connections between data assets. It provides users with a comprehensive view of how data points are linked across systems, offering a clear picture of data lineage– from origin to current state. The visibility into the data journey is invaluable in manufacturing, where understanding the flow of information between production data, supply chain metrics, and customer feedback is critical. By tracing the lineage of data, you can quickly assess its accuracy, relevance, and context, leading to more precise insights and informed decision-making.

Quick Access to Quality Data Through Data Marketplace

A data marketplace provides a centralized hub where you can easily search, discover, and access high-quality data. This self-service model empowers your teams to find the information they need without relying on IT, accelerating time to insight. The result? Faster product development cycles, improved process efficiency, and enhanced decision-making capabilities.

User-Friendly Interface With Natural Language Search

Modern data discovery platforms prioritize user experience with intuitive, user-friendly interfaces. Features like natural language search allow users to query data using everyday language, making it easier for non-technical users to find what they need. This democratizes access to data across the organization, fostering a culture of data-driven decision-making.

Low Total Cost of Ownership (TCO)

Traditional metadata management solutions often come with a hefty price tag due to high infrastructure costs and ongoing maintenance. In contrast, modern data discovery tools are designed to minimize TCO with automated features, cloud-based deployment, and reduced need for manual intervention. This means more efficient operations and a greater return on investment.

Benefits

By leveraging a comprehensive data discovery solution, manufacturers can achieve several key benefits:

Enhanced Innovation

With quick access to quality data, teams can identify trends and insights that drive product development and process optimization.

Faster Time to Market

Automated implementation and seamless data connectivity reduce the time required to gather and analyze data, enabling faster decision-making.

Improved Operational Efficiency

Advanced metadata management and knowledge graphs help streamline data governance, ensuring that users have access to reliable, high-quality data.

Increased Competitiveness

A user-friendly data marketplace democratizes data access, empowering teams to make data-driven decisions and stay ahead of industry trends.

Cost Savings

With low TCO and reduced dependency on manual processes, manufacturers can maximize their resources and allocate budgets towards strategic initiatives.

Data is more than just a resource—it’s a catalyst for innovation. By embracing advanced metadata management and data discovery solutions, you can find, trust, and access data. This not only accelerates time to market but also drives operational efficiency and boosts competitiveness. With powerful features like API-led automation, a data discovery knowledge graph, and an intuitive data marketplace, you’ll be well-equipped to navigate the challenges of Industry 4.0 and beyond.

Call to Action

Ready to accelerate your innovation journey? Explore how Actian can transform your manufacturing processes and give you a competitive edge.

Learn more about how our advanced data discovery solutions can help you unlock the full potential of your data. Sign up for a live product demo and Q&A

Kasey Nolan

About Kasey Nolan

Kasey Nolan is Solutions Product Marketing Manager at Actian, aligning sales and marketing in IaaS and edge compute technologies. With a decade of experience bridging cloud services and enterprise needs, Kasey drives messaging around core use cases and solutions. She has authored solution briefs and contributed to events focused on cloud transformation. Her Actian blog posts explore how to map customer challenges to product offerings, highlighting real-world deployments. Read her articles for guidance on matching technology to business goals.
Data Governance

Federated Data Governance Explained

Actian Corporation

December 6, 2024

federated data governance blog hero

Summary

This blog post from Actian delves into the concept of federated data governance, highlighting its significance in balancing centralized oversight with decentralized execution. It emphasizes the importance of this model in managing complex data landscapes, especially in large organizations.

Key Takeaways:

  • Hybrid Governance Model: Federated data governance combines centralized policies with decentralized execution, allowing for both standardization and flexibility in data management.
  • Enhanced Collaboration: This approach fosters collaboration across various departments, ensuring that data governance is aligned with business needs while maintaining compliance and security.
  • Scalability: Federated governance models are scalable, making them suitable for organizations of varying sizes and complexities, facilitating efficient data management across diverse systems.

Databases can streamline your business’s processes and make accessing important information easy, fast, and reliable. Organizing the information in those databases can be challenging, especially if you’re trying to migrate information from old systems into newer models. Federated data governance can help you easily organize data without sacrificing the independence necessary for each department to input and structure information in a way they can use.

A federated system is an architectural approach that allows multiple independent databases or data sources to work together as a unified whole while maintaining their autonomy. For end users, this allows access to view and use data from other systems without having to move data to a central repository. Let’s take a closer look.

What is Federated Data Governance?

Strictly speaking, federated data governance is a method of organizing data sources with common standards for items such as security, compliance, and interoperability. This means each data source within an organization share clear and centralized standards or guidelines that govern how the information is managed. Governance standards are usually decided by a data committee or council.

With governance in place, data owners are empowered with the ability to access and share data as well as provide context to data sets that will improve its usability. Individual teams are able to use in a way that works for their needs and often self-0serve access to data without the need for IT resources.

The goal of federated governance is to ensure compliance across data domains without hampering individuals’ ability to create and manage databases in a way that works for them, thereby helping businesses conquer the chaos of data management.

How the Federated Approach to Data Governance Works

Federated governance can be applied to many situations and industries to help streamline processes for both large and small teams. Here are some of the common ways businesses in different sectors can utilize this method of organization: 

  • Healthcare: For healthcare providers with multiple clinics or locations, federated governance can allow each clinic to access key information while still maintaining patient privacy and adhering to HIPAA regulations. Each clinic is able to access and update its individual databases without exposing sensitive data. 
  • Hospitality: Individual hotels or bed and breakfasts can manage their own data according to their needs while still adhering to the parent company’s overarching structure and standards. This helps boost consistency across subsidiaries while allowing individual hotels to do things such as flagging high profile customers.  
  • Finance: Banks and lenders use federated governance to enhance the security of each branch’s and each department’s data while ensuring customer data can be easily shared. This allows the bank to uncover potential opportunities to service customers while also improving customer experience.   

This form of governance can be customized to fit each organization’s unique needs regardless of size.  

Federated vs. Centralized Data Governance: Key Differences

Though both federated and centralized data systems work to help companies manage and organize their information, they do so in different ways. As we mentioned before, federated systems are flexible and allow each team or department to structure their data in the way that makes the most sense for their needs while still adhering to an overarching set of standards. Centralized systems are far more rigid. Only one department oversees the company’s data management, restricting each team’s ability to customize systems.   

The chart below outlines the key differences between federated and centralized systems. 

 

 

Federated Data Systems 

Centralized Data Systems 

Ownership  Each team or department establishes its own processes as long as they follow the company’s standards.  One team organizes and manages the company’s data regardless of department or purpose. 
Flexibility  Gives each team independence in data management.  Restricts each team’s data management efforts. 
Best For  Small businesses with small databases   Large established companies or new companies looking to grow quickly 

 

The Benefits of Using Federated Data Governance in Your Organization

If you’re currently using another form of data governance in your organization, transitioning into a new framework can feel overwhelming. However, switching to a federated data governance system can offer a few unique benefits that are not provided by other methods. This includes but is not limited to the following: 

  • Faster Decisions: When individuals can manage their data based on their needs, they can make decisions and changes faster. This cuts down on missed opportunities and may help businesses save on labor. 
  • Enhanced Compliance: Empowering each team to manage their own data as needed within the approved guidelines and processes increases compliance with company standards. Teams will be less likely to manage data in other places or house sensitive information in locations that aren’t as secure. 
  • Clear Ownership: Each department is responsible for its own data sets. This makes it easier for companies and team members to know who is responsible for maintenance and upkeep and cuts down on stress when trying to identify the best person to fix specific problems. 

How to Implement a Federated Data Model

Implementing a federated data model in your business takes time, but the end result is worth the effort. Here are the steps you need to follow to create a model that works.

1. Assign Roles and Responsibilities

Since you’re giving individual departments or team members the ability to oversee their domain’s data, it’s essential that they understand what’s expected of them. Take the time to assign clear roles and responsibilities to those individuals. You’ll want to outline who owns specific tasks, and the exact responsibilities each person involved has with regard to the data. This will help hold everyone accountable and reduce the risk of tasks falling through the cracks.

2. Define the Scope and Goals of the Data Governance Program

Figure out what you want the data governance program to accomplish. This will help you set your priorities and figure out the structure you need your program to include. Keep in mind that you may want to consider both your short-term and long-term goals when you’re planning your model. By having a clearer vision of where you want to go as your business expands, you’ll be better able to establish a framework that accommodates that growth.

3. Create a Framework

Once you understand your goals, you can create a framework for your federated data model within your existing data mesh by developing a new and more functional framework than you’re currently using. This framework should explain the key policies you need to uphold and how those policies will be followed at every level. Part of this framework may need to include key performance indicators to help you measure success and establish processes to monitor and report your data management efforts.

4. Develop Quality Standards

Think about your primary concerns and what you want your new model to help you improve compared to your old data management system. This can help you establish a benchmark for data quality standards for your new model. For many businesses, this involves setting minimum requirements for data security, privacy, and internal accessibility.

5. Choose a Data Catalog

Your data catalog helps your business actively manage metadata in a single, easy-to-access location. Remember, your metadata makes it easier for your team to search through your datasets and identify the information they need. As you add new information to your model, your metadata grows, so your catalog needs to be able to grow with your company, just like your framework.

Explore your options and choose a data catalog that can integrate with your current systems well without sacrificing security and scalability. Actian Data Intelligence Platform can take the stress out of managing your metadata.

6. Establish Clear Lines of Communication

Each person involved in the data model needs to be able to communicate with each other to make sure everyone is on the same page. As you develop your framework, create a roadmap of communication for your team. Make sure each person knows who the other stakeholders are, how to reach them, and their responsibilities. You may find it beneficial to schedule recurring meetings to keep everyone up to date.

7. Train Your Team

Once the new model is in place, you’ll want to get everyone up to speed quickly. Host training sessions and encourage each domain owner to share their expertise with other owners regularly. Consider hosting webinars, masterclasses, and other training sessions at least once a quarter or as new features are implemented.

Implementing Federated Data Governance Can Simplify Your Business

Federated data governance aims to make managing complex data sets easier for your teams, whether your business is a large multi-branch organization or a small but rapidly growing startup. Once implemented, your team will be better able to update, interpret, and manage data relevant to their needs in a dynamic and agile way.

At Actian, our team wants to take the stress out of implementing new data management models. Join a free demo of our tools and see how easy it can be to take control of your data management processes.

actian avatar logo

About Actian Corporation

Actian empowers enterprises to confidently manage and govern data at scale. Actian data intelligence solutions help streamline complex data environments and accelerate the delivery of AI-ready data. Designed to be flexible, Actian solutions integrate seamlessly and perform reliably across on-premises, cloud, and hybrid environments. Learn more about Actian, the data division of HCLSoftware, at actian.com.
Data Management

From Silos to Synergy: Data Discovery for Manufacturing

Kasey Nolan

December 4, 2024

from silos to synergy blue circle infinity

Introduction to Data Discovery for Manufacturing

Manufacturing organizations often struggle with data silos, where valuable information remains trapped within departmental systems. This hinders the ability to make informed decisions across the entire value chain. Data discovery, facilitated by a data catalog and an enterprise data marketplace, empowers businesses to break down these silos. By providing a centralized and integrated view of all your data assets, these solutions unlock the potential for improved decision-making, increased efficiency, and enhanced competitiveness within the manufacturing sector.

The Problem: Data Silos Impede Visibility

In your organization, each department maintains its own critical datasets– finance compiles detailed financial reports, sales leverages CRM data, marketing analyzes campaign performance, and operations tracks supply chain metrics. But here’s the challenge: how confident are you that you even know what data is available, who owns it, or if it’s quality?

The issue goes beyond traditional data silos. It’s not just that the data is isolated– it’s that your teams are unaware of what data even exists. This lack of visibility creates a blind spot. Without a clear understanding of your company’s data landscape, you face inefficiencies, inconsistent analysis, and missed opportunities. Departments and up duplicating work, using outdated or unreliable data, and making decisions based on incomplete information.

The absence of a unified approach to data discovery and cataloging means that even if the data is technically accessible, it remains hidden in plain sight, trapped in disparate systems without any context or clarity. Without a comprehensive search engine for your data, your organization will struggle to:

  • Identify data sources: You can’t leverage data if you don’t know it exists. Without visibility into all available datasets, valuable information often remains unused, limiting your ability to make fully informed decisions.
  • Access data quality: Even when you find the data, how do you know it’s accurate and up-to-date? Lack of metadata means you can’t evaluate the quality or relevance of the information, leading to analysis based on faulty data.
  • Understand data ownership: when it’s unclear who owns or manages specific datasets, you waste time tracking down information and validating its source. This confusion slows down projects and introduces unnecessary friction. 

The Solution

Now, imagine the transformative potential if your team could search for and discover all available data across your organization as easily as using a search engine. Implementing a robust metadata management strategy—including data lineage, discovery, and cataloging—bridges the gaps between disparate datasets, enabling you to understand what data exists, its quality, and how it can be used. Instead of chasing down reports or sifting through isolated systems, your teams gain an integrated view of your company’s data assets.

  • Data Lineage provides a clear map of how data flows through your systems, from its origin to its current state. It allows you to trace the journey of your data, ensuring you know where it came from, how it’s been transformed, and if it can be trusted. This transparency is crucial for verifying data quality and making accurate, data-driven decisions.
  • Data Discovery enables teams to quickly search through your company’s data landscape, finding relevant datasets without needing to know the specific source system. It’s like having a powerful search tool that surfaces all available data, complete with context about its quality and ownership, helping your team unlock valuable insights faster.
  • A Comprehensive Data Catalog serves as a central hub for all your metadata, documenting information about the datasets, their context, quality, and relationships. It acts as a single source of truth, making it easy for any team member to understand what data is available, who owns it, and how it can be used effectively.

Revolutionizing Your Operations With Metadata Management

This approach can transform the way each department operates, fostering a culture of informed decision-making and reducing inefficiencies:

  • Finance gains immediate visibility into relevant sales data, customer demand forecasts, and historical trends, allowing for more accurate budgeting and financial planning. With data lineage, your finance team can verify the source and integrity of financial metrics, ensuring compliance and minimizing risks.
  • Sales can easily search for and access up-to-date product data, customer insights, and market analysis, all without needing to navigate complex systems. A comprehensive data catalog simplifies the process of finding the most relevant datasets, enabling your sales team to tailor their pitches and close deals faster.
  • Marketing benefits from an integrated view of customer behavior, campaign performance, and product success. Using data discovery, your marketing team can identify the most impactful campaigns and refine strategies based on real-time feedback, driving greater engagement and ROI.
  • Supply Chain Leaders can trace inventory data back to its origin, gaining full visibility into shipments, supplier performance, and potential disruptions. With data lineage, they understand the data’s history and quality, allowing for proactive adjustments and optimized procurement.
  • Manufacturing Managers have access to a clear, unified view of production data, demand forecasts, and operational metrics. The data catalog offers a streamlined way to integrate insights from across the company, enabling better decision-making in scheduling, resource allocation, and quality management.
  • Operations gains a comprehensive understanding of the entire production workflow, from raw materials to delivery. Data discovery and lineage provide the necessary context for making quick adjustments, ensuring seamless production and minimizing delays.

This strategy isn’t about collecting more data—it’s about creating a clearer, more reliable picture of your entire business. By investing in a data catalog, you turn fragmented insights into a cohesive, navigable map that guides your strategic decisions with clarity and confidence. It’s the difference between flying blind and having a comprehensive navigation system that leads you directly to success.

The Benefits: From Fragmentation to Unified Insight

When you prioritize data intelligence with a catalog as a cornerstone, your organization gains access to a powerful suite of benefits:

  1. Enhanced Decision-Making: With a unified view of all data sources, your team can make well-informed decisions based on real-time insights. Data lineage allows you to trace back the origin of key metrics, ensuring the accuracy and reliability of your analysis.
  2. Improved Collaboration Across Teams: With centralized metadata and clear data relationships, every department has access to the same information, reducing silos and fostering a culture of collaboration.
  3. Greater Efficiency and Reduced Redundancies: By eliminating duplicate efforts and streamlining data access, your teams can focus on strategic initiatives rather than time-consuming data searches.
  4. Proactive Risk Management: Full visibility into data flow and origins enables you to identify potential issues before they escalate, minimizing disruptions and maintaining smooth operations.
  5. Increased Compliance and Data Governance: Data lineage provides a transparent trail for auditing purposes, ensuring your organization meets regulatory requirements and maintains data integrity.

Conclusion

Data silos are more than just an operational inconvenience—they are a barrier to your company’s growth and innovation. By embracing data cataloging, lineage, and governance, you empower your teams to collaborate seamlessly, leverage accurate insights, and make strategic decisions with confidence. It is time to break down the barriers, integrate your metadata, and unlock the full potential of your organization’s data.

Call to Action

Are you ready to eliminate data silos and gain a unified view of your operations? Discover the power of metadata management with our comprehensive platform. Visit our website today to learn more and sign up for a live product demo and Q&A.

Kasey Nolan

About Kasey Nolan

Kasey Nolan is Solutions Product Marketing Manager at Actian, aligning sales and marketing in IaaS and edge compute technologies. With a decade of experience bridging cloud services and enterprise needs, Kasey drives messaging around core use cases and solutions. She has authored solution briefs and contributed to events focused on cloud transformation. Her Actian blog posts explore how to map customer challenges to product offerings, highlighting real-world deployments. Read her articles for guidance on matching technology to business goals.