+

Data Feast Weekly

Get weekly insights on modern data delivered to your inbox, straight from our hand-picked curations!

Back

Could Data Products Save (Real) Money?

7 min
|
Value Evaluation Model, Calculation Functions, Product Remediation, and more!
Jul 25, 2024
Data Strategy
,
  and
,
  and
Originally published on
Modern Data 101 Newsletter,
the following is a revised edition.

It is important to remember that managing data as a product guarantees interoperability, facilitates data and code reuse, and promotes an "as a product" approach, which includes considering the profitability of our digital product.

The steps we will follow are:

  1. Pre-requisites: Data Taxonomy and Selection of Drivers
  2. Tuning: Classification by Functional Domains and Calculation Function
  3. Monitoring & Remediation: Analysis of Behaviors and Trends, and Applicable Actions

While the 'data as a product' paradigm might theoretically assign absolute economic values to calculate the profitability of our digital product, this is quite complex (if not unachievable) because it requires complete synergies among enterprise architecture layers. In the article, I will explore how this complexity may not necessarily hinder the initial purpose.

Overview of Distilling the Profitability of Data Products | Image Source: Author Insights + MD101 Archives

I. Pre-requisites


We will analyze all preparatory aspects, whether technological or methodological.

Data Taxonomy

Assuming a basic level of observability is in place (both for determining costs and for usage), the real differentiating requirement is methodological. It is crucial to have a domain-driven organization (not necessarily a mesh) and a taxonomy that can assign meaning to the various data products. This taxonomy should also be distributed across different data products (data contracts) with potential impacts on the data governance model (data product owner).

In this sense, it is strongly recommended to have a federated modeling service to manage and guarantee this within the data architecture (regardless of whether it is distributed or not).

This is necessary because we will use the single element of the taxonomy, i.e., the entity data (and not the single data product), as the calculation element.

📝 Related Reads
- Domain-Driven Data Product Models + Metrics & Entities as First-Class Citizens in a Data Ecosystem

- Taxonomy (Classifying Data for Consistency and Clarity)

Selection of Drivers

Secondly, it is necessary to select an evaluation model, for example, from this work:

Evaluating Data Benefit Index | Image Source: Author

From which we can consider some drivers:

  • Effort: Implementation cost, execution cost of pipelines & structures, the hourly cost of involved figures (data steward, data product owner), number of data products/output ports available;
  • Consumption: increased cloud consumption due to accesses, number of applications, number of accesses to data products, and related number of users;
  • Business Value: Categorization of the data entity and categorization of the applications using it.

Obviously, each business context might select different drivers, but these seem sufficiently agnostic to be considered universal.

It is important to emphasize that not all drivers are economically quantifiable, so it will be necessary to reason in terms of function points or correlations.

Economical and Non-Economical Aspects of Standard Drivers | Image Source: Author

II. Classification of Domains


At this point, two actions are necessary:

Identification of Domains

This action is of fundamental importance because it will serve to identify the outliers (see below). In theory, the domains should come either from data governance or from the enterprise architecture group as instantiations of business processes.

Whichever choice is applied, it is important to reach the point where all data products have at least one business entity and are, therefore, allocable in a functional domain (for consumption data products, a series of scenarios dependent on the applied data governance model opens up).

Calculation Function

This activity is undoubtedly the most evident. After several months of simulations on real data, I concluded that there is no ideal mathematical formula but that we need to work in an incremental approach until we are able to achieve a distribution that should look like the following diagram:

A real distribution based on real-world data (modified to grant anonymity) | Image Source: Author

Some verification thumb rules adopted are:

  • High-value entities should be records or related to the core business
  • All domain entities should be clusterable (at least 85% of the distribution)

At this point, you should be able to obtain a clustering like follows

Distribution of Domains on the Graph | Image Source: Author

III. Monitoring and Remediation


To explain the benefits of the approaches, I provide some examples from the dataset I am using:

Image Source: Author

Where Avg Ratio is computed as the average ratio of (Consumption + Value) : (Effort) axis (similar to Data Benefit Index). This makes sense, in my opinion, that Reference data is usually more valuable than IoT data.

Monitoring & Remediation - First Run

The analyses that can be performed are applicable both during the setup phase and recurring (I will apply them only to the "Reference" domain).

Analysis of Unused Products

140 out of 230 products initially appeared unused (Consumption + Value equal to 0). Several causes for this phenomenon exist:

  • Ghost IT: The product is used but not tracked;
  • Enabling Entity: The entity is not directly used but serves other products, either as a technical object or as a model hierarchy root;
  • Outlier: Entities that are out of the confidence range. In our case, there were specifically two entities wrongly attributed to the domain that had anomalous behaviors.
  • Over-engineered Data Product(s): each product should have a shaped market and a consolidated user base. In data management, we could have multiple products that are over-engineered, and that could be shrunk/merged with others.

Analysis of Entities on Products

An analysis of product/entity assignments reveals that several entities are duplicated without a real reason other than the sedimentation of application layers. For these integrations, it is possible to consider the dismissal of duplicates or their integration into the correct entity (of the reference domain).

Products Managed by Non-Master Domains

A very common phenomenon (but not relevant for a reference domain), in this case, it is also possible to consider reassignments and dismissals.

Applying these normalization actions leads to the following changes:

Image Source: Author

In detail, the benefits are due to the following actions:

  • #Entity: Entities are rationalized both through the dismissal of irrelevant ones and through consolidation into more relevant entities;
  • #Product: The reduction is mainly due to the merging of derived products into masters and a rationalization of output ports;
  • %Usage: Obviously increases to the maximum as unused objects are dismissed. It is important to note that many technical and service objects are not dismissed but "incorporated" into the relevant business products (thereby also contributing to the increase in unit cost);
  • Total Cost: Obviously reduced due mainly to the dismissal of duplicate and unusable structures;
  • Product Cost: Tends to increase both due to the added cost of enabling structures and the unification of multiple products.

Monitoring and Remediation Post Setup and Tuning

Following the initial run and the setup and tuning activities, it is necessary to consider how products and entities move within the diagram over time. Without proactive management, each entity could experience two seemingly opposite phenomena:

  • Enablement of New Accesses/Users: This increases both its value and partially its costs.
  • Increase in Operating and Management Costs: Due to growing data volume, tickets, and versioning/ports.

Both phenomena have a net result that shifts our product or data entity along the effort axis with a proportional decrease in generated value until it reaches the non-exercisable threshold.

An entity tends to increase its- cost and decrease its proportional value | Image Source: Author

This evolution could lead our entity outside the domain clustering, thus affecting it.

An entity that used to be a top performer could worsen its ranking over time.

This clustering approach allows evaluating not only the absolute worst-performing products but also enables the analysis of point decay (e.g., percentage) according to calculation functions that can be defined as needed (and which we won't elaborate on here). Share

Two Categories of Corrective Solutions


  1. Actions Aimed at Improving Relevance
    Such as creating new output ports or unifying multiple products (which, although increasing unit cost, reduces overall cost in the larger economy).
  2. Optimization Actions
    These can include decommissioning (e.g., archiving historical data), re-architecting or decommissioning unnecessary functionalities (e.g., input and output ports or components of data contracts), or reducing SLAs.

These two contributions have a positive vector summation that improves the entity's performance and, thus, its intrinsic profitability.

Image Source: Author

📝 Related Reads
Versioning, Cataloging, and Decommissioning Data Products

Conclusion and Next Steps


The approach is certainly enabling as it allows for rationalization of the information assets both for the quantifiable component and for the intangible one (product quality, clear ownership).

However, it should be evaluated how to assign a business benefit for each initiative (theorized in enterprise architecture frameworks) and how to extend the model to products for which improvement actions cannot be applied (external products or those managed by regulatory constraints).

Suggested Further Monitoring Actions

  1. Periodic Review of Data Taxonomy and Driver Selection: Regularly update the taxonomy and evaluation drivers to reflect changes in business processes and data usage patterns.
    More on Drivers/Metrics & Taxonomy/Semantics
  2. Stakeholder Feedback Loop: Create a feedback loop with key stakeholders, including data stewards and product owners, to gather insights and make necessary adjustments to the data product management strategy.
    More on Feedback Loops
  3. Data Democratization and Integration with a Marketplace: Facilitate wider access to data by democratizing it and integrating it with a data marketplace. This allows for better and more fitting utilization of data products, encourages innovation, and provides opportunities for external users to leverage the data, thus potentially creating new revenue streams.
    More on Marketplaces (Data Products)

Thanks for reading Modern Data 101! Subscribe for free to receive new posts and support our work.


🔏 Author Connect


Find me on LinkedIn