Exploring Data and Analytics Services

Objective

After completing this lesson, you will be able to describe different SAP BTP tools for data and analytics.

Introduction to the Lesson: Exploring Data and Analytics

This lesson first provides a discussion of the terms data fabric, data mesh and data product. Building off of those terms it then provides a comprehensive overview of several SAP tools that can be used for data management and analytics, specifically SAP HANA Cloud, SAP Datasphere, SAP Master Data Governance and SAP Analytics Cloud. How these tools support the integration, management, and analysis of data to gain valuable business insights and make data-driven decisions is explored.

This lesson contains the following topics:

  • Data Terminology
  • Data and Analytics Tools at a glance
  • SAP Datasphere Overview
  • SAP Analytics Cloud Overview
  • SAP Datasphere: A Deeper Dive
  • SAP Analytics Cloud: A Deeper Dive

Data Terminology

Data Terminology

Data Terminology

Organizations need data and analytics to make better, data-driven decisions, improve operational efficiency, enhance customer understanding, and gain a competitive edge. The data and analytics services in SAP BTP will help in identifying trends, optimizing processes, predicting future outcomes, and mitigating risks, ultimately leading to increased attainment of organizational goals related to areas such as revenue and profitability, customer satisfaction and sustainability.

There are several terms that have come into usage over the years when talking about data. Let's briefly define them. The main terms are:

  • Data Warehouse
  • Data Lake
  • Data Fabric
  • Business Data Fabric
  • Data Mesh
  • Data Product

Data Warehouse

Analogy: A well-organized library: Books (data) are carefully categorized, cataloged, and placed on specific shelves (tables/schemas) so you can easily find exactly what you need for a research paper (a business report).

Definition:This is one of the first terms in the list and has been used for many decades. Data warehouses have been around for many years and businesses utilize them for reporting, analysis and decision making by taking advantage of its repository features. Data warehouses integrate and store structured data (i.e., database tables, spreadsheets and the like) which typically comes from all corners of the organization and from many different source systems. Whereas data warehouses typically represent a repository for the organization as a whole, data marts are typically smaller and represent the needs of a specific business department or function such as finance or marketing. Otherwise the two are the same.

Primary Use Case: Business Intelligence, corporate reporting, and performance dashboards.

Data Lake

Analogy: A large, natural reservoir of water: Water (data) flows in from many rivers (sources) in its raw form. You can later take this water and filter it for drinking (reporting), use it for farming (machine learning), or analyze its composition (data science).

Definition: A newer term data lakes compliment data warehouses by being a repository for not only structured data but also semi-structured and unstructured data. By "semi-structured" and "unstructured" we mean data sources such as social media, mobile apps, intelligent sensors and even websites.

Primary Use Case: Big data analytics, machine learning model training, and data science exploration.

Data Warehouse versus Data Lake

It is important to understand that data warehouses and data lakes are complimentary and organizations do not have to choose between them. Both are an important part of a comprehensive data and analytics strategy. A few interesting technical differences between them include:

  • Data is typically stored in a data warehouses via an "Extract, Transformation and Loading (ETL)" approach. The relevant data is "extracted" from a source system, "transformed" (i.e., cleaned and converted) and then "loaded" into the data warehouse. With data lakes the raw data is typically stored as files or some sort of object storage with a unique key enabling it to be "found" when queried.
  • To access data from a data warehouse typically specialized tools optimized for that data warehouse are used. For example when using SAP Datasphere (covered later in this lesson) is used as a data warehouse SAP Analytics Cloud (also covered later in this lesson) can be used for analytical reporting. While the same specialized tools could exist for a data lake there are several open source frameworks such as Apache Hadoop and Apache Spark that are available for usage.

Data Fabric

Analogy: A smart universal adapter or a city's electrical grid: You don't care which power plant generated the electricity; you just plug your device into the wall outlet (the fabric's access point) and get standardized power. The grid intelligently manages and routes the electricity for you.

Definition: A Data Fabric is an architectural approach that acts as an intelligent, integrated layer over disparate data sources. It doesn't necessarily move the data but provides a unified way to access, manage, and govern it, no matter where it resides (in a data warehouse, data lake, or operational database).

Primary Use Case: Providing a unified, real-time view of all enterprise data to data consumers, breaking down data silos without creating another massive repository.

A data fabric architecture links, manages, and controls data across systems and applications, providing a centralized and uniform view of the data. This applies to both the teams and the systems - at every point in your organization. The most important layers of a data fabric architecture include:

  • Data Connection Layer: You can think of data connectors as bridges that connect different systems where the data is stored (for example, databases, applications, sensors) at a central point. In this way, all data sets can be analyzed from a central point.

  • Data Management Layer: The aim here is to ensure that the data is structured, secure, and of high quality. This includes tasks such as data integration (merging data from different sources), data governance (defining rules for the use and management of data), and data security (protecting sensitive data from unauthorized access).

  • Data Modeling & Semantic Layer: This step helps you to understand the data by creating a common language for the data from the different systems. This involves creating a model and a semantic layer. While the model describes the data, the semantic layer represents the agreed language for the "story" of the data.

  • Data Processing & Analytics Layer: This is where the data is processed and analyzed to gain important insights. This includes tasks such as data warehousing (storage of large amounts of data), data streaming (continuous processing during creation), and data visualization (easy-to-understand presentation of the data).

  • Automated Data Management Layer: While data analysis is at the center of automation in various business areas, in the architecture context it is more about the efficient and consistent management of data. This includes the automation of tasks such as data integration, data governance, and data security. Through automation, companies can reduce errors, save time, and improve data quality.

Putting the "Business" In Data Fabric

Analogy: A GPS for your business data: A regular map (Data Fabric) shows you all the roads. A GPS (Business Data Fabric) understands your goal (e.g., "get to the nearest coffee shop"), overlays points of interest (business context), and gives you turn-by-turn directions (actionable insights).

A Business Data Fabric is an evolution of the Data Fabric concept that is specifically designed around business context and meaning. It goes beyond technical connectivity to weave a semantic layer over the data, making it understandable and usable for business users, not just technical staff.

Primary Use Case: Enabling true self-service analytics for business departments, ensuring that data is not just accessible but also meaningful and trustworthy from a business perspective.

Data Mesh

Analogy: A marketplace of data products: Instead of one giant supermarket (central data team), you have a market with different stalls (business domains like "Marketing," "Sales," "Logistics"). Each stall owner is an expert in their goods (their data) and is responsible for packaging them as high-quality, ready-to-use products for others to consume.

Definition: A data mesh is similar to a data fabric in some ways. The main difference is that whereas a data fabric tends to be centralized for the entire organization bringing together data across the entire organization, the focus for data meshes is relatively speaking more decentralized and either more department or more functional specific (similar as to data marts discussed earlier in the lesson). Data meshes promote the concept of "data autonomy" where different teams can claim and manage their own data and services making decisions independently based on that data and their needs.

Primary Use Case: Scaling data management in large, complex organizations to avoid the bottlenecks of a centralized data team and increase data agility and quality.

Data Product

From Raw Data to Business Value

Data product overview.

Challenges With Traditional Data Access

Challenges sometimes exist with having access to data for reporting and decision making purposes:

  • Complexity: Database table models are sometimes highly normalized with non intuitive table names. Business context can sometimes be buried in application logic.
  • Silos: Data can be locked within separate areas (i.e., SD, MM, FI, CO). This can create challenges when trying to achieve an end-to-end view (e.g., "Order-to-Cash").
  • Dependency on Experts: Business users often cannot access data directly. They must file a request with IT or a specialized team, sometimes leading to long delays which can impact fast decision making.
  • Inconsistent Extracts: Different teams sometimes extract the same data (e.g., customer data) in slightly different ways, which can lead to multiple "versions of the truth".

How the Data Product Concept Solves These Challenges

Think of a physical product you buy. It’s not raw materials; it's a finished good that is reliable, documented (user manual), and easy to use. A data product is the same. It is a self-contained, reusable, and trustworthy data asset designed for a specific business purpose.

A Data Product is not just a database table or a raw data extract. It is a composite of:

  • Data: The core information itself (e.g., sales order details).
  • Code: The logic to ingest, clean, transform, and serve the data.
  • Infrastructure: The underlying platform to store and compute the data.
  • Metadata & Governance: The "labeling" that makes it understandable, trustworthy, and secure. This includes lineage, ownership, quality metrics, and access policies.

Applying the data product concept directly addresses these pain points. It's a move fromextracting data from SAP products to serving curated data products from the entire SAP ecosystem.

Note the differences in the table below where on the left are traditional aspects and on the right are data product advantages.

"Can you build me a report of sales orders?""The 'Validated Sales Orders' Data Product is available for consumption."
One-off, custom ABAP extractors or ETL jobs.Reusable, versioned, and supported data asset with an API.
IT owns the extraction pipeline.The Sales domain team owns the data product's quality and definition.
Data is raw and needs interpretation.Data is cleaned, enriched, and has business-friendly names.
Slow, project-based delivery.Agile, self-service access for consumers.

Outcome

The aim is now to create concrete solution architectures from the technical and business use cases that can then be implemented. In the following, a closer look at selected use cases is taken.

SAP BTP Solution Diagram

In the graphic we see a data fabric architecture outlined. SAP Datasphere provides the infrastructure. HR, finance and workflow data is provided from source systems (through an ETL process). Utilization of the data happens through two apps, a workflow service app and a headcount request app running in Node.js and SAP Analytics Cloud respectively.

Data and Analytics Tools at a Glance

Data and analytics capabilities.

The data and analytics category in SAP BTP includes the following tools:

  • SAP HANA Cloud
  • SAP Datasphere
  • SAP Master Data Governance
  • SAP Analytics Cloud

SAP HANA Cloud: At a Glance

Sample overview

SAP HANA Cloud is a cloud-native database as a service (DBaaS) designed to power modern applications and analytics across enterprise data. It combines the performance and reliability of SAP HANA with the flexibility and scalability of the cloud, enabling businesses to build intelligent, data-driven solutions.

SAP HANA Cloud supports multi-model engines, allowing it to handle diverse workloads such as relational, document store, geospatial, knowledge graph, vector, and time series data. This makes it suitable for both analytical (OLAP) and transactional (OLTP) use cases. Developers can leverage its capabilities to create applications that integrate generative AI, are context-aware, and securely connect to business-critical data.

One of the key advantages of SAP HANA Cloud is its fully managed nature, which eliminates the need for managing hardware, operating systems, backups, and other maintenance tasks. It offers elastic scalability, enabling businesses to adjust computing and storage resources based on demand, while ensuring high availability and compliance.

SAP HANA Cloud integrates seamlessly with SAP and third-party data sources using technologies like Smart Data Access (SDA) and Smart Data Integration (SDI), as well as protocols like ODBC, REST, and JDBC.

SAP Datasphere: At a Glance

Example for SAP Datasphere

SAP Datasphere is a unified data service that provides a business data fabric for accessing, integrating, modeling, and distributing mission-critical business data across SAP and non-SAP landscapes. It enables data professionals to deliver seamless and scalable access to data, preserving business context and logic. It's the evolution of SAP Data Warehouse Cloud, incorporating new features like data integration, cataloging, and semantic modeling. A few key features of SAP Datasphere are:

  • Unified Data Experience - SAP Datasphere provides a single point of access to data from various sources, both SAP and non-SAP, cloud and on-premise
  • Data Integration and Modeling - SAP Datasphere offers tools for integrating, harmonizing, and modeling data from diverse sources
  • Data Virtualization - SAP Datasphere allows access to data without physically replicating it, ensuring consistency and reducing data duplication.

SAP Master Data Management: At a Glance

Example of a master data consolidation process.

Master Data Management (MDM) is a concept that enables organizations to manage and govern critical business data, like customer, product, and vendor information, ensuring consistency and accuracy across various systems and processes. It aims to create a single, trusted source of truth for master data, improving data quality, reducing errors, and enhancing operational efficiency. SAP BTP has two complimentary services dealing with MDM:

  • Master Data Integration
  • SAP Master Data Governance, cloud edition

SAP Master Data Integration

SAP Master Data Integration

The SAP Master Data Integration service helps business applications to arrive at a consistent view on master data. All SAP applications produce and/or consume master data. Since business processes are typically not local to a single application, master data needs to be replicated between these applications.

Customer landscapes not utilizing master data integration, are currently characterized by many point-to-point integrations between multiple different applications. These point-to-point integrations lead to an inherently complex landscape that is hard to maintain as well as to extend with new applications. As a natural consequence this also leads to data inconsistencies between applications.

SAP Master Data Integration provides a solution to reduce this complexity by migrating from point-to-point integrations towards a centralized, hub & spoke based approach. Instead of creating multiple custom integrations per application, applications connect to SAP Master Data Integration centrally for any kind of master data replication. These integrations are based on the SAP One Domain Model, which is the common model for business objects across different SAP applications. SAP One Domain Model enables applications to "speak the same language" during data exchange, while keeping their local representation untouched. This removes the complexity from mapping models between various applications, towards a situation in which each application just has a native single mapping towards the SAP One Domain Model effectively eliminating manual data mapping and enabling the delivery of ready-to-deploy integrations. This allows for easier extensibility and reduces overall complexity and maintenance efforts. If desired, organizations can build custom applications on top of SAP Master Data Integration as its data is persisted in SAP HANA Cloud.

Since SAP Master Data Integration persists the data, it is possible to add a new application to a customer landscape without any large analysis of available master data in the landscape. Initial loads towards these solutions are supported and the application will eventually arrive at a consistent state according to the customers filter settings for this application. These filters allow the customer to specify exactly which master data objects and which fields of these objects are supposed to be replicated to an individual downstream application.

Many SAP SaaS products integrate with SAP Master Data Integration out of the box. For example SAP Cloud ERP uses the master data objects "Bank" and "Business Partner" while SAP SuccessFactors uses "Job Classification" and "Workforce Person". When out of the box integration is not available, organizations are recommended to use SAP Integration Suite to integrate applications with SAP Master Data Integration.

SAP Master Data Governance, cloud edition

SAP Master Data Governance, cloud edition

As just mentioned applications can be built on top of SAP Master Data Integration. SAP Master Data Governance, cloud edition (MDG) is an SAP BTP service which offers the maintenance of high level master data quality, specifically offering capabilities to manage core attributes of business partners and to evaluate their quality. Its capabilities are organized around three pillars:

  • Central Governance - The creation and maintenance of business partners incl. contact persons and relationships in a workflow-driven process including the ability to process multiple BPs simultaneously, and to use reference data from external data providers.
  • Consolidation - The creation of a single view for accurate analytics and operational insight continuously or on request. Business partners can be loaded from files or received via APIs. Duplicates can be detected and merged into a best record.
  • Data Quality Management - The definition, enforcement, monitoring, and improvement of the quality of business partner master data. As part of this capability validation rules can be managed collaboratively and used in data quality evaluations, consolidations, and central governance.

SAP Analytics Cloud: At a Glance

SAP Analytics Cloud

SAP Analytics Cloud (SAC) is a unified cloud analytics solution. It's designed to provide business intelligence (BI), predictive analytics, and planning capabilities all within a single, cloud-based platform. Think of it as a one-stop shop for understanding your data, predicting future outcomes, and planning for the future. It's basic features are:

  • Unified Platform
  • Real-Time Insights
  • Improved Collaboration
  • Increased Efficiency
  • Reduced Costs
  • Data-Driven Decision Making

SAP Datasphere: A Deeper Dive

SAP Datasphere is a unified service for data integration, cataloging, semantic modeling, data warehousing, and virtualization of workloads for SAP and non-SAP data. With SAP Datasphere, data experts can easily distribute business-critical data across their organization's entire data landscape while maintaining business context and logic. SAP Datasphere is the next generation of the SAP Data Warehouse Cloud.

Overview of SAP Datasphere.

Capabilities

A few of the capabilities of SAP Datasphere include:

  • Access Authoritative Data
  • Enrichment of all data projects
  • Simplify your data landscape

Access Authoritative Data

SAP Datasphere enables direct access to business-critical data and metadata - and the critical relationships between them - from SAP applications. With the new SAP Datasphere Analytic Model, data experts can easily solve complex requirements for semantic modeling of business data. The automatic reuse of semantic definitions and associations from SAP applications shortens time-to-value and simplifies the steps that data experts must take across the organization to deliver business-critical data. SAP Datasphere Catalog also makes it easier to find, manage, and control data automatically. In addition, cataloging of data stored in SAP Analytics Cloud and SAP Datasphere is offered at no additional cost.

Sample View

Enrichment of All Data Projects

With SAP Datasphere, organizations can bring all their data together and perform transformations that ensure all stakeholders are always working with the same data. SAP Datasphere supports both data federation and the new fast data replication, providing powerful options for combining data sources and harmonizing heterogeneous data in a semantic business model for customers' different data landscapes. With the SAP Datasphere Marketplace, access to industry data is now even easier: more than 3,000 curated data sets can be purchased at the touch of a button to quickly enrich data projects with up-to-date, trustworthy data. SAP announced partnerships with industry-leading open data partners to tightly integrate their data and AI platforms with SAP Datasphere, enabling customers to combine SAP data with other data sources like never before.

Data marketplace

Simplify Your Data Landscape

SAP Datasphere runs in any cloud and hybrid environment. Thanks to this flexibility, data experts can securely access all their data, no matter where it is located. In addition, customers can maximize their investment in on-premise solutions such as SAP Business Warehouse (SAP BW). SAP Datasphere, BW bridge, enables the reuse of SAP BW models and connectors and offers customers of SAP NetWeaver BW and SAP BW/4HANA a transition path to the public cloud. Once data has been integrated, harmonized, and enhanced in SAP Datasphere, data professionals can create a "virtual workspace" with SAP Datasphere Spaces - a capability to organize, group, and manage data and control the level of visibility and collaboration. These virtual workspaces can then be made available to data consumers that rely on federated data, such as marketing, sales, supply chain or partner data, and AI teams.

Space management

Building a Powerful Open Data Ecosystem Around SAP Datasphere

Business Data Fabric

SAP Datasphere and its open data ecosystem form the technological basis for a business data fabric. As mentioned earlier in the lesson a business data fabric focuses on providing an integrated, semantically rich data layer on top of the underlying data landscapes to enable seamless and scalable access to data without duplication.

For a more in depth discussion of SAP Datasphere please see:

SAP Analytics Cloud: A Deeper Dive

SAP Analytics Cloud is an all-in-one cloud product offered as software as a service (SaaS) for business intelligence (BI), planning, and predictive analytics. It integrates trustworthy AI for automated reporting and business planning and it includes numerous functions such as data visualization and modeling. Built natively on SAP Business Technology Platform (SAP BTP), it provides a unified and secure public cloud experience to maximize data-driven decision-making.

Major Capabilities

Stories

A story is a presentation-style document that uses charts, visualizations, text, and images to describe data.

Sample Story

Analytic Applications

With SAP Analytics Cloud, analytics designer, you can create analytic applications for data analysis and data planning. In analytic applications, you can configure the behavior of the UI elements with a set of specific script API events and specify which actions must take place when events are triggered.

Create an analytic application

With analytic applications, you can:

  • Build a data-driven analytic application. You can visualize your data, pick a model, add widgets, and add custom logic to your analytic application.
  • Create and use planning applications that support both manual and automated data entry and changes to data.

Data Analyzer

In SAP Analytics Cloud, you can start ad-hoc analyses with a data analyzer - a predefined ready-to-run service for multidimensional, pivot styled ad-hoc analysis for SAP BW Live queries, SAP HANA Live views, SAP Datasphere models, and SAP Analytics Cloud models. SAP BW queries and SAP HANA views can be accessed directly and no additional model needs to be created. The data can also be displayed in charts.

Sample data report.

Models

In SAP Analytics Cloud, a model is a representation of large amounts of business data from source systems. It defines measures and dimensions that are used to build visualizations, filters, and calculations in stories.

Formulas and Calculations

In SAP Analytics Cloud, you can create formulas to perform calculations on values and members of the account dimensions. You can also use predefined formulas, functions, conditions, and operators to build up a formula in Modeler.

Planning

SAP Analytics Cloud has many planning tools for collaboratively creating and updating model data to carry out planning tasks. At its simplest, planning involves typing values directly into table cells in planning models.

For a more in depth discussion of SAP Analytics Cloud please see: Exploring SAP Analytics Cloud

Summary

In summary think of the data & analytics dimension of SAP BTP as the central nervous system for an enterprise's data. Its primary goal is to empower organizations to transform raw data from any source—SAP or non-SAP, cloud or on-premise—into actionable, real-time insights and intelligent business decisions.