Skip to main content Link Menu Expand (external link) Document Search Copy Copied

Task 02: Ingest SAP data

Introduction

Ingesting SAP data into Microsoft Fabric involves configuring both cloud and on-premises integration components, preparing datasets, running pipelines, and validating the results across the Silver and Gold layers of a medallion architecture. This process brings together Azure Data Factory, the Business Process Solutions (BPS) workload, SAP source systems, and Power BI to provide a complete end-to-end data ingestion and analytics experience.

Description

In this task, you’ll perform the full workflow required to extract and process SAP data in Microsoft Fabric. You’ll configure the SAP source system in Azure Data Factory, validate the self-hosted integration runtime, set up and activate a dataset, run pipelines that extract metadata and data, and verify that the Silver and Gold lakehouses are populated. You will then deploy Power BI reports, configure a semantic model connection for refreshes, and import required lakehouse views to support downstream insights.

Success criteria

  • An SAP source system has been configured in Azure Data Factory.
  • A self-hosted integration runtime is running and connected.
  • A dataset has been configured and activated using an SAP template.
  • Pipelines have been run to extract metadata and data into the Silver lakehouse.
  • The Silver lakehouse contains tables populated from SAP.
  • Data has been processed into the Gold lakehouse.
  • Power BI reports have been deployed using the imported dataset.
  • A Fabric connection has been configured to enable semantic model refresh.
  • Required lakehouse views have been created and imported successfully.

Duration

Estimated time to complete this exercise: 20 minutes

Key steps

01: Configure an SAP source for Azure Data Factory

  1. Return to the web page that displays Fabric or open a new browser tab and go to app.powerbi.com.

  2. In the left pane, select Workspaces and then select WS57383073.

    gl88nljo.jpg

  3. Select the BPS57383073 BPS resource.

    u46ikubi.jpg

  4. On the BPS57383073 page, on the command bar, select Configure source system.

    8qlcz1be.jpg

  5. In the Configure source system pane, select + New source system.

    mfkcz3mv.jpg

  6. Configure the fields in the New source system page by using the values in the following table. Leave all other settings at the default values.

    Setting Value
    Source type SAP
    Source system name AdfSource57383073
    System type SAP S/4HANA
    Source system version SAP S/4HANA 2023
    Connection type Azure Data Factory
    Azure Subscription ID insert your cloud subscription
    Location insert your location
    Resource group insert your resource group

    y3nhtjks.jpg

  7. Continue configuring the fields in the New source system page by using the values in the following table. Leave all other settings at the default values.

    Setting Value
    Hostname 10.15.0.6
    SAP Instance number 01
    SAP Client ID 400
    Username DATAEX
    Password Welcome1
    Service Principal - Name BPS-SAPapp57383073
    Service Principal - Object ID @lab.Variable(ObjectID)
    Service Principal - Client ID @lab.Variable(ClientID)
    Service Principal - Secret @lab.Variable(Secret)

    Here you’d use the ID’s saved in your notepad file.

  8. Select Create. The Configure source system pane displays the status for the deployment status.

    In the Configure source system dialog, on the command bar, select Refresh periodically to monitor progress.

    It can take 5-10 minutes for the process to complete.

    q5czhdti.jpg

  9. Close the Configure source system pane.

Task 02: Configure a self-hosted integration runtime

  1. Open a new browser tab and go to https://adf.azure.com/.

  2. On the Azure Data Factory page, in the Name field, select the available value and then select Continue.

    In the lab, this resource was created for you when you configured the Azure Data Factory source in the previous task. If you’re using your own tenant, you’ll have to set up this yourself.

    ns4x8ivj.jpg

  3. In the left pane, select Manage (the toolbox icon). Then, in the Connections section, select Integration runtimes.

    17azlb1c.jpg

  4. Locate the resource of type Self-Hosted and verify that the status is Running.

    9m6cmovj.jpg

    The Running status indicates that the integration runtime is connected to and communicating with the SAP system.

{: .warning } This was set up for you in the hands on lab, if you wish to use this, you have to set it up yourself.

  1. Leave the Azure Data Factory page open. You will use the page again later in the workshop.

03: Configure a dataset

  1. Return to the web page that displays Fabric or open a new browser tab and go to app.powerbi.com.

  2. In the left pane, select Workspaces and then select WS57383073.

    gl88nljo.jpg

  3. Select the BPS57383073 BPS resource.

    u46ikubi.jpg

  4. On the BPS57383073 page, on the command bar, select Set up your datasets.

    v5spi4rn.jpg

  5. You may see the Additional authentication or authorization required dialog. If you do, select Continue and sign in by using your tenant’s credentials.

  6. On the Datasets page, on the command bar, select Import from template.

    68jwsemc.jpg

  7. In the Import dataset from template dialog, select SAP templates:

    Setting Value
    Source system name AdfSource57383073
    Dataset name SAPds57383073

    spendfxifhh7c.jpg

  8. In the Filter by keyword field, enter Spend.

  9. Select the Procure to pay - Spend Insights tile and then select Import.

    You may see multiple versions of the template. Each version targets a specific version of SAP S4HANA. Be sure to select the S4HANA2023 version.

    jb7d0xmi.jpg

    The Spend Insights dataset forms the foundation for Spend Insights dashboard. It provides a consolidated view of organizational expenditure across multiple dimensions.

    The dataset also enables monitoring of spending behavior, identification of cost-saving opportunities, and assurance of compliance with procurement policies within a single reporting layer.

  10. In the list of datasets, select SAPds57383073 and then, on the command bar, select Activate datasets.

    On the command bar, you may need to select Refresh to see the newly created dataset.

    yvujfca6.jpg

    Datasets are created in a disabled state. You must activate the dataset to use it.

  11. In the Activate selected dialog, select Activate.

    89z0e5h7.jpg

  12. It may take 1-3 minutes for the provisioning to complete. Do not proceed to the next task until the provisioning state for the dataset is Succeeded.

  13. Leave the Fabric page open. You will use the page again later in the workshop.

04: Run a pipeline to extract field metadata

  1. Return to the Azure Data Factory page or open a new browser tab and go to https://adf.azure.com/.

  2. On the Azure Data Factory page, in the Name field, select the available value and then select Continue.

  3. In the left pane, select Author (the pencil icon). Expand the Pipelines section and then expand Business Process Solutions.

    r4sri5co.jpg

  4. In the Business Process Solutions section, select the Get field metadata pipeline.

    g3drz4el.jpg

  5. On the command bar, select Add Trigger and then select Trigger now.

    ucfa94p8.jpg

  6. In the Pipeline run pane, select OK.

    yerxshfd.jpg n3un0kcn.jpg

  7. On the command bar, select Notifications (the bell icon).

    dyavy4re.jpg

  8. In the Notifications pane, select View pipeline run. This allows you to monitor to pipeline process.

  9. Do not proceed to the next task until the pipeline completes successfully.

    The pipeline populates the metadata table. It may take 15-20 minutes for the pipeline to complete.

  10. Leave the Fabric page open. You will use the page again later in the workshop.

05: Extract data

  1. Return to the Azure Data Factory page or open a new browser tab and go to https://adf.azure.com/.

  2. On the Azure Data Factory page, in the Name field, select the available value and then select Continue.

  3. In the left pane, select Author (the pencil icon). Expand the Pipelines section and then expand Business Process Solutions.

    r4sri5co.jpg

  4. In the Business Process Solutions section, select the Orchestration Master pipeline.

    5ppkpqy9.jpg

  5. On the command bar, select Add trigger and then select Trigger now.

  6. In the Pipeline run pane, select OK.

    yerxshfd.jpg n3un0kcn.jpg

  7. On the command bar, select Notifications (the bell icon).

    dyavy4re.jpg

  8. In the Notifications pane, select View pipeline run. This allows you to monitor to pipeline process.

  9. Do not proceed to the next task until the pipeline completes successfully.

    The pipeline populates the metadata table. It may take 2-3 minutes for the pipeline to complete.

06: Verify that the Silver lakehouse contains tables

  1. Return to the web page that displays Fabric or open a new browser tab and go to app.powerbi.com.

  2. In the left pane, select Workspaces and then select WS57383073.

    gl88nljo.jpg

  3. Locate and select the silver lakehouse.

    z6jhyy4z.jpg yoq1vayg.jpg

  4. In the Explorer pane, expand Tables. Verify that multiple tables are present.

    2i1lpoej.jpg

  5. Leave the Fabric page open. You will use the page again later in the workshop.

07: Process data

The pipeline has a single activity that automates a notebook.

sf033ghh.jpg

You can view the ei_orchestration_pipeline_full_processing notebook in the workspace to see all the operations that the code performs.

  1. Return to the web page that displays Fabric or open a new browser tab and go to app.powerbi.com.

  2. In the left pane, select Workspaces and then select WS57383073.

    gl88nljo.jpg

  3. Locate and select the ei_orchestration_pipeline_full_processing pipeline.

    xy3np9o9.jpg mb40yoig.jpg

  4. On the pipeline page, on the command bar, select Run All.

    lvcd2wbv.jpg

  5. Wait for the pipeline to complete before proceeding with the next task.

  6. Leave the Fabric page open. You will use the page again later in the workshop.

08: Verify that the Gold lakehouse contains tables

  1. Return to the web page that displays Fabric or open a new browser tab and go to app.powerbi.com.

  2. In the left pane, select Workspaces and then select WS57383073.

    gl88nljo.jpg

  3. Locate and select the gold lakehouse.

  4. In the Explorer pane, expand Tables. Verify that multiple tables are present.

  5. Leave the Fabric page open. You will use the page again later in the workshop.

09: Deploy Power BI reports

  1. Return to the web page that displays Fabric or open a new browser tab and go to app.powerbi.com.

  2. In the left pane, select Workspaces and then select WS57383073.

    gl88nljo.jpg

  3. Select the BPS57383073 BPS resource.

    u46ikubi.jpg

  4. On the command bar, select Insights.

    at9ntubb.jpg

  5. Select + New Insights.

    lmvvia1z.jpg

  6. In the Filter by keyword field, enter Spend.

  7. Select the Procure to Pay - Spend Insights Report tile and then select Deploy.

    k5fg94py.jpg

  8. In the New Insights dialog, in the Name field, enter Spend_Insights_Report.

  9. In the Select source system field, select AdfSource57383073 and then select Deploy.

  10. Wait for the, and then review the data.

    y7ndxb5y.jpg

  11. Once the deployment completes, you should be able to see the report in your workspace.

10: Configure the connection to allow semantic model refreshes

  1. Return to the web page that displays Fabric or open a new browser tab and go to app.powerbi.com.

  2. In the left pane, select Workspaces and then select WS57383073.

    gl88nljo.jpg

  3. Select the BPS57383073 BPS resource.

    u46ikubi.jpg

  4. On the command bar, select Insights.

    at9ntubb.jpg

  5. Select the Deployed Insights tab.

  6. Select the semantic model. Then, in the File menu, select Settings.

    83ikihsa.jpg

  7. Expand the Gateway and cloud connections section. Then, in the Cloud connections field, select Create a connection.

    ydibeorv.jpg

  8. In the Connection name field, enter Report_Connection.

  9. In the Authentication method field, select OAuth 2.0.

    y4vneo59.jpg

  10. Select Edit credentials. Select your credentials and then select Create.

  11. Expand the Gateway and cloud connections section. Then, in the Cloud connections field, select Report_Connection.

11: Import lakehouse views

  1. On the Fabric page, in the left pane, select Workspaces and then select WS57383073.

    gl88nljo.jpg

  2. Locate and select the ei_gold_view_creation notebook.

  3. On the command bar, select Run all. Wait for code execution to complete.

    vzdtvl9j.jpg

Congratulations!

You’ve completed this lab!