Skip to main content Link Menu Expand (external link) Document Search Copy Copied

Task 2.1: Set up an Azure Databricks environment

Integrate OneLake with Databricks:

  • Use OneLake with existing data lakes using Shortcuts
  • Use and land data directly in OneLake

In this exercise, you will use second option Use and land data directly in OneLake.

Contoso already had some of their compute workload on Azure Databricks. You don’t need to migrate any of that workload to work with Fabric. You can simply use the OneLake endpoint to mount the storage and work with the same data directly from the Lakehouse for their analytical and ML operations.

Currently there are two ways to authenticate OneLake.

  • Credential passthrough
  • Service Principal approach (In this lab, you will use Service Principal approach)
  1. Open new tab in your browser and sign in to the Azure Databricks Workspace, , select https://@lab.Variable(workspaceurl) and press ENTER.

  2. Select Sign in with Azure AD.

    task-2.1.new7.png

    Skip or Close any popups that appear.

  3. Download the necessary notebooks to your local machine.

  4. Import the notebooks to your Azure Databricks environment.

  5. In the left navigation pane, select Workspace, select Workspace in the Workspace navigation menu and then select the 01_Setup-OneLake_Integration_with_Databrick notebook.

    Skip or Close any popups you see.

  6. In the cell named OneLake Path or cmd 2, replace “#WORKSPACE_NAME#” with the current Fabric workspace name , select ContosoSales@lab.LabInstance.Id.

    Select Workflows

  7. Select Run all. A new window will pop up.

  8. Select Attach and run to start executing the notebook.

  9. Once the setup notebook runs successfully, mounting to the OneLake is complete.

    It will take a few minutes for the execution to complete. You can check the last cell of notebook for the last execution time to verify.