Note
Access to this page requires authorization. You can try signing in or changing directories.
Access to this page requires authorization. You can try changing directories.
In this quickstart, you'll build a data pipeline that moves a sample dataset into a Lakehouse. It's a simple way to see how pipeline copy activities work and how to load data into a Lakehouse.
Tip
You can also use a Copy job to move data from one place to another. Check out this decision guide to help you pick the right tool.
Prerequisites
Before you begin, make sure you have the following setup:
- A Microsoft Fabric tenant account with an active subscription. Create a free account.
- Make sure you have a Microsoft Fabric enabled Workspace: Create a workspace.
Create a data pipeline
Go to Power BI.
Select the Power BI icon in the lower left, then choose Fabric to open the Microsoft Fabric homepage.
Go to your Microsoft Fabric workspace. If you made a new workspace as a prerequisite, use that one.
Select New item, pick Data pipeline, and enter a name for your pipeline.
Copy data with your data pipeline
In your data pipeline, select Copy data assistant.
Choose the Sample data tab at the top of the data source browser page, then select the Public Holidays sample data, and then Next.
On the Connect to data source page of the assistant, the preview for the Public Holidays sample data is displayed, and then select Next.
To configure your destination, select Lakehouse.
Enter a Lakehouse name, then select Create and connect.
Configure and map your source data to the destination Lakehouse table. Select Tables for the Root folder and Load to new table for Load settings. Provide a Table name and select Next.
Review your copy activity settings, then select Save + run to finish. You can go back and change any settings if you need to. If you just want to save your pipeline without running it right away, clear the Start data transfer immediately checkbox.
The Copy activity now appears in your pipeline. When you select the Copy data activity, you'll see all its settings—including advanced options—in the tabs below the canvas.
Run and schedule your data pipeline
If you didn't choose to Save + run on the Review + save page of the Copy data assistant, switch to the Home tab and select Run. A confirmation dialog is displayed. Then select Save and run to start the activity.
You can watch your pipeline run and see the results on the Output tab below the canvas. To check the details, select the activity name in the output list.
The run details page shows how much data your pipeline read and wrote, along with other helpful info about the run.
You can set your pipeline to run on a schedule. For example, select Schedule to open the scheduling options, then pick how often you want it to run—for example, every 15 minutes.
Related content
This quickstart walked you through copying sample data into a Lakehouse using a data pipeline as a simple way to get hands-on with pipelines and see how easy it is to move data.
Next, learn how to monitor your pipeline runs and keep an eye on your data.