Note
Access to this page requires authorization. You can try signing in or changing directories.
Access to this page requires authorization. You can try changing directories.
Important
The notebook experience for Delta Live Tables development is in Public Preview.
This article describes features in Azure Databricks notebooks that assist in the development and debugging of Delta Live Tables code.
When you work on a Python or SQL notebook that is the source code for an existing Delta Live Tables pipeline, you can connect the notebook directly to the pipeline. When the notebook is connected to the pipeline, the following features are available:
- Start and validate the pipeline from the notebook.
- View the pipeline's dataflow graph and event log for the latest update in the notebook.
- View pipeline diagnostics in the notebook editor.
- View the status of the pipeline's cluster in the notebook.
- Access the Delta Live Tables UI from the notebook.
- You must have an existing Delta Live Tables pipeline with a Python or SQL notebook as source code.
- You must either be the owner of the pipeline or have the
CAN_MANAGE
privilege.
- The features covered in this article are only available in Azure Databricks notebooks. Workspace files are not supported.
- The web terminal is not available when attached to a pipeline. As a result, it is not visible as a tab in the bottom panel.
Inside the notebook, click on the drop-down menu used to select compute. The drop-down menu shows all your Delta Live Tables pipelines with this notebook as source code. To connect the notebook to a pipeline, select it from the list.
To easily understand the state of your pipeline's cluster, its status is shown in the compute drop-down menu with a green color to indicate that the cluster is running.
You can validate the pipeline to check for syntax errors in your source code without processing any data.
To validate a pipeline, do one of the following:
- In the top-right corner of the notebook, click Validate.
- Press
Shift+Enter
in any notebook cell. - In a cell's dropdown menu, click Validate Pipeline.
Note
If you attempt to validate your pipeline while an existing update is already running, a dialog box displays asking if you want to terminate the existing update. If you click Yes, the existing update stops, and a validate update automatically starts.
A pipeline update does the following: starts a cluster, discovers and validates all the tables and views defined, and creates or updates tables and views with the most recent data available.
To start an update of your pipeline, click the Start button in the top-right corner of the notebook.
The top panel in the notebook displays whether a pipeline update is:
- Starting
- Validating
- Stopping
After a pipeline has been started or validated, any errors are shown inline with a red underline. Hover over an error to see more information.
When attached to a pipeline, there is a Delta Live Tables event log tab at the bottom of the notebook.
To view a pipeline's dataflow graph, use the Delta Live Tables graph tab at the bottom of the notebook. Selecting a node in the graph displays its schema in the right panel.
To easily jump to the Delta Live Tables UI, use the menu in the top-right corner of the notebook.
The driver logs and Spark UI associated with the pipeline being developed can be easily accessed from the notebook's View menu.