Note
Access to this page requires authorization. You can try signing in or changing directories.
Access to this page requires authorization. You can try changing directories.
Note
This information applies to Databricks CLI versions 0.205 and above. The Databricks CLI is in Public Preview.
Databricks CLI use is subject to the Databricks License and Databricks Privacy Notice, including any Usage Data provisions.
The pipelines
command group within the Databricks CLI allows you to create, edit, delete, start, and view details about pipelines. See Lakeflow Declarative Pipelines.
databricks pipelines create
Create a new data processing pipeline based on the requested configuration. If successful, this command returns the ID of the new pipeline.
databricks pipelines create [flags]
Arguments
None
Options
--json JSON
The inline JSON string or the @path to the JSON file with the request body.
databricks pipelines delete
Delete a pipeline.
databricks pipelines delete PIPELINE_ID [flags]
Arguments
PIPELINE_ID
The pipeline to delete.
Options
databricks pipelines get
Get a pipeline.
databricks pipelines get PIPELINE_ID [flags]
Arguments
PIPELINE_ID
The pipeline to get.
Options
databricks pipelines get-update
Get an update from an active pipeline.
databricks pipelines get-update PIPELINE_ID UPDATE_ID [flags]
Arguments
PIPELINE_ID
The ID of the pipeline.
UPDATE_ID
The ID of the update.
Options
databricks pipelines list-pipeline-events
Retrieve events for a pipeline.
databricks pipelines list-pipeline-events PIPELINE_ID [flags]
Arguments
PIPELINE_ID
The pipeline to retrieve events for.
Options
--filter string
Criteria to select a subset of results, expressed using a SQL-like syntax.
--max-results int
Max number of entries to return in a single page.
--page-token string
Page token returned by previous call.
databricks pipelines list-pipelines
List pipelines defined in the Delta Live Tables system.
databricks pipelines list-pipelines [flags]
Arguments
None
Options
--filter string
Select a subset of results based on the specified criteria.
--max-results int
The maximum number of entries to return in a single page.
--page-token string
Page token returned by previous call.
databricks pipelines list-updates
List updates for an active pipeline.
databricks pipelines list-updates PIPELINE_ID [flags]
Arguments
PIPELINE_ID
The pipeline to return updates for.
Options
--max-results int
Max number of entries to return in a single page.
--page-token string
Page token returned by previous call.
--until-update-id string
If present, returns updates until and including this update_id.
databricks pipelines start-update
Start a new update for the pipeline. If there is already an active update for the pipeline, the request will fail and the active update will remain running.
databricks pipelines start-update PIPELINE_ID [flags]
Arguments
PIPELINE_ID
The pipeline to start an update for.
Options
--cause StartUpdateCause
Supported values: [API_CALL, JOB_TASK, RETRY_ON_FAILURE, SCHEMA_CHANGE, SERVICE_UPGRADE, USER_ACTION]
--full-refresh
If true, this update will reset all tables before running.
--json JSON
The inline JSON string or the @path to the JSON file with the request body.
--validate-only
If true, this update only validates the correctness of pipeline source code but does not materialize or publish any datasets.
databricks pipelines stop
Stop the pipeline by canceling the active update. If there is no active update for the pipeline, this request is a no-op.
databricks pipelines stop PIPELINE_ID [flags]
Arguments
PIPELINE_ID
The pipeline to stop.
Options
--no-wait
do not wait to reach IDLE state
--timeout duration
maximum amount of time to reach IDLE state (default 20m0s)
databricks pipelines update
Update a pipeline with the supplied configuration.
databricks pipelines update PIPELINE_ID [flags]
Arguments
PIPELINE_ID
Unique identifier for this pipeline.
Options
--allow-duplicate-names
If false, deployment will fail if name has changed and conflicts the name of another pipeline.
--budget-policy-id string
Budget policy of this pipeline.
--catalog string
A catalog in Unity Catalog to publish data from this pipeline to.
--channel string
DLT Release Channel that specifies which version to use.
--continuous
Whether the pipeline is continuous or triggered.
--development
Whether the pipeline is in Development mode.
--edition string
Pipeline product edition.
--expected-last-modified int
If present, the last-modified time of the pipeline settings before the edit.
--id string
Unique identifier for this pipeline.
--json JSON
The inline JSON string or the @path to the JSON file with the request body.
--name string
Friendly identifier for this pipeline.
--photon
Whether Photon is enabled for this pipeline.
--pipeline-id string
Unique identifier for this pipeline.
--schema string
The default schema (database) where tables are read from or published to.
--serverless
Whether serverless compute is enabled for this pipeline.
--storage string
DBFS root directory for storing checkpoints and tables.
--target string
Target schema (database) to add tables in this pipeline to.
databricks pipelines get-permission-levels
Get pipeline permission levels.
databricks pipelines get-permission-levels PIPELINE_ID [flags]
Arguments
PIPELINE_ID
The pipeline for which to get or manage permissions.
Options
databricks pipelines get-permissions
Get the permissions of a pipeline. Pipelines can inherit permissions from their root object.
databricks pipelines get-permissions PIPELINE_ID [flags]
Arguments
PIPELINE_ID
The pipeline for which to get or manage permissions.
Options
databricks pipelines set-permissions
Set pipeline permissions.
Sets permissions on an object, replacing existing permissions if they exist. Deletes all direct permissions if none are specified. Objects can inherit permissions from their root object.
databricks pipelines set-permissions PIPELINE_ID [flags]
Arguments
PIPELINE_ID
The pipeline for which to get or manage permissions.
Options
--json JSON
The inline JSON string or the @path to the JSON file with the request body.
databricks pipelines update-permissions
Update the permissions on a pipeline. Pipelines can inherit permissions from their root object.
databricks pipelines update-permissions PIPELINE_ID [flags]
Arguments
PIPELINE_ID
The pipeline for which to get or manage permissions.
Options
--json JSON
The inline JSON string or the @path to the JSON file with the request body.
Global flags
--debug
Whether to enable debug logging.
-h
or --help
Display help for the Databricks CLI or the related command group or the related command.
--log-file
string
A string representing the file to write output logs to. If this flag is not specified then the default is to write output logs to stderr.
--log-format
format
The log format type, text
or json
. The default value is text
.
--log-level
string
A string representing the log format level. If not specified then the log format level is disabled.
-o, --output
type
The command output type, text
or json
. The default value is text
.
-p, --profile
string
The name of the profile in the ~/.databrickscfg
file to use to run the command. If this flag is not specified then if it exists, the profile named DEFAULT
is used.
--progress-format
format
The format to display progress logs: default
, append
, inplace
, or json
-t, --target
string
If applicable, the bundle target to use