DagsterDocs

Dagster CLI

dagster asset

Commands for working with Dagster assets.

dagster asset [OPTIONS] COMMAND [ARGS]...

wipe

Eliminate asset key indexes from event logs.

Warning: Cannot be undone.

Usage:
dagster asset wipe –all
dagster asset wipe <unstructured_asset_key_name>
dagster asset wipe <json_string_of_structured_asset_key>
dagster asset wipe [OPTIONS] [KEY]...

Options

--all

Eliminate all asset key indexes

Arguments

KEY

Optional argument(s)

dagster debug

Commands for debugging Dagster pipeline runs.

dagster debug [OPTIONS] COMMAND [ARGS]...

export

Export the relevant artifacts for a pipeline run to a file.

dagster debug export [OPTIONS] RUN_ID OUTPUT_FILE

Arguments

RUN_ID

Required argument

OUTPUT_FILE

Required argument

dagster instance

Commands for working with the current Dagster instance.

dagster instance [OPTIONS] COMMAND [ARGS]...

info

List the information about the current instance.

dagster instance info [OPTIONS]

migrate

Automatically migrate an out of date instance.

dagster instance migrate [OPTIONS]

reindex

Rebuild index over historical runs for performance.

dagster instance reindex [OPTIONS]

dagster pipeline

Commands for working with Dagster pipelines.

dagster pipeline [OPTIONS] COMMAND [ARGS]...

backfill

Backfill a partitioned pipeline.

This commands targets a partitioned pipeline. The pipeline and partition set must be defined in a repository, which can be specified in a number of ways:

  1. dagster pipeline backfill -p <<pipeline_name>> (works if .workspace.yaml exists)

  2. dagster pipeline backfill -p <<pipeline_name>> -w path/to/workspace.yaml

  3. dagster pipeline backfill -f /path/to/file.py -a define_some_repo -p <<pipeline_name>>

  4. dagster pipeline backfill -m a_module.submodule -a define_some_repo -p <<pipeline_name>>

dagster pipeline backfill [OPTIONS]

Options

-p, --pipeline <pipeline>

Pipeline within the repository, necessary if more than one pipeline is present.

-l, --location <location>

RepositoryLocation within the workspace, necessary if more than one location is present.

-r, --repository <repository>

Repository within the workspace, necessary if more than one repository is present.

--use-ssl

Use a secure channel when connecting to the gRPC server

--grpc-host <grpc_host>

Host to use to connect to gRPC server, defaults to localhost

--grpc-socket <grpc_socket>

Named socket to use to connect to gRPC server

--grpc-port <grpc_port>

Port to use to connect to gRPC server

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-w, --workspace <workspace>

Path to workspace file. Argument can be provided multiple times.

--empty-workspace

Allow an empty workspace

--partitions <partitions>

Comma-separated list of partition names that we want to backfill

--partition-set <partition_set>

The name of the partition set over which we want to backfill.

--all <all>

Specify to select all partitions to backfill.

--from <from>

Specify a start partition for this backfill job

Example: dagster pipeline backfill log_daily_stats –from 20191101

--to <to>

Specify an end partition for this backfill job

Example: dagster pipeline backfill log_daily_stats –to 20191201

--tags <tags>

JSON string of tags to use for this pipeline run

--noprompt

execute

Execute a pipeline.

This commands targets a pipeline. The pipeline can be specified in a number of ways:

  1. dagster pipeline execute -f /path/to/file.py -a define_some_pipeline

  2. dagster pipeline execute -m a_module.submodule -a define_some_pipeline

  3. dagster pipeline execute -f /path/to/file.py -a define_some_repo -p <<pipeline_name>>

  4. dagster pipeline execute -m a_module.submodule -a define_some_repo -p <<pipeline_name>>

dagster pipeline execute [OPTIONS]

Options

-p, --pipeline <pipeline>

Pipeline within the repository, necessary if more than one pipeline is present.

-r, --repository <repository>

Repository name, necessary if more than one repository is present.

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-c, --config <config>

Specify one or more run config files. These can also be file patterns. If more than one run config file is captured then those files are merged. Files listed first take precedence. They will smash the values of subsequent files at the key-level granularity. If the file is a pattern then you must enclose it in double quotes

Example: dagster pipeline execute -f hello_world.py -p pandas_hello_world -c “pandas_hello_world/*.yaml”

You can also specify multiple files:

Example: dagster pipeline execute -f hello_world.py -p pandas_hello_world -c pandas_hello_world/solids.yaml -c pandas_hello_world/env.yaml

--preset <preset>

Specify a preset to use for this pipeline. Presets are defined on pipelines under preset_defs.

--mode <mode>

The name of the mode in which to execute the pipeline.

--tags <tags>

JSON string of tags to use for this pipeline run

-s, --solid-selection <solid_selection>

Specify the solid subselection to execute. It can be multiple clauses separated by commas.Examples: - “some_solid” will execute “some_solid” itself - “*some_solid” will execute “some_solid” and all its ancestors (upstream dependencies) - “*some_solid+++” will execute “some_solid”, all its ancestors, and its descendants (downstream dependencies) within 3 levels down - “*some_solid,other_solid_a,other_solid_b+” will execute “some_solid” and all its ancestors, “other_solid_a” itself, and “other_solid_b” and its direct child solids

launch

Launch a pipeline using the run launcher configured on the Dagster instance.

This commands targets a pipeline. The pipeline can be specified in a number of ways:

  1. dagster pipeline launch -p <<pipeline_name>> (works if .workspace.yaml exists)

  2. dagster pipeline launch -p <<pipeline_name>> -w path/to/workspace.yaml

  3. dagster pipeline launch -f /path/to/file.py -a define_some_pipeline

  4. dagster pipeline launch -m a_module.submodule -a define_some_pipeline

  5. dagster pipeline launch -f /path/to/file.py -a define_some_repo -p <<pipeline_name>>

  6. dagster pipeline launch -m a_module.submodule -a define_some_repo -p <<pipeline_name>>

dagster pipeline launch [OPTIONS]

Options

-p, --pipeline <pipeline>

Pipeline within the repository, necessary if more than one pipeline is present.

-l, --location <location>

RepositoryLocation within the workspace, necessary if more than one location is present.

-r, --repository <repository>

Repository within the workspace, necessary if more than one repository is present.

--use-ssl

Use a secure channel when connecting to the gRPC server

--grpc-host <grpc_host>

Host to use to connect to gRPC server, defaults to localhost

--grpc-socket <grpc_socket>

Named socket to use to connect to gRPC server

--grpc-port <grpc_port>

Port to use to connect to gRPC server

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-w, --workspace <workspace>

Path to workspace file. Argument can be provided multiple times.

--empty-workspace

Allow an empty workspace

-c, --config <config>

Specify one or more run config files. These can also be file patterns. If more than one run config file is captured then those files are merged. Files listed first take precedence. They will smash the values of subsequent files at the key-level granularity. If the file is a pattern then you must enclose it in double quotes

Example: dagster pipeline launch -f hello_world.py -p pandas_hello_world -c “pandas_hello_world/*.yaml”

You can also specify multiple files:

Example: dagster pipeline launch -f hello_world.py -p pandas_hello_world -c pandas_hello_world/solids.yaml -c pandas_hello_world/env.yaml

--config-json <config_json>

JSON string of run config to use for this pipeline run. Cannot be used with -c / –config.

--preset <preset>

Specify a preset to use for this pipeline. Presets are defined on pipelines under preset_defs.

--mode <mode>

The name of the mode in which to execute the pipeline.

--tags <tags>

JSON string of tags to use for this pipeline run

-s, --solid-selection <solid_selection>

Specify the solid subselection to launch. It can be multiple clauses separated by commas.Examples: - “some_solid” will launch “some_solid” itself - “*some_solid” will launch “some_solid” and all its ancestors (upstream dependencies) - “*some_solid+++” will launch “some_solid”, all its ancestors, and its descendants (downstream dependencies) within 3 levels down - “*some_solid,other_solid_a,other_solid_b+” will launch “some_solid” and all its ancestors, “other_solid_a” itself, and “other_solid_b” and its direct child solids

--run-id <run_id>

The ID to give to the launched pipeline run

list

List the pipelines in a repository. Can only use ONE of –workspace/-w, –python-file/-f, –module-name/-m, –grpc-port, –grpc-socket.

dagster pipeline list [OPTIONS]

Options

-l, --location <location>

RepositoryLocation within the workspace, necessary if more than one location is present.

-r, --repository <repository>

Repository within the workspace, necessary if more than one repository is present.

--use-ssl

Use a secure channel when connecting to the gRPC server

--grpc-host <grpc_host>

Host to use to connect to gRPC server, defaults to localhost

--grpc-socket <grpc_socket>

Named socket to use to connect to gRPC server

--grpc-port <grpc_port>

Port to use to connect to gRPC server

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-w, --workspace <workspace>

Path to workspace file. Argument can be provided multiple times.

--empty-workspace

Allow an empty workspace

list_versions

Display the freshness of memoized results for the given pipeline.

This commands targets a pipeline. The pipeline can be specified in a number of ways:

  1. dagster pipeline list_versions -f /path/to/file.py -a define_some_pipeline

  2. dagster pipeline list_versions -m a_module.submodule -a define_some_pipeline

  3. dagster pipeline list_versions -f /path/to/file.py -a define_some_repo -p <<pipeline_name>>

  4. dagster pipeline list_versions -m a_module.submodule -a define_some_repo -p <<pipeline_name>>

dagster pipeline list_versions [OPTIONS]

Options

-p, --pipeline <pipeline>

Pipeline within the repository, necessary if more than one pipeline is present.

-r, --repository <repository>

Repository name, necessary if more than one repository is present.

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-c, --config <config>

Specify one or more run config files. These can also be file patterns. If more than one run config file is captured then those files are merged. Files listed first take precedence. They will smash the values of subsequent files at the key-level granularity. If the file is a pattern then you must enclose it in double quotes

Example: dagster pipeline list_versions -f hello_world.py -p pandas_hello_world -c “pandas_hello_world/*.yaml”

You can also specify multiple files:

Example: dagster pipeline list_versions -f hello_world.py -p pandas_hello_world -c pandas_hello_world/solids.yaml -c pandas_hello_world/env.yaml

--preset <preset>

Specify a preset to use for this pipeline. Presets are defined on pipelines under preset_defs.

--mode <mode>

The name of the mode in which to execute the pipeline.

print

Print a pipeline.

This commands targets a pipeline. The pipeline can be specified in a number of ways:

  1. dagster pipeline print -p <<pipeline_name>> (works if .workspace.yaml exists)

  2. dagster pipeline print -p <<pipeline_name>> -w path/to/workspace.yaml

  3. dagster pipeline print -f /path/to/file.py -a define_some_pipeline

  4. dagster pipeline print -m a_module.submodule -a define_some_pipeline

  5. dagster pipeline print -f /path/to/file.py -a define_some_repo -p <<pipeline_name>>

  6. dagster pipeline print -m a_module.submodule -a define_some_repo -p <<pipeline_name>>

dagster pipeline print [OPTIONS]

Options

--verbose
-p, --pipeline <pipeline>

Pipeline within the repository, necessary if more than one pipeline is present.

-l, --location <location>

RepositoryLocation within the workspace, necessary if more than one location is present.

-r, --repository <repository>

Repository within the workspace, necessary if more than one repository is present.

--use-ssl

Use a secure channel when connecting to the gRPC server

--grpc-host <grpc_host>

Host to use to connect to gRPC server, defaults to localhost

--grpc-socket <grpc_socket>

Named socket to use to connect to gRPC server

--grpc-port <grpc_port>

Port to use to connect to gRPC server

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-w, --workspace <workspace>

Path to workspace file. Argument can be provided multiple times.

--empty-workspace

Allow an empty workspace

scaffold_config

Scaffold the config for a pipeline.

This commands targets a pipeline. The pipeline can be specified in a number of ways:

  1. dagster pipeline scaffold_config -f /path/to/file.py -a define_some_pipeline

  2. dagster pipeline scaffold_config -m a_module.submodule -a define_some_pipeline

  3. dagster pipeline scaffold_config -f /path/to/file.py -a define_some_repo -p <<pipeline_name>>

  4. dagster pipeline scaffold_config -m a_module.submodule -a define_some_repo -p <<pipeline_name>>

dagster pipeline scaffold_config [OPTIONS]

Options

-p, --pipeline <pipeline>

Pipeline within the repository, necessary if more than one pipeline is present.

-r, --repository <repository>

Repository name, necessary if more than one repository is present.

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

--print-only-required

dagster run

Commands for working with Dagster pipeline runs.

dagster run [OPTIONS] COMMAND [ARGS]...

delete

Delete a run by id and its associated event logs. Warning: Cannot be undone

dagster run delete [OPTIONS] RUN_ID

Options

-f, --force

Skip prompt to delete run.

Arguments

RUN_ID

Required argument

list

List the runs in the current Dagster instance.

dagster run list [OPTIONS]

Options

--limit <limit>

Only list a specified number of runs

wipe

Eliminate all run history and event logs. Warning: Cannot be undone.

dagster run wipe [OPTIONS]

Options

-f, --force

Skip prompt to delete run history and event logs.

dagster schedule

Commands for working with Dagster schedules.

dagster schedule [OPTIONS] COMMAND [ARGS]...

debug

Debug information about the scheduler.

dagster schedule debug [OPTIONS]

list

List all schedules that correspond to a repository.

dagster schedule list [OPTIONS]

Options

-l, --location <location>

RepositoryLocation within the workspace, necessary if more than one location is present.

-r, --repository <repository>

Repository within the workspace, necessary if more than one repository is present.

--use-ssl

Use a secure channel when connecting to the gRPC server

--grpc-host <grpc_host>

Host to use to connect to gRPC server, defaults to localhost

--grpc-socket <grpc_socket>

Named socket to use to connect to gRPC server

--grpc-port <grpc_port>

Port to use to connect to gRPC server

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-w, --workspace <workspace>

Path to workspace file. Argument can be provided multiple times.

--empty-workspace

Allow an empty workspace

--running

Filter for running schedules

--stopped

Filter for stopped schedules

--name

Only display schedule schedule names

logs

Get logs for a schedule.

dagster schedule logs [OPTIONS] [SCHEDULE_NAME]...

Options

-l, --location <location>

RepositoryLocation within the workspace, necessary if more than one location is present.

-r, --repository <repository>

Repository within the workspace, necessary if more than one repository is present.

--use-ssl

Use a secure channel when connecting to the gRPC server

--grpc-host <grpc_host>

Host to use to connect to gRPC server, defaults to localhost

--grpc-socket <grpc_socket>

Named socket to use to connect to gRPC server

--grpc-port <grpc_port>

Port to use to connect to gRPC server

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-w, --workspace <workspace>

Path to workspace file. Argument can be provided multiple times.

--empty-workspace

Allow an empty workspace

Arguments

SCHEDULE_NAME

Optional argument(s)

preview

Preview changes that will be performed by dagster schedule up.

dagster schedule preview [OPTIONS]

Options

-l, --location <location>

RepositoryLocation within the workspace, necessary if more than one location is present.

-r, --repository <repository>

Repository within the workspace, necessary if more than one repository is present.

--use-ssl

Use a secure channel when connecting to the gRPC server

--grpc-host <grpc_host>

Host to use to connect to gRPC server, defaults to localhost

--grpc-socket <grpc_socket>

Named socket to use to connect to gRPC server

--grpc-port <grpc_port>

Port to use to connect to gRPC server

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-w, --workspace <workspace>

Path to workspace file. Argument can be provided multiple times.

--empty-workspace

Allow an empty workspace

restart

Restart a running schedule.

dagster schedule restart [OPTIONS] [SCHEDULE_NAME]...

Options

--restart-all-running

restart previously running schedules

-l, --location <location>

RepositoryLocation within the workspace, necessary if more than one location is present.

-r, --repository <repository>

Repository within the workspace, necessary if more than one repository is present.

--use-ssl

Use a secure channel when connecting to the gRPC server

--grpc-host <grpc_host>

Host to use to connect to gRPC server, defaults to localhost

--grpc-socket <grpc_socket>

Named socket to use to connect to gRPC server

--grpc-port <grpc_port>

Port to use to connect to gRPC server

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-w, --workspace <workspace>

Path to workspace file. Argument can be provided multiple times.

--empty-workspace

Allow an empty workspace

Arguments

SCHEDULE_NAME

Optional argument(s)

start

Start an existing schedule.

dagster schedule start [OPTIONS] [SCHEDULE_NAME]...

Options

--start-all

start all schedules

-l, --location <location>

RepositoryLocation within the workspace, necessary if more than one location is present.

-r, --repository <repository>

Repository within the workspace, necessary if more than one repository is present.

--use-ssl

Use a secure channel when connecting to the gRPC server

--grpc-host <grpc_host>

Host to use to connect to gRPC server, defaults to localhost

--grpc-socket <grpc_socket>

Named socket to use to connect to gRPC server

--grpc-port <grpc_port>

Port to use to connect to gRPC server

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-w, --workspace <workspace>

Path to workspace file. Argument can be provided multiple times.

--empty-workspace

Allow an empty workspace

Arguments

SCHEDULE_NAME

Optional argument(s)

stop

Stop an existing schedule.

dagster schedule stop [OPTIONS] [SCHEDULE_NAME]...

Options

-l, --location <location>

RepositoryLocation within the workspace, necessary if more than one location is present.

-r, --repository <repository>

Repository within the workspace, necessary if more than one repository is present.

--use-ssl

Use a secure channel when connecting to the gRPC server

--grpc-host <grpc_host>

Host to use to connect to gRPC server, defaults to localhost

--grpc-socket <grpc_socket>

Named socket to use to connect to gRPC server

--grpc-port <grpc_port>

Port to use to connect to gRPC server

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-w, --workspace <workspace>

Path to workspace file. Argument can be provided multiple times.

--empty-workspace

Allow an empty workspace

Arguments

SCHEDULE_NAME

Optional argument(s)

up

Update the Dagster instance’s representation of schedules.

After running this command, the internal representation will match the list of ScheduleDefinitions defined in the repository.

To preview what changes will be applied, one of the following commands can be used:
dagster schedule up –preview
dagster schedule preview

Note by default, new ScheduleDefinitions in the Dagster instance will not start running until they are explicitly started.

dagster schedule up [OPTIONS]

Options

--preview

Preview changes

-l, --location <location>

RepositoryLocation within the workspace, necessary if more than one location is present.

-r, --repository <repository>

Repository within the workspace, necessary if more than one repository is present.

--use-ssl

Use a secure channel when connecting to the gRPC server

--grpc-host <grpc_host>

Host to use to connect to gRPC server, defaults to localhost

--grpc-socket <grpc_socket>

Named socket to use to connect to gRPC server

--grpc-port <grpc_port>

Port to use to connect to gRPC server

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-w, --workspace <workspace>

Path to workspace file. Argument can be provided multiple times.

--empty-workspace

Allow an empty workspace

wipe

Delete the schedule history and turn off all schedules.

dagster schedule wipe [OPTIONS]

Options

-l, --location <location>

RepositoryLocation within the workspace, necessary if more than one location is present.

-r, --repository <repository>

Repository within the workspace, necessary if more than one repository is present.

--use-ssl

Use a secure channel when connecting to the gRPC server

--grpc-host <grpc_host>

Host to use to connect to gRPC server, defaults to localhost

--grpc-socket <grpc_socket>

Named socket to use to connect to gRPC server

--grpc-port <grpc_port>

Port to use to connect to gRPC server

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-w, --workspace <workspace>

Path to workspace file. Argument can be provided multiple times.

--empty-workspace

Allow an empty workspace

dagster sensor

Commands for working with Dagster sensors.

dagster sensor [OPTIONS] COMMAND [ARGS]...

cursor

Set the cursor value for an existing sensor.

dagster sensor cursor [OPTIONS] [SENSOR_NAME]...

Options

--set <set>

Set the cursor value for the sensor context

--delete

Delete the existing cursor value for the sensor context

-l, --location <location>

RepositoryLocation within the workspace, necessary if more than one location is present.

-r, --repository <repository>

Repository within the workspace, necessary if more than one repository is present.

--use-ssl

Use a secure channel when connecting to the gRPC server

--grpc-host <grpc_host>

Host to use to connect to gRPC server, defaults to localhost

--grpc-socket <grpc_socket>

Named socket to use to connect to gRPC server

--grpc-port <grpc_port>

Port to use to connect to gRPC server

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-w, --workspace <workspace>

Path to workspace file. Argument can be provided multiple times.

--empty-workspace

Allow an empty workspace

Arguments

SENSOR_NAME

Optional argument(s)

list

List all sensors that correspond to a repository.

dagster sensor list [OPTIONS]

Options

-l, --location <location>

RepositoryLocation within the workspace, necessary if more than one location is present.

-r, --repository <repository>

Repository within the workspace, necessary if more than one repository is present.

--use-ssl

Use a secure channel when connecting to the gRPC server

--grpc-host <grpc_host>

Host to use to connect to gRPC server, defaults to localhost

--grpc-socket <grpc_socket>

Named socket to use to connect to gRPC server

--grpc-port <grpc_port>

Port to use to connect to gRPC server

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-w, --workspace <workspace>

Path to workspace file. Argument can be provided multiple times.

--empty-workspace

Allow an empty workspace

--running

Filter for running sensors

--stopped

Filter for stopped sensors

--name

Only display sensor sensor names

preview

Preview an existing sensor execution.

dagster sensor preview [OPTIONS] [SENSOR_NAME]...

Options

--since <since>

Set the last_completion_time value as a timestamp float for the sensor context

--last_run_key <last_run_key>

Set the last_run_key value for the sensor context

--cursor <cursor>

Set the cursor value for the sensor context

-l, --location <location>

RepositoryLocation within the workspace, necessary if more than one location is present.

-r, --repository <repository>

Repository within the workspace, necessary if more than one repository is present.

--use-ssl

Use a secure channel when connecting to the gRPC server

--grpc-host <grpc_host>

Host to use to connect to gRPC server, defaults to localhost

--grpc-socket <grpc_socket>

Named socket to use to connect to gRPC server

--grpc-port <grpc_port>

Port to use to connect to gRPC server

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-w, --workspace <workspace>

Path to workspace file. Argument can be provided multiple times.

--empty-workspace

Allow an empty workspace

Arguments

SENSOR_NAME

Optional argument(s)

start

Start an existing sensor.

dagster sensor start [OPTIONS] [SENSOR_NAME]...

Options

--start-all

start all sensors

-l, --location <location>

RepositoryLocation within the workspace, necessary if more than one location is present.

-r, --repository <repository>

Repository within the workspace, necessary if more than one repository is present.

--use-ssl

Use a secure channel when connecting to the gRPC server

--grpc-host <grpc_host>

Host to use to connect to gRPC server, defaults to localhost

--grpc-socket <grpc_socket>

Named socket to use to connect to gRPC server

--grpc-port <grpc_port>

Port to use to connect to gRPC server

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-w, --workspace <workspace>

Path to workspace file. Argument can be provided multiple times.

--empty-workspace

Allow an empty workspace

Arguments

SENSOR_NAME

Optional argument(s)

stop

Stop an existing sensor.

dagster sensor stop [OPTIONS] [SENSOR_NAME]...

Options

-l, --location <location>

RepositoryLocation within the workspace, necessary if more than one location is present.

-r, --repository <repository>

Repository within the workspace, necessary if more than one repository is present.

--use-ssl

Use a secure channel when connecting to the gRPC server

--grpc-host <grpc_host>

Host to use to connect to gRPC server, defaults to localhost

--grpc-socket <grpc_socket>

Named socket to use to connect to gRPC server

--grpc-port <grpc_port>

Port to use to connect to gRPC server

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-w, --workspace <workspace>

Path to workspace file. Argument can be provided multiple times.

--empty-workspace

Allow an empty workspace

Arguments

SENSOR_NAME

Optional argument(s)

dagster-graphql

Run a GraphQL query against the dagster interface to a specified repository or pipeline.

Can only use ONE of –workspace/-w, –python-file/-f, –module-name/-m, –grpc-port, –grpc-socket.

Examples:

  1. dagster-graphql

  2. dagster-graphql -y path/to/workspace.yaml

  3. dagster-graphql -f path/to/file.py -a define_repo

  4. dagster-graphql -m some_module -a define_repo

  5. dagster-graphql -f path/to/file.py -a define_pipeline

  6. dagster-graphql -m some_module -a define_pipeline

dagster-graphql [OPTIONS]

Options

--version

Show the version and exit.

-t, --text <text>

GraphQL document to execute passed as a string

-f, --file <file>

GraphQL document to execute passed as a file

-p, --predefined <predefined>

GraphQL document to execute, from a predefined set provided by dagster-graphql.

Options

launchPipelineExecution

-v, --variables <variables>

A JSON encoded string containing the variables for GraphQL execution.

-r, --remote <remote>

A URL for a remote instance running dagit server to send the GraphQL request to.

-o, --output <output>

A file path to store the GraphQL response to. This flag is useful when making pipeline execution queries, since pipeline execution causes logs to print to stdout and stderr.

--ephemeral-instance

Use an ephemeral DagsterInstance instead of resolving via DAGSTER_HOME

--empty-workspace

Allow an empty workspace

-w, --workspace <workspace>

Path to workspace file. Argument can be provided multiple times.

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

--grpc-port <grpc_port>

Port to use to connect to gRPC server

--grpc-socket <grpc_socket>

Named socket to use to connect to gRPC server

--grpc-host <grpc_host>

Host to use to connect to gRPC server, defaults to localhost

--use-ssl

Use a secure channel when connecting to the gRPC server

dagit

Run dagit. Loads a repository or pipeline.

Can only use ONE of –workspace/-w, –python-file/-f, –module-name/-m, –grpc-port, –grpc-socket.

Examples:

  1. dagit (works if .workspace.yaml exists)

  2. dagit -w path/to/workspace.yaml

  3. dagit -f path/to/file.py

  4. dagit -f path/to/file.py -d path/to/working_directory

  5. dagit -m some_module

  6. dagit -f path/to/file.py -a define_repo

  7. dagit -m some_module -a define_repo

  8. dagit -p 3333

Options can also provide arguments via environment variables prefixed with DAGIT

For example, DAGIT_PORT=3333 dagit

dagit [OPTIONS]

Options

--use-ssl

Use a secure channel when connecting to the gRPC server

--grpc-host <grpc_host>

Host to use to connect to gRPC server, defaults to localhost

--grpc-socket <grpc_socket>

Named socket to use to connect to gRPC server

--grpc-port <grpc_port>

Port to use to connect to gRPC server

-a, --attribute <attribute>

Attribute that is either a 1) repository or pipeline or 2) a function that returns a repository or pipeline

-m, --module-name <module_name>

Specify module where repository or pipeline function lives

--package-name <package_name>

Specify installed Python package where repository or pipeline function lives

-f, --python-file <python_file>

Specify python file where repository or pipeline function lives

-d, --working-directory <working_directory>

Specify working directory to use when loading the repository or pipeline. Can only be used along with -f/–python-file

--empty-working-directory

Indicates that the working directory should be empty and should not set to the current directory as a default

-w, --workspace <workspace>

Path to workspace file. Argument can be provided multiple times.

--empty-workspace

Allow an empty workspace

-h, --host <host>

Host to run server on

Default

127.0.0.1

-p, --port <port>

Port to run server on, default is 3000

-l, --path-prefix <path_prefix>

The path prefix where Dagit will be hosted (eg: /dagit)

Default

--db-statement-timeout <db_statement_timeout>

The timeout in milliseconds to set on database statements sent to the DagsterInstance. Not respected in all configurations.

Default

5000

--read-only

Start Dagit in read-only mode, where all mutations such as launching runs and turning schedules on/off are turned off.

--suppress-warnings

Filter all warnings when hosting Dagit.

--version

Show the version and exit.