- The
FreshnessPolicy
object now supports a cron_schedule_timezone
argument. AssetsDefinition.from_graph
now supports a freshness_policies_by_output_name
parameter.- The
@asset_sensor
will now display an informative SkipReason
when no new materializations have been created since the last sensor tick. AssetsDefinition
now has a to_source_asset
method, which returns a representation of this asset as a SourceAsset
.- You can now designate assets as inputs to ops within a graph or graph-based job. E.g.
from dagster import asset, job, op
@asset
def emails_to_send():
...
@op
def send_emails(emails) -> None:
...
@job
def send_emails_job():
send_emails(emails_to_send.to_source_asset())
- Added a
--dagit-host/-h
argument to the dagster dev
command to allow customization of the host where Dagit runs. - [dagster-snowflake, dagster-duckdb] Database I/O managers (Snowflake, DuckDB) now support static partitions, multi-partitions, and dynamic partitions.
- Previously, if a description was provided for an op that backed a multi-asset, the op’s description would override the descriptions in Dagit for the individual assets. This has been fixed.
- Sometimes, when applying an
input_manager_key
to an asset’s input, incorrect resource config could be used when loading that input. This has been fixed. - Previously, the backfill page errored when partitions definitions changed for assets that had been backfilled. This has been fixed.
- When displaying materialized partitions for multipartitioned assets, Dagit would error if a dimension had zero partitions. This has been fixed.
- [dagster-k8s] Fixed an issue where setting
runK8sConfig
in the Dagster Helm chart would not pass configuration through to pods launched using the k8s_job_executor
. - [dagster-k8s] Previously, using the
execute_k8s_job
op downstream of a dynamic output would result in k8s jobs with duplicate names being created. This has been fixed. - [dagster-snowflake] Previously, if the schema for storing outputs didn’t exist, the Snowflake I/O manager would fail. Now it creates the schema.
- Removed the experimental, undocumented
asset_key
, asset_partitions
, and asset_partitions_defs
arguments on Out
. @multi_asset
no longer accepts Out
values in the dictionary passed to its outs
argument. This was experimental and deprecated. Instead, use AssetOut
.- The experimental, undocumented
top_level_resources
argument to the repository
decorator has been renamed to _top_level_resources
to emphasize that it should not be set manually.
load_asset_values
now accepts resource configuration (thanks @Nintorac!)- Previously, when using the
UPathIOManager
, paths with the "."
character in them would be incorrectly truncated, which could result in multiple distinct objects being written to the same path. This has been fixed. (Thanks @spenczar!)
- [dagster-dbt] Added documentation to our dbt Cloud integration to cache the loading of software-defined assets from a dbt Cloud job.
- Revamped the introduction to the Partitions concepts page to make it clear that non-time-window partitions are equally encouraged.
- In Navigation, moved the Partitions and Backfill concept pages to their own section underneath Concepts.
- Moved the Running Dagster locally guide from Deployment to Guides to reflect that OSS and Cloud users can follow it.
- Added a new guide covering asset versioning and caching.
- Assets with time-window
PartitionsDefinition
s (e.g. HourlyPartitionsDefinition
, DailyPartitionsDefinition
) may now have a FreshnessPolicy
. - [dagster-dbt] When using
load_assets_from_dbt_project
or load_assets_from_dbt_manifest
with dbt-core>=1.4
, AssetMaterialization
events will be emitted as the dbt command executes, rather than waiting for dbt to complete before emitting events. - [dagster-aws] When run monitoring detects that a run unexpectedly crashed or failed to start, an error message in the run’s event log will include log messages from the ECS task for that run to help diagnose the cause of the failure.
- [dagster-airflow] added
make_ephemeral_airflow_db_resource
which returns a ResourceDefinition
for a local only airflow database for use in migrated airflow DAGs - Made some performance improvements for job run queries which can be applied by running
dagster instance migrate
. - [dagit] System tags (code + logical versions) are now shown in the asset sidebar and on the asset details page.
- [dagit] Source assets that have never been observed are presented more clearly on the asset graph.
- [dagit] The number of materialized and missing partitions are shown on the asset graph and in the asset catalog for partitioned assets.
- [dagit] Databricks-backed assets are now shown on the asset graph with a small “Databricks” logo.
- Fixed a bug where materializations of part of the asset graph did not construct required resource keys correctly.
- Fixed an issue where
observable_source_asset
incorrectly required its function to have a context
argument. - Fixed an issue with serialization of freshness policies, which affected cacheable assets that included these policies such as those from
dagster-airbyte
- [dagster-dbt] Previously, the
dagster-dbt
integration was incompatible with dbt-core>=1.4
. This has been fixed. - [dagster-dbt]
load_assets_from_dbt_cloud_job
will now avoid unnecessarily generating docs when compiling a manifest for the job. Compile runs will no longer be kicked off for jobs not managed by this integration. - Previously for multipartitioned assets,
context.asset_partition_key
returned a string instead of a MultiPartitionKey
. This has been fixed. - [dagster-k8s] Fixed an issue where pods launched by the
k8s_job_executor
would sometimes unexpectedly fail due to transient 401 errors in certain kubernetes clusters. - Fix a bug with nth-weekday-of-the-month handling in cron schedules.
- [dagster-airflow]
load_assets_from_airflow_dag
no longer creates airflow db resource definitions, as a user you will need to provide them on Definitions
directly
- The
partitions_fn
argument of the DynamicPartitionsDefinition
class is now deprecated and will be removed in 2.0.0.
- [dagster-wandb] A new integration with Weights & Biases allows you to orchestrate your MLOps pipelines and maintain ML assets with Dagster.
- Postgres has been updated to 14.6 for Dagster’s helm chart. Thanks @DustyShap!
- Typo fixed in docs. Thanks @C0DK!
- You can now pass a callable directly to
asset
(rather than using @asset
in decorator form) to create an asset. Thanks @ns-finkelstein!
- New “Asset versioning and caching” guide
- [dagster-snowflake] The Snowflake guide has been updated to include PySpark dataframes
- [dagster-snowflake] The Snowflake guide has been updated to include private key authentication
- [dagster-airflow] The Airflow migration guide has been update to include more detailed instructions and considerations for making a migration
- The
dagster-airflow
library as been moved to 1.x.x to denote the stability of its api's going forward. - [dagster-airflow]
make_schedules_and_jobs_from_airflow_dag_bag
has been added to allow for more fine grained composition of your transformed airflow DAGs into Dagster. - [dagster-airflow] Airflow dag task
retries
and retry_delay
configuration are now converted to op RetryPolicies with all make_dagster_*
apis.
- Fixed an issue where cron schedules using a form like
0 5 * * mon#1
to execute on a certain day of the week each month executed every week instead. - [dagit] Fixed an issue where the asset lineage page sometimes timed out while loading large asset graphs.
- Fixed an issue where the partitions page sometimes failed to load for partitioned asset jobs.
- [dagster-airflow] The
use_airflow_template_context
, mock_xcom
and use_ephemeral_airflow_db
params have been dropped, by default all make_dagster_*
apis now use a run-scoped airflow db, similiar to how use_ephemeral_airflow_db
worked. - [dagster-airflow]
make_airflow_dag
has been removed. - [dagster-airflow]
make_airflow_dag_for_operator
has been removed. - [dagster-airflow]
make_airflow_dag_containerized
has been removed. - [dagster-airflow]
airflow_operator_to_op
has been removed. - [dagster-airflow]
make_dagster_repo_from_airflow_dags_path
has been removed. - [dagster-airflow]
make_dagster_repo_from_airflow_dag_bag
has been removed. - [dagster-airflow]
make_dagster_repo_from_airflow_example_dags
has been removed. - [dagster-airflow] The naming convention for ops generated from airflow tasks has been changed to
${dag_id}__${task_id}
from airflow_${task_id}_${unique_int}
. - [dagster-airflow] The naming convention for jobs generated from airflow dags has been changed to
${dag_id}
from airflow_${dag_id}
.
- Definitions now accepts Executor instances in its executor argument, not just ExecutorDefinitions.
@multi_asset_sensor
now accepts a request_assets
parameter, which allows it to directly request that assets be materialized, instead of requesting a run of a job.- Improved the performance of instantiating a
Definitions
when using large numbers of assets or many asset jobs. - The job passed to
build_schedule_from_partitioned_job
no longer needs to have a partitions_def
directly assigned to it. Instead, Dagster will infer from the partitions from the assets it targets. OpExecutionContext.asset_partition_keys_for_output
no longer requires an argument to specify the default output.- The “Reload all” button on the Code Locations page in Dagit will now detect changes to a
pyproject.toml
file that were made while Dagit was running. Previously, Dagit needed to be restarted in order for such changes to be shown. get_run_record_by_id
has been added to DagsterInstance
to provide easier access to RunRecord
objects which expose the start_time
and end_time
of the run.- [dagit] In the “Materialize” modal, you can now choose to pass a range of asset partitions to a single run rather than launching a backfill.
- [dagster-docker] Added a
docker_container_op
op and execute_docker_container_op
helper function for running ops that launch arbitrary Docker containers. See the docs for more information. - [dagster-snowflake-pyspark] The Snowflake I/O manager now supports PySpark DataFrames.
- [dagster-k8s] The Docker images include in the Dagster Helm chart are now built on the most recently released
python:3.x-slim
base image.
- Previously, the
build_asset_reconciliation_sensor
could time out when evaluating ticks over large selections of assets, or assets with many partitions. A series of performance improvements should make this much less likely. - Fixed a bug that caused a failure when using
run_request_for_partition
in a sensor that targeted multiple jobs created via define_asset_job
. - The cost of importing
dagster
has been reduced. - Issues preventing “re-execute from failure” from working correctly with dynamic graphs have been fixed.
- [dagit] In Firefox, Dagit no longer truncates text unnecessarily in some cases.
- [dagit] Dagit’s asset graph now allows you to click “Materialize” without rendering the graph if you have too many assets to display.
- [dagit] Fixed a bug that stopped the backfill page from loading when assets that had previously been backfilled no longer had a
PartitionsDefinition
. - [dagster-k8s] Fixed an issue where
k8s_job_op
raised an Exception when running pods with multiple containers. - [dagster-airbyte] Loosened credentials masking for Airbyte managed ingestion, fixing the Hubspot source, thanks @joel-olazagasti!
- [dagster-airbyte] When using managed ingestion, Airbyte now pulls all source types available to the instance rather than the workspace, thanks @emilija-omnisend!
- [dagster-airbyte] Fixed an issue which arose when attaching freshness policies to Airbyte assets and using the multiprocessing executor.
- [dagster-fivetran] Added the ability to force assets to be output for all specified Fivetran tables during a sync in the case that a sync’s API outputs are missing one or more tables.
- The
asset_keys
and asset_selection
parameters of the experimental @multi_asset_sensor
decorator have been replaced with a monitored_assets
parameter. This helps disambiguate them from the new request_assets
parameter.
- A broken docs link in snowflake_quickstart has been fixed, thanks @clayheaton!
- Troubleshooting help added to helm deployment guide, thanks @adam-bloom!
StaticPartitionMapping
is now serializable, thanks @AlexanderVR!- [dagster-fivetran]
build_fivetran_assets
now supports group_name
, thanks @toddy86! - [dagster-azure]
AzureBlobComputeManager
now supports authentication via DefaultAzureCredential
, thanks @mpicard!
- [dagster-airflow] added a new api
load_assets_from_airflow_dag
that creates graph-backed, partitioned, assets based on the provided Airflow DAG.
- Large asset graphs can now be materialized in Dagit without needing to first enter an asset subset. Previously, if you wanted to materialize every asset in such a graph, you needed to first enter
*
as the asset selection before materializing the assets. - Added a pin of the
sqlalchemy
package to <2.0.0
due to a breaking change in that version. - Added a pin of the
dbt-core
package to <1.4.0
due to breaking changes in that release that affected the Dagster dbt integration. We plan to remove this pin in the next release. - Added a pin of the
jupyter-client
package to <8.0
due to an issue with the most recent release causing hangs while running dagstermill ops.
- Fixed an issue where the Backfills page in Dagit didn't show partition status for some backfills.
- [dagster-aws] Fixed an issue where the
EcsRunLauncher
sometimes waited much longer than intended before retrying after a failure launching a run. - [dagster-mysql] Fixed an issue where some implementations of MySQL storage were raising invalid version errors.