Dagster

Latest version: v1.8.7

Safety actively analyzes 663899 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 38 of 49

0.10.1

Not secure
**Community Contributions**

- Reduced image size of `k8s-example` by 25% (104 MB) (thanks alex-treebeard and mrdavidlaing!)
- [dagster-snowflake] `snowflake_resource` can now be configured to use the SQLAlchemy connector (thanks basilvetas!)

**New**

- When setting `userDeployments.deployments` in the Helm chart, `replicaCount` now defaults to 1 if not specified.

**Bugfixes**

- Fixed an issue where the Dagster daemon process couldn’t launch runs in repository locations containing more than one repository.
- Fixed an issue where Helm chart was not correctly templating `env`, `envConfigMaps`, and `envSecrets`.

**Documentation**

- Added new [troubleshooting guide](https://legacy-docs.dagster.io/troubleshooting) for problems encountered while using the `QueuedRunCoordinator` to limit run concurrency.
- Added documentation for the sensor command-line interface.

0.10.0

Not secure
Major Changes

- A **native scheduler** with support for exactly-once, fault tolerant, timezone-aware scheduling.
A new Dagster daemon process has been added to manage your schedules and sensors with a
reconciliation loop, ensuring that all runs are executed exactly once, even if the Dagster daemon
experiences occasional failure. See the
[Migration Guide](https://github.com/dagster-io/dagster/blob/master/MIGRATION.md) for
instructions on moving from `SystemCronScheduler` or `K8sScheduler` to the new scheduler.
- **First-class sensors**, built on the new Dagster daemon, allow you to instigate runs based on
changes in external state - for example, files on S3 or assets materialized by other Dagster
pipelines. See the [Sensors Overview](http://docs.dagster.io/overview/schedules-sensors/sensors)
for more information.
- Dagster now supports **pipeline run queueing**. You can apply instance-level run concurrency
limits and prioritization rules by adding the QueuedRunCoordinator to your Dagster instance. See
the [Run Concurrency Overview](http://docs.dagster.io/overview/pipeline-runs/limiting-run-concurrency)
for more information.
- The `IOManager` abstraction provides a new, streamlined primitive for granular control over where
and how solid outputs are stored and loaded. This is intended to replace the (deprecated)
intermediate/system storage abstractions, See the
[IO Manager Overview](http://docs.dagster.io/overview/io-managers/io-managers) for more
information.
- A new **Partitions page** in Dagit lets you view your your pipeline runs organized by partition.
You can also **launch backfills from Dagit** and monitor them from this page.
- A new **Instance Status page** in Dagit lets you monitor the health of your Dagster instance,
with repository location information, daemon statuses, instance-level schedule and sensor
information, and linkable instance configuration.
- **Resources can now declare their dependencies on other resources** via the
`required_resource_keys` parameter on `resource`.
- Our support for deploying on **Kubernetes** is now mature and battle-tested Our Helm chart is
now easier to configure and deploy, and we’ve made big investments in observability and
reliability. You can view Kubernetes interactions in the structured event log and use Dagit to
help you understand what’s happening in your deployment. The defaults in the Helm chart will
give you graceful degradation and failure recovery right out of the box.
- Experimental support for **dynamic orchestration** with the new `DynamicOutputDefinition` API.
Dagster can now map the downstream dependencies over a dynamic output at runtime.

Breaking Changes

**Dropping Python 2 support**

- We’ve dropped support for Python 2.7, based on community usage and enthusiasm for Python 3-native
public APIs.

**Removal of deprecated APIs**

_These APIs were marked for deprecation with warnings in the 0.9.0 release, and have been removed in
the 0.10.0 release._

- The decorator `input_hydration_config` has been removed. Use the `dagster_type_loader` decorator
instead.
- The decorator `output_materialization_config` has been removed. Use `dagster_type_materializer`
instead.
- The system storage subsystem has been removed. This includes `SystemStorageDefinition`,
`system_storage`, and `default_system_storage_defs` . Use the new `IOManagers` API instead. See
the [IO Manager Overview](http://docs.dagster.io/overview/io-managers/io-managers) for more
information.
- The `config_field` argument on decorators and definitions classes has been removed and replaced
with `config_schema`. This is a drop-in rename.
- The argument `step_keys_to_execute` to the functions `reexecute_pipeline` and
`reexecute_pipeline_iterator` has been removed. Use the `step_selection` argument to select
subsets for execution instead.
- Repositories can no longer be loaded using the legacy `repository` key in your `workspace.yaml`;
use `load_from` instead. See the
[Workspaces Overview](https://docs.dagster.io/concepts/code-locations/workspace-files) for
documentation about how to define a workspace.

**Breaking API Changes**

- `SolidExecutionResult.compute_output_event_dict` has been renamed to
`SolidExecutionResult.compute_output_events_dict`. A solid execution result is returned from
methods such as `result_for_solid`. Any call sites will need to be updated.
- The `.compute` suffix is no longer applied to step keys. Step keys that were previously named
`my_solid.compute` will now be named `my_solid`. If you are using any API method that takes a
step_selection argument, you will need to update the step keys accordingly.
- The `pipeline_def` property has been removed from the `InitResourceContext` passed to functions
decorated with `resource`.

**Dagstermill**

- If you are using `define_dagstermill_solid` with the `output_notebook` parameter set to `True`,
you will now need to provide a file manager resource (subclass of
`dagster.core.storage.FileManager`) on your pipeline mode under the resource key `"file_manager"`,
e.g.:

python
from dagster import ModeDefinition, local_file_manager, pipeline
from dagstermill import define_dagstermill_solid

my_dagstermill_solid = define_dagstermill_solid("my_dagstermill_solid", output_notebook=True, ...)

pipeline(mode_defs=[ModeDefinition(resource_defs={"file_manager": local_file_manager})])
def my_dagstermill_pipeline():
my_dagstermill_solid(...)


**Helm Chart**

- The schema for the `scheduler` values in the helm chart has changed. Instead of a simple toggle
on/off, we now require an explicit `scheduler.type` to specify usage of the
`DagsterDaemonScheduler`, `K8sScheduler`, or otherwise. If your specified `scheduler.type` has
required config, these fields must be specified under `scheduler.config`.
- `snake_case` fields have been changed to `camelCase`. Please update your `values.yaml` as follows:
- `pipeline_run` → `pipelineRun`
- `dagster_home` → `dagsterHome`
- `env_secrets` → `envSecrets`
- `env_config_maps` → `envConfigMaps`
- The Helm values `celery` and `k8sRunLauncher` have now been consolidated under the Helm value
`runLauncher` for simplicity. Use the field `runLauncher.type` to specify usage of the
`K8sRunLauncher`, `CeleryK8sRunLauncher`, or otherwise. By default, the `K8sRunLauncher` is
enabled.
- All Celery message brokers (i.e. RabbitMQ and Redis) are disabled by default. If you are using
the `CeleryK8sRunLauncher`, you should explicitly enable your message broker of choice.
- `userDeployments` are now enabled by default.

Core

- Event log messages streamed to `stdout` and `stderr` have been streamlined to be a single line
per event.
- Experimental support for memoization and versioning lets you execute pipelines incrementally,
selecting which solids need to be rerun based on runtime criteria and versioning their outputs
with configurable identifiers that capture their upstream dependencies.

To set up memoized step selection, users can provide a `MemoizableIOManager`, whose `has_output`
function decides whether a given solid output needs to be computed or already exists. To execute
a pipeline with memoized step selection, users can supply the `dagster/is_memoized_run` run tag
to `execute_pipeline`.

To set the version on a solid or resource, users can supply the `version` field on the definition.
To access the derived version for a step output, users can access the `version` field on the
`OutputContext` passed to the `handle_output` and `load_input` methods of `IOManager` and the
`has_output` method of `MemoizableIOManager`.

- Schedules that are executed using the new `DagsterDaemonScheduler` can now execute in any
timezone by adding an `execution_timezone` parameter to the schedule. Daylight Savings Time
transitions are also supported. See the
[Schedules Overview](http://docs.dagster.io/overview/schedules-sensors/schedules#timezones) for
more information and examples.

Dagit

- Countdown and refresh buttons have been added for pages with regular polling queries (e.g. Runs,
Schedules).
- Confirmation and progress dialogs are now presented when performing run terminations and
deletions. Additionally, hanging/orphaned runs can now be forced to terminate, by selecting
"Force termination immediately" in the run termination dialog.
- The Runs page now shows counts for "Queued" and "In progress" tabs, and individual run pages
show timing, tags, and configuration metadata.
- The backfill experience has been improved with means to view progress and terminate the entire
backfill via the partition set page. Additionally, errors related to backfills are now surfaced
more clearly.
- Shortcut hints are no longer displayed when attempting to use the screen capture command.
- The asset page has been revamped to include a table of events and enable organizing events by
partition. Asset key escaping issues in other views have been fixed as well.
- Miscellaneous bug fixes, frontend performance tweaks, and other improvements are also included.

Kubernetes/Helm

- The [Dagster Kubernetes documentation](https://legacy-docs.dagster.io/deploying/kubernetes) has been refreshed.

**Helm**

- We've added schema validation to our Helm chart. You can now check that your values YAML file is
correct by running:

bash
helm lint helm/dagster -f helm/dagster/values.yaml


- Added support for resource annotations throughout our Helm chart.
- Added Helm deployment of the dagster daemon & daemon scheduler.
- Added Helm support for configuring a compute log manager in your dagster instance.
- User code deployments now include a user `ConfigMap` by default.
- Changed the default liveness probe for Dagit to use `httpGet "/dagit_info"` instead of
`tcpSocket:80`

**Dagster-K8s [Kubernetes]**

- Added support for user code deployments on Kubernetes.
- Added support for tagging pipeline executions.
- Fixes to support version 12.0.0 of the Python Kubernetes client.
- Improved implementation of Kubernetes+Dagster retries.
- Many logging improvements to surface debugging information and failures in the structured event
log.

**Dagster-Celery-K8s**

- Improved interrupt/termination handling in Celery workers.

Integrations & Libraries

- Added a new `dagster-docker` library with a `DockerRunLauncher` that launches each run in its own
Docker container. (See [Deploying with Docker docs](https://docs.dagster.io/examples/deploy_docker)
for an example.)
- Added support for AWS Athena. (Thanks jmsanders!)
- Added mocks for AWS S3, Athena, and Cloudwatch in tests. (Thanks jmsanders!)
- Allow setting of S3 endpoint through env variables. (Thanks marksteve!)
- Various bug fixes and new features for the Azure, Databricks, and Dask integrations.
- Added a `create_databricks_job_solid` for creating solids that launch Databricks jobs.

0.9.22.post0

Not secure
**Bugfixes**

- [Dask] Pin dask[dataframe] to <=2.30.0 and distributed to <=2.30.1

0.9.22

Not secure
**New**

- When using a solid selection in the Dagit Playground, non-matching solids are hidden in the
RunPreview panel.
- The CLI command dagster pipeline launch now accepts --run-id

**Bugfixes**

- [Helm/K8s] Fixed whitespacing bug in ingress.yaml Helm template.

0.9.21

Not secure
**Community Contributions**

- Fixed helm chart to only add flower to the K8s ingress when enabled (thanks [PenguinToast](https://github.com/PenguinToast)!)
- Updated helm chart to use more lenient timeouts for liveness probes on user code deployments (thanks [PenguinToast](https://github.com/PenguinToast)!)

**Bugfixes**

- [Helm/K8s] Due to Flower being incompatible with Celery 5.0, the Helm chart for Dagster now uses a specific image `mher/flower:0.9.5` for the Flower pod.

0.9.20

Not secure
**New**

- [Dagit] Show recent runs on individual schedule pages
- [Dagit] It’s no longer required to run `dagster schedule up` or press the Reconcile button before turning on a new schedule for the first time
- [Dagit] Various improvements to the asset view. Expanded the Last Materialization Event view. Expansions to the materializations over time view, allowing for both a list view and a graphical view of materialization data.

**Community Contributions**

- Updated many dagster-aws tests to use mocked resources instead of depending on real cloud resources, making it possible to run these tests locally. (thanks jmsanders!)

**Bugfixes**

- fixed an issue with retries in step launchers
- [Dagit] bugfixes and improvements
- Fixed an issue where dagit sometimes left hanging processes behind after exiting

**Experimental**

- [K8s] The dagster daemon is now optionally deployed by the helm chart. This enables run-level queuing with the QueuedRunCoordinator.

Page 38 of 49

Links

Releases

Has known vulnerabilities

© 2024 Safety CLI Cybersecurity Inc. All Rights Reserved.