Great-expectations

Latest version: v1.2.6

Safety actively analyzes 688463 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 43 of 46

0.7.4

* Fix numerous rendering bugs and formatting issues for rendering documentation.
* Add support for pandas extension dtypes in pandas backend of expect_column_values_to_be_of_type and expect_column_values_to_be_in_type_list and fix bug affecting some dtype-based checks.
* Add datetime and boolean column-type detection in BasicDatasetProfiler.
* Improve BasicDatasetProfiler performance by disabling interactive evaluation when output of expectation is not immediately used for determining next expectations in profile.
* Add support for rendering expectation_suite and expectation_level notes from meta in docs.
* Fix minor formatting issue in readthedocs documentation.

0.7.3

* BREAKING: Harmonize expect_column_values_to_be_of_type and expect_column_values_to_be_in_type_list semantics in
Pandas with other backends, including support for None type and type_list parameters to support profiling.
*These type expectations now rely exclusively on native python or numpy type names.*
* Add configurable support for Custom DataAsset modules to DataContext
* Improve support for setting and inheriting custom data_asset_type names
* Add tooltips with expectations backing data elements to rendered documentation
* Allow better selective disabling of tests (thanks RoyalITS)
* Fix documentation build errors causing missing code blocks on readthedocs
* Update the parameter naming system in DataContext to reflect data_asset_name *and* expectation_suite_name
* Change scary warning about discarding expectations to be clearer, less scary, and only in log
* Improve profiler support for boolean types, value_counts, and type detection
* Allow user to specify data_assets to profile via CLI
* Support CLI rendering of expectation_suite and EVR-based documentation

0.7.2

* Improved error detection and handling in CLI "add datasource" feature
* Fixes in rendering of profiling results (descriptive renderer of validation results)
* Query Generator of SQLAlchemy datasource adds tables in non-default schemas to the data asset namespace
* Added convenience methods to display HTML renderers of sections in Jupyter notebooks
* Implemented prescriptive rendering of expectations for most expectation types

0.7.1

v.0.7.1
------------

* Added documentation/tutorials/videos for onboarding and new profiling and documentation features
* Added prescriptive documentation built from expectation suites
* Improved index, layout, and navigation of data context HTML documentation site
* Bug fix: non-Python files were not included in the package
* Improved the rendering logic to gracefully deal with failed expectations
* Improved the basic dataset profiler to be more resilient
* Implement expect_column_values_to_be_of_type, expect_column_values_to_be_in_type_list for SparkDFDataset
* Updated CLI with a new documentation command and improved profile and render commands
* Expectation suites and validation results within a data context are saved in a more readable form (with indentation)
* Improved compatibility between SparkDatasource and InMemoryGenerator
* Optimization for Pandas column type checking
* Optimization for Spark duplicate value expectation (thanks orenovadia!)
* Default run_id format no longer includes ":" and specifies UTC time
* Other internal improvements and bug fixes

0.7

and a large number of improvements, including breaking API changes.

The core vocabulary of expectations remains consistent. Upgrading to
the new version of GE will primarily require changes to code that
uses data contexts; existing expectation suites will require only changes
to top-level names.

* Major update of Data Contexts. Data Contexts now offer significantly \
more support for building and maintaining expectation suites and \
interacting with existing pipeline systems, including providing a namespace for objects.\
They can handle integrating, registering, and storing validation results, and
provide a namespace for data assets, making **batches** first-class citizens in GE.
Read more: :ref:`data_context` or :py:mod:`great_expectations.data_context`

* Major refactor of autoinspect. Autoinspect is now built around a module
called "profile" which provides a class-based structure for building
expectation suites. There is no longer a default "autoinspect_func" --
calling autoinspect requires explicitly passing the desired profiler. See :ref:`profiling`

* New "Compile to Docs" feature produces beautiful documentation from expectations and expectation
validation reports, helping keep teams on the same page.

* Name clarifications: we've stopped using the overloaded terms "expectations
config" and "config" and instead use "expectation suite" to refer to a
collection (or suite!) of expectations that can be used for validating a
data asset.

- Expectation Suites include several top level keys that are useful \
for organizing content in a data context: data_asset_name, \
expectation_suite_name, and data_asset_type. When a data_asset is \
validated, those keys will be placed in the `meta` key of the \
validation result.

* Major enhancement to the CLI tool including `init`, `render` and more flexibility with `validate`

* Added helper notebooks to make it easy to get started. Each notebook acts as a combination of \
tutorial and code scaffolding, to help you quickly learn best practices by applying them to \
your own data.

* Relaxed constraints on expectation parameter values, making it possible to declare many column
aggregate expectations in a way that is always "vacuously" true, such as
``expect_column_values_to_be_between`` ``None`` and ``None``. This makes it possible to progressively
tighten expectations while using them as the basis for profiling results and documentation.

* Enabled caching on dataset objects by default.

* Bugfixes and improvements:

* New expectations:

* expect_column_quantile_values_to_be_between
* expect_column_distinct_values_to_be_in_set

* Added support for ``head`` method on all current backends, returning a PandasDataset
* More implemented expectations for SparkDF Dataset with optimizations

* expect_column_values_to_be_between
* expect_column_median_to_be_between
* expect_column_value_lengths_to_be_between

* Optimized histogram fetching for SqlalchemyDataset and SparkDFDataset
* Added cross-platform internal partition method, paving path for improved profiling
* Fixed bug with outputstrftime not being honored in PandasDataset
* Fixed series naming for column value counts
* Standardized naming for expect_column_values_to_be_of_type
* Standardized and made explicit use of sample normalization in stdev calculation
* Added from_dataset helper
* Internal testing improvements
* Documentation reorganization and improvements
* Introduce custom exceptions for more detailed error logs

0.7.0

Page 43 of 46

Links

Releases

Has known vulnerabilities

© 2024 Safety CLI Cybersecurity Inc. All Rights Reserved.