Expectation
An Expectation is a verifiable assertion about data. Expectations enhance communication about your data and improve quality for data applications. They help you take the implicit assumptions about your data and make them explicit. Using Expectations helps reduce trips to domain experts and avoids leaving insights about data on the "cutting room floor."
Great Expectations' built-in library includes more than 50 common Expectations, such as:
-
expect_column_values_to_not_be_null
-
expect_column_values_to_match_regex
-
expect_column_values_to_be_unique
-
expect_column_values_to_match_strftime_format
-
expect_table_row_count_to_be_between
-
expect_column_median_to_be_between
For a full list of available Expectations, see the Expectation Gallery.
Great Expectations is a framework for defining Expectations and running them against your data. Like assertions in traditional Python unit tests, Expectations provide a flexible, declarative language for describing expected behavior. Unlike traditional unit tests, Great Expectations applies Expectations to data instead of code. For example, you could define an Expectation that a column contains no null values, and Great Expectations would run that Expectation against your data, and report if a null value was found.
Relationship to other objects
Expectations are grouped into Expectation SuitesA collection of verifiable assertions about data., and are stored in an Expectation StoreA connector to store and retrieve information about collections of verifiable assertions about data.. CheckpointsThe primary means for validating data in a production deployment of Great Expectations. rely on Expectation Suites to ValidateThe act of applying an Expectation Suite to a Batch. data.
Use cases
Expectations are obviously a fundamental component of the Create Expectations step in working with Great Expectations. There are two points at which you will have direct interaction with them. The first is when you are creating new Expectations. The second is when you are editing them. Expectations are not meant to be static: the recommended best practice is an iterative process where your Expectations are edited as your data, and your understanding of that data, change. For further information on this process, please see our overview on the Create Expectations process, and our related how-to guides.
When you create your Checkpoints, you will be able to configure them to use specific Expectation Suites. Other than setting up this configuration (or arranging to pass Expectation Suites at runtime) you will not need to directly interact with the Expectations themselves. Instead, when you run your Checkpoint it will handle the use of the Expectations in any of its Expectation Suites to Validate the data indicated in its Batch Request/s. This will be done under the hood, with the Validation ResultsGenerated when data is Validated against an Expectation or Expectation Suite. that are generated being passed along to the Checkpoint's (optional) ActionsA Python class with a run method that takes a Validation Result and does something with it for further processing.
Customization
Expectations are especially useful when they capture critical aspects of data understanding that analysts and practitioners know based on its semantic meaning. It's common to want to extend Great Expectations with application or domain specific Expectations. For example:
expect_column_text_to_be_in_english
expect_column_value_to_be_valid_icd_code
These Expectations aren't included in the default set, but could be very useful for specific applications.
Fear not! Great Expectations is designed for customization and extensibility.
Building custom Expectations is easy and allows your custom logic to become part of the validation, documentation, and even profiling workflows that make Great Expectations stand out. See the guides on creating custom Expectations for more information on building Expectations and updating Data ContextThe primary entry point for a Great Expectations deployment, with configurations and methods for all supporting components. configurations to automatically load BatchesA selection of records from a Data Asset. of data with custom Data AssetsA collection of records within a Datasource which is usually named based on the underlying data system and sliced to correspond to a desired specification..
Distributional Expectations
Distributional Expectations rely on expected
distributions or "partition objects", which
are built from intervals for continuous data or
categorical classes and their associated weights, in
order to help identify when new datasets or samples
may be different from expected. Distributional
Expectations represent specific Expectation types,
such as
expect_column_kl_divergence_to_be_less_than
. You should use Distributional Expectations in the
same way as other Expectations: to help accelerate
identification of risks and changes to a modeled
system or disruptions to a complex upstream data feed.
For more information, see reference guide on Distributional Expectations.
Conditional Expectations
Conditional Expectations are those that are intended
to be applied not to an entire dataset, but to a
particular subset of the data. Additionally,
Conditional Expectations include those where what one
expects of one variable depends on the value of
another. An example of this would be an Expectation
that a column not have a null value only if another
column's value falls into a specific subset.
Conditional Expectations represent a facet of map
Expectations, including such things as
expect_column_values_to_be_in_set
.
Conditional Expectations are experimentally available for Pandas, Spark, and SQLAlchemy backends.
For more information on these Expectations, please see our reference guide for Conditional Expectations.
Limitations
Unfortunately, not all Expectations are implemented for all source data systems. To view a list of available Expectations, see the Expectations Gallery.
Likewise, Conditional Expectations are considered experimental and may exhibit unexpected behavior when utilized with Spark and SQLAlchemy backends. You can reference our documentation on Conditional Expectations for more information.
Access
You may directly access Expectations as part of the interactive workflow for creating new Expectations. For further details on this process, see guide on how to create and edit Expectations with instant feedback from a sample Batch of data.
Create
Generating Expectations is one of the most important parts of using Great Expectations effectively, and there are a variety of methods for generating and encoding Expectations. When Expectations are encoded in the Great Expectations format, they become shareable and persistent sources of truth about how data was expected to behave-and how it actually did.
There are several paths to generating Expectations:
- Automated inspection of datasets. Data Assistants produce Expectation Suites for data validation. The goal is automated inspection, which can produce Expectations that can be used to validate future batches of data.
- Expertise. Rich experience from subject-matter experts, Analysts, and data owners is often a critical source of Expectations. Interviewing experts and encoding their tacit knowledge of common distributions, values, or failure conditions can be can excellent way to generate Expectations.
- Exploratory Analysis. Using Great Expectations in an exploratory analysis workflow (e.g. within Jupyter Notebooks)is an important way to develop experience with both raw and derived datasets and generate useful and testable Expectations about characteristics that may be important for the data's eventual purpose, whether reporting or feeding another downstream model or data system.
For more information on these methods, see overview guide for creating Expectations.
Configure
Most of the time you will not need to directly interact with an Expectation's configurations. However, advanced users may have circumstances in which it is desirable to define Expectations based purely on domain knowledge, without comparing against underlying data. To do this, you will need to directly write an Expectation's configuration. For details on how to do this, please reference our guide on how to create and edit Expectations based on domain knowledge, without inspecting data directly.
The other occasion when you may want to edit an
Expectation's configuration is when you need to
edit the result_format
of a
Custom ExpectationAn extension of the `Expectation` class,
developed outside of the Great Expectations
library.
The result_format
parameter may be either
a string or a dictionary which specifies the fields to
return in result
. For further details,
please see
our reference guide on the result_format
parameter.
Results
All Expectations return a JSON-serializable dictionary
when evaluated, which consists of four standard
(though optional, depending on the type of Expectation
in question) arguments. These are:
result_format
,
include_config
,
catch_exceptions
, and meta
.
For a more detailed explanation as to what each of
these arguments consists of and which Expectations use
them, please see
our reference guide on standard arguments.
Domain and Success Keys
A domain makes it possible to address a specific set of data, such as a table, query result, column in a table or dataframe, or even a Metric computed on a previous Batch of data. It does this by describing the locale of data to which a Metric or Expectation applies.
A domain is defined by a set of key-value pairs. The
domain keys are the keys that
uniquely define the domain for an Expectation. They
vary depending on the Expectation; for example, many
Expectations apply to data in a single
column
, but others apply to data from
multiple columns or to properties that do not apply to
a column at all.
An Expectation also defines success keys that specify the values of metrics that determine when the Expectation will succeed.
For example, the
expect_column_values_to_be_in_set
Expectation relies on the batch_id
,
table
, column
, and
row_condition
domain keys to determine what data
are described by a particular configuration, and the
value_set
and mostly
success keys to evaluate whether the
Expectation is actually met for that data.
Note: The batch_id and table domain keys are often omitted when running a validation, because the Expectation is being applied to a single batch and table. However, they must be provided in cases where they could be ambiguous.
Metrics use a similar concept: they also use the same kind of domain keys as Expectations, but instead of success keys, we call the keys that determine a Metric's value its value keys.