LogoLogo
GitHubSlackDataCebo
  • SDMetrics
  • Getting Started
    • Installation
    • Quickstart
    • Metadata
      • Single Table Metadata
      • Multi Table Metadata
      • Sequential Metadata
  • Reports
    • Quality Report
      • What's included?
      • Single Table API
      • Multi Table API
    • Diagnostic Report
      • What's included?
      • Single Table API
      • Multi Table API
    • Other Reports
    • Visualization Utilities
  • Metrics
    • Diagnostic Metrics
      • BoundaryAdherence
      • CardinalityBoundaryAdherence
      • CategoryAdherence
      • KeyUniqueness
      • ReferentialIntegrity
      • TableStructure
    • Quality Metrics
      • CardinalityShapeSimilarity
      • CategoryCoverage
      • ContingencySimilarity
      • CorrelationSimilarity
      • KSComplement
      • MissingValueSimilarity
      • RangeCoverage
      • SequenceLengthSimilarity
      • StatisticMSAS
      • StatisticSimilarity
      • TVComplement
    • Privacy Metrics
      • DCRBaselineProtection
      • DCROverfittingProtection
      • DisclosureProtection
      • DisclosureProtectionEstimate
      • CategoricalCAP
    • ML Augmentation Metrics
      • BinaryClassifierPrecisionEfficacy
      • BinaryClassifierRecallEfficacy
    • Metrics in Beta
      • CSTest
      • Data Likelihood
        • BNLikelihood
        • BNLogLikelihood
        • GMLikelihood
      • Detection: Sequential
      • Detection: Single Table
      • InterRowMSAS
      • ML Efficacy: Sequential
      • ML Efficacy: Single Table
        • Binary Classification
        • Multiclass Classification
        • Regression
      • NewRowSynthesis
      • * OutlierCoverage
      • Privacy Against Inference
      • * SmoothnessSimilarity
  • Resources
    • Citation
    • Contributions
      • Defining your metric
      • Development
      • Release FAQs
    • Enterprise
      • Domain Specific Reports
    • Blog
Powered by GitBook
On this page
  • Data Compatibility
  • Score
  • How does it work?
  • Usage
  • FAQs
  • References
  1. Metrics
  2. Quality Metrics

ContingencySimilarity

PreviousCategoryCoverageNextCorrelationSimilarity

Last updated 2 months ago

This metric computes the similarity of a pair of categorical columns between the real and synthetic datasets -- aka it compares 2D distributions.

Data Compatibility

  • Categorical: This metric is meant for discrete, categorical data

  • Boolean: This metric works well on boolean data

  • Numerical: This metric discretizes numerical data into bins

  • Datetime: This metric discretizes continuous datetime values into bins

To use this metric, both of the columns must be compatible. If there are missing values in the columns, the metric will treat them as an additional, single category.

Score

(best) 1.0: The contingency table is exactly the same between the real vs. synthetic data

(worst) 0.0: The contingency table is as different as can be

The plots below show an example of fictitious real and synthetic data with ContingencySimilarity=0.92.

How does it work?

For a pair of columns, A and B, the test computes a normalized contingency table [1] for the real and synthetic data. This table describes the proportion of rows that have each combination of categories in A and B.

Then, it computes the difference between the contingency tables using the Total Variation Distance [2]. Finally, we subtract the distance from 1 to ensure that a high score means high similarity. The process is summarized by the formula below.

In the formula, α describes all the possible categories in column A and β describes all the possible categories in column B. Meanwhile, R and S refer to the real and synthetic frequencies for those categories.

Usage

To manually run this metric, access the column_pairs module and use the compute method.

from sdmetrics.column_pairs import ContingencySimilarity

ContingencySimilarity.compute(
    real_data=real_table[['column_1', 'column_2']],
    synthetic_data=synthetic_table[['column_1', 'column_2']]
)

Parameters

  • (required) real_data: A pandas.DataFrame object containing 2 columns of real data

  • (required) synthetic_data: A pandas.DataFrame object containing 2 columns of synthetic data

  • num_rows_subsample: The number of rows to subsample before running this metric. Use this option to get an estimate of the full ContingencySimilarity score with faster performance.

    • (default) None: Do not subsample the rows. Use the full dataset to compute the score.

    • <integer>: Randomly subsample the provided number of rows for both the real and the synthetic datasets before computing the metric

  • continuous_column_names: A list of column names that represent continuous values. Such columns will be discretized into bins before applying this metric.

    • (default) None: None of the columns are continuous

    • [<column names>]: Each column name in the list will be discretized into bins before applying this metric.

  • num_discrete_bins: The number of discrete bins to create for continuous columns

    • (default) 10: Discretize continuous columns into 10 bins

    • <int>: Discretize continuous columns into the number of bins provided

FAQs

Is there an equivalent to this metric for numerical columns?

If you want to compute a similarity between 1 numerical and 1 categorical column, provide the column names in the continuous_column_names parameter. This discretizes the values into histogram bins, and then treats the column as categorical. Note that this approach will no longer factor in the order in the numerical values.

Can you compare trends between 3 or more columns?

Currently, the SDMetrics library does not support any explicit similarity metrics for higher order trends.

How can I speed up this metric?

This metric computes a full contingency table for all combinations of values in the 2 columns, for the real and synthetic data. This process may take some time if your datasets are large and if there are a large number of possible values.

To speed up the metric, we recommend using the num_rows_subsample parameter that subsamples the real and synthetic data before computing the contingency table. Ultimately, this score will be an estimate of the true score (based on the whole data), but in practice, we do not see significant changes if you make sure to keep a couple thousand rows.

An additional way to speed up the metric is, if you have continuous columns, decrease the number of discrete bins to limit the size of the contingency table. Note that this may increase your score because you are counting larger variations of numbers as a single bin.

Technical Notes: Association

It is possible to use an association measure to quantify whether a contingency table is biased towards certain categories. There are many ways to compute association, a common coefficient being Cramer's V [3].

The association score measures the degree of bias but not its direction. This can lead to misleading results when comparing real and synthetic data. In the example below, both the real and synthetic data have the same, high association score even though they are biased towards different categories. If we simply compared association scores, we would erroneously conclude that the similarity is high, at 1.0.

This is why SDMetrics does not use association metrics to compare real and synthetic data.

References

score=1−12∑α∈A∑β∈B∣Sα,β−Rα,β∣score = 1 - \frac{1}{2}\sum_{\alpha \in A}\sum_{\beta \in B} |S_{\alpha, \beta} - R_{\alpha, \beta}|score=1−21​α∈A∑​β∈B∑​∣Sα,β​−Rα,β​∣

Recommended Usage: The applies this metric to every pair of compatible columns and provides visualizations to understand the score.

If you want to compute the similarity between two numerical columns, use the metric.

You may be interested in browsing through the experimental or metrics, which can factor in 3 or more columns when computing their score.

[1]

[2]

[3]

Quality Report
CorrelationSimilarity
ML efficacy
detection
https://en.wikipedia.org/wiki/Contingency_table
https://en.wikipedia.org/wiki/Total_variation_distance_of_probability_measures
https://en.wikipedia.org/wiki/Cram%C3%A9r%27s_V
In this contingency table, a categorical column describing a country (vertical) is compared with a boolean column describing whether a user is subscribed (horizontal). The real and synthetic data have similar breakdowns for each (country, subscribed) combination, so the contingency similarity is high at 0.92. They are not exactly the same, so the score is <1.
The association of the real and synthetic data are both high because there is a significant bias in the categories. In the real data, subscribed users are biased to be in the US but in the synthetic data, they are biased to be in Mexico.