52 results for “topic:data-contracts”
OpenMetadata is a unified metadata platform for data discovery, data observability, and data governance powered by a central metadata repository, in-depth column level lineage, and seamless team collaboration.
Data Contracts engine for the modern data stack. https://www.soda.io
Home of the Open Data Contract Standard (ODCS).
Code-Native Data Privacy
A curated list of awesome blogs, videos, tools and resources about Data Contracts
Easy and flexible data contracts
A simple and easy to use Data Validation library for Python.
The DBT of ML, as Aligned describes data dependencies in ML systems, and reduce technical data debt
Data policy IN, dynamic view OUT: PACE is the Policy As Code Engine. It helps you to programatically create and apply a data policy to a processing platform like Databricks, Snowflake or BigQuery (or plain 'ol Postgres, even!) with definitions imported from Collibra, Datahub, ODD and the like.
A kedro plugin to use pandera in your kedro projects
DCEE is a lightweight Python framework for validating data against contracts and enforcing SLA rules. Built on pandas and boto3, it provides simple, fast data validation without heavy dependencies.
The Data Store API Specification Repository
Business model representation automation
Open-source data contract enforcement — define, sync dbt, validate, block, report. Built on ODCS v3.1 + DuckDB.
High-performance, API-compatible drop-in replacement for Confluent Schema Registry. Avro, Protobuf & JSON Schema with PostgreSQL, MySQL, Cassandra storage. Built-in RBAC, client side field encryption, data contracts, audit logging. No Kafka dependency. Apache 2.0.
Declarative data quality engine. Define checks in YAML, run anywhere.
Open-source, contract-driven data quality validation. Shift-left enforcement at the point of write — before data enters your pipeline.
Amadeus Dataspace Connector built on top of the Eclipse EDC components to support distributed data exchange ecosystems
Kafka Schema Evolution & Contract Enforcement demo with Avro, Schema Registry and Spring Kafka.
Data quality that just works. 3 lines of code, any data source, 10x faster. Snowflake, Databricks, Fabric, BigQuery, S3, Parquet & 16+ connectors.
Let's be honest - most data pipeline frameworks treat types as suggestions. Config files are strings. Schemas are "validated" at runtime. Data quality is an afterthought. So, let's do differently
Data contract management and drift detection for ML/LLM pipelines -- automatic schema inference, validation, and statistical profiling with CI/CD integration.
Contracts and schemas for a multi-repo knowledge pipeline: events, sessions, summaries, digests, and publishing buses.
This is a repo designed to show the workflow of a data contract, with pre-commit hooks and GitHub Actions on how to have the contract power a data platform
Versioned JSON Schemas for FINULIO data contracts (validation, compatibility, traceability)
Catch bad data between Metaflow steps before it corrupts your pipeline
A production-style e-commerce lakehouse pipeline showcasing Bronze → Silver → Gold transformations with dbt, DuckDB, Polars, BigQuery, and Airflow. Built to demonstrate data contracts, quality gates, spec-driven orchestration, and observability in a modern analytics stack.
PoC for Soda Contracts against Vertica DB
Schema drift detection API for data pipelines
2025-26 DSCI-524 Group 26 - Lightweight Python package for defining, validating, and comparing data contracts for tabular datasets.