stack-profile:data-quality-governance
Data Quality / Governance Stack (Great Expectations, dbt, Airflow, PostgreSQL, Python) overview
A data quality and governance platform that embeds validation checkpoints throughout the data pipeline. dbt models define tested transformations with built-in schema and data tests. Airflow orchestrates pipeline DAGs with quality gates that halt downstream processing on validation failure. Python scripts run custom quality rules — completeness checks, referential integrity, distribution drift detection — and publish results to a PostgreSQL-backed quality dashboard. SQLAlchemy provides the data access layer. Designed for data teams in regulated industries where data quality is auditable and pipeline failures must be traceable to specific validation rules. The tradeoff is pipeline velocity — adding quality gates increases end-to-end latency and requires balancing thoroughness against freshness SLAs for downstream consumers.
Attributes
Outgoing edges
- domain:data-engineering·DomainData Engineering
- domain:data-science·DomainData Science
- language:python·LanguagePython
- language:sql·LanguageSQL
- tool:airflow·ToolApache Airflow
- library:sqlalchemy·LibrarySQLAlchemy
- library:pandas·Librarypandas
- library:pydantic·LibraryPydantic
- library:pytest·Librarypytest
- tool:docker·ToolDocker
- workflow:data-quality-monitoring·WorkflowData Quality Monitoring
- workflow:dbt-model-review·Workflowdbt Model Review
- skill-area:data-quality·SkillAreaData Quality
- skill-area:dbt-modeling·SkillAreadbt Modeling
- skill-area:etl-pipelines·SkillAreaETL Pipelines
- skill-area:data-governance·SkillAreaData Governance
- skill-area:data-lineage·SkillAreaData Lineage
- role:data-engineer·RoleData Engineer
- role:analytics-engineer·RoleAnalytics Engineer
- role:data-scientist·RoleData Scientist