Agentic AI Atlasby a5c.ai
OverviewWikiGraphFor AgentsEdgesSearchWorkspace
/
GitHubDocsDiscord
iiRecord
Agentic AI Atlas · Data Quality / Governance Stack (Great Expectations, dbt, Airflow, PostgreSQL, Python)
stack-profile:data-quality-governancea5c.ai
Search record views/
Record · tabs

Available views

II.Record viewspp. 1 - 1
overviewjsongraph
II.
StackProfile overview

stack-profile:data-quality-governance

Reference · live

Data Quality / Governance Stack (Great Expectations, dbt, Airflow, PostgreSQL, Python) overview

A data quality and governance platform that embeds validation checkpoints throughout the data pipeline. dbt models define tested transformations with built-in schema and data tests. Airflow orchestrates pipeline DAGs with quality gates that halt downstream processing on validation failure. Python scripts run custom quality rules — completeness checks, referential integrity, distribution drift detection — and publish results to a PostgreSQL-backed quality dashboard. SQLAlchemy provides the data access layer. Designed for data teams in regulated industries where data quality is auditable and pipeline failures must be traceable to specific validation rules. The tradeoff is pipeline velocity — adding quality gates increases end-to-end latency and requires balancing thoroughness against freshness SLAs for downstream consumers.

StackProfileOutgoing · 20Incoming · 0

Attributes

displayName
Data Quality / Governance Stack (Great Expectations, dbt, Airflow, PostgreSQL, Python)
description
A data quality and governance platform that embeds validation checkpoints throughout the data pipeline. dbt models define tested transformations with built-in schema and data tests. Airflow orchestrates pipeline DAGs with quality gates that halt downstream processing on validation failure. Python scripts run custom quality rules — completeness checks, referential integrity, distribution drift detection — and publish results to a PostgreSQL-backed quality dashboard. SQLAlchemy provides the data access layer. Designed for data teams in regulated industries where data quality is auditable and pipeline failures must be traceable to specific validation rules. The tradeoff is pipeline velocity — adding quality gates increases end-to-end latency and requires balancing thoroughness against freshness SLAs for downstream consumers.
composes
  • language:python
  • language:sql
  • tool:airflow
  • library:sqlalchemy
  • library:pandas
  • library:pydantic

Outgoing edges

applies_to2
  • domain:data-engineering·DomainData Engineering
  • domain:data-science·DomainData Science
composed_of8
  • language:python·LanguagePython
  • language:sql·LanguageSQL
  • tool:airflow·ToolApache Airflow
  • library:sqlalchemy·LibrarySQLAlchemy
  • library:pandas·Librarypandas
  • library:pydantic·LibraryPydantic
  • library:pytest·Librarypytest
  • tool:docker·ToolDocker
follows_workflow2
  • workflow:data-quality-monitoring·WorkflowData Quality Monitoring
  • workflow:dbt-model-review·Workflowdbt Model Review
requires_skill_area5
  • skill-area:data-quality·SkillAreaData Quality
  • skill-area:dbt-modeling·SkillAreadbt Modeling
  • skill-area:etl-pipelines·SkillAreaETL Pipelines
  • skill-area:data-governance·SkillAreaData Governance
  • skill-area:data-lineage·SkillAreaData Lineage
used_by_role3
  • role:data-engineer·RoleData Engineer
  • role:analytics-engineer·RoleAnalytics Engineer
  • role:data-scientist·RoleData Scientist

Incoming edges

None.

Related pages

No related wiki pages for this record.

Shortcuts

Open in graph
Browse node kind