iiRecord
Agentic AI Atlas · model-evaluation
lib-process:data-science-ml--model-evaluationa5c.ai
II.
LibraryProcess JSON

lib-process:data-science-ml--model-evaluation

Structured · live

model-evaluation json

Inspect the normalized record payload exactly as the atlas UI reads it.

File · generated-library/processes.yamlCluster · generated-library
Record JSON
{
  "id": "lib-process:data-science-ml--model-evaluation",
  "_kind": "LibraryProcess",
  "_file": "generated-library/processes.yaml",
  "_cluster": "generated-library",
  "attributes": {
    "displayName": "model-evaluation",
    "description": "Model Evaluation and Validation Framework - Comprehensive model assessment across multiple dimensions\nincluding performance metrics, robustness testing, fairness analysis, explainability, and production readiness checks\nwith iterative validation loops and quality gates.",
    "libraryPath": "library/specializations/data-science-ml/model-evaluation.js",
    "specialization": "data-science-ml",
    "references": [
      "- Google ML Testing: https://developers.google.com/machine-learning/testing-debugging\n- Model Cards: https://arxiv.org/abs/1810.03993\n- Fairness Indicators: https://www.tensorflow.org/responsible_ai/fairness_indicators/guide\n- SHAP (SHapley Additive exPlanations): https://github.com/slundberg/shap\n- Model Validation Best Practices: https://cloud.google.com/architecture/mlops-continuous-delivery-and-automation-pipelines-in-machine-learning"
    ],
    "example": "const result = await orchestrate('specializations/data-science-ml/model-evaluation', {\n  modelPath: 'models/trained/churn-predictor-v2.pkl',\n  testDataPath: 'data/test/churn_test.csv',\n  modelType: 'classification',\n  targetMetrics: { accuracy: 0.85, f1_score: 0.80, auc_roc: 0.88 },\n  validationLevel: 'comprehensive',\n  fairnessAttributes: ['age_group', 'gender', 'region'],\n  explainabilityRequired: true\n});",
    "usesAgents": [
      "general-purpose"
    ]
  },
  "outgoingEdges": [
    {
      "from": "lib-process:data-science-ml--model-evaluation",
      "to": "domain:data-science",
      "kind": "lib_applies_to_domain",
      "attributes": {
        "weight": 1
      }
    },
    {
      "from": "lib-process:data-science-ml--model-evaluation",
      "to": "role:data-scientist",
      "kind": "lib_involves_role",
      "attributes": {
        "weight": 1
      }
    },
    {
      "from": "lib-process:data-science-ml--model-evaluation",
      "to": "workflow:data-pipeline-deployment",
      "kind": "lib_implements_workflow",
      "attributes": {
        "weight": 1
      }
    },
    {
      "from": "lib-process:data-science-ml--model-evaluation",
      "to": "specialization:data-science-ml",
      "kind": "lib_belongs_to_specialization",
      "attributes": {
        "weight": 0.9
      }
    }
  ],
  "incomingEdges": []
}