iiRecord
Agentic AI Atlas · agent-evaluation-framework
lib-process:ai-agents-conversational--agent-evaluation-frameworka5c.ai
II.
LibraryProcess JSON

lib-process:ai-agents-conversational--agent-evaluation-framework

Structured · live

agent-evaluation-framework json

Inspect the normalized record payload exactly as the atlas UI reads it.

File · generated-library/processes.yamlCluster · generated-library
Record JSON
{
  "id": "lib-process:ai-agents-conversational--agent-evaluation-framework",
  "_kind": "LibraryProcess",
  "_file": "generated-library/processes.yaml",
  "_cluster": "generated-library",
  "attributes": {
    "displayName": "agent-evaluation-framework",
    "description": "Agent Evaluation Framework Implementation - Comprehensive process for evaluating agent performance including\nsuccess metrics, task completion rates, reasoning quality, tool use accuracy, and LLM-as-judge evaluation.",
    "libraryPath": "library/specializations/ai-agents-conversational/agent-evaluation-framework.js",
    "specialization": "ai-agents-conversational",
    "references": [
      "- LangSmith Evaluation: https://docs.smith.langchain.com/evaluation\n- AgentBench: https://github.com/THUDM/AgentBench\n- LLM-as-Judge: https://arxiv.org/abs/2306.05685"
    ],
    "example": "const result = await orchestrate('specializations/ai-agents-conversational/agent-evaluation-framework', {\n  agentName: 'research-agent',\n  evaluationTypes: ['task-completion', 'reasoning-quality', 'tool-use'],\n  benchmarks: ['AgentBench', 'custom']\n});",
    "usesAgents": [
      "agent-evaluator",
      "test-developer",
      "metrics-developer",
      "llm-judge-developer",
      "benchmark-developer",
      "dashboard-developer"
    ]
  },
  "outgoingEdges": [
    {
      "from": "lib-process:ai-agents-conversational--agent-evaluation-framework",
      "to": "domain:software-engineering",
      "kind": "lib_applies_to_domain",
      "attributes": {
        "weight": 1
      }
    },
    {
      "from": "lib-process:ai-agents-conversational--agent-evaluation-framework",
      "to": "workflow:agent-evaluation-cycle",
      "kind": "lib_implements_workflow",
      "attributes": {
        "weight": 1
      }
    },
    {
      "from": "lib-process:ai-agents-conversational--agent-evaluation-framework",
      "to": "workflow:agent-evaluation-cycle",
      "kind": "lib_implements_workflow",
      "attributes": {
        "weight": 0.7
      }
    },
    {
      "from": "lib-process:ai-agents-conversational--agent-evaluation-framework",
      "to": "specialization:ai-agents-conversational",
      "kind": "lib_belongs_to_specialization",
      "attributes": {
        "weight": 0.9
      }
    },
    {
      "from": "lib-process:ai-agents-conversational--agent-evaluation-framework",
      "to": "lib-agent:ai-agents-conversational--agent-evaluator",
      "kind": "uses_agent",
      "attributes": {
        "weight": 0.8
      }
    }
  ],
  "incomingEdges": []
}