II.
Workflow JSON
Structured · liveworkflow:prompt-engineering-iteration
Prompt Engineering Iteration json
Inspect the normalized record payload exactly as the atlas UI reads it.
{
"id": "workflow:prompt-engineering-iteration",
"_kind": "Workflow",
"_file": "domain/workflows/workflows-technical-depth.yaml",
"_cluster": "domain",
"attributes": {
"displayName": "Prompt Engineering Iteration",
"description": "Focused iteration loop for systematically improving the quality and reliability of\nprompts powering AI-driven product features. The ML engineer hypothesises changes\nto the system prompt, few-shot examples, or instruction structure, then runs an\nautomated evaluation suite to measure impact against quality metrics. A version-\ncontrolled prompt registry tracks all variants and their eval scores. Winning\ncandidates graduate to a shadow-traffic A/B test in production, and only prompts\nthat improve user-facing metrics are promoted to the live configuration.\n",
"workflowKind": "development",
"triggerType": "on-demand",
"typicalCadence": "per-sprint",
"complexity": "moderate"
},
"outgoingEdges": [
{
"from": "workflow:prompt-engineering-iteration",
"to": "role:ml-engineer",
"kind": "involves_role"
},
{
"from": "workflow:prompt-engineering-iteration",
"to": "role:backend-engineer",
"kind": "involves_role"
},
{
"from": "workflow:prompt-engineering-iteration",
"to": "role:product-manager",
"kind": "involves_role"
},
{
"from": "workflow:prompt-engineering-iteration",
"to": "domain:software-engineering",
"kind": "applies_to_domain"
}
],
"incomingEdges": [
{
"from": "stack-profile:agentic-rag",
"to": "workflow:prompt-engineering-iteration",
"kind": "follows_workflow"
},
{
"from": "stack-profile:voice-ai-agent",
"to": "workflow:prompt-engineering-iteration",
"kind": "follows_workflow"
},
{
"from": "stack-profile:ai-code-review-pipeline",
"to": "workflow:prompt-engineering-iteration",
"kind": "follows_workflow"
},
{
"from": "stack-profile:autonomous-agent-fleet",
"to": "workflow:prompt-engineering-iteration",
"kind": "follows_workflow"
},
{
"from": "stack-profile:prompt-engineering-workbench",
"to": "workflow:prompt-engineering-iteration",
"kind": "follows_workflow"
},
{
"from": "stack-profile:ai-safety-guardrails",
"to": "workflow:prompt-engineering-iteration",
"kind": "follows_workflow"
},
{
"from": "tool:anthropic",
"to": "workflow:prompt-engineering-iteration",
"kind": "supports_work",
"attributes": {
"confidence": "high",
"evidence": "Claude prompt and tool-use behavior is directly relevant to prompt iteration."
}
},
{
"from": "tool:openai",
"to": "workflow:prompt-engineering-iteration",
"kind": "supports_work",
"attributes": {
"confidence": "high",
"evidence": "Prompt and response behavior are direct inputs to prompt iteration."
}
},
{
"from": "lib-process:ai-agents-conversational--conversation-quality-testing",
"to": "workflow:prompt-engineering-iteration",
"kind": "lib_implements_workflow",
"attributes": {
"weight": 0.5
}
},
{
"from": "lib-process:ai-agents-conversational--prompt-engineering-workflow",
"to": "workflow:prompt-engineering-iteration",
"kind": "lib_implements_workflow",
"attributes": {
"weight": 0.7
}
}
]
}