II.
LibraryProcess JSON
Structured · livelib-process:gpu-programming--ml-inference-optimization
specializations/gpu-programming/ml-inference-optimization json
Inspect the normalized record payload exactly as the atlas UI reads it.
{
"id": "lib-process:gpu-programming--ml-inference-optimization",
"_kind": "LibraryProcess",
"_file": "generated-library/processes.yaml",
"_cluster": "generated-library",
"attributes": {
"displayName": "specializations/gpu-programming/ml-inference-optimization",
"description": "Machine Learning Inference Optimization - Workflow for optimizing GPU-accelerated ML model\ninference for production deployment, covering quantization, batching, and kernel fusion.",
"libraryPath": "library/specializations/gpu-programming/ml-inference-optimization.js",
"specialization": "gpu-programming",
"references": [
"- TensorRT Documentation: https://docs.nvidia.com/deeplearning/tensorrt/",
"- ONNX Runtime: https://onnxruntime.ai/",
"- Quantization: https://pytorch.org/docs/stable/quantization.html"
],
"example": "const result = await orchestrate('specializations/gpu-programming/ml-inference-optimization', {\n modelName: 'resnet50',\n framework: 'pytorch',\n targetLatency: 5,\n quantization: 'int8'\n});",
"usesAgents": [
"ml-inference-optimizer"
]
},
"outgoingEdges": [
{
"from": "lib-process:gpu-programming--ml-inference-optimization",
"to": "skill-area:cuda-kernels",
"kind": "lib_requires_skill_area",
"attributes": {
"weight": 1
}
},
{
"from": "lib-process:gpu-programming--ml-inference-optimization",
"to": "skill-area:compute-shaders",
"kind": "lib_requires_skill_area",
"attributes": {
"weight": 0.7
}
},
{
"from": "lib-process:gpu-programming--ml-inference-optimization",
"to": "domain:scientific-computing",
"kind": "lib_applies_to_domain",
"attributes": {
"weight": 1
}
},
{
"from": "lib-process:gpu-programming--ml-inference-optimization",
"to": "role:computational-scientist",
"kind": "lib_involves_role",
"attributes": {
"weight": 1
}
},
{
"from": "lib-process:gpu-programming--ml-inference-optimization",
"to": "role:ml-engineer",
"kind": "lib_involves_role",
"attributes": {
"weight": 0.7
}
},
{
"from": "lib-process:gpu-programming--ml-inference-optimization",
"to": "workflow:ml-model-lifecycle",
"kind": "lib_implements_workflow",
"attributes": {
"weight": 1
}
},
{
"from": "lib-process:gpu-programming--ml-inference-optimization",
"to": "specialization:gpu-programming",
"kind": "lib_belongs_to_specialization",
"attributes": {
"weight": 1
}
},
{
"from": "lib-process:gpu-programming--ml-inference-optimization",
"to": "lib-agent:gpu-programming--ml-inference-optimizer",
"kind": "uses_agent",
"attributes": {
"weight": 0.8
}
}
],
"incomingEdges": []
}