II.
SkillArea overview
Reference · liveskill-area:llm-infrastructure
LLM Infrastructure overview
Building and operating infrastructure for large language model workloads — GPU cluster management, inference serving frameworks, model routing, token budget management, and multi-tenant LLM gateway design.
Attributes
displayName
LLM Infrastructure
description
Building and operating infrastructure for large language model workloads —
GPU cluster management, inference serving frameworks, model routing, token
budget management, and multi-tenant LLM gateway design.
domains
expertiseLevels
- intermediate
- expert
Outgoing edges
applies_to1
- domain:ml-ops·DomainMLOps
Incoming edges
prerequisite_for_learning1
- skill-area:model-serving·SkillAreaModel Serving
requires_skill_area1
- stack-profile:llm-fine-tuning·StackProfileLLM Fine-Tuning Stack (PyTorch, HuggingFace, PEFT/LoRA, W&B, vLLM)
tool_used_by2
- tool:vllm·ToolvLLM
- tool:triton-inference·ToolTriton Inference Server