Prerequisites
Confirm these prerequisites before you change the node:- go-livepeer is installed and running as a transcoding orchestrator on Arbitrum mainnet
- the node already serves video workloads successfully
- Docker is installed with
nvidia-container-toolkitenabled - the GPU has at least 4 GB of free VRAM for a small AI pipeline
- the model directory is available at
~/.lpData/models
Start with Install go-livepeer when you are building a node from scratch.
Check your hardware
AI inference runs in a separate Docker container alongside the transcoding process. Shared GPUs divide VRAM between video work and AI workloads, so verify available memory first.- 4 GB:
image-to-text - 6 GB:
segment-anything-2 - 8 GB:
llm - 12 GB:
audio-to-text - 16 GB+:
image-to-video - 20 GB:
image-to-image - 24 GB:
text-to-image
Pull the AI runner image
Pull the default AI runner image before changing the startup command:segment-anything-2:
Configure aiModels.json
Create the file at~/.lpData/aiModels.json:
pipeline: pipeline name such astext-to-image,audio-to-text, orllmmodel_id: Hugging Face model identifierprice_per_unit: integer wei price or supported USD stringwarm: loads the model into VRAM on startupurl: external runner endpoint when the model is hosted outside go-livepeertoken: bearer token for an authenticated external runner
Keep one warm model per GPU for the initial setup. Keep additional models cold until demand justifies them.
Update the startup command
Add three flags to the existing startup command:-aiWorker-aiModels-aiModelsDir
-aiModelsDir must point at the host path. go-livepeer passes that path directly into the runner containers it spawns.Verify AI is active
Within a few seconds of startup, warm models should trigger a managed-container log line:aiModels.jsonis valid JSON- the model weights exist under
-aiModelsDir - the Docker socket is mounted in Docker deployments
images array. After on-chain capability advertisement is configured, the AI pipeline also appears on the node profile in the Livepeer Explorer.
Choose your AI path
The AI runner is active. Choose the workload path you want to specialise in next.Set up batch AI inference
Configure image, audio, and video generation pipelines with model downloads, pricing, and activation guidance.
Set up Cascade AI
Configure ComfyStream for persistent video processing with the GPU allocation needed for live workloads.
Related
- Workload Options - compare workloads before choosing a path
- AI Inference Operations - review advanced
aiModels.jsonoptions, multi-GPU setups, and external runners