Skip to content

MLMP Internal Documentation

Welcome to the MLMP Docs!

Inference Architecture

For details about how inference is implemented (Open WebUI, LiteLLM, Exo, LM Studio, and the Mac Studio topology), see: