Skip to content

serve

Serve your app locally for development and testing. Reads configuration from your project (e.g., pyproject.toml or llama_deploy.yaml) and starts the Python API server, optionally proxying your UI in dev.

See also: Deployment Config Reference and UI build and dev integration.

Terminal window
llamactl serve [DEPLOYMENT_FILE] [options]
  • DEPLOYMENT_FILE defaults to . (current directory). Provide a path to a specific deployment file or directory if needed.
  • --no-install: Skip installing Python and JS dependencies
  • --no-reload: Disable API server auto‑reload on code changes
  • --no-open-browser: Do not open the browser automatically
  • --preview: Build the UI to static files and serve them (production‑like)
  • --port <int>: Port for the API server
  • --ui-port <int>: Port for the UI proxy in dev
  • Prepares the server environment (installs dependencies unless --no-install)
  • In dev mode (default), proxies your UI dev server and reloads on change
  • In preview mode, builds the UI to static files and serves them without a proxy