Skip to content

edgible ai

Higher-level helpers for running a local LLM (Ollama) and exposing it as an Edgible application. Convenient for getting a private API endpoint to a model you’re hosting yourself.

End-to-end setup: install Ollama, pull a model, optionally install Open WebUI, optionally expose either as Edgible applications.

Terminal window
edgible ai setup --model llama3 --auto-install
edgible ai setup --model llama3 --expose-ollama --setup-webui
FlagDescription
--model <name>Model to pull (e.g. llama3, mistral).
--auto-installInstall Ollama if missing.
--local-onlyDon’t create Edgible applications.
--expose-ollamaPublish the Ollama API as an application.
--setup-webuiInstall Open WebUI on the device.
--device-id <id>Device to place on (defaults to local).
--ollama-device-id <id>Override the placement of just the Ollama workload.
--webui-device-id <id>Override the placement of just the WebUI workload.
--gateway-ids <ids>Pin to specific gateways.
--webui-deployment <type>How WebUI runs (docker or managed-process).
--non-interactiveFail rather than prompt.

Stop Ollama and any UI you started with setup.

Terminal window
edgible ai stop

Start Open WebUI locally pointing at a (possibly different) Ollama URL.

Terminal window
edgible ai serve --port 3200 --ollama-url http://localhost:11434
FlagDescription
--port <n>Port WebUI listens on. Default 3200.
--ollama-url <url>Where to reach Ollama.
-d, --detachedRun in the background. Default true.

Show whether Ollama and WebUI are running.

Terminal window
edgible ai status

Run a one-shot prompt against Ollama and print the response.

Terminal window
edgible ai test --model llama3

Stop services and remove Edgible applications.

Terminal window
edgible ai teardown --stop-ollama --remove-volumes
FlagDescription
--stop-ollamaAlso stop Ollama on the device.
--remove-volumesAlso delete model storage volumes.