seperate routing for llama & stack

This commit is contained in:
Bryan Ramos 2026-04-14 00:41:39 -04:00
parent 3feb5ddc6b
commit 46adf8e9f0

View file

@ -140,15 +140,16 @@ in
useACMEHost = domain; useACMEHost = domain;
forceSSL = true; forceSSL = true;
# Web UI — llama.cpp chat interface (browser) # Web UI + llama.cpp API (browser, /v1/* calls from the UI)
# Auth handled by llama.cpp itself (--api-key flag) # Auth handled by llama.cpp itself (--api-key flag)
locations."/" = { locations."/" = {
proxyPass = "http://192.168.0.23:8000"; proxyPass = "http://192.168.0.23:8000";
proxyWebsockets = true; proxyWebsockets = true;
}; };
# API — Llama Stack (opencode, programmatic clients) # Llama Stack API (opencode, programmatic clients)
locations."/v1/" = { # Clients use baseURL: https://ai.ramos.codes/stack/v1
locations."/stack/v1/" = {
proxyPass = "http://192.168.0.23:8321/v1/"; proxyPass = "http://192.168.0.23:8321/v1/";
proxyWebsockets = true; proxyWebsockets = true;
extraConfig = apiKeyAuth + '' extraConfig = apiKeyAuth + ''
@ -158,7 +159,7 @@ in
}; };
# Llama Stack beta API # Llama Stack beta API
locations."/v1beta/" = { locations."/stack/v1beta/" = {
proxyPass = "http://192.168.0.23:8321/v1beta/"; proxyPass = "http://192.168.0.23:8321/v1beta/";
proxyWebsockets = true; proxyWebsockets = true;
extraConfig = apiKeyAuth + '' extraConfig = apiKeyAuth + ''