seperate routing for llama & stack

This commit is contained in:
Bryan Ramos 2026-04-14 00:41:39 -04:00
parent 3feb5ddc6b
commit 46adf8e9f0

View file

@ -140,15 +140,16 @@ in
useACMEHost = domain;
forceSSL = true;
# Web UI — llama.cpp chat interface (browser)
# Web UI + llama.cpp API (browser, /v1/* calls from the UI)
# Auth handled by llama.cpp itself (--api-key flag)
locations."/" = {
proxyPass = "http://192.168.0.23:8000";
proxyWebsockets = true;
};
# API — Llama Stack (opencode, programmatic clients)
locations."/v1/" = {
# Llama Stack API (opencode, programmatic clients)
# Clients use baseURL: https://ai.ramos.codes/stack/v1
locations."/stack/v1/" = {
proxyPass = "http://192.168.0.23:8321/v1/";
proxyWebsockets = true;
extraConfig = apiKeyAuth + ''
@ -158,7 +159,7 @@ in
};
# Llama Stack beta API
locations."/v1beta/" = {
locations."/stack/v1beta/" = {
proxyPass = "http://192.168.0.23:8321/v1beta/";
proxyWebsockets = true;
extraConfig = apiKeyAuth + ''