service: ollama + open-webui
This commit is contained in:
@@ -4,20 +4,25 @@ let
|
||||
ollamahostname = "llm.kabtop.de";
|
||||
in
|
||||
{
|
||||
services.ollama = {
|
||||
enable = true;
|
||||
listenAddress = "127.0.0.1:11434";
|
||||
};
|
||||
virtualisation.oci-containers.containers."open-webui" = {
|
||||
autoStart = true;
|
||||
image = "ghcr.io/open-webui/open-webui:ollama";
|
||||
volumes = [
|
||||
"/var/lib/open-webui:/app/backend/data"
|
||||
];
|
||||
hostname = "open-webui";
|
||||
ports = [ "8081:8080" ];
|
||||
};
|
||||
|
||||
services.nginx = {
|
||||
virtualHosts = {
|
||||
ollamahostname = {
|
||||
enableACME = true;
|
||||
forceSSL = true;
|
||||
listen = [ {
|
||||
addr = "127.0.0.1"; port = 11434;
|
||||
} ];
|
||||
};
|
||||
};
|
||||
services = {
|
||||
nginx = {
|
||||
virtualHosts = {
|
||||
${ollamahostname} = {
|
||||
enableACME = true;
|
||||
forceSSL = true;
|
||||
locations."/".proxyPass = "http://localhost:8081";
|
||||
};
|
||||
};
|
||||
};
|
||||
};
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user