local base = import "base/base.jsonnet"; local images = import "values/images.jsonnet"; local url = import "values/url.jsonnet"; local prompts = import "prompts/slm.jsonnet"; { "ollama-model":: "gemma2:9b", "ollama-url":: "${OLLAMA_HOST}", "text-completion" +: { create:: function(engine) local container = engine.container("text-completion") .with_image(images.trustgraph) .with_command([ "text-completion-ollama", "-p", url.pulsar, "-m", $["ollama-model"], "-r", $["ollama-url"], ]) .with_limits("0.5", "128M") .with_reservations("0.1", "128M"); local containerSet = engine.containers( "text-completion", [ container ] ); local service = engine.internalService(containerSet) .with_port(8080, 8080, "metrics"); engine.resources([ containerSet, service, ]) }, "text-completion-rag" +: { create:: function(engine) local container = engine.container("text-completion-rag") .with_image(images.trustgraph) .with_command([ "text-completion-ollama", "-p", url.pulsar, "-m", $["ollama-model"], "-r", $["ollama-url"], "-i", "non-persistent://tg/request/text-completion-rag", "-o", "non-persistent://tg/response/text-completion-rag-response", ]) .with_limits("0.5", "128M") .with_reservations("0.1", "128M"); local containerSet = engine.containers( "text-completion-rag", [ container ] ); local service = engine.internalService(containerSet) .with_port(8080, 8080, "metrics"); engine.resources([ containerSet, service, ]) } } + prompts