Three threads, all reinforcing the contract's system-level vs.
workspace-association distinction.
WS Mux service routing
- tg-show-flows (and any workspace-level service over the WS) was
failing with "unknown service" because the post-refactor Mux
unconditionally looked up flow-service:<kind>. Now branches on
the envelope's flow field: with flow → flow-service:<kind>;
without flow → <kind>:<op> from the inner body; with bare op
lookup for service=iam. Resource and parameters come from the
matched op's own extractors — same path the HTTP endpoints take.
Optional workspace on system-level user/key ops
- list-users returns the deployment-wide list when no workspace is
supplied, filters when one is. get-user, update-user,
disable-user, enable-user, delete-user, reset-password,
create-api-key, list-api-keys, revoke-api-key all treat workspace
as an optional integrity check rather than a required argument.
- create-user keeps workspace required — there it's the new user's
home-workspace binding, a parameter rather than an address.
- API keys reclassified as SYSTEM-level resources. By the same
reasoning that makes users system-level, an API key is a
credential record on a deployment-wide registry; the workspace it
authenticates to is a property, not a containment.
Self-service surface
- whoami: returns the caller's own user record. AUTHENTICATED-only;
no users:read capability required. Foundation for UI affordances
that depend on the caller's permissions.
- bootstrap-status: POST /api/v1/auth/bootstrap-status, PUBLIC,
side-effect-free. Returns {bootstrap_available: bool} so a
first-run UI can decide whether to render setup without consuming
the bootstrap op.
- Gateway now injects actor=identity.handle on every authenticated
forward to iam-svc (IamEndpoint and WS Mux iam path), overwriting
any caller-supplied value. Underpins whoami, audit logging, and
future regime-side decisions that need actor identity.
- tg-whoami and tg-update-user CLIs.
Spec polish
- iam-contract.md: actor-injection rule documented; whoami /
bootstrap-status added to operations list; permission-scope
framing tightened (workspace scope is a property of the grant,
not the user or role).
- iam.md: self-service section; gateway flow gains the actor-
injection step; role section reframed so iam-svc constraints
don't leak into contract-level prose.
- iam-protocol.md: ops table updated for whoami, bootstrap-status,
optional-workspace pattern; bootstrap_available added to the
IamResponse listing.
44 KiB
| layout | title | parent |
|---|---|---|
| default | Identity and Access Management | Tech Specs |
Identity and Access Management
Problem Statement
TrustGraph has no meaningful identity or access management. The system
relies on a single shared gateway token for authentication and an
honour-system user query parameter for data isolation. This creates
several problems:
-
No user identity. There are no user accounts, no login, and no way to know who is making a request. The
userfield in message metadata is a caller-supplied string with no validation — any client can claim to be any user. -
No access control. A valid gateway token grants unrestricted access to every endpoint, every user's data, every collection, and every administrative operation. There is no way to limit what an authenticated caller can do.
-
No credential isolation. All callers share one static token. There is no per-user credential, no token expiration, and no rotation mechanism. Revoking access means changing the shared token, which affects all callers.
-
Data isolation is unenforced. Storage backends (Cassandra, Neo4j, Qdrant) filter queries by
userandcollection, but the gateway does not prevent a caller from specifying another user's identity. Cross-user data access is trivial. -
No audit trail. There is no logging of who accessed what. Without user identity, audit logging is impossible.
These gaps make the system unsuitable for multi-user deployments, multi-tenant SaaS, or any environment where access needs to be controlled or audited.
Current State
Authentication
The API gateway supports a single shared token configured via the
GATEWAY_SECRET environment variable or --api-token CLI argument. If
unset, authentication is disabled entirely. When enabled, every HTTP
endpoint requires an Authorization: Bearer <token> header. WebSocket
connections pass the token as a query parameter.
Implementation: trustgraph-flow/trustgraph/gateway/auth.py
class Authenticator:
def __init__(self, token=None, allow_all=False):
self.token = token
self.allow_all = allow_all
def permitted(self, token, roles):
if self.allow_all: return True
if self.token != token: return False
return True
The roles parameter is accepted but never evaluated. All authenticated
requests have identical privileges.
MCP tool configurations support an optional per-tool auth-token for
service-to-service authentication with remote MCP servers. These are
static, system-wide tokens — not per-user credentials. See
mcp-tool-bearer-token.md for details.
User identity
The user field is passed explicitly by the caller as a query parameter
(e.g. ?user=trustgraph) or set by CLI tools. It flows through the
system in the core Metadata dataclass:
@dataclass
class Metadata:
id: str = ""
root: str = ""
user: str = ""
collection: str = ""
There is no user registration, login, user database, or session management.
Data isolation
The user + collection pair is used at the storage layer to partition
data:
- Cassandra: queries filter by
userandcollectioncolumns - Neo4j: queries filter by
userandcollectionproperties - Qdrant: vector search filters by
userandcollectionmetadata
| Layer | Isolation mechanism | Enforced by |
|---|---|---|
| Gateway | Single shared token | Authenticator class |
| Message metadata | user + collection fields |
Caller (honour system) |
| Cassandra | Column filters on user, collection |
Query layer |
| Neo4j | Property filters on user, collection |
Query layer |
| Qdrant | Metadata filters on user, collection |
Query layer |
| Pub/sub topics | Per-flow topic namespacing | Flow service |
The storage-layer isolation depends on all queries correctly filtering by
user and collection. There is no gateway-level enforcement preventing
a caller from querying another user's data by passing a different user
parameter.
Configuration and secrets
| Setting | Source | Default | Purpose |
|---|---|---|---|
GATEWAY_SECRET |
Env var | Empty (auth disabled) | Gateway bearer token |
--api-token |
CLI arg | None | Gateway bearer token (overrides env) |
PULSAR_API_KEY |
Env var | None | Pub/sub broker auth |
MCP auth-token |
Config service | None | Per-tool MCP server auth |
No secrets are encrypted at rest. The gateway token and MCP tokens are stored and transmitted in plaintext (aside from any transport-layer encryption such as TLS).
Capabilities that do not exist
- Per-user authentication (JWT, OAuth, SAML, API keys per user)
- User accounts or user management
- Role-based access control (RBAC)
- Attribute-based access control (ABAC)
- Per-user or per-workspace API keys
- Token expiration or rotation
- Session management
- Per-user rate limiting
- Audit logging of user actions
- Permission checks preventing cross-user data access
- Multi-workspace credential isolation
Key files
| File | Purpose |
|---|---|
trustgraph-flow/trustgraph/gateway/auth.py |
Authenticator class |
trustgraph-flow/trustgraph/gateway/service.py |
Gateway init, token config |
trustgraph-flow/trustgraph/gateway/endpoint/*.py |
Per-endpoint auth checks |
trustgraph-base/trustgraph/schema/core/metadata.py |
Metadata dataclass with user field |
Technical Design
Design principles
-
Auth at the edge. The gateway is the single enforcement point. Internal services trust the gateway and do not re-authenticate. This avoids distributing credential validation across dozens of microservices.
-
Identity from credentials, not from callers. The gateway derives user identity from authentication credentials. Callers can no longer self-declare their identity via query parameters.
-
Workspace isolation by default. Every authenticated user belongs to a workspace. All data operations are scoped to that workspace. Cross-workspace access is not possible through the API.
-
Extensible API contract. The API accepts an optional workspace parameter on every request. This allows the same protocol to support single-workspace deployments today and multi-workspace extensions in the future without breaking changes.
-
Simple roles, not fine-grained permissions. A small number of predefined roles controls what operations a user can perform. This is sufficient for the current API surface and avoids the complexity of per-resource permission management.
Authentication
The gateway supports two credential types. Both are carried as a Bearer
token in the Authorization header for HTTP requests. The gateway
distinguishes them by format.
For WebSocket connections, credentials are not passed in the URL or headers. Instead, the client authenticates after connecting by sending an auth message as the first frame:
Client: opens WebSocket to /api/v1/socket
Server: accepts connection (unauthenticated state)
Client: sends {"type": "auth", "token": "tg_abc123..."}
Server: validates token
success → {"type": "auth-ok", "workspace": "acme"}
failure → {"type": "auth-failed", "error": "invalid token"}
The server rejects all non-auth messages until authentication succeeds. The socket remains open on auth failure, allowing the client to retry with a different token without reconnecting. The client can also send a new auth message at any time to re-authenticate — for example, to refresh an expiring JWT or to switch workspace. The resolved identity (handle, workspace, principal_id, source) is updated on each successful auth.
API keys
For programmatic access: CLI tools, scripts, and integrations.
- Opaque tokens (e.g.
tg_a1b2c3d4e5f6...). Not JWTs — short, simple, easy to paste into CLI tools and headers. - Each user has one or more API keys.
- Keys are stored hashed (SHA-256 with salt) in the IAM service. The plaintext key is returned once at creation time and cannot be retrieved afterwards.
- Keys can be revoked individually without affecting other users.
- Keys optionally have an expiry date. Expired keys are rejected.
On each request, the gateway resolves an API key by:
- Hashing the token.
- Checking a local cache (hash → identity).
- On cache miss, calling the IAM service to resolve.
- Caching the result with a short TTL (e.g. 60 seconds).
Revoked keys stop working when the cache entry expires. No push invalidation is needed.
JWTs (login sessions)
For interactive access via the UI or WebSocket connections.
- A user logs in with username and password. The gateway forwards the request to the IAM service, which validates the credentials and returns a signed JWT.
- The JWT carries identity-binding claims only — user id (
sub) and the workspace this credential authenticates to. No roles, no policy state. Per the IAM contract, all policy decisions go throughauthorise; the gateway never reads roles or other regime-internal state from the credential. - The gateway validates JWTs locally using the IAM service's public signing key — no service call needed for the authentication step; authorisation calls remain per-request (cached per the contract's caching rules).
- Token expiry is enforced by standard JWT validation at the time the request (or WebSocket connection) is made.
- For long-lived WebSocket connections, the JWT is validated at connect time only. The connection remains authenticated for its lifetime.
The IAM service manages the signing key. The gateway fetches the public key at startup (or on first JWT encounter) and caches it.
Login endpoint
POST /api/v1/auth/login
{
"username": "alice",
"password": "..."
}
→ {
"token": "eyJ...",
"expires": "2026-04-20T19:00:00Z"
}
The gateway forwards this to the IAM service, which validates credentials and returns a signed JWT. The gateway returns the JWT to the caller.
Self-service: whoami and bootstrap-status
Two side-effect-free probes that exist to support UI affordances without giving the caller broad read access:
-
POST /api/v1/iamwith{"operation": "whoami"}— authenticated only. Returns the caller's own user record (id, username, name, email, workspace, roles, enabled, must_change_password, created). Nousers:readcapability is required, because every authenticated caller can read themselves. The gateway populatesactoron the request from the authenticated identity, so the regime resolves "the caller" without taking a target argument. -
POST /api/v1/auth/bootstrap-status— public, side-effect-free. Returns{"bootstrap_available": true|false}.trueiff iam-svc is inbootstrapmode and its tables are empty (i.e. an unconsumedbootstrapcall would currently succeed). Exists so a first-run UI can decide whether to render the setup flow without invoking the consumingbootstrapop.
IAM service delegation
The gateway stays thin. Its authentication logic is:
- Extract Bearer token from header (or query param for WebSocket).
- If the token has JWT format (dotted structure), validate the signature locally and extract claims.
- Otherwise, treat as an API key: hash it and check the local cache. On cache miss, call the IAM service to resolve.
- If neither succeeds, return 401.
All user management, key management, credential validation, and token signing logic lives in the IAM service. The gateway is a generic enforcement point that can be replaced without changing the IAM service.
No legacy token support
The existing GATEWAY_SECRET shared token is removed. All
authentication uses API keys or JWTs. On first start, the bootstrap
process creates a default workspace and admin user with an initial API
key.
Identity, credentials, and workspace binding
The gateway never asks "which workspace does this user belong to?". That question forces every IAM regime to expose a user-to-workspace mapping, which prevents regimes where the relationship is many-to-many or doesn't exist (e.g. SSO with IdP-driven workspace selection). Instead, the gateway asks "which workspace does this credential authenticate to?" — a question every regime can answer in its own terms.
A credential (API key, JWT, OIDC token, etc.) is bound to a workspace at issue time. The IAM regime decides what binding means:
- OSS regime — each user has a home workspace; credentials issued to that user are bound to that workspace. A 1:1 user-to-workspace constraint is an internal data-model decision, not a contract assertion.
- Multi-workspace regime (future / enterprise) — a user with access to several workspaces gets a different credential per workspace. Each credential authenticates to exactly one workspace; the relationship between user and workspace is a regime-internal detail the gateway does not see.
When the gateway authenticates a credential, the IAM regime returns
an Identity whose workspace is the workspace this credential is
for. That value — not "the user's workspace" — is what the gateway
uses for default-fill-in and as input to the IAM authorise call.
Identity surface
What the gateway holds after authenticate:
| Field | Purpose |
|---|---|
handle |
Opaque token quoted back when calling authorise. Regime-defined. |
workspace |
The workspace this credential authenticates to. Used as the default if a request omits workspace. |
principal_id |
Stable identifier for audit logging (a user id, sub claim, service account id). Never used for authorisation. |
source |
How the credential was presented (api-key, jwt). Logged with audit events; not policy input. |
Anything else — roles, claims, group memberships, policy attributes
— stays inside the regime and is reachable only via authorise.
See iam-contract.md for the full contract.
OSS user record
The OSS regime stores the following per user. These fields are OSS-implementation specifics, not part of the contract.
| Field | Type | Description |
|---|---|---|
id |
string | Unique user identifier (UUID) |
name |
string | Display name |
email |
string | Email address (optional) |
workspace |
string | Home workspace; default binding for issued credentials |
roles |
list[string] | Assigned roles (e.g. ["reader"]) |
enabled |
bool | Whether the user can authenticate |
created |
datetime | Account creation timestamp |
The workspace field on a user record is the default binding
used when issuing credentials, not a constraint visible to the
gateway. An enterprise regime may have no user records at all
(authentication delegated to an IdP).
Workspaces
A workspace is an isolated data boundary — a tenancy scope in which
users, flows, configuration, documents, and knowledge graphs live.
Workspaces map to storage-layer isolation: the user field in
Metadata, the corresponding Cassandra keyspace, the Qdrant
collection prefix, the Neo4j property filter.
Workspace is the most prominent component of an operation's resource scope: when a request says "do X to Y", workspace is part of "Y". Listing users, creating flows, querying the graph — all of these target a specific workspace.
| Field | Type | Description |
|---|---|---|
id |
string | Unique workspace identifier |
name |
string | Display name |
enabled |
bool | Whether the workspace is active |
created |
datetime | Creation timestamp |
Default-fill-in
If a request omits workspace, the gateway fills it in from the
authenticated identity's bound workspace (identity.workspace)
before any IAM check runs. IAM never receives an unresolved
workspace; every authorise call sees a concrete value.
Authorisation
Whether the resolved workspace is permitted to be operated on by
this caller is an IAM decision, not a gateway one. The gateway
calls authorise(identity, capability, {workspace: ..., ...}) and
relays the answer. In the OSS regime, the regime checks whether
the caller's permission grants for <capability> include this
workspace — see capabilities.md. In other
regimes the decision could come from group mappings, policies,
relationship tuples, or anything else the regime models.
Request anatomy
The shape of a request — where workspace appears, where flow appears, where parameters live — follows from the level of the resource being operated on, not from any single property of the request like its URL or its required capability.
Resources live at one of three levels (see also the resource model
in iam-contract.md):
| Resource level | Examples | Resource address |
|---|---|---|
| System | The user registry, the workspace registry, the IAM signing key, the audit log | empty {} |
| Workspace | A workspace's config, flow definitions, library, knowledge cores, collections | {workspace: ...} |
| Flow | A flow's knowledge graph, agent state, LLM context, embeddings, MCP context | {workspace: ..., flow: ...} |
For the gateway-to-bus mapping this dictates where workspace lives in the message, but only when workspace is part of the resource address. Workspace can also appear as an operation parameter on system-level resources (see below).
Workspace as address vs. parameter
Two distinct roles, two distinct locations:
- Workspace as address component. Workspace identifies the
thing being operated on. Used for workspace-level and flow-level
resources. Lives in the addressing layer of the message — the
URL path for HTTP, or the WebSocket envelope alongside
flowfor flow-scoped operations sent through the Mux. - Workspace as operation parameter. Workspace is data the
operation acts on, while the resource itself is system-level.
Used for operations on the user registry (
create-user with workspace association W), the workspace registry (create- workspace W), and other system-level operations that happen to reference a workspace. Lives in the request body or inner WS payload alongside the operation's other parameters.
The two roles never coexist on the same operation. Either the
operation addresses something within a workspace (workspace is in
the address) or it operates on a system-level resource with
workspace as a parameter (workspace is in the body) or workspace
is irrelevant (system-level operations like bootstrap,
rotate-signing-key, login itself).
Where workspace lives, by request type
| Request type | Resource level | Workspace lives in |
|---|---|---|
Flow-scoped data plane (agent, graph-rag, llm, embeddings, mcp, etc.) |
Flow | Envelope alongside flow (WS) or URL path (HTTP) — part of the address |
Workspace-scoped control plane (config, library, knowledge, collection-management, flow lifecycle) |
Workspace | Body / inner request — part of the address |
User registry ops (create-user, list-users, disable-user, etc.) |
System | Body — as a parameter (the user's workspace association or a list filter) |
Workspace registry ops (create-workspace, list-workspaces, etc.) |
System | Body — as a parameter (the workspace identifier in workspace_record) |
Credential ops (create-api-key, revoke-api-key, change-password, reset-password) |
System | Body — as a parameter on ops that have one; absent on change-password (target is the caller's identity) |
System ops (bootstrap, login, rotate-signing-key, get-signing-key-public) |
System | Not present at all |
The classification is deliberate. Users are a global concept that have a workspace; they don't live in one. An OSS regime has 1:1 user-to-workspace; a multi-workspace regime maps a user to many workspaces; an SSO regime might delegate workspace membership to an IdP entirely. The gateway treats user-registry operations as system-level so the contract is the same across regimes — the workspace association is a parameter the regime interprets in its own terms.
HTTP
HTTP routes by URL path, so the address lives in the URL. Per-operation REST shape:
- Flow-level:
POST /api/v1/workspaces/{w}/flows/{f}/services/{kind}—workspaceandfloware URL components. - Workspace-level:
POST /api/v1/workspaces/{w}/config,/api/v1/workspaces/{w}/library, etc. —workspaceis a URL component. - System-level:
POST /api/v1/users,/api/v1/workspaces, etc. — no workspace in URL; if the operation references one, it's a field in the body.
/api/v1/iam is itself registry-driven: the body's operation
field is looked up against the registry to obtain the capability,
resource shape, and parameter shape per operation, rather than
gating the whole endpoint with a single coarse capability.
WebSocket Mux
The Mux envelope is the addressing layer for flow-scoped
operations. For workspace-level and system-level operations the
envelope routes by service only, and the inner request payload
carries the address components or parameters as appropriate. See
iam-contract.md for the operation-registry
mechanism the Mux uses to know which fields to read.
Roles and access control
Roles are an OSS-regime concept and live entirely in the IAM
service. The gateway does not enumerate or check them — it asks
authorise(identity, capability, resource, parameters) per
request and the regime maps the caller's roles to a decision.
The OSS regime ships three roles:
| Role | Capabilities granted |
|---|---|
reader |
Read capabilities on data and config (graph:read, documents:read, rows:read, config:read, flows:read, knowledge:read, collections:read, keys:self, plus the per-service caps agent, llm, embeddings, mcp). |
writer |
All reader capabilities, plus graph:write, documents:write, rows:write, knowledge:write, collections:write. |
admin |
All writer capabilities, plus config:write, flows:write, users:read, users:write, users:admin, keys:admin, workspaces:admin, iam:admin, metrics:read. |
Workspace scope is a property of the grant, not of the user or
role. In the OSS regime each capability granted by reader /
writer is scoped to the workspace the user record is associated
with; capabilities granted by admin are scoped to * (every
workspace). A user is a system-level object — they don't "live
in" a workspace, they hold permissions whose scope happens to
reference one.
The OSS regime is deliberately limited to one workspace association per user; future regimes are free to grant the same user different permissions in different workspaces, or use a non-workspace scope entirely. This is regime-internal — neither the contract nor the gateway carries an assumption either way.
The gateway gates each endpoint by capability, not by role.
Capabilities are declared per operation in the gateway's operation
registry; see iam-contract.md for the
registry mechanism and capabilities.md for
the capability vocabulary.
IAM service
The IAM service is a backend service that implements the
IAM contract — authenticate, authorise, and
the management operations the gateway forwards. It is the
authority for identity, credential validation, and access decisions.
The gateway treats it as a black box behind the contract; nothing
in the gateway is regime-specific.
The OSS distribution ships one IAM regime: a role-based service
backed by Cassandra, described in
iam-protocol.md. Enterprise / future regimes
can replace this implementation without changing the gateway, the
wire protocol between gateway and backends, or the capability
vocabulary — see the contract spec for the abstraction the gateway
is wired against and the implementation notes for what other
regimes look like.
OSS data model
The OSS regime stores users, workspaces, API keys, and signing keys in Cassandra. This is an OSS regime implementation detail; it is not part of the contract. Other regimes will have different (or no) data models.
iam_workspaces (
id text PRIMARY KEY,
name text,
enabled boolean,
created timestamp
)
iam_users (
id text PRIMARY KEY,
workspace text,
name text,
email text,
password_hash text,
roles set<text>,
enabled boolean,
created timestamp
)
iam_api_keys (
key_hash text PRIMARY KEY,
user_id text,
name text,
expires timestamp,
created timestamp
)
A secondary index on iam_api_keys.user_id supports listing a user's
keys.
Responsibilities
- User CRUD (create, list, update, disable)
- Workspace CRUD (create, list, update, disable)
- API key management (create, revoke, list)
- API key resolution (hash → user/workspace/roles)
- Credential validation (username/password → signed JWT)
- JWT signing key management (initialise, rotate)
- Bootstrap (create default workspace and admin user on first start)
Communication
The IAM service communicates via the standard request/response pub/sub pattern, the same as the config service. The gateway calls it to resolve API keys and to handle login requests. User management operations (create user, revoke key, etc.) also go through the IAM service.
Error policy
External error responses carry no diagnostic detail for authentication or access-control failures. The goal is to give an attacker probing the endpoint no signal about which condition they tripped.
| Category | HTTP | Body | WebSocket frame |
|---|---|---|---|
| Authentication failure | 401 Unauthorized |
{"error": "auth failure"} |
{"type": "auth-failed", "error": "auth failure"} |
| Access control failure | 403 Forbidden |
{"error": "access denied"} |
{"error": "access denied"} (endpoint-specific frame type) |
"Authentication failure" covers missing credential, malformed credential, invalid signature, expired token, revoked API key, and unknown API key — all indistinguishable to the caller.
"Access control failure" covers role insufficient, workspace mismatch, user disabled, and workspace disabled — all indistinguishable to the caller.
Server-side logging is richer. The audit log records the specific
reason ("workspace-mismatch: user alice assigned 'acme', requested 'beta'", "role-insufficient: admin required, user has writer",
etc.) for operators and post-incident forensics. These messages never
appear in responses.
Other error classes (bad request, internal error) remain descriptive
because they do not reveal anything about the auth or access-control
surface — e.g. "missing required field 'workspace'" or
"invalid JSON" is fine.
Gateway changes
The current Authenticator class is replaced with a thin
authentication+authorisation middleware that delegates to the IAM
service per the IAM contract. The gateway performs no role check
itself — authorisation is asked of the regime via authorise.
For HTTP requests:
- Extract Bearer token from the
Authorizationheader. - If the token has JWT format (dotted structure):
- Validate signature locally using the cached public key.
- Build an
Identityfromsubandworkspaceclaims (no other claims are consulted).
- Otherwise, treat as an API key:
- Hash the token and check the local cache.
- On cache miss, call the IAM service to resolve to an
Identity(handle, workspace, principal_id, source). - Cache the result with a short TTL.
- If neither succeeds, return 401.
- Look up the operation in the gateway's operation registry to get
(capability, resource_level, extractors). Build the resource address (system / workspace / flow level) and parameters from the request. - Default-fill the workspace into the body when the operation is workspace- or flow-level (so downstream code sees a single canonical address); the resource address keeps its supplied value.
- Call
authorise(identity, capability, resource, parameters). On allow, forward the request; on deny, return 403. On regime error, fail closed (401 / 503 per deployment). - Cache the decision per the contract's caching rules (clamped above by a deployment-set ceiling).
- For requests forwarded to iam-svc, set
actoron the body fromidentity.handle, overwriting any caller-supplied value. Seeiam-contract.md.
For WebSocket connections:
- Accept the connection in an unauthenticated state.
- Wait for an auth message (
{"type": "auth", "token": "..."}). - Validate the token using the same logic as steps 1-3 above.
- On success, attach the resolved identity to the connection and
send
{"type": "auth-ok", ...}. - On failure, send
{"type": "auth-failed", ...}but keep the socket open. - Reject all non-auth messages until authentication succeeds.
- Accept new auth messages at any time to re-authenticate.
- For each subsequent request frame, look up
flow-service:<service>in the registry and callauthoriseagainst the{workspace, flow}resource — same authority gateway HTTP callers see, evaluated per-frame.
CLI changes
CLI tools authenticate with API keys:
--api-keyargument on all CLI tools, replacing--api-token.tg-create-workspace,tg-list-workspacesfor workspace management.tg-create-user,tg-list-users,tg-disable-userfor user management.tg-create-api-key,tg-list-api-keys,tg-revoke-api-keyfor key management.--workspaceargument on tools that operate on workspace-scoped data.- The API key is passed as a Bearer token in the same way as the current shared token, so the transport protocol is unchanged.
Audit logging
With user identity established, the gateway logs:
- Timestamp, user ID, workspace, endpoint, HTTP method, response status.
- Audit logs are written to the standard logging output (structured JSON). Integration with external log aggregation (Loki, ELK) is a deployment concern, not an application concern.
Config service changes
All configuration is workspace-scoped (see data-ownership-model.md). The config service needs to support this.
Schema change
The config table adds workspace as a key dimension:
config (
workspace text,
class text,
key text,
value text,
PRIMARY KEY ((workspace, class), key)
)
Request format
Config requests add a workspace field at the request level. The
existing (type, key) structure is unchanged within each workspace.
Get:
{
"operation": "get",
"workspace": "workspace-a",
"keys": [{"type": "prompt", "key": "rag-prompt"}]
}
Put:
{
"operation": "put",
"workspace": "workspace-a",
"values": [{"type": "prompt", "key": "rag-prompt", "value": "..."}]
}
List (all keys of a type within a workspace):
{
"operation": "list",
"workspace": "workspace-a",
"type": "prompt"
}
Delete:
{
"operation": "delete",
"workspace": "workspace-a",
"keys": [{"type": "prompt", "key": "rag-prompt"}]
}
The workspace is set by:
- Gateway — from the authenticated user's workspace for API-facing requests.
- Internal services — explicitly, based on
Metadata.userfrom the message being processed, or_systemfor operational config.
System config namespace
Processor-level operational config (logging levels, connection strings,
resource limits) is not workspace-specific. This stays in a reserved
_system workspace that is not associated with any user workspace.
Services read system config at startup without needing a workspace
context.
Config change notifications
The config notify mechanism pushes change notifications via pub/sub when config is updated. A single update may affect multiple workspaces and multiple config types. The notification message carries a dict of changes keyed by config type, with each value being the list of affected workspaces:
{
"version": 42,
"changes": {
"prompt": ["workspace-a", "workspace-b"],
"schema": ["workspace-a"]
}
}
System config changes use the reserved _system workspace:
{
"version": 43,
"changes": {
"logging": ["_system"]
}
}
This structure is keyed by type because handlers register by type. A
handler registered for prompt looks up "prompt" directly and gets
the list of affected workspaces — no iteration over unrelated types.
Config change handlers
The current on_config hook mechanism needs two modes to support shared
processing services:
-
Workspace-scoped handlers — notify when a config type changes in a specific workspace. The handler looks up its registered type in the changes dict and checks if its workspace is in the list. Used by the gateway and by services that serve a single workspace.
-
Global handlers — notify when a config type changes in any workspace. The handler looks up its registered type in the changes dict and gets the full list of affected workspaces. Used by shared processing services (prompt-rag, agent manager, etc.) that serve all workspaces. Each workspace in the list tells the handler which cache entry to update rather than reloading everything.
Per-workspace config caching
Shared services that handle messages from multiple workspaces maintain a
per-workspace config cache. When a message arrives, the service looks up
the config for the workspace identified in Metadata.user. If the
workspace is not yet cached, the service fetches its config on demand.
Config change notifications update the relevant cache entry.
Flow and queue isolation
Flows are workspace-owned. When two workspaces start flows with the same name and blueprint, their queues must be separate to prevent data mixing.
Flow blueprint templates currently use {id} (flow instance ID) and
{class} (blueprint name) as template variables in queue names. A new
{workspace} variable is added so queue names include the workspace:
Current queue names (no workspace isolation):
flow:tg:document-load:{id} → flow:tg:document-load:default
request:tg:embeddings:{class} → request:tg:embeddings:everything
With workspace isolation:
flow:tg:{workspace}:document-load:{id} → flow:tg:ws-a:document-load:default
request:tg:{workspace}:embeddings:{class} → request:tg:ws-a:embeddings:everything
The flow service substitutes {workspace} from the authenticated
workspace when starting a flow, the same way it substitutes {id} and
{class} today.
Processing services are shared infrastructure — they consume from
workspace-specific queues but are not themselves workspace-aware. The
workspace is carried in Metadata.user on every message, so services
know which workspace's data they are processing.
Blueprint templates need updating to include {workspace} in all queue
name patterns. For migration, the flow service can inject the workspace
into queue names automatically if the template does not include
{workspace}, defaulting to the legacy behaviour for existing
blueprints.
See flow-class-definition.md for the full blueprint template specification.
What changes and what doesn't
Changes:
| Component | Change |
|---|---|
gateway/auth.py |
Replace Authenticator with new auth middleware |
gateway/service.py |
Initialise IAM client, configure JWT validation |
gateway/endpoint/*.py |
Add role requirement per endpoint |
| Metadata propagation | Gateway sets user from workspace, ignores query param |
| Config service | Add workspace dimension to config schema |
| Config table | PRIMARY KEY ((workspace, class), key) |
| Config request/response schema | Add workspace field |
| Config notify messages | Include workspace ID in change notifications |
on_config handlers |
Support workspace-scoped and global modes |
| Shared services | Per-workspace config caching |
| Flow blueprints | Add {workspace} template variable to queue names |
| Flow service | Substitute {workspace} when starting flows |
| CLI tools | New user management commands, --api-key argument |
| Cassandra schema | New iam_workspaces, iam_users, iam_api_keys tables |
Does not change:
| Component | Reason |
|---|---|
| Internal service-to-service pub/sub | Services trust the gateway |
Metadata dataclass |
user field continues to carry workspace identity |
| Storage-layer isolation | Same user + collection filtering |
| Message serialisation | No schema changes |
Migration
This is a breaking change. Existing deployments must be reconfigured:
GATEWAY_SECRETis removed. Authentication requires API keys or JWT login tokens.- The
?user=query parameter is removed. Workspace identity comes from authentication. - On first start, the IAM service bootstraps a default workspace and admin user. The initial API key is output to the service log.
- Operators create additional workspaces and users via CLI tools.
- Flow blueprints must be updated to include
{workspace}in queue name patterns. - Config data must be migrated to include the workspace dimension.
Extension points
The design includes deliberate extension points for future capabilities. These are not implemented but the architecture does not preclude them:
- Multi-workspace access. Users could be granted access to additional workspaces beyond their primary assignment. The workspace validation step checks a grant list instead of a single assignment.
- Workspace resolver. Workspace resolution on each authenticated
request — "given this user and this requested workspace, which
workspace (if any) may the request operate on?" — is encapsulated
in a single pluggable resolver. The open-source edition ships a
resolver that permits only the user's single assigned workspace;
enterprise editions that implement multi-workspace access swap in a
resolver that consults a permitted set. The wire protocol (the
optional
workspacefield on the authenticated request) is identical in both editions, so clients written against one edition work unchanged against the other. - Rules-based access control. A separate access control service could evaluate fine-grained policies (per-collection permissions, operation-level restrictions, time-based access). The gateway delegates authorisation decisions to this service.
- External identity provider integration. SAML, LDAP, and OIDC flows (group mapping, claims-based role assignment) could be added to the IAM service.
- Cross-workspace administration. A
superadminrole for platform operators who manage multiple workspaces. - Delegated workspace provisioning. APIs for programmatic workspace creation and user onboarding.
These extensions are additive — they extend the validation logic without changing the request/response protocol. The gateway can be replaced with an alternative implementation that supports these capabilities while the IAM service and backend services remain unchanged.
Implementation plan
Workspace support is a prerequisite for auth — users are assigned to workspaces, config is workspace-scoped, and flows use workspace in queue names. Implementing workspaces first allows the structural changes to be tested end-to-end without auth complicating debugging.
Phase 1: Workspace support (no auth)
All workspace-scoped data and processing changes. The system works with workspaces but no authentication — callers pass workspace as a parameter, honour system. This allows full end-to-end testing: multiple workspaces with separate flows, config, queues, and data.
Config service
- Update config client API to accept a workspace parameter on all requests
- Update config storage schema to add workspace as a key dimension
- Update config notification API to report changes as a dict of type → workspace list
- Update the processor base class to understand workspaces in config notifications (workspace-scoped and global handler modes)
- Update all processors to implement workspace-aware config handling (per-workspace config caching, on-demand fetch)
Flow and queue isolation
- Update flow blueprints to include
{workspace}in all queue name patterns - Update the flow service to substitute
{workspace}when starting flows - Update all built-in blueprints to include
{workspace}
CLI tools (workspace support)
- Add
--workspaceargument to CLI tools that operate on workspace-scoped data - Add
tg-create-workspace,tg-list-workspacescommands
Phase 2: Authentication and access control
With workspaces working, add the IAM service and lock down the gateway.
IAM service
A new service handling identity and access management on behalf of the API gateway:
- Add workspace table support (CRUD, enable/disable)
- Add user table support (CRUD, enable/disable, workspace assignment)
- Add roles support (role assignment, role validation)
- Add API key support (create, revoke, list, hash storage)
- Add ability to initialise a JWT signing key for token grants
- Add token grant endpoint: user/password login returns a signed JWT
- Add bootstrap/initialisation mechanism: ability to set the signing key and create the initial workspace + admin user on first start
API gateway integration
- Add IAM middleware to the API gateway replacing the current
Authenticator - Add local JWT validation (public key from IAM service)
- Add API key resolution with local cache (hash → user/workspace/roles, cache miss calls IAM service, short TTL)
- Add login endpoint forwarding to IAM service
- Add workspace resolution: validate requested workspace against user assignment
- Add role-based endpoint access checks
- Add user management API endpoints (forwarded to IAM service)
- Add audit logging (user ID, workspace, endpoint, method, status)
- WebSocket auth via first-message protocol (auth message after connect, socket stays open on failure, re-auth supported)
CLI tools (auth support)
- Add
tg-create-user,tg-list-users,tg-disable-usercommands - Add
tg-create-api-key,tg-list-api-keys,tg-revoke-api-keycommands - Replace
--api-tokenwith--api-keyon existing CLI tools
Bootstrap and cutover
- Create default workspace and admin user on first start if IAM tables are empty
- Remove
GATEWAY_SECRETand?user=query parameter support
Design Decisions
IAM data store
IAM data is stored in dedicated Cassandra tables owned by the IAM service, not in the config service. Reasons:
- Security isolation. The config service has a broad, generic protocol. An access control failure on the config service could expose credentials. A dedicated IAM service with a purpose-built protocol limits the attack surface and makes security auditing clearer.
- Data model fit. IAM needs indexed lookups (API key hash → user,
list keys by user). The config service's
(workspace, type, key) → valuemodel stores opaque JSON strings with no secondary indexes. - Scope. IAM data is global (workspaces, users, keys). Config is workspace-scoped. Mixing global and workspace-scoped data in the same store adds complexity.
- Audit. IAM operations (key creation, revocation, login attempts) are security events that should be logged separately from general config changes.
Deferred to future design
- OIDC integration. External identity provider support (SAML, LDAP, OIDC) is left for future implementation. The extension points section describes where this fits architecturally.
- API key scoping. API keys could be scoped to specific collections within a workspace rather than granting workspace-wide access. To be designed when the need arises.
References
- IAM Contract Specification — the gateway↔IAM regime abstraction this design is wired against.
- IAM Service Protocol Specification — the OSS regime's wire-level protocol.
- Capability Vocabulary Specification — the
capability strings the gateway uses as
authoriseinput. - Data Ownership and Information Separation
- MCP Tool Bearer Token Specification
- Multi-Tenant Support Specification
- Neo4j User Collection Isolation