All checks were successful
Extend .woodpecker/nomad-validate.yml with three new fail-closed steps
that guard every artifact under vault/policies/ and vault/roles.yaml
before it can land:
4. vault-policy-fmt — cp+fmt+diff idempotence check (vault 1.18.5
has no `policy fmt -check` flag, so we
build the non-destructive check out of
`vault policy fmt` on a /tmp copy + diff
against the original)
5. vault-policy-validate — HCL syntax + capability validation via
`vault policy write` against an inline
dev-mode Vault server (no offline
`policy validate` subcommand exists;
dev-mode writes are ephemeral so this is
a validator, not a deploy)
6. vault-roles-validate — yamllint + PyYAML-based role→policy
reference check (every role's `policy:`
field must match a vault/policies/*.hcl
basename; also checks the four required
fields name/policy/namespace/job_id)
Secret-scan coverage for vault/policies/*.hcl is already provided by
the P11 gate (.woodpecker/secret-scan.yml) via its `vault/**/*` trigger
path — this pipeline intentionally does NOT duplicate that gate to
avoid the inline-heredoc / YAML-parse failure mode that sank the prior
attempt at this issue (PR #896).
Trigger paths extended: `vault/policies/**` and `vault/roles.yaml`.
`lib/init/nomad/vault-*.sh` is already covered by the existing
`lib/init/nomad/**` glob.
Docs: nomad/AGENTS.md and vault/policies/AGENTS.md updated with the
policy lifecycle, the CI enforcement table, and the common failure
modes authors will see.
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
181 lines
9.6 KiB
Markdown
181 lines
9.6 KiB
Markdown
# vault/policies/ — Agent Instructions
|
||
|
||
HashiCorp Vault ACL policies for the disinto factory. One `.hcl` file per
|
||
policy; the basename (minus `.hcl`) is the Vault policy name applied to it.
|
||
Synced into Vault by `tools/vault-apply-policies.sh` (idempotent — see the
|
||
script header for the contract).
|
||
|
||
This directory is part of the **Nomad+Vault migration (Step 2)** — see
|
||
issues #879–#884. Policies attach to Nomad jobs via workload identity in
|
||
S2.4; this PR only lands the files + apply script.
|
||
|
||
## Naming convention
|
||
|
||
| Prefix | Audience | KV scope |
|
||
|---|---|---|
|
||
| `service-<name>.hcl` | Long-running platform services (forgejo, woodpecker) | `kv/data/disinto/shared/<name>/*` |
|
||
| `bot-<name>.hcl` | Per-agent jobs (dev, review, gardener, …) | `kv/data/disinto/bots/<name>/*` + shared forge URL |
|
||
| `runner-<TOKEN>.hcl` | Per-secret policy for vault-runner ephemeral dispatch | exactly one `kv/data/disinto/runner/<TOKEN>` path |
|
||
| `dispatcher.hcl` | Long-running edge dispatcher | `kv/data/disinto/runner/*` + `kv/data/disinto/shared/ops-repo/*` |
|
||
|
||
The KV mount name `kv/` is the convention this migration uses (mounted as
|
||
KV v2). Vault addresses KV v2 data at `kv/data/<path>` and metadata at
|
||
`kv/metadata/<path>` — policies that need `list` always target the
|
||
`metadata` path; reads target `data`.
|
||
|
||
## Policy → KV path summary
|
||
|
||
| Policy | Reads |
|
||
|---|---|
|
||
| `service-forgejo` | `kv/data/disinto/shared/forgejo/*` |
|
||
| `service-woodpecker` | `kv/data/disinto/shared/woodpecker/*` |
|
||
| `bot-<role>` (dev, review, gardener, architect, planner, predictor, supervisor, vault, dev-qwen) | `kv/data/disinto/bots/<role>/*` + `kv/data/disinto/shared/forge/*` |
|
||
| `runner-<TOKEN>` (GITHUB\_TOKEN, CODEBERG\_TOKEN, CLAWHUB\_TOKEN, DEPLOY\_KEY, NPM\_TOKEN, DOCKER\_HUB\_TOKEN) | `kv/data/disinto/runner/<TOKEN>` (exactly one) |
|
||
| `dispatcher` | `kv/data/disinto/runner/*` + `kv/data/disinto/shared/ops-repo/*` |
|
||
|
||
## Why one policy per runner secret
|
||
|
||
`vault-runner` (Step 5) reads each action TOML's `secrets = [...]` list
|
||
and composes only those `runner-<NAME>` policies onto the per-dispatch
|
||
ephemeral token. Wildcards or batched policies would hand the runner more
|
||
secrets than the action declared — defeats AD-006 (least-privilege per
|
||
external action). Adding a new declarable secret = adding one new
|
||
`runner-<NAME>.hcl` here + extending the SECRETS allow-list in vault-action
|
||
validation.
|
||
|
||
## Adding a new policy
|
||
|
||
1. Drop a file matching one of the four naming patterns above. Use an
|
||
existing file in the same family as the template — comment header,
|
||
capability list, and KV path layout should match the family.
|
||
2. Run `vault policy fmt <file>` locally so the formatting matches what
|
||
the CI fmt-check (step 4 of `.woodpecker/nomad-validate.yml`) will
|
||
accept. The fmt check runs non-destructively in CI but a dirty file
|
||
fails the step; running `fmt` locally before pushing is the fastest
|
||
path.
|
||
3. Add the matching entry to `../roles.yaml` (see "JWT-auth roles" below)
|
||
so the CI role-reference check (step 6) stays green.
|
||
4. Run `tools/vault-apply-policies.sh --dry-run` to confirm the new
|
||
basename appears in the planned-work list with the expected SHA.
|
||
5. Run `tools/vault-apply-policies.sh` against a Vault instance to
|
||
create it; re-run to confirm it reports `unchanged`.
|
||
|
||
## JWT-auth roles (S2.3)
|
||
|
||
Policies are inert until a Vault token carrying them is minted. In this
|
||
migration that mint path is JWT auth — Nomad jobs exchange their
|
||
workload-identity JWT for a Vault token via
|
||
`auth/jwt-nomad/role/<name>` → `token_policies = ["<policy>"]`. The
|
||
role bindings live in [`../roles.yaml`](../roles.yaml); the script that
|
||
enables the auth method + writes the config + applies roles is
|
||
[`lib/init/nomad/vault-nomad-auth.sh`](../../lib/init/nomad/vault-nomad-auth.sh).
|
||
The applier is [`tools/vault-apply-roles.sh`](../../tools/vault-apply-roles.sh).
|
||
|
||
### Role → policy naming convention
|
||
|
||
Role name == policy name, 1:1. `vault/roles.yaml` carries one entry per
|
||
`vault/policies/*.hcl` file:
|
||
|
||
```yaml
|
||
roles:
|
||
- name: service-forgejo # Vault role
|
||
policy: service-forgejo # ACL policy attached to minted tokens
|
||
namespace: default # bound_claims.nomad_namespace
|
||
job_id: forgejo # bound_claims.nomad_job_id
|
||
```
|
||
|
||
The role name is what jobspecs reference via `vault { role = "..." }` —
|
||
keep it identical to the policy basename so an S2.1↔S2.3 drift (new
|
||
policy without a role, or vice versa) shows up in one directory review,
|
||
not as a runtime "permission denied" at job placement.
|
||
|
||
`bound_claims.nomad_job_id` is the actual `job "..."` name in the
|
||
jobspec, which may differ from the policy name (e.g. policy
|
||
`service-forgejo` binds to job `forgejo`). Update it when each bot's or
|
||
runner's jobspec lands.
|
||
|
||
### Adding a new service
|
||
|
||
1. Write `vault/policies/<name>.hcl` using the naming-table family that
|
||
fits (`service-`, `bot-`, `runner-`, or standalone).
|
||
2. Add a matching entry to `vault/roles.yaml` with all four fields
|
||
(`name`, `policy`, `namespace`, `job_id`).
|
||
3. Apply both — either in one shot via `lib/init/nomad/vault-nomad-auth.sh`
|
||
(policies → roles → nomad SIGHUP), or granularly via
|
||
`tools/vault-apply-policies.sh` + `tools/vault-apply-roles.sh`.
|
||
4. Reference the role in the consuming jobspec's `vault { role = "<name>" }`.
|
||
|
||
### Token shape
|
||
|
||
All roles share the same token shape, hardcoded in
|
||
`tools/vault-apply-roles.sh`:
|
||
|
||
| Field | Value |
|
||
|---|---|
|
||
| `bound_audiences` | `["vault.io"]` — matches `default_identity.aud` in `nomad/server.hcl` |
|
||
| `token_type` | `service` — auto-revoked when the task exits |
|
||
| `token_ttl` | `1h` |
|
||
| `token_max_ttl` | `24h` |
|
||
|
||
Bumping any of these is a knowing, repo-wide change. Per-role overrides
|
||
would let one service's tokens outlive the others — add a field to
|
||
`vault/roles.yaml` and the applier at the same time if that ever
|
||
becomes necessary.
|
||
|
||
## Policy lifecycle
|
||
|
||
Adding a policy that an actual workload consumes is a three-step chain;
|
||
the CI pipeline guards each link.
|
||
|
||
1. **Add the policy HCL** — `vault/policies/<name>.hcl`, formatted with
|
||
`vault policy fmt`. Capabilities must be drawn from the Vault-recognized
|
||
set (`read`, `list`, `create`, `update`, `delete`, `patch`, `sudo`,
|
||
`deny`); a typo fails CI step 5 (HCL written to an inline dev-mode Vault
|
||
via `vault policy write` — a real parser, not a regex).
|
||
2. **Update `../roles.yaml`** — add a JWT-auth role entry whose `policy:`
|
||
field matches the new basename (without `.hcl`). CI step 6 re-checks
|
||
every role in this file against the policy set, so a drift between the
|
||
two directories fails the step.
|
||
3. **Reference from a Nomad jobspec** — add `vault { role = "<name>" }` in
|
||
`nomad/jobs/<service>.hcl` (owned by S2.4). Policies do not take effect
|
||
until a Nomad job asks for a token via that role.
|
||
|
||
See the "Adding a new service" walkthrough below for the applier-script
|
||
flow once steps 1–3 are committed.
|
||
|
||
## CI enforcement (`.woodpecker/nomad-validate.yml`)
|
||
|
||
The pipeline triggers on any PR touching `vault/policies/**`,
|
||
`vault/roles.yaml`, or `lib/init/nomad/vault-*.sh` and runs four
|
||
vault-scoped checks (in addition to the nomad-scoped steps already in
|
||
place):
|
||
|
||
| Step | Tool | What it catches |
|
||
|---|---|---|
|
||
| 4. `vault-policy-fmt` | `vault policy fmt` + `diff` | formatting drift — trailing whitespace, wrong indentation, missing newlines |
|
||
| 5. `vault-policy-validate` | `vault policy write` against inline dev Vault | HCL syntax errors, unknown stanzas, invalid capability names (e.g. `"frobnicate"`), malformed `path "..." {}` blocks |
|
||
| 6. `vault-roles-validate` | yamllint + PyYAML | roles.yaml syntax drift, missing required fields, role→policy references with no matching `.hcl` |
|
||
| P11 | `lib/secret-scan.sh` via `.woodpecker/secret-scan.yml` | literal secret leaked into a policy HCL (rare copy-paste mistake) — already covers `vault/**/*`, no duplicate step here |
|
||
|
||
All four steps are fail-closed — any error blocks merge. The pipeline
|
||
pins `hashicorp/vault:1.18.5` (matching `lib/init/nomad/install.sh`);
|
||
bumping the runtime version without bumping the CI image is a CI-caught
|
||
drift.
|
||
|
||
## Common failure modes
|
||
|
||
| Symptom in CI logs | Root cause | Fix |
|
||
|---|---|---|
|
||
| `vault-policy-fmt: … is not formatted — run 'vault policy fmt <file>'` | Trailing whitespace / mixed indent in an HCL file | `vault policy fmt <file>` locally and re-commit |
|
||
| `vault-policy-validate: … failed validation` plus a `policy` error from Vault | Unknown capability (e.g. `"frobnicate"`), unknown stanza, malformed `path` block | Fix the HCL; valid capabilities are `read`, `list`, `create`, `update`, `delete`, `patch`, `sudo`, `deny` |
|
||
| `vault-roles-validate: ERROR: role 'X' references policy 'Y' but vault/policies/Y.hcl does not exist` | A role's `policy:` field does not match any file basename in `vault/policies/` | Either add the missing policy HCL or fix the typo in `roles.yaml` |
|
||
| `vault-roles-validate: ERROR: role entry missing required field 'Z'` | A role in `roles.yaml` is missing one of `name`, `policy`, `namespace`, `job_id` | Add the field; all four are required |
|
||
| P11 `secret-scan: detected potential secret …` on a `.hcl` file | A literal token/password was pasted into a policy | Policies must name KV paths, not carry secret values — move the literal into KV (S2.2) and have the policy grant `read` on the path |
|
||
|
||
## What this directory does NOT own
|
||
|
||
- **Attaching policies to Nomad jobs.** That's S2.4 (#882) via the
|
||
jobspec `template { vault { policies = […] } }` stanza — the role
|
||
name in `vault { role = "..." }` is what binds the policy.
|
||
- **Writing the secret values themselves.** That's S2.2 (#880) via
|
||
`tools/vault-import.sh`.
|