#!/usr/bin/env bash # formula-session.sh — Shared helpers for formula-driven cron agents # # Provides reusable functions for the common cron-wrapper + tmux-session # pattern used by planner-run.sh, predictor-run.sh, gardener-run.sh, and supervisor-run.sh. # # Functions: # acquire_cron_lock LOCK_FILE — PID lock with stale cleanup # check_memory [MIN_MB] — skip if available RAM too low # load_formula FORMULA_FILE — sets FORMULA_CONTENT # build_context_block FILE [FILE ...] — sets CONTEXT_BLOCK # start_formula_session SESSION WORKDIR PHASE_FILE — create tmux + claude # build_prompt_footer [EXTRA_API] — sets PROMPT_FOOTER (API ref + env + phase) # run_formula_and_monitor AGENT [TIMEOUT] [CALLBACK] — session start, inject, monitor, log # formula_phase_callback PHASE — standard crash-recovery callback # # Requires: lib/agent-session.sh sourced first (for create_agent_session, # agent_kill_session, agent_inject_into_session). # Globals used by formula_phase_callback: SESSION_NAME, PHASE_FILE, # PROJECT_REPO_ROOT, PROMPT (set by the calling script). # ── Cron guards ────────────────────────────────────────────────────────── # acquire_cron_lock LOCK_FILE # Acquires a PID lock. Exits 0 if another instance is running. # Sets an EXIT trap to clean up the lock file. acquire_cron_lock() { _CRON_LOCK_FILE="$1" if [ -f "$_CRON_LOCK_FILE" ]; then local lock_pid lock_pid=$(cat "$_CRON_LOCK_FILE" 2>/dev/null || true) if [ -n "$lock_pid" ] && kill -0 "$lock_pid" 2>/dev/null; then log "run: already running (PID $lock_pid)" exit 0 fi rm -f "$_CRON_LOCK_FILE" fi echo $$ > "$_CRON_LOCK_FILE" trap 'rm -f "$_CRON_LOCK_FILE"' EXIT } # check_memory [MIN_MB] # Exits 0 (skip) if available memory is below MIN_MB (default 2000). check_memory() { local min_mb="${1:-2000}" local avail_mb avail_mb=$(free -m | awk '/Mem:/{print $7}') if [ "${avail_mb:-0}" -lt "$min_mb" ]; then log "run: skipping — only ${avail_mb}MB available (need ${min_mb})" exit 0 fi } # ── Agent identity resolution ──────────────────────────────────────────── # resolve_agent_identity # Resolves the agent identity (user login) from the FORGE_TOKEN. # Exports AGENT_IDENTITY (user login string). # Returns 0 on success, 1 on failure. resolve_agent_identity() { if [ -z "${FORGE_TOKEN:-}" ]; then log "WARNING: FORGE_TOKEN not set, cannot resolve agent identity" return 1 fi local forge_url="${FORGE_URL:-http://localhost:3000}" AGENT_IDENTITY=$(curl -sf --max-time 10 \ -H "Authorization: token ${FORGE_TOKEN}" \ "${forge_url}/api/v1/user" 2>/dev/null | jq -r '.login // empty' 2>/dev/null) || true if [ -z "$AGENT_IDENTITY" ]; then log "WARNING: failed to resolve agent identity from FORGE_TOKEN" return 1 fi log "Resolved agent identity: ${AGENT_IDENTITY}" return 0 } # ── .profile repo management ────────────────────────────────────────────── # ensure_profile_repo [AGENT_IDENTITY] # Clones or pulls the agent's .profile repo to a local cache dir. # Requires: FORGE_TOKEN, FORGE_URL. # Exports PROFILE_REPO_PATH (local cache path) and PROFILE_FORMULA_PATH. # Returns 0 on success, 1 on failure (falls back gracefully). ensure_profile_repo() { local agent_identity="${1:-${AGENT_IDENTITY:-}}" if [ -z "$agent_identity" ]; then # Try to resolve from FORGE_TOKEN if ! resolve_agent_identity; then log "WARNING: cannot resolve agent identity, skipping .profile repo" return 1 fi agent_identity="$AGENT_IDENTITY" fi # Define cache directory: /home/agent/data/.profile/{agent-name} PROFILE_REPO_PATH="${HOME:-/home/agent}/data/.profile/${agent_identity}" # Build clone URL from FORGE_URL and agent identity local forge_url="${FORGE_URL:-http://localhost:3000}" local auth_url auth_url=$(printf '%s' "$forge_url" | sed "s|://|://$(whoami):${FORGE_TOKEN}@|") local clone_url="${auth_url}/${agent_identity}/.profile.git" # Check if already cached and up-to-date if [ -d "${PROFILE_REPO_PATH}/.git" ]; then log "Pulling .profile repo: ${agent_identity}/.profile" if git -C "$PROFILE_REPO_PATH" fetch origin --quiet 2>/dev/null; then git -C "$PROFILE_REPO_PATH" checkout main --quiet 2>/dev/null || \ git -C "$PROFILE_REPO_PATH" checkout master --quiet 2>/dev/null || true git -C "$PROFILE_REPO_PATH" pull --ff-only origin main --quiet 2>/dev/null || \ git -C "$PROFILE_REPO_PATH" pull --ff-only origin master --quiet 2>/dev/null || true log ".profile repo pulled: ${PROFILE_REPO_PATH}" else log "WARNING: failed to pull .profile repo, using cached version" fi else log "Cloning .profile repo: ${agent_identity}/.profile -> ${PROFILE_REPO_PATH}" if git clone --quiet "$clone_url" "$PROFILE_REPO_PATH" 2>/dev/null; then log ".profile repo cloned: ${PROFILE_REPO_PATH}" else log "WARNING: failed to clone .profile repo ${agent_identity}/.profile — falling back to formulas/" return 1 fi fi # Set formula path from .profile PROFILE_FORMULA_PATH="${PROFILE_REPO_PATH}/formula.toml" return 0 } # _profile_has_repo # Checks if the agent has a .profile repo by querying Forgejo API. # Returns 0 if repo exists, 1 otherwise. _profile_has_repo() { local agent_identity="${1:-${AGENT_IDENTITY:-}}" if [ -z "$agent_identity" ]; then if ! resolve_agent_identity; then return 1 fi agent_identity="$AGENT_IDENTITY" fi local forge_url="${FORGE_URL:-http://localhost:3000}" local api_url="${forge_url}/api/v1/repos/${agent_identity}/.profile" # Check if repo exists via API (returns 200 if exists, 404 if not) if curl -sf -o /dev/null -w "%{http_code}" \ -H "Authorization: token ${FORGE_TOKEN}" \ "$api_url" >/dev/null 2>&1; then return 0 fi return 1 } # _count_undigested_journals # Counts journal entries in .profile/journal/ excluding archive/ # Returns count via stdout. _count_undigested_journals() { if [ ! -d "${PROFILE_REPO_PATH:-}/journal" ]; then echo "0" return fi find "${PROFILE_REPO_PATH}/journal" -maxdepth 1 -name "*.md" -type f ! -path "*/archive/*" 2>/dev/null | wc -l } # _profile_digest_journals # Runs a claude -p one-shot to digest undigested journals into lessons-learned.md # Returns 0 on success, 1 on failure. _profile_digest_journals() { local agent_identity="${1:-${AGENT_IDENTITY:-}}" local model="${2:-${CLAUDE_MODEL:-opus}}" if [ -z "$agent_identity" ]; then if ! resolve_agent_identity; then return 1 fi agent_identity="$AGENT_IDENTITY" fi local journal_dir="${PROFILE_REPO_PATH}/journal" local knowledge_dir="${PROFILE_REPO_PATH}/knowledge" local lessons_file="${knowledge_dir}/lessons-learned.md" # Collect undigested journal entries local journal_entries="" if [ -d "$journal_dir" ]; then for jf in "$journal_dir"/*.md; do [ -f "$jf" ] || continue # Skip archived entries [[ "$jf" == */archive/* ]] && continue local basename basename=$(basename "$jf") journal_entries="${journal_entries} ### ${basename} $(cat "$jf") " done fi if [ -z "$journal_entries" ]; then log "profile: no undigested journals to digest" return 0 fi # Read existing lessons if available local existing_lessons="" if [ -f "$lessons_file" ]; then existing_lessons=$(cat "$lessons_file") fi # Build prompt for digestion local digest_prompt="You are digesting journal entries from a developer agent's work sessions. ## Task Condense these journal entries into abstract, transferable lessons. Rewrite lessons-learned.md entirely. ## Constraints - Hard cap: 2KB maximum - Abstract: patterns and heuristics, not specific issues or file paths - Transferable: must help with future unseen work, not just recall past work - Drop the least transferable lessons if over limit ## Existing lessons-learned.md (if any) ${existing_lessons:-} ## Journal entries to digest ${journal_entries} ## Output Write the complete, rewritten lessons-learned.md content below. No preamble, no explanation — just the file content." # Run claude -p one-shot with same model as agent local output output=$(claude -p "$digest_prompt" \ --output-format json \ --dangerously-skip-permissions \ --max-tokens 1000 \ ${model:+--model "$model"} \ 2>>"$LOGFILE" || echo '{"result":"error"}') # Extract content from JSON response local lessons_content lessons_content=$(printf '%s' "$output" | jq -r '.result // empty' 2>/dev/null || echo "") if [ -z "$lessons_content" ]; then log "profile: failed to digest journals" return 1 fi # Ensure knowledge directory exists mkdir -p "$knowledge_dir" # Write the lessons file (full rewrite) printf '%s\n' "$lessons_content" > "$lessons_file" log "profile: wrote lessons-learned.md (${#lessons_content} bytes)" # Move digested journals to archive (if any were processed) if [ -d "$journal_dir" ]; then mkdir -p "${journal_dir}/archive" local archived=0 for jf in "$journal_dir"/*.md; do [ -f "$jf" ] || continue [[ "$jf" == */archive/* ]] && continue local basename basename=$(basename "$jf") mv "$jf" "${journal_dir}/archive/${basename}" 2>/dev/null && archived=$((archived + 1)) done if [ "$archived" -gt 0 ]; then log "profile: archived ${archived} journal entries" fi fi return 0 } # _profile_commit_and_push MESSAGE [FILE ...] # Commits and pushes changes to .profile repo. _profile_commit_and_push() { local msg="$1" shift local files=("$@") if [ ! -d "${PROFILE_REPO_PATH:-}/.git" ]; then return 1 fi ( cd "$PROFILE_REPO_PATH" || return 1 if [ ${#files[@]} -gt 0 ]; then git add "${files[@]}" else git add -A fi if ! git diff --cached --quiet 2>/dev/null; then git config user.name "${AGENT_IDENTITY}" || true git config user.email "${AGENT_IDENTITY}@users.noreply.codeberg.org" || true git commit -m "$msg" --no-verify 2>/dev/null || true git push origin main --quiet 2>/dev/null || git push origin master --quiet 2>/dev/null || true fi ) } # profile_load_lessons # Pre-session: loads lessons-learned.md into LESSONS_CONTEXT for prompt injection. # Lazy digestion: if >10 undigested journals exist, runs claude -p to digest them. # Returns 0 on success, 1 if agent has no .profile repo (silent no-op). # Requires: ensure_profile_repo() called, AGENT_IDENTITY, FORGE_TOKEN, FORGE_URL, CLAUDE_MODEL. # Exports: LESSONS_CONTEXT (the lessons file content, hard-capped at 2KB). profile_load_lessons() { # Check if agent has .profile repo if ! _profile_has_repo; then return 0 # Silent no-op fi # Pull .profile repo if ! ensure_profile_repo; then return 0 # Silent no-op fi # Check journal count for lazy digestion trigger local journal_count journal_count=$(_count_undigested_journals) if [ "${journal_count:-0}" -gt 10 ]; then log "profile: digesting ${journal_count} undigested journals" if ! _profile_digest_journals; then log "profile: warning — journal digestion failed" fi fi # Read lessons-learned.md (hard cap at 2KB) local lessons_file="${PROFILE_REPO_PATH}/knowledge/lessons-learned.md" LESSONS_CONTEXT="" if [ -f "$lessons_file" ]; then local lessons_content lessons_content=$(head -c 2048 "$lessons_file" 2>/dev/null) || lessons_content="" if [ -n "$lessons_content" ]; then # shellcheck disable=SC2034 # exported to caller for prompt injection LESSONS_CONTEXT="## Lessons learned (from .profile/knowledge/lessons-learned.md) ${lessons_content}" log "profile: loaded lessons-learned.md (${#lessons_content} bytes)" fi fi return 0 } # profile_write_journal ISSUE_NUM ISSUE_TITLE OUTCOME [FILES_CHANGED] # Post-session: writes a reflection journal entry after work completes. # Returns 0 on success, 1 on failure. # Requires: AGENT_IDENTITY, FORGE_TOKEN, FORGE_URL, CLAUDE_MODEL. # Args: # $1 - ISSUE_NUM: The issue number worked on # $2 - ISSUE_TITLE: The issue title # $3 - OUTCOME: Session outcome (merged, blocked, failed, etc.) # $4 - FILES_CHANGED: Optional comma-separated list of files changed profile_write_journal() { local issue_num="$1" local issue_title="$2" local outcome="$3" local files_changed="${4:-}" # Check if agent has .profile repo if ! _profile_has_repo; then return 0 # Silent no-op fi # Pull .profile repo if ! ensure_profile_repo; then return 0 # Silent no-op fi # Build session summary local session_summary="" if [ -n "$files_changed" ]; then session_summary="Files changed: ${files_changed} " fi session_summary="${session_summary}Outcome: ${outcome}" # Build reflection prompt local reflection_prompt="You are reflecting on a development session. Write a concise journal entry about transferable lessons learned. ## Session context - Issue: #${issue_num} — ${issue_title} - Outcome: ${outcome} ${session_summary} ## Task Write a journal entry focused on what you learned that would help you do similar work better next time. ## Constraints - Be concise (100-200 words) - Focus on transferable lessons, not a summary of what you did - Abstract patterns and heuristics, not specific issue/file references - One concise entry, not a list ## Output Write the journal entry below. Use markdown format." # Run claude -p one-shot with same model as agent local output output=$(claude -p "$reflection_prompt" \ --output-format json \ --dangerously-skip-permissions \ --max-tokens 500 \ ${CLAUDE_MODEL:+--model "$CLAUDE_MODEL"} \ 2>>"$LOGFILE" || echo '{"result":"error"}') # Extract content from JSON response local journal_content journal_content=$(printf '%s' "$output" | jq -r '.result // empty' 2>/dev/null || echo "") if [ -z "$journal_content" ]; then log "profile: failed to write journal entry" return 1 fi # Ensure journal directory exists local journal_dir="${PROFILE_REPO_PATH}/journal" mkdir -p "$journal_dir" # Write journal entry (append if exists) local journal_file="${journal_dir}/issue-${issue_num}.md" if [ -f "$journal_file" ]; then printf '\n---\n\n' >> "$journal_file" fi printf '%s\n' "$journal_content" >> "$journal_file" log "profile: wrote journal entry for issue #${issue_num}" # Commit and push to .profile repo _profile_commit_and_push "journal: issue #${issue_num} reflection" "journal/issue-${issue_num}.md" return 0 } # ── Formula loading ────────────────────────────────────────────────────── # load_formula FORMULA_FILE # Reads formula TOML into FORMULA_CONTENT. Exits 1 if missing. load_formula() { local formula_file="$1" if [ ! -f "$formula_file" ]; then log "ERROR: formula not found: $formula_file" exit 1 fi # shellcheck disable=SC2034 # consumed by the calling script FORMULA_CONTENT=$(cat "$formula_file") } # load_formula_or_profile [ROLE] [FORMULA_FILE] # Tries to load formula from .profile repo first, falls back to formulas/.toml. # Requires: AGENT_IDENTITY, ensure_profile_repo() available. # Exports: FORMULA_CONTENT, FORMULA_SOURCE (either ".profile" or "formulas/"). # Returns 0 on success, 1 on failure. load_formula_or_profile() { local role="${1:-}" local fallback_formula="${2:-}" # Try to load from .profile repo if [ -n "$AGENT_IDENTITY" ] && ensure_profile_repo "$AGENT_IDENTITY"; then if [ -f "$PROFILE_FORMULA_PATH" ]; then log "formula source: .profile (${PROFILE_FORMULA_PATH})" # shellcheck disable=SC2034 FORMULA_CONTENT="$(cat "$PROFILE_FORMULA_PATH")" FORMULA_SOURCE=".profile" return 0 else log "WARNING: .profile repo exists but formula.toml not found at ${PROFILE_FORMULA_PATH}" fi fi # Fallback to formulas/.toml if [ -n "$fallback_formula" ]; then if [ -f "$fallback_formula" ]; then log "formula source: formulas/ (fallback) — ${fallback_formula}" # shellcheck disable=SC2034 FORMULA_CONTENT="$(cat "$fallback_formula")" FORMULA_SOURCE="formulas/" return 0 else log "ERROR: formula not found in .profile and fallback file not found: $fallback_formula" return 1 fi fi # No fallback specified but role provided — construct fallback path if [ -n "$role" ]; then fallback_formula="${FACTORY_ROOT}/formulas/${role}.toml" if [ -f "$fallback_formula" ]; then log "formula source: formulas/ (fallback) — ${fallback_formula}" # shellcheck disable=SC2034 FORMULA_CONTENT="$(cat "$fallback_formula")" # shellcheck disable=SC2034 FORMULA_SOURCE="formulas/" return 0 fi fi # No fallback specified log "ERROR: formula not found in .profile and no fallback specified" return 1 } # build_context_block FILE [FILE ...] # Reads each file from $PROJECT_REPO_ROOT and builds CONTEXT_BLOCK. # Files prefixed with "ops:" are read from $OPS_REPO_ROOT instead. build_context_block() { CONTEXT_BLOCK="" local ctx ctx_path ctx_label for ctx in "$@"; do case "$ctx" in ops:*) ctx_label="${ctx#ops:}" ctx_path="${OPS_REPO_ROOT}/${ctx_label}" ;; *) ctx_label="$ctx" ctx_path="${PROJECT_REPO_ROOT}/${ctx}" ;; esac if [ -f "$ctx_path" ]; then CONTEXT_BLOCK="${CONTEXT_BLOCK} ### ${ctx_label} $(cat "$ctx_path") " fi done } # ── Ops repo helpers ───────────────────────────────────────────────── # ensure_ops_repo # Clones or pulls the ops repo so agents can read/write operational data. # Requires: OPS_REPO_ROOT, FORGE_OPS_REPO, FORGE_URL, FORGE_TOKEN. # No-op if OPS_REPO_ROOT already exists and is up-to-date. ensure_ops_repo() { local ops_root="${OPS_REPO_ROOT:-}" [ -n "$ops_root" ] || return 0 if [ -d "${ops_root}/.git" ]; then # Pull latest from primary branch git -C "$ops_root" fetch origin "${PRIMARY_BRANCH}" --quiet 2>/dev/null || true git -C "$ops_root" checkout "${PRIMARY_BRANCH}" --quiet 2>/dev/null || true git -C "$ops_root" pull --ff-only origin "${PRIMARY_BRANCH}" --quiet 2>/dev/null || true return 0 fi # Clone from Forgejo local ops_repo="${FORGE_OPS_REPO:-}" [ -n "$ops_repo" ] || return 0 local forge_url="${FORGE_URL:-http://localhost:3000}" local clone_url if [ -n "${FORGE_TOKEN:-}" ]; then local auth_url auth_url=$(printf '%s' "$forge_url" | sed "s|://|://$(whoami):${FORGE_TOKEN}@|") clone_url="${auth_url}/${ops_repo}.git" else clone_url="${forge_url}/${ops_repo}.git" fi log "Cloning ops repo: ${ops_repo} -> ${ops_root}" if git clone --quiet "$clone_url" "$ops_root" 2>/dev/null; then log "Ops repo cloned: ${ops_root}" else log "WARNING: failed to clone ops repo ${ops_repo} — creating local directory" mkdir -p "$ops_root" fi } # ops_commit_and_push MESSAGE [FILE ...] # Stage, commit, and push changes in the ops repo. # If no files specified, stages all changes. ops_commit_and_push() { local msg="$1" shift local ops_root="${OPS_REPO_ROOT:-}" [ -d "${ops_root}/.git" ] || return 0 ( cd "$ops_root" || return if [ $# -gt 0 ]; then git add "$@" else git add -A fi if ! git diff --cached --quiet; then git commit -m "$msg" git push origin "${PRIMARY_BRANCH}" --quiet 2>/dev/null || true fi ) } # ── Session management ─────────────────────────────────────────────────── # start_formula_session SESSION WORKDIR PHASE_FILE # Kills stale session, resets phase file, creates a per-agent git worktree # for session isolation, and creates a new tmux + claude session in it. # Sets _FORMULA_SESSION_WORKDIR to the worktree path (or original workdir # on fallback). Callers must clean up via remove_formula_worktree after # the session ends. # Returns 0 on success, 1 on failure. start_formula_session() { local session="$1" workdir="$2" phase_file="$3" agent_kill_session "$session" rm -f "$phase_file" # Create per-agent git worktree for session isolation. # Each agent gets its own CWD so Claude Code treats them as separate # projects — no resume collisions between sequential formula runs. _FORMULA_SESSION_WORKDIR="/tmp/disinto-${session}" # Clean up any stale worktree from a previous run git -C "$workdir" worktree remove "$_FORMULA_SESSION_WORKDIR" --force 2>/dev/null || true if git -C "$workdir" worktree add "$_FORMULA_SESSION_WORKDIR" HEAD --detach 2>/dev/null; then log "Created worktree: ${_FORMULA_SESSION_WORKDIR}" else log "WARNING: worktree creation failed — falling back to ${workdir}" _FORMULA_SESSION_WORKDIR="$workdir" fi log "Creating tmux session: ${session}" if ! create_agent_session "$session" "$_FORMULA_SESSION_WORKDIR" "$phase_file"; then log "ERROR: failed to create tmux session ${session}" return 1 fi } # remove_formula_worktree # Removes the worktree created by start_formula_session if it differs from # PROJECT_REPO_ROOT. Safe to call multiple times. No-op if no worktree was created. remove_formula_worktree() { if [ -n "${_FORMULA_SESSION_WORKDIR:-}" ] \ && [ "$_FORMULA_SESSION_WORKDIR" != "${PROJECT_REPO_ROOT:-}" ]; then git -C "$PROJECT_REPO_ROOT" worktree remove "$_FORMULA_SESSION_WORKDIR" --force 2>/dev/null || true log "Removed worktree: ${_FORMULA_SESSION_WORKDIR}" fi } # formula_phase_callback PHASE # Standard crash-recovery phase callback for formula sessions. # Requires globals: SESSION_NAME, PHASE_FILE, PROJECT_REPO_ROOT, PROMPT. # Uses _FORMULA_CRASH_COUNT (auto-initialized) for single-retry limit. # shellcheck disable=SC2154 # SESSION_NAME, PHASE_FILE, PROJECT_REPO_ROOT, PROMPT set by caller formula_phase_callback() { local phase="$1" log "phase: ${phase}" case "$phase" in PHASE:crashed) if [ "${_FORMULA_CRASH_COUNT:-0}" -gt 0 ]; then log "ERROR: session crashed again after recovery — giving up" return 0 fi _FORMULA_CRASH_COUNT=$(( ${_FORMULA_CRASH_COUNT:-0} + 1 )) log "WARNING: tmux session died unexpectedly — attempting recovery" if create_agent_session "${_MONITOR_SESSION:-$SESSION_NAME}" "${_FORMULA_SESSION_WORKDIR:-$PROJECT_REPO_ROOT}" "$PHASE_FILE" 2>/dev/null; then agent_inject_into_session "${_MONITOR_SESSION:-$SESSION_NAME}" "$PROMPT" log "Recovery session started" else log "ERROR: could not restart session after crash" fi ;; PHASE:done|PHASE:failed|PHASE:escalate|PHASE:merged) agent_kill_session "${_MONITOR_SESSION:-$SESSION_NAME}" ;; esac } # ── Stale crashed worktree cleanup ───────────────────────────────────────── # cleanup_stale_crashed_worktrees [MAX_AGE_HOURS] # Thin wrapper around worktree_cleanup_stale() from lib/worktree.sh. # Kept for backwards compatibility with existing callers. # Requires: lib/worktree.sh sourced. cleanup_stale_crashed_worktrees() { worktree_cleanup_stale "${1:-24}" } # ── Scratch file helpers (compaction survival) ──────────────────────────── # build_scratch_instruction SCRATCH_FILE # Returns a prompt block instructing Claude to periodically flush context # to a scratch file so understanding survives context compaction. build_scratch_instruction() { local scratch_file="$1" cat <<_SCRATCH_EOF_ ## Context scratch file (compaction survival) Periodically (every 10-15 tool calls), write a summary of: - What you have discovered so far - Decisions made and why - What remains to do to: ${scratch_file} If this file existed at session start, its contents have already been injected into your prompt above. This file is ephemeral — not evidence or permanent memory, just a compaction survival mechanism. _SCRATCH_EOF_ } # read_scratch_context SCRATCH_FILE # If the scratch file exists, returns a context block for prompt injection. # Returns empty string if the file does not exist. read_scratch_context() { local scratch_file="$1" if [ -f "$scratch_file" ]; then printf '## Previous context (from scratch file)\n%s\n' "$(head -c 8192 "$scratch_file")" fi } # ── Graph report helper ─────────────────────────────────────────────────── # build_graph_section # Runs build-graph.py and sets GRAPH_SECTION to a markdown block containing # the JSON report. Sets GRAPH_SECTION="" on failure (non-fatal). # Requires globals: PROJECT_NAME, FACTORY_ROOT, PROJECT_REPO_ROOT, LOG_FILE. build_graph_section() { local report="/tmp/${PROJECT_NAME}-graph-report.json" # shellcheck disable=SC2034 # consumed by the calling script's PROMPT GRAPH_SECTION="" if python3 "$FACTORY_ROOT/lib/build-graph.py" \ --project-root "$PROJECT_REPO_ROOT" \ --output "$report" 2>>"$LOG_FILE"; then # shellcheck disable=SC2034 local report_content report_content="$(cat "$report")" # shellcheck disable=SC2034 GRAPH_SECTION=" ## Structural analysis \`\`\`json ${report_content} \`\`\`" log "graph report generated: $(jq -r '.stats | "\(.nodes) nodes, \(.edges) edges"' "$report")" else log "WARN: build-graph.py failed — continuing without structural analysis" fi } # ── SDK helpers ─────────────────────────────────────────────────────────── # build_sdk_prompt_footer [EXTRA_API_LINES] # Like build_prompt_footer but omits the phase protocol section (SDK mode). # Sets PROMPT_FOOTER. build_sdk_prompt_footer() { # shellcheck disable=SC2034 # consumed by build_prompt_footer PHASE_FILE="" # not used in SDK mode build_prompt_footer "${1:-}" PROMPT_FOOTER="${PROMPT_FOOTER%%## Phase protocol*}" } # formula_worktree_setup WORKTREE # Creates an isolated worktree for synchronous formula execution. # Fetches primary branch, cleans stale worktree, creates new one, and # sets an EXIT trap for cleanup. # Requires globals: PROJECT_REPO_ROOT, PRIMARY_BRANCH. formula_worktree_setup() { local worktree="$1" cd "$PROJECT_REPO_ROOT" || return git fetch origin "$PRIMARY_BRANCH" 2>/dev/null || true worktree_cleanup "$worktree" git worktree add "$worktree" "origin/${PRIMARY_BRANCH}" --detach 2>/dev/null # shellcheck disable=SC2064 # expand worktree now, not at trap time trap "worktree_cleanup '$worktree'" EXIT } # ── Prompt + monitor helpers ────────────────────────────────────────────── # build_prompt_footer [EXTRA_API_LINES] # Assembles the common forge API reference + environment + phase protocol # block for formula prompts. Sets PROMPT_FOOTER. # Pass additional API endpoint lines (pre-formatted, newline-prefixed) via $1. # Requires globals: FORGE_API, FACTORY_ROOT, PROJECT_REPO_ROOT, # PRIMARY_BRANCH, PHASE_FILE. build_prompt_footer() { local extra_api="${1:-}" # shellcheck disable=SC2034 # consumed by the calling script's PROMPT PROMPT_FOOTER="## Forge API reference Base URL: ${FORGE_API} Auth header: -H \"Authorization: token \${FORGE_TOKEN}\" Read issue: curl -sf -H \"Authorization: token \${FORGE_TOKEN}\" '${FORGE_API}/issues/{number}' | jq '.body' Create issue: curl -sf -X POST -H \"Authorization: token \${FORGE_TOKEN}\" -H 'Content-Type: application/json' '${FORGE_API}/issues' -d '{\"title\":\"...\",\"body\":\"...\",\"labels\":[LABEL_ID]}'${extra_api} List labels: curl -sf -H \"Authorization: token \${FORGE_TOKEN}\" '${FORGE_API}/labels' NEVER echo or include the actual token value in output — always reference \${FORGE_TOKEN}. ## Environment FACTORY_ROOT=${FACTORY_ROOT} PROJECT_REPO_ROOT=${PROJECT_REPO_ROOT} OPS_REPO_ROOT=${OPS_REPO_ROOT} PRIMARY_BRANCH=${PRIMARY_BRANCH} PHASE_FILE=${PHASE_FILE} ## Phase protocol (REQUIRED) When all work is done: echo 'PHASE:done' > '${PHASE_FILE}' On unrecoverable error: printf 'PHASE:failed\nReason: %s\n' 'describe error' > '${PHASE_FILE}'" } # run_formula_and_monitor AGENT_NAME [TIMEOUT] # Starts the formula session, injects PROMPT, monitors phase, and logs result. # Requires globals: SESSION_NAME, PHASE_FILE, PROJECT_REPO_ROOT, PROMPT, # FORGE_REPO, CLAUDE_MODEL (exported). # shellcheck disable=SC2154 # SESSION_NAME, PHASE_FILE, PROJECT_REPO_ROOT, PROMPT set by caller run_formula_and_monitor() { local agent_name="$1" local timeout="${2:-7200}" local callback="${3:-formula_phase_callback}" if ! start_formula_session "$SESSION_NAME" "$PROJECT_REPO_ROOT" "$PHASE_FILE"; then exit 1 fi # Write phase protocol to context file for compaction survival if [ -n "${PROMPT_FOOTER:-}" ]; then write_compact_context "$PHASE_FILE" "$PROMPT_FOOTER" fi agent_inject_into_session "$SESSION_NAME" "$PROMPT" log "Prompt sent to tmux session" log "Monitoring phase file: ${PHASE_FILE}" _FORMULA_CRASH_COUNT=0 monitor_phase_loop "$PHASE_FILE" "$timeout" "$callback" FINAL_PHASE=$(read_phase "$PHASE_FILE") log "Final phase: ${FINAL_PHASE:-none}" if [ "$FINAL_PHASE" != "PHASE:done" ]; then case "${_MONITOR_LOOP_EXIT:-}" in idle_prompt) log "${agent_name}: Claude returned to prompt without writing phase signal" ;; idle_timeout) log "${agent_name}: timed out with no phase signal" ;; *) log "${agent_name} finished without PHASE:done (phase: ${FINAL_PHASE:-none}, exit: ${_MONITOR_LOOP_EXIT:-})" ;; esac fi # Preserve worktree on crash for debugging; clean up on success if [ "${_MONITOR_LOOP_EXIT:-}" = "crashed" ]; then worktree_preserve "${_FORMULA_SESSION_WORKDIR:-}" "crashed (agent=${agent_name})" else remove_formula_worktree fi log "--- ${agent_name^} run done ---" }