2026-03-18 02:05:54 +00:00
|
|
|
#!/usr/bin/env bash
|
|
|
|
|
# ci-helpers.sh — Shared CI helper functions
|
|
|
|
|
#
|
|
|
|
|
# Source from any script: source "$(dirname "$0")/../lib/ci-helpers.sh"
|
2026-03-19 09:03:03 +00:00
|
|
|
# ci_passed() requires: WOODPECKER_REPO_ID (from env.sh / project config)
|
|
|
|
|
# classify_pipeline_failure() requires: woodpecker_api() (defined in env.sh)
|
2026-03-18 02:05:54 +00:00
|
|
|
|
2026-03-19 13:48:00 +00:00
|
|
|
# diff_has_code_files — check if file list (stdin, one per line) contains code files
|
|
|
|
|
# Non-code paths: docs/*, formulas/*, evidence/*, *.md
|
|
|
|
|
# Returns 0 if any code file found, 1 if all files are non-code.
|
|
|
|
|
diff_has_code_files() {
|
|
|
|
|
while IFS= read -r f; do
|
|
|
|
|
[ -z "$f" ] && continue
|
|
|
|
|
case "$f" in
|
|
|
|
|
docs/*|formulas/*|evidence/*) continue ;;
|
|
|
|
|
*.md) continue ;;
|
|
|
|
|
*) return 0 ;;
|
|
|
|
|
esac
|
|
|
|
|
done
|
|
|
|
|
return 1
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
# ci_required_for_pr <pr_number> — check if CI is needed for this PR
|
|
|
|
|
# Returns 0 if PR has code files (CI required), 1 if non-code only (CI not required).
|
|
|
|
|
ci_required_for_pr() {
|
|
|
|
|
local pr_num="$1"
|
2026-03-20 06:55:00 +00:00
|
|
|
local files all_json
|
|
|
|
|
all_json=$(codeberg_api_all "/pulls/${pr_num}/files") || return 0
|
|
|
|
|
files=$(printf '%s' "$all_json" | jq -r '.[].filename' 2>/dev/null) || return 0
|
2026-03-19 13:48:00 +00:00
|
|
|
if [ -z "$files" ]; then
|
|
|
|
|
return 0 # empty file list — require CI as safety default
|
|
|
|
|
fi
|
|
|
|
|
echo "$files" | diff_has_code_files
|
|
|
|
|
}
|
|
|
|
|
|
2026-03-18 02:05:54 +00:00
|
|
|
# ci_passed <state> — check if CI is passing (or no CI configured)
|
|
|
|
|
# Returns 0 if state is "success", or if no CI is configured and
|
|
|
|
|
# state is empty/pending/unknown.
|
|
|
|
|
ci_passed() {
|
|
|
|
|
local state="$1"
|
|
|
|
|
if [ "$state" = "success" ]; then return 0; fi
|
|
|
|
|
if [ "${WOODPECKER_REPO_ID:-2}" = "0" ] && { [ -z "$state" ] || [ "$state" = "pending" ] || [ "$state" = "unknown" ]; }; then
|
|
|
|
|
return 0 # no CI configured
|
|
|
|
|
fi
|
|
|
|
|
return 1
|
|
|
|
|
}
|
2026-03-19 08:51:30 +00:00
|
|
|
|
2026-03-20 19:19:29 +00:00
|
|
|
# is_infra_step <step_name> <exit_code> [log_data]
|
|
|
|
|
# Checks whether a single CI step failure matches infra heuristics.
|
|
|
|
|
# Returns 0 (infra) with reason on stdout, or 1 (not infra).
|
|
|
|
|
#
|
|
|
|
|
# Heuristics (union of P2e and classify_pipeline_failure patterns):
|
|
|
|
|
# - Clone/git step with exit 128 → connection failure / rate limit
|
|
|
|
|
# - Any step with exit 137 → OOM / killed by signal 9
|
|
|
|
|
# - Log patterns: connection timeout, docker pull timeout, TLS handshake timeout
|
|
|
|
|
is_infra_step() {
|
|
|
|
|
local sname="$1" ecode="$2" log_data="${3:-}"
|
|
|
|
|
|
|
|
|
|
# Clone/git step exit 128 → Codeberg connection failure / rate limit
|
|
|
|
|
if { [[ "$sname" == *clone* ]] || [[ "$sname" == git* ]]; } && [ "$ecode" = "128" ]; then
|
|
|
|
|
echo "${sname} exit 128 (connection failure)"
|
|
|
|
|
return 0
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
# Exit 137 → OOM / killed by signal 9
|
|
|
|
|
if [ "$ecode" = "137" ]; then
|
|
|
|
|
echo "${sname} exit 137 (OOM/signal 9)"
|
|
|
|
|
return 0
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
# Log-pattern matching for infra issues
|
|
|
|
|
if [ -n "$log_data" ] && \
|
|
|
|
|
printf '%s' "$log_data" | grep -qiE 'Failed to connect|connection timed out|docker pull.*timeout|TLS handshake timeout'; then
|
|
|
|
|
echo "${sname}: log matches infra pattern (timeout/connection)"
|
|
|
|
|
return 0
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
return 1
|
|
|
|
|
}
|
|
|
|
|
|
2026-03-19 08:51:30 +00:00
|
|
|
# classify_pipeline_failure <repo_id> <pipeline_num>
|
2026-03-20 19:19:29 +00:00
|
|
|
# Classifies a pipeline's failure type by inspecting failed steps.
|
|
|
|
|
# Uses is_infra_step() for per-step classification (exit codes + log patterns).
|
|
|
|
|
# Outputs "infra <reason>" if any failed step matches infra heuristics.
|
2026-03-19 08:51:30 +00:00
|
|
|
# Outputs "code" otherwise (including when steps cannot be determined).
|
|
|
|
|
# Returns 0 for infra, 1 for code or unclassifiable.
|
|
|
|
|
classify_pipeline_failure() {
|
|
|
|
|
local repo_id="$1" pip_num="$2"
|
2026-03-20 19:19:29 +00:00
|
|
|
local pip_json failed_steps _sname _ecode _spid _reason _log_data
|
2026-03-19 08:51:30 +00:00
|
|
|
|
|
|
|
|
pip_json=$(woodpecker_api "/repos/${repo_id}/pipelines/${pip_num}" 2>/dev/null) || {
|
|
|
|
|
echo "code"; return 1
|
|
|
|
|
}
|
|
|
|
|
|
2026-03-20 19:19:29 +00:00
|
|
|
# Extract failed steps: name, exit_code, pid
|
2026-03-19 08:51:30 +00:00
|
|
|
failed_steps=$(printf '%s' "$pip_json" | jq -r '
|
|
|
|
|
.workflows[]?.children[]? |
|
|
|
|
|
select(.state == "failure" or .state == "error" or .state == "killed") |
|
2026-03-20 19:19:29 +00:00
|
|
|
"\(.name)\t\(.exit_code)\t\(.pid)"' 2>/dev/null)
|
2026-03-19 08:51:30 +00:00
|
|
|
|
|
|
|
|
if [ -z "$failed_steps" ]; then
|
|
|
|
|
echo "code"; return 1
|
|
|
|
|
fi
|
|
|
|
|
|
2026-03-20 19:19:29 +00:00
|
|
|
while IFS=$'\t' read -r _sname _ecode _spid; do
|
2026-03-19 08:51:30 +00:00
|
|
|
[ -z "$_sname" ] && continue
|
2026-03-20 19:19:29 +00:00
|
|
|
|
|
|
|
|
# Check name+exit_code patterns (no log fetch needed)
|
|
|
|
|
if _reason=$(is_infra_step "$_sname" "$_ecode"); then
|
|
|
|
|
echo "infra ${_reason}"
|
|
|
|
|
return 0
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
# Fetch step logs and check log patterns
|
|
|
|
|
if [ -n "$_spid" ] && [ "$_spid" != "null" ]; then
|
|
|
|
|
_log_data=$(woodpecker_api "/repos/${repo_id}/logs/${pip_num}/${_spid}" \
|
|
|
|
|
--max-time 15 2>/dev/null \
|
|
|
|
|
| jq -r '.[].data // empty' 2>/dev/null | tail -200 || true)
|
|
|
|
|
if [ -n "$_log_data" ]; then
|
|
|
|
|
if _reason=$(is_infra_step "$_sname" "$_ecode" "$_log_data"); then
|
|
|
|
|
echo "infra ${_reason}"
|
|
|
|
|
return 0
|
|
|
|
|
fi
|
|
|
|
|
fi
|
2026-03-19 08:51:30 +00:00
|
|
|
fi
|
|
|
|
|
done <<< "$failed_steps"
|
|
|
|
|
|
|
|
|
|
echo "code"
|
|
|
|
|
return 1
|
|
|
|
|
}
|