Skip to content

ci: bump actions/setup-python from 5 to 6#1

Open
dependabot[bot] wants to merge 1 commit intomainfrom
dependabot/github_actions/actions/setup-python-6
Open

ci: bump actions/setup-python from 5 to 6#1
dependabot[bot] wants to merge 1 commit intomainfrom
dependabot/github_actions/actions/setup-python-6

Conversation

@dependabot
Copy link
Copy Markdown

@dependabot dependabot Bot commented on behalf of github Apr 22, 2026

Bumps actions/setup-python from 5 to 6.

Release notes

Sourced from actions/setup-python's releases.

v6.0.0

What's Changed

Breaking Changes

Make sure your runner is on version v2.327.1 or later to ensure compatibility with this release. See Release Notes

Enhancements:

Bug fixes:

Dependency updates:

New Contributors

Full Changelog: actions/setup-python@v5...v6.0.0

v5.6.0

What's Changed

Full Changelog: actions/setup-python@v5...v5.6.0

v5.5.0

What's Changed

Enhancements:

Bug fixes:

... (truncated)

Commits
  • a309ff8 Bump urllib3 from 2.6.0 to 2.6.3 in /tests/data (#1264)
  • bfe8cc5 Upgrade @​actions dependencies to Node 24 compatible versions (#1259)
  • 4f41a90 Bump urllib3 from 2.5.0 to 2.6.0 in /tests/data (#1253)
  • 83679a8 Bump @​types/node from 24.1.0 to 24.9.1 and update macos-13 to macos-15-intel ...
  • bfc4944 Bump prettier from 3.5.3 to 3.6.2 (#1234)
  • 97aeb3e Bump requests from 2.32.2 to 2.32.4 in /tests/data (#1130)
  • 443da59 Bump actions/publish-action from 0.3.0 to 0.4.0 & Documentation update for pi...
  • cfd55ca graalpy: add graalpy early-access and windows builds (#880)
  • bba65e5 Bump typescript from 5.4.2 to 5.9.3 and update docs/advanced-usage.md (#1094)
  • 18566f8 Improve wording and "fix example" (remove 3.13) on testing against pre-releas...
  • Additional commits viewable in compare view

Dependabot compatibility score

Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting @dependabot rebase.


Dependabot commands and options

You can trigger Dependabot actions by commenting on this PR:

  • @dependabot rebase will rebase this PR
  • @dependabot recreate will recreate this PR, overwriting any edits that have been made to it
  • @dependabot show <dependency name> ignore conditions will show all of the ignore conditions of the specified dependency
  • @dependabot ignore this major version will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself)
  • @dependabot ignore this minor version will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself)
  • @dependabot ignore this dependency will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)

Bumps [actions/setup-python](https://github.com/actions/setup-python) from 5 to 6.
- [Release notes](https://github.com/actions/setup-python/releases)
- [Commits](actions/setup-python@v5...v6)

---
updated-dependencies:
- dependency-name: actions/setup-python
  dependency-version: '6'
  dependency-type: direct:production
  update-type: version-update:semver-major
...

Signed-off-by: dependabot[bot] <support@github.com>
@dependabot @github
Copy link
Copy Markdown
Author

dependabot Bot commented on behalf of github Apr 22, 2026

Labels

The following labels could not be found: area/ci, dependencies. Please create them before Dependabot can add them to a pull request.

Please fix the above issues or remove invalid values from dependabot.yml.

@dependabot dependabot Bot requested a review from GareBear99 as a code owner April 22, 2026 17:19
GareBear99 added a commit that referenced this pull request Apr 22, 2026
…ique corpus

Mirror of the evidence the operator captured on its first non-toy run. This
doc is the audit trail that pairs each row in data/critique/operator_reviews.jsonl
with its Portfolio issue, its verdict, and its ingest manifest, so Gate v2
training can cite specific live-deployment events as the cause of any
critique-slice improvement (not just 'the model got better').

docs/OPERATOR_EVIDENCE.md
- Header with links back to LIVE_DEPLOYMENT_LEARNING.md and the operator-
  side FIRST_LIVE_RUN.md for the same run.
- Entry #1: FreeEQ8 (Portfolio issue #1, 2026-04-22).
  * Target, depth, focus.
  * Verdict: yellow / address feedback.
  * Observation: 51 files, 1.79 MB, full root file + dir listing, top
    extensions.
  * Finding verbatim: sparse symbols in three JUCE headers.
  * Exact JSONL shape that lands in data/critique/operator_reviews.jsonl.
  * Ingest manifest from scripts/ingest_operator_reviews.py --strict
    (exit 0, 1 record accepted).
  * Phase mapping (Phase 0 PROVED on FreeEQ8; Phases 2-5 blocked only by
    billing hold + three PATs, not by code).
- 'How to read this doc going forward' explains that new entries are
  appended chronologically and form the evidence chain.

README.md
- Adds 'Operator evidence log' to the Table of contents between the
  Live-deployment anchor and 'What this is'.
- Adds a 'Live-run evidence' line at the end of the Live-deployment
  section pointing at OPERATOR_EVIDENCE.md.

Co-Authored-By: Oz <oz-agent@warp.dev>
GareBear99 added a commit that referenced this pull request Apr 22, 2026
Defines the six acceptance stages LLMBuilder uses to validate that live-
deployment data from gh-ai-operator actually improves the critique
capability. Professional engineering-validation framing throughout.

Stages:
  1. Contract correctness (schema) -- PASSED, automated via round-trip CI.
  2. Sample representativeness -- diversity requirements across first 50
     records (difficulty range, distinct verdicts, distinct target_urls).
  3. Provenance auditability -- every record traceable to a public event
     (PASSED for entry #1: FreeEQ8, Portfolio issue #1).
  4. Self-consistency -- two runs against the same target at the same
     commit must produce identical verdict and Jaccard >= 0.7 on findings.
  5. A/B proof of learning on the critique slice -- two identical
     candidates, only training data differs. Pass iff delta_critique > 0
     AND max_regression <= 0.5 pp on any other slice.
  6. Blind evaluation against a human reviewer -- 10 held-out repos, 20
     randomized critiques, 3 axes (specificity, usefulness, invention
     absence). Pass iff enriched candidate wins >=6/10 on at least two
     of three axes.

Each stage has explicit pass criteria and explicit failure modes. No stage
passes by assertion. Results get appended to docs/OPERATOR_EVIDENCE.md
with the date, the Portfolio issues involved, and the commands/run IDs
that produced the measurement.

Stages 1 and 3 pass today. Stages 2 and 4-6 require the live-deployment
secrets set and sufficient ingested records to accumulate.

Co-Authored-By: Oz <oz-agent@warp.dev>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

0 participants