From d81b9264579a4e990bcc5c06021aa5e513c2cf9a Mon Sep 17 00:00:00 2001 From: joncrall Date: Thu, 26 Mar 2026 15:53:11 -0400 Subject: [PATCH 1/2] Add publish to pypi action --- .github/workflows/publish.yml | 33 ++++++++++++++++++++++++++++++--- 1 file changed, 30 insertions(+), 3 deletions(-) diff --git a/.github/workflows/publish.yml b/.github/workflows/publish.yml index 051b5ee94..ef893f762 100644 --- a/.github/workflows/publish.yml +++ b/.github/workflows/publish.yml @@ -1,12 +1,13 @@ -name: Create GitHub Release +name: Publish release on: push: tags: - - 'v*.*.*' + - "v*.*.*" jobs: - create-release: + build: + name: Build distributions and create GitHub release runs-on: ubuntu-latest permissions: contents: write @@ -21,6 +22,12 @@ jobs: - name: Build package run: uv build + - name: Upload distribution artifacts + uses: actions/upload-artifact@v4 + with: + name: python-package-distributions + path: dist/ + - name: Create GitHub Release uses: softprops/action-gh-release@v2 with: @@ -30,3 +37,23 @@ jobs: dist/*.tar.gz env: GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }} + + publish-to-pypi: + name: Publish to PyPI + needs: build + runs-on: ubuntu-latest + environment: + name: pypi + url: https://pypi.org/p/every-eval-ever + permissions: + id-token: write + + steps: + - name: Download distribution artifacts + uses: actions/download-artifact@v4 + with: + name: python-package-distributions + path: dist/ + + - name: Publish distributions to PyPI + uses: pypa/gh-action-pypi-publish@release/v1 From 702c87cdf279dc75bd02cb69b4c94a8e61a13676 Mon Sep 17 00:00:00 2001 From: joncrall Date: Thu, 26 Mar 2026 16:09:27 -0400 Subject: [PATCH 2/2] Add pypi badges to the README --- README.md | 3 +++ 1 file changed, 3 insertions(+) diff --git a/README.md b/README.md index 262a2ca11..5b9dec696 100644 --- a/README.md +++ b/README.md @@ -1,5 +1,8 @@ # Every Eval Ever +[![PyPI version](https://img.shields.io/pypi/v/every-eval-ever.svg)](https://pypi.org/project/every-eval-ever/) +[![PyPI downloads](https://img.shields.io/pypi/dm/every-eval-ever.svg)](https://pypistats.org/packages/every-eval-ever) + > [EvalEval Coalition](https://evalevalai.com) — "We are a researcher community developing scientifically grounded research outputs and robust deployment infrastructure for broader impact evaluations." **Every Eval Ever** is a shared schema and crowdsourced eval database. It defines a standardized metadata format for storing AI evaluation results — from leaderboard scrapes and research papers to local evaluation runs — so that results from different frameworks can be compared, reproduced, and reused. The three components that make it work: