id,seq,job,job_label,repo,repo_label,uses,name,with,run,env,if 133396,1,26772,mirror,107914493,datasette,zofrex/mirror-branch@ea152f124954fa4eb26eea3fe0dbe313a3a08d94,"Mirror to ""master""","{""target-branch"": ""master"", ""force"": false}",,, 133397,2,26772,mirror,107914493,datasette,zofrex/mirror-branch@ea152f124954fa4eb26eea3fe0dbe313a3a08d94,"Mirror to ""main""","{""target-branch"": ""main"", ""force"": false}",,, 138802,1,27858,deploy-branch-preview,107914493,datasette,actions/checkout@v3,,,,, 138803,2,27858,deploy-branch-preview,107914493,datasette,actions/setup-python@v4,Set up Python 3.11,"{""python-version"": ""3.11""}",,, 138804,3,27858,deploy-branch-preview,107914493,datasette,,Install dependencies,,"pip install datasette-publish-vercel ",, 138805,4,27858,deploy-branch-preview,107914493,datasette,,Deploy the preview,,"export BRANCH=""${{ github.event.inputs.branch }}"" wget https://latest.datasette.io/fixtures.db datasette publish vercel fixtures.db \ --branch $BRANCH \ --project ""datasette-preview-$BRANCH"" \ --token $VERCEL_TOKEN \ --scope datasette \ --about ""Preview of $BRANCH"" \ --about_url ""https://github.com/simonw/datasette/tree/$BRANCH"" ","{""VERCEL_TOKEN"": ""${{ secrets.BRANCH_PREVIEW_VERCEL_TOKEN }}""}", 138806,1,27859,deploy,107914493,datasette,actions/checkout@v3,Check out datasette,,,, 138807,2,27859,deploy,107914493,datasette,actions/setup-python@v4,Set up Python,"{""python-version"": ""3.9""}",,, 138808,3,27859,deploy,107914493,datasette,actions/cache@v3,Configure pip caching,"{""path"": ""~/.cache/pip"", ""key"": ""${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}"", ""restore-keys"": ""${{ runner.os }}-pip-\n""}",,, 138809,4,27859,deploy,107914493,datasette,,Install Python dependencies,,"python -m pip install --upgrade pip python -m pip install -e .[test] python -m pip install -e .[docs] python -m pip install sphinx-to-sqlite==0.1a1 ",, 138810,5,27859,deploy,107914493,datasette,,Run tests,,"pytest -n auto -m ""not serial"" pytest -m ""serial"" ",,${{ github.ref == 'refs/heads/main' }} 138811,6,27859,deploy,107914493,datasette,,Build fixtures.db and other files needed to deploy the demo,,"python tests/fixtures.py \ fixtures.db \ fixtures-config.json \ fixtures-metadata.json \ plugins \ --extra-db-filename extra_database.db",, 138812,7,27859,deploy,107914493,datasette,,Build docs.db,,"cd docs DISABLE_SPHINX_INLINE_TABS=1 sphinx-build -b xml . _build sphinx-to-sqlite ../docs.db _build cd ..",,${{ github.ref == 'refs/heads/main' }} 138813,8,27859,deploy,107914493,datasette,,Set up the alternate-route demo,,"echo ' from datasette import hookimpl @hookimpl def startup(datasette): db = datasette.get_database(""fixtures2"") db.route = ""alternative-route"" ' > plugins/alternative_route.py cp fixtures.db fixtures2.db ",, 138814,9,27859,deploy,107914493,datasette,,And the counters writable canned query demo,,"cat > plugins/counters.py <=0.2.2' \ --service ""datasette-latest$SUFFIX"" \ --secret $LATEST_DATASETTE_SECRET","{""LATEST_DATASETTE_SECRET"": ""${{ secrets.LATEST_DATASETTE_SECRET }}""}", 138817,12,27859,deploy,107914493,datasette,,Deploy to docs as well (only for main),,"# Deploy docs.db to a different service datasette publish cloudrun docs.db \ --branch=$GITHUB_SHA \ --version-note=$GITHUB_SHA \ --extra-options=""--setting template_debug 1"" \ --service=datasette-docs-latest",,${{ github.ref == 'refs/heads/main' }} 138818,1,27860,documentation-links,107914493,datasette,readthedocs/actions/preview@v1,,"{""project-slug"": ""datasette""}",,, 138819,1,27861,prettier,107914493,datasette,actions/checkout@v2,Check out repo,,,, 138820,2,27861,prettier,107914493,datasette,actions/cache@v2,Configure npm caching,"{""path"": ""~/.npm"", ""key"": ""${{ runner.OS }}-npm-${{ hashFiles('**/package-lock.json') }}"", ""restore-keys"": ""${{ runner.OS }}-npm-\n""}",,, 138821,3,27861,prettier,107914493,datasette,,Install dependencies,,npm ci,, 138822,4,27861,prettier,107914493,datasette,,Run prettier,,npm run prettier -- --check,, 138823,1,27862,test,107914493,datasette,actions/checkout@v3,,,,, 138824,2,27862,test,107914493,datasette,actions/setup-python@v4,Set up Python ${{ matrix.python-version }},"{""python-version"": ""${{ matrix.python-version }}""}",,, 138825,3,27862,test,107914493,datasette,actions/cache@v3,Configure pip caching,"{""path"": ""~/.cache/pip"", ""key"": ""${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}"", ""restore-keys"": ""${{ runner.os }}-pip-\n""}",,, 138826,4,27862,test,107914493,datasette,,Install dependencies,,"pip install -e '.[test]' ",, 138827,5,27862,test,107914493,datasette,,Run tests,,"pytest ",, 138828,1,27863,deploy,107914493,datasette,actions/checkout@v3,,,,, 138829,2,27863,deploy,107914493,datasette,actions/setup-python@v4,Set up Python,"{""python-version"": ""3.11""}",,, 138830,3,27863,deploy,107914493,datasette,actions/cache@v3,Configure pip caching,"{""path"": ""~/.cache/pip"", ""key"": ""${{ runner.os }}-publish-pip-${{ hashFiles('**/setup.py') }}"", ""restore-keys"": ""${{ runner.os }}-publish-pip-\n""}",,, 138831,4,27863,deploy,107914493,datasette,,Install dependencies,,"pip install setuptools wheel twine ",, 138832,5,27863,deploy,107914493,datasette,,Publish,,"python setup.py sdist bdist_wheel twine upload dist/* ","{""TWINE_USERNAME"": ""__token__"", ""TWINE_PASSWORD"": ""${{ secrets.PYPI_TOKEN }}""}", 138833,1,27864,deploy_static_docs,107914493,datasette,actions/checkout@v2,,,,, 138834,2,27864,deploy_static_docs,107914493,datasette,actions/setup-python@v2,Set up Python,"{""python-version"": ""3.9""}",,, 138835,3,27864,deploy_static_docs,107914493,datasette,actions/cache@v2,Configure pip caching,"{""path"": ""~/.cache/pip"", ""key"": ""${{ runner.os }}-publish-pip-${{ hashFiles('**/setup.py') }}"", ""restore-keys"": ""${{ runner.os }}-publish-pip-\n""}",,, 138836,4,27864,deploy_static_docs,107914493,datasette,,Install dependencies,,"python -m pip install -e .[docs] python -m pip install sphinx-to-sqlite==0.1a1 ",, 138837,5,27864,deploy_static_docs,107914493,datasette,,Build docs.db,,"cd docs DISABLE_SPHINX_INLINE_TABS=1 sphinx-build -b xml . _build sphinx-to-sqlite ../docs.db _build cd ..",, 138838,6,27864,deploy_static_docs,107914493,datasette,google-github-actions/setup-gcloud@v0,Set up Cloud Run,"{""version"": ""318.0.0"", ""service_account_email"": ""${{ secrets.GCP_SA_EMAIL }}"", ""service_account_key"": ""${{ secrets.GCP_SA_KEY }}""}",,, 138839,7,27864,deploy_static_docs,107914493,datasette,,Deploy stable-docs.datasette.io to Cloud Run,,"gcloud config set run/region us-central1 gcloud config set project datasette-222320 datasette publish cloudrun docs.db \ --service=datasette-docs-stable",, 138840,1,27865,deploy_docker,107914493,datasette,actions/checkout@v2,,,,, 138841,2,27865,deploy_docker,107914493,datasette,,Build and push to Docker Hub,,"sleep 60 # Give PyPI time to make the new release available docker login -u $DOCKER_USER -p $DOCKER_PASS export REPO=datasetteproject/datasette docker build -f Dockerfile \ -t $REPO:${GITHUB_REF#refs/tags/} \ --build-arg VERSION=${GITHUB_REF#refs/tags/} . docker tag $REPO:${GITHUB_REF#refs/tags/} $REPO:latest docker push $REPO:${GITHUB_REF#refs/tags/} docker push $REPO:latest","{""DOCKER_USER"": ""${{ secrets.DOCKER_USER }}"", ""DOCKER_PASS"": ""${{ secrets.DOCKER_PASS }}""}", 138842,1,27866,deploy_docker,107914493,datasette,actions/checkout@v2,,,,, 138843,2,27866,deploy_docker,107914493,datasette,,Build and push to Docker Hub,,"docker login -u $DOCKER_USER -p $DOCKER_PASS export REPO=datasetteproject/datasette docker build -f Dockerfile \ -t $REPO:${VERSION_TAG} \ --build-arg VERSION=${VERSION_TAG} . docker push $REPO:${VERSION_TAG}","{""DOCKER_USER"": ""${{ secrets.DOCKER_USER }}"", ""DOCKER_PASS"": ""${{ secrets.DOCKER_PASS }}"", ""VERSION_TAG"": ""${{ github.event.inputs.version_tag }}""}", 138844,1,27867,spellcheck,107914493,datasette,actions/checkout@v4,,,,, 138845,2,27867,spellcheck,107914493,datasette,actions/setup-python@v4,Set up Python,"{""python-version"": ""3.11"", ""cache"": ""pip"", ""cache-dependency-path"": ""**/setup.py""}",,, 138846,3,27867,spellcheck,107914493,datasette,,Install dependencies,,"pip install -e '.[docs]' ",, 138847,4,27867,spellcheck,107914493,datasette,,Check spelling,,"codespell README.md --ignore-words docs/codespell-ignore-words.txt codespell docs/*.rst --ignore-words docs/codespell-ignore-words.txt codespell datasette -S datasette/static --ignore-words docs/codespell-ignore-words.txt ",, 138848,1,27868,test,107914493,datasette,actions/checkout@v2,Check out datasette,,,, 138849,2,27868,test,107914493,datasette,actions/setup-python@v2,Set up Python,"{""python-version"": 3.9}",,, 138850,3,27868,test,107914493,datasette,actions/cache@v2,Configure pip caching,"{""path"": ""~/.cache/pip"", ""key"": ""${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}"", ""restore-keys"": ""${{ runner.os }}-pip-\n""}",,, 138851,4,27868,test,107914493,datasette,,Install Python dependencies,,"python -m pip install --upgrade pip python -m pip install -e .[test] python -m pip install pytest-cov ",, 138852,5,27868,test,107914493,datasette,,Run tests,,"ls -lah cat .coveragerc pytest --cov=datasette --cov-config=.coveragerc --cov-report xml:coverage.xml --cov-report term ls -lah",, 138853,6,27868,test,107914493,datasette,codecov/codecov-action@v1,Upload coverage report,"{""token"": ""${{ secrets.CODECOV_TOKEN }}"", ""file"": ""coverage.xml""}",,, 138854,1,27869,test,107914493,datasette,actions/checkout@v3,,,,, 138855,2,27869,test,107914493,datasette,actions/setup-python@v3,Set up Python 3.10,"{""python-version"": ""3.10"", ""cache"": ""pip"", ""cache-dependency-path"": ""**/setup.py""}",,, 138856,3,27869,test,107914493,datasette,actions/cache@v2,Cache Playwright browsers,"{""path"": ""~/.cache/ms-playwright/"", ""key"": ""${{ runner.os }}-browsers""}",,, 138857,4,27869,test,107914493,datasette,,Install Playwright dependencies,,"pip install shot-scraper build shot-scraper install ",, 138858,5,27869,test,107914493,datasette,,Run test,,"./test-in-pyodide-with-shot-scraper.sh ",, 138859,1,27870,test,107914493,datasette,actions/checkout@v3,,,,, 138860,2,27870,test,107914493,datasette,actions/setup-python@v4,Set up Python ${{ matrix.python-version }},"{""python-version"": ""${{ matrix.python-version }}"", ""allow-prereleases"": true}",,, 138861,3,27870,test,107914493,datasette,actions/cache@v3,Configure pip caching,"{""path"": ""~/.cache/pip"", ""key"": ""${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}"", ""restore-keys"": ""${{ runner.os }}-pip-\n""}",,, 138862,4,27870,test,107914493,datasette,,Build extension for --load-extension test,,(cd tests && gcc ext.c -fPIC -shared -o ext.so),, 138863,5,27870,test,107914493,datasette,,Install dependencies,,"pip install -e '.[test]' pip freeze ",, 138864,6,27870,test,107914493,datasette,,Run tests,,"pytest -n auto -m ""not serial"" pytest -m ""serial"" # And the test that exceeds a localhost HTTPS server tests/test_datasette_https_server.sh ",, 138865,7,27870,test,107914493,datasette,,Install docs dependencies on Python 3.9+,,"pip install -e '.[docs]' ",,matrix.python-version != '3.8' 138866,8,27870,test,107914493,datasette,,Check if cog needs to be run,,"cog --check docs/*.rst ",,matrix.python-version != '3.8' 138867,9,27870,test,107914493,datasette,,Check if blacken-docs needs to be run,,"# This fails on syntax errors, or a diff was applied blacken-docs -l 60 docs/*.rst ",,matrix.python-version != '3.8' 138868,10,27870,test,107914493,datasette,,Test DATASETTE_LOAD_PLUGINS,,"pip install datasette-init datasette-json-html tests/test-datasette-load-plugins.sh ",, 138869,1,27871,build,107914493,datasette,actions/checkout@v2,,,,, 138870,2,27871,build,107914493,datasette,mxschmitt/action-tmate@v3,Setup tmate session,,,, 138871,1,27872,build,107914493,datasette,actions/checkout@v2,,,,, 138872,2,27872,build,107914493,datasette,mxschmitt/action-tmate@v3,Setup tmate session,,,,