github
id | seq | job | repo | uses | name | with | run | env | if |
---|---|---|---|---|---|---|---|---|---|
138834 | 2 | 27864 | 107914493 | actions/setup-python@v2 | Set up Python | { "python-version": "3.9" } |
|||
138835 | 3 | 27864 | 107914493 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-publish-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-publish-pip-\n" } |
|||
138838 | 6 | 27864 | 107914493 | google-github-actions/setup-gcloud@v0 | Set up Cloud Run | { "version": "318.0.0", "service_account_email": "${{ secrets.GCP_SA_EMAIL }}", "service_account_key": "${{ secrets.GCP_SA_KEY }}" } |
|||
138840 | 1 | 27865 | 107914493 | actions/checkout@v2 | |||||
138842 | 1 | 27866 | 107914493 | actions/checkout@v2 | |||||
138844 | 1 | 27867 | 107914493 | actions/checkout@v4 | |||||
138845 | 2 | 27867 | 107914493 | actions/setup-python@v4 | Set up Python | { "python-version": "3.11", "cache": "pip", "cache-dependency-path": "**/setup.py" } |
|||
138848 | 1 | 27868 | 107914493 | actions/checkout@v2 | Check out datasette | ||||
138849 | 2 | 27868 | 107914493 | actions/setup-python@v2 | Set up Python | { "python-version": 3.9 } |
|||
138850 | 3 | 27868 | 107914493 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138853 | 6 | 27868 | 107914493 | codecov/codecov-action@v1 | Upload coverage report | { "token": "${{ secrets.CODECOV_TOKEN }}", "file": "coverage.xml" } |
|||
138854 | 1 | 27869 | 107914493 | actions/checkout@v3 | |||||
138855 | 2 | 27869 | 107914493 | actions/setup-python@v3 | Set up Python 3.10 | { "python-version": "3.10", "cache": "pip", "cache-dependency-path": "**/setup.py" } |
|||
138856 | 3 | 27869 | 107914493 | actions/cache@v2 | Cache Playwright browsers | { "path": "~/.cache/ms-playwright/", "key": "${{ runner.os }}-browsers" } |
|||
138859 | 1 | 27870 | 107914493 | actions/checkout@v3 | |||||
138860 | 2 | 27870 | 107914493 | actions/setup-python@v4 | Set up Python ${{ matrix.python-version }} | { "python-version": "${{ matrix.python-version }}", "allow-prereleases": true } |
|||
138861 | 3 | 27870 | 107914493 | actions/cache@v3 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138869 | 1 | 27871 | 107914493 | actions/checkout@v2 | |||||
138870 | 2 | 27871 | 107914493 | mxschmitt/action-tmate@v3 | Setup tmate session | ||||
138871 | 1 | 27872 | 107914493 | actions/checkout@v2 | |||||
138872 | 2 | 27872 | 107914493 | mxschmitt/action-tmate@v3 | Setup tmate session | ||||
138873 | 1 | 27873 | 140912432 | actions/checkout@v2 | Checkout repository | { "fetch-depth": 2 } |
|||
138875 | 3 | 27873 | 140912432 | github/codeql-action/init@v1 | Initialize CodeQL | { "languages": "${{ matrix.language }}" } |
|||
138876 | 4 | 27873 | 140912432 | github/codeql-action/autobuild@v1 | Autobuild | ||||
138877 | 5 | 27873 | 140912432 | github/codeql-action/analyze@v1 | Perform CodeQL Analysis | ||||
138878 | 1 | 27874 | 140912432 | readthedocs/actions/preview@v1 | { "project-slug": "sqlite-utils" } |
||||
138879 | 1 | 27875 | 140912432 | actions/checkout@v3 | |||||
138880 | 2 | 27875 | 140912432 | actions/setup-python@v4 | Set up Python ${{ matrix.python-version }} | { "python-version": "${{ matrix.python-version }}" } |
|||
138881 | 3 | 27875 | 140912432 | actions/cache@v3 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138884 | 1 | 27876 | 140912432 | actions/checkout@v3 | |||||
138885 | 2 | 27876 | 140912432 | actions/setup-python@v4 | Set up Python | { "python-version": "3.12" } |
|||
138886 | 3 | 27876 | 140912432 | actions/cache@v3 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-publish-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-publish-pip-\n" } |
|||
138889 | 1 | 27877 | 140912432 | actions/checkout@v2 | |||||
138890 | 2 | 27877 | 140912432 | actions/setup-python@v2 | Set up Python ${{ matrix.python-version }} | { "python-version": 3.9 } |
|||
138891 | 3 | 27877 | 140912432 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138894 | 1 | 27878 | 140912432 | actions/checkout@v2 | Check out repo | ||||
138895 | 2 | 27878 | 140912432 | actions/setup-python@v2 | Set up Python | { "python-version": 3.9 } |
|||
138896 | 3 | 27878 | 140912432 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138900 | 7 | 27878 | 140912432 | codecov/codecov-action@v1 | Upload coverage report | { "token": "${{ secrets.CODECOV_TOKEN }}", "file": "coverage.xml" } |
|||
138901 | 1 | 27879 | 140912432 | actions/checkout@v3 | |||||
138902 | 2 | 27879 | 140912432 | actions/setup-python@v4 | Set up Python ${{ matrix.python-version }} | { "python-version": "${{ matrix.python-version }}" } |
|||
138903 | 3 | 27879 | 140912432 | actions/cache@v3 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138817 | 12 | 27859 | 107914493 | Deploy to docs as well (only for main) | # Deploy docs.db to a different service datasette publish cloudrun docs.db \ --branch=$GITHUB_SHA \ --version-note=$GITHUB_SHA \ --extra-options="--setting template_debug 1" \ --service=datasette-docs-latest | ${{ github.ref == 'refs/heads/main' }} | |||
138706 | 6 | 27837 | 207052882 | Fetch previous copy of database | # So we can keep track of when we first saw each dependent repo wget https://github-to-sqlite.dogsheep.net/github.db | !contains(github.event.head_commit.message, 'REFRESH_DB') | |||
138867 | 9 | 27870 | 107914493 | Check if blacken-docs needs to be run | # This fails on syntax errors, or a diff was applied blacken-docs -l 60 docs/*.rst | matrix.python-version != '3.8' | |||
138909 | 9 | 27879 | 140912432 | Build extension for --load-extension test | (cd tests && gcc ext.c -fPIC -shared -o ext.so && ls -lah) | matrix.os == 'ubuntu-latest' | |||
138862 | 4 | 27870 | 107914493 | Build extension for --load-extension test | (cd tests && gcc ext.c -fPIC -shared -o ext.so) | ||||
138858 | 5 | 27869 | 107914493 | Run test | ./test-in-pyodide-with-shot-scraper.sh | ||||
138913 | 13 | 27879 | 140912432 | Check formatting | black . --check | ||||
138814 | 9 | 27859 | 107914493 | And the counters writable canned query demo | cat > plugins/counters.py <<EOF from datasette import hookimpl @hookimpl def startup(datasette): db = datasette.add_memory_database("counters") async def inner(): await db.execute_write("create table if not exists counters (name text primary key, value integer)") await db.execute_write("insert or ignore into counters (name, value) values ('counter_a', 0)") await db.execute_write("insert or ignore into counters (name, value) values ('counter_b', 0)") await db.execute_write("insert or ignore into counters (name, value) values ('counter_c', 0)") return inner @hookimpl def canned_queries(database): if database == "counters": queries = {} for name in ("counter_a", "counter_b", "counter_c"): queries["increment_{}".format(name)] = { "sql": "update counters set value = value + 1 where name = '{}'".format(name), "on_success_message_sql": "select 'Counter {name} incremented to ' || value from counters where name = '{name}'".format(name=name), "write": True, } queries["decrement_{}".format(name)] = { "sql": "update counters set value = value - 1 where name = '{}'".format(name), "on_success_message_sql": "select 'Counter {name} decremented to ' || value from counters where name = '{name}'".format(name=name), "write": True, } return queries EOF | ||||
138812 | 7 | 27859 | 107914493 | Build docs.db | cd docs DISABLE_SPHINX_INLINE_TABS=1 sphinx-build -b xml . _build sphinx-to-sqlite ../docs.db _build cd .. | ${{ github.ref == 'refs/heads/main' }} | |||
138837 | 5 | 27864 | 107914493 | Build docs.db | cd docs DISABLE_SPHINX_INLINE_TABS=1 sphinx-build -b xml . _build sphinx-to-sqlite ../docs.db _build cd .. | ||||
138847 | 4 | 27867 | 107914493 | Check spelling | codespell README.md --ignore-words docs/codespell-ignore-words.txt codespell docs/*.rst --ignore-words docs/codespell-ignore-words.txt codespell datasette -S datasette/static --ignore-words docs/codespell-ignore-words.txt | ||||
138893 | 5 | 27877 | 140912432 | Check spelling | codespell docs/*.rst --ignore-words docs/codespell-ignore-words.txt codespell sqlite_utils --ignore-words docs/codespell-ignore-words.txt | ||||
138665 | 5 | 27829 | 611552758 | Check if README is up-to-date | cog --check README.md | ||||
138914 | 14 | 27879 | 140912432 | Check if cog needs to be run | cog --check README.md docs/*.rst | matrix.python-version != '3.7' | |||
138866 | 8 | 27870 | 107914493 | Check if cog needs to be run | cog --check docs/*.rst | matrix.python-version != '3.8' | |||
138843 | 2 | 27866 | 107914493 | Build and push to Docker Hub | docker login -u $DOCKER_USER -p $DOCKER_PASS export REPO=datasetteproject/datasette docker build -f Dockerfile \ -t $REPO:${VERSION_TAG} \ --build-arg VERSION=${VERSION_TAG} . docker push $REPO:${VERSION_TAG} | { "DOCKER_USER": "${{ secrets.DOCKER_USER }}", "DOCKER_PASS": "${{ secrets.DOCKER_PASS }}", "VERSION_TAG": "${{ github.event.inputs.version_tag }}" } |
|||
138705 | 5 | 27837 | 207052882 | Create auth.json | echo "{\"github_personal_token\": \"$GITHUB_ACCESS_TOKEN\"}" > auth.json | { "GITHUB_ACCESS_TOKEN": "${{ secrets.GH_TOKEN }}" } |
|||
138813 | 8 | 27859 | 107914493 | Set up the alternate-route demo | echo ' from datasette import hookimpl @hookimpl def startup(datasette): db = datasette.get_database("fixtures2") db.route = "alternative-route" ' > plugins/alternative_route.py cp fixtures.db fixtures2.db | ||||
138805 | 4 | 27858 | 107914493 | Deploy the preview | export BRANCH="${{ github.event.inputs.branch }}" wget https://latest.datasette.io/fixtures.db datasette publish vercel fixtures.db \ --branch $BRANCH \ --project "datasette-preview-$BRANCH" \ --token $VERCEL_TOKEN \ --scope datasette \ --about "Preview of $BRANCH" \ --about_url "https://github.com/simonw/datasette/tree/$BRANCH" | { "VERCEL_TOKEN": "${{ secrets.BRANCH_PREVIEW_VERCEL_TOKEN }}" } |
|||
138912 | 12 | 27879 | 140912432 | run flake8 if Python 3.8 or higher | flake8 | matrix.python-version >= 3.8 | |||
138839 | 7 | 27864 | 107914493 | Deploy stable-docs.datasette.io to Cloud Run | gcloud config set run/region us-central1 gcloud config set project datasette-222320 datasette publish cloudrun docs.db \ --service=datasette-docs-stable | ||||
138711 | 11 | 27837 | 207052882 | Deploy to Cloud Run | gcloud config set run/region us-central1 gcloud config set project datasette-222320 datasette publish cloudrun github.db \ -m demo-metadata.json \ --service github-to-sqlite \ --install=datasette-search-all>=0.3 \ --install=datasette-render-markdown>=1.1.2 \ --install=datasette-pretty-json>=0.2.2 \ --install=datasette-json-html \ --install=datasette-vega \ --install=datasette-render-images \ --install=datasette-graphql \ --install=datasette-atom | ||||
138816 | 11 | 27859 | 107914493 | Deploy to Cloud Run | gcloud config set run/region us-central1 gcloud config set project datasette-222320 export SUFFIX="-${GITHUB_REF#refs/heads/}" export SUFFIX=${SUFFIX#-main} # Replace 1.0 with one-dot-zero in SUFFIX export SUFFIX=${SUFFIX//1.0/one-dot-zero} datasette publish cloudrun fixtures.db fixtures2.db extra_database.db \ -m fixtures-metadata.json \ --plugins-dir=plugins \ --branch=$GITHUB_SHA \ --version-note=$GITHUB_SHA \ --extra-options="--setting template_debug 1 --setting trace_debug 1 --crossdb" \ --install 'datasette-ephemeral-tables>=0.2.2' \ --service "datasette-latest$SUFFIX" \ --secret $LATEST_DATASETTE_SECRET | { "LATEST_DATASETTE_SECRET": "${{ secrets.LATEST_DATASETTE_SECRET }}" } |
|||
138874 | 2 | 27873 | 140912432 | git checkout HEAD^2 | ${{ github.event_name == 'pull_request' }} | ||||
138724 | 3 | 27840 | 207052882 | Commit and push if README changed | git diff git config --global user.email "readme-bot@example.com" git config --global user.name "README-bot" git diff --quiet || (git add README.md && git commit -m "Updated README") git push | ||||
138796 | 3 | 27856 | 206156866 | Commit and push if README changed | git diff git config --global user.email "readme-bot@example.com" git config --global user.name "README-bot" git diff --quiet || (git add README.md && git commit -m "Updated README") git push | ||||
138707 | 7 | 27837 | 207052882 | Fetch the repos | github-to-sqlite repos github.db dogsheep sqlite-utils tables --counts github.db | ||||
138852 | 5 | 27868 | 107914493 | Run tests | ls -lah cat .coveragerc pytest --cov=datasette --cov-config=.coveragerc --cov-report xml:coverage.xml --cov-report term ls -lah | ||||
138899 | 6 | 27878 | 140912432 | Run tests | ls -lah pytest --cov=sqlite_utils --cov-report xml:coverage.xml --cov-report term ls -lah | ||||
138911 | 11 | 27879 | 140912432 | run mypy | mypy sqlite_utils tests | ||||
138821 | 3 | 27861 | 107914493 | Install dependencies | npm ci | ||||
138822 | 4 | 27861 | 107914493 | Run prettier | npm run prettier -- --check | ||||
138723 | 2 | 27840 | 207052882 | Update TOC | npx markdown-toc README.md -i | ||||
138795 | 2 | 27856 | 206156866 | Update TOC | npx markdown-toc README.md -i | ||||
138655 | 3 | 27827 | 611552758 | Install dependencies | pip install '.[test]' | ||||
138663 | 3 | 27829 | 611552758 | Install dependencies | pip install '.[test]' | ||||
138762 | 3 | 27848 | 213286752 | Install dependencies | pip install '.[test]' | ||||
138770 | 3 | 27850 | 213286752 | Install dependencies | pip install '.[test]' | ||||
138846 | 3 | 27867 | 107914493 | Install dependencies | pip install -e '.[docs]' | ||||
138865 | 7 | 27870 | 107914493 | Install docs dependencies on Python 3.9+ | pip install -e '.[docs]' | matrix.python-version != '3.8' | |||
138892 | 4 | 27877 | 140912432 | Install dependencies | pip install -e '.[docs]' | ||||
138904 | 4 | 27879 | 140912432 | Install dependencies | pip install -e '.[test,mypy,flake8]' | ||||
138669 | 4 | 27830 | 197431109 | Install dependencies | pip install -e '.[test]' | ||||
138679 | 4 | 27832 | 197431109 | Install dependencies | pip install -e '.[test]' | ||||
138689 | 4 | 27834 | 303218369 | Install dependencies | pip install -e '.[test]' | ||||
138699 | 4 | 27836 | 303218369 | Install dependencies | pip install -e '.[test]' | ||||
138715 | 4 | 27838 | 207052882 | Install dependencies | pip install -e '.[test]' | ||||
138728 | 4 | 27841 | 207052882 | Install dependencies | pip install -e '.[test]' | ||||
138733 | 4 | 27842 | 248903544 | Install dependencies | pip install -e '.[test]' | ||||
138743 | 4 | 27844 | 248903544 | Install dependencies | pip install -e '.[test]' | ||||
138748 | 4 | 27845 | 197882382 | Install dependencies | pip install -e '.[test]' | ||||
138758 | 4 | 27847 | 197882382 | Install dependencies | pip install -e '.[test]' | ||||
138774 | 3 | 27851 | 205429375 | Install dependencies | pip install -e '.[test]' | ||||
138782 | 3 | 27853 | 205429375 | Install dependencies | pip install -e '.[test]' | ||||
138787 | 4 | 27854 | 206156866 | Install dependencies | pip install -e '.[test]' | ||||
138800 | 4 | 27857 | 206156866 | Install dependencies | pip install -e '.[test]' | ||||
138826 | 4 | 27862 | 107914493 | Install dependencies | pip install -e '.[test]' | ||||
138882 | 4 | 27875 | 140912432 | Install dependencies | pip install -e '.[test]' | ||||
138863 | 5 | 27870 | 107914493 | Install dependencies | pip install -e '.[test]' pip freeze |