steps
264 rows sorted by name
This data as json, CSV (advanced)
Suggested facets: seq, uses, with, env, if
run >30 ✖
- pytest 19
- pip install -e '.[test]' 16
- pip install setuptools wheel twine 9
- python setup.py sdist bdist_wheel twine upload dist/* 9
- pip install '.[test]' 4
- pip install -e '.[docs]' 3
- pytest -v 3
- cd docs DISABLE_SPHINX_INLINE_TABS=1 sphinx-build -b xml . _build sphinx-to-sqlite ../docs.db _build cd .. 2
- git diff git config --global user.email "readme-bot@example.com" git config --global user.name "README-bot" git diff --quiet || (git add README.md && git commit -m "Updated README") git push 2
- npx markdown-toc README.md -i 2
- pip install setuptools wheel twine build 2
- python -m build twine upload dist/* 2
- python -m pip install --upgrade pip python -m pip install -e .[test] python -m pip install pytest-cov 2
- sudo apt-get install libsqlite3-mod-spatialite 2
- # Deploy docs.db to a different service datasette publish cloudrun docs.db \ --branch=$GITHUB_SHA \ --version-note=$GITHUB_SHA \ --extra-options="--setting template_debug 1" \ --service=datasette-docs-latest 1
- # So we can keep track of when we first saw each dependent repo wget https://github-to-sqlite.dogsheep.net/github.db 1
- # This fails on syntax errors, or a diff was applied blacken-docs -l 60 docs/*.rst 1
- (cd tests && gcc ext.c -fPIC -shared -o ext.so && ls -lah) 1
- (cd tests && gcc ext.c -fPIC -shared -o ext.so) 1
- ./test-in-pyodide-with-shot-scraper.sh 1
- black . --check 1
- cat > plugins/counters.py <<EOF from datasette import hookimpl @hookimpl def startup(datasette): db = datasette.add_memory_database("counters") async def inner(): await db.execute_write("create table if not exists counters (name text primary key, value integer)") await db.execute_write("insert or ignore into counters (name, value) values ('counter_a', 0)") await db.execute_write("insert or ignore into counters (name, value) values ('counter_b', 0)") await db.execute_write("insert or ignore into counters (name, value) values ('counter_c', 0)") return inner @hookimpl def canned_queries(database): if database == "counters": queries = {} for name in ("counter_a", "counter_b", "counter_c"): queries["increment_{}".format(name)] = { "sql": "update counters set value = value + 1 where name = '{}'".format(name), "on_success_message_sql": "select 'Counter {name} incremented to ' || value from counters where name = '{name}'".format(name=name), "write": True, } queries["decrement_{}".format(name)] = { "sql": "update counters set value = value - 1 where name = '{}'".format(name), "on_success_message_sql": "select 'Counter {name} decremented to ' || value from counters where name = '{name}'".format(name=name), "write": True, } return queries EOF 1
- codespell README.md --ignore-words docs/codespell-ignore-words.txt codespell docs/*.rst --ignore-words docs/codespell-ignore-words.txt codespell datasette -S datasette/static --ignore-words docs/codespell-ignore-words.txt 1
- codespell docs/*.rst --ignore-words docs/codespell-ignore-words.txt codespell sqlite_utils --ignore-words docs/codespell-ignore-words.txt 1
- cog --check README.md 1
- cog --check README.md docs/*.rst 1
- cog --check docs/*.rst 1
- docker login -u $DOCKER_USER -p $DOCKER_PASS export REPO=datasetteproject/datasette docker build -f Dockerfile \ -t $REPO:${VERSION_TAG} \ --build-arg VERSION=${VERSION_TAG} . docker push $REPO:${VERSION_TAG} 1
- echo "{\"github_personal_token\": \"$GITHUB_ACCESS_TOKEN\"}" > auth.json 1
- echo ' from datasette import hookimpl @hookimpl def startup(datasette): db = datasette.get_database("fixtures2") db.route = "alternative-route" ' > plugins/alternative_route.py cp fixtures.db fixtures2.db 1
- …
repo 12 ✖
id | seq | job | repo | uses | name ▼ | with | run | env | if |
---|---|---|---|---|---|---|---|---|---|
138653 | 1 | test 27827 | apple-notes-to-sqlite 611552758 | actions/checkout@v3 | |||||
138657 | 1 | deploy 27828 | apple-notes-to-sqlite 611552758 | actions/checkout@v3 | |||||
138661 | 1 | test 27829 | apple-notes-to-sqlite 611552758 | actions/checkout@v3 | |||||
138666 | 1 | test 27830 | dogsheep-beta 197431109 | actions/checkout@v2 | |||||
138671 | 1 | deploy 27831 | dogsheep-beta 197431109 | actions/checkout@v2 | |||||
138676 | 1 | test 27832 | dogsheep-beta 197431109 | actions/checkout@v2 | |||||
138681 | 1 | build 27833 | dogsheep-photos 256834907 | actions/checkout@v2 | |||||
138686 | 1 | test 27834 | evernote-to-sqlite 303218369 | actions/checkout@v2 | |||||
138691 | 1 | deploy 27835 | evernote-to-sqlite 303218369 | actions/checkout@v2 | |||||
138696 | 1 | test 27836 | evernote-to-sqlite 303218369 | actions/checkout@v2 | |||||
138709 | 9 | scheduled 27837 | github-to-sqlite 207052882 | actions/upload-artifact@v2 | { "path": "github.db" } |
||||
138712 | 1 | test 27838 | github-to-sqlite 207052882 | actions/checkout@v2 | |||||
138717 | 1 | deploy 27839 | github-to-sqlite 207052882 | actions/checkout@v2 | |||||
138725 | 1 | test 27841 | github-to-sqlite 207052882 | actions/checkout@v2 | |||||
138730 | 1 | test 27842 | hacker-news-to-sqlite 248903544 | actions/checkout@v2 | |||||
138735 | 1 | deploy 27843 | hacker-news-to-sqlite 248903544 | actions/checkout@v2 | |||||
138740 | 1 | test 27844 | hacker-news-to-sqlite 248903544 | actions/checkout@v2 | |||||
138745 | 1 | test 27845 | healthkit-to-sqlite 197882382 | actions/checkout@v2 | |||||
138750 | 1 | deploy 27846 | healthkit-to-sqlite 197882382 | actions/checkout@v2 | |||||
138755 | 1 | test 27847 | healthkit-to-sqlite 197882382 | actions/checkout@v2 | |||||
138760 | 1 | test 27848 | pocket-to-sqlite 213286752 | actions/checkout@v3 | |||||
138764 | 1 | deploy 27849 | pocket-to-sqlite 213286752 | actions/checkout@v3 | |||||
138768 | 1 | test 27850 | pocket-to-sqlite 213286752 | actions/checkout@v3 | |||||
138772 | 1 | test 27851 | swarm-to-sqlite 205429375 | actions/checkout@v3 | |||||
138776 | 1 | deploy 27852 | swarm-to-sqlite 205429375 | actions/checkout@v3 | |||||
138780 | 1 | test 27853 | swarm-to-sqlite 205429375 | actions/checkout@v3 | |||||
138784 | 1 | test 27854 | twitter-to-sqlite 206156866 | actions/checkout@v2 | |||||
138789 | 1 | deploy 27855 | twitter-to-sqlite 206156866 | actions/checkout@v2 | |||||
138797 | 1 | test 27857 | twitter-to-sqlite 206156866 | actions/checkout@v2 | |||||
138802 | 1 | deploy-branch-preview 27858 | datasette 107914493 | actions/checkout@v3 | |||||
138818 | 1 | documentation-links 27860 | datasette 107914493 | readthedocs/actions/preview@v1 | { "project-slug": "datasette" } |
||||
138823 | 1 | test 27862 | datasette 107914493 | actions/checkout@v3 | |||||
138828 | 1 | deploy 27863 | datasette 107914493 | actions/checkout@v3 | |||||
138833 | 1 | deploy_static_docs 27864 | datasette 107914493 | actions/checkout@v2 | |||||
138840 | 1 | deploy_docker 27865 | datasette 107914493 | actions/checkout@v2 | |||||
138842 | 1 | deploy_docker 27866 | datasette 107914493 | actions/checkout@v2 | |||||
138844 | 1 | spellcheck 27867 | datasette 107914493 | actions/checkout@v4 | |||||
138854 | 1 | test 27869 | datasette 107914493 | actions/checkout@v3 | |||||
138859 | 1 | test 27870 | datasette 107914493 | actions/checkout@v3 | |||||
138869 | 1 | build 27871 | datasette 107914493 | actions/checkout@v2 | |||||
138871 | 1 | build 27872 | datasette 107914493 | actions/checkout@v2 | |||||
138874 | 2 | Analyze 27873 | sqlite-utils 140912432 | git checkout HEAD^2 | ${{ github.event_name == 'pull_request' }} | ||||
138878 | 1 | documentation-links 27874 | sqlite-utils 140912432 | readthedocs/actions/preview@v1 | { "project-slug": "sqlite-utils" } |
||||
138879 | 1 | test 27875 | sqlite-utils 140912432 | actions/checkout@v3 | |||||
138884 | 1 | deploy 27876 | sqlite-utils 140912432 | actions/checkout@v3 | |||||
138889 | 1 | spellcheck 27877 | sqlite-utils 140912432 | actions/checkout@v2 | |||||
138901 | 1 | test 27879 | sqlite-utils 140912432 | actions/checkout@v3 | |||||
138814 | 9 | deploy 27859 | datasette 107914493 | And the counters writable canned query demo | cat > plugins/counters.py <<EOF from datasette import hookimpl @hookimpl def startup(datasette): db = datasette.add_memory_database("counters") async def inner(): await db.execute_write("create table if not exists counters (name text primary key, value integer)") await db.execute_write("insert or ignore into counters (name, value) values ('counter_a', 0)") await db.execute_write("insert or ignore into counters (name, value) values ('counter_b', 0)") await db.execute_write("insert or ignore into counters (name, value) values ('counter_c', 0)") return inner @hookimpl def canned_queries(database): if database == "counters": queries = {} for name in ("counter_a", "counter_b", "counter_c"): queries["increment_{}".format(name)] = { "sql": "update counters set value = value + 1 where name = '{}'".format(name), "on_success_message_sql": "select 'Counter {name} incremented to ' || value from counters where name = '{name}'".format(name=name), "write": True, } queries["decrement_{}".format(name)] = { "sql": "update counters set value = value - 1 where name = '{}'".format(name), "on_success_message_sql": "select 'Counter {name} decremented to ' || value from counters where name = '{name}'".format(name=name), "write": True, } return queries EOF | ||||
138876 | 4 | Analyze 27873 | sqlite-utils 140912432 | github/codeql-action/autobuild@v1 | Autobuild | ||||
138841 | 2 | deploy_docker 27865 | datasette 107914493 | Build and push to Docker Hub | sleep 60 # Give PyPI time to make the new release available docker login -u $DOCKER_USER -p $DOCKER_PASS export REPO=datasetteproject/datasette docker build -f Dockerfile \ -t $REPO:${GITHUB_REF#refs/tags/} \ --build-arg VERSION=${GITHUB_REF#refs/tags/} . docker tag $REPO:${GITHUB_REF#refs/tags/} $REPO:latest docker push $REPO:${GITHUB_REF#refs/tags/} docker push $REPO:latest | { "DOCKER_USER": "${{ secrets.DOCKER_USER }}", "DOCKER_PASS": "${{ secrets.DOCKER_PASS }}" } |
|||
138843 | 2 | deploy_docker 27866 | datasette 107914493 | Build and push to Docker Hub | docker login -u $DOCKER_USER -p $DOCKER_PASS export REPO=datasetteproject/datasette docker build -f Dockerfile \ -t $REPO:${VERSION_TAG} \ --build-arg VERSION=${VERSION_TAG} . docker push $REPO:${VERSION_TAG} | { "DOCKER_USER": "${{ secrets.DOCKER_USER }}", "DOCKER_PASS": "${{ secrets.DOCKER_PASS }}", "VERSION_TAG": "${{ github.event.inputs.version_tag }}" } |
|||
138812 | 7 | deploy 27859 | datasette 107914493 | Build docs.db | cd docs DISABLE_SPHINX_INLINE_TABS=1 sphinx-build -b xml . _build sphinx-to-sqlite ../docs.db _build cd .. | ${{ github.ref == 'refs/heads/main' }} | |||
138837 | 5 | deploy_static_docs 27864 | datasette 107914493 | Build docs.db | cd docs DISABLE_SPHINX_INLINE_TABS=1 sphinx-build -b xml . _build sphinx-to-sqlite ../docs.db _build cd .. | ||||
138862 | 4 | test 27870 | datasette 107914493 | Build extension for --load-extension test | (cd tests && gcc ext.c -fPIC -shared -o ext.so) | ||||
138909 | 9 | test 27879 | sqlite-utils 140912432 | Build extension for --load-extension test | (cd tests && gcc ext.c -fPIC -shared -o ext.so && ls -lah) | matrix.os == 'ubuntu-latest' | |||
138811 | 6 | deploy 27859 | datasette 107914493 | Build fixtures.db and other files needed to deploy the demo | python tests/fixtures.py \ fixtures.db \ fixtures-config.json \ fixtures-metadata.json \ plugins \ --extra-db-filename extra_database.db | ||||
138856 | 3 | test 27869 | datasette 107914493 | actions/cache@v2 | Cache Playwright browsers | { "path": "~/.cache/ms-playwright/", "key": "${{ runner.os }}-browsers" } |
|||
138913 | 13 | test 27879 | sqlite-utils 140912432 | Check formatting | black . --check | ||||
138665 | 5 | test 27829 | apple-notes-to-sqlite 611552758 | Check if README is up-to-date | cog --check README.md | ||||
138867 | 9 | test 27870 | datasette 107914493 | Check if blacken-docs needs to be run | # This fails on syntax errors, or a diff was applied blacken-docs -l 60 docs/*.rst | matrix.python-version != '3.8' | |||
138866 | 8 | test 27870 | datasette 107914493 | Check if cog needs to be run | cog --check docs/*.rst | matrix.python-version != '3.8' | |||
138914 | 14 | test 27879 | sqlite-utils 140912432 | Check if cog needs to be run | cog --check README.md docs/*.rst | matrix.python-version != '3.7' | |||
138806 | 1 | deploy 27859 | datasette 107914493 | actions/checkout@v3 | Check out datasette | ||||
138848 | 1 | test 27868 | datasette 107914493 | actions/checkout@v2 | Check out datasette | ||||
138701 | 1 | scheduled 27837 | github-to-sqlite 207052882 | actions/checkout@v2 | Check out repo | ||||
138722 | 1 | build 27840 | github-to-sqlite 207052882 | actions/checkout@v2 | Check out repo | ||||
138794 | 1 | build 27856 | twitter-to-sqlite 206156866 | actions/checkout@v2 | Check out repo | ||||
138819 | 1 | prettier 27861 | datasette 107914493 | actions/checkout@v2 | Check out repo | ||||
138894 | 1 | test 27878 | sqlite-utils 140912432 | actions/checkout@v2 | Check out repo | ||||
138847 | 4 | spellcheck 27867 | datasette 107914493 | Check spelling | codespell README.md --ignore-words docs/codespell-ignore-words.txt codespell docs/*.rst --ignore-words docs/codespell-ignore-words.txt codespell datasette -S datasette/static --ignore-words docs/codespell-ignore-words.txt | ||||
138893 | 5 | spellcheck 27877 | sqlite-utils 140912432 | Check spelling | codespell docs/*.rst --ignore-words docs/codespell-ignore-words.txt codespell sqlite_utils --ignore-words docs/codespell-ignore-words.txt | ||||
138873 | 1 | Analyze 27873 | sqlite-utils 140912432 | actions/checkout@v2 | Checkout repository | { "fetch-depth": 2 } |
|||
138724 | 3 | build 27840 | github-to-sqlite 207052882 | Commit and push if README changed | git diff git config --global user.email "readme-bot@example.com" git config --global user.name "README-bot" git diff --quiet || (git add README.md && git commit -m "Updated README") git push | ||||
138796 | 3 | build 27856 | twitter-to-sqlite 206156866 | Commit and push if README changed | git diff git config --global user.email "readme-bot@example.com" git config --global user.name "README-bot" git diff --quiet || (git add README.md && git commit -m "Updated README") git push | ||||
138820 | 2 | prettier 27861 | datasette 107914493 | actions/cache@v2 | Configure npm caching | { "path": "~/.npm", "key": "${{ runner.OS }}-npm-${{ hashFiles('**/package-lock.json') }}", "restore-keys": "${{ runner.OS }}-npm-\n" } |
|||
138668 | 3 | test 27830 | dogsheep-beta 197431109 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138673 | 3 | deploy 27831 | dogsheep-beta 197431109 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-publish-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-publish-pip-\n" } |
|||
138678 | 3 | test 27832 | dogsheep-beta 197431109 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138683 | 3 | build 27833 | dogsheep-photos 256834907 | actions/cache@v1 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138688 | 3 | test 27834 | evernote-to-sqlite 303218369 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138693 | 3 | deploy 27835 | evernote-to-sqlite 303218369 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-publish-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-publish-pip-\n" } |
|||
138698 | 3 | test 27836 | evernote-to-sqlite 303218369 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138703 | 3 | scheduled 27837 | github-to-sqlite 207052882 | actions/cache@v1 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138714 | 3 | test 27838 | github-to-sqlite 207052882 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138719 | 3 | deploy 27839 | github-to-sqlite 207052882 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-publish-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-publish-pip-\n" } |
|||
138727 | 3 | test 27841 | github-to-sqlite 207052882 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138732 | 3 | test 27842 | hacker-news-to-sqlite 248903544 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138737 | 3 | deploy 27843 | hacker-news-to-sqlite 248903544 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-publish-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-publish-pip-\n" } |
|||
138742 | 3 | test 27844 | hacker-news-to-sqlite 248903544 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138747 | 3 | test 27845 | healthkit-to-sqlite 197882382 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138752 | 3 | deploy 27846 | healthkit-to-sqlite 197882382 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-publish-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-publish-pip-\n" } |
|||
138757 | 3 | test 27847 | healthkit-to-sqlite 197882382 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138786 | 3 | test 27854 | twitter-to-sqlite 206156866 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138791 | 3 | deploy 27855 | twitter-to-sqlite 206156866 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-publish-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-publish-pip-\n" } |
|||
138799 | 3 | test 27857 | twitter-to-sqlite 206156866 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138808 | 3 | deploy 27859 | datasette 107914493 | actions/cache@v3 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138825 | 3 | test 27862 | datasette 107914493 | actions/cache@v3 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
|||
138830 | 3 | deploy 27863 | datasette 107914493 | actions/cache@v3 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-publish-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-publish-pip-\n" } |
|||
138835 | 3 | deploy_static_docs 27864 | datasette 107914493 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-publish-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-publish-pip-\n" } |
|||
138850 | 3 | test 27868 | datasette 107914493 | actions/cache@v2 | Configure pip caching | { "path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n" } |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [steps] ( [id] INTEGER PRIMARY KEY, [seq] INTEGER, [job] INTEGER REFERENCES [jobs]([id]), [repo] INTEGER REFERENCES [repos]([id]), [uses] TEXT, [name] TEXT, [with] TEXT, [run] TEXT , [env] TEXT, [if] TEXT); CREATE INDEX [idx_steps_repo] ON [steps] ([repo]); CREATE INDEX [idx_steps_job] ON [steps] ([job]);