issue_comments
4 rows where "created_at" is on date 2020-08-16, issue = 679808124 and "updated_at" is on date 2020-08-16 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: created_at (date), updated_at (date)
issue 1
- Move CI to GitHub Issues · 4 ✖
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | issue | performed_via_github_app |
---|---|---|---|---|---|---|---|---|---|---|---|
674589472 | https://github.com/simonw/datasette/issues/940#issuecomment-674589472 | https://api.github.com/repos/simonw/datasette/issues/940 | MDEyOklzc3VlQ29tbWVudDY3NDU4OTQ3Mg== | simonw 9599 | 2020-08-16T23:05:57Z | 2020-08-16T23:05:57Z | OWNER | When I figure this out I'll update the https://github.com/simonw/datasette-plugin/blob/main/datasette-%7B%7Bcookiecutter.hyphenated%7D%7D/.github/workflows/publish.yml default workflow to do this - right now it runs the tests once on just a single version of Python as part of the package deploy to PyPI step. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Move CI to GitHub Issues 679808124 | |
674589321 | https://github.com/simonw/datasette/issues/940#issuecomment-674589321 | https://api.github.com/repos/simonw/datasette/issues/940 | MDEyOklzc3VlQ29tbWVudDY3NDU4OTMyMQ== | simonw 9599 | 2020-08-16T23:04:34Z | 2020-08-16T23:04:34Z | OWNER |
|
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Move CI to GitHub Issues 679808124 | |
674589035 | https://github.com/simonw/datasette/issues/940#issuecomment-674589035 | https://api.github.com/repos/simonw/datasette/issues/940 | MDEyOklzc3VlQ29tbWVudDY3NDU4OTAzNQ== | simonw 9599 | 2020-08-16T23:02:23Z | 2020-08-16T23:02:23Z | OWNER | I'd like to set these up as different workflows that depend on each other, if that's possible. I want to start three test runs in parallel (on three different Python versions), then if all three pass kick off the PyPI push (without running more tests), then if that passes do the Docker build and push. |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Move CI to GitHub Issues 679808124 | |
674566618 | https://github.com/simonw/datasette/issues/940#issuecomment-674566618 | https://api.github.com/repos/simonw/datasette/issues/940 | MDEyOklzc3VlQ29tbWVudDY3NDU2NjYxOA== | simonw 9599 | 2020-08-16T19:20:58Z | 2020-08-16T19:20:58Z | OWNER | I need to figure out how to build and push the Docker image on releases. Here's the Travis code for that: https://github.com/simonw/datasette/blob/52eabb019d4051084b21524bd0fd9c2731126985/.travis.yml#L38-L47 |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Move CI to GitHub Issues 679808124 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [issue] INTEGER REFERENCES [issues]([id]) , [performed_via_github_app] TEXT); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
user 1