issue_comments
1 row where author_association = "OWNER" and "created_at" is on date 2019-04-09 sorted by updated_at descending
This data as json, CSV (advanced)
Suggested facets: updated_at (date)
id | html_url | issue_url | node_id | user | created_at | updated_at ▲ | author_association | body | reactions | issue | performed_via_github_app |
---|---|---|---|---|---|---|---|---|---|---|---|
481310295 | https://github.com/simonw/datasette/issues/420#issuecomment-481310295 | https://api.github.com/repos/simonw/datasette/issues/420 | MDEyOklzc3VlQ29tbWVudDQ4MTMxMDI5NQ== | simonw 9599 | 2019-04-09T15:50:52Z | 2019-04-09T15:50:52Z | OWNER | Efficient row counts are even more important for the The row counts on those pages don't have to be precise, so one option is for me to calculate them and cache them occasionally. I could even have a dedicated thread which just does the counting? In #422 I've figured out a mechanism for getting accurate or lower-bound counts within a time limit (accurate if possible, lower-bound otherwise). |
{ "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 } |
Fix all the places that currently use .inspect() data 421971339 |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [issue_comments] ( [html_url] TEXT, [issue_url] TEXT, [id] INTEGER PRIMARY KEY, [node_id] TEXT, [user] INTEGER REFERENCES [users]([id]), [created_at] TEXT, [updated_at] TEXT, [author_association] TEXT, [body] TEXT, [reactions] TEXT, [issue] INTEGER REFERENCES [issues]([id]) , [performed_via_github_app] TEXT); CREATE INDEX [idx_issue_comments_issue] ON [issue_comments] ([issue]); CREATE INDEX [idx_issue_comments_user] ON [issue_comments] ([user]);
user 1