{"html_url": "https://github.com/simonw/datasette/issues/1190#issuecomment-1699925224", "issue_url": "https://api.github.com/repos/simonw/datasette/issues/1190", "id": 1699925224, "node_id": "IC_kwDOBm6k_c5lUszo", "user": {"value": 9599, "label": "simonw"}, "created_at": "2023-08-30T22:16:38Z", "updated_at": "2023-08-30T22:16:38Z", "author_association": "OWNER", "body": "This is going to happen in this tool instead:\r\n- https://github.com/simonw/dclient", "reactions": "{\"total_count\": 1, \"+1\": 0, \"-1\": 0, \"laugh\": 0, \"hooray\": 1, \"confused\": 0, \"heart\": 0, \"rocket\": 0, \"eyes\": 0}", "issue": {"value": 787098146, "label": "`datasette publish upload` mechanism for uploading databases to an existing Datasette instance"}, "performed_via_github_app": null} {"html_url": "https://github.com/simonw/sqlite-utils/issues/532#issuecomment-1539009453", "issue_url": "https://api.github.com/repos/simonw/sqlite-utils/issues/532", "id": 1539009453, "node_id": "IC_kwDOCGYnMM5bu2ut", "user": {"value": 9599, "label": "simonw"}, "created_at": "2023-05-08T20:30:29Z", "updated_at": "2023-05-08T20:30:42Z", "author_association": "OWNER", "body": "Here's an improvement:\r\n```\r\n% sqlite-utils insert /tmp/b.db blah /tmp/blah.txt\r\n [####################################] 100%\r\nError: Invalid JSON - use --csv for CSV or --tsv for TSV files\r\n\r\nJSON error: Expecting value: line 1 column 1 (char 0)\r\n```", "reactions": "{\"total_count\": 1, \"+1\": 0, \"-1\": 0, \"laugh\": 0, \"hooray\": 1, \"confused\": 0, \"heart\": 0, \"rocket\": 0, \"eyes\": 0}", "issue": {"value": 1620254998, "label": "Show more information when JSON can't be imported with sqlite-utils insert"}, "performed_via_github_app": null} {"html_url": "https://github.com/simonw/datasette/issues/1886#issuecomment-1312898318", "issue_url": "https://api.github.com/repos/simonw/datasette/issues/1886", "id": 1312898318, "node_id": "IC_kwDOBm6k_c5OQT0O", "user": {"value": 19851673, "label": "eigenfoo"}, "created_at": "2022-11-14T00:52:16Z", "updated_at": "2022-11-14T00:52:16Z", "author_association": "NONE", "body": "I'm a cryptic crossword enthusiast and have spent a lot of time scraping and parsing cryptic crossword clues from various blogs, forums and publications. The result is over **half a million clues from cryptic crosswords over the past twelve years**, including the clue, answer, puzzle date, puzzle name and a link to the original source. This is all hosted using Datasette, which has been a delight to use: https://cryptics.georgeho.org/\r\n\r\nThis dataset is a significant work of crossword archivism and scholarship, as acquiring historical crosswords and structuring their contents require focused effort and tedious cleaning that few are willing to do for such trivial data - for example, according to [this 2004 selection guide](https://cryptics.georgeho.org/static/documents/Selection_AppendixE_v2.pdf), the Library of Congress explicitly does not collect crossword puzzles. Anecdotally, I know that many constructors/setters of cryptic crosswords use this dataset as a resource, and some even simply call it \"the database\" - this is probably one of the most impactful data projects I've worked on!", "reactions": "{\"total_count\": 1, \"+1\": 0, \"-1\": 0, \"laugh\": 0, \"hooray\": 1, \"confused\": 0, \"heart\": 0, \"rocket\": 0, \"eyes\": 0}", "issue": {"value": 1447050738, "label": "Call for birthday presents: if you're using Datasette, let us know how you're using it here"}, "performed_via_github_app": null} {"html_url": "https://github.com/simonw/datasette/issues/1817#issuecomment-1258756231", "issue_url": "https://api.github.com/repos/simonw/datasette/issues/1817", "id": 1258756231, "node_id": "IC_kwDOBm6k_c5LBxiH", "user": {"value": 9599, "label": "simonw"}, "created_at": "2022-09-26T23:19:34Z", "updated_at": "2022-09-26T23:19:34Z", "author_association": "OWNER", "body": "This is a good idea - it's something I should do before Datasette 1.0.\r\n\r\nI was a tiny bit worried about compatibility (Datasette is 3.7+) but it looks like they have been in Python since 3.0!", "reactions": "{\"total_count\": 1, \"+1\": 0, \"-1\": 0, \"laugh\": 0, \"hooray\": 1, \"confused\": 0, \"heart\": 0, \"rocket\": 0, \"eyes\": 0}", "issue": {"value": 1384273985, "label": "Expose `sql` and `params` arguments to various plugin hooks"}, "performed_via_github_app": null} {"html_url": "https://github.com/simonw/sqlite-utils/pull/480#issuecomment-1232089808", "issue_url": "https://api.github.com/repos/simonw/sqlite-utils/issues/480", "id": 1232089808, "node_id": "IC_kwDOCGYnMM5JcDLQ", "user": {"value": 9599, "label": "simonw"}, "created_at": "2022-08-30T19:43:02Z", "updated_at": "2022-08-30T19:43:02Z", "author_association": "OWNER", "body": "Yeah this seems like a reasonable addition to me. \r\n\r\nNeeds a test, which can go next to this one: https://github.com/simonw/sqlite-utils/blob/087753cd42c406f1e060c1822dcd9b5fda3d60f4/tests/test_fts.py#L561", "reactions": "{\"total_count\": 1, \"+1\": 0, \"-1\": 0, \"laugh\": 0, \"hooray\": 1, \"confused\": 0, \"heart\": 0, \"rocket\": 0, \"eyes\": 0}", "issue": {"value": 1355433619, "label": "search_sql add include_rank option"}, "performed_via_github_app": null} {"html_url": "https://github.com/simonw/sqlite-utils/issues/467#issuecomment-1224386951", "issue_url": "https://api.github.com/repos/simonw/sqlite-utils/issues/467", "id": 1224386951, "node_id": "IC_kwDOCGYnMM5I-qmH", "user": {"value": 9599, "label": "simonw"}, "created_at": "2022-08-23T17:20:07Z", "updated_at": "2022-08-23T17:20:07Z", "author_association": "OWNER", "body": "Example of that prototype working:\r\n```pycon\r\n>>> from sqlite_utils import Database\r\n>>> db = Database(memory=True)\r\n>>> db[\"dogs\"].create({\"id\": int, \"name\": str}, pk=\"id\")\r\n