home / github

Menu
  • Search all tables
  • GraphQL API

issue_comments

Table actions
  • GraphQL API for issue_comments

14 rows where issue = 1079149656, "updated_at" is on date 2021-12-18 and user = 9599 sorted by updated_at descending

✎ View and edit SQL

This data as json, CSV (advanced)

Suggested facets: created_at (date), updated_at (date)

user 1

  • simonw · 14 ✖

issue 1

  • Optimize all those calls to index_list and foreign_key_list · 14 ✖

author_association 1

  • OWNER 14
id html_url issue_url node_id user created_at updated_at ▲ author_association body reactions issue performed_via_github_app
997272223 https://github.com/simonw/datasette/issues/1555#issuecomment-997272223 https://api.github.com/repos/simonw/datasette/issues/1555 IC_kwDOBm6k_c47cSqf simonw 9599 2021-12-18T19:17:13Z 2021-12-18T19:17:13Z OWNER

That's a good optimization. Still need to deal with the huge flurry of PRAGMA queries though before I can consider this done.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
Optimize all those calls to index_list and foreign_key_list 1079149656  
997267416 https://github.com/simonw/datasette/issues/1555#issuecomment-997267416 https://api.github.com/repos/simonw/datasette/issues/1555 IC_kwDOBm6k_c47cRfY simonw 9599 2021-12-18T18:44:53Z 2021-12-18T18:45:28Z OWNER

Rather than adding a executemany=True parameter, I'm now thinking a better design might be to have three methods:

  • db.execute_write(sql, params=None, block=False)
  • db.execute_writescript(sql, block=False)
  • db.execute_writemany(sql, params_seq, block=False)
{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
Optimize all those calls to index_list and foreign_key_list 1079149656  
997266100 https://github.com/simonw/datasette/issues/1555#issuecomment-997266100 https://api.github.com/repos/simonw/datasette/issues/1555 IC_kwDOBm6k_c47cRK0 simonw 9599 2021-12-18T18:40:02Z 2021-12-18T18:40:02Z OWNER

The implementation of cursor.executemany() looks very efficient - it turns into a call to this C function with multiple set to 1: https://github.com/python/cpython/blob/e002bbc6cce637171fb2b1391ffeca8643a13843/Modules/_sqlite/cursor.c#L468-L469

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
Optimize all those calls to index_list and foreign_key_list 1079149656  
997262475 https://github.com/simonw/datasette/issues/1555#issuecomment-997262475 https://api.github.com/repos/simonw/datasette/issues/1555 IC_kwDOBm6k_c47cQSL simonw 9599 2021-12-18T18:34:18Z 2021-12-18T18:34:18Z OWNER

Using executescript=True that call now takes 1.89ms to create all of those tables.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
Optimize all those calls to index_list and foreign_key_list 1079149656  
997248364 https://github.com/simonw/datasette/issues/1555#issuecomment-997248364 https://api.github.com/repos/simonw/datasette/issues/1555 IC_kwDOBm6k_c47cM1s simonw 9599 2021-12-18T18:20:10Z 2021-12-18T18:20:10Z OWNER

Idea: teach execute_write to accept an optional executescript=True parameter, like this: ```diff diff --git a/datasette/database.py b/datasette/database.py index 468e936..1a424f5 100644 --- a/datasette/database.py +++ b/datasette/database.py @@ -94,10 +94,14 @@ class Database: f"file:{self.path}{qs}", uri=True, check_same_thread=False )

  • async def execute_write(self, sql, params=None, block=False):
  • async def execute_write(self, sql, params=None, executescript=False, block=False):
  • assert not executescript and params, "Cannot use params with executescript=True" def _inner(conn): with conn:
  • return conn.execute(sql, params or [])
  • if executescript:
  • return conn.executescript(sql)
  • else:
  • return conn.execute(sql, params or [])
     with trace("sql", database=self.name, sql=sql.strip(), params=params):
         results = await self.execute_write_fn(_inner, block=block)
    

    ```

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
Optimize all those calls to index_list and foreign_key_list 1079149656  
997245301 https://github.com/simonw/datasette/issues/1555#issuecomment-997245301 https://api.github.com/repos/simonw/datasette/issues/1555 IC_kwDOBm6k_c47cMF1 simonw 9599 2021-12-18T18:17:04Z 2021-12-18T18:17:04Z OWNER

One downside of conn.executescript() is that it won't be picked up by the tracing mechanism - in fact nothing that uses await db.execute_write_fn(fn, block=True) or await db.execute_fn(fn, block=True) gets picked up by tracing.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
Optimize all those calls to index_list and foreign_key_list 1079149656  
997241969 https://github.com/simonw/datasette/issues/1555#issuecomment-997241969 https://api.github.com/repos/simonw/datasette/issues/1555 IC_kwDOBm6k_c47cLRx simonw 9599 2021-12-18T18:13:04Z 2021-12-18T18:13:04Z OWNER

Also: running all of those CREATE TABLE IF NOT EXISTS in a single call to conn.executescript() rather than as separate queries may speed things up too.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
Optimize all those calls to index_list and foreign_key_list 1079149656  
997241645 https://github.com/simonw/datasette/issues/1555#issuecomment-997241645 https://api.github.com/repos/simonw/datasette/issues/1555 IC_kwDOBm6k_c47cLMt simonw 9599 2021-12-18T18:12:26Z 2021-12-18T18:12:26Z OWNER

A simpler optimization would be just to turn all of those column and index reads into a single efficient UNION query against each database, then figure out the most efficient pattern to send them all as writes in one go as opposed to calling .execute_write() in a loop.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
Optimize all those calls to index_list and foreign_key_list 1079149656  
997235086 https://github.com/simonw/datasette/issues/1555#issuecomment-997235086 https://api.github.com/repos/simonw/datasette/issues/1555 IC_kwDOBm6k_c47cJmO simonw 9599 2021-12-18T17:30:13Z 2021-12-18T17:30:13Z OWNER

Now that trace sees write queries (#1568) it's clear that there is a whole lot more DB activity then I had realized:

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
Optimize all those calls to index_list and foreign_key_list 1079149656  
997234858 https://github.com/simonw/datasette/issues/1555#issuecomment-997234858 https://api.github.com/repos/simonw/datasette/issues/1555 IC_kwDOBm6k_c47cJiq simonw 9599 2021-12-18T17:28:44Z 2021-12-18T17:28:44Z OWNER

Maybe it would be worth exploring attaching each DB in turn to the _internal connection in order to perform these queries faster.

I'm a bit worried about leaks though: the internal database isn't meant to be visible, even temporarily attaching another DB to it could cause SQL queries against that DB to be able to access the internal data.

So maybe instead the _internal connection gets to connect to the other DBs? There's a maximum of ten there I think, which is good for most but not all cases. But the cases with the most connected databases will see the worst performance!

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
Optimize all those calls to index_list and foreign_key_list 1079149656  
997128508 https://github.com/simonw/datasette/issues/1555#issuecomment-997128508 https://api.github.com/repos/simonw/datasette/issues/1555 IC_kwDOBm6k_c47bvk8 simonw 9599 2021-12-18T02:33:57Z 2021-12-18T02:33:57Z OWNER

Here's why - trace only applies to read, not write SQL operations: https://github.com/simonw/datasette/blob/7c8f8aa209e4ba7bf83976f8495d67c28fbfca24/datasette/database.py#L209-L211

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
Optimize all those calls to index_list and foreign_key_list 1079149656  
997128368 https://github.com/simonw/datasette/issues/1555#issuecomment-997128368 https://api.github.com/repos/simonw/datasette/issues/1555 IC_kwDOBm6k_c47bviw simonw 9599 2021-12-18T02:32:43Z 2021-12-18T02:32:43Z OWNER

I wonder why the INSERT INTO queries don't show up in that ?trace=1 view?

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
Optimize all those calls to index_list and foreign_key_list 1079149656  
997128251 https://github.com/simonw/datasette/issues/1555#issuecomment-997128251 https://api.github.com/repos/simonw/datasette/issues/1555 IC_kwDOBm6k_c47bvg7 simonw 9599 2021-12-18T02:31:51Z 2021-12-18T02:31:51Z OWNER

I was thinking it might even be possible to convert this into a insert into tables select from ... query:

https://github.com/simonw/datasette/blob/c00f29affcafce8314366852ba1a0f5a7dd25690/datasette/utils/internal_db.py#L102-L112

But the SELECT runs against a separate database from the INSERT INTO, so I would have to setup a cross-database connection for this which feels a little too complicated.

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
Optimize all those calls to index_list and foreign_key_list 1079149656  
997128080 https://github.com/simonw/datasette/issues/1555#issuecomment-997128080 https://api.github.com/repos/simonw/datasette/issues/1555 IC_kwDOBm6k_c47bveQ simonw 9599 2021-12-18T02:30:19Z 2021-12-18T02:30:19Z OWNER

I think all of these queries happen in one place - in the populate_schema_tables() function - so optimizing them might be localized to just that area of the code, which would be nice:

https://github.com/simonw/datasette/blob/c00f29affcafce8314366852ba1a0f5a7dd25690/datasette/utils/internal_db.py#L97-L183

{
    "total_count": 0,
    "+1": 0,
    "-1": 0,
    "laugh": 0,
    "hooray": 0,
    "confused": 0,
    "heart": 0,
    "rocket": 0,
    "eyes": 0
}
Optimize all those calls to index_list and foreign_key_list 1079149656  

Advanced export

JSON shape: default, array, newline-delimited, object

CSV options:

CREATE TABLE [issue_comments] (
   [html_url] TEXT,
   [issue_url] TEXT,
   [id] INTEGER PRIMARY KEY,
   [node_id] TEXT,
   [user] INTEGER REFERENCES [users]([id]),
   [created_at] TEXT,
   [updated_at] TEXT,
   [author_association] TEXT,
   [body] TEXT,
   [reactions] TEXT,
   [issue] INTEGER REFERENCES [issues]([id])
, [performed_via_github_app] TEXT);
CREATE INDEX [idx_issue_comments_issue]
                ON [issue_comments] ([issue]);
CREATE INDEX [idx_issue_comments_user]
                ON [issue_comments] ([user]);
Powered by Datasette · Queries took 483.545ms · About: github-to-sqlite