Opened 3 days ago

Last modified 3 days ago

#36143 new Cleanup/optimization

DatabaseOperations.bulk_batch_size() is overly protective on SQLite in most cases

Reported by: Sarah Boyce Owned by:
Component: Database layer (models, ORM) Version: 5.1
Severity: Normal Keywords:
Cc: Simon Charette Triage Stage: Accepted
Has patch: no Needs documentation: no
Needs tests: no Patch needs improvement: no
Easy pickings: no UI/UX: no

Description (last modified by Sarah Boyce)

DatabaseOperations.bulk_batch_size() for SQLite is dependent on DatabaseFeatures.max_query_params which is currently set as 999.

However, for SQLite versions >= 3.32, this limit is increased to 32,766. See: https://www.sqlite.org/limits.html#max_variable_number

SQLITE_MAX_VARIABLE_NUMBER, which defaults to 999 for SQLite versions prior to 3.32.0 (2020-05-22) or 32766 for SQLite versions after 3.32.0.

This means that is some cases (assuming there isn't another database limit being hit), we can increase the maximum batch size for SQLite.

For illustration purposes, on SQLite >= 3.32, if the value of DatabaseFeatures.max_query_params is increased, the following test would pass:

  • tests/bulk_create/tests.py

    diff --git a/tests/bulk_create/tests.py b/tests/bulk_create/tests.py
    index 7b86a2def5..ce9123ce97 100644
    a b class BulkCreateTests(TestCase):  
    286286        objs = [Country(name=f"Country {i}") for i in range(1000)]
    287287        fields = ["name", "iso_two_letter", "description"]
    288288        max_batch_size = max(connection.ops.bulk_batch_size(fields, objs), 1)
    289         with self.assertNumQueries(ceil(len(objs) / max_batch_size)):
     289        with self.assertNumQueries(1):
    290290            Country.objects.bulk_create(objs, batch_size=max_batch_size + 1)
    291291
    292292    @skipUnlessDBFeature("has_bulk_insert")

PR discussion references:

Ticket which sparked the discussion/discovery: #36118

Change History (3)

comment:1 by Sarah Boyce, 3 days ago

Description: modified (diff)

comment:2 by Jacob Walls, 3 days ago

Triage Stage: UnreviewedAccepted

Sage showed a query to get the most accurate limit based on compilation flags. Some discussion needed about whether and how to do that (e.g. @cached_property? What about people asserting a query count in tests?)

comment:3 by Simon Charette, 3 days ago

I think we should use class level cached properties for that to avoid per-connection overhead (as Python can only be built against one version of SQLite at a time) and that we should expect query counts tests that don't take into account bulk_batch_size return value to require to be adjusted.

Note: See TracTickets for help on using tickets.
Back to Top