Add benchmarking test suite and greatly improve performance in a few cases (#948)
* Add benchmarking test suite * Improve amortized time of model relation loads with a large number of rows * Improve performance of loading models with many related models * Improve performance of loading models with many related models to O(N)ish * Fix bug where N model creation with shared related model would build in N^2 time * Lower blocking time for queryset results * Add docstrings and streamline hash code Co-authored-by: haydeec1 <Eric.Haydel@jhuapl.edu>
This commit is contained in:
91
benchmarks/test_benchmark_create.py
Normal file
91
benchmarks/test_benchmark_create.py
Normal file
@ -0,0 +1,91 @@
|
||||
import random
|
||||
import string
|
||||
|
||||
import pytest
|
||||
|
||||
from benchmarks.conftest import Author, Book, Publisher
|
||||
|
||||
pytestmark = pytest.mark.asyncio
|
||||
|
||||
|
||||
@pytest.mark.parametrize("num_models", [10, 20, 40])
|
||||
async def test_creating_models_individually(aio_benchmark, num_models: int):
|
||||
@aio_benchmark
|
||||
async def create(num_models: int):
|
||||
authors = []
|
||||
for idx in range(0, num_models):
|
||||
author = await Author.objects.create(
|
||||
name="".join(random.sample(string.ascii_letters, 5)),
|
||||
score=random.random() * 100,
|
||||
)
|
||||
authors.append(author)
|
||||
return authors
|
||||
|
||||
authors = create(num_models)
|
||||
for author in authors:
|
||||
assert author.id is not None
|
||||
|
||||
|
||||
@pytest.mark.parametrize("num_models", [10, 20, 40])
|
||||
async def test_creating_individually_with_related_models(
|
||||
aio_benchmark, num_models: int, author: Author, publisher: Publisher
|
||||
):
|
||||
@aio_benchmark
|
||||
async def create_with_related_models(
|
||||
author: Author, publisher: Publisher, num_models: int
|
||||
):
|
||||
books = []
|
||||
for idx in range(0, num_models):
|
||||
book = await Book.objects.create(
|
||||
author=author,
|
||||
publisher=publisher,
|
||||
title="".join(random.sample(string.ascii_letters, 5)),
|
||||
year=random.randint(0, 2000),
|
||||
)
|
||||
books.append(book)
|
||||
|
||||
return books
|
||||
|
||||
books = create_with_related_models(
|
||||
author=author, publisher=publisher, num_models=num_models
|
||||
)
|
||||
|
||||
for book in books:
|
||||
assert book.id is not None
|
||||
|
||||
|
||||
@pytest.mark.parametrize("num_models", [10, 20, 40])
|
||||
async def test_get_or_create_when_create(aio_benchmark, num_models: int):
|
||||
@aio_benchmark
|
||||
async def get_or_create(num_models: int):
|
||||
authors = []
|
||||
for idx in range(0, num_models):
|
||||
author, created = await Author.objects.get_or_create(
|
||||
name="".join(random.sample(string.ascii_letters, 5)),
|
||||
score=random.random() * 100,
|
||||
)
|
||||
assert created
|
||||
authors.append(author)
|
||||
return authors
|
||||
|
||||
authors = get_or_create(num_models)
|
||||
for author in authors:
|
||||
assert author.id is not None
|
||||
|
||||
|
||||
@pytest.mark.parametrize("num_models", [10, 20, 40])
|
||||
async def test_update_or_create_when_create(aio_benchmark, num_models: int):
|
||||
@aio_benchmark
|
||||
async def update_or_create(num_models: int):
|
||||
authors = []
|
||||
for idx in range(0, num_models):
|
||||
author = await Author.objects.update_or_create(
|
||||
name="".join(random.sample(string.ascii_letters, 5)),
|
||||
score=random.random() * 100,
|
||||
)
|
||||
authors.append(author)
|
||||
return authors
|
||||
|
||||
authors = update_or_create(num_models)
|
||||
for author in authors:
|
||||
assert author.id is not None
|
||||
Reference in New Issue
Block a user