Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Remove leaky lru_cache #2277

Merged
merged 2 commits into from
Jun 6, 2024
Merged

Conversation

mrcljx
Copy link
Contributor

@mrcljx mrcljx commented Jun 5, 2024

Problem

After calling the endpoint, the instance db_session was not freed, causing a warning to be emitted.

def endpoint(request):
  data = MySchema(context={"db_session": create_db_session()}).load(request.json)
  return "OK"

The reason is that the MySchema instance is kept in a global LRU cache. Calling endpoint multiple times will eventually release the older db_session instances, though.

Analysis

The @lru_cache(max_size=8) decorator was used on a method, causing instances of self to be cached and outlive their intended scope. This led to unexpected behavior where instances persisted beyond a single web request, retaining references and consuming memory unnecessarily.

This problem was called out by PyCQA/flake8-bugbear#310 but ignored via noqa.

Options

  1. Use cachetools.cachedmethod (requires a third party dependency)
  2. Implement manually
  3. Remove LRU cache

Proposal

The LRU cache was introduced in #1309 (the PR reported 3% performance increase on the test base). I'm not sure the LRU cache actually has that much of an effect in real world scenarios:

  1. The cost of two tuple constructions and two dict lookups is small in the first place
  2. If more than 2 schema instances are used, max_size is exhausted, leading to cache misses.
    • A cold load causes 3 misses, 0 hits (warm -> 3 hits)
    • A cold dump causes 2 misses, 0 hits (warm -> 3 hits).
  3. A more complex schema guarantees 100% misses, as too many instances are involved.
class Child(ma.Schema):
  pass

class Parent(ma.Schema):
  a = ma.fields.Nested(Child)
  b = ma.fields.Nested(Child)
  c = ma.fields.Nested(Child)

ma.Schema._has_processors.cache_clear()
p = Parent()
p.load({"a": {}, "b": {}, "c": {}})
p.load({"a": {}, "b": {}, "c": {}})
ma.Schema._has_processors.cache_info() # CacheInfo(hits=0, misses=24, maxsize=8, currsize=8)

Copy link
Member

@sloria sloria left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

good catch! i hadn't considered that side effect of context objects remaining in the global LRU cache. and i generally agree that the CPU cost in typical usage scenarios is small.

i'm good with this, but i'll give a moment to @lafrech @deckar01 to look if they want. otherwise i'll plan to merge and release this over the weekend

@sloria
Copy link
Member

sloria commented Jun 5, 2024

in the meantime, do you mind adding yourself to AUTHORS.rst please?

Copy link
Member

@lafrech lafrech left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I agree with the rationale.

Thanks.

@deckar01
Copy link
Member

deckar01 commented Jun 5, 2024

TLDR: 👍

I tested this using our dump benchmark. Indexing dict with a tuple is pretty slow, but that's mostly what lru_cache does under the hood, so it can't help much. lru_cache is actually significantly slower in pypy. I also benchmarked a hacky refactor that replaces the tuple keys with separate dicts for many hooks and tag-only keys.

(usec/dump) cpython (3.10) pypy (3.10)
mrcljx:remove-lru-cache 475 46
marshmallow-code:dev 475 51
deckar01:fast-hook-key 470 36

The benchmark shows ~30% of the hot path is tuple hashing, at least in pypy. I'm not sure the 1% cpython difference is worth the increased code complexity of my current refactor. Something to punt to a follow up issue maybe.

@mrcljx mrcljx force-pushed the remove-lru-cache branch from 2586230 to 4d165a8 Compare June 5, 2024 23:34
@mrcljx
Copy link
Contributor Author

mrcljx commented Jun 5, 2024

Thanks for the benchmark! I updated AUTHORS.rst.

@sloria sloria enabled auto-merge (squash) June 6, 2024 01:50
@sloria sloria merged commit 99103a6 into marshmallow-code:dev Jun 6, 2024
7 of 8 checks passed
@mrcljx mrcljx deleted the remove-lru-cache branch June 6, 2024 06:51
@deckar01 deckar01 mentioned this pull request Jun 6, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants