-
Notifications
You must be signed in to change notification settings - Fork 2.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Improve dependency resolution performance by not clearing caches when backtracking #7950
Improve dependency resolution performance by not clearing caches when backtracking #7950
Conversation
Wow, this looks quite promising. However, in some cases, it seems dependency resolution takes significantly longer now. Example pyproject.toml[tool.poetry]
name = "app"
version = "0.1.0"
description = ""
authors = ["Your Name <[email protected]>"]
[tool.poetry.dependencies]
python = "^3.10"
beautifulsoup4 = ">=4.7.1"
boto3 = ">=1.22.12"
botocore = ">=1.25.12"
celery = ">=4.4.7"
click = ">=8.0.4"
confluent-kafka = ">=1.9.2"
croniter = ">=0.3.37"
datadog = ">=0.29.3"
django-crispy-forms = ">=1.14.0"
django-pg-zero-downtime-migrations = ">=0.11"
Django = ">=2.2.28"
djangorestframework = ">=3.12.4"
drf-spectacular = ">=0.22.1"
email-reply-parser = ">=0.5.12"
google-api-core = ">=2.10.1"
google-auth = ">=1.35.0"
google-cloud-bigtable = ">=2.11.3"
google-cloud-core = ">=2.3.2"
google-cloud-functions = ">=1.8.1"
google-cloud-pubsub = ">=2.13.6"
google-cloud-spanner = ">=3.20.0"
google-cloud-storage = ">=2.5.0"
googleapis-common-protos = ">=1.56.4"
google-crc32c = ">=1.3.0"
isodate = ">=0.6.1"
jsonschema = ">=3.2.0"
lxml = ">=4.6.5"
maxminddb = ">=2.0.3"
mistune = ">=2.0.3"
mmh3 = ">=3.0.0"
packaging = ">=21.3"
parsimonious = ">=0.8.0"
petname = ">=2.6"
phonenumberslite = ">=8.12.0"
Pillow = ">=9.2.0"
progressbar2 = ">=3.41.0"
python-rapidjson = ">=1.4"
psycopg2-binary = ">=2.8.6"
PyJWT = ">=2.4.0"
python-dateutil = ">=2.8.1"
python-memcached = ">=1.59"
python-u2flib-server = ">=5.0.0"
fido2 = ">=0.9.2"
python3-saml = ">=1.14.0"
PyYAML = ">=5.4"
rb = ">=1.9.0"
redis-py-cluster = ">=2.1.0"
redis = ">=3.4.1"
requests-oauthlib = ">=1.2.0"
requests = ">=2.25.1"
rfc3339-validator = ">=0.1.2"
rfc3986-validator = ">=0.1.1"
sentry-arroyo = ">=2.4.0"
sentry-relay = ">=0.8.15"
sentry-sdk = ">=1.11.0"
snuba-sdk = ">=1.0.3"
simplejson = ">=3.17.6"
statsd = ">=3.3"
structlog = ">=21.1.0"
symbolic = ">=10.2.0"
toronado = ">=0.1.0"
typing-extensions = ">=3.10.0.2"
ua-parser = ">=0.10.0"
unidiff = ">=0.7.4"
urllib3 = { version = ">=1.26.9,<2", extras = ["brotli"] }
brotli = ">=1.0.9"
pyuwsgi = "==2.0.20.0"
zstandard = ">=0.18.0"
msgpack = ">=1.0.4"
cryptography = ">=38.0.3"
billiard = ">=3.6.4"
kombu = ">=4.6.11"
grpcio = ">=1.47.0"
hiredis = ">=0.3.1"
cffi = ">=1.15.0"
cssutils = ">=2.4.0"
cssselect = ">=1.0.3"
phabricator = ">=0.7.0"
selenium = ">=4.1.5"
sqlparse = ">=0.2.4,<=0.3.0"
In another project (unfortunately containing some private dependencies) time for dependency resolution increases from about 200 s to 360 s. (Measurements with warm cache.) |
speculating wildly: my first worry on looking at this is the linear scan through the many caches. Perhaps that's what the problem is in the examples where things get worse |
(I think that what is implemented here could more compactly be a https://docs.python.org/3/library/collections.html#collections.ChainMap - but I don't think that would help with the linear scan, if that is indeed the problem) |
Agreed, I think that's probably the cause. I initially wrote this change with one big set that I updated rather than the linear scan of many caches, but changed to this approach since the performance seemed the same in my tests and it was a bit simpler. Most of my tests were with a small number of packages though, so I bet it makes a much larger difference in these cases where the decision levels get much larger. I'll take a look at this later today and see if that approach works better. Thanks for providing that repro! |
Regarding the difference with pre-releasesWe allow pre-releases for a constraint that can only be satisfied by a pre-release even if pre-releases were not allowed explicity, i.e.
The upper branch uses poetry/src/poetry/repositories/repository.py Lines 46 to 61 in 277fa60
I don't think the bottom branch filters them out per se, but only if they are not allowed by the constraint. The branch uses
This change may be problematic if only pre-releases can satisfy a constraint. Maybe, you can verify the pre-release filtering with an example constraint like |
I've made two changes which should improve things and pushed these as separate commits. The main difference seemed to actually be the Here are the timings on my machine for your reproduction (testing with Timings for repro from #7950 (comment)
For me this gets the performance back down to matching master, but if you're able to re-test with these changes (on this repro or your private repo) that would be appreciated. It could be that some combinations of CPU or IO speed cause this branch to regress timings on some machines but not others. You may want to try at least 2 runs as I noticed the first run is often 1-2 seconds slower for me. As you can see in the table above, the Timings for
For now I decided to keep it due to the small above improvement, but if it causes problems it wouldn't hurt too much to remove. (FYI, I updated my test assertions branch at https://github.com/chriskuehl/poetry/commits/dependency-resolution-perf-with-tests-v2 for the updated PR.) |
Ahh, thanks for the explanation. That helps a lot -- I assumed it must have been filtering those out in subsequent iterations, but it's just that the function above may return additional entries as the constraint gets tighter. I think the existing cache implementation has some inconsistencies around pre-release package version which may make it hard to exactly match its behavior. The bottom branch does filtering only and doesn't have any kind of "if package version list is empty, add in compatible pre-release versions" logic, so its results may not match if the cache was disabled and we always called For example if I have a constraint on black that forces a pre-release version at the top level: [tool.poetry.dependencies]
python = "^3.11"
black = "<22" ...then I can successfully But if instead I have the same restrictive black dependency declared at the second level, like this: [tool.poetry.dependencies]
python = "^3.11"
black = "*"
# This is a wheel I created that has `black<22` as its only requirement.
pkg-that-requires-old-black = {path = "test-pkg/dist/pkg_that_requires_old_black-1.0.0-py3-none-any.whl"} then
If I disable the I think I could fix this in this PR by adding some logic like: diff --git a/src/poetry/mixology/version_solver.py b/src/poetry/mixology/version_solver.py
index ec44b0fa..c149508c 100644
--- a/src/poetry/mixology/version_solver.py
+++ b/src/poetry/mixology/version_solver.py
@@ -82,7 +82,7 @@ class DependencyCache:
for p in cache_entries[-1]
if dependency.constraint.allows(p.package.version)
]
- else:
+ if not cache_entries or not packages:
packages = self.provider.search_for(dependency)
return packages That patch needs more testing and to verify its performance impact, but seems to work (fixes the install error for the repro above with Do you think it makes sense to add that to this branch? It does seem more correct to me, but it won't match 100% the behavior in master. |
I can confirm that the time for dependency resolution for the repro is as fast as on master. Testing with my private repo (multiple runs), it's still about 5 seconds slower than master. However, I think that's neglectable considering the resolution time of about 200 s.
Good catch. I'd say that's a bug.
I just tested the change with some projects and did not experience any performance regression. If it's the same for you, it should be good.
It would probably be the best, if we add the fix and a test in a separate PR (before this one?). The fix should be similar without the performance improvement of this PR. A test in |
By the way, I tend to keep the 4 commits separated when merging this into master, so please rebase (not merge) in case you want to update your branch (and don't squash the commits 😉). |
Created a new PR: #7978 |
491b29a
to
19c267f
Compare
|
19c267f
to
f19b20e
Compare
The backport to
To backport manually, run these commands in your terminal: # Fetch latest updates from GitHub
git fetch
# Create a new working tree
git worktree add .worktrees/backport-1.5 1.5
# Navigate to the new working tree
cd .worktrees/backport-1.5
# Create a new branch
git switch --create backport-7950-to-1.5
# Cherry-pick the merged commit of this pull request and resolve the conflicts
git cherry-pick -x --mainline 1 92179755f8c16aa93d65ad708a4e6d3dbd15957c
# Push it to GitHub
git push --set-upstream origin backport-7950-to-1.5
# Go back to the original working tree
cd ../..
# Delete the working tree
git worktree remove .worktrees/backport-1.5 Then, create a pull request where the |
…backtracking (python-poetry#7950) (cherry picked from commit fba1309)
…on-poetry#7950) (cherry picked from commit f54864e)
…atibilities (python-poetry#7950) (cherry picked from commit 0adc1c5)
…ough lots of levels (python-poetry#7950) (cherry picked from commit 9217975)
(cherry picked from commit f54864e)
This pull request has been automatically locked since there has not been any recent activity after it was closed. Please open a new issue for related bugs. |
This PR improves dependency resolution performance for conflicts involving packages with many versions. It works by preserving parts of the
DependencyCache
and thecontradicted_incompatibilities
caches when backtracking.On my machine with fully populated caches, the performance of
poetry update urllib3
(in a repo where this causes conflicts withboto3
/botocore
) goes from ~58 minutes to ~3 minutes with these changes.Background
I've been investigating reports of slow dependency resolution after urllib3 2.x was released, similar to #7856 (comment). On my machine it takes almost an hour for Poetry to find a solution even with populated caches. I noticed that the time spent evaluating each solution increases a lot (evaluating the 10th solution takes 20ms/solution; by the 1000th solution it's taking 7000ms/solution).
After a bunch of profiling, I found two possible improvements around preserving these two caches. Currently both of these caches are cleared when backtracking which is causing a huge amount of duplicate work (almost all of the time is spent in
VersionRange.allows
according to flameprof).Some reading of the algorithm docs suggests to me that we could improve this a lot by reverting the cache states back to the state of the decision level we backtrack to rather than wiping out the cache together. I updated
DependencyCache
and thecontradicted_incompatibilities
to track what decision level entries are added in and delete only the rolled back levels when backtracking.The two cache updates are independent changes; I made them in two separate commits so that they can be merged or tested individually if desired.
Performance improvements
All changes were tested using this reproduction repo and a warm cache using
poetry install && poetry update -vv urllib3
. All resulted in the same final solution after 1116 solutions evaluated.master
DependencyCache
onlycontradicted_incompatibilities
cache onlyHere is a graph of time taken per solution evaluated:
You can see that time taken to evaluate each solution still increases, but much more slowly (~200ms at solution 1115 compared to ~9000ms on master).
Does this actually work?
I think it does, but I am not 100% sure and could use help from someone with a better understanding of the algorithm to confirm.
To try to verify, I implemented a test mode which runs both the old and new cache behavior and asserts the results of each cache operation are identical: https://github.com/chriskuehl/poetry/commits/dependency-resolution-perf-with-tests. I tested this in a bunch of repos and randomly added/updated/removed big dependency trees with only one difference discovered which I'll describe below. From the linked branch you can set
VERIFY_INCOMPATIBILITY_CACHE=1
andVERIFY_DEPENDENCY_CACHE=1
when running Poetry operations to enable the assertions.The one difference from status quo I found
I found only one difference when asserting the results from the two caches:
DependencyCache.search_for
can return pre-release versions (for dependencies which do not allow them) on cache misses, but does not return them on cache hits. This happens because (even on master) the upper branch here does not filter out prerelease packages (it filters by the version range constraint only), while the bottom branch filters them out:poetry/src/poetry/mixology/version_solver.py
Lines 58 to 64 in 3602b21
I don't think this is a problem since they will be discarded later, but it does make my assertions sometimes fail as the caches return a different result now that the updated cache has a much higher hit rate. If I temporarily apply this patch then all assertion errors go away:
This makes me think it is not an issue? Note that I did not add this patch to this branch and instead kept the status quo behavior.