fix: poly check performance for large repos #355
Merged
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Description
The
poly check
command performs AST parsing to find any missing bricks or third-party imports. A module that already have been parsed should be cached during the session. This is done since before, but with the default max size oflru_cache
. For large Polylith workspaces, the cache need to be bigger.Motivation and Context
See discussion #354
How Has This Been Tested?
✅ CI
✅ running the
poly check
command for a repo with 18+ projects, 350+ bricks and with large python files (20K LOC).Types of changes
Checklist: