From 5ab9b3a15e0095a8186ca3870cb9f9eade83833d Mon Sep 17 00:00:00 2001 From: Omkar Phansopkar Date: Wed, 18 Oct 2023 15:42:56 +0530 Subject: [PATCH 01/28] Added docs server script, dark mode & copybutton for docs Signed-off-by: Omkar Phansopkar --- .github/workflows/docs-ci.yml | 3 --- docs/Makefile | 8 ++++++++ docs/make.bat | 12 ++++++++++++ docs/scripts/doc8_style_check.sh | 0 docs/source/conf.py | 4 ++++ setup.cfg | 3 +++ 6 files changed, 27 insertions(+), 3 deletions(-) mode change 100644 => 100755 docs/scripts/doc8_style_check.sh diff --git a/.github/workflows/docs-ci.yml b/.github/workflows/docs-ci.yml index 511b7c28..ada779bf 100644 --- a/.github/workflows/docs-ci.yml +++ b/.github/workflows/docs-ci.yml @@ -20,9 +20,6 @@ jobs: with: python-version: ${{ matrix.python-version }} - - name: Give permission to run scripts - run: chmod +x ./docs/scripts/doc8_style_check.sh - - name: Install Dependencies run: pip install -e .[docs] diff --git a/docs/Makefile b/docs/Makefile index d0c3cbf1..788b0396 100644 --- a/docs/Makefile +++ b/docs/Makefile @@ -5,6 +5,7 @@ # from the environment for the first two. SPHINXOPTS ?= SPHINXBUILD ?= sphinx-build +SPHINXAUTOBUILD = sphinx-autobuild SOURCEDIR = source BUILDDIR = build @@ -14,6 +15,13 @@ help: .PHONY: help Makefile +# Run the development server using sphinx-autobuild +docs: + @echo + @echo "Starting up the docs server..." + @echo + $(SPHINXAUTOBUILD) --port 8000 --watch ${SOURCEDIR} $(SOURCEDIR) "$(BUILDDIR)/html" $(SPHINXOPTS) $(O) + # Catch-all target: route all unknown targets to Sphinx using the new # "make mode" option. $(O) is meant as a shortcut for $(SPHINXOPTS). %: Makefile diff --git a/docs/make.bat b/docs/make.bat index 6247f7e2..4a3c1a48 100644 --- a/docs/make.bat +++ b/docs/make.bat @@ -7,11 +7,16 @@ REM Command file for Sphinx documentation if "%SPHINXBUILD%" == "" ( set SPHINXBUILD=sphinx-build ) +if "%SPHINXAUTOBUILD%" == "" ( + set SPHINXAUTOBUILD=sphinx-autobuild +) set SOURCEDIR=source set BUILDDIR=build if "%1" == "" goto help +if "%1" == "docs" goto docs + %SPHINXBUILD% >NUL 2>NUL if errorlevel 9009 ( echo. @@ -28,6 +33,13 @@ if errorlevel 9009 ( %SPHINXBUILD% -M %1 %SOURCEDIR% %BUILDDIR% %SPHINXOPTS% %O% goto end +:docs +@echo +@echo Starting up the docs server... +@echo +%SPHINXAUTOBUILD% --port 8000 --watch %SOURCEDIR% %SOURCEDIR% %BUILDDIR%\html %SPHINXOPTS% %O% +goto end + :help %SPHINXBUILD% -M help %SOURCEDIR% %BUILDDIR% %SPHINXOPTS% %O% diff --git a/docs/scripts/doc8_style_check.sh b/docs/scripts/doc8_style_check.sh old mode 100644 new mode 100755 diff --git a/docs/source/conf.py b/docs/source/conf.py index 918d62c1..54e5e665 100644 --- a/docs/source/conf.py +++ b/docs/source/conf.py @@ -30,6 +30,10 @@ extensions = [ "sphinx.ext.intersphinx", "sphinx_reredirects", + 'sphinx_rtd_theme', + "sphinx_rtd_dark_mode", + "sphinx.ext.extlinks", + "sphinx_copybutton", ] diff --git a/setup.cfg b/setup.cfg index d6c7da7d..bd0e58a7 100644 --- a/setup.cfg +++ b/setup.cfg @@ -62,4 +62,7 @@ docs = sphinx-rtd-theme>=1.0.0 sphinx-reredirects >= 0.1.2 doc8>=0.11.2 + sphinx-autobuild + sphinx-rtd-dark-mode>=1.3.0 + sphinx-copybutton From 0a9d983650bf042a5bd2c277711b637979e566f1 Mon Sep 17 00:00:00 2001 From: "John M. Horan" Date: Mon, 20 Nov 2023 16:46:54 -0800 Subject: [PATCH 02/28] Update CSS to widen page and handle mobile #84 Reference: https://github.com/nexB/skeleton/issues/84 Signed-off-by: John M. Horan --- docs/source/_static/theme_overrides.css | 363 +----------------- .../_static/theme_overrides_SUPERSEDED.css | 353 +++++++++++++++++ docs/source/conf.py | 15 +- 3 files changed, 380 insertions(+), 351 deletions(-) create mode 100644 docs/source/_static/theme_overrides_SUPERSEDED.css diff --git a/docs/source/_static/theme_overrides.css b/docs/source/_static/theme_overrides.css index 9662d63a..de5ae433 100644 --- a/docs/source/_static/theme_overrides.css +++ b/docs/source/_static/theme_overrides.css @@ -1,353 +1,26 @@ -body { - color: #000000; -} - -p { - margin-bottom: 10px; -} - -.wy-plain-list-disc, .rst-content .section ul, .rst-content .toctree-wrapper ul, article ul { - margin-bottom: 10px; -} - -.custom_header_01 { - color: #cc0000; - font-size: 22px; - font-weight: bold; - line-height: 50px; -} - -h1, h2, h3, h4, h5, h6 { - margin-bottom: 20px; - margin-top: 20px; -} - -h5 { - font-size: 18px; - color: #000000; - font-style: italic; - margin-bottom: 10px; -} - -h6 { - font-size: 15px; - color: #000000; - font-style: italic; - margin-bottom: 10px; -} - -/* custom admonitions */ -/* success */ -.custom-admonition-success .admonition-title { - color: #000000; - background: #ccffcc; - border-radius: 5px 5px 0px 0px; -} -div.custom-admonition-success.admonition { - color: #000000; - background: #ffffff; - border: solid 1px #cccccc; - border-radius: 5px; - box-shadow: 1px 1px 5px 3px #d8d8d8; - margin: 20px 0px 30px 0px; -} - -/* important */ -.custom-admonition-important .admonition-title { - color: #000000; - background: #ccffcc; - border-radius: 5px 5px 0px 0px; - border-bottom: solid 1px #000000; -} -div.custom-admonition-important.admonition { - color: #000000; - background: #ffffff; - border: solid 1px #cccccc; - border-radius: 5px; - box-shadow: 1px 1px 5px 3px #d8d8d8; - margin: 20px 0px 30px 0px; -} - -/* caution */ -.custom-admonition-caution .admonition-title { - color: #000000; - background: #ffff99; - border-radius: 5px 5px 0px 0px; - border-bottom: solid 1px #e8e8e8; -} -div.custom-admonition-caution.admonition { - color: #000000; - background: #ffffff; - border: solid 1px #cccccc; - border-radius: 5px; - box-shadow: 1px 1px 5px 3px #d8d8d8; - margin: 20px 0px 30px 0px; -} - -/* note */ -.custom-admonition-note .admonition-title { - color: #ffffff; - background: #006bb3; - border-radius: 5px 5px 0px 0px; -} -div.custom-admonition-note.admonition { - color: #000000; - background: #ffffff; - border: solid 1px #cccccc; - border-radius: 5px; - box-shadow: 1px 1px 5px 3px #d8d8d8; - margin: 20px 0px 30px 0px; -} - -/* todo */ -.custom-admonition-todo .admonition-title { - color: #000000; - background: #cce6ff; - border-radius: 5px 5px 0px 0px; - border-bottom: solid 1px #99ccff; -} -div.custom-admonition-todo.admonition { - color: #000000; - background: #ffffff; - border: solid 1px #99ccff; - border-radius: 5px; - box-shadow: 1px 1px 5px 3px #d8d8d8; - margin: 20px 0px 30px 0px; -} - -/* examples */ -.custom-admonition-examples .admonition-title { - color: #000000; - background: #ffe6cc; - border-radius: 5px 5px 0px 0px; - border-bottom: solid 1px #d8d8d8; -} -div.custom-admonition-examples.admonition { - color: #000000; - background: #ffffff; - border: solid 1px #cccccc; - border-radius: 5px; - box-shadow: 1px 1px 5px 3px #d8d8d8; - margin: 20px 0px 30px 0px; -} - +/* this is the container for the pages */ .wy-nav-content { max-width: 100%; - padding-right: 100px; - padding-left: 100px; - background-color: #f2f2f2; -} - -div.rst-content { - background-color: #ffffff; - border: solid 1px #e5e5e5; - padding: 20px 40px 20px 40px; -} - -.rst-content .guilabel { - border: 1px solid #ffff99; - background: #ffff99; - font-size: 100%; - font-weight: normal; - border-radius: 4px; - padding: 2px 0px; - margin: auto 2px; - vertical-align: middle; -} - -.rst-content kbd { - font-family: SFMono-Regular,Menlo,Monaco,Consolas,"Liberation Mono","Courier New",Courier,monospace; - border: solid 1px #d8d8d8; - background-color: #f5f5f5; - padding: 0px 3px; - border-radius: 3px; -} - -.wy-nav-content-wrap a { - color: #0066cc; - text-decoration: none; -} -.wy-nav-content-wrap a:hover { - color: #0099cc; - text-decoration: underline; -} - -.wy-nav-top a { - color: #ffffff; -} - -/* Based on numerous similar approaches e.g., https://github.com/readthedocs/sphinx_rtd_theme/issues/117 and https://rackerlabs.github.io/docs-rackspace/tools/rtd-tables.html -- but remove form-factor limits to enable table wrap on full-size and smallest-size form factors */ -.wy-table-responsive table td { - white-space: normal !important; -} - -.rst-content table.docutils td, -.rst-content table.docutils th { - padding: 5px 10px 5px 10px; -} -.rst-content table.docutils td p, -.rst-content table.docutils th p { - font-size: 14px; - margin-bottom: 0px; -} -.rst-content table.docutils td p cite, -.rst-content table.docutils th p cite { - font-size: 14px; - background-color: transparent; -} - -.colwidths-given th { - border: solid 1px #d8d8d8 !important; -} -.colwidths-given td { - border: solid 1px #d8d8d8 !important; -} - -/*handles single-tick inline code*/ -.wy-body-for-nav cite { - color: #000000; - background-color: transparent; - font-style: normal; - font-family: "Courier New"; - font-size: 13px; - padding: 3px 3px 3px 3px; -} - -.rst-content pre.literal-block, .rst-content div[class^="highlight"] pre, .rst-content .linenodiv pre { - font-family: SFMono-Regular,Menlo,Monaco,Consolas,"Liberation Mono","Courier New",Courier,monospace; - font-size: 13px; - overflow: visible; - white-space: pre-wrap; - color: #000000; -} - -.rst-content pre.literal-block, .rst-content div[class^='highlight'] { - background-color: #f8f8f8; - border: solid 1px #e8e8e8; -} - -/* This enables inline code to wrap. */ -code, .rst-content tt, .rst-content code { - white-space: pre-wrap; - padding: 2px 3px 1px; - border-radius: 3px; - font-size: 13px; - background-color: #ffffff; -} - -/* use this added class for code blocks attached to bulleted list items */ -.highlight-top-margin { - margin-top: 20px !important; -} - -/* change color of inline code block */ -span.pre { - color: #e01e5a; -} - -.wy-body-for-nav blockquote { - margin: 1em 0; - padding-left: 1em; - border-left: 4px solid #ddd; - color: #000000; -} - -/* Fix the unwanted top and bottom padding inside a nested bulleted/numbered list */ -.rst-content .section ol p, .rst-content .section ul p { - margin-bottom: 0px; -} - -/* add spacing between bullets for legibility */ -.rst-content .section ol li, .rst-content .section ul li { - margin-bottom: 5px; -} - -.rst-content .section ol li:first-child, .rst-content .section ul li:first-child { - margin-top: 5px; -} - -/* but exclude the toctree bullets */ -.rst-content .toctree-wrapper ul li, .rst-content .toctree-wrapper ul li:first-child { + padding: 0px 40px 0px 0px; margin-top: 0px; - margin-bottom: 0px; } -/* remove extra space at bottom of multine list-table cell */ -.rst-content .line-block { - margin-left: 0px; - margin-bottom: 0px; - line-height: 24px; +.wy-nav-content-wrap { + border-right: solid 1px; } -/* fix extra vertical spacing in page toctree */ -.rst-content .toctree-wrapper ul li ul, article ul li ul { - margin-top: 0; - margin-bottom: 0; -} - -/* this is used by the genindex added via layout.html (see source/_templates/) to sidebar toc */ -.reference.internal.toc-index { - color: #d9d9d9; -} - -.reference.internal.toc-index.current { - background-color: #ffffff; - color: #000000; - font-weight: bold; -} - -.toc-index-div { - border-top: solid 1px #000000; - margin-top: 10px; - padding-top: 5px; -} - -.indextable ul li { - font-size: 14px; - margin-bottom: 5px; -} - -/* The next 2 fix the poor vertical spacing in genindex.html (the alphabetized index) */ -.indextable.genindextable { - margin-bottom: 20px; -} - -div.genindex-jumpbox { - margin-bottom: 10px; -} - -/* rst image classes */ - -.clear-both { - clear: both; - } - -.float-left { - float: left; - margin-right: 20px; -} - -img { - border: solid 1px #e8e8e8; -} - -/* These are custom and need to be defined in conf.py to access in all pages, e.g., '.. role:: red' */ -.img-title { - color: #000000; - /* neither padding nor margin works for vertical spacing bc it's a span -- line-height does, sort of */ - line-height: 3.0; - font-style: italic; - font-weight: 600; -} - -.img-title-para { - color: #000000; - margin-top: 20px; - margin-bottom: 0px; - font-style: italic; - font-weight: 500; -} - -.red { - color: red; +div.rst-content { + max-width: 1300px; + border: 0; + padding: 0px 80px 10px 80px; + margin-left: 50px; +} + +@media (max-width: 768px) { + div.rst-content { + max-width: 1300px; + border: 0; + padding: 0px 10px 10px 10px; + margin-left: 0px; + } } diff --git a/docs/source/_static/theme_overrides_SUPERSEDED.css b/docs/source/_static/theme_overrides_SUPERSEDED.css new file mode 100644 index 00000000..9662d63a --- /dev/null +++ b/docs/source/_static/theme_overrides_SUPERSEDED.css @@ -0,0 +1,353 @@ +body { + color: #000000; +} + +p { + margin-bottom: 10px; +} + +.wy-plain-list-disc, .rst-content .section ul, .rst-content .toctree-wrapper ul, article ul { + margin-bottom: 10px; +} + +.custom_header_01 { + color: #cc0000; + font-size: 22px; + font-weight: bold; + line-height: 50px; +} + +h1, h2, h3, h4, h5, h6 { + margin-bottom: 20px; + margin-top: 20px; +} + +h5 { + font-size: 18px; + color: #000000; + font-style: italic; + margin-bottom: 10px; +} + +h6 { + font-size: 15px; + color: #000000; + font-style: italic; + margin-bottom: 10px; +} + +/* custom admonitions */ +/* success */ +.custom-admonition-success .admonition-title { + color: #000000; + background: #ccffcc; + border-radius: 5px 5px 0px 0px; +} +div.custom-admonition-success.admonition { + color: #000000; + background: #ffffff; + border: solid 1px #cccccc; + border-radius: 5px; + box-shadow: 1px 1px 5px 3px #d8d8d8; + margin: 20px 0px 30px 0px; +} + +/* important */ +.custom-admonition-important .admonition-title { + color: #000000; + background: #ccffcc; + border-radius: 5px 5px 0px 0px; + border-bottom: solid 1px #000000; +} +div.custom-admonition-important.admonition { + color: #000000; + background: #ffffff; + border: solid 1px #cccccc; + border-radius: 5px; + box-shadow: 1px 1px 5px 3px #d8d8d8; + margin: 20px 0px 30px 0px; +} + +/* caution */ +.custom-admonition-caution .admonition-title { + color: #000000; + background: #ffff99; + border-radius: 5px 5px 0px 0px; + border-bottom: solid 1px #e8e8e8; +} +div.custom-admonition-caution.admonition { + color: #000000; + background: #ffffff; + border: solid 1px #cccccc; + border-radius: 5px; + box-shadow: 1px 1px 5px 3px #d8d8d8; + margin: 20px 0px 30px 0px; +} + +/* note */ +.custom-admonition-note .admonition-title { + color: #ffffff; + background: #006bb3; + border-radius: 5px 5px 0px 0px; +} +div.custom-admonition-note.admonition { + color: #000000; + background: #ffffff; + border: solid 1px #cccccc; + border-radius: 5px; + box-shadow: 1px 1px 5px 3px #d8d8d8; + margin: 20px 0px 30px 0px; +} + +/* todo */ +.custom-admonition-todo .admonition-title { + color: #000000; + background: #cce6ff; + border-radius: 5px 5px 0px 0px; + border-bottom: solid 1px #99ccff; +} +div.custom-admonition-todo.admonition { + color: #000000; + background: #ffffff; + border: solid 1px #99ccff; + border-radius: 5px; + box-shadow: 1px 1px 5px 3px #d8d8d8; + margin: 20px 0px 30px 0px; +} + +/* examples */ +.custom-admonition-examples .admonition-title { + color: #000000; + background: #ffe6cc; + border-radius: 5px 5px 0px 0px; + border-bottom: solid 1px #d8d8d8; +} +div.custom-admonition-examples.admonition { + color: #000000; + background: #ffffff; + border: solid 1px #cccccc; + border-radius: 5px; + box-shadow: 1px 1px 5px 3px #d8d8d8; + margin: 20px 0px 30px 0px; +} + +.wy-nav-content { + max-width: 100%; + padding-right: 100px; + padding-left: 100px; + background-color: #f2f2f2; +} + +div.rst-content { + background-color: #ffffff; + border: solid 1px #e5e5e5; + padding: 20px 40px 20px 40px; +} + +.rst-content .guilabel { + border: 1px solid #ffff99; + background: #ffff99; + font-size: 100%; + font-weight: normal; + border-radius: 4px; + padding: 2px 0px; + margin: auto 2px; + vertical-align: middle; +} + +.rst-content kbd { + font-family: SFMono-Regular,Menlo,Monaco,Consolas,"Liberation Mono","Courier New",Courier,monospace; + border: solid 1px #d8d8d8; + background-color: #f5f5f5; + padding: 0px 3px; + border-radius: 3px; +} + +.wy-nav-content-wrap a { + color: #0066cc; + text-decoration: none; +} +.wy-nav-content-wrap a:hover { + color: #0099cc; + text-decoration: underline; +} + +.wy-nav-top a { + color: #ffffff; +} + +/* Based on numerous similar approaches e.g., https://github.com/readthedocs/sphinx_rtd_theme/issues/117 and https://rackerlabs.github.io/docs-rackspace/tools/rtd-tables.html -- but remove form-factor limits to enable table wrap on full-size and smallest-size form factors */ +.wy-table-responsive table td { + white-space: normal !important; +} + +.rst-content table.docutils td, +.rst-content table.docutils th { + padding: 5px 10px 5px 10px; +} +.rst-content table.docutils td p, +.rst-content table.docutils th p { + font-size: 14px; + margin-bottom: 0px; +} +.rst-content table.docutils td p cite, +.rst-content table.docutils th p cite { + font-size: 14px; + background-color: transparent; +} + +.colwidths-given th { + border: solid 1px #d8d8d8 !important; +} +.colwidths-given td { + border: solid 1px #d8d8d8 !important; +} + +/*handles single-tick inline code*/ +.wy-body-for-nav cite { + color: #000000; + background-color: transparent; + font-style: normal; + font-family: "Courier New"; + font-size: 13px; + padding: 3px 3px 3px 3px; +} + +.rst-content pre.literal-block, .rst-content div[class^="highlight"] pre, .rst-content .linenodiv pre { + font-family: SFMono-Regular,Menlo,Monaco,Consolas,"Liberation Mono","Courier New",Courier,monospace; + font-size: 13px; + overflow: visible; + white-space: pre-wrap; + color: #000000; +} + +.rst-content pre.literal-block, .rst-content div[class^='highlight'] { + background-color: #f8f8f8; + border: solid 1px #e8e8e8; +} + +/* This enables inline code to wrap. */ +code, .rst-content tt, .rst-content code { + white-space: pre-wrap; + padding: 2px 3px 1px; + border-radius: 3px; + font-size: 13px; + background-color: #ffffff; +} + +/* use this added class for code blocks attached to bulleted list items */ +.highlight-top-margin { + margin-top: 20px !important; +} + +/* change color of inline code block */ +span.pre { + color: #e01e5a; +} + +.wy-body-for-nav blockquote { + margin: 1em 0; + padding-left: 1em; + border-left: 4px solid #ddd; + color: #000000; +} + +/* Fix the unwanted top and bottom padding inside a nested bulleted/numbered list */ +.rst-content .section ol p, .rst-content .section ul p { + margin-bottom: 0px; +} + +/* add spacing between bullets for legibility */ +.rst-content .section ol li, .rst-content .section ul li { + margin-bottom: 5px; +} + +.rst-content .section ol li:first-child, .rst-content .section ul li:first-child { + margin-top: 5px; +} + +/* but exclude the toctree bullets */ +.rst-content .toctree-wrapper ul li, .rst-content .toctree-wrapper ul li:first-child { + margin-top: 0px; + margin-bottom: 0px; +} + +/* remove extra space at bottom of multine list-table cell */ +.rst-content .line-block { + margin-left: 0px; + margin-bottom: 0px; + line-height: 24px; +} + +/* fix extra vertical spacing in page toctree */ +.rst-content .toctree-wrapper ul li ul, article ul li ul { + margin-top: 0; + margin-bottom: 0; +} + +/* this is used by the genindex added via layout.html (see source/_templates/) to sidebar toc */ +.reference.internal.toc-index { + color: #d9d9d9; +} + +.reference.internal.toc-index.current { + background-color: #ffffff; + color: #000000; + font-weight: bold; +} + +.toc-index-div { + border-top: solid 1px #000000; + margin-top: 10px; + padding-top: 5px; +} + +.indextable ul li { + font-size: 14px; + margin-bottom: 5px; +} + +/* The next 2 fix the poor vertical spacing in genindex.html (the alphabetized index) */ +.indextable.genindextable { + margin-bottom: 20px; +} + +div.genindex-jumpbox { + margin-bottom: 10px; +} + +/* rst image classes */ + +.clear-both { + clear: both; + } + +.float-left { + float: left; + margin-right: 20px; +} + +img { + border: solid 1px #e8e8e8; +} + +/* These are custom and need to be defined in conf.py to access in all pages, e.g., '.. role:: red' */ +.img-title { + color: #000000; + /* neither padding nor margin works for vertical spacing bc it's a span -- line-height does, sort of */ + line-height: 3.0; + font-style: italic; + font-weight: 600; +} + +.img-title-para { + color: #000000; + margin-top: 20px; + margin-bottom: 0px; + font-style: italic; + font-weight: 500; +} + +.red { + color: red; +} diff --git a/docs/source/conf.py b/docs/source/conf.py index 54e5e665..7771ff09 100644 --- a/docs/source/conf.py +++ b/docs/source/conf.py @@ -30,7 +30,7 @@ extensions = [ "sphinx.ext.intersphinx", "sphinx_reredirects", - 'sphinx_rtd_theme', + "sphinx_rtd_theme", "sphinx_rtd_dark_mode", "sphinx.ext.extlinks", "sphinx_copybutton", @@ -47,7 +47,10 @@ intersphinx_mapping = { "aboutcode": ("https://aboutcode.readthedocs.io/en/latest/", None), - "scancode-workbench": ("https://scancode-workbench.readthedocs.io/en/develop/", None), + "scancode-workbench": ( + "https://scancode-workbench.readthedocs.io/en/develop/", + None, + ), } @@ -82,7 +85,9 @@ "conf_py_path": "/docs/source/", # path in the checkout to the docs root } -html_css_files = ["_static/theme_overrides.css"] +html_css_files = [ + "theme_overrides.css", +] # If true, "Created using Sphinx" is shown in the HTML footer. Default is True. @@ -108,6 +113,4 @@ # -- Options for LaTeX output ------------------------------------------------- -latex_elements = { - 'classoptions': ',openany,oneside' -} \ No newline at end of file +latex_elements = {"classoptions": ",openany,oneside"} From 4e36fc601eaa17bde0d2a4bebfed70d7bde28e7c Mon Sep 17 00:00:00 2001 From: "John M. Horan" Date: Tue, 16 Jan 2024 12:22:54 -0800 Subject: [PATCH 03/28] Delete theme_overrides_SUPERSEDED.css as no longer needed #84 Reference: https://github.com/nexB/skeleton/issues/84 Signed-off-by: John M. Horan --- .../_static/theme_overrides_SUPERSEDED.css | 353 ------------------ 1 file changed, 353 deletions(-) delete mode 100644 docs/source/_static/theme_overrides_SUPERSEDED.css diff --git a/docs/source/_static/theme_overrides_SUPERSEDED.css b/docs/source/_static/theme_overrides_SUPERSEDED.css deleted file mode 100644 index 9662d63a..00000000 --- a/docs/source/_static/theme_overrides_SUPERSEDED.css +++ /dev/null @@ -1,353 +0,0 @@ -body { - color: #000000; -} - -p { - margin-bottom: 10px; -} - -.wy-plain-list-disc, .rst-content .section ul, .rst-content .toctree-wrapper ul, article ul { - margin-bottom: 10px; -} - -.custom_header_01 { - color: #cc0000; - font-size: 22px; - font-weight: bold; - line-height: 50px; -} - -h1, h2, h3, h4, h5, h6 { - margin-bottom: 20px; - margin-top: 20px; -} - -h5 { - font-size: 18px; - color: #000000; - font-style: italic; - margin-bottom: 10px; -} - -h6 { - font-size: 15px; - color: #000000; - font-style: italic; - margin-bottom: 10px; -} - -/* custom admonitions */ -/* success */ -.custom-admonition-success .admonition-title { - color: #000000; - background: #ccffcc; - border-radius: 5px 5px 0px 0px; -} -div.custom-admonition-success.admonition { - color: #000000; - background: #ffffff; - border: solid 1px #cccccc; - border-radius: 5px; - box-shadow: 1px 1px 5px 3px #d8d8d8; - margin: 20px 0px 30px 0px; -} - -/* important */ -.custom-admonition-important .admonition-title { - color: #000000; - background: #ccffcc; - border-radius: 5px 5px 0px 0px; - border-bottom: solid 1px #000000; -} -div.custom-admonition-important.admonition { - color: #000000; - background: #ffffff; - border: solid 1px #cccccc; - border-radius: 5px; - box-shadow: 1px 1px 5px 3px #d8d8d8; - margin: 20px 0px 30px 0px; -} - -/* caution */ -.custom-admonition-caution .admonition-title { - color: #000000; - background: #ffff99; - border-radius: 5px 5px 0px 0px; - border-bottom: solid 1px #e8e8e8; -} -div.custom-admonition-caution.admonition { - color: #000000; - background: #ffffff; - border: solid 1px #cccccc; - border-radius: 5px; - box-shadow: 1px 1px 5px 3px #d8d8d8; - margin: 20px 0px 30px 0px; -} - -/* note */ -.custom-admonition-note .admonition-title { - color: #ffffff; - background: #006bb3; - border-radius: 5px 5px 0px 0px; -} -div.custom-admonition-note.admonition { - color: #000000; - background: #ffffff; - border: solid 1px #cccccc; - border-radius: 5px; - box-shadow: 1px 1px 5px 3px #d8d8d8; - margin: 20px 0px 30px 0px; -} - -/* todo */ -.custom-admonition-todo .admonition-title { - color: #000000; - background: #cce6ff; - border-radius: 5px 5px 0px 0px; - border-bottom: solid 1px #99ccff; -} -div.custom-admonition-todo.admonition { - color: #000000; - background: #ffffff; - border: solid 1px #99ccff; - border-radius: 5px; - box-shadow: 1px 1px 5px 3px #d8d8d8; - margin: 20px 0px 30px 0px; -} - -/* examples */ -.custom-admonition-examples .admonition-title { - color: #000000; - background: #ffe6cc; - border-radius: 5px 5px 0px 0px; - border-bottom: solid 1px #d8d8d8; -} -div.custom-admonition-examples.admonition { - color: #000000; - background: #ffffff; - border: solid 1px #cccccc; - border-radius: 5px; - box-shadow: 1px 1px 5px 3px #d8d8d8; - margin: 20px 0px 30px 0px; -} - -.wy-nav-content { - max-width: 100%; - padding-right: 100px; - padding-left: 100px; - background-color: #f2f2f2; -} - -div.rst-content { - background-color: #ffffff; - border: solid 1px #e5e5e5; - padding: 20px 40px 20px 40px; -} - -.rst-content .guilabel { - border: 1px solid #ffff99; - background: #ffff99; - font-size: 100%; - font-weight: normal; - border-radius: 4px; - padding: 2px 0px; - margin: auto 2px; - vertical-align: middle; -} - -.rst-content kbd { - font-family: SFMono-Regular,Menlo,Monaco,Consolas,"Liberation Mono","Courier New",Courier,monospace; - border: solid 1px #d8d8d8; - background-color: #f5f5f5; - padding: 0px 3px; - border-radius: 3px; -} - -.wy-nav-content-wrap a { - color: #0066cc; - text-decoration: none; -} -.wy-nav-content-wrap a:hover { - color: #0099cc; - text-decoration: underline; -} - -.wy-nav-top a { - color: #ffffff; -} - -/* Based on numerous similar approaches e.g., https://github.com/readthedocs/sphinx_rtd_theme/issues/117 and https://rackerlabs.github.io/docs-rackspace/tools/rtd-tables.html -- but remove form-factor limits to enable table wrap on full-size and smallest-size form factors */ -.wy-table-responsive table td { - white-space: normal !important; -} - -.rst-content table.docutils td, -.rst-content table.docutils th { - padding: 5px 10px 5px 10px; -} -.rst-content table.docutils td p, -.rst-content table.docutils th p { - font-size: 14px; - margin-bottom: 0px; -} -.rst-content table.docutils td p cite, -.rst-content table.docutils th p cite { - font-size: 14px; - background-color: transparent; -} - -.colwidths-given th { - border: solid 1px #d8d8d8 !important; -} -.colwidths-given td { - border: solid 1px #d8d8d8 !important; -} - -/*handles single-tick inline code*/ -.wy-body-for-nav cite { - color: #000000; - background-color: transparent; - font-style: normal; - font-family: "Courier New"; - font-size: 13px; - padding: 3px 3px 3px 3px; -} - -.rst-content pre.literal-block, .rst-content div[class^="highlight"] pre, .rst-content .linenodiv pre { - font-family: SFMono-Regular,Menlo,Monaco,Consolas,"Liberation Mono","Courier New",Courier,monospace; - font-size: 13px; - overflow: visible; - white-space: pre-wrap; - color: #000000; -} - -.rst-content pre.literal-block, .rst-content div[class^='highlight'] { - background-color: #f8f8f8; - border: solid 1px #e8e8e8; -} - -/* This enables inline code to wrap. */ -code, .rst-content tt, .rst-content code { - white-space: pre-wrap; - padding: 2px 3px 1px; - border-radius: 3px; - font-size: 13px; - background-color: #ffffff; -} - -/* use this added class for code blocks attached to bulleted list items */ -.highlight-top-margin { - margin-top: 20px !important; -} - -/* change color of inline code block */ -span.pre { - color: #e01e5a; -} - -.wy-body-for-nav blockquote { - margin: 1em 0; - padding-left: 1em; - border-left: 4px solid #ddd; - color: #000000; -} - -/* Fix the unwanted top and bottom padding inside a nested bulleted/numbered list */ -.rst-content .section ol p, .rst-content .section ul p { - margin-bottom: 0px; -} - -/* add spacing between bullets for legibility */ -.rst-content .section ol li, .rst-content .section ul li { - margin-bottom: 5px; -} - -.rst-content .section ol li:first-child, .rst-content .section ul li:first-child { - margin-top: 5px; -} - -/* but exclude the toctree bullets */ -.rst-content .toctree-wrapper ul li, .rst-content .toctree-wrapper ul li:first-child { - margin-top: 0px; - margin-bottom: 0px; -} - -/* remove extra space at bottom of multine list-table cell */ -.rst-content .line-block { - margin-left: 0px; - margin-bottom: 0px; - line-height: 24px; -} - -/* fix extra vertical spacing in page toctree */ -.rst-content .toctree-wrapper ul li ul, article ul li ul { - margin-top: 0; - margin-bottom: 0; -} - -/* this is used by the genindex added via layout.html (see source/_templates/) to sidebar toc */ -.reference.internal.toc-index { - color: #d9d9d9; -} - -.reference.internal.toc-index.current { - background-color: #ffffff; - color: #000000; - font-weight: bold; -} - -.toc-index-div { - border-top: solid 1px #000000; - margin-top: 10px; - padding-top: 5px; -} - -.indextable ul li { - font-size: 14px; - margin-bottom: 5px; -} - -/* The next 2 fix the poor vertical spacing in genindex.html (the alphabetized index) */ -.indextable.genindextable { - margin-bottom: 20px; -} - -div.genindex-jumpbox { - margin-bottom: 10px; -} - -/* rst image classes */ - -.clear-both { - clear: both; - } - -.float-left { - float: left; - margin-right: 20px; -} - -img { - border: solid 1px #e8e8e8; -} - -/* These are custom and need to be defined in conf.py to access in all pages, e.g., '.. role:: red' */ -.img-title { - color: #000000; - /* neither padding nor margin works for vertical spacing bc it's a span -- line-height does, sort of */ - line-height: 3.0; - font-style: italic; - font-weight: 600; -} - -.img-title-para { - color: #000000; - margin-top: 20px; - margin-bottom: 0px; - font-style: italic; - font-weight: 500; -} - -.red { - color: red; -} From 7d74b8a3c98761293cd133d543e4d58a525dc7bf Mon Sep 17 00:00:00 2001 From: Ayan Sinha Mahapatra Date: Thu, 18 Jan 2024 17:11:14 +0530 Subject: [PATCH 04/28] Fix top padding for rst content Signed-off-by: Ayan Sinha Mahapatra --- docs/source/_static/theme_overrides.css | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docs/source/_static/theme_overrides.css b/docs/source/_static/theme_overrides.css index de5ae433..5863ccf5 100644 --- a/docs/source/_static/theme_overrides.css +++ b/docs/source/_static/theme_overrides.css @@ -12,7 +12,7 @@ div.rst-content { max-width: 1300px; border: 0; - padding: 0px 80px 10px 80px; + padding: 10px 80px 10px 80px; margin-left: 50px; } From 008d521aec51e5983f6d6a2adc4efa7fd92159cf Mon Sep 17 00:00:00 2001 From: Ayan Sinha Mahapatra Date: Mon, 19 Feb 2024 15:21:45 +0530 Subject: [PATCH 05/28] Update CI runners and python version Signed-off-by: Ayan Sinha Mahapatra --- .github/workflows/docs-ci.yml | 2 +- .github/workflows/pypi-release.yml | 8 ++++---- azure-pipelines.yml | 22 +++++++++++++++------- 3 files changed, 20 insertions(+), 12 deletions(-) diff --git a/.github/workflows/docs-ci.yml b/.github/workflows/docs-ci.yml index ada779bf..8c2abfe9 100644 --- a/.github/workflows/docs-ci.yml +++ b/.github/workflows/docs-ci.yml @@ -4,7 +4,7 @@ on: [push, pull_request] jobs: build: - runs-on: ubuntu-20.04 + runs-on: ubuntu-22.04 strategy: max-parallel: 4 diff --git a/.github/workflows/pypi-release.yml b/.github/workflows/pypi-release.yml index 95857301..d2206c87 100644 --- a/.github/workflows/pypi-release.yml +++ b/.github/workflows/pypi-release.yml @@ -21,10 +21,10 @@ on: jobs: build-pypi-distribs: name: Build and publish library to PyPI - runs-on: ubuntu-20.04 + runs-on: ubuntu-22.04 steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 - name: Set up Python uses: actions/setup-python@v4 with: @@ -47,7 +47,7 @@ jobs: name: Create GH release needs: - build-pypi-distribs - runs-on: ubuntu-20.04 + runs-on: ubuntu-22.04 steps: - name: Download built archives @@ -67,7 +67,7 @@ jobs: name: Create PyPI release needs: - create-gh-release - runs-on: ubuntu-20.04 + runs-on: ubuntu-22.04 steps: - name: Download built archives diff --git a/azure-pipelines.yml b/azure-pipelines.yml index 764883de..373b78cd 100644 --- a/azure-pipelines.yml +++ b/azure-pipelines.yml @@ -11,7 +11,7 @@ jobs: parameters: job_name: ubuntu20_cpython image_name: ubuntu-20.04 - python_versions: ['3.7', '3.8', '3.9', '3.10', '3.11'] + python_versions: ['3.8', '3.9', '3.10', '3.11', '3.12'] test_suites: all: venv/bin/pytest -n 2 -vvs @@ -19,7 +19,7 @@ jobs: parameters: job_name: ubuntu22_cpython image_name: ubuntu-22.04 - python_versions: ['3.7', '3.8', '3.9', '3.10', '3.11'] + python_versions: ['3.8', '3.9', '3.10', '3.11', '3.12'] test_suites: all: venv/bin/pytest -n 2 -vvs @@ -27,7 +27,7 @@ jobs: parameters: job_name: macos11_cpython image_name: macOS-11 - python_versions: ['3.7', '3.8', '3.9', '3.10', '3.11'] + python_versions: ['3.8', '3.9', '3.10', '3.11', '3.12'] test_suites: all: venv/bin/pytest -n 2 -vvs @@ -35,7 +35,7 @@ jobs: parameters: job_name: macos12_cpython image_name: macOS-12 - python_versions: ['3.7', '3.8', '3.9', '3.10', '3.11'] + python_versions: ['3.8', '3.9', '3.10', '3.11', '3.12'] test_suites: all: venv/bin/pytest -n 2 -vvs @@ -43,7 +43,15 @@ jobs: parameters: job_name: macos13_cpython image_name: macOS-13 - python_versions: ['3.7', '3.8', '3.9', '3.10', '3.11'] + python_versions: ['3.8', '3.9', '3.10', '3.11', '3.12'] + test_suites: + all: venv/bin/pytest -n 2 -vvs + + - template: etc/ci/azure-posix.yml + parameters: + job_name: macos14_cpython + image_name: macOS-14 + python_versions: ['3.8', '3.9', '3.10', '3.11', '3.12'] test_suites: all: venv/bin/pytest -n 2 -vvs @@ -51,7 +59,7 @@ jobs: parameters: job_name: win2019_cpython image_name: windows-2019 - python_versions: ['3.7', '3.8', '3.9', '3.10', '3.11'] + python_versions: ['3.8', '3.9', '3.10', '3.11', '3.12'] test_suites: all: venv\Scripts\pytest -n 2 -vvs @@ -59,6 +67,6 @@ jobs: parameters: job_name: win2022_cpython image_name: windows-2022 - python_versions: ['3.7', '3.8', '3.9', '3.10', '3.11'] + python_versions: ['3.8', '3.9', '3.10', '3.11', '3.12'] test_suites: all: venv\Scripts\pytest -n 2 -vvs From ba64d0744f0c18cf50de8f22c4e577557cc20b5d Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 6 May 2024 18:11:05 +0000 Subject: [PATCH 06/28] Bump jinja2 from 3.0.3 to 3.1.4 Bumps [jinja2](https://github.com/pallets/jinja) from 3.0.3 to 3.1.4. - [Release notes](https://github.com/pallets/jinja/releases) - [Changelog](https://github.com/pallets/jinja/blob/main/CHANGES.rst) - [Commits](https://github.com/pallets/jinja/compare/3.0.3...3.1.4) --- updated-dependencies: - dependency-name: jinja2 dependency-type: direct:production ... Signed-off-by: dependabot[bot] --- requirements.txt | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/requirements.txt b/requirements.txt index 78a6b533..26214b6a 100644 --- a/requirements.txt +++ b/requirements.txt @@ -33,7 +33,7 @@ intbitset==3.1.0 isodate==0.6.1 jaraco.functools==3.4.0 javaproperties==0.8.1 -Jinja2==3.0.3 +Jinja2==3.1.4 jsonstreams==0.6.0 license-expression==21.6.14 lxml==4.9.1 From 324394bbd3bf676813c0b95840cd3e5fac5aa0ec Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Tue, 21 May 2024 02:03:46 +0000 Subject: [PATCH 07/28] --- updated-dependencies: - dependency-name: requests dependency-type: direct:production ... Signed-off-by: dependabot[bot] --- requirements.txt | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/requirements.txt b/requirements.txt index 78a6b533..d307ddd8 100644 --- a/requirements.txt +++ b/requirements.txt @@ -63,7 +63,7 @@ pytz==2022.1 PyYAML==6.0 rdflib==5.0.0 regipy==2.3.1 -requests==2.27.1 +requests==2.32.0 rpm-inspector-rpm==4.16.1.3.210404 saneyaml==0.5.2 six==1.16.0 From 227fba6e1b47a3370f75ed7b4f5c9d114ab43b6c Mon Sep 17 00:00:00 2001 From: Philippe Ombredanne Date: Fri, 14 Jun 2024 17:41:46 +0200 Subject: [PATCH 08/28] Update install command Reference: https://github.com/nexB/fetchcode/issues/117#issuecomment-2165440171 Reported-by: Stuck @Wassaf001 Signed-off-by: Philippe Ombredanne --- README.rst | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/README.rst b/README.rst index 7bb19a77..df744442 100644 --- a/README.rst +++ b/README.rst @@ -26,9 +26,9 @@ Clone the repo:: git clone https://github.com/nexB/fetchcode -Then install all the requirements using:: +Then install all the requirements using this command (on POSIX):: - configure --dev + ./configure --dev Running tests From 124da3dcef0d95a6f6aa76ed849f47ada25b83e2 Mon Sep 17 00:00:00 2001 From: Ayan Sinha Mahapatra Date: Mon, 1 Jul 2024 15:11:21 +0530 Subject: [PATCH 09/28] Replace deprecated macos CI runners Replace macos-11 runners with macos-14 runners. Reference: https://github.com/actions/runner-images?tab=readme-ov-file#available-images Reference: https://github.com/nexB/skeleton/issues/89 Signed-off-by: Ayan Sinha Mahapatra --- azure-pipelines.yml | 16 ++++++++-------- 1 file changed, 8 insertions(+), 8 deletions(-) diff --git a/azure-pipelines.yml b/azure-pipelines.yml index 373b78cd..c2a3b522 100644 --- a/azure-pipelines.yml +++ b/azure-pipelines.yml @@ -25,24 +25,24 @@ jobs: - template: etc/ci/azure-posix.yml parameters: - job_name: macos11_cpython - image_name: macOS-11 + job_name: macos12_cpython + image_name: macOS-12 python_versions: ['3.8', '3.9', '3.10', '3.11', '3.12'] test_suites: all: venv/bin/pytest -n 2 -vvs - template: etc/ci/azure-posix.yml parameters: - job_name: macos12_cpython - image_name: macOS-12 + job_name: macos13_cpython + image_name: macOS-13 python_versions: ['3.8', '3.9', '3.10', '3.11', '3.12'] test_suites: all: venv/bin/pytest -n 2 -vvs - template: etc/ci/azure-posix.yml parameters: - job_name: macos13_cpython - image_name: macOS-13 + job_name: macos14_cpython_arm64 + image_name: macOS-14 python_versions: ['3.8', '3.9', '3.10', '3.11', '3.12'] test_suites: all: venv/bin/pytest -n 2 -vvs @@ -50,8 +50,8 @@ jobs: - template: etc/ci/azure-posix.yml parameters: job_name: macos14_cpython - image_name: macOS-14 - python_versions: ['3.8', '3.9', '3.10', '3.11', '3.12'] + image_name: macOS-14-large + python_versions: ['3.8', '3.8', '3.9', '3.10', '3.12'] test_suites: all: venv/bin/pytest -n 2 -vvs From be4e14d414cf4f7112b529dc71f7abccc9dcf24a Mon Sep 17 00:00:00 2001 From: Ayan Sinha Mahapatra Date: Mon, 1 Jul 2024 16:00:40 +0530 Subject: [PATCH 10/28] Update minimum required python version to 3.8 Signed-off-by: Ayan Sinha Mahapatra --- setup.cfg | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/setup.cfg b/setup.cfg index bd0e58a7..a8e20c5d 100644 --- a/setup.cfg +++ b/setup.cfg @@ -38,7 +38,7 @@ zip_safe = false setup_requires = setuptools_scm[toml] >= 4 -python_requires = >=3.7 +python_requires = >=3.8 install_requires = From b06c4b49410340be331567c27b1a708ed888df58 Mon Sep 17 00:00:00 2001 From: Chin Yeung Li Date: Wed, 14 Aug 2024 13:42:35 +0800 Subject: [PATCH 11/28] Remove the deprecated macos_11 Signed-off-by: Chin Yeung Li --- azure-pipelines.yml | 104 ++++++++++++++++++++------------------------ 1 file changed, 47 insertions(+), 57 deletions(-) diff --git a/azure-pipelines.yml b/azure-pipelines.yml index 764883de..913c0301 100644 --- a/azure-pipelines.yml +++ b/azure-pipelines.yml @@ -1,4 +1,3 @@ - ################################################################################ # We use Azure to run the full tests suites on multiple Python 3.x # on multiple Windows, macOS and Linux versions all on 64 bits @@ -6,59 +5,50 @@ ################################################################################ jobs: - - - template: etc/ci/azure-posix.yml - parameters: - job_name: ubuntu20_cpython - image_name: ubuntu-20.04 - python_versions: ['3.7', '3.8', '3.9', '3.10', '3.11'] - test_suites: - all: venv/bin/pytest -n 2 -vvs - - - template: etc/ci/azure-posix.yml - parameters: - job_name: ubuntu22_cpython - image_name: ubuntu-22.04 - python_versions: ['3.7', '3.8', '3.9', '3.10', '3.11'] - test_suites: - all: venv/bin/pytest -n 2 -vvs - - - template: etc/ci/azure-posix.yml - parameters: - job_name: macos11_cpython - image_name: macOS-11 - python_versions: ['3.7', '3.8', '3.9', '3.10', '3.11'] - test_suites: - all: venv/bin/pytest -n 2 -vvs - - - template: etc/ci/azure-posix.yml - parameters: - job_name: macos12_cpython - image_name: macOS-12 - python_versions: ['3.7', '3.8', '3.9', '3.10', '3.11'] - test_suites: - all: venv/bin/pytest -n 2 -vvs - - - template: etc/ci/azure-posix.yml - parameters: - job_name: macos13_cpython - image_name: macOS-13 - python_versions: ['3.7', '3.8', '3.9', '3.10', '3.11'] - test_suites: - all: venv/bin/pytest -n 2 -vvs - - - template: etc/ci/azure-win.yml - parameters: - job_name: win2019_cpython - image_name: windows-2019 - python_versions: ['3.7', '3.8', '3.9', '3.10', '3.11'] - test_suites: - all: venv\Scripts\pytest -n 2 -vvs - - - template: etc/ci/azure-win.yml - parameters: - job_name: win2022_cpython - image_name: windows-2022 - python_versions: ['3.7', '3.8', '3.9', '3.10', '3.11'] - test_suites: - all: venv\Scripts\pytest -n 2 -vvs + - template: etc/ci/azure-posix.yml + parameters: + job_name: ubuntu20_cpython + image_name: ubuntu-20.04 + python_versions: ["3.7", "3.8", "3.9", "3.10", "3.11"] + test_suites: + all: venv/bin/pytest -n 2 -vvs + + - template: etc/ci/azure-posix.yml + parameters: + job_name: ubuntu22_cpython + image_name: ubuntu-22.04 + python_versions: ["3.7", "3.8", "3.9", "3.10", "3.11"] + test_suites: + all: venv/bin/pytest -n 2 -vvs + + - template: etc/ci/azure-posix.yml + parameters: + job_name: macos12_cpython + image_name: macOS-12 + python_versions: ["3.7", "3.8", "3.9", "3.10", "3.11"] + test_suites: + all: venv/bin/pytest -n 2 -vvs + + - template: etc/ci/azure-posix.yml + parameters: + job_name: macos13_cpython + image_name: macOS-13 + python_versions: ["3.7", "3.8", "3.9", "3.10", "3.11"] + test_suites: + all: venv/bin/pytest -n 2 -vvs + + - template: etc/ci/azure-win.yml + parameters: + job_name: win2019_cpython + image_name: windows-2019 + python_versions: ["3.7", "3.8", "3.9", "3.10", "3.11"] + test_suites: + all: venv\Scripts\pytest -n 2 -vvs + + - template: etc/ci/azure-win.yml + parameters: + job_name: win2022_cpython + image_name: windows-2022 + python_versions: ["3.7", "3.8", "3.9", "3.10", "3.11"] + test_suites: + all: venv\Scripts\pytest -n 2 -vvs From 8b2125c2c3e7a86fb6f28be1ed34278fb9df0b30 Mon Sep 17 00:00:00 2001 From: Chin Yeung Li Date: Wed, 14 Aug 2024 13:50:28 +0800 Subject: [PATCH 12/28] #122 - Update link references of ownership from nexB to aboutcode-org Signed-off-by: Chin Yeung Li --- NOTICE | 2 +- README.rst | 8 +- configure | 2 +- configure.bat | 2 +- setup.cfg | 2 +- src/fetchcode/__init__.py | 12 +- src/fetchcode/package.py | 62 ++- src/fetchcode/package_util.py | 11 +- src/fetchcode/package_versions.py | 10 +- src/fetchcode/packagedcode_models.py | 27 +- src/fetchcode/utils.py | 5 +- src/fetchcode/vcs/__init__.py | 2 +- src/fetchcode/vcs/git.py | 2 +- .../dirlisting/regenerate_mock_data.py | 396 +++++++++--------- .../package/github/regenerate_mock_data.py | 5 +- .../package_versions/regenerate_mock_data.py | 8 +- tests/test_fetch.py | 2 +- tests/test_package.py | 68 ++- tests/test_package_versions.py | 8 +- tests/test_vcs.py | 73 ++-- tests/test_vcs_git.py | 21 +- 21 files changed, 410 insertions(+), 318 deletions(-) diff --git a/NOTICE b/NOTICE index 65936b2b..cbdaef79 100644 --- a/NOTICE +++ b/NOTICE @@ -2,7 +2,7 @@ # Copyright (c) nexB Inc. and others. # SPDX-License-Identifier: Apache-2.0 # -# Visit https://aboutcode.org and https://github.com/nexB/ for support and download. +# Visit https://aboutcode.org and https://github.com/aboutcode-org/ for support and download. # ScanCode is a trademark of nexB Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); diff --git a/README.rst b/README.rst index df744442..2c107598 100644 --- a/README.rst +++ b/README.rst @@ -7,7 +7,7 @@ systems. It can work using plain HTTP and FTP URLs, as well as as used in Python pip and as specified in `SPDX Package Download Location `_ -Homepage and support: https://github.com/nexB/fetchcode +Homepage and support: https://github.com/aboutcode-org/fetchcode Why FetchCode? @@ -24,7 +24,7 @@ Development installation Clone the repo:: - git clone https://github.com/nexB/fetchcode + git clone https://github.com/aboutcode-org/fetchcode Then install all the requirements using this command (on POSIX):: @@ -45,13 +45,13 @@ Usage Fetch a code archive and get a ``fetchcode.fetch.Response`` object back:: >>> from fetchcode import fetch - >>> f = fetch('https://github.com/nexB/fetchcode/archive/ab65b2e645c889887227ea49eb3332d885fd0a54.zip') + >>> f = fetch('https://github.com/aboutcode-org/fetchcode/archive/ab65b2e645c889887227ea49eb3332d885fd0a54.zip') >>> f.location '/tmp/tmp_cm02xsg' >>> f.content_type 'application/zip' >>> f.url - 'https://github.com/nexB/fetchcode/archive/ab65b2e645c889887227ea49eb3332d885fd0a54.zip' + 'https://github.com/aboutcode-org/fetchcode/archive/ab65b2e645c889887227ea49eb3332d885fd0a54.zip' Fetch some package metadata and get a ``fetchcode.packagedcode_models.Package`` object back:: diff --git a/configure b/configure index 926a894e..22d92885 100755 --- a/configure +++ b/configure @@ -3,7 +3,7 @@ # Copyright (c) nexB Inc. and others. All rights reserved. # SPDX-License-Identifier: Apache-2.0 # See http://www.apache.org/licenses/LICENSE-2.0 for the license text. -# See https://github.com/nexB/ for support or download. +# See https://github.com/aboutcode-org/ for support or download. # See https://aboutcode.org for more information about nexB OSS projects. # diff --git a/configure.bat b/configure.bat index 5e95b311..5b9a9d68 100644 --- a/configure.bat +++ b/configure.bat @@ -4,7 +4,7 @@ @rem Copyright (c) nexB Inc. and others. All rights reserved. @rem SPDX-License-Identifier: Apache-2.0 @rem See http://www.apache.org/licenses/LICENSE-2.0 for the license text. -@rem See https://github.com/nexB/ for support or download. +@rem See https://github.com/aboutcode-org/ for support or download. @rem See https://aboutcode.org for more information about nexB OSS projects. diff --git a/setup.cfg b/setup.cfg index 200c39cb..e40f38ac 100644 --- a/setup.cfg +++ b/setup.cfg @@ -6,7 +6,7 @@ license = Apache-2.0 description = fetchcode is a library to reliably fetch code via HTTP, FTP and version control systems. long_description = file:README.rst long_description_content_type = text/x-rst -url = https://github.com/nexB/fetchcode +url = https://github.com/aboutcode-org/fetchcode author = nexB. Inc. and others author_email = info@aboutcode.org diff --git a/src/fetchcode/__init__.py b/src/fetchcode/__init__.py index c573991f..115e4a7e 100644 --- a/src/fetchcode/__init__.py +++ b/src/fetchcode/__init__.py @@ -1,5 +1,5 @@ # fetchcode is a free software tool from nexB Inc. and others. -# Visit https://github.com/nexB/fetchcode for support and download. +# Visit https://github.com/aboutcode-org/fetchcode for support and download. # # Copyright (c) nexB Inc. and others. All rights reserved. # http://nexb.com and http://aboutcode.org @@ -41,7 +41,7 @@ def __init__(self, location, content_type, size, url): def fetch_http(url, location): """ Return a `Response` object built from fetching the content at a HTTP/HTTPS based `url` URL string - saving the content in a file at `location` + saving the content in a file at `location` """ r = requests.get(url) with open(location, 'wb') as f: @@ -51,7 +51,8 @@ def fetch_http(url, location): size = r.headers.get('content-length') size = int(size) if size else None - resp = Response(location=location, content_type=content_type, size=size, url=url) + resp = Response(location=location, + content_type=content_type, size=size, url=url) return resp @@ -59,7 +60,7 @@ def fetch_http(url, location): def fetch_ftp(url, location): """ Return a `Response` object built from fetching the content at a FTP based `url` URL string - saving the content in a file at `location` + saving the content in a file at `location` """ url_parts = urlparse(url) @@ -84,7 +85,8 @@ def fetch_ftp(url, location): ftp.retrbinary(file, f.write) ftp.close() - resp = Response(location=location, content_type=content_type, size=size, url=url) + resp = Response(location=location, + content_type=content_type, size=size, url=url) return resp diff --git a/src/fetchcode/package.py b/src/fetchcode/package.py index 22d50b9c..edba3733 100644 --- a/src/fetchcode/package.py +++ b/src/fetchcode/package.py @@ -1,5 +1,5 @@ # fetchcode is a free software tool from nexB Inc. and others. -# Visit https://github.com/nexB/fetchcode for support and download. +# Visit https://github.com/aboutcode-org/fetchcode for support and download. # # Copyright (c) nexB Inc. and others. All rights reserved. # http://nexb.com and http://aboutcode.org @@ -92,7 +92,8 @@ def get_cargo_data_from_purl(purl): ) versions = response.get("versions", []) for version in versions: - version_purl = PackageURL(type=purl.type, name=name, version=version.get("num")) + version_purl = PackageURL( + type=purl.type, name=name, version=version.get("num")) dl_path = version.get("dl_path") if dl_path: download_url = f"{base_url}/{dl_path}" @@ -355,7 +356,8 @@ def get_gnu_data_from_purl(purl): purl = PackageURL.from_string(purl) source_archive_url = f"https://ftp.gnu.org/pub/gnu/{purl.name}/" version_regex_template = r"^({}-)(?P[\w.-]*)(.tar.gz)$" - version_regex = re.compile(version_regex_template.format(re.escape(purl.name))) + version_regex = re.compile( + version_regex_template.format(re.escape(purl.name))) yield from extract_packages_from_listing( purl, source_archive_url, version_regex, [] @@ -427,7 +429,8 @@ def get_package_info(cls, package_url): else: for version, data in UDHCP_RELEASES.items(): - purl = PackageURL(type="generic", name="udhcp", version=version) + purl = PackageURL( + type="generic", name="udhcp", version=version) yield Package( homepage_url=cls.source_url, download_url=data["url"], @@ -481,7 +484,8 @@ class UtilLinuxDirectoryListedSource(DirectoryListedSource): class BusyBoxDirectoryListedSource(DirectoryListedSource): source_url = "https://www.busybox.net/downloads/" # Source archive ex: busybox-1.2.3.tar.bz2 - source_archive_regex = re.compile(r"^(busybox-)(?P[\w.-]*)(.tar.bz2)$") + source_archive_regex = re.compile( + r"^(busybox-)(?P[\w.-]*)(.tar.bz2)$") is_nested = False ignored_files_and_dir = [] @@ -489,7 +493,8 @@ class BusyBoxDirectoryListedSource(DirectoryListedSource): class UclibcDirectoryListedSource(DirectoryListedSource): source_url = "https://www.uclibc.org/downloads/" # Source archive ex: uClibc-1.2.3.tar.gz - source_archive_regex = re.compile(r"^(uClibc-)(?P[\w.-]*)(.tar.gz)$") + source_archive_regex = re.compile( + r"^(uClibc-)(?P[\w.-]*)(.tar.gz)$") is_nested = False ignored_files_and_dir = [] @@ -497,7 +502,8 @@ class UclibcDirectoryListedSource(DirectoryListedSource): class UclibcNGDirectoryListedSource(DirectoryListedSource): source_url = "https://downloads.uclibc-ng.org/releases/" # Source archive ex: uClibc-ng-1.2.3.tar.gz - source_archive_regex = re.compile(r"^(uClibc-ng-)(?P[\w.-]*)(.tar.gz)$") + source_archive_regex = re.compile( + r"^(uClibc-ng-)(?P[\w.-]*)(.tar.gz)$") is_nested = True ignored_files_and_dir = [] @@ -505,7 +511,8 @@ class UclibcNGDirectoryListedSource(DirectoryListedSource): class Bzip2DirectoryListedSource(DirectoryListedSource): source_url = "https://sourceware.org/pub/bzip2/" # Source archive ex: bzip2-1.2.3.tar.gz - source_archive_regex = re.compile(r"^(bzip2-)(?P[\w.-]*)(.tar.gz)$") + source_archive_regex = re.compile( + r"^(bzip2-)(?P[\w.-]*)(.tar.gz)$") is_nested = False ignored_files_and_dir = [] @@ -513,7 +520,8 @@ class Bzip2DirectoryListedSource(DirectoryListedSource): class OpenSSHDirectoryListedSource(DirectoryListedSource): source_url = "https://cdn.openbsd.org/pub/OpenBSD/OpenSSH/" # Source archive ex: openssh-1.2.3.tar.gz - source_archive_regex = re.compile(r"^(openssh-)(?P[\w.-]*)(.tgz|.tar.gz)$") + source_archive_regex = re.compile( + r"^(openssh-)(?P[\w.-]*)(.tgz|.tar.gz)$") is_nested = False ignored_files_and_dir = [] @@ -531,7 +539,8 @@ class DnsmasqDirectoryListedSource(DirectoryListedSource): class EbtablesDirectoryListedSource(DirectoryListedSource): source_url = "https://www.netfilter.org/pub/ebtables/" # Source archive ex: ebtables-1.2.3.tar.gz - source_archive_regex = re.compile(r"^(ebtables-)(?P[\w.-]*)(.tar.gz)$") + source_archive_regex = re.compile( + r"^(ebtables-)(?P[\w.-]*)(.tar.gz)$") is_nested = False ignored_files_and_dir = [] @@ -539,7 +548,8 @@ class EbtablesDirectoryListedSource(DirectoryListedSource): class HostapdDirectoryListedSource(DirectoryListedSource): source_url = "https://w1.fi/releases/" # Source archive ex: hostapd-1.2.3.tar.gz - source_archive_regex = re.compile(r"^(hostapd-)(?P[\w.-]*)(.tar.gz)$") + source_archive_regex = re.compile( + r"^(hostapd-)(?P[\w.-]*)(.tar.gz)$") is_nested = False ignored_files_and_dir = [] @@ -557,7 +567,8 @@ class Iproute2DirectoryListedSource(DirectoryListedSource): class IptablesDirectoryListedSource(DirectoryListedSource): source_url = "https://www.netfilter.org/pub/iptables/" # Source archive ex: iptables-1.2.3.tar.bz2 - source_archive_regex = re.compile(r"^(iptables-)(?P[\w.-]*)(.tar.bz2)$") + source_archive_regex = re.compile( + r"^(iptables-)(?P[\w.-]*)(.tar.bz2)$") is_nested = False ignored_files_and_dir = [] @@ -565,7 +576,8 @@ class IptablesDirectoryListedSource(DirectoryListedSource): class LibnlDirectoryListedSource(DirectoryListedSource): source_url = "https://www.infradead.org/~tgr/libnl/files/" # Source archive ex: libnl-1.2.3.tar.gz - source_archive_regex = re.compile(r"^(libnl-)(?P[\w.-]*)(.tar.gz)$") + source_archive_regex = re.compile( + r"^(libnl-)(?P[\w.-]*)(.tar.gz)$") is_nested = False ignored_files_and_dir = [] @@ -573,7 +585,8 @@ class LibnlDirectoryListedSource(DirectoryListedSource): class LighttpdDirectoryListedSource(DirectoryListedSource): source_url = "https://download.lighttpd.net/lighttpd/releases-1.4.x/" # Source archive ex: lighttpd-1.2.3.tar.gz - source_archive_regex = re.compile(r"^(lighttpd-)(?P[\w.-]*)(.tar.gz)$") + source_archive_regex = re.compile( + r"^(lighttpd-)(?P[\w.-]*)(.tar.gz)$") is_nested = False ignored_files_and_dir = [] @@ -601,7 +614,8 @@ class WpaSupplicantDirectoryListedSource(DirectoryListedSource): class SyslinuxDirectoryListedSource(DirectoryListedSource): source_url = "https://mirrors.edge.kernel.org/pub/linux/utils/boot/syslinux/" # Source archive ex: syslinux-1.2.3.tar.gz - source_archive_regex = re.compile(r"^(syslinux-)(?P[\w.-]*)(.tar.gz)$") + source_archive_regex = re.compile( + r"^(syslinux-)(?P[\w.-]*)(.tar.gz)$") is_nested = False ignored_files_and_dir = [] @@ -609,7 +623,8 @@ class SyslinuxDirectoryListedSource(DirectoryListedSource): class SyslinuxDirectoryListedSource(DirectoryListedSource): source_url = "https://mirrors.edge.kernel.org/pub/linux/utils/boot/syslinux/" # Source archive ex: syslinux-1.2.3.tar.gz - source_archive_regex = re.compile(r"^(syslinux-)(?P[\w.-]*)(.tar.gz)$") + source_archive_regex = re.compile( + r"^(syslinux-)(?P[\w.-]*)(.tar.gz)$") is_nested = False ignored_files_and_dir = [] @@ -646,7 +661,8 @@ class DropbearDirectoryListedSource(DirectoryListedSource): class SambaDirectoryListedSource(DirectoryListedSource): source_url = "https://download.samba.org/pub/samba/stable/" # Source archive ex: samba-1.2.3.tar.gz - source_archive_regex = re.compile(r"^(samba-)(?P[\w.-]*)(.tar.gz)$") + source_archive_regex = re.compile( + r"^(samba-)(?P[\w.-]*)(.tar.gz)$") is_nested = False ignored_files_and_dir = [] @@ -654,7 +670,8 @@ class SambaDirectoryListedSource(DirectoryListedSource): class MtdUtilsDirectoryListedSource(DirectoryListedSource): source_url = "https://infraroot.at/pub/mtd/" # Source archive ex: mtd-utils-1.2.3.tar.bz2 - source_archive_regex = re.compile(r"^(mtd-utils-)(?P[\w.-]*)(.tar.bz2)$") + source_archive_regex = re.compile( + r"^(mtd-utils-)(?P[\w.-]*)(.tar.bz2)$") is_nested = False ignored_files_and_dir = [] @@ -662,7 +679,8 @@ class MtdUtilsDirectoryListedSource(DirectoryListedSource): class BareboxDirectoryListedSource(DirectoryListedSource): source_url = "https://www.barebox.org/download/" # Source archive ex: barebox-1.2.3.tar.bz2 - source_archive_regex = re.compile(r"^(barebox-)(?P[\w.-]*)(.tar.bz2)$") + source_archive_regex = re.compile( + r"^(barebox-)(?P[\w.-]*)(.tar.bz2)$") is_nested = False ignored_files_and_dir = [] @@ -670,7 +688,8 @@ class BareboxDirectoryListedSource(DirectoryListedSource): class LinuxDirectoryListedSource(DirectoryListedSource): source_url = "https://mirrors.edge.kernel.org/pub/linux/kernel/" # Source archive ex: linux-1.2.3.tar.gz - source_archive_regex = re.compile(r"^(linux-)(?P[\w.-]*)(.tar.gz)$") + source_archive_regex = re.compile( + r"^(linux-)(?P[\w.-]*)(.tar.gz)$") is_nested = True ignored_files_and_dir = [ "Historic/", @@ -692,7 +711,8 @@ class E2fsprogsDirectoryListedSource(DirectoryListedSource): "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/" ) # Source archive ex: e2fsprogs-1.2.3.tar.gz - source_archive_regex = re.compile(r"^(e2fsprogs-)(?P[\w.-]*)(.tar.gz)$") + source_archive_regex = re.compile( + r"^(e2fsprogs-)(?P[\w.-]*)(.tar.gz)$") is_nested = True ignored_files_and_dir = ["testing/"] diff --git a/src/fetchcode/package_util.py b/src/fetchcode/package_util.py index 282f9261..2acb1b95 100644 --- a/src/fetchcode/package_util.py +++ b/src/fetchcode/package_util.py @@ -1,5 +1,5 @@ # fetchcode is a free software tool from nexB Inc. and others. -# Visit https://github.com/nexB/fetchcode for support and download. +# Visit https://github.com/aboutcode-org/fetchcode for support and download. # # Copyright (c) nexB Inc. and others. All rights reserved. # http://nexb.com and http://aboutcode.org @@ -165,7 +165,8 @@ class SquashfsToolsGitHubSource(GitHubSource): class PupnpGitHubSource(GitHubSource): - version_regex = re.compile(r"\brelease-?(?P(?:\d+(\.\d+){1,2}))\b") + version_regex = re.compile( + r"\brelease-?(?P(?:\d+(\.\d+){1,2}))\b") ignored_tag_regex = None @@ -180,7 +181,8 @@ class BpftoolGitHubSource(GitHubSource): class SqliteGitHubSource(GitHubSource): - version_regex = re.compile(r"\bversion-?(?P(?:\d+(\.\d+){1,2}))\b") + version_regex = re.compile( + r"\bversion-?(?P(?:\d+(\.\d+){1,2}))\b") ignored_tag_regex = None @@ -190,7 +192,8 @@ class LlvmGitHubSource(GitHubSource): class RpmGitHubSource(GitHubSource): - version_regex = re.compile(r"rpm-(?P[^-]+(?:-(?!release).*)?|-release)") + version_regex = re.compile( + r"rpm-(?P[^-]+(?:-(?!release).*)?|-release)") ignored_tag_regex = None diff --git a/src/fetchcode/package_versions.py b/src/fetchcode/package_versions.py index 0a565d04..d2720b3d 100644 --- a/src/fetchcode/package_versions.py +++ b/src/fetchcode/package_versions.py @@ -1,5 +1,5 @@ # fetchcode is a free software tool from nexB Inc. and others. -# Visit https://github.com/nexB/fetchcode for support and download. +# Visit https://github.com/aboutcode-org/fetchcode for support and download. # # Copyright (c) nexB Inc. and others. All rights reserved. # http://nexb.com and http://aboutcode.org @@ -317,7 +317,8 @@ def get_golang_versions_from_purl(purl): break if response is None or escaped_pkg is None or trimmed_pkg is None: - logger.error(f"Error while fetching versions for {package_slug!r} from goproxy") + logger.error( + f"Error while fetching versions for {package_slug!r} from goproxy") return for version_info in response.split("\n"): @@ -347,7 +348,7 @@ def trim_go_url_path(url_path: str) -> Optional[str]: # some advisories contains this prefix in package name, e.g. https://github.com/advisories/GHSA-7h6j-2268-fhcm go_url_prefix = "https://pkg.go.dev/" if url_path.startswith(go_url_prefix): - url_path = url_path[len(go_url_prefix) :] + url_path = url_path[len(go_url_prefix):] parsed_url_path = urlparse(url_path) path = parsed_url_path.path @@ -408,7 +409,8 @@ def fetch_version_info(version_info: str, escaped_pkg: str) -> Optional[PackageV f"Error while fetching version info for {escaped_pkg}/{escaped_ver} " f"from goproxy:\n{traceback.format_exc()}" ) - release_date = dateparser.parse(response.get("Time", "")) if response else None + release_date = dateparser.parse( + response.get("Time", "")) if response else None return PackageVersion(value=version, release_date=release_date) diff --git a/src/fetchcode/packagedcode_models.py b/src/fetchcode/packagedcode_models.py index fd48ae78..f76cae3c 100644 --- a/src/fetchcode/packagedcode_models.py +++ b/src/fetchcode/packagedcode_models.py @@ -1,7 +1,7 @@ -# Copied from https://github.com/nexB/scancode-toolkit/blob/b4ea9c640f8ee4ed8851b5618c6d223bb1c02d47/src/packagedcode/models.py +# Copied from https://github.com/aboutcode-org/scancode-toolkit/blob/b4ea9c640f8ee4ed8851b5618c6d223bb1c02d47/src/packagedcode/models.py # # Copyright (c) 2017 nexB Inc. and others. All rights reserved. -# http://nexb.com and https://github.com/nexB/scancode-toolkit/ +# http://nexb.com and https://github.com/aboutcode-org/scancode-toolkit/ # The ScanCode software is licensed under the Apache License version 2.0. # Data generated with ScanCode require an acknowledgment. # ScanCode is a trademark of nexB Inc. @@ -21,7 +21,7 @@ # ScanCode should be considered or used as legal advice. Consult an Attorney # for any legal advice. # ScanCode is a free software code scanning tool from nexB Inc. and others. -# Visit https://github.com/nexB/scancode-toolkit/ for support and download. +# Visit https://github.com/aboutcode-org/scancode-toolkit/ for support and download. from __future__ import absolute_import from __future__ import print_function @@ -144,7 +144,7 @@ class Party(BaseModel): validator=choices(PARTY_TYPES), label='party type', help='the type of this party: One of: ' - +', '.join(p for p in PARTY_TYPES if p)) + + ', '.join(p for p in PARTY_TYPES if p)) role = String( label='party role', @@ -284,7 +284,8 @@ def set_purl(self, package_url): if not isinstance(package_url, PackageURL): package_url = PackageURL.from_string(package_url) - attribs = ['type', 'namespace', 'name', 'version', 'qualifiers', 'subpath'] + attribs = ['type', 'namespace', 'name', + 'version', 'qualifiers', 'subpath'] for att in attribs: self_val = getattr(self, att) purl_val = getattr(package_url, att) @@ -302,7 +303,8 @@ def to_dict(self, **kwargs): mapping['repository_download_url'] = self.repository_download_url() mapping['api_data_url'] = self.api_data_url() if self.qualifiers: - mapping['qualifiers'] = normalize_qualifiers(self.qualifiers, encode=False) + mapping['qualifiers'] = normalize_qualifiers( + self.qualifiers, encode=False) return mapping @classmethod @@ -399,7 +401,7 @@ class Package(BasePackage): api_url = String( label='API URL', - help='URL of API for this package.') + help='URL of API for this package.') size = Integer( default=None, @@ -433,8 +435,8 @@ class Package(BasePackage): vcs_url = String( help='a URL to the VCS repository in the SPDX form of: ' 'https://github.com/nexb/scancode-toolkit.git@405aaa4b3 ' - 'See SPDX specification "Package Download Location" ' - 'at https://spdx.org/spdx-specification-21-web-version#h.49x2ik5 ') + 'See SPDX specification "Package Download Location" ' + 'at https://spdx.org/spdx-specification-21-web-version#h.49x2ik5 ') copyright = String( label='Copyright', @@ -608,7 +610,8 @@ class DebianPackage(Package): metafiles = ('*.control',) extensions = ('.deb',) filetypes = ('debian binary package',) - mimetypes = ('application/x-archive', 'application/vnd.debian.binary-package',) + mimetypes = ('application/x-archive', + 'application/vnd.debian.binary-package',) default_type = 'deb' @@ -670,6 +673,8 @@ class IvyJar(JavaJar): default_primary_language = 'Java' # FIXME: move to bower.py + + @attr.s() class BowerPackage(Package): metafiles = ('bower.json',) @@ -842,4 +847,4 @@ class SquashfsPackage(Package): default_type = 'squashfs' -# TODO: Add VM images formats(VMDK, OVA, OVF, VDI, etc) and Docker/other containers \ No newline at end of file +# TODO: Add VM images formats(VMDK, OVA, OVF, VDI, etc) and Docker/other containers diff --git a/src/fetchcode/utils.py b/src/fetchcode/utils.py index 1e5ab842..3116caca 100644 --- a/src/fetchcode/utils.py +++ b/src/fetchcode/utils.py @@ -1,5 +1,5 @@ # fetchcode is a free software tool from nexB Inc. and others. -# Visit https://github.com/nexB/fetchcode for support and download. +# Visit https://github.com/aboutcode-org/fetchcode for support and download. # # Copyright (c) nexB Inc. and others. All rights reserved. # http://nexb.com and http://aboutcode.org @@ -141,7 +141,8 @@ def github_response(graphql_query): headers = {"Authorization": f"bearer {gh_token}"} endpoint = "https://api.github.com/graphql" - response = requests.post(endpoint, headers=headers, json=graphql_query).json() + response = requests.post(endpoint, headers=headers, + json=graphql_query).json() message = response.get("message") if message and message == "Bad credentials": diff --git a/src/fetchcode/vcs/__init__.py b/src/fetchcode/vcs/__init__.py index db9cb4da..ef429ef5 100644 --- a/src/fetchcode/vcs/__init__.py +++ b/src/fetchcode/vcs/__init__.py @@ -1,5 +1,5 @@ # fetchcode is a free software tool from nexB Inc. and others. -# Visit https://github.com/nexB/fetchcode for support and download. +# Visit https://github.com/aboutcode-org/fetchcode for support and download. # # Copyright (c) nexB Inc. and others. All rights reserved. # http://nexb.com and http://aboutcode.org diff --git a/src/fetchcode/vcs/git.py b/src/fetchcode/vcs/git.py index 9d9c6cca..f602516c 100644 --- a/src/fetchcode/vcs/git.py +++ b/src/fetchcode/vcs/git.py @@ -1,5 +1,5 @@ # fetchcode is a free software tool from nexB Inc. and others. -# Visit https://github.com/nexB/fetchcode for support and download. +# Visit https://github.com/aboutcode-org/fetchcode for support and download. # # Copyright (c) nexB Inc. and others. All rights reserved. # http://nexb.com and http://aboutcode.org diff --git a/tests/data/package/dirlisting/regenerate_mock_data.py b/tests/data/package/dirlisting/regenerate_mock_data.py index a3161665..067b8308 100644 --- a/tests/data/package/dirlisting/regenerate_mock_data.py +++ b/tests/data/package/dirlisting/regenerate_mock_data.py @@ -1,5 +1,5 @@ # fetchcode is a free software tool from nexB Inc. and others. -# Visit https://github.com/nexB/fetchcode for support and download. +# Visit https://github.com/aboutcode-org/fetchcode for support and download. # Copyright (c) nexB Inc. and others. All rights reserved. # http://nexb.com and http://aboutcode.org @@ -601,204 +601,204 @@ }, ], }, - { - "purl": "pkg:generic/e2fsprogs", - "sources": [ - { - "filename": "generic/e2fsprogs/index.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/", - }, - { - "filename": "generic/e2fsprogs/0.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.1/", - }, - { - "filename": "generic/e2fsprogs/1.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.10/", - }, - { - "filename": "generic/e2fsprogs/2.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.11/", - }, - { - "filename": "generic/e2fsprogs/3.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.12/", - }, - { - "filename": "generic/e2fsprogs/4.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.13/", - }, - { - "filename": "generic/e2fsprogs/5.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.2/", - }, - { - "filename": "generic/e2fsprogs/6.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.3/", - }, - { - "filename": "generic/e2fsprogs/7.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.4/", - }, - { - "filename": "generic/e2fsprogs/8.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.5/", - }, - { - "filename": "generic/e2fsprogs/9.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.6/", - }, - { - "filename": "generic/e2fsprogs/10.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.7/", - }, - { - "filename": "generic/e2fsprogs/11.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.8/", - }, - { - "filename": "generic/e2fsprogs/12.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.9/", - }, - { - "filename": "generic/e2fsprogs/13.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.43/", - }, - { - "filename": "generic/e2fsprogs/14.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.43.1/", - }, - { - "filename": "generic/e2fsprogs/15.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.43.2/", - }, - { - "filename": "generic/e2fsprogs/16.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.43.3/", - }, - { - "filename": "generic/e2fsprogs/17.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.43.4/", - }, - { - "filename": "generic/e2fsprogs/18.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.43.5/", - }, - { - "filename": "generic/e2fsprogs/19.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.43.6/", - }, - { - "filename": "generic/e2fsprogs/20.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.43.7/", - }, - { - "filename": "generic/e2fsprogs/21.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.43.8/", - }, - { - "filename": "generic/e2fsprogs/22.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.43.9/", - }, - { - "filename": "generic/e2fsprogs/23.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.44.0/", - }, - { - "filename": "generic/e2fsprogs/24.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.44.1/", - }, - { - "filename": "generic/e2fsprogs/25.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.44.2/", - }, - { - "filename": "generic/e2fsprogs/26.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.44.3/", - }, - { - "filename": "generic/e2fsprogs/27.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.44.3-rc2/", - }, - { - "filename": "generic/e2fsprogs/28.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.44.4/", - }, - { - "filename": "generic/e2fsprogs/29.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.44.5/", - }, - { - "filename": "generic/e2fsprogs/30.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.44.6/", - }, - { - "filename": "generic/e2fsprogs/31.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.45.0/", - }, - { - "filename": "generic/e2fsprogs/32.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.45.1/", - }, - { - "filename": "generic/e2fsprogs/33.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.45.2/", - }, - { - "filename": "generic/e2fsprogs/34.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.45.3/", - }, - { - "filename": "generic/e2fsprogs/35.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.45.4/", - }, - { - "filename": "generic/e2fsprogs/36.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.45.5/", - }, - { - "filename": "generic/e2fsprogs/37.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.45.6/", - }, - { - "filename": "generic/e2fsprogs/38.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.45.7/", - }, - { - "filename": "generic/e2fsprogs/39.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.46.0/", - }, - { - "filename": "generic/e2fsprogs/40.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.46.1/", - }, - { - "filename": "generic/e2fsprogs/41.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.46.2/", - }, - { - "filename": "generic/e2fsprogs/42.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.46.3/", - }, - { - "filename": "generic/e2fsprogs/43.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.46.4/", - }, - { - "filename": "generic/e2fsprogs/44.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.46.5/", - }, - { - "filename": "generic/e2fsprogs/45.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.46.6/", - }, - { - "filename": "generic/e2fsprogs/46.html", - "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.47.0/", - }, + { + "purl": "pkg:generic/e2fsprogs", + "sources": [ + { + "filename": "generic/e2fsprogs/index.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/", + }, + { + "filename": "generic/e2fsprogs/0.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.1/", + }, + { + "filename": "generic/e2fsprogs/1.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.10/", + }, + { + "filename": "generic/e2fsprogs/2.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.11/", + }, + { + "filename": "generic/e2fsprogs/3.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.12/", + }, + { + "filename": "generic/e2fsprogs/4.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.13/", + }, + { + "filename": "generic/e2fsprogs/5.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.2/", + }, + { + "filename": "generic/e2fsprogs/6.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.3/", + }, + { + "filename": "generic/e2fsprogs/7.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.4/", + }, + { + "filename": "generic/e2fsprogs/8.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.5/", + }, + { + "filename": "generic/e2fsprogs/9.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.6/", + }, + { + "filename": "generic/e2fsprogs/10.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.7/", + }, + { + "filename": "generic/e2fsprogs/11.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.8/", + }, + { + "filename": "generic/e2fsprogs/12.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.42.9/", + }, + { + "filename": "generic/e2fsprogs/13.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.43/", + }, + { + "filename": "generic/e2fsprogs/14.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.43.1/", + }, + { + "filename": "generic/e2fsprogs/15.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.43.2/", + }, + { + "filename": "generic/e2fsprogs/16.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.43.3/", + }, + { + "filename": "generic/e2fsprogs/17.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.43.4/", + }, + { + "filename": "generic/e2fsprogs/18.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.43.5/", + }, + { + "filename": "generic/e2fsprogs/19.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.43.6/", + }, + { + "filename": "generic/e2fsprogs/20.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.43.7/", + }, + { + "filename": "generic/e2fsprogs/21.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.43.8/", + }, + { + "filename": "generic/e2fsprogs/22.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.43.9/", + }, + { + "filename": "generic/e2fsprogs/23.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.44.0/", + }, + { + "filename": "generic/e2fsprogs/24.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.44.1/", + }, + { + "filename": "generic/e2fsprogs/25.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.44.2/", + }, + { + "filename": "generic/e2fsprogs/26.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.44.3/", + }, + { + "filename": "generic/e2fsprogs/27.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.44.3-rc2/", + }, + { + "filename": "generic/e2fsprogs/28.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.44.4/", + }, + { + "filename": "generic/e2fsprogs/29.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.44.5/", + }, + { + "filename": "generic/e2fsprogs/30.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.44.6/", + }, + { + "filename": "generic/e2fsprogs/31.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.45.0/", + }, + { + "filename": "generic/e2fsprogs/32.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.45.1/", + }, + { + "filename": "generic/e2fsprogs/33.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.45.2/", + }, + { + "filename": "generic/e2fsprogs/34.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.45.3/", + }, + { + "filename": "generic/e2fsprogs/35.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.45.4/", + }, + { + "filename": "generic/e2fsprogs/36.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.45.5/", + }, + { + "filename": "generic/e2fsprogs/37.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.45.6/", + }, + { + "filename": "generic/e2fsprogs/38.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.45.7/", + }, + { + "filename": "generic/e2fsprogs/39.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.46.0/", + }, + { + "filename": "generic/e2fsprogs/40.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.46.1/", + }, + { + "filename": "generic/e2fsprogs/41.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.46.2/", + }, + { + "filename": "generic/e2fsprogs/42.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.46.3/", + }, + { + "filename": "generic/e2fsprogs/43.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.46.4/", + }, + { + "filename": "generic/e2fsprogs/44.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.46.5/", + }, + { + "filename": "generic/e2fsprogs/45.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.46.6/", + }, + { + "filename": "generic/e2fsprogs/46.html", + "url": "https://mirrors.edge.kernel.org/pub/linux/kernel/people/tytso/e2fsprogs/v1.47.0/", + }, - ], -}, + ], + }, ] diff --git a/tests/data/package/github/regenerate_mock_data.py b/tests/data/package/github/regenerate_mock_data.py index 41467300..9a2b33ef 100644 --- a/tests/data/package/github/regenerate_mock_data.py +++ b/tests/data/package/github/regenerate_mock_data.py @@ -1,5 +1,5 @@ # fetchcode is a free software tool from nexB Inc. and others. -# Visit https://github.com/nexB/fetchcode for support and download. +# Visit https://github.com/aboutcode-org/fetchcode for support and download. # Copyright (c) nexB Inc. and others. All rights reserved. # http://nexb.com and http://aboutcode.org @@ -41,7 +41,8 @@ def fetch_github_mock_data(owner, name, subdir): while True: response = github_response(graphql_query) refs = response["data"]["repository"]["refs"] - mock_data_file = data_location / f"{subdir}/github_mock_data_{file_count}.json" + mock_data_file = data_location / \ + f"{subdir}/github_mock_data_{file_count}.json" with open(mock_data_file, "w") as file: json.dump(response, file, indent=2) diff --git a/tests/data/package_versions/regenerate_mock_data.py b/tests/data/package_versions/regenerate_mock_data.py index 68ee76d2..9dbcaefb 100644 --- a/tests/data/package_versions/regenerate_mock_data.py +++ b/tests/data/package_versions/regenerate_mock_data.py @@ -1,5 +1,5 @@ # fetchcode is a free software tool from nexB Inc. and others. -# Visit https://github.com/nexB/fetchcode for support and download. +# Visit https://github.com/aboutcode-org/fetchcode for support and download. # Copyright (c) nexB Inc. and others. All rights reserved. # http://nexb.com and http://aboutcode.org @@ -140,7 +140,8 @@ def fetch_golang_mock_data(): url = f"https://proxy.golang.org/github.com/blockloop/scan/@v/list" version_list = get_response(url=url, content_type="text") for version in version_list.split(): - file_name = data_location / f"golang/versions/golang_mock_{version}_data.json" + file_name = data_location / \ + f"golang/versions/golang_mock_{version}_data.json" response = get_response( url=f"https://proxy.golang.org/github.com/blockloop/scan/@v/{version}.info", content_type="json", @@ -168,7 +169,8 @@ def fetch_github_mock_data(): while True: response = github_response(graphql_query) refs = response["data"]["repository"]["refs"] - mock_data_file = data_location / f"github/github_mock_data_{file_count}.json" + mock_data_file = data_location / \ + f"github/github_mock_data_{file_count}.json" with open(mock_data_file, "w") as file: json.dump(response, file, indent=4) diff --git a/tests/test_fetch.py b/tests/test_fetch.py index 09b0bf4a..6366da36 100644 --- a/tests/test_fetch.py +++ b/tests/test_fetch.py @@ -1,5 +1,5 @@ # fetchcode is a free software tool from nexB Inc. and others. -# Visit https://github.com/nexB/fetchcode for support and download. +# Visit https://github.com/aboutcode-org/fetchcode for support and download. # # Copyright (c) nexB Inc. and others. All rights reserved. # http://nexb.com and http://aboutcode.org diff --git a/tests/test_package.py b/tests/test_package.py index 68c93507..fa1f5ea8 100644 --- a/tests/test_package.py +++ b/tests/test_package.py @@ -1,5 +1,5 @@ # fetchcode is a free software tool from nexB Inc. and others. -# Visit https://github.com/nexB/fetchcode for support and download. +# Visit https://github.com/aboutcode-org/fetchcode for support and download. # Copyright (c) nexB Inc. and others. All rights reserved. # http://nexb.com and http://aboutcode.org @@ -120,7 +120,8 @@ def test_packages_github_source_avahi( test_data = [ "tests/data/package/github/avahi/github_mock_data_1.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/avahi/github_mock_data_0.json" ) @@ -138,7 +139,8 @@ def test_packages_github_source_avahi( test_data = [ "tests/data/package/github/avahi/github_mock_data_1.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/avahi/github_mock_data_0.json" ) @@ -156,7 +158,8 @@ def test_packages_github_source_bpftool( test_data = [ "tests/data/package/github/bpftool/github_mock_data_1.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/bpftool/github_mock_data_0.json" ) @@ -174,7 +177,8 @@ def test_packages_github_source_brotli( test_data = [ "tests/data/package/github/brotli/github_mock_data_1.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/brotli/github_mock_data_0.json" ) @@ -192,7 +196,8 @@ def test_packages_github_source_dosfstools( test_data = [ "tests/data/package/github/dosfstools/github_mock_data_1.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/dosfstools/github_mock_data_0.json" ) @@ -210,7 +215,8 @@ def test_packages_github_source_genext2fs( test_data = [ "tests/data/package/github/genext2fs/github_mock_data_1.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/genext2fs/github_mock_data_0.json" ) @@ -228,7 +234,8 @@ def test_packages_github_source_inotify_tools( test_data = [ "tests/data/package/github/inotify-tools/github_mock_data_1.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/inotify-tools/github_mock_data_0.json" ) @@ -248,7 +255,8 @@ def test_packages_github_source_llvm_project( "tests/data/package/github/llvm-project/github_mock_data_2.json", "tests/data/package/github/llvm-project/github_mock_data_3.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/llvm-project/github_mock_data_0.json" ) @@ -266,7 +274,8 @@ def test_packages_github_source_miniupnpc( test_data = [ "tests/data/package/github/miniupnp/github_mock_data_1.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/miniupnp/github_mock_data_0.json" ) @@ -284,7 +293,8 @@ def test_packages_github_source_miniupnpd( test_data = [ "tests/data/package/github/miniupnp/github_mock_data_1.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/miniupnp/github_mock_data_0.json" ) @@ -302,7 +312,8 @@ def test_packages_github_source_minissdpd( test_data = [ "tests/data/package/github/miniupnp/github_mock_data_1.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/miniupnp/github_mock_data_0.json" ) @@ -319,7 +330,8 @@ def test_packages_github_source_nix(self, mock_github_response, mock_get_respons "tests/data/package/github/nix/github_mock_data_1.json", "tests/data/package/github/nix/github_mock_data_2.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/nix/github_mock_data_0.json" ) @@ -337,7 +349,8 @@ def test_packages_github_source_pupnp( test_data = [ "tests/data/package/github/pupnp/github_mock_data_1.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/pupnp/github_mock_data_0.json" ) @@ -360,7 +373,8 @@ def test_packages_github_source_cpython( "tests/data/package/github/cpython/github_mock_data_5.json", "tests/data/package/github/cpython/github_mock_data_6.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/cpython/github_mock_data_0.json" ) @@ -377,7 +391,8 @@ def test_packages_github_source_rpm(self, mock_github_response, mock_get_respons "tests/data/package/github/rpm/github_mock_data_1.json", "tests/data/package/github/rpm/github_mock_data_2.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/rpm/github_mock_data_0.json" ) @@ -395,7 +410,8 @@ def test_packages_github_source_shadow( test_data = [ "tests/data/package/github/shadow/github_mock_data_1.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/shadow/github_mock_data_0.json" ) @@ -414,7 +430,8 @@ def test_packages_github_source_sqlite( "tests/data/package/github/sqlite/github_mock_data_1.json", "tests/data/package/github/sqlite/github_mock_data_2.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/sqlite/github_mock_data_0.json" ) @@ -432,7 +449,8 @@ def test_packages_github_source_squashfs_tools( test_data = [ "tests/data/package/github/squashfs-tools/github_mock_data_1.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/squashfs-tools/github_mock_data_0.json" ) @@ -450,7 +468,8 @@ def test_packages_github_source_wireless_tools( test_data = [ "tests/data/package/github/wireless-tools/github_mock_data_1.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/wireless-tools/github_mock_data_0.json" ) @@ -472,7 +491,8 @@ def test_packages_github_source_uboot( "tests/data/package/github/u-boot/github_mock_data_4.json", "tests/data/package/github/u-boot/github_mock_data_5.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/u-boot/github_mock_data_0.json" ) @@ -490,7 +510,8 @@ def test_packages_github_source_erofs_utils( test_data = [ "tests/data/package/github/erofs-utils/github_mock_data_1.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/erofs-utils/github_mock_data_0.json" ) @@ -511,7 +532,8 @@ def test_packages_github_source_openssl( "tests/data/package/github/openssl/github_mock_data_3.json", "tests/data/package/github/openssl/github_mock_data_4.json", ] - mock_github_response.side_effect = [file_json(file) for file in test_data] + mock_github_response.side_effect = [ + file_json(file) for file in test_data] mock_get_response.return_value = file_json( "tests/data/package/github/openssl/github_mock_data_0.json" ) diff --git a/tests/test_package_versions.py b/tests/test_package_versions.py index 1b5ad8d4..6c17c7b3 100644 --- a/tests/test_package_versions.py +++ b/tests/test_package_versions.py @@ -1,5 +1,5 @@ # fetchcode is a free software tool from nexB Inc. and others. -# Visit https://github.com/nexB/fetchcode for support and download. +# Visit https://github.com/aboutcode-org/fetchcode for support and download. # Copyright (c) nexB Inc. and others. All rights reserved. # http://nexb.com and http://aboutcode.org @@ -23,7 +23,8 @@ from fetchcode.package_versions import versions -FETCHCODE_REGEN_TEST_FIXTURES = os.getenv("FETCHCODE_REGEN_TEST_FIXTURES", False) +FETCHCODE_REGEN_TEST_FIXTURES = os.getenv( + "FETCHCODE_REGEN_TEST_FIXTURES", False) data_location = Path(__file__).parent / "data" / "package_versions" @@ -190,7 +191,8 @@ def test_get_golang_versions_from_purl(mock_get_response): for version in version_list.split(): side_effect.append( get_json_data( - data_location / f"golang/versions/golang_mock_{version}_data.json" + data_location / + f"golang/versions/golang_mock_{version}_data.json" ) ) diff --git a/tests/test_vcs.py b/tests/test_vcs.py index 869bf800..9c22b0f8 100644 --- a/tests/test_vcs.py +++ b/tests/test_vcs.py @@ -1,5 +1,5 @@ # fetchcode is a free software tool from nexB Inc. and others. -# Visit https://github.com/nexB/fetchcode for support and download. +# Visit https://github.com/aboutcode-org/fetchcode for support and download. # # Copyright (c) nexB Inc. and others. All rights reserved. # http://nexb.com and http://aboutcode.org @@ -25,32 +25,56 @@ def obtain(dest, url): pass + @pytest.mark.parametrize( "url, vcs_type, domain", [ - pytest.param("git+http://github.com/jamesor/mongoose-versioner", "git", "github.com", id="git_http"), - pytest.param("git://github.com/jamesor/mongoose-versioner", "git", "github.com", id="git"), - pytest.param("git+https://github.com/jamesor/mongoose-versioner", "git", "github.com", id="git_https"), - pytest.param("git+ssh://github.com/jamesor/mongoose-versioner", "git", "github.com", id="git_ssh"), - pytest.param("git+file://github.com/jamesor/mongoose-versioner", "git", "github.com", id="git_file"), - pytest.param("git+git://github.com/jamesor/mongoose-versioner", "git", "github.com", id="git_git"), - pytest.param("bzr+http://gitlab.com/jamesor/mongoose-versioner", "bzr", "gitlab.com", id="bzr_http"), - pytest.param("bzr+https://gitlab.com/jamesor/mongoose-versioner", "bzr", "gitlab.com", id="bzr_https"), - pytest.param("bzr://gitlab.com/jamesor/mongoose-versioner", "bzr", "gitlab.com", id="bzr"), - pytest.param("bzr+ssh://gitlab.com/jamesor/mongoose-versioner", "bzr", "gitlab.com", id="bzr_ssh"), - pytest.param("bzr+ftp://gitlab.com/jamesor/mongoose-versioner", "bzr", "gitlab.com", id="bzr_ftp"), - pytest.param("bzr+sftp://gitlab.com/jamesor/mongoose-versioner", "bzr", "gitlab.com", id="bzr_sftp"), - pytest.param("bzr+lp://gitlab.com/jamesor/mongoose-versioner", "bzr", "gitlab.com", id="bzr_lp"), - pytest.param("hg://bitbucket.com/jamesor/mongoose-versioner", "hg", "bitbucket.com", id="hg"), - pytest.param("hg+file://bitbucket.com/jamesor/mongoose-versioner", "hg", "bitbucket.com", id="hg_file"), - pytest.param("hg+http://bitbucket.com/jamesor/mongoose-versioner", "hg", "bitbucket.com", id="hg_http"), - pytest.param("hg+https://bitbucket.com/jamesor/mongoose-versioner", "hg", "bitbucket.com", id="hg_https"), - pytest.param("hg+ssh://bitbucket.com/jamesor/mongoose-versioner", "hg", "bitbucket.com", id="hg_ssh"), - pytest.param("hg+static-http://bitbucket.com/jamesor/mongoose-versioner", "hg", "bitbucket.com", id="hg_static_http"), - pytest.param("svn://bitbucket.com/jamesor/mongoose-versioner", "svn", "bitbucket.com", id="svn"), - pytest.param("svn+http://bitbucket.com/jamesor/mongoose-versioner", "svn", "bitbucket.com", id="svn_http"), - pytest.param("svn+https://bitbucket.com/jamesor/mongoose-versioner", "svn", "bitbucket.com", id="svn_https"), - pytest.param("svn+svn://bitbucket.com/jamesor/mongoose-versioner", "svn", "bitbucket.com", id="svn_svn") + pytest.param("git+http://github.com/jamesor/mongoose-versioner", + "git", "github.com", id="git_http"), + pytest.param("git://github.com/jamesor/mongoose-versioner", + "git", "github.com", id="git"), + pytest.param("git+https://github.com/jamesor/mongoose-versioner", + "git", "github.com", id="git_https"), + pytest.param("git+ssh://github.com/jamesor/mongoose-versioner", + "git", "github.com", id="git_ssh"), + pytest.param("git+file://github.com/jamesor/mongoose-versioner", + "git", "github.com", id="git_file"), + pytest.param("git+git://github.com/jamesor/mongoose-versioner", + "git", "github.com", id="git_git"), + pytest.param("bzr+http://gitlab.com/jamesor/mongoose-versioner", + "bzr", "gitlab.com", id="bzr_http"), + pytest.param("bzr+https://gitlab.com/jamesor/mongoose-versioner", + "bzr", "gitlab.com", id="bzr_https"), + pytest.param("bzr://gitlab.com/jamesor/mongoose-versioner", + "bzr", "gitlab.com", id="bzr"), + pytest.param("bzr+ssh://gitlab.com/jamesor/mongoose-versioner", + "bzr", "gitlab.com", id="bzr_ssh"), + pytest.param("bzr+ftp://gitlab.com/jamesor/mongoose-versioner", + "bzr", "gitlab.com", id="bzr_ftp"), + pytest.param("bzr+sftp://gitlab.com/jamesor/mongoose-versioner", + "bzr", "gitlab.com", id="bzr_sftp"), + pytest.param("bzr+lp://gitlab.com/jamesor/mongoose-versioner", + "bzr", "gitlab.com", id="bzr_lp"), + pytest.param("hg://bitbucket.com/jamesor/mongoose-versioner", + "hg", "bitbucket.com", id="hg"), + pytest.param("hg+file://bitbucket.com/jamesor/mongoose-versioner", + "hg", "bitbucket.com", id="hg_file"), + pytest.param("hg+http://bitbucket.com/jamesor/mongoose-versioner", + "hg", "bitbucket.com", id="hg_http"), + pytest.param("hg+https://bitbucket.com/jamesor/mongoose-versioner", + "hg", "bitbucket.com", id="hg_https"), + pytest.param("hg+ssh://bitbucket.com/jamesor/mongoose-versioner", + "hg", "bitbucket.com", id="hg_ssh"), + pytest.param("hg+static-http://bitbucket.com/jamesor/mongoose-versioner", + "hg", "bitbucket.com", id="hg_static_http"), + pytest.param("svn://bitbucket.com/jamesor/mongoose-versioner", + "svn", "bitbucket.com", id="svn"), + pytest.param("svn+http://bitbucket.com/jamesor/mongoose-versioner", + "svn", "bitbucket.com", id="svn_http"), + pytest.param("svn+https://bitbucket.com/jamesor/mongoose-versioner", + "svn", "bitbucket.com", id="svn_https"), + pytest.param("svn+svn://bitbucket.com/jamesor/mongoose-versioner", + "svn", "bitbucket.com", id="svn_svn") ], ) @mock.patch("fetchcode.vcs.vcs.get_backend_for_scheme") @@ -60,6 +84,7 @@ def test_fetch_via_vcs_returns_response(mock_backend, url, vcs_type, domain): assert response.vcs_type == vcs_type assert response.domain == domain + def test_fetch_with_invalid_scheme(): invalid_urls = [ "https://github.com/TG1999/fetchcode", diff --git a/tests/test_vcs_git.py b/tests/test_vcs_git.py index 3972b768..8434abf4 100644 --- a/tests/test_vcs_git.py +++ b/tests/test_vcs_git.py @@ -1,5 +1,5 @@ # fetchcode is a free software tool from nexB Inc. and others. -# Visit https://github.com/nexB/fetchcode for support and download. +# Visit https://github.com/aboutcode-org/fetchcode for support and download. # # Copyright (c) nexB Inc. and others. All rights reserved. # http://nexb.com and http://aboutcode.org @@ -28,12 +28,18 @@ def obtain(dest, url): @pytest.mark.parametrize( "url, vcs_type, domain", [ - pytest.param("git+http://github.com/jamesor/mongoose-versioner", "git", "github.com", id="git_http"), - pytest.param("git://github.com/jamesor/mongoose-versioner", "git", "github.com", id="git"), - pytest.param("git+https://github.com/jamesor/mongoose-versioner", "git", "github.com", id="git_https"), - pytest.param("git+ssh://github.com/jamesor/mongoose-versioner", "git", "github.com", id="git_ssh"), - pytest.param("git+file://github.com/jamesor/mongoose-versioner", "git", "github.com", id="git_file"), - pytest.param("git+git://github.com/jamesor/mongoose-versioner", "git", "github.com", id="git_git") + pytest.param("git+http://github.com/jamesor/mongoose-versioner", + "git", "github.com", id="git_http"), + pytest.param("git://github.com/jamesor/mongoose-versioner", + "git", "github.com", id="git"), + pytest.param("git+https://github.com/jamesor/mongoose-versioner", + "git", "github.com", id="git_https"), + pytest.param("git+ssh://github.com/jamesor/mongoose-versioner", + "git", "github.com", id="git_ssh"), + pytest.param("git+file://github.com/jamesor/mongoose-versioner", + "git", "github.com", id="git_file"), + pytest.param("git+git://github.com/jamesor/mongoose-versioner", + "git", "github.com", id="git_git") ], ) @mock.patch("fetchcode.vcs.git.vcs.get_backend") @@ -43,6 +49,7 @@ def test_fetch_via_vcs_returns_response(mock_backend, url, vcs_type, domain): assert response.vcs_type == vcs_type assert response.domain == domain + def test_fetch_with_git_invalid_scheme(): invalid_urls = [ "https://github.com/TG1999/fetchcode", From 647cff12326a3106aadfa33b6e6c89dfe26b42be Mon Sep 17 00:00:00 2001 From: Chin Yeung Li Date: Wed, 14 Aug 2024 13:59:08 +0800 Subject: [PATCH 13/28] #122 - Update changelog Signed-off-by: Chin Yeung Li --- CHANGELOG.rst | 9 ++++++--- 1 file changed, 6 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.rst b/CHANGELOG.rst index e93d0c9a..4a8a8064 100644 --- a/CHANGELOG.rst +++ b/CHANGELOG.rst @@ -1,6 +1,10 @@ Changelog ========= +v0.5.2 +------ +- Update link references of ownership from nexB to aboutcode-org + v0.5.1 ------- @@ -9,7 +13,7 @@ v0.5.1 v0.5.0 ------- -- FetchCode now supports retrieving package info for following generic packages: +- FetchCode now supports retrieving package info for following generic packages: * pkg:generic/linux * pkg:generic/mtd-utils * pkg:generic/barebox @@ -42,7 +46,7 @@ v0.5.0 v0.4.0 ------- -- FetchCode now supports retrieving package info for following generic packages: +- FetchCode now supports retrieving package info for following generic packages: ``busybox``, ``bzip2``, ``dnsmasq``, ``dropbear``, ``ebtables``, ``hostapd``, ``ipkg``, ``iproute2``, ``iptables``, ``libnl``, ``lighttpd``, ``nftables``, ``openssh``, ``samba``, ``syslinux``, ``toybox``, ``uclibc``, ``uclibc-ng``, ``util-linux`` and ``wpa_supplicant``. @@ -63,4 +67,3 @@ v0.1.0 --------- First, initial release. - \ No newline at end of file From 9c57f340d22d8891a5614a93553b20d75e2f3136 Mon Sep 17 00:00:00 2001 From: Chin Yeung Li Date: Tue, 20 Aug 2024 16:46:20 +0800 Subject: [PATCH 14/28] Update link references of ownership from nexB to aboutcode-org Signed-off-by: Chin Yeung Li --- Makefile | 4 +- NOTICE | 2 +- configure | 2 +- configure.bat | 2 +- docs/source/conf.py | 2 +- docs/source/contribute/contrib_doc.rst | 2 +- docs/source/skeleton-usage.rst | 2 +- etc/scripts/check_thirdparty.py | 5 +- etc/scripts/fetch_thirdparty.py | 19 ++++-- etc/scripts/gen_requirements.py | 2 +- etc/scripts/gen_requirements_dev.py | 2 +- etc/scripts/utils_dejacode.py | 11 ++-- etc/scripts/utils_requirements.py | 11 ++-- etc/scripts/utils_thirdparty.py | 89 +++++++++++++++++--------- setup.cfg | 2 +- tests/test_skeleton_codestyle.py | 2 +- 16 files changed, 100 insertions(+), 59 deletions(-) diff --git a/Makefile b/Makefile index cc36c355..94451b33 100644 --- a/Makefile +++ b/Makefile @@ -4,7 +4,7 @@ # ScanCode is a trademark of nexB Inc. # SPDX-License-Identifier: Apache-2.0 # See http://www.apache.org/licenses/LICENSE-2.0 for the license text. -# See https://github.com/nexB/skeleton for support or download. +# See https://github.com/aboutcode-org/skeleton for support or download. # See https://aboutcode.org for more information about nexB OSS projects. # @@ -35,7 +35,7 @@ check: @echo "-> Run pycodestyle (PEP8) validation" @${ACTIVATE} pycodestyle --max-line-length=100 --exclude=.eggs,venv,lib,thirdparty,docs,migrations,settings.py,.cache . @echo "-> Run isort imports ordering validation" - @${ACTIVATE} isort --sl --check-only -l 100 setup.py src tests . + @${ACTIVATE} isort --sl --check-only -l 100 setup.py src tests . @echo "-> Run black validation" @${ACTIVATE} black --check --check -l 100 src tests setup.py diff --git a/NOTICE b/NOTICE index 65936b2b..cbdaef79 100644 --- a/NOTICE +++ b/NOTICE @@ -2,7 +2,7 @@ # Copyright (c) nexB Inc. and others. # SPDX-License-Identifier: Apache-2.0 # -# Visit https://aboutcode.org and https://github.com/nexB/ for support and download. +# Visit https://aboutcode.org and https://github.com/aboutcode-org/ for support and download. # ScanCode is a trademark of nexB Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); diff --git a/configure b/configure index 926a894e..22d92885 100755 --- a/configure +++ b/configure @@ -3,7 +3,7 @@ # Copyright (c) nexB Inc. and others. All rights reserved. # SPDX-License-Identifier: Apache-2.0 # See http://www.apache.org/licenses/LICENSE-2.0 for the license text. -# See https://github.com/nexB/ for support or download. +# See https://github.com/aboutcode-org/ for support or download. # See https://aboutcode.org for more information about nexB OSS projects. # diff --git a/configure.bat b/configure.bat index 5e95b311..5b9a9d68 100644 --- a/configure.bat +++ b/configure.bat @@ -4,7 +4,7 @@ @rem Copyright (c) nexB Inc. and others. All rights reserved. @rem SPDX-License-Identifier: Apache-2.0 @rem See http://www.apache.org/licenses/LICENSE-2.0 for the license text. -@rem See https://github.com/nexB/ for support or download. +@rem See https://github.com/aboutcode-org/ for support or download. @rem See https://aboutcode.org for more information about nexB OSS projects. diff --git a/docs/source/conf.py b/docs/source/conf.py index 7771ff09..8c88fa2c 100644 --- a/docs/source/conf.py +++ b/docs/source/conf.py @@ -43,7 +43,7 @@ # This points to aboutcode.readthedocs.io # In case of "undefined label" ERRORS check docs on intersphinx to troubleshoot -# Link was created at commit - https://github.com/nexB/aboutcode/commit/faea9fcf3248f8f198844fe34d43833224ac4a83 +# Link was created at commit - https://github.com/aboutcode-org/aboutcode/commit/faea9fcf3248f8f198844fe34d43833224ac4a83 intersphinx_mapping = { "aboutcode": ("https://aboutcode.readthedocs.io/en/latest/", None), diff --git a/docs/source/contribute/contrib_doc.rst b/docs/source/contribute/contrib_doc.rst index 13882e10..5640db26 100644 --- a/docs/source/contribute/contrib_doc.rst +++ b/docs/source/contribute/contrib_doc.rst @@ -12,7 +12,7 @@ To get started, create or identify a working directory on your local machine. Open that directory and execute the following command in a terminal session:: - git clone https://github.com/nexB/skeleton.git + git clone https://github.com/aboutcode-org/skeleton.git That will create an ``/skeleton`` directory in your working directory. Now you can install the dependencies in a virtualenv:: diff --git a/docs/source/skeleton-usage.rst b/docs/source/skeleton-usage.rst index cde23dcd..6cb4cc5f 100644 --- a/docs/source/skeleton-usage.rst +++ b/docs/source/skeleton-usage.rst @@ -118,7 +118,7 @@ corrected. You can check to see if your corrections are valid by running: Once the wheels are collected and the ABOUT files are generated and correct, upload them to thirdparty.aboutcode.org/pypi by placing the wheels and ABOUT files from the thirdparty directory to the pypi directory at -https://github.com/nexB/thirdparty-packages +https://github.com/aboutcode-org/thirdparty-packages Usage after project initialization diff --git a/etc/scripts/check_thirdparty.py b/etc/scripts/check_thirdparty.py index b052f25b..2daded94 100644 --- a/etc/scripts/check_thirdparty.py +++ b/etc/scripts/check_thirdparty.py @@ -5,7 +5,7 @@ # ScanCode is a trademark of nexB Inc. # SPDX-License-Identifier: Apache-2.0 # See http://www.apache.org/licenses/LICENSE-2.0 for the license text. -# See https://github.com/nexB/skeleton for support or download. +# See https://github.com/aboutcode-org/skeleton for support or download. # See https://aboutcode.org for more information about nexB OSS projects. # import click @@ -17,7 +17,8 @@ @click.option( "-d", "--dest", - type=click.Path(exists=True, readable=True, path_type=str, file_okay=False), + type=click.Path(exists=True, readable=True, + path_type=str, file_okay=False), required=True, help="Path to the thirdparty directory to check.", ) diff --git a/etc/scripts/fetch_thirdparty.py b/etc/scripts/fetch_thirdparty.py index eedf05c6..3f9ff527 100644 --- a/etc/scripts/fetch_thirdparty.py +++ b/etc/scripts/fetch_thirdparty.py @@ -5,7 +5,7 @@ # ScanCode is a trademark of nexB Inc. # SPDX-License-Identifier: Apache-2.0 # See http://www.apache.org/licenses/LICENSE-2.0 for the license text. -# See https://github.com/nexB/skeleton for support or download. +# See https://github.com/aboutcode-org/skeleton for support or download. # See https://aboutcode.org for more information about nexB OSS projects. # @@ -55,7 +55,8 @@ "-d", "--dest", "dest_dir", - type=click.Path(exists=True, readable=True, path_type=str, file_okay=False), + type=click.Path(exists=True, readable=True, + path_type=str, file_okay=False), metavar="DIR", default=utils_thirdparty.THIRDPARTY_DIR, show_default=True, @@ -224,7 +225,8 @@ def fetch_thirdparty( environments = None if wheels: evts = itertools.product(python_versions, operating_systems) - environments = [utils_thirdparty.Environment.from_pyver_and_os(pyv, os) for pyv, os in evts] + environments = [utils_thirdparty.Environment.from_pyver_and_os( + pyv, os) for pyv, os in evts] # Collect PyPI repos repos = [] @@ -260,13 +262,14 @@ def fetch_thirdparty( repos=repos, ) if not fetched: - wheels_or_sdist_not_found[f"{name}=={version}"].append(environment) + wheels_or_sdist_not_found[f"{name}=={version}"].append( + environment) if TRACE: print(f" NOT FOUND") if (sdists or (f"{name}=={version}" in wheels_or_sdist_not_found and name in sdist_only) - ): + ): if TRACE: print(f" ==> Fetching sdist: {name}=={version}") @@ -289,7 +292,8 @@ def fetch_thirdparty( sdist_missing = sdists and "sdist" in dists and not name in wheel_only if sdist_missing: mia.append(f"SDist missing: {nv} {dists}") - wheels_missing = wheels and any(d for d in dists if d != "sdist") and not name in sdist_only + wheels_missing = wheels and any( + d for d in dists if d != "sdist") and not name in sdist_only if wheels_missing: mia.append(f"Wheels missing: {nv} {dists}") @@ -299,7 +303,8 @@ def fetch_thirdparty( raise Exception(mia) print(f"==> FETCHING OR CREATING ABOUT AND LICENSE FILES") - utils_thirdparty.fetch_abouts_and_licenses(dest_dir=dest_dir, use_cached_index=use_cached_index) + utils_thirdparty.fetch_abouts_and_licenses( + dest_dir=dest_dir, use_cached_index=use_cached_index) utils_thirdparty.clean_about_files(dest_dir=dest_dir) # check for problems diff --git a/etc/scripts/gen_requirements.py b/etc/scripts/gen_requirements.py index 07e26f77..2b65ae80 100644 --- a/etc/scripts/gen_requirements.py +++ b/etc/scripts/gen_requirements.py @@ -5,7 +5,7 @@ # ScanCode is a trademark of nexB Inc. # SPDX-License-Identifier: Apache-2.0 # See http://www.apache.org/licenses/LICENSE-2.0 for the license text. -# See https://github.com/nexB/skeleton for support or download. +# See https://github.com/aboutcode-org/skeleton for support or download. # See https://aboutcode.org for more information about nexB OSS projects. # import argparse diff --git a/etc/scripts/gen_requirements_dev.py b/etc/scripts/gen_requirements_dev.py index 12cc06d3..5db1c48e 100644 --- a/etc/scripts/gen_requirements_dev.py +++ b/etc/scripts/gen_requirements_dev.py @@ -5,7 +5,7 @@ # ScanCode is a trademark of nexB Inc. # SPDX-License-Identifier: Apache-2.0 # See http://www.apache.org/licenses/LICENSE-2.0 for the license text. -# See https://github.com/nexB/skeleton for support or download. +# See https://github.com/aboutcode-org/skeleton for support or download. # See https://aboutcode.org for more information about nexB OSS projects. # import argparse diff --git a/etc/scripts/utils_dejacode.py b/etc/scripts/utils_dejacode.py index c42e6c93..652252d4 100644 --- a/etc/scripts/utils_dejacode.py +++ b/etc/scripts/utils_dejacode.py @@ -5,7 +5,7 @@ # ScanCode is a trademark of nexB Inc. # SPDX-License-Identifier: Apache-2.0 # See http://www.apache.org/licenses/LICENSE-2.0 for the license text. -# See https://github.com/nexB/skeleton for support or download. +# See https://github.com/aboutcode-org/skeleton for support or download. # See https://aboutcode.org for more information about nexB OSS projects. # import io @@ -33,7 +33,8 @@ def can_do_api_calls(): if not DEJACODE_API_KEY and DEJACODE_API_URL: - print("DejaCode DEJACODE_API_KEY and DEJACODE_API_URL not configured. Doing nothing") + print( + "DejaCode DEJACODE_API_KEY and DEJACODE_API_URL not configured. Doing nothing") return False else: return True @@ -68,7 +69,8 @@ def get_package_data(distribution): return results[0] elif len_results > 1: - print(f"More than 1 entry exists, review at: {DEJACODE_API_URL_PACKAGES}") + print( + f"More than 1 entry exists, review at: {DEJACODE_API_URL_PACKAGES}") else: print("Could not find package:", distribution.download_url) @@ -149,7 +151,8 @@ def find_latest_dejacode_package(distribution): # there was no exact match, find the latest version # TODO: consider the closest version rather than the latest # or the version that has the best data - with_versions = [(packaging_version.parse(p["version"]), p) for p in packages] + with_versions = [(packaging_version.parse(p["version"]), p) + for p in packages] with_versions = sorted(with_versions) latest_version, latest_package_version = sorted(with_versions)[-1] print( diff --git a/etc/scripts/utils_requirements.py b/etc/scripts/utils_requirements.py index 0fc25a35..1c502390 100644 --- a/etc/scripts/utils_requirements.py +++ b/etc/scripts/utils_requirements.py @@ -5,7 +5,7 @@ # ScanCode is a trademark of nexB Inc. # SPDX-License-Identifier: Apache-2.0 # See http://www.apache.org/licenses/LICENSE-2.0 for the license text. -# See https://github.com/nexB/skeleton for support or download. +# See https://github.com/aboutcode-org/skeleton for support or download. # See https://aboutcode.org for more information about nexB OSS projects. # @@ -102,7 +102,8 @@ def lock_dev_requirements( all_req_nvs = get_required_name_versions(all_req_lines) dev_only_req_nvs = {n: v for n, v in all_req_nvs if n not in main_names} - new_reqs = "\n".join(f"{n}=={v}" for n, v in sorted(dev_only_req_nvs.items())) + new_reqs = "\n".join( + f"{n}=={v}" for n, v in sorted(dev_only_req_nvs.items())) with open(dev_requirements_file, "w") as fo: fo.write(new_reqs) @@ -113,10 +114,12 @@ def get_installed_reqs(site_packages_dir): as a text. """ if not os.path.exists(site_packages_dir): - raise Exception(f"site_packages directory: {site_packages_dir!r} does not exists") + raise Exception( + f"site_packages directory: {site_packages_dir!r} does not exists") # Also include these packages in the output with --all: wheel, distribute, # setuptools, pip - args = ["pip", "freeze", "--exclude-editable", "--all", "--path", site_packages_dir] + args = ["pip", "freeze", "--exclude-editable", + "--all", "--path", site_packages_dir] return subprocess.check_output(args, encoding="utf-8") diff --git a/etc/scripts/utils_thirdparty.py b/etc/scripts/utils_thirdparty.py index addf8e5e..46dc7289 100644 --- a/etc/scripts/utils_thirdparty.py +++ b/etc/scripts/utils_thirdparty.py @@ -5,7 +5,7 @@ # ScanCode is a trademark of nexB Inc. # SPDX-License-Identifier: Apache-2.0 # See http://www.apache.org/licenses/LICENSE-2.0 for the license text. -# See https://github.com/nexB/skeleton for support or download. +# See https://github.com/aboutcode-org/skeleton for support or download. # See https://aboutcode.org for more information about nexB OSS projects. # import email @@ -245,9 +245,11 @@ def download_wheel(name, version, environment, dest_dir=THIRDPARTY_DIR, repos=tu package = repo.get_package_version(name=name, version=version) if not package: if TRACE_DEEP: - print(f" download_wheel: No package in {repo.index_url} for {name}=={version}") + print( + f" download_wheel: No package in {repo.index_url} for {name}=={version}") continue - supported_wheels = list(package.get_supported_wheels(environment=environment)) + supported_wheels = list( + package.get_supported_wheels(environment=environment)) if not supported_wheels: if TRACE_DEEP: print( @@ -291,7 +293,8 @@ def download_sdist(name, version, dest_dir=THIRDPARTY_DIR, repos=tuple()): if not package: if TRACE_DEEP: - print(f" download_sdist: No package in {repo.index_url} for {name}=={version}") + print( + f" download_sdist: No package in {repo.index_url} for {name}=={version}") continue sdist = package.sdist if not sdist: @@ -300,7 +303,8 @@ def download_sdist(name, version, dest_dir=THIRDPARTY_DIR, repos=tuple()): continue if TRACE_DEEP: - print(f" download_sdist: Getting sdist from index (or cache): {sdist.download_url}") + print( + f" download_sdist: Getting sdist from index (or cache): {sdist.download_url}") fetched_sdist_filename = package.sdist.download(dest_dir=dest_dir) if fetched_sdist_filename: @@ -533,7 +537,8 @@ def get_best_download_url(self, repos=tuple()): repos = DEFAULT_PYPI_REPOS for repo in repos: - package = repo.get_package_version(name=self.name, version=self.version) + package = repo.get_package_version( + name=self.name, version=self.version) if not package: if TRACE: print( @@ -772,7 +777,8 @@ def load_remote_about_data(self): if notice_text: about_data["notice_text"] = notice_text except RemoteNotFetchedException: - print(f"Failed to fetch NOTICE file: {self.notice_download_url}") + print( + f"Failed to fetch NOTICE file: {self.notice_download_url}") return self.load_about_data(about_data) def get_checksums(self, dest_dir=THIRDPARTY_DIR): @@ -821,9 +827,11 @@ def fetch_license_files(self, dest_dir=THIRDPARTY_DIR, use_cached_index=False): Fetch license files if missing in `dest_dir`. Return True if license files were fetched. """ - urls = LinksRepository.from_url(use_cached_index=use_cached_index).links + urls = LinksRepository.from_url( + use_cached_index=use_cached_index).links errors = [] - extra_lic_names = [l.get("file") for l in self.extra_data.get("licenses", {})] + extra_lic_names = [l.get("file") + for l in self.extra_data.get("licenses", {})] extra_lic_names += [self.extra_data.get("license_file")] extra_lic_names = [ln for ln in extra_lic_names if ln] lic_names = [f"{key}.LICENSE" for key in self.get_license_keys()] @@ -834,7 +842,8 @@ def fetch_license_files(self, dest_dir=THIRDPARTY_DIR, use_cached_index=False): try: # try remotely first - lic_url = get_license_link_for_filename(filename=filename, urls=urls) + lic_url = get_license_link_for_filename( + filename=filename, urls=urls) fetch_and_save( path_or_url=lic_url, @@ -911,7 +920,8 @@ def load_pkginfo_data(self, dest_dir=THIRDPARTY_DIR): c for c in classifiers if c.startswith("License") ] license_expression = get_license_expression(declared_license) - other_classifiers = [c for c in classifiers if not c.startswith("License")] + other_classifiers = [ + c for c in classifiers if not c.startswith("License")] holder = raw_data["Author"] holder_contact = raw_data["Author-email"] @@ -953,7 +963,8 @@ def update(self, data, overwrite=False, keep_extra=True): package_url = data.get("package_url") if package_url: purl_from_data = packageurl.PackageURL.from_string(package_url) - purl_from_self = packageurl.PackageURL.from_string(self.package_url) + purl_from_self = packageurl.PackageURL.from_string( + self.package_url) if purl_from_data != purl_from_self: print( f"Invalid dist update attempt, no same same purl with dist: " @@ -1003,7 +1014,8 @@ def get_license_link_for_filename(filename, urls): if not path_or_url: raise Exception(f"Missing link to file: {filename}") if not len(path_or_url) == 1: - raise Exception(f"Multiple links to file: {filename}: \n" + "\n".join(path_or_url)) + raise Exception( + f"Multiple links to file: {filename}: \n" + "\n".join(path_or_url)) return path_or_url[0] @@ -1397,7 +1409,8 @@ def packages_from_dir(cls, directory): """ base = os.path.abspath(directory) - paths = [os.path.join(base, f) for f in os.listdir(base) if f.endswith(EXTENSIONS)] + paths = [os.path.join(base, f) + for f in os.listdir(base) if f.endswith(EXTENSIONS)] if TRACE_ULTRA_DEEP: print("packages_from_dir: paths:", paths) @@ -1458,7 +1471,8 @@ def dists_from_paths_or_urls(cls, paths_or_urls): dists = [] if TRACE_ULTRA_DEEP: print(" ###paths_or_urls:", paths_or_urls) - installable = [f for f in paths_or_urls if f.endswith(EXTENSIONS_INSTALLABLE)] + installable = [f for f in paths_or_urls if f.endswith( + EXTENSIONS_INSTALLABLE)] for path_or_url in installable: try: dist = Distribution.from_path_or_url(path_or_url) @@ -1476,7 +1490,8 @@ def dists_from_paths_or_urls(cls, paths_or_urls): ) except InvalidDistributionFilename: if TRACE_DEEP: - print(f" Skipping invalid distribution from: {path_or_url}") + print( + f" Skipping invalid distribution from: {path_or_url}") continue return dists @@ -1525,7 +1540,8 @@ class Environment: implementation = attr.ib( type=str, default="cp", - metadata=dict(help="Python implementation supported by this environment."), + metadata=dict( + help="Python implementation supported by this environment."), repr=False, ) @@ -1539,7 +1555,8 @@ class Environment: platforms = attr.ib( type=list, default=attr.Factory(list), - metadata=dict(help="List of platform tags supported by this environment."), + metadata=dict( + help="List of platform tags supported by this environment."), repr=False, ) @@ -1623,7 +1640,8 @@ class PypiSimpleRepository: fetched_package_normalized_names = attr.ib( type=set, default=attr.Factory(set), - metadata=dict(help="A set of already fetched package normalized names."), + metadata=dict( + help="A set of already fetched package normalized names."), ) use_cached_index = attr.ib( @@ -1654,10 +1672,12 @@ def _get_package_versions_map(self, name): self.packages[normalized_name] = versions except RemoteNotFetchedException as e: if TRACE: - print(f"failed to fetch package name: {name} from: {self.index_url}:\n{e}") + print( + f"failed to fetch package name: {name} from: {self.index_url}:\n{e}") if not versions and TRACE: - print(f"WARNING: package {name} not found in repo: {self.index_url}") + print( + f"WARNING: package {name} not found in repo: {self.index_url}") return versions @@ -1842,7 +1862,8 @@ def get(self, path_or_url, as_text=True, force=False): if force or not os.path.exists(cached): if TRACE_DEEP: print(f" FILE CACHE MISS: {path_or_url}") - content = get_file_content(path_or_url=path_or_url, as_text=as_text) + content = get_file_content( + path_or_url=path_or_url, as_text=as_text) wmode = "w" if as_text else "wb" with open(cached, wmode) as fo: fo.write(content) @@ -1864,7 +1885,8 @@ def get_file_content(path_or_url, as_text=True): if path_or_url.startswith("https://"): if TRACE_DEEP: print(f"Fetching: {path_or_url}") - _headers, content = get_remote_file_content(url=path_or_url, as_text=as_text) + _headers, content = get_remote_file_content( + url=path_or_url, as_text=as_text) return content elif path_or_url.startswith("file://") or ( @@ -1930,7 +1952,8 @@ def get_remote_file_content( ) else: - raise RemoteNotFetchedException(f"Failed HTTP request from {url} with {status}") + raise RemoteNotFetchedException( + f"Failed HTTP request from {url} with {status}") if headers_only: return response.headers, None @@ -2021,7 +2044,8 @@ def get_other_dists(_package, _dist): # if has key data we may look to improve later, but we can move on if local_dist.has_key_metadata(): local_dist.save_about_and_notice_files(dest_dir=dest_dir) - local_dist.fetch_license_files(dest_dir=dest_dir, use_cached_index=use_cached_index) + local_dist.fetch_license_files( + dest_dir=dest_dir, use_cached_index=use_cached_index) continue # lets try to get from another dist of the same local package @@ -2033,7 +2057,8 @@ def get_other_dists(_package, _dist): # if has key data we may look to improve later, but we can move on if local_dist.has_key_metadata(): local_dist.save_about_and_notice_files(dest_dir=dest_dir) - local_dist.fetch_license_files(dest_dir=dest_dir, use_cached_index=use_cached_index) + local_dist.fetch_license_files( + dest_dir=dest_dir, use_cached_index=use_cached_index) continue # try to get another version of the same package that is not our version @@ -2044,7 +2069,8 @@ def get_other_dists(_package, _dist): ] other_local_version = other_local_packages and other_local_packages[-1] if other_local_version: - latest_local_dists = list(other_local_version.get_distributions()) + latest_local_dists = list( + other_local_version.get_distributions()) for latest_local_dist in latest_local_dists: latest_local_dist.load_about_data(dest_dir=dest_dir) if not latest_local_dist.has_key_metadata(): @@ -2070,7 +2096,8 @@ def get_other_dists(_package, _dist): # if has key data we may look to improve later, but we can move on if local_dist.has_key_metadata(): local_dist.save_about_and_notice_files(dest_dir=dest_dir) - local_dist.fetch_license_files(dest_dir=dest_dir, use_cached_index=use_cached_index) + local_dist.fetch_license_files( + dest_dir=dest_dir, use_cached_index=use_cached_index) continue # try to get a latest version of the same package that is not our version @@ -2111,7 +2138,8 @@ def get_other_dists(_package, _dist): # if local_dist.has_key_metadata() or not local_dist.has_key_metadata(): local_dist.save_about_and_notice_files(dest_dir) - lic_errs = local_dist.fetch_license_files(dest_dir, use_cached_index=use_cached_index) + lic_errs = local_dist.fetch_license_files( + dest_dir, use_cached_index=use_cached_index) if not local_dist.has_key_metadata(): print(f"Unable to add essential ABOUT data for: {local_dist}") @@ -2259,7 +2287,8 @@ def find_problems( for dist in package.get_distributions(): dist.load_about_data(dest_dir=dest_dir) - abpth = os.path.abspath(os.path.join(dest_dir, dist.about_filename)) + abpth = os.path.abspath(os.path.join( + dest_dir, dist.about_filename)) if not dist.has_key_metadata(): print(f" Missing key ABOUT data in file://{abpth}") if "classifiers" in dist.extra_data: diff --git a/setup.cfg b/setup.cfg index a8e20c5d..ef7d369b 100644 --- a/setup.cfg +++ b/setup.cfg @@ -6,7 +6,7 @@ license = Apache-2.0 description = skeleton long_description = file:README.rst long_description_content_type = text/x-rst -url = https://github.com/nexB/skeleton +url = https://github.com/aboutcode-org/skeleton author = nexB. Inc. and others author_email = info@aboutcode.org diff --git a/tests/test_skeleton_codestyle.py b/tests/test_skeleton_codestyle.py index 2eb6e558..b4ce8c16 100644 --- a/tests/test_skeleton_codestyle.py +++ b/tests/test_skeleton_codestyle.py @@ -3,7 +3,7 @@ # ScanCode is a trademark of nexB Inc. # SPDX-License-Identifier: Apache-2.0 # See http://www.apache.org/licenses/LICENSE-2.0 for the license text. -# See https://github.com/nexB/skeleton for support or download. +# See https://github.com/aboutcode-org/skeleton for support or download. # See https://aboutcode.org for more information about nexB OSS projects. # From 7360047c01c43ae4d1c4deea4f8cdeb3dcebaa70 Mon Sep 17 00:00:00 2001 From: Keshav Priyadarshi Date: Tue, 3 Sep 2024 17:52:01 +0530 Subject: [PATCH 15/28] Improve PyPI test fixture Signed-off-by: Keshav Priyadarshi --- tests/data/package_versions/pypi.json | 1344 +- .../data/package_versions/pypi_mock_data.json | 19561 +++++----------- tests/test_package_versions.py | 2 +- 3 files changed, 6055 insertions(+), 14852 deletions(-) diff --git a/tests/data/package_versions/pypi.json b/tests/data/package_versions/pypi.json index 620fa2b8..259bf0e3 100644 --- a/tests/data/package_versions/pypi.json +++ b/tests/data/package_versions/pypi.json @@ -1,1362 +1,334 @@ [ { - "value": "1.1.3", - "release_date": "2010-12-23T05:14:23.509436+00:00" + "value": "0.7.32", + "release_date": "2016-02-22T06:13:06.774750+00:00" }, { - "value": "1.1.4", - "release_date": "2011-02-09T04:13:07.000075+00:00" + "value": "0.7.33", + "release_date": "2016-03-01T18:55:48.013515+00:00" }, { - "value": "1.10", - "release_date": "2016-08-01T18:32:16.280614+00:00" + "value": "0.8", + "release_date": "2015-10-29T21:14:49.182641+00:00" }, { - "value": "1.10.1", - "release_date": "2016-09-01T23:18:18.672706+00:00" + "value": "0.8.1", + "release_date": "2015-10-29T21:06:19.206843+00:00" }, { - "value": "1.10.2", - "release_date": "2016-10-01T20:05:31.330942+00:00" + "value": "0.8.10", + "release_date": "2015-10-29T21:06:38.680140+00:00" }, { - "value": "1.10.3", - "release_date": "2016-11-01T13:57:16.055061+00:00" + "value": "0.8.11", + "release_date": "2015-10-29T21:06:57.940593+00:00" }, { - "value": "1.10.4", - "release_date": "2016-12-01T23:46:50.215935+00:00" + "value": "0.8.12", + "release_date": "2015-10-29T21:07:20.546698+00:00" }, { - "value": "1.10.5", - "release_date": "2017-01-04T19:23:00.596664+00:00" + "value": "0.8.13", + "release_date": "2015-10-29T21:07:39.207404+00:00" }, { - "value": "1.10.6", - "release_date": "2017-03-01T13:37:40.243134+00:00" + "value": "0.8.14", + "release_date": "2015-10-29T21:07:58.530827+00:00" }, { - "value": "1.10.7", - "release_date": "2017-04-04T14:27:54.235551+00:00" + "value": "0.8.15", + "release_date": "2015-10-29T21:08:22.566012+00:00" }, { - "value": "1.10.8", - "release_date": "2017-09-05T15:31:58.221021+00:00" + "value": "0.8.16", + "release_date": "2015-10-29T21:08:39.373749+00:00" }, { - "value": "1.10a1", - "release_date": "2016-05-20T12:24:59.952686+00:00" + "value": "0.8.17", + "release_date": "2015-10-29T21:08:57.975915+00:00" }, { - "value": "1.10b1", - "release_date": "2016-06-22T01:15:17.267637+00:00" + "value": "0.8.18", + "release_date": "2015-10-29T21:09:20.067292+00:00" }, { - "value": "1.10rc1", - "release_date": "2016-07-18T18:05:05.503584+00:00" + "value": "0.8.19", + "release_date": "2015-10-29T21:09:38.790874+00:00" }, { - "value": "1.11", - "release_date": "2017-04-04T16:00:04.407084+00:00" + "value": "0.8.2", + "release_date": "2015-10-29T21:09:57.412044+00:00" }, { - "value": "1.11.1", - "release_date": "2017-05-06T13:26:38.108379+00:00" + "value": "0.8.20", + "release_date": "2015-10-29T21:10:39.796073+00:00" }, { - "value": "1.11.10", - "release_date": "2018-02-01T14:40:24.121117+00:00" + "value": "0.8.21", + "release_date": "2015-10-29T21:11:02.588393+00:00" }, { - "value": "1.11.11", - "release_date": "2018-03-06T14:16:18.949597+00:00" + "value": "0.8.22", + "release_date": "2015-10-29T21:11:21.857612+00:00" }, { - "value": "1.11.12", - "release_date": "2018-04-03T02:46:00.567935+00:00" + "value": "0.8.23", + "release_date": "2015-10-29T21:11:40.771833+00:00" }, { - "value": "1.11.13", - "release_date": "2018-05-02T01:54:13.881452+00:00" + "value": "0.8.24", + "release_date": "2016-02-22T06:41:51.694397+00:00" }, { - "value": "1.11.14", - "release_date": "2018-07-02T09:02:11.027707+00:00" + "value": "0.8.25", + "release_date": "2016-03-01T19:01:52.507584+00:00" }, { - "value": "1.11.15", - "release_date": "2018-08-01T13:45:14.692609+00:00" + "value": "0.8.26", + "release_date": "2016-04-13T21:12:52.594112+00:00" }, { - "value": "1.11.16", - "release_date": "2018-10-01T09:22:18.040195+00:00" + "value": "0.8.27", + "release_date": "2016-10-18T01:59:51.726624+00:00" }, { - "value": "1.11.17", - "release_date": "2018-12-03T17:03:07.563036+00:00" + "value": "0.8.28", + "release_date": "2017-02-09T23:06:06.737009+00:00" }, { - "value": "1.11.18", - "release_date": "2019-01-04T14:10:56.890031+00:00" + "value": "0.8.29", + "release_date": "2017-02-16T02:03:10.400432+00:00" }, { - "value": "1.11.2", - "release_date": "2017-06-01T16:50:13.913846+00:00" + "value": "0.8.3", + "release_date": "2015-10-29T21:12:00.978974+00:00" }, { - "value": "1.11.20", - "release_date": "2019-02-11T15:10:54.841634+00:00" + "value": "0.8.4", + "release_date": "2015-10-29T21:12:35.979475+00:00" }, { - "value": "1.11.21", - "release_date": "2019-06-03T10:11:17.443138+00:00" + "value": "0.8.5", + "release_date": "2015-10-29T21:12:52.525366+00:00" }, { - "value": "1.11.22", - "release_date": "2019-07-01T07:19:29.492277+00:00" + "value": "0.8.6", + "release_date": "2015-10-29T21:13:20.969507+00:00" }, { - "value": "1.11.23", - "release_date": "2019-08-01T09:04:43.456757+00:00" + "value": "0.8.7", + "release_date": "2015-10-29T21:13:40.939660+00:00" }, { - "value": "1.11.24", - "release_date": "2019-09-02T07:18:46.393062+00:00" + "value": "0.8.8", + "release_date": "2015-10-29T21:14:00.169901+00:00" }, { - "value": "1.11.25", - "release_date": "2019-10-01T08:36:51.238785+00:00" + "value": "0.8.9", + "release_date": "2015-10-29T21:14:42.996197+00:00" }, { - "value": "1.11.26", - "release_date": "2019-11-04T08:33:25.858722+00:00" + "value": "0.9", + "release_date": "2015-10-29T21:04:02.806960+00:00" }, { - "value": "1.11.27", - "release_date": "2019-12-18T08:59:19.797864+00:00" + "value": "0.9.1", + "release_date": "2016-02-22T06:49:11.661244+00:00" }, { - "value": "1.11.28", - "release_date": "2020-02-03T09:50:51.569042+00:00" + "value": "0.9.2", + "release_date": "2016-03-01T19:05:05.497377+00:00" }, { - "value": "1.11.29", - "release_date": "2020-03-04T09:32:02.383798+00:00" + "value": "0.9.3", + "release_date": "2016-04-13T21:14:46.224818+00:00" }, { - "value": "1.11.3", - "release_date": "2017-07-01T23:25:09.072231+00:00" + "value": "0.9.4", + "release_date": "2016-10-18T02:17:04.972690+00:00" }, { - "value": "1.11.4", - "release_date": "2017-08-01T12:25:19.627007+00:00" + "value": "0.9.5", + "release_date": "2017-02-09T23:14:25.588925+00:00" }, { - "value": "1.11.5", - "release_date": "2017-09-05T15:19:03.327101+00:00" + "value": "0.9.6", + "release_date": "2017-02-16T02:07:31.869735+00:00" }, { - "value": "1.11.6", - "release_date": "2017-10-05T18:24:42.635754+00:00" + "value": "0.9.7", + "release_date": "2017-05-26T04:49:30.085130+00:00" }, { - "value": "1.11.7", - "release_date": "2017-11-02T01:26:35.309616+00:00" + "value": "0.9.8", + "release_date": "2017-06-19T23:11:07.980310+00:00" }, { - "value": "1.11.8", - "release_date": "2017-12-02T14:21:08.831473+00:00" + "value": "0.9.9", + "release_date": "2017-08-01T19:55:53.085622+00:00" }, { - "value": "1.11.9", - "release_date": "2018-01-02T01:02:08.883938+00:00" + "value": "1.0", + "release_date": "2017-11-21T07:29:52.294894+00:00" }, { - "value": "1.11a1", - "release_date": "2017-01-18T01:01:35.823808+00:00" + "value": "1.0.1", + "release_date": "2017-12-05T23:40:07.861672+00:00" }, { - "value": "1.11b1", - "release_date": "2017-02-20T23:21:50.835517+00:00" + "value": "1.0.10", + "release_date": "2019-02-19T10:36:04.556612+00:00" }, { - "value": "1.11rc1", - "release_date": "2017-03-21T22:56:03.250177+00:00" + "value": "1.0.11", + "release_date": "2019-03-19T07:53:06.333147+00:00" }, { - "value": "1.2", - "release_date": "2010-05-17T20:04:28.174094+00:00" + "value": "1.0.12", + "release_date": "2019-06-12T05:21:43.341447+00:00" }, { - "value": "1.2.1", - "release_date": "2010-05-24T21:19:28.440044+00:00" + "value": "1.0.13", + "release_date": "2020-06-30T06:35:47.512638+00:00" }, { - "value": "1.2.2", - "release_date": "2010-09-09T02:41:25.682749+00:00" + "value": "1.0.14", + "release_date": "2020-12-10T09:50:55.012440+00:00" }, { - "value": "1.2.3", - "release_date": "2010-09-11T08:50:39.445072+00:00" + "value": "1.0.15", + "release_date": "2021-01-20T09:11:15.176142+00:00" }, { - "value": "1.2.4", - "release_date": "2010-12-23T05:15:19.624230+00:00" + "value": "1.0.16", + "release_date": "2021-02-12T07:22:49.844813+00:00" }, { - "value": "1.2.5", - "release_date": "2011-02-09T04:08:37.395797+00:00" + "value": "1.0.17", + "release_date": "2021-04-14T02:11:20.188424+00:00" }, { - "value": "1.2.6", - "release_date": "2011-09-10T03:42:09.759203+00:00" + "value": "1.0.18", + "release_date": "2021-06-15T07:17:02.882111+00:00" }, { - "value": "1.2.7", - "release_date": "2011-09-11T03:05:18.996286+00:00" + "value": "1.0.2", + "release_date": "2018-01-09T08:40:36.936214+00:00" }, { - "value": "1.3", - "release_date": "2011-03-23T06:09:12.606484+00:00" + "value": "1.0.3", + "release_date": "2018-02-06T07:52:11.555307+00:00" }, { - "value": "1.3.1", - "release_date": "2011-09-10T03:36:21.376323+00:00" + "value": "1.0.4", + "release_date": "2018-03-20T06:40:21.420724+00:00" }, { - "value": "1.3.2", - "release_date": "2012-07-30T23:02:55.545349+00:00" + "value": "1.0.5", + "release_date": "2018-03-23T07:12:22.151189+00:00" }, { - "value": "1.3.3", - "release_date": "2012-08-01T22:08:20.092282+00:00" + "value": "1.0.6", + "release_date": "2018-06-05T07:54:08.985947+00:00" }, { - "value": "1.3.4", - "release_date": "2013-03-05T22:33:47.875562+00:00" + "value": "1.0.7", + "release_date": "2018-10-12T07:40:17.609873+00:00" }, { - "value": "1.3.5", - "release_date": "2012-12-10T21:39:30.210465+00:00" + "value": "1.0.8", + "release_date": "2018-11-06T11:05:03.005040+00:00" }, { - "value": "1.3.6", - "release_date": "2013-02-19T20:32:04.436478+00:00" + "value": "1.0.9", + "release_date": "2019-01-29T09:23:50.140504+00:00" }, { - "value": "1.3.7", - "release_date": "2013-02-20T20:03:48.025529+00:00" - }, - { - "value": "1.4", - "release_date": "2012-03-23T18:00:23.693465+00:00" - }, - { - "value": "1.4.1", - "release_date": "2012-07-30T22:48:27.374116+00:00" - }, - { - "value": "1.4.10", - "release_date": "2013-11-06T14:21:25.765558+00:00" - }, - { - "value": "1.4.11", - "release_date": "2014-04-21T22:40:17.318319+00:00" - }, - { - "value": "1.4.12", - "release_date": "2014-04-28T20:30:21.125270+00:00" - }, - { - "value": "1.4.13", - "release_date": "2014-05-14T18:27:42.778726+00:00" - }, - { - "value": "1.4.14", - "release_date": "2014-08-20T20:01:35.076618+00:00" - }, - { - "value": "1.4.15", - "release_date": "2014-09-02T20:44:06.366428+00:00" - }, - { - "value": "1.4.16", - "release_date": "2014-10-22T16:37:17.950676+00:00" - }, - { - "value": "1.4.17", - "release_date": "2015-01-03T02:20:41.434868+00:00" - }, - { - "value": "1.4.18", - "release_date": "2015-01-13T18:54:01.979098+00:00" - }, - { - "value": "1.4.19", - "release_date": "2015-01-27T17:11:15.646882+00:00" - }, - { - "value": "1.4.2", - "release_date": "2012-10-17T22:18:35.480417+00:00" - }, - { - "value": "1.4.20", - "release_date": "2015-03-19T00:03:58.032976+00:00" - }, - { - "value": "1.4.21", - "release_date": "2015-07-08T19:56:26.333839+00:00" - }, - { - "value": "1.4.22", - "release_date": "2015-08-18T17:22:09.242532+00:00" - }, - { - "value": "1.4.3", - "release_date": "2012-12-10T21:46:28.825133+00:00" - }, - { - "value": "1.4.4", - "release_date": "2013-02-19T20:27:55.134186+00:00" - }, - { - "value": "1.4.5", - "release_date": "2013-02-20T19:54:40.773224+00:00" - }, - { - "value": "1.4.6", - "release_date": "2013-08-13T16:52:54.160398+00:00" - }, - { - "value": "1.4.7", - "release_date": "2013-09-11T01:18:42.411587+00:00" - }, - { - "value": "1.4.8", - "release_date": "2013-09-15T06:22:11.681231+00:00" - }, - { - "value": "1.4.9", - "release_date": "2013-10-25T04:38:13.629577+00:00" - }, - { - "value": "1.5", - "release_date": "2013-02-26T19:30:37.100371+00:00" - }, - { - "value": "1.5.1", - "release_date": "2013-03-28T20:57:18.760489+00:00" - }, - { - "value": "1.5.10", - "release_date": "2014-09-02T20:51:16.978621+00:00" - }, - { - "value": "1.5.11", - "release_date": "2014-10-22T16:45:00.838415+00:00" - }, - { - "value": "1.5.12", - "release_date": "2015-01-03T02:09:17.471850+00:00" - }, - { - "value": "1.5.2", - "release_date": "2013-08-13T16:54:03.787628+00:00" - }, - { - "value": "1.5.3", - "release_date": "2013-09-11T01:26:50.200599+00:00" - }, - { - "value": "1.5.4", - "release_date": "2013-09-15T06:30:37.726078+00:00" - }, - { - "value": "1.5.5", - "release_date": "2013-10-25T04:32:41.565014+00:00" - }, - { - "value": "1.5.6", - "release_date": "2014-04-21T22:53:10.449788+00:00" - }, - { - "value": "1.5.7", - "release_date": "2014-04-28T20:35:50.148784+00:00" - }, - { - "value": "1.5.8", - "release_date": "2014-05-14T18:36:01.585251+00:00" - }, - { - "value": "1.5.9", - "release_date": "2014-08-20T20:10:42.568983+00:00" - }, - { - "value": "1.6", - "release_date": "2013-11-06T15:01:29.487525+00:00" - }, - { - "value": "1.6.1", - "release_date": "2013-12-12T20:04:35.467654+00:00" - }, - { - "value": "1.6.10", - "release_date": "2015-01-13T18:48:53.435836+00:00" - }, - { - "value": "1.6.11", - "release_date": "2015-03-18T23:58:02.690545+00:00" - }, - { - "value": "1.6.2", - "release_date": "2014-02-06T21:51:23.493703+00:00" - }, - { - "value": "1.6.3", - "release_date": "2014-04-21T23:12:25.701466+00:00" - }, - { - "value": "1.6.4", - "release_date": "2014-04-28T20:40:52.661964+00:00" - }, - { - "value": "1.6.5", - "release_date": "2014-05-14T18:34:11.991240+00:00" - }, - { - "value": "1.6.6", - "release_date": "2014-08-20T20:17:23.703605+00:00" - }, - { - "value": "1.6.7", - "release_date": "2014-09-02T20:56:17.794128+00:00" - }, - { - "value": "1.6.8", - "release_date": "2014-10-22T16:50:38.114205+00:00" - }, - { - "value": "1.6.9", - "release_date": "2015-01-03T01:52:35.812923+00:00" - }, - { - "value": "1.7", - "release_date": "2014-09-02T21:10:05.229716+00:00" - }, - { - "value": "1.7.1", - "release_date": "2014-10-22T16:57:16.294800+00:00" - }, - { - "value": "1.7.10", - "release_date": "2015-08-18T17:15:47.478514+00:00" - }, - { - "value": "1.7.11", - "release_date": "2015-11-24T17:20:22.836539+00:00" - }, - { - "value": "1.7.2", - "release_date": "2015-01-03T01:37:40.931378+00:00" - }, - { - "value": "1.7.3", - "release_date": "2015-01-13T18:39:56.760858+00:00" - }, - { - "value": "1.7.4", - "release_date": "2015-01-27T17:22:47.486790+00:00" - }, - { - "value": "1.7.5", - "release_date": "2015-02-25T13:58:38.544144+00:00" - }, - { - "value": "1.7.6", - "release_date": "2015-03-09T15:30:50.859703+00:00" - }, - { - "value": "1.7.7", - "release_date": "2015-03-18T23:49:18.866382+00:00" - }, - { - "value": "1.7.8", - "release_date": "2015-05-01T20:42:55.907784+00:00" - }, - { - "value": "1.7.9", - "release_date": "2015-07-08T21:33:04.180643+00:00" - }, - { - "value": "1.8", - "release_date": "2015-04-01T20:12:48.080635+00:00" - }, - { - "value": "1.8.1", - "release_date": "2015-05-01T20:36:45.639030+00:00" - }, - { - "value": "1.8.10", - "release_date": "2016-03-01T17:10:34.083677+00:00" - }, - { - "value": "1.8.11", - "release_date": "2016-03-05T18:36:25.916378+00:00" - }, - { - "value": "1.8.12", - "release_date": "2016-04-01T17:54:40.693793+00:00" - }, - { - "value": "1.8.13", - "release_date": "2016-05-02T22:51:42.144726+00:00" - }, - { - "value": "1.8.14", - "release_date": "2016-07-18T18:38:36.642604+00:00" - }, - { - "value": "1.8.15", - "release_date": "2016-09-26T18:30:30.224525+00:00" - }, - { - "value": "1.8.16", - "release_date": "2016-11-01T14:09:38.084055+00:00" - }, - { - "value": "1.8.17", - "release_date": "2016-12-01T23:03:42.965764+00:00" - }, - { - "value": "1.8.18", - "release_date": "2017-04-04T14:07:49.913438+00:00" - }, - { - "value": "1.8.19", - "release_date": "2018-03-06T14:23:15.360659+00:00" - }, - { - "value": "1.8.2", - "release_date": "2015-05-20T18:02:30.286562+00:00" - }, - { - "value": "1.8.3", - "release_date": "2015-07-08T19:43:35.397608+00:00" - }, - { - "value": "1.8.4", - "release_date": "2015-08-18T17:06:55.158421+00:00" - }, - { - "value": "1.8.5", - "release_date": "2015-10-04T00:06:18.616258+00:00" - }, - { - "value": "1.8.6", - "release_date": "2015-11-04T17:03:52.915842+00:00" - }, - { - "value": "1.8.7", - "release_date": "2015-11-24T17:29:07.982583+00:00" - }, - { - "value": "1.8.8", - "release_date": "2016-01-02T14:28:34.130412+00:00" - }, - { - "value": "1.8.9", - "release_date": "2016-02-01T17:24:29.386552+00:00" - }, - { - "value": "1.8a1", - "release_date": "2015-01-16T22:25:13.083005+00:00" - }, - { - "value": "1.8b1", - "release_date": "2015-02-25T13:42:42.788820+00:00" - }, - { - "value": "1.8b2", - "release_date": "2015-03-09T15:55:16.864430+00:00" - }, - { - "value": "1.8c1", - "release_date": "2015-03-18T23:39:34.451697+00:00" - }, - { - "value": "1.9", - "release_date": "2015-12-01T23:55:55.779479+00:00" - }, - { - "value": "1.9.1", - "release_date": "2016-01-02T13:50:42.992870+00:00" - }, - { - "value": "1.9.10", - "release_date": "2016-09-26T18:34:45.917564+00:00" - }, - { - "value": "1.9.11", - "release_date": "2016-11-01T14:02:27.855507+00:00" - }, - { - "value": "1.9.12", - "release_date": "2016-12-01T23:16:35.055179+00:00" - }, - { - "value": "1.9.13", - "release_date": "2017-04-04T14:15:10.543830+00:00" - }, - { - "value": "1.9.2", - "release_date": "2016-02-01T17:17:34.977864+00:00" - }, - { - "value": "1.9.3", - "release_date": "2016-03-01T17:01:04.656339+00:00" - }, - { - "value": "1.9.4", - "release_date": "2016-03-05T14:32:11.209251+00:00" - }, - { - "value": "1.9.5", - "release_date": "2016-04-01T17:47:24.975699+00:00" - }, - { - "value": "1.9.6", - "release_date": "2016-05-02T22:33:50.459797+00:00" - }, - { - "value": "1.9.7", - "release_date": "2016-06-04T23:44:08.110124+00:00" - }, - { - "value": "1.9.8", - "release_date": "2016-07-18T18:20:13.055798+00:00" - }, - { - "value": "1.9.9", - "release_date": "2016-08-01T18:10:24.753259+00:00" - }, - { - "value": "1.9a1", - "release_date": "2015-09-24T00:20:01.735219+00:00" - }, - { - "value": "1.9b1", - "release_date": "2015-10-20T01:17:13.619306+00:00" - }, - { - "value": "1.9rc1", - "release_date": "2015-11-16T21:10:10.276918+00:00" - }, - { - "value": "1.9rc2", - "release_date": "2015-11-24T17:35:35.276135+00:00" - }, - { - "value": "2.0", - "release_date": "2017-12-02T15:12:05.947934+00:00" - }, - { - "value": "2.0.1", - "release_date": "2018-01-02T00:51:04.317791+00:00" - }, - { - "value": "2.0.10", - "release_date": "2019-01-04T14:03:19.511618+00:00" - }, - { - "value": "2.0.12", - "release_date": "2019-02-11T15:11:02.427442+00:00" - }, - { - "value": "2.0.13", - "release_date": "2019-02-12T10:50:09.191763+00:00" - }, - { - "value": "2.0.2", - "release_date": "2018-02-01T14:30:22.407202+00:00" - }, - { - "value": "2.0.3", - "release_date": "2018-03-06T14:06:37.383785+00:00" - }, - { - "value": "2.0.4", - "release_date": "2018-04-03T02:39:41.734614+00:00" - }, - { - "value": "2.0.5", - "release_date": "2018-05-02T01:34:52.880781+00:00" - }, - { - "value": "2.0.6", - "release_date": "2018-06-01T15:32:11.407175+00:00" - }, - { - "value": "2.0.7", - "release_date": "2018-07-02T09:02:19.534930+00:00" - }, - { - "value": "2.0.8", - "release_date": "2018-08-01T13:52:08.556325+00:00" - }, - { - "value": "2.0.9", - "release_date": "2018-10-01T09:22:23.082215+00:00" - }, - { - "value": "2.0a1", - "release_date": "2017-09-22T18:09:22.828675+00:00" - }, - { - "value": "2.0b1", - "release_date": "2017-10-17T02:00:54.980990+00:00" - }, - { - "value": "2.0rc1", - "release_date": "2017-11-15T23:51:54.051974+00:00" - }, - { - "value": "2.1", - "release_date": "2018-08-01T14:11:34.715690+00:00" - }, - { - "value": "2.1.1", - "release_date": "2018-08-31T08:42:20.929702+00:00" - }, - { - "value": "2.1.10", - "release_date": "2019-07-01T07:19:36.568033+00:00" - }, - { - "value": "2.1.11", - "release_date": "2019-08-01T09:04:51.690400+00:00" - }, - { - "value": "2.1.12", - "release_date": "2019-09-02T07:18:54.146400+00:00" - }, - { - "value": "2.1.13", - "release_date": "2019-10-01T08:36:59.308397+00:00" - }, - { - "value": "2.1.14", - "release_date": "2019-11-04T08:33:33.296744+00:00" - }, - { - "value": "2.1.15", - "release_date": "2019-12-02T08:57:57.603087+00:00" - }, - { - "value": "2.1.2", - "release_date": "2018-10-01T09:22:28.985017+00:00" - }, - { - "value": "2.1.3", - "release_date": "2018-11-01T14:36:54.164080+00:00" - }, - { - "value": "2.1.4", - "release_date": "2018-12-03T17:03:23.910206+00:00" - }, - { - "value": "2.1.5", - "release_date": "2019-01-04T13:53:03.556241+00:00" - }, - { - "value": "2.1.7", - "release_date": "2019-02-11T15:11:10.020675+00:00" - }, - { - "value": "2.1.8", - "release_date": "2019-04-01T09:19:04.186926+00:00" - }, - { - "value": "2.1.9", - "release_date": "2019-06-03T10:11:23.821054+00:00" - }, - { - "value": "2.1a1", - "release_date": "2018-05-18T01:01:19.030204+00:00" + "value": "2.0", + "release_date": "2020-12-17T04:59:25.614106+00:00" }, { - "value": "2.1b1", - "release_date": "2018-06-18T23:55:36.588389+00:00" + "value": "2.1", + "release_date": "2021-01-20T09:53:55.441808+00:00" }, { - "value": "2.1rc1", - "release_date": "2018-07-18T17:35:02.726810+00:00" + "value": "2.1.1", + "release_date": "2021-01-20T10:50:04.217950+00:00" }, { "value": "2.2", - "release_date": "2019-04-01T12:47:42.240453+00:00" + "release_date": "2021-03-02T09:27:28.223673+00:00" }, { "value": "2.2.1", - "release_date": "2019-05-01T06:57:47.651991+00:00" - }, - { - "value": "2.2.10", - "release_date": "2020-02-03T09:50:57.805388+00:00" - }, - { - "value": "2.2.11", - "release_date": "2020-03-04T09:32:08.970734+00:00" - }, - { - "value": "2.2.12", - "release_date": "2020-04-01T07:59:21.455385+00:00" - }, - { - "value": "2.2.13", - "release_date": "2020-06-03T09:36:39.587482+00:00" - }, - { - "value": "2.2.14", - "release_date": "2020-07-01T04:50:13.815347+00:00" - }, - { - "value": "2.2.15", - "release_date": "2020-08-03T07:23:37.474778+00:00" - }, - { - "value": "2.2.16", - "release_date": "2020-09-01T09:14:36.209966+00:00" - }, - { - "value": "2.2.17", - "release_date": "2020-11-02T08:12:44.286790+00:00" - }, - { - "value": "2.2.18", - "release_date": "2021-02-01T09:28:29.173011+00:00" - }, - { - "value": "2.2.19", - "release_date": "2021-02-19T09:08:15.073175+00:00" + "release_date": "2021-05-20T22:40:46.954786+00:00" }, { "value": "2.2.2", - "release_date": "2019-06-03T10:11:32.109900+00:00" - }, - { - "value": "2.2.20", - "release_date": "2021-04-06T07:35:02.866945+00:00" - }, - { - "value": "2.2.21", - "release_date": "2021-05-04T08:47:51.750991+00:00" - }, - { - "value": "2.2.22", - "release_date": "2021-05-06T07:40:41.663188+00:00" - }, - { - "value": "2.2.23", - "release_date": "2021-05-13T07:36:51.046462+00:00" - }, - { - "value": "2.2.24", - "release_date": "2021-06-02T08:54:12.637481+00:00" - }, - { - "value": "2.2.25", - "release_date": "2021-12-07T07:34:54.595264+00:00" - }, - { - "value": "2.2.26", - "release_date": "2022-01-04T09:53:29.324996+00:00" - }, - { - "value": "2.2.27", - "release_date": "2022-02-01T07:56:27.211309+00:00" - }, - { - "value": "2.2.28", - "release_date": "2022-04-11T07:53:05.463179+00:00" + "release_date": "2021-06-15T07:22:30.602407+00:00" }, { "value": "2.2.3", - "release_date": "2019-07-01T07:19:43.693346+00:00" - }, - { - "value": "2.2.4", - "release_date": "2019-08-01T09:05:00.463982+00:00" + "release_date": "2021-06-29T06:39:50.246408+00:00" }, { - "value": "2.2.5", - "release_date": "2019-09-02T07:19:02.248056+00:00" + "value": "2.3", + "release_date": "2022-02-08T11:23:56.597596+00:00" }, { - "value": "2.2.6", - "release_date": "2019-10-01T08:37:07.504139+00:00" + "value": "2.3.1", + "release_date": "2022-06-14T06:32:56.736553+00:00" }, { - "value": "2.2.7", - "release_date": "2019-11-04T08:33:42.021959+00:00" + "value": "2.3.2", + "release_date": "2022-07-27T04:51:45.416890+00:00" }, { - "value": "2.2.8", - "release_date": "2019-12-02T08:58:05.035772+00:00" + "value": "2.3.3", + "release_date": "2022-07-27T05:09:03.986835+00:00" }, { - "value": "2.2.9", - "release_date": "2019-12-18T08:59:27.308468+00:00" - }, - { - "value": "2.2a1", - "release_date": "2019-01-17T15:35:59.171725+00:00" - }, - { - "value": "2.2b1", - "release_date": "2019-02-11T10:34:19.912823+00:00" - }, - { - "value": "2.2rc1", - "release_date": "2019-03-18T08:57:35.243662+00:00" + "value": "2.3.4", + "release_date": "2022-09-29T07:10:49.861732+00:00" }, { "value": "3.0", - "release_date": "2019-12-02T11:13:17.709458+00:00" - }, - { - "value": "3.0.1", - "release_date": "2019-12-18T08:59:35.114312+00:00" - }, - { - "value": "3.0.10", - "release_date": "2020-09-01T09:14:40.538593+00:00" - }, - { - "value": "3.0.11", - "release_date": "2020-11-02T08:12:49.355431+00:00" - }, - { - "value": "3.0.12", - "release_date": "2021-02-01T09:28:35.717844+00:00" - }, - { - "value": "3.0.13", - "release_date": "2021-02-19T09:08:22.462879+00:00" - }, - { - "value": "3.0.14", - "release_date": "2021-04-06T07:35:07.952242+00:00" - }, - { - "value": "3.0.2", - "release_date": "2020-01-02T07:22:21.750853+00:00" - }, - { - "value": "3.0.3", - "release_date": "2020-02-03T09:51:04.774633+00:00" - }, - { - "value": "3.0.4", - "release_date": "2020-03-04T09:32:15.777733+00:00" - }, - { - "value": "3.0.5", - "release_date": "2020-04-01T07:59:27.770337+00:00" - }, - { - "value": "3.0.6", - "release_date": "2020-05-04T05:26:41.532361+00:00" - }, - { - "value": "3.0.7", - "release_date": "2020-06-03T09:36:44.234306+00:00" - }, - { - "value": "3.0.8", - "release_date": "2020-07-01T04:50:38.186817+00:00" - }, - { - "value": "3.0.9", - "release_date": "2020-08-03T07:23:44.286778+00:00" - }, - { - "value": "3.0a1", - "release_date": "2019-09-10T09:19:39.974331+00:00" - }, - { - "value": "3.0b1", - "release_date": "2019-10-14T10:21:46.008282+00:00" - }, - { - "value": "3.0rc1", - "release_date": "2019-11-18T08:51:19.037478+00:00" + "release_date": "2022-10-04T21:44:10.883330+00:00" }, { "value": "3.1", - "release_date": "2020-08-04T08:07:06.894968+00:00" - }, - { - "value": "3.1.1", - "release_date": "2020-09-01T09:14:45.550781+00:00" - }, - { - "value": "3.1.10", - "release_date": "2021-05-06T07:40:18.974792+00:00" - }, - { - "value": "3.1.11", - "release_date": "2021-05-13T07:36:56.726950+00:00" - }, - { - "value": "3.1.12", - "release_date": "2021-06-02T08:54:32.877136+00:00" - }, - { - "value": "3.1.13", - "release_date": "2021-07-01T07:40:04.542907+00:00" - }, - { - "value": "3.1.14", - "release_date": "2021-12-07T07:35:00.760884+00:00" - }, - { - "value": "3.1.2", - "release_date": "2020-10-01T05:38:34.306154+00:00" - }, - { - "value": "3.1.3", - "release_date": "2020-11-02T08:12:54.518784+00:00" - }, - { - "value": "3.1.4", - "release_date": "2020-12-01T06:03:38.760397+00:00" - }, - { - "value": "3.1.5", - "release_date": "2021-01-04T07:54:40.039501+00:00" - }, - { - "value": "3.1.6", - "release_date": "2021-02-01T09:28:42.846948+00:00" - }, - { - "value": "3.1.7", - "release_date": "2021-02-19T09:08:29.394194+00:00" - }, - { - "value": "3.1.8", - "release_date": "2021-04-06T07:35:13.000931+00:00" - }, - { - "value": "3.1.9", - "release_date": "2021-05-04T08:48:12.197387+00:00" - }, - { - "value": "3.1a1", - "release_date": "2020-05-14T09:41:11.988119+00:00" - }, - { - "value": "3.1b1", - "release_date": "2020-06-15T08:15:33.357243+00:00" - }, - { - "value": "3.1rc1", - "release_date": "2020-07-20T06:38:29.308366+00:00" + "release_date": "2022-11-02T06:37:00.468916+00:00" }, { "value": "3.2", - "release_date": "2021-04-06T09:33:26.902785+00:00" - }, - { - "value": "3.2.1", - "release_date": "2021-05-04T08:48:26.422878+00:00" - }, - { - "value": "3.2.10", - "release_date": "2021-12-07T07:35:06.857668+00:00" - }, - { - "value": "3.2.11", - "release_date": "2022-01-04T09:53:34.580658+00:00" + "release_date": "2023-02-07T07:14:55.780452+00:00" }, { - "value": "3.2.12", - "release_date": "2022-02-01T07:56:32.000097+00:00" - }, - { - "value": "3.2.13", - "release_date": "2022-04-11T07:53:10.596399+00:00" - }, - { - "value": "3.2.14", - "release_date": "2022-07-04T07:57:28.551095+00:00" - }, - { - "value": "3.2.15", - "release_date": "2022-08-03T07:38:21.696991+00:00" - }, - { - "value": "3.2.16", - "release_date": "2022-10-04T07:54:27.079681+00:00" - }, - { - "value": "3.2.17", - "release_date": "2023-02-01T09:56:01.031856+00:00" - }, - { - "value": "3.2.18", - "release_date": "2023-02-14T08:25:40.252336+00:00" - }, - { - "value": "3.2.19", - "release_date": "2023-05-03T12:58:31.141156+00:00" - }, - { - "value": "3.2.2", - "release_date": "2021-05-06T07:40:03.835147+00:00" - }, - { - "value": "3.2.20", - "release_date": "2023-07-03T07:57:23.346933+00:00" - }, - { - "value": "3.2.21", - "release_date": "2023-09-04T10:58:25.702642+00:00" - }, - { - "value": "3.2.22", - "release_date": "2023-10-04T15:00:00.548503+00:00" - }, - { - "value": "3.2.23", - "release_date": "2023-11-01T06:59:19.806570+00:00" - }, - { - "value": "3.2.3", - "release_date": "2021-05-13T07:37:01.485953+00:00" - }, - { - "value": "3.2.4", - "release_date": "2021-06-02T08:54:46.382181+00:00" - }, - { - "value": "3.2.5", - "release_date": "2021-07-01T07:40:10.126225+00:00" - }, - { - "value": "3.2.6", - "release_date": "2021-08-02T06:28:42.678399+00:00" - }, - { - "value": "3.2.7", - "release_date": "2021-09-01T05:57:20.280283+00:00" - }, - { - "value": "3.2.8", - "release_date": "2021-10-05T07:46:25.737120+00:00" - }, - { - "value": "3.2.9", - "release_date": "2021-11-01T09:32:25.457689+00:00" - }, - { - "value": "3.2a1", - "release_date": "2021-01-19T13:04:37.298129+00:00" - }, - { - "value": "3.2b1", - "release_date": "2021-02-19T09:35:43.063787+00:00" - }, - { - "value": "3.2rc1", - "release_date": "2021-03-18T13:56:15.907546+00:00" + "value": "3.3", + "release_date": "2023-06-20T05:03:33.610935+00:00" }, { "value": "4.0", - "release_date": "2021-12-07T09:20:02.897592+00:00" - }, - { - "value": "4.0.1", - "release_date": "2022-01-04T09:53:39.880362+00:00" - }, - { - "value": "4.0.10", - "release_date": "2023-02-14T08:25:45.089209+00:00" - }, - { - "value": "4.0.2", - "release_date": "2022-02-01T07:56:37.646212+00:00" - }, - { - "value": "4.0.3", - "release_date": "2022-03-01T08:47:28.425336+00:00" - }, - { - "value": "4.0.4", - "release_date": "2022-04-11T07:53:16.406304+00:00" - }, - { - "value": "4.0.5", - "release_date": "2022-06-01T12:22:18.618899+00:00" - }, - { - "value": "4.0.6", - "release_date": "2022-07-04T07:57:34.578441+00:00" - }, - { - "value": "4.0.7", - "release_date": "2022-08-03T07:38:26.373780+00:00" - }, - { - "value": "4.0.8", - "release_date": "2022-10-04T07:54:33.062795+00:00" - }, - { - "value": "4.0.9", - "release_date": "2023-02-01T09:56:06.045214+00:00" + "release_date": "2023-10-17T07:55:37.827462+00:00" }, { - "value": "4.0a1", - "release_date": "2021-09-21T19:09:14.295332+00:00" - }, - { - "value": "4.0b1", - "release_date": "2021-10-25T09:23:22.644895+00:00" - }, - { - "value": "4.0rc1", - "release_date": "2021-11-22T06:38:01.843947+00:00" - }, - { - "value": "4.1", - "release_date": "2022-08-03T08:40:25.070462+00:00" - }, - { - "value": "4.1.1", - "release_date": "2022-09-05T05:02:34.094711+00:00" - }, - { - "value": "4.1.10", - "release_date": "2023-07-03T07:57:30.301266+00:00" - }, - { - "value": "4.1.11", - "release_date": "2023-09-04T10:58:30.124274+00:00" - }, - { - "value": "4.1.12", - "release_date": "2023-10-04T14:59:09.851212+00:00" - }, - { - "value": "4.1.13", - "release_date": "2023-11-01T06:59:24.955376+00:00" - }, - { - "value": "4.1.2", - "release_date": "2022-10-04T07:54:38.403977+00:00" - }, - { - "value": "4.1.3", - "release_date": "2022-11-01T06:18:21.116130+00:00" - }, - { - "value": "4.1.4", - "release_date": "2022-12-06T09:16:52.386734+00:00" - }, - { - "value": "4.1.5", - "release_date": "2023-01-02T07:34:49.562776+00:00" - }, - { - "value": "4.1.6", - "release_date": "2023-02-01T09:56:11.799732+00:00" - }, - { - "value": "4.1.7", - "release_date": "2023-02-14T08:25:50.105773+00:00" - }, - { - "value": "4.1.8", - "release_date": "2023-04-05T06:11:09.369362+00:00" - }, - { - "value": "4.1.9", - "release_date": "2023-05-03T12:58:36.244311+00:00" - }, - { - "value": "4.1a1", - "release_date": "2022-05-18T05:54:38.575881+00:00" - }, - { - "value": "4.1b1", - "release_date": "2022-06-21T09:20:06.847874+00:00" - }, - { - "value": "4.1rc1", - "release_date": "2022-07-19T09:02:07.093043+00:00" - }, - { - "value": "4.2", - "release_date": "2023-04-03T08:36:16.829178+00:00" - }, - { - "value": "4.2.1", - "release_date": "2023-05-03T12:58:41.313440+00:00" - }, - { - "value": "4.2.2", - "release_date": "2023-06-05T14:09:38.470129+00:00" - }, - { - "value": "4.2.3", - "release_date": "2023-07-03T07:57:37.448508+00:00" - }, - { - "value": "4.2.4", - "release_date": "2023-08-01T17:30:23.968800+00:00" - }, - { - "value": "4.2.5", - "release_date": "2023-09-04T10:58:34.288156+00:00" - }, - { - "value": "4.2.6", - "release_date": "2023-10-04T14:58:41.808770+00:00" - }, - { - "value": "4.2.7", - "release_date": "2023-11-01T06:59:30.228988+00:00" - }, - { - "value": "4.2.8", - "release_date": "2023-12-04T08:34:50.201677+00:00" - }, - { - "value": "4.2a1", - "release_date": "2023-01-17T09:39:25.092445+00:00" - }, - { - "value": "4.2b1", - "release_date": "2023-02-20T08:49:47.262950+00:00" - }, - { - "value": "4.2rc1", - "release_date": "2023-03-20T07:32:00.502267+00:00" + "value": "5.0", + "release_date": "2024-06-06T08:38:33.983937+00:00" }, { - "value": "5.0", - "release_date": "2023-12-04T13:12:50.251293+00:00" + "value": "5.0.1", + "release_date": "2024-06-07T18:44:31.459744+00:00" }, { - "value": "5.0a1", - "release_date": "2023-09-18T22:48:42.066135+00:00" + "value": "5.0.2", + "release_date": "2024-06-12T07:07:06.679773+00:00" }, { - "value": "5.0b1", - "release_date": "2023-10-23T18:57:57.673233+00:00" + "value": "5.1", + "release_date": "2024-07-02T06:08:58.290996+00:00" }, { - "value": "5.0rc1", - "release_date": "2023-11-20T12:40:32.210081+00:00" + "value": "5.1.1", + "release_date": "2024-08-20T04:56:19.606276+00:00" } ] \ No newline at end of file diff --git a/tests/data/package_versions/pypi_mock_data.json b/tests/data/package_versions/pypi_mock_data.json index 96b56d6c..945fdd66 100644 --- a/tests/data/package_versions/pypi_mock_data.json +++ b/tests/data/package_versions/pypi_mock_data.json @@ -1,13673 +1,5904 @@ { "info": { - "author": "Django Software Foundation", - "author_email": "foundation@djangoproject.com", - "bugtrack_url": null, - "classifiers": [ - "Development Status :: 5 - Production/Stable", - "Environment :: Web Environment", - "Framework :: Django", - "Intended Audience :: Developers", - "License :: OSI Approved :: BSD License", - "Operating System :: OS Independent", - "Programming Language :: Python", - "Programming Language :: Python :: 3", - "Programming Language :: Python :: 3 :: Only", - "Programming Language :: Python :: 3.10", - "Programming Language :: Python :: 3.11", - "Programming Language :: Python :: 3.12", - "Topic :: Internet :: WWW/HTTP", - "Topic :: Internet :: WWW/HTTP :: Dynamic Content", - "Topic :: Internet :: WWW/HTTP :: WSGI", - "Topic :: Software Development :: Libraries :: Application Frameworks", - "Topic :: Software Development :: Libraries :: Python Modules" - ], - "description": "======\nDjango\n======\n\nDjango is a high-level Python web framework that encourages rapid development\nand clean, pragmatic design. Thanks for checking it out.\n\nAll documentation is in the \"``docs``\" directory and online at\nhttps://docs.djangoproject.com/en/stable/. If you're just getting started,\nhere's how we recommend you read the docs:\n\n* First, read ``docs/intro/install.txt`` for instructions on installing Django.\n\n* Next, work through the tutorials in order (``docs/intro/tutorial01.txt``,\n ``docs/intro/tutorial02.txt``, etc.).\n\n* If you want to set up an actual deployment server, read\n ``docs/howto/deployment/index.txt`` for instructions.\n\n* You'll probably want to read through the topical guides (in ``docs/topics``)\n next; from there you can jump to the HOWTOs (in ``docs/howto``) for specific\n problems, and check out the reference (``docs/ref``) for gory details.\n\n* See ``docs/README`` for instructions on building an HTML version of the docs.\n\nDocs are updated rigorously. If you find any problems in the docs, or think\nthey should be clarified in any way, please take 30 seconds to fill out a\nticket here: https://code.djangoproject.com/newticket\n\nTo get more help:\n\n* Join the ``#django`` channel on ``irc.libera.chat``. Lots of helpful people\n hang out there. `Webchat is available `_.\n\n* Join the django-users mailing list, or read the archives, at\n https://groups.google.com/group/django-users.\n\n* Join the `Django Discord community `_.\n\n* Join the community on the `Django Forum `_.\n\nTo contribute to Django:\n\n* Check out https://docs.djangoproject.com/en/dev/internals/contributing/ for\n information about getting involved.\n\nTo run Django's test suite:\n\n* Follow the instructions in the \"Unit tests\" section of\n ``docs/internals/contributing/writing-code/unit-tests.txt``, published online at\n https://docs.djangoproject.com/en/dev/internals/contributing/writing-code/unit-tests/#running-the-unit-tests\n\nSupporting the Development of Django\n====================================\n\nDjango's development depends on your contributions.\n\nIf you depend on Django, remember to support the Django Software Foundation: https://www.djangoproject.com/fundraising/\n", - "description_content_type": "", - "docs_url": null, - "download_url": "", - "downloads": { - "last_day": -1, - "last_month": -1, - "last_week": -1 - }, - "home_page": "https://www.djangoproject.com/", - "keywords": "", - "license": "BSD-3-Clause", - "maintainer": "", - "maintainer_email": "", - "name": "Django", - "package_url": "https://pypi.org/project/Django/", - "platform": null, - "project_url": "https://pypi.org/project/Django/", - "project_urls": { - "Documentation": "https://docs.djangoproject.com/", - "Funding": "https://www.djangoproject.com/fundraising/", - "Homepage": "https://www.djangoproject.com/", - "Release notes": "https://docs.djangoproject.com/en/stable/releases/", - "Source": "https://github.com/django/django", - "Tracker": "https://code.djangoproject.com/" - }, - "release_url": "https://pypi.org/project/Django/5.0/", - "requires_dist": [ - "asgiref >=3.7.0", - "sqlparse >=0.3.1", - "tzdata ; sys_platform == \"win32\"", - "argon2-cffi >=19.1.0 ; extra == 'argon2'", - "bcrypt ; extra == 'bcrypt'" - ], - "requires_python": ">=3.10", - "summary": "A high-level Python web framework that encourages rapid development and clean, pragmatic design.", - "version": "5.0", - "yanked": false, - "yanked_reason": null + "author": "Beanbag, Inc.", + "author_email": "reviewboard@googlegroups.com", + "bugtrack_url": null, + "classifiers": [ + "Development Status :: 5 - Production/Stable", + "Environment :: Web Environment", + "Framework :: Django", + "Framework :: Django :: 4.2", + "Intended Audience :: Developers", + "License :: OSI Approved :: MIT License", + "Operating System :: OS Independent", + "Programming Language :: Python", + "Programming Language :: Python :: 3", + "Programming Language :: Python :: 3.10", + "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", + "Programming Language :: Python :: 3.8", + "Programming Language :: Python :: 3.9", + "Topic :: Software Development", + "Topic :: Software Development :: Libraries :: Python Modules" + ], + "description": "Djblets\n=======\n\nDjblets is a large collection of general and special-purpose building blocks\ndesigned to help with the development of web applications written using\nDjango_ and Python.\n\nThe following modules are available. These contain classes, functions,\ntemplate tags, templates, etc. that can be used by your own codebase.\n\n* djblets.auth_ -\n Authentication-related utilities for registration, login rate limiting, and\n other auth-related uses\n\n* djblets.avatars_ -\n Avatar rendering with flexible backends (supporting Gravatars, custom URLs,\n file uploads, or custom options)\n\n* djblets.cache_ -\n Helpers for working with client-side and server-side caching needs\n\n* djblets.conditions_ -\n User-defined condition rules under which actions should be performed\n\n* djblets.configforms_ -\n Category-based, multi-form configuration pages\n\n* djblets.datagrid_ -\n Customizable grids for displaying information, with custom columns\n\n* djblets.db_ -\n Specialized fields, validation, and query operations for databases\n\n* djblets.extensions_ -\n Extension framework, allowing third-party developers to extend your product\n or service\n\n* djblets.features_ -\n Feature flags for enabling/disabling functionality based on any criteria\n\n* djblets.forms_ -\n Specialized fields and widgets, enhanced form rendering, and\n dictionary-backed form data\n\n* djblets.gravatars_ -\n Low-level functions and template tags for injecting Gravatars_ into pages\n\n* djblets.http_ -\n Utilities for working with HTTP requests and responses.\n\n* djblets.integrations_ -\n Framework for integrating with third-party services and offering unlimited\n numbers of user-defined configurations\n\n* djblets.log_ -\n Enhanced logging capabilities and log viewing\n\n* djblets.mail_ -\n Enhanced Mail sending with DMARC checks and send-on-behalf-of-user\n functionality\n\n* djblets.markdown_ -\n Markdown rendering for pages and e-mails, with WYSIWYG editing/rendering\n support\n\n* djblets.pipeline_ -\n Pipeline_ compilers for ES6 JavaScript and optimized LessCSS support\n\n* djblets.privacy_ -\n Privacy-by-design support, allowing consent to be requested and tracked\n and personal information redacted\n\n* djblets.recaptcha_ -\n Mixins and form widgets for reCAPTCHA_ integration\n\n* djblets.registries_ -\n Base support for defining in-code registries, which tracks and allows lookup\n of custom-registered objects\n\n* djblets.secrets_ -\n Uilities and infrastructure for encryption/decryption and token generation.\n\n* djblets.siteconfig_ -\n In-database site configuration and settings, with Django settings mappings\n\n* djblets.template_ -\n Loaders for intelligent template caching and utilities for working with\n template caches and state\n\n* djblets.testing_ -\n Utilities for enhancing unit tests and defining smarter test runners\n\n* djblets.urls_ -\n Flexible root-level URL handlers, dynamic URL patterns that can be changed\n at runtime, and more\n\n* djblets.util_ -\n An assortment of misc. utility functions and template tags\n\n* djblets.views_ -\n Class-based View mixins for controlling caching and more complex dispatch\n behavior\n\n* djblets.webapi_ -\n Foundation for building fully-featured, consisent, maintainable REST APIs\n\nWe built and maintain Djblets as part of the `Review Board`_ code review\nproduct and Splat_ bug tracker at Beanbag_.\n\nSee the documentation_ for guides and code references for working with\nDjblets.\n\n\n.. _Beanbag: https://www.beanbaginc.com/\n.. _Django: https://www.djangoproject.com/\n.. _GDPR: https://www.eugdpr.org/\n.. _Gravatars: https://gravatars.com/\n.. _Pipeline: https://django-pipeline.readthedocs.io/en/latest/\n.. _reCAPTCHA: https://www.google.com/recaptcha/\n.. _Review Board: https://www.reviewboard.org/\n.. _Splat: https://www.hellosplat.com/\n.. _documentation: https://www.reviewboard.org/docs/djblets/latest/\n\n.. _djblets.auth:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-auth\n.. _djblets.avatars:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-avatars\n.. _djblets.cache:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-cache\n.. _djblets.conditions:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-conditions\n.. _djblets.configforms:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-configforms\n.. _djblets.datagrid:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-datagrid\n.. _djblets.db:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-db\n.. _djblets.extensions:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-extensions\n.. _djblets.features:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-features\n.. _djblets.feedview:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-feedview\n.. _djblets.forms:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-forms\n.. _djblets.gravatars:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-gravatars\n.. _djblets.http:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-http\n.. _djblets.integrations:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-integrations\n.. _djblets.log:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-log\n.. _djblets.mail:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-mail\n.. _djblets.markdown:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-markdown\n.. _djblets.pipeline:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-pipeline\n.. _djblets.privacy:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-privacy\n.. _djblets.recaptcha:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-recaptcha\n.. _djblets.registries:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-registries\n.. _djblets.secrets:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-secrets\n.. _djblets.siteconfig:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-siteconfig\n.. _djblets.template:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-template\n.. _djblets.testing:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-testing\n.. _djblets.urls:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-urls\n.. _djblets.util:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-util\n.. _djblets.views:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-views\n.. _djblets.webapi:\n https://www.reviewboard.org/docs/djblets/latest/coderef/#coderef-djblets-webapi\n\n\nCompatibility\n=============\n\nDjblets 5.x (release-5.x_) supports Python 3.8-3.12 and Django 4.2.\n\nDjblets 4.x (release-4.x_) supports Python 3.8-3.12 and Django 3.2.\n\nDjblets 3.x (release-3.x_) supports Python 3.7-3.11 and Django 3.2.\n\nDjblets 2.x (release-2.x_) supports Python 2.7, 3.5, and 3.6, and Django\n1.6, 1.8, 1.0. 1,0. and 1.11.\n\nDjblets 1.x (release-1.0.x_) supports Python 2.7 and Django 1.6.\n\nDjblets 0.9 (release-0.9.x_) supports Python 2.6 and 2.7, and Django 1.6.\n\nSee the `release notes`_ for information on the latest public releases.\n\n\n.. _release-0.9.x: https://github.com/djblets/djblets/tree/release-0.9.x\n.. _release-1.0.x: https://github.com/djblets/djblets/tree/release-1.0.x\n.. _release-2.x: https://github.com/djblets/djblets/tree/release-2.x\n.. _release-3.x: https://github.com/djblets/djblets/tree/release-3.x\n.. _release-4.x: https://github.com/djblets/djblets/tree/release-4.x\n.. _release-5.x: https://github.com/djblets/djblets/tree/release-5.x\n.. _release notes: https://www.reviewboard.org/docs/releasenotes/djblets/\n\n\nInstalling Djblets\n==================\n\nWe provide source builds and Python Wheels for Djblets.\n\nWe recommend you use Wheels unless you have a reason to use source builds\n(which requires a proper build setup for static media).\n\nTo install Wheels via pip::\n\n $ pip install Djblets\n\n\nGetting Support\n===============\n\nWe can help you with Djblets-related development over on our `Review Board\ndevelopment list`_.\n\n\n.. _Review Board development list:\n https://groups.google.com/group/reviewboard-dev\n\n\nReporting Bugs\n==============\n\nHit a bug? Let us know by\n`filing a bug report `_.\n\nYou can also look through the\n`existing bug reports `_ to see if anyone\nelse has already filed the bug.\n\n\nContributing\n============\n\nAre you a developer? Do you want to integrate Djblets into your project and\ncontribute back? Great! Let's help get you started.\n\nFirst off, we have a few handy guides:\n\n* `Review Board Contributor Guide`_ -\n This generally applies to Djblets as well.\n\nWe accept patches on `reviews.reviewboard.org\n`_. (Please note that we *do not* accept pull\nrequests.)\n\n.. _Review Board Contributor Guide:\n https://www.notion.so/reviewboard/Review-Board-45d228fb07a0459b84fee509ac054cec\n\n\nRelated Projects\n================\n\n* `Review Board`_ -\n Our dedicated open source code review product for teams of all sizes.\n", + "description_content_type": "text/x-rst", + "docs_url": null, + "download_url": "https://downloads.reviewboard.org/releases/Djblets/5.1/", + "downloads": { + "last_day": -1, + "last_month": -1, + "last_week": -1 + }, + "dynamic": null, + "home_page": "https://www.reviewboard.org/downloads/djblets/", + "keywords": null, + "license": "MIT", + "maintainer": null, + "maintainer_email": null, + "name": "Djblets", + "package_url": "https://pypi.org/project/Djblets/", + "platform": null, + "project_url": "https://pypi.org/project/Djblets/", + "project_urls": { + "Download": "https://downloads.reviewboard.org/releases/Djblets/5.1/", + "Homepage": "https://www.reviewboard.org/downloads/djblets/" + }, + "provides_extra": null, + "release_url": "https://pypi.org/project/Djblets/5.1.1/", + "requires_dist": [ + "cryptography>=41.0.4", + "django-pipeline~=3.1.0", + "Django~=4.2.15", + "dnspython>=2.3.0", + "housekeeping~=1.1", + "importlib-metadata>=6.6", + "importlib-resources>=5.9", + "packaging>=23.1", + "Pillow>=6.2", + "publicsuffixlist~=0.10.0", + "python-dateutil>=2.7", + "pytz", + "typing-extensions>=4.4" + ], + "requires_python": ">=3.8", + "summary": "A collection of useful classes and functions for developing large-scale Django-based web applications.", + "version": "5.1.1", + "yanked": false, + "yanked_reason": null }, - "last_serial": 20903581, + "last_serial": 24642017, "releases": { - "1.0.1": [], - "1.0.2": [], - "1.0.3": [], - "1.0.4": [], - "1.1": [], - "1.1.1": [], - "1.1.2": [], - "1.1.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "8f1f74aa91b56dea5847b62e11ce6737db82c6446561bddc20ca80fa5df025cc", - "md5": "52848c23dbc120fe0b2a8e7189b20306", - "sha256": "0e5034cf8046ba77c62e95a45d776d2c59998b26f181ceaf5cec516115e3f85a" - }, - "downloads": -1, - "filename": "Django-1.1.3.tar.gz", - "has_sig": false, - "md5_digest": "52848c23dbc120fe0b2a8e7189b20306", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 5748608, - "upload_time": "2010-12-23T05:14:23", - "upload_time_iso_8601": "2010-12-23T05:14:23.509436Z", - "url": "https://files.pythonhosted.org/packages/8f/1f/74aa91b56dea5847b62e11ce6737db82c6446561bddc20ca80fa5df025cc/Django-1.1.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.1.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "0001c29275c88671d5e4089388c54ecbd72ed64f8d472067f765e52f767d472a", - "md5": "e818668acc4de944f85e494ac80f1e7d", - "sha256": "1f9d48a741f98951e65818e167c84c407d1c322efcfd4cb419384773ea793dee" - }, - "downloads": -1, - "filename": "Django-1.1.4.tar.gz", - "has_sig": false, - "md5_digest": "e818668acc4de944f85e494ac80f1e7d", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 5750441, - "upload_time": "2011-02-09T04:13:07", - "upload_time_iso_8601": "2011-02-09T04:13:07.000075Z", - "url": "https://files.pythonhosted.org/packages/00/01/c29275c88671d5e4089388c54ecbd72ed64f8d472067f765e52f767d472a/Django-1.1.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.10": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "4b4c059f68d8f029f7054e4e6bb0b1ed2fde7f28d07a3727325727d5a95ae1b8", - "md5": "36e17cd1a0255258e1dec1bbb8808040", - "sha256": "9c60f4a801bf7c26bd6824c1062550c12c373344116703461c18cc258f8c9320" - }, - "downloads": -1, - "filename": "Django-1.10-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "36e17cd1a0255258e1dec1bbb8808040", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6795167, - "upload_time": "2016-08-01T18:32:07", - "upload_time_iso_8601": "2016-08-01T18:32:07.351674Z", - "url": "https://files.pythonhosted.org/packages/4b/4c/059f68d8f029f7054e4e6bb0b1ed2fde7f28d07a3727325727d5a95ae1b8/Django-1.10-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "185c3cd8989b2226c55a1faf66f1a110e76cba6e6ca5d9dd15fb469fb636f378", - "md5": "939e4d989b93a4e12e4ec5d98fcdb4f5", - "sha256": "46b868d68e5fd69dd9e05a0a7900df91786097e30b2aa6f065dd7fa3b22f7005" - }, - "downloads": -1, - "filename": "Django-1.10.tar.gz", - "has_sig": false, - "md5_digest": "939e4d989b93a4e12e4ec5d98fcdb4f5", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7691063, - "upload_time": "2016-08-01T18:32:16", - "upload_time_iso_8601": "2016-08-01T18:32:16.280614Z", - "url": "https://files.pythonhosted.org/packages/18/5c/3cd8989b2226c55a1faf66f1a110e76cba6e6ca5d9dd15fb469fb636f378/Django-1.10.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.10.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "6ccfd6ab0edb891865ef86b3e3d7290c162f57c363cf880099bbe94229806f56", - "md5": "6b50546050424bf01fd9687de3096855", - "sha256": "3d689905cd0635bbb33b87f9a5df7ca70a3db206faae4ec58cda5e7f5f47050d" - }, - "downloads": -1, - "filename": "Django-1.10.1-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "6b50546050424bf01fd9687de3096855", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6796229, - "upload_time": "2016-09-01T23:17:41", - "upload_time_iso_8601": "2016-09-01T23:17:41.185068Z", - "url": "https://files.pythonhosted.org/packages/6c/cf/d6ab0edb891865ef86b3e3d7290c162f57c363cf880099bbe94229806f56/Django-1.10.1-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "0a9ee76cca958089cd0317ab46cb91f0ed36274900e48829c949b2e33d2a4469", - "md5": "037d07e126eecc15a5fbf5221dd4081b", - "sha256": "d6e6c5b25cb67f46afd7c82f536529b11981183423dad8932e15bce93d1a24f3" - }, - "downloads": -1, - "filename": "Django-1.10.1.tar.gz", - "has_sig": false, - "md5_digest": "037d07e126eecc15a5fbf5221dd4081b", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7700057, - "upload_time": "2016-09-01T23:18:18", - "upload_time_iso_8601": "2016-09-01T23:18:18.672706Z", - "url": "https://files.pythonhosted.org/packages/0a/9e/e76cca958089cd0317ab46cb91f0ed36274900e48829c949b2e33d2a4469/Django-1.10.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.10.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "8a0946f790104abca7eb93786139d3adde9366b1afd59a77b583a1f310dc8cbd", - "md5": "0b29f13cc5907dcf6f06649ce77be7c2", - "sha256": "4d48ab8e84a7c8b2bc4b2f4f199bc3a8bfcc9cbdbc29e355ac5c44a501d73a1a" - }, - "downloads": -1, - "filename": "Django-1.10.2-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "0b29f13cc5907dcf6f06649ce77be7c2", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6832895, - "upload_time": "2016-10-01T20:05:18", - "upload_time_iso_8601": "2016-10-01T20:05:18.594291Z", - "url": "https://files.pythonhosted.org/packages/8a/09/46f790104abca7eb93786139d3adde9366b1afd59a77b583a1f310dc8cbd/Django-1.10.2-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "579e59444485f092b6ed4f1931e7d2e13b67fdab967c041d02f58a0d1dab8c23", - "md5": "5342e77374b2acd2eafa86d2bb68f8c9", - "sha256": "e127f12a0bfb34843b6e8c82f91e26fff6445a7ca91d222c0794174cf97cbce1" - }, - "downloads": -1, - "filename": "Django-1.10.2.tar.gz", - "has_sig": false, - "md5_digest": "5342e77374b2acd2eafa86d2bb68f8c9", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7724987, - "upload_time": "2016-10-01T20:05:31", - "upload_time_iso_8601": "2016-10-01T20:05:31.330942Z", - "url": "https://files.pythonhosted.org/packages/57/9e/59444485f092b6ed4f1931e7d2e13b67fdab967c041d02f58a0d1dab8c23/Django-1.10.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.10.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "0eab16abddb9ab7ee46a26e04a0c8ba1f02b9412a77927dec699c1af6d0070f8", - "md5": "dd66eaec09d7a3810c40b01c53535b37", - "sha256": "94426cc28d8721fbf13c333053f08d32427671a4ca7986f7030fc82bdf9c88c1" - }, - "downloads": -1, - "filename": "Django-1.10.3-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "dd66eaec09d7a3810c40b01c53535b37", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6833335, - "upload_time": "2016-11-01T13:56:54", - "upload_time_iso_8601": "2016-11-01T13:56:54.139706Z", - "url": "https://files.pythonhosted.org/packages/0e/ab/16abddb9ab7ee46a26e04a0c8ba1f02b9412a77927dec699c1af6d0070f8/Django-1.10.3-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "4d6bcf3edad0526851d1fd6dd56c9cc94f2be090489c39d9666ca4ad980312e2", - "md5": "70e4e0e6b2b38e782436e4eb7eb6ff39", - "sha256": "6f92f08dee8a1bd7680e098a91bf5acd08b5cdfe74137f695b60fd79f4478c30" - }, - "downloads": -1, - "filename": "Django-1.10.3.tar.gz", - "has_sig": false, - "md5_digest": "70e4e0e6b2b38e782436e4eb7eb6ff39", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7733727, - "upload_time": "2016-11-01T13:57:16", - "upload_time_iso_8601": "2016-11-01T13:57:16.055061Z", - "url": "https://files.pythonhosted.org/packages/4d/6b/cf3edad0526851d1fd6dd56c9cc94f2be090489c39d9666ca4ad980312e2/Django-1.10.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.10.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "7137581a00bbc4571526ce88ef517c0c02ca7575ac2ae8a3671161d2aa14b740", - "md5": "28f2a0607bb52dac9c1b168b374de1cd", - "sha256": "a8e1a552205cda15023c39ecf17f7e525e96c5b0142e7879e8bd0c445351f2cc" - }, - "downloads": -1, - "filename": "Django-1.10.4-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "28f2a0607bb52dac9c1b168b374de1cd", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6833471, - "upload_time": "2016-12-01T23:46:26", - "upload_time_iso_8601": "2016-12-01T23:46:26.502027Z", - "url": "https://files.pythonhosted.org/packages/71/37/581a00bbc4571526ce88ef517c0c02ca7575ac2ae8a3671161d2aa14b740/Django-1.10.4-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "3b146c1e7508b1342afde8e80f50a55d6b305c0755c702f741db6094924f7499", - "md5": "65aa2a1bd3b3f08b16a4cd368472d520", - "sha256": "fff7f062e510d812badde7cfc57745b7779edb4d209b2bc5ea8d954c22305c2b" - }, - "downloads": -1, - "filename": "Django-1.10.4.tar.gz", - "has_sig": false, - "md5_digest": "65aa2a1bd3b3f08b16a4cd368472d520", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7735213, - "upload_time": "2016-12-01T23:46:50", - "upload_time_iso_8601": "2016-12-01T23:46:50.215935Z", - "url": "https://files.pythonhosted.org/packages/3b/14/6c1e7508b1342afde8e80f50a55d6b305c0755c702f741db6094924f7499/Django-1.10.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.10.5": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "4560faa28a1d17f879f9dbef28f249e4e9a8dd1d29ae78409516b4b8b6c3ebab", - "md5": "6892778eea81f14acd58d883f10f3d9f", - "sha256": "4541a60834f28f308ee7b6e96400feca905fb0de473eb9dad6847e98a36d86d4" - }, - "downloads": -1, - "filename": "Django-1.10.5-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "6892778eea81f14acd58d883f10f3d9f", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6833796, - "upload_time": "2017-01-04T19:22:17", - "upload_time_iso_8601": "2017-01-04T19:22:17.889078Z", - "url": "https://files.pythonhosted.org/packages/45/60/faa28a1d17f879f9dbef28f249e4e9a8dd1d29ae78409516b4b8b6c3ebab/Django-1.10.5-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "c3c26096bf5d0caa4e3d5b985ac72e3a0c795e37fa7407d6c85460b2a105b467", - "md5": "3fce02f1e6461fec21f1f15ea7489924", - "sha256": "0db89374b691b9c8b057632a6cd64b18d08db2f4d63b4d4af6024267ab965f8b" - }, - "downloads": -1, - "filename": "Django-1.10.5.tar.gz", - "has_sig": false, - "md5_digest": "3fce02f1e6461fec21f1f15ea7489924", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7734715, - "upload_time": "2017-01-04T19:23:00", - "upload_time_iso_8601": "2017-01-04T19:23:00.596664Z", - "url": "https://files.pythonhosted.org/packages/c3/c2/6096bf5d0caa4e3d5b985ac72e3a0c795e37fa7407d6c85460b2a105b467/Django-1.10.5.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.10.6": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "b9bb723f78e6f6aea78590331eba4e42b8a09c33ce154204a942525a91101d0b", - "md5": "31a63e4c21a4e12d5ebbafc137523e40", - "sha256": "2cfb83859bfaa10e2bd586340bead27c69fdcaa21fa683a008cc712482c26726" - }, - "downloads": -1, - "filename": "Django-1.10.6-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "31a63e4c21a4e12d5ebbafc137523e40", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6833948, - "upload_time": "2017-03-01T13:37:27", - "upload_time_iso_8601": "2017-03-01T13:37:27.613779Z", - "url": "https://files.pythonhosted.org/packages/b9/bb/723f78e6f6aea78590331eba4e42b8a09c33ce154204a942525a91101d0b/Django-1.10.6-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "1d07fb81c7ed26abbfadd84185be80b5b949219948c4bfd7c30c5c1436d5fd7d", - "md5": "aaf0e61104bca75f2dea179d666537cf", - "sha256": "7a6ebe254ab126510da143628d019ca8d6da2de49d7682bf046c03713a3c2c61" - }, - "downloads": -1, - "filename": "Django-1.10.6.tar.gz", - "has_sig": false, - "md5_digest": "aaf0e61104bca75f2dea179d666537cf", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7734864, - "upload_time": "2017-03-01T13:37:40", - "upload_time_iso_8601": "2017-03-01T13:37:40.243134Z", - "url": "https://files.pythonhosted.org/packages/1d/07/fb81c7ed26abbfadd84185be80b5b949219948c4bfd7c30c5c1436d5fd7d/Django-1.10.6.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.10.7": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "e5e7bdcc0837a2e7ccb1a37be9e5e6e6da642cec5fe9fc1f9ac37dd397c91f74", - "md5": "ed23475695d32c176e12b6e2a1fbe1aa", - "sha256": "e68fd450154ad7ee2c88472bb812350490232462adc6e3c6bcb544abe5212134" - }, - "downloads": -1, - "filename": "Django-1.10.7-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "ed23475695d32c176e12b6e2a1fbe1aa", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6834495, - "upload_time": "2017-04-04T14:27:40", - "upload_time_iso_8601": "2017-04-04T14:27:40.297406Z", - "url": "https://files.pythonhosted.org/packages/e5/e7/bdcc0837a2e7ccb1a37be9e5e6e6da642cec5fe9fc1f9ac37dd397c91f74/Django-1.10.7-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "15b4d4bb7313e02386bd23a60e1eb5670321313fb67289c6f36ec43bce747aff", - "md5": "693dfeabad62c561cb205900d32c2a98", - "sha256": "593d779dbc2350a245c4f76d26bdcad58a39895e87304fe6d725bbdf84b5b0b8" - }, - "downloads": -1, - "filename": "Django-1.10.7.tar.gz", - "has_sig": false, - "md5_digest": "693dfeabad62c561cb205900d32c2a98", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7737654, - "upload_time": "2017-04-04T14:27:54", - "upload_time_iso_8601": "2017-04-04T14:27:54.235551Z", - "url": "https://files.pythonhosted.org/packages/15/b4/d4bb7313e02386bd23a60e1eb5670321313fb67289c6f36ec43bce747aff/Django-1.10.7.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.10.8": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "bb9f2c20639ac635a83123ddffd91ba15001cb0d04e74fbb08f31fb57e490dab", - "md5": "76640241d7aa59c87a5095fbebc7e2a1", - "sha256": "ffdc7e938391ae3c2ee8ff82e0b4444e4e6bb15c99d00770285233d42aaf33d6" - }, - "downloads": -1, - "filename": "Django-1.10.8-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "76640241d7aa59c87a5095fbebc7e2a1", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6834486, - "upload_time": "2017-09-05T15:31:48", - "upload_time_iso_8601": "2017-09-05T15:31:48.077227Z", - "url": "https://files.pythonhosted.org/packages/bb/9f/2c20639ac635a83123ddffd91ba15001cb0d04e74fbb08f31fb57e490dab/Django-1.10.8-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "091713a0cd29f603a4a51b06f7cdc9466fd7bfc48aa20ae2aa80f79d3ad9ba7d", - "md5": "d140e63b9f704ab375d052c40f9d8e76", - "sha256": "d4ef83bd326573c00972cb9429beb396d210341a636e4b816fc9b3f505c498bb" - }, - "downloads": -1, - "filename": "Django-1.10.8.tar.gz", - "has_sig": false, - "md5_digest": "d140e63b9f704ab375d052c40f9d8e76", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7739226, - "upload_time": "2017-09-05T15:31:58", - "upload_time_iso_8601": "2017-09-05T15:31:58.221021Z", - "url": "https://files.pythonhosted.org/packages/09/17/13a0cd29f603a4a51b06f7cdc9466fd7bfc48aa20ae2aa80f79d3ad9ba7d/Django-1.10.8.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.10a1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "0250b210dc6206e9c61a25a05acb76fc2f09101d9031ff0fb4d137f746e2e419", - "md5": "2e29930f401031ecf33e36bfa4739245", - "sha256": "1a8b6ad1f8fabbbd2e1ef8fb54dfe5f9a0b4908c642cccee20d58cfd7c0a3f7e" - }, - "downloads": -1, - "filename": "Django-1.10a1-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "2e29930f401031ecf33e36bfa4739245", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6600566, - "upload_time": "2016-05-20T12:16:44", - "upload_time_iso_8601": "2016-05-20T12:16:44.951411Z", - "url": "https://files.pythonhosted.org/packages/02/50/b210dc6206e9c61a25a05acb76fc2f09101d9031ff0fb4d137f746e2e419/Django-1.10a1-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "ae983b27e0a3c53ec0d6727eb19da46eed5705fa9250ed28ec0d1df48778c401", - "md5": "25f3210c15f3bf9bb7e4c33adfbcb952", - "sha256": "a53bbf8be7be60a9479295ab2bef375c1e25ae777d00ff0fea5ac2e347aa5c76" - }, - "downloads": -1, - "filename": "Django-1.10a1.tar.gz", - "has_sig": false, - "md5_digest": "25f3210c15f3bf9bb7e4c33adfbcb952", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7543001, - "upload_time": "2016-05-20T12:24:59", - "upload_time_iso_8601": "2016-05-20T12:24:59.952686Z", - "url": "https://files.pythonhosted.org/packages/ae/98/3b27e0a3c53ec0d6727eb19da46eed5705fa9250ed28ec0d1df48778c401/Django-1.10a1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.10b1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "e68c142b08d2dc89aec2b74ad1d37f943ec50c73d4afce12ea6c0c568403ab22", - "md5": "907c0f7f7b6ac716e46312015dfa9e1a", - "sha256": "3dee9e77e12d3edc30aed96e5522632d8ded656845c4e3e804dab8c60a937478" - }, - "downloads": -1, - "filename": "Django-1.10b1-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "907c0f7f7b6ac716e46312015dfa9e1a", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6641407, - "upload_time": "2016-06-22T01:15:05", - "upload_time_iso_8601": "2016-06-22T01:15:05.240779Z", - "url": "https://files.pythonhosted.org/packages/e6/8c/142b08d2dc89aec2b74ad1d37f943ec50c73d4afce12ea6c0c568403ab22/Django-1.10b1-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "02bed10613977c37674ca3b7f6db7105fac9a104b7765ede4a2f1445fabc2873", - "md5": "c2f10b2d1453adc68e37132c3304317a", - "sha256": "d8ef9aef259d68a452d5ae1a6f60793e8c10c609dbbe9e7412d47ac21e6d4245" - }, - "downloads": -1, - "filename": "Django-1.10b1.tar.gz", - "has_sig": false, - "md5_digest": "c2f10b2d1453adc68e37132c3304317a", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7600640, - "upload_time": "2016-06-22T01:15:17", - "upload_time_iso_8601": "2016-06-22T01:15:17.267637Z", - "url": "https://files.pythonhosted.org/packages/02/be/d10613977c37674ca3b7f6db7105fac9a104b7765ede4a2f1445fabc2873/Django-1.10b1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.10rc1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "79a2988b57157526dcbdf78501c68ba6409b7863381a6cc6bc06424e07e134a2", - "md5": "c951d6a11587d5e9f2b5e0a5cca96915", - "sha256": "ccb60ae7804bc451e42b39e6863fc916de8c1fd9a681426e4d9fc9a1abf8bd44" - }, - "downloads": -1, - "filename": "Django-1.10rc1-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "c951d6a11587d5e9f2b5e0a5cca96915", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6785247, - "upload_time": "2016-07-18T18:04:51", - "upload_time_iso_8601": "2016-07-18T18:04:51.589122Z", - "url": "https://files.pythonhosted.org/packages/79/a2/988b57157526dcbdf78501c68ba6409b7863381a6cc6bc06424e07e134a2/Django-1.10rc1-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "320ba54e4d4922545b0deb6808d4af0bb78010c0ca4d3109608ce6675f4f0ea1", - "md5": "daf478a2459e54ba28e5ec600d669960", - "sha256": "26d08f62284d838598bc45671af6e6dba880d54fff3c14aa6aa78ba5519aeac0" - }, - "downloads": -1, - "filename": "Django-1.10rc1.tar.gz", - "has_sig": false, - "md5_digest": "daf478a2459e54ba28e5ec600d669960", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7687388, - "upload_time": "2016-07-18T18:05:05", - "upload_time_iso_8601": "2016-07-18T18:05:05.503584Z", - "url": "https://files.pythonhosted.org/packages/32/0b/a54e4d4922545b0deb6808d4af0bb78010c0ca4d3109608ce6675f4f0ea1/Django-1.10rc1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "47a6078ebcbd49b19e22fd560a2348cfc5cec9e5dcfe3c4fad8e64c9865135bb", - "md5": "4a191de100babe2bf88f205982d48e57", - "sha256": "0120b3b60760fb0617848b58aaa9702c0bf963320ed472f0879c5c55ab75b64a" - }, - "downloads": -1, - "filename": "Django-1.11-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "4a191de100babe2bf88f205982d48e57", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6942445, - "upload_time": "2017-04-04T15:59:30", - "upload_time_iso_8601": "2017-04-04T15:59:30.713294Z", - "url": "https://files.pythonhosted.org/packages/47/a6/078ebcbd49b19e22fd560a2348cfc5cec9e5dcfe3c4fad8e64c9865135bb/Django-1.11-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "7943ed9ca4d69f35b5e64f2ecad73f75a8529a9c6f0d562e5af9a1f65beda355", - "md5": "5008d266f198c2fe761916139162a0c2", - "sha256": "b6f3b864944276b4fd1d099952112696558f78b77b39188ac92b6c5e80152c30" - }, - "downloads": -1, - "filename": "Django-1.11.tar.gz", - "has_sig": false, - "md5_digest": "5008d266f198c2fe761916139162a0c2", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7853479, - "upload_time": "2017-04-04T16:00:04", - "upload_time_iso_8601": "2017-04-04T16:00:04.407084Z", - "url": "https://files.pythonhosted.org/packages/79/43/ed9ca4d69f35b5e64f2ecad73f75a8529a9c6f0d562e5af9a1f65beda355/Django-1.11.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "2b2c019d6d5f7ed2889082ed96f849bf462c57265087a3a568a19b0d4c53bc55", - "md5": "a300c34f63f1b5b1a57447b89fecbb85", - "sha256": "bb3109a31cfa016e5f234223665f80fc06107f2169afb9f6dc8828295db73547" - }, - "downloads": -1, - "filename": "Django-1.11.1-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "a300c34f63f1b5b1a57447b89fecbb85", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6943932, - "upload_time": "2017-05-06T13:26:23", - "upload_time_iso_8601": "2017-05-06T13:26:23.402737Z", - "url": "https://files.pythonhosted.org/packages/2b/2c/019d6d5f7ed2889082ed96f849bf462c57265087a3a568a19b0d4c53bc55/Django-1.11.1-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "226824fd855343f218e016bbb04b1a5a2dc1be191a4cbd4d3cdabb13d1c2a371", - "md5": "c625789a32f87fda28df0a71c1b3e324", - "sha256": "bbcefdf822eeef2cd04718ebcc24dd2ecf47407258cfcde2b4f95df57ce33a8c" - }, - "downloads": -1, - "filename": "Django-1.11.1.tar.gz", - "has_sig": false, - "md5_digest": "c625789a32f87fda28df0a71c1b3e324", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7857868, - "upload_time": "2017-05-06T13:26:38", - "upload_time_iso_8601": "2017-05-06T13:26:38.108379Z", - "url": "https://files.pythonhosted.org/packages/22/68/24fd855343f218e016bbb04b1a5a2dc1be191a4cbd4d3cdabb13d1c2a371/Django-1.11.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.10": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "3244a9afb2b7dd641d89341a2126020bbda9201d270e23e4ffe2601eeaaabcfd", - "md5": "dc59370ed297d818fe90b873fb22c901", - "sha256": "ac4c797a328a5ac8777ad61bcd00da279773455cc78b4058de2a9842a0eb6ee8" - }, - "downloads": -1, - "filename": "Django-1.11.10-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "dc59370ed297d818fe90b873fb22c901", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6949488, - "upload_time": "2018-02-01T14:40:14", - "upload_time_iso_8601": "2018-02-01T14:40:14.674673Z", - "url": "https://files.pythonhosted.org/packages/32/44/a9afb2b7dd641d89341a2126020bbda9201d270e23e4ffe2601eeaaabcfd/Django-1.11.10-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "58bfbc99750dcb155639e0b1d3f9dd4729b12c1d4f1bc37610b56a0cdcbcc66a", - "md5": "f306015e16a8d5024dbac923ac34fffb", - "sha256": "22383567385a9c406d8a5ce080a2694c82c6b733e157922197e8b393bb3aacd9" - }, - "downloads": -1, - "filename": "Django-1.11.10.tar.gz", - "has_sig": false, - "md5_digest": "f306015e16a8d5024dbac923ac34fffb", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7881348, - "upload_time": "2018-02-01T14:40:24", - "upload_time_iso_8601": "2018-02-01T14:40:24.121117Z", - "url": "https://files.pythonhosted.org/packages/58/bf/bc99750dcb155639e0b1d3f9dd4729b12c1d4f1bc37610b56a0cdcbcc66a/Django-1.11.10.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.11": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "d5bf2cd5eb314aa2b89855c01259c94dc48dbd9be6c269370c1f7ae4979e6e2f", - "md5": "47f8dbee713f14b6480c319e9eaf500e", - "sha256": "fd186d544c7c2f835668cf11f77be071307c9eb22615a5b3a16bdb14c8357f41" - }, - "downloads": -1, - "filename": "Django-1.11.11-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "47f8dbee713f14b6480c319e9eaf500e", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6949575, - "upload_time": "2018-03-06T14:15:57", - "upload_time_iso_8601": "2018-03-06T14:15:57.141867Z", - "url": "https://files.pythonhosted.org/packages/d5/bf/2cd5eb314aa2b89855c01259c94dc48dbd9be6c269370c1f7ae4979e6e2f/Django-1.11.11-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "736dbf75a9854bc9a9af65d06a19fb89e194bbfd9c925aea2bf818dd2452b7b5", - "md5": "f57c1946db67fe15a5c35166235d0c37", - "sha256": "74077d7309b48b97dacdac2dfb35c968028becf00a7a684e7f29b2af1b980edc" - }, - "downloads": -1, - "filename": "Django-1.11.11.tar.gz", - "has_sig": false, - "md5_digest": "f57c1946db67fe15a5c35166235d0c37", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7961491, - "upload_time": "2018-03-06T14:16:18", - "upload_time_iso_8601": "2018-03-06T14:16:18.949597Z", - "url": "https://files.pythonhosted.org/packages/73/6d/bf75a9854bc9a9af65d06a19fb89e194bbfd9c925aea2bf818dd2452b7b5/Django-1.11.11.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.12": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "92561f30c1e6a58b0c97c492461148edcbece2c6e43dcc3529695165744349ee", - "md5": "9f48e21c7c0054ffb06c629d041d98d4", - "sha256": "056fe5b9e1f8f7fed9bb392919d64f6b33b3a71cfb0f170a90ee277a6ed32bc2" - }, - "downloads": -1, - "filename": "Django-1.11.12-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "9f48e21c7c0054ffb06c629d041d98d4", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6948451, - "upload_time": "2018-04-03T02:45:43", - "upload_time_iso_8601": "2018-04-03T02:45:43.053920Z", - "url": "https://files.pythonhosted.org/packages/92/56/1f30c1e6a58b0c97c492461148edcbece2c6e43dcc3529695165744349ee/Django-1.11.12-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "b73c30a9e5c9c21c5ac298684f00aca2479f3d7c43c5013f3ea17d0595ed90b7", - "md5": "af669c68c00382780c05d0e7a77b0d48", - "sha256": "4d398c7b02761e234bbde490aea13ea94cb539ceeb72805b72303f348682f2eb" - }, - "downloads": -1, - "filename": "Django-1.11.12.tar.gz", - "has_sig": false, - "md5_digest": "af669c68c00382780c05d0e7a77b0d48", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7882396, - "upload_time": "2018-04-03T02:46:00", - "upload_time_iso_8601": "2018-04-03T02:46:00.567935Z", - "url": "https://files.pythonhosted.org/packages/b7/3c/30a9e5c9c21c5ac298684f00aca2479f3d7c43c5013f3ea17d0595ed90b7/Django-1.11.12.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.13": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "254dc8228419346a0e84aec202a43e181afc6572b861d38f8a0306dbce6abef0", - "md5": "6a723c5239e19f70fb797cbaccc01192", - "sha256": "18986bcffe69653a84eaf1faa1fa5a7eded32cee41cfecc77fdc65a3e046404d" - }, - "downloads": -1, - "filename": "Django-1.11.13-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "6a723c5239e19f70fb797cbaccc01192", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6948641, - "upload_time": "2018-05-02T01:54:01", - "upload_time_iso_8601": "2018-05-02T01:54:01.504059Z", - "url": "https://files.pythonhosted.org/packages/25/4d/c8228419346a0e84aec202a43e181afc6572b861d38f8a0306dbce6abef0/Django-1.11.13-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "f9e3b9226c6a12bd5984daa68d41f066df0021a511bfbc86b83cdfc7dce956d9", - "md5": "ebdac613143ebdca911d5cef326fdc53", - "sha256": "46adfe8e0abe4d1f026c1086889970b611aec492784fbdfbdaabc2457360a4a5" - }, - "downloads": -1, - "filename": "Django-1.11.13.tar.gz", - "has_sig": false, - "md5_digest": "ebdac613143ebdca911d5cef326fdc53", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7884529, - "upload_time": "2018-05-02T01:54:13", - "upload_time_iso_8601": "2018-05-02T01:54:13.881452Z", - "url": "https://files.pythonhosted.org/packages/f9/e3/b9226c6a12bd5984daa68d41f066df0021a511bfbc86b83cdfc7dce956d9/Django-1.11.13.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.14": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "bfe0e659df5b5b82299fffd8b3df2910c99351b9308b8f45f5702cc4cdf946e9", - "md5": "82813473ac4008f14470c68a4c081edd", - "sha256": "b7f77c0d168de4c4ad30a02ae31b9dca04fb3c10472f04918d5c02b4117bba68" - }, - "downloads": -1, - "filename": "Django-1.11.14-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "82813473ac4008f14470c68a4c081edd", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6950140, - "upload_time": "2018-07-02T09:01:54", - "upload_time_iso_8601": "2018-07-02T09:01:54.170037Z", - "url": "https://files.pythonhosted.org/packages/bf/e0/e659df5b5b82299fffd8b3df2910c99351b9308b8f45f5702cc4cdf946e9/Django-1.11.14-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "dc35273514b5eb2201d5183dd138c240b12e1d822d2358b1e9436545b66d2d76", - "md5": "38e82b59a1c27bbf98ccf0564ead7426", - "sha256": "eb9271f0874f53106a2719c0c35ce67631f6cc27cf81a60c6f8c9817b35a3f6e" - }, - "downloads": -1, - "filename": "Django-1.11.14.tar.gz", - "has_sig": false, - "md5_digest": "38e82b59a1c27bbf98ccf0564ead7426", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7850578, - "upload_time": "2018-07-02T09:02:11", - "upload_time_iso_8601": "2018-07-02T09:02:11.027707Z", - "url": "https://files.pythonhosted.org/packages/dc/35/273514b5eb2201d5183dd138c240b12e1d822d2358b1e9436545b66d2d76/Django-1.11.14.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.15": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "f81c31112c778b7a56ce18e3fff5e8915719fbe1cd3476c1eef557dddacfac8b", - "md5": "d72f2d11b2c8f6cdced3d4c682d7a6fb", - "sha256": "8176ac7985fe6737ce3d6b2531b4a2453cb7c3377c9db00bacb2b3320f4a1311" - }, - "downloads": -1, - "filename": "Django-1.11.15-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "d72f2d11b2c8f6cdced3d4c682d7a6fb", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6949154, - "upload_time": "2018-08-01T13:45:07", - "upload_time_iso_8601": "2018-08-01T13:45:07.143406Z", - "url": "https://files.pythonhosted.org/packages/f8/1c/31112c778b7a56ce18e3fff5e8915719fbe1cd3476c1eef557dddacfac8b/Django-1.11.15-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "43b5b44286e56a5211d37b4058dcd5e62835afa5ce5aa6a38b56bd04c0d01cbc", - "md5": "9c25bc2575a2cd357bcc5764f809d29d", - "sha256": "b18235d82426f09733d2de9910cee975cf52ff05e5f836681eb957d105a05a40" - }, - "downloads": -1, - "filename": "Django-1.11.15.tar.gz", - "has_sig": false, - "md5_digest": "9c25bc2575a2cd357bcc5764f809d29d", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7843843, - "upload_time": "2018-08-01T13:45:14", - "upload_time_iso_8601": "2018-08-01T13:45:14.692609Z", - "url": "https://files.pythonhosted.org/packages/43/b5/b44286e56a5211d37b4058dcd5e62835afa5ce5aa6a38b56bd04c0d01cbc/Django-1.11.15.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.16": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "44e7872bbf76aa16b7a061698d75325dac023285db33db4bda8ba8fe5d3bb356", - "md5": "70439596428f6c669266d75a69efe297", - "sha256": "37f5876c1fbfd66085001f4c06fa0bf96ef05442c53daf8d4294b6f29e7fa6b8" - }, - "downloads": -1, - "filename": "Django-1.11.16-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "70439596428f6c669266d75a69efe297", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6950378, - "upload_time": "2018-10-01T09:22:02", - "upload_time_iso_8601": "2018-10-01T09:22:02.053553Z", - "url": "https://files.pythonhosted.org/packages/44/e7/872bbf76aa16b7a061698d75325dac023285db33db4bda8ba8fe5d3bb356/Django-1.11.16-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "351d59836bce4c9cfded261e21c0abd6a4629de6d289522d0fd928117d8eb985", - "md5": "e380fba4b67c360dc99337c648f56b4a", - "sha256": "29268cc47816a44f27308e60f71da635f549c47d8a1d003b28de55141df75791" - }, - "downloads": -1, - "filename": "Django-1.11.16.tar.gz", - "has_sig": false, - "md5_digest": "e380fba4b67c360dc99337c648f56b4a", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7852514, - "upload_time": "2018-10-01T09:22:18", - "upload_time_iso_8601": "2018-10-01T09:22:18.040195Z", - "url": "https://files.pythonhosted.org/packages/35/1d/59836bce4c9cfded261e21c0abd6a4629de6d289522d0fd928117d8eb985/Django-1.11.16.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.17": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "092b6c2d363e3d46307251a9d6bf74ec28543805bbcadf56ca729f4a04846914", - "md5": "a638d81ed5a9f3e72aaa623208a0328e", - "sha256": "f1a961b954d96bb24b397db4c35e9a128d12e044d6b57984c122282b592d508d" - }, - "downloads": -1, - "filename": "Django-1.11.17-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "a638d81ed5a9f3e72aaa623208a0328e", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6950469, - "upload_time": "2018-12-03T17:02:50", - "upload_time_iso_8601": "2018-12-03T17:02:50.863101Z", - "url": "https://files.pythonhosted.org/packages/09/2b/6c2d363e3d46307251a9d6bf74ec28543805bbcadf56ca729f4a04846914/Django-1.11.17-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "a3609d004f544297259b9467a1429e48d1fe1bda990aeb80744afaccb34aab4a", - "md5": "7ca3b663495a78895f014573f37db606", - "sha256": "a787ee66f4b4cf8ed753661cabcec603989677fa3a107fcb7f15511a44bdb483" - }, - "downloads": -1, - "filename": "Django-1.11.17.tar.gz", - "has_sig": false, - "md5_digest": "7ca3b663495a78895f014573f37db606", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7853439, - "upload_time": "2018-12-03T17:03:07", - "upload_time_iso_8601": "2018-12-03T17:03:07.563036Z", - "url": "https://files.pythonhosted.org/packages/a3/60/9d004f544297259b9467a1429e48d1fe1bda990aeb80744afaccb34aab4a/Django-1.11.17.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.18": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "e0eb6dc122c6d0a82263bd26bebae3cdbafeb99a7281aa1dae57ca1f645a9872", - "md5": "45315e6544510143dfcb066584447ecc", - "sha256": "7ee7d93d407f082e3849c8d10da50ff5b488af37ed1b0066a22dee5f2709ed16" - }, - "downloads": -1, - "filename": "Django-1.11.18-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "45315e6544510143dfcb066584447ecc", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6949226, - "upload_time": "2019-01-04T14:10:44", - "upload_time_iso_8601": "2019-01-04T14:10:44.776909Z", - "url": "https://files.pythonhosted.org/packages/e0/eb/6dc122c6d0a82263bd26bebae3cdbafeb99a7281aa1dae57ca1f645a9872/Django-1.11.18-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "90847981bdfcfa80fe81df5325899f9fc1cbebce1fbe4fac092a32dca00d0ab2", - "md5": "ef734560a81a8c0eb535e7a46205bd72", - "sha256": "73cca1dac154e749b39cc91a54dc876109eb0512a5c6804986495305047066a5" - }, - "downloads": -1, - "filename": "Django-1.11.18.tar.gz", - "has_sig": false, - "md5_digest": "ef734560a81a8c0eb535e7a46205bd72", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7847617, - "upload_time": "2019-01-04T14:10:56", - "upload_time_iso_8601": "2019-01-04T14:10:56.890031Z", - "url": "https://files.pythonhosted.org/packages/90/84/7981bdfcfa80fe81df5325899f9fc1cbebce1fbe4fac092a32dca00d0ab2/Django-1.11.18.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "d1e64ac2f5c9bdc9c82eb48e86a6190e2579be3d10c1afe457993c54cb7d5bc5", - "md5": "84cc396d30b88b549313bfccd8bb2a5a", - "sha256": "cbd3944599086518cdcf3235e90230c8b8a9c5476b20447ee57b313dba14f67a" - }, - "downloads": -1, - "filename": "Django-1.11.2-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "84cc396d30b88b549313bfccd8bb2a5a", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6947046, - "upload_time": "2017-06-01T16:47:48", - "upload_time_iso_8601": "2017-06-01T16:47:48.983642Z", - "url": "https://files.pythonhosted.org/packages/d1/e6/4ac2f5c9bdc9c82eb48e86a6190e2579be3d10c1afe457993c54cb7d5bc5/Django-1.11.2-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "c0314bffd9183066eea645430114419c30b030b599320da8246701b81c6a78d2", - "md5": "f089f1f86d25f2b78f6cf36478d4edd1", - "sha256": "3c5b070482df4f9e5750539dc1824d353729ee423fd410c579b8cd3dea5b0617" - }, - "downloads": -1, - "filename": "Django-1.11.2.tar.gz", - "has_sig": false, - "md5_digest": "f089f1f86d25f2b78f6cf36478d4edd1", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7865109, - "upload_time": "2017-06-01T16:50:13", - "upload_time_iso_8601": "2017-06-01T16:50:13.913846Z", - "url": "https://files.pythonhosted.org/packages/c0/31/4bffd9183066eea645430114419c30b030b599320da8246701b81c6a78d2/Django-1.11.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.20": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "8e1f20bbc601c442d02cc8d9b25a399a18ef573077e3350acdf5da3743ff7da1", - "md5": "f597fa6d128cda70649a2147819d2b5e", - "sha256": "0a73696e0ac71ee6177103df984f9c1e07cd297f080f8ec4dc7c6f3fb74395b5" - }, - "downloads": -1, - "filename": "Django-1.11.20-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "f597fa6d128cda70649a2147819d2b5e", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6949426, - "upload_time": "2019-02-11T15:10:36", - "upload_time_iso_8601": "2019-02-11T15:10:36.880499Z", - "url": "https://files.pythonhosted.org/packages/8e/1f/20bbc601c442d02cc8d9b25a399a18ef573077e3350acdf5da3743ff7da1/Django-1.11.20-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "992a6cb6fdae67a101e19cd02b1af75131eee51b8dcd0cc22c9cfdd2270b5715", - "md5": "096091c29c00f36cce4356054119b702", - "sha256": "43a99da08fee329480d27860d68279945b7d8bf7b537388ee2c8938c709b2041" - }, - "downloads": -1, - "filename": "Django-1.11.20.tar.gz", - "has_sig": false, - "md5_digest": "096091c29c00f36cce4356054119b702", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7846576, - "upload_time": "2019-02-11T15:10:54", - "upload_time_iso_8601": "2019-02-11T15:10:54.841634Z", - "url": "https://files.pythonhosted.org/packages/99/2a/6cb6fdae67a101e19cd02b1af75131eee51b8dcd0cc22c9cfdd2270b5715/Django-1.11.20.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.21": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "a2849f66e359ba8e63cf9b54f6815ed55188dda43cd1cc951a8bb95542dee956", - "md5": "7d3c49ebed0c18280675b879b57d75bc", - "sha256": "aae1b776d78cc3f492afda405b9b9d322b27761442997456c158687d7a0610a1" - }, - "downloads": -1, - "filename": "Django-1.11.21-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "7d3c49ebed0c18280675b879b57d75bc", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6949509, - "upload_time": "2019-06-03T10:10:58", - "upload_time_iso_8601": "2019-06-03T10:10:58.728571Z", - "url": "https://files.pythonhosted.org/packages/a2/84/9f66e359ba8e63cf9b54f6815ed55188dda43cd1cc951a8bb95542dee956/Django-1.11.21-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "51153498e4b783cda329a823f3d474c857d91b2fdfc1c739bb7f616f17d748ca", - "md5": "9a659a9dd9f5900fe75c7fbc4ce1b6a3", - "sha256": "ba723e524facffa2a9d8c2e9116db871e16b9207e648e1d3e4af8aae1167b029" - }, - "downloads": -1, - "filename": "Django-1.11.21.tar.gz", - "has_sig": false, - "md5_digest": "9a659a9dd9f5900fe75c7fbc4ce1b6a3", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7847136, - "upload_time": "2019-06-03T10:11:17", - "upload_time_iso_8601": "2019-06-03T10:11:17.443138Z", - "url": "https://files.pythonhosted.org/packages/51/15/3498e4b783cda329a823f3d474c857d91b2fdfc1c739bb7f616f17d748ca/Django-1.11.21.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.22": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "7022237da71dc112f2bba335c18380bc403fba430c44cc4da088824e77652738", - "md5": "53b7bb6b6257f71772301cd3c500b5b8", - "sha256": "94395804ad80f68d66090a74d68ff2583b43333e1785a026c2aa10cf161642a6" - }, - "downloads": -1, - "filename": "Django-1.11.22-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "53b7bb6b6257f71772301cd3c500b5b8", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6949503, - "upload_time": "2019-07-01T07:19:13", - "upload_time_iso_8601": "2019-07-01T07:19:13.174236Z", - "url": "https://files.pythonhosted.org/packages/70/22/237da71dc112f2bba335c18380bc403fba430c44cc4da088824e77652738/Django-1.11.22-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "2f967d56b16388e8686ef8e2cb330204f247a90e6f008849dad7ce61c9c21c84", - "md5": "d3a20b27a0cfb562bac46a06605b29af", - "sha256": "830d5d40a1705089502bba70605ab3246831440ffc16d1501dfeeef5f4b9c845" - }, - "downloads": -1, - "filename": "Django-1.11.22.tar.gz", - "has_sig": false, - "md5_digest": "d3a20b27a0cfb562bac46a06605b29af", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7972885, - "upload_time": "2019-07-01T07:19:29", - "upload_time_iso_8601": "2019-07-01T07:19:29.492277Z", - "url": "https://files.pythonhosted.org/packages/2f/96/7d56b16388e8686ef8e2cb330204f247a90e6f008849dad7ce61c9c21c84/Django-1.11.22.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.23": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "61cbe3c6bfccdf23c48dd4ce014b96178aa048b9450739eaa5f11d4d23d9d5d6", - "md5": "f15b59e066032c23e28b2e86b9914a79", - "sha256": "c85b8c95366e187ca0581d45a6e508107ca4bd38cb45c24aa09d3572074c523d" - }, - "downloads": -1, - "filename": "Django-1.11.23-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "f15b59e066032c23e28b2e86b9914a79", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6949572, - "upload_time": "2019-08-01T09:04:24", - "upload_time_iso_8601": "2019-08-01T09:04:24.934545Z", - "url": "https://files.pythonhosted.org/packages/61/cb/e3c6bfccdf23c48dd4ce014b96178aa048b9450739eaa5f11d4d23d9d5d6/Django-1.11.23-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "570e722252684d409626d8e6f1aeb4790e7ea06d115498c98d260281022468b5", - "md5": "ded95be58e57d0fa65b03e36b1566265", - "sha256": "52a66d7f8b036d02da0a4472359e8be1727424fc1e4b4f5c684ef97de7b569e1" - }, - "downloads": -1, - "filename": "Django-1.11.23.tar.gz", - "has_sig": false, - "md5_digest": "ded95be58e57d0fa65b03e36b1566265", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7849738, - "upload_time": "2019-08-01T09:04:43", - "upload_time_iso_8601": "2019-08-01T09:04:43.456757Z", - "url": "https://files.pythonhosted.org/packages/57/0e/722252684d409626d8e6f1aeb4790e7ea06d115498c98d260281022468b5/Django-1.11.23.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.24": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "3ac1dbd77256695f4b4e12b5d2c917a35963db11ce5df19c8ea6cd136b2ed54d", - "md5": "f4a9dc38b1a82d1387fe64705092db14", - "sha256": "ffd89b89a2ee860ee521f054225044f52676825be4b61168d2842d44fcf457d3" - }, - "downloads": -1, - "filename": "Django-1.11.24-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "f4a9dc38b1a82d1387fe64705092db14", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6949576, - "upload_time": "2019-09-02T07:18:28", - "upload_time_iso_8601": "2019-09-02T07:18:28.734799Z", - "url": "https://files.pythonhosted.org/packages/3a/c1/dbd77256695f4b4e12b5d2c917a35963db11ce5df19c8ea6cd136b2ed54d/Django-1.11.24-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "a67904502b30769680fe2538a046125afabae5c46399ffc4e18fb44abfa338c6", - "md5": "e5aec60146bf7c4d929d54a0dedb8e56", - "sha256": "215c27453f775b6b1add83a185f76c2e2ab711d17786a6704bd62eabd93f89e3" - }, - "downloads": -1, - "filename": "Django-1.11.24.tar.gz", - "has_sig": false, - "md5_digest": "e5aec60146bf7c4d929d54a0dedb8e56", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7976020, - "upload_time": "2019-09-02T07:18:46", - "upload_time_iso_8601": "2019-09-02T07:18:46.393062Z", - "url": "https://files.pythonhosted.org/packages/a6/79/04502b30769680fe2538a046125afabae5c46399ffc4e18fb44abfa338c6/Django-1.11.24.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.25": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "3ced06a81a65fa00f766f2dbda94d09e946aa65c23e6d7ca3532984627a6c75a", - "md5": "885c067d43c1a2e9191712882b6e7ef0", - "sha256": "2f31cdaaeffdb1728614c4ede5f7101c5770aa90b471c1299d4a203bf00f3b05" - }, - "downloads": -1, - "filename": "Django-1.11.25-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "885c067d43c1a2e9191712882b6e7ef0", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6949587, - "upload_time": "2019-10-01T08:36:31", - "upload_time_iso_8601": "2019-10-01T08:36:31.683239Z", - "url": "https://files.pythonhosted.org/packages/3c/ed/06a81a65fa00f766f2dbda94d09e946aa65c23e6d7ca3532984627a6c75a/Django-1.11.25-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "8d6eca1eaf0a03e7921b020e9a8bd2e9ec441f6d958050be8571951e5556cf77", - "md5": "48f8657c570e39adab89e7cd09699c67", - "sha256": "5314e8586285d532b7aa5c6d763b0248d9a977a37efec86d30f0212b82e8ef66" - }, - "downloads": -1, - "filename": "Django-1.11.25.tar.gz", - "has_sig": false, - "md5_digest": "48f8657c570e39adab89e7cd09699c67", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7845344, - "upload_time": "2019-10-01T08:36:51", - "upload_time_iso_8601": "2019-10-01T08:36:51.238785Z", - "url": "https://files.pythonhosted.org/packages/8d/6e/ca1eaf0a03e7921b020e9a8bd2e9ec441f6d958050be8571951e5556cf77/Django-1.11.25.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.26": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "cf19632a613bc37bbf890f9323ba09374ce9af1d70bb4cba7ff4d3e5e0991b47", - "md5": "dcba8ced9c278e6ddeca667ad4b806ea", - "sha256": "83615ecf738ed710e90ca0dff719b269cc93128c5f0f361370d4e52f42209d16" - }, - "downloads": -1, - "filename": "Django-1.11.26-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "dcba8ced9c278e6ddeca667ad4b806ea", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6949593, - "upload_time": "2019-11-04T08:33:08", - "upload_time_iso_8601": "2019-11-04T08:33:08.180128Z", - "url": "https://files.pythonhosted.org/packages/cf/19/632a613bc37bbf890f9323ba09374ce9af1d70bb4cba7ff4d3e5e0991b47/Django-1.11.26-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "1ecc3226f5b935841bf8c1a1387cc0cba9770f0c12df9aab4460a63036765b23", - "md5": "858e5417a10ce565a15d6e4a2ea0ee37", - "sha256": "861db7f82436ab43e1411832ed8dca81fc5fc0f7c2039c7e07a080a63092fb44" - }, - "downloads": -1, - "filename": "Django-1.11.26.tar.gz", - "has_sig": false, - "md5_digest": "858e5417a10ce565a15d6e4a2ea0ee37", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7976282, - "upload_time": "2019-11-04T08:33:25", - "upload_time_iso_8601": "2019-11-04T08:33:25.858722Z", - "url": "https://files.pythonhosted.org/packages/1e/cc/3226f5b935841bf8c1a1387cc0cba9770f0c12df9aab4460a63036765b23/Django-1.11.26.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.27": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "a62bfbd71ae0980c899c0df70779d36c5897a6b56518eb5942ddd53b0b969b30", - "md5": "f18cd55578581166080cc7e04dd626cc", - "sha256": "372faee5b93c92f19e9d65f52b278a1b689d3e3b4a7d9d30db73a78ebc729770" - }, - "downloads": -1, - "filename": "Django-1.11.27-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "f18cd55578581166080cc7e04dd626cc", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6949848, - "upload_time": "2019-12-18T08:59:02", - "upload_time_iso_8601": "2019-12-18T08:59:02.278881Z", - "url": "https://files.pythonhosted.org/packages/a6/2b/fbd71ae0980c899c0df70779d36c5897a6b56518eb5942ddd53b0b969b30/Django-1.11.27-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "8a26d4dd19084cb1a8d4c7b566024e3fd80b25aa50bccef3a8e5db1d300d4fba", - "md5": "e75626654c7d92ff8bafa2a36d137372", - "sha256": "20111383869ad1b11400c94b0c19d4ab12975316cd058eabd17452e0546169b8" - }, - "downloads": -1, - "filename": "Django-1.11.27.tar.gz", - "has_sig": false, - "md5_digest": "e75626654c7d92ff8bafa2a36d137372", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7976980, - "upload_time": "2019-12-18T08:59:19", - "upload_time_iso_8601": "2019-12-18T08:59:19.797864Z", - "url": "https://files.pythonhosted.org/packages/8a/26/d4dd19084cb1a8d4c7b566024e3fd80b25aa50bccef3a8e5db1d300d4fba/Django-1.11.27.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.28": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "6acdd14d70ad55850e3bd656eb1cc235730e855120c18b882b7cbd383216723d", - "md5": "103fe7af9f88d6c621026b8f9d284d1b", - "sha256": "a3b01cdff845a43830d7ccacff55e0b8ff08305a4cbf894517a686e53ba3ad2d" - }, - "downloads": -1, - "filename": "Django-1.11.28-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "103fe7af9f88d6c621026b8f9d284d1b", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6949861, - "upload_time": "2020-02-03T09:50:36", - "upload_time_iso_8601": "2020-02-03T09:50:36.736416Z", - "url": "https://files.pythonhosted.org/packages/6a/cd/d14d70ad55850e3bd656eb1cc235730e855120c18b882b7cbd383216723d/Django-1.11.28-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "52bee4bfd6db49d6b94112668ef3dcfb027c8717729a8daebf5c9fd19a4c5115", - "md5": "8a21a5148aece7f6110d6ff3a9f57652", - "sha256": "b33ce35f47f745fea6b5aa3cf3f4241069803a3712d423ac748bd673a39741eb" - }, - "downloads": -1, - "filename": "Django-1.11.28.tar.gz", - "has_sig": false, - "md5_digest": "8a21a5148aece7f6110d6ff3a9f57652", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7852525, - "upload_time": "2020-02-03T09:50:51", - "upload_time_iso_8601": "2020-02-03T09:50:51.569042Z", - "url": "https://files.pythonhosted.org/packages/52/be/e4bfd6db49d6b94112668ef3dcfb027c8717729a8daebf5c9fd19a4c5115/Django-1.11.28.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.29": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "4949178daa8725d29c475216259eb19e90b2aa0b8c0431af8c7e9b490ae6481d", - "md5": "d0ca2dbbdcca6e49536d527fbc32e4ea", - "sha256": "014e3392058d94f40569206a24523ce254d55ad2f9f46c6550b0fe2e4f94cf3f" - }, - "downloads": -1, - "filename": "Django-1.11.29-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "d0ca2dbbdcca6e49536d527fbc32e4ea", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6949937, - "upload_time": "2020-03-04T09:31:46", - "upload_time_iso_8601": "2020-03-04T09:31:46.566802Z", - "url": "https://files.pythonhosted.org/packages/49/49/178daa8725d29c475216259eb19e90b2aa0b8c0431af8c7e9b490ae6481d/Django-1.11.29-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "68ab2278a4a9404fac661be1be9627f11336613149e07fc4df0b6e929cc9f300", - "md5": "e725953dfc63ea9e3b5b0898a8027bd7", - "sha256": "4200aefb6678019a0acf0005cd14cfce3a5e6b9b90d06145fcdd2e474ad4329c" - }, - "downloads": -1, - "filename": "Django-1.11.29.tar.gz", - "has_sig": false, - "md5_digest": "e725953dfc63ea9e3b5b0898a8027bd7", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7977916, - "upload_time": "2020-03-04T09:32:02", - "upload_time_iso_8601": "2020-03-04T09:32:02.383798Z", - "url": "https://files.pythonhosted.org/packages/68/ab/2278a4a9404fac661be1be9627f11336613149e07fc4df0b6e929cc9f300/Django-1.11.29.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "fecaa7b35a0f5088f26b1ef3c7add57161a7d387a4cbd30db01c1091aa87e207", - "md5": "51588493d8557726a7dd75a3f5ffb74b", - "sha256": "c69e0c0416f2376b677830304d4c5fa8793b9c815af77be659a3c50d1f46c2e6" - }, - "downloads": -1, - "filename": "Django-1.11.3-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "51588493d8557726a7dd75a3f5ffb74b", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6947525, - "upload_time": "2017-07-01T23:24:53", - "upload_time_iso_8601": "2017-07-01T23:24:53.311560Z", - "url": "https://files.pythonhosted.org/packages/fe/ca/a7b35a0f5088f26b1ef3c7add57161a7d387a4cbd30db01c1091aa87e207/Django-1.11.3-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "4523a5dbb0cfee731549032034d2666cabba47a447f811f706cff82fd1947efc", - "md5": "fe23bb9ae450698451ac82cb636bb882", - "sha256": "9ef9de0a957245ed3a29c4162ed2fd493252ca249a755f9e2b4a9be82caf8f6b" - }, - "downloads": -1, - "filename": "Django-1.11.3.tar.gz", - "has_sig": false, - "md5_digest": "fe23bb9ae450698451ac82cb636bb882", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7872014, - "upload_time": "2017-07-01T23:25:09", - "upload_time_iso_8601": "2017-07-01T23:25:09.072231Z", - "url": "https://files.pythonhosted.org/packages/45/23/a5dbb0cfee731549032034d2666cabba47a447f811f706cff82fd1947efc/Django-1.11.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "fcfb01e0084061c50f1160c2db5565ff1c3d8d76f2a76f67cd282835ee64e04a", - "md5": "71cf96f790b1e729c8c1a95304971341", - "sha256": "6fd30e05dc9af265f7d7d10cfb0efa013e6236db0853c9f47c74c585587c5a57" - }, - "downloads": -1, - "filename": "Django-1.11.4-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "71cf96f790b1e729c8c1a95304971341", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6947863, - "upload_time": "2017-08-01T12:24:53", - "upload_time_iso_8601": "2017-08-01T12:24:53.575224Z", - "url": "https://files.pythonhosted.org/packages/fc/fb/01e0084061c50f1160c2db5565ff1c3d8d76f2a76f67cd282835ee64e04a/Django-1.11.4-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "cea03f59e798179b23f813250c79ee31a346aeecd4fa09ae05b639647086f5f3", - "md5": "c851d892cd5ad3a90808703c4f36e3fe", - "sha256": "abe86e67dda9897a1536a727ed57dbefb5a42b41943be3b116fe3edab4c07bb2" - }, - "downloads": -1, - "filename": "Django-1.11.4.tar.gz", - "has_sig": false, - "md5_digest": "c851d892cd5ad3a90808703c4f36e3fe", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7870752, - "upload_time": "2017-08-01T12:25:19", - "upload_time_iso_8601": "2017-08-01T12:25:19.627007Z", - "url": "https://files.pythonhosted.org/packages/ce/a0/3f59e798179b23f813250c79ee31a346aeecd4fa09ae05b639647086f5f3/Django-1.11.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.5": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "182db477232dd619d81766064cd07ba5b35e956ff8a8c5c5d41754e0392b96e3", - "md5": "6380d5fb6ede4847dc186a09ccc7b538", - "sha256": "89162f70a74aac62a53f975128faba6099a7ef2c9d8140a41ae9d6210bda05cd" - }, - "downloads": -1, - "filename": "Django-1.11.5-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "6380d5fb6ede4847dc186a09ccc7b538", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6948177, - "upload_time": "2017-09-05T15:18:53", - "upload_time_iso_8601": "2017-09-05T15:18:53.403268Z", - "url": "https://files.pythonhosted.org/packages/18/2d/b477232dd619d81766064cd07ba5b35e956ff8a8c5c5d41754e0392b96e3/Django-1.11.5-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "069f7f07816842ad8020d3bdcfbedc568314e0739bc3de435bc034874b6f3e39", - "md5": "8cef0d42aabacbc414ec4fbbb6056f3c", - "sha256": "1836878162dfdf865492bacfdff0321e4ee8f1e7d51d93192546000b54982b29" - }, - "downloads": -1, - "filename": "Django-1.11.5.tar.gz", - "has_sig": false, - "md5_digest": "8cef0d42aabacbc414ec4fbbb6056f3c", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7875054, - "upload_time": "2017-09-05T15:19:03", - "upload_time_iso_8601": "2017-09-05T15:19:03.327101Z", - "url": "https://files.pythonhosted.org/packages/06/9f/7f07816842ad8020d3bdcfbedc568314e0739bc3de435bc034874b6f3e39/Django-1.11.5.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.6": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "8233f9d2871f3aed5062661711bf91b3ebb03daa52cc0e1c37925f3e0c4508c5", - "md5": "89322a57ced871be6e794a9a63a897a2", - "sha256": "7ab6a9c798a5f9f359ee6da3677211f883fb02ef32cebe9b29751eb7a871febf" - }, - "downloads": -1, - "filename": "Django-1.11.6-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "89322a57ced871be6e794a9a63a897a2", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6948250, - "upload_time": "2017-10-05T18:21:24", - "upload_time_iso_8601": "2017-10-05T18:21:24.388540Z", - "url": "https://files.pythonhosted.org/packages/82/33/f9d2871f3aed5062661711bf91b3ebb03daa52cc0e1c37925f3e0c4508c5/Django-1.11.6-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "1326f3841e00663027ba7cf7ce7ba2cabb682a83cf0629bef013d70bebefa69d", - "md5": "68f6981fbf05549b2014da483d52d1ba", - "sha256": "c3b42ca1efa1c0a129a9e863134cc3fe705c651dea3a04a7998019e522af0c60" - }, - "downloads": -1, - "filename": "Django-1.11.6.tar.gz", - "has_sig": false, - "md5_digest": "68f6981fbf05549b2014da483d52d1ba", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7874450, - "upload_time": "2017-10-05T18:24:42", - "upload_time_iso_8601": "2017-10-05T18:24:42.635754Z", - "url": "https://files.pythonhosted.org/packages/13/26/f3841e00663027ba7cf7ce7ba2cabb682a83cf0629bef013d70bebefa69d/Django-1.11.6.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.7": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "15d8b17afdcd527026d2f1acd30ac33406e6b22c0f573a3c14b2d9e0bd7df945", - "md5": "d53ee997bba37fd2ccb94888fdd02dcc", - "sha256": "75ce405d60f092f6adf904058d023eeea0e6d380f8d9c36134bac73da736023d" - }, - "downloads": -1, - "filename": "Django-1.11.7-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "d53ee997bba37fd2ccb94888fdd02dcc", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6948299, - "upload_time": "2017-11-02T01:26:27", - "upload_time_iso_8601": "2017-11-02T01:26:27.795201Z", - "url": "https://files.pythonhosted.org/packages/15/d8/b17afdcd527026d2f1acd30ac33406e6b22c0f573a3c14b2d9e0bd7df945/Django-1.11.7-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "c3a858bf5ce8f54b8fd9aa0de10288600cf71c6b779d519e301f4b0de8c06259", - "md5": "cd5cf6f9e9d2c7b05747ee7d4154b131", - "sha256": "8918e392530d8fc6965a56af6504229e7924c27265893f3949aa0529cd1d4b99" - }, - "downloads": -1, - "filename": "Django-1.11.7.tar.gz", - "has_sig": false, - "md5_digest": "cd5cf6f9e9d2c7b05747ee7d4154b131", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7877132, - "upload_time": "2017-11-02T01:26:35", - "upload_time_iso_8601": "2017-11-02T01:26:35.309616Z", - "url": "https://files.pythonhosted.org/packages/c3/a8/58bf5ce8f54b8fd9aa0de10288600cf71c6b779d519e301f4b0de8c06259/Django-1.11.7.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.8": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "7e365266e0c51ee9b953d60ea8ea1fea10e268b1368f9c0ad08e2ff76ee9c1b5", - "md5": "0f6557754e9347aeccf0f379b15a1264", - "sha256": "fad46f44f6f4de66aacaa92e7753dbc4fe3ae834aa2daffaca0bf16c64798186" - }, - "downloads": -1, - "filename": "Django-1.11.8-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "0f6557754e9347aeccf0f379b15a1264", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6949370, - "upload_time": "2017-12-02T14:20:51", - "upload_time_iso_8601": "2017-12-02T14:20:51.619888Z", - "url": "https://files.pythonhosted.org/packages/7e/36/5266e0c51ee9b953d60ea8ea1fea10e268b1368f9c0ad08e2ff76ee9c1b5/Django-1.11.8-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "b09eb1939fc389c091f17e725a7bd11a161db8fea8d632af708cba3b4e2deb94", - "md5": "e8b68d44b87a3de36e13547ec2722af2", - "sha256": "fed3e79bb5a3a8d5eb054c7a1ec1de229ef3f43335a67821cc3e489e9582f711" - }, - "downloads": -1, - "filename": "Django-1.11.8.tar.gz", - "has_sig": false, - "md5_digest": "e8b68d44b87a3de36e13547ec2722af2", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7879685, - "upload_time": "2017-12-02T14:21:08", - "upload_time_iso_8601": "2017-12-02T14:21:08.831473Z", - "url": "https://files.pythonhosted.org/packages/b0/9e/b1939fc389c091f17e725a7bd11a161db8fea8d632af708cba3b4e2deb94/Django-1.11.8.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11.9": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "c8a6291039f0ce4b9818e0399359866337e6dfe9c0e23d8d94f00e657edbfcd8", - "md5": "005d40c630a677a3527fc72ff7da7179", - "sha256": "90952c46d2b7b042db00e98b05f5dd97a5775822948d46fd82ff074d8ac75853" - }, - "downloads": -1, - "filename": "Django-1.11.9-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "005d40c630a677a3527fc72ff7da7179", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6949531, - "upload_time": "2018-01-02T01:01:55", - "upload_time_iso_8601": "2018-01-02T01:01:55.893297Z", - "url": "https://files.pythonhosted.org/packages/c8/a6/291039f0ce4b9818e0399359866337e6dfe9c0e23d8d94f00e657edbfcd8/Django-1.11.9-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "31b78ad017c3e81635bb12c6d41b56fdbf4bb52eb30aea7f45cfeea61607bab8", - "md5": "08ad028fc50ee961dea35e1e1f657b65", - "sha256": "353d129f22e1d24980d6061666f435781141c2dfd852f14ffc8a670175821034" - }, - "downloads": -1, - "filename": "Django-1.11.9.tar.gz", - "has_sig": false, - "md5_digest": "08ad028fc50ee961dea35e1e1f657b65", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7879870, - "upload_time": "2018-01-02T01:02:08", - "upload_time_iso_8601": "2018-01-02T01:02:08.883938Z", - "url": "https://files.pythonhosted.org/packages/31/b7/8ad017c3e81635bb12c6d41b56fdbf4bb52eb30aea7f45cfeea61607bab8/Django-1.11.9.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.11a1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "327e884dfbac0d640b428bd800aa4204ed69a39ec155eacf8d6586c8e77f91c6", - "md5": "29b521397678e62525aedc8da4c03b09", - "sha256": "edb0cb0ae9120a21ec729de7138106acd1f0737daa63b9cd8e2d739a7b6198dc" - }, - "downloads": -1, - "filename": "Django-1.11a1-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "29b521397678e62525aedc8da4c03b09", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6870097, - "upload_time": "2017-01-18T01:01:35", - "upload_time_iso_8601": "2017-01-18T01:01:35.823808Z", - "url": "https://files.pythonhosted.org/packages/32/7e/884dfbac0d640b428bd800aa4204ed69a39ec155eacf8d6586c8e77f91c6/Django-1.11a1-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - } - ], - "1.11b1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "41c168dd27946b03a3d756b0ff665baad25aee1f59918891d86ab76764209208", - "md5": "4f6aa7c92e80488096f8b2c6b25c879d", - "sha256": "fbc7ffaa45a4a67cb45f77dbd94e8eceecebe1d0959fe9c665dfbf28b41899e6" - }, - "downloads": -1, - "filename": "Django-1.11b1-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "4f6aa7c92e80488096f8b2c6b25c879d", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6869998, - "upload_time": "2017-02-20T23:21:50", - "upload_time_iso_8601": "2017-02-20T23:21:50.835517Z", - "url": "https://files.pythonhosted.org/packages/41/c1/68dd27946b03a3d756b0ff665baad25aee1f59918891d86ab76764209208/Django-1.11b1-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - } - ], - "1.11rc1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "9df5e437d325edcf762854f58174d53a06dcaffa211388ba8f0c4726f6b0b1d4", - "md5": "b7c9d578e71d8e98450e615cd23bc2d7", - "sha256": "621245da55de87d9cd8527bb3b354d60643d420347a75adb3ebe07750f4c2dae" - }, - "downloads": -1, - "filename": "Django-1.11rc1-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "b7c9d578e71d8e98450e615cd23bc2d7", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6870876, - "upload_time": "2017-03-21T22:55:53", - "upload_time_iso_8601": "2017-03-21T22:55:53.651688Z", - "url": "https://files.pythonhosted.org/packages/9d/f5/e437d325edcf762854f58174d53a06dcaffa211388ba8f0c4726f6b0b1d4/Django-1.11rc1-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "2507b915b82e345c5189ba999dcfdf96d8de7026166c5e16f0e7030ffbe5f6cd", - "md5": "080665995b322de22a12d0596c081856", - "sha256": "fde8e20117d942e7c1a3c23bb00ab6caf38aefc31c3ca28e961c8b67bd576f2e" - }, - "downloads": -1, - "filename": "Django-1.11rc1.tar.gz", - "has_sig": false, - "md5_digest": "080665995b322de22a12d0596c081856", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7796930, - "upload_time": "2017-03-21T22:56:03", - "upload_time_iso_8601": "2017-03-21T22:56:03.250177Z", - "url": "https://files.pythonhosted.org/packages/25/07/b915b82e345c5189ba999dcfdf96d8de7026166c5e16f0e7030ffbe5f6cd/Django-1.11rc1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "8ed7c31ff2b5564090955c9c67aa41c7d920f31a3fac019205747835b89dc5bd", - "md5": "98fa833fdabcdd78d00245aead66c174", - "sha256": "3d4b18dfa0ef181ef85fea7be98a763906f767bc320694c98280c52c5af0bd83" - }, - "downloads": -1, - "filename": "Django-1.2.tar.gz", - "has_sig": false, - "md5_digest": "98fa833fdabcdd78d00245aead66c174", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6246202, - "upload_time": "2010-05-17T20:04:28", - "upload_time_iso_8601": "2010-05-17T20:04:28.174094Z", - "url": "https://files.pythonhosted.org/packages/8e/d7/c31ff2b5564090955c9c67aa41c7d920f31a3fac019205747835b89dc5bd/Django-1.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.2.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "abf7974021fdfd71a419526704e404447aaf3b0f303a6cdfde87478a09dd3a49", - "md5": "2351efb20f6b7b5d9ce80fa4cb1bd9ca", - "sha256": "eaa29f2344568cc871c4517a348de0d5c39fbd055b4c998cd4a80601bb51e7b9" - }, - "downloads": -1, - "filename": "Django-1.2.1.tar.gz", - "has_sig": false, - "md5_digest": "2351efb20f6b7b5d9ce80fa4cb1bd9ca", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6248006, - "upload_time": "2010-05-24T21:19:28", - "upload_time_iso_8601": "2010-05-24T21:19:28.440044Z", - "url": "https://files.pythonhosted.org/packages/ab/f7/974021fdfd71a419526704e404447aaf3b0f303a6cdfde87478a09dd3a49/Django-1.2.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.2.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "c911cfdc58670b7f01ae94cd2b256c07892109fc5fffd0c5a613393891426cbe", - "md5": "9cdbf79a31988ace9ef2ab4ede890136", - "sha256": "803831781dbe9802de079c6735b7f5ecd7edf2ea8d91cb031e9b29c720d3d1ba" - }, - "downloads": -1, - "filename": "Django-1.2.2.tar.gz", - "has_sig": false, - "md5_digest": "9cdbf79a31988ace9ef2ab4ede890136", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6304356, - "upload_time": "2010-09-09T02:41:25", - "upload_time_iso_8601": "2010-09-09T02:41:25.682749Z", - "url": "https://files.pythonhosted.org/packages/c9/11/cfdc58670b7f01ae94cd2b256c07892109fc5fffd0c5a613393891426cbe/Django-1.2.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.2.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "3b3bf98695d298c983c37aa5153014acca0d3714bbc08ce6f55cffa4b00b6fb4", - "md5": "10bfb5831bcb4d3b1e6298d0e41d6603", - "sha256": "cb830f6038b78037647150d977f6cd5cf2bfd731f1788ecf8758a03c213a0f84" - }, - "downloads": -1, - "filename": "Django-1.2.3.tar.gz", - "has_sig": false, - "md5_digest": "10bfb5831bcb4d3b1e6298d0e41d6603", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6306760, - "upload_time": "2010-09-11T08:50:39", - "upload_time_iso_8601": "2010-09-11T08:50:39.445072Z", - "url": "https://files.pythonhosted.org/packages/3b/3b/f98695d298c983c37aa5153014acca0d3714bbc08ce6f55cffa4b00b6fb4/Django-1.2.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.2.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "02794135356ed5f7266080fa6ca34015bc58eab25a0b0773eeaeb32a97422ef4", - "md5": "b0e67d3d6447f7eb1ce6392b9465a183", - "sha256": "0f06cccd4ca92173b958dd80edff35035888f15554be425e5d6d55c7f94a8381" - }, - "downloads": -1, - "filename": "Django-1.2.4.tar.gz", - "has_sig": false, - "md5_digest": "b0e67d3d6447f7eb1ce6392b9465a183", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6357270, - "upload_time": "2010-12-23T05:15:19", - "upload_time_iso_8601": "2010-12-23T05:15:19.624230Z", - "url": "https://files.pythonhosted.org/packages/02/79/4135356ed5f7266080fa6ca34015bc58eab25a0b0773eeaeb32a97422ef4/Django-1.2.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.2.5": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "0325f57cca579f5cc621501e69d73e7baa155f32203e13adf6f67fc322b37f06", - "md5": "e031ea3d00996035e49e4bfa86e07c40", - "sha256": "649387248296386b589c4a8bf91d34590b43f93b6ebfe6cefbea0ddf4641ccd6" - }, - "downloads": -1, - "filename": "Django-1.2.5.tar.gz", - "has_sig": false, - "md5_digest": "e031ea3d00996035e49e4bfa86e07c40", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6379313, - "upload_time": "2011-02-09T04:08:37", - "upload_time_iso_8601": "2011-02-09T04:08:37.395797Z", - "url": "https://files.pythonhosted.org/packages/03/25/f57cca579f5cc621501e69d73e7baa155f32203e13adf6f67fc322b37f06/Django-1.2.5.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.2.6": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "19a1569de723efc22ea15498c34112cd00025c3277c131298bb8a1e9d3e6071a", - "md5": "bff9fc7d871c0b5e6ce1a7babd16847b", - "sha256": "a554a902f3a170239a982f750a973013c01fe65206641bd8a658726081f670ed" - }, - "downloads": -1, - "filename": "Django-1.2.6.tar.gz", - "has_sig": false, - "md5_digest": "bff9fc7d871c0b5e6ce1a7babd16847b", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6399890, - "upload_time": "2011-09-10T03:42:09", - "upload_time_iso_8601": "2011-09-10T03:42:09.759203Z", - "url": "https://files.pythonhosted.org/packages/19/a1/569de723efc22ea15498c34112cd00025c3277c131298bb8a1e9d3e6071a/Django-1.2.6.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.2.7": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "116a64ded77a5134cde391ddd6f034977c21e3024012f4599e8f643664eef394", - "md5": "902fe294a2f7b16e5e1dee42d458c2ba", - "sha256": "912b6b9223e2eaa64912f01e0e3b0cddc1d16007a2a6f30b206a96a8c901298a" - }, - "downloads": -1, - "filename": "Django-1.2.7.tar.gz", - "has_sig": false, - "md5_digest": "902fe294a2f7b16e5e1dee42d458c2ba", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6400234, - "upload_time": "2011-09-11T03:05:18", - "upload_time_iso_8601": "2011-09-11T03:05:18.996286Z", - "url": "https://files.pythonhosted.org/packages/11/6a/64ded77a5134cde391ddd6f034977c21e3024012f4599e8f643664eef394/Django-1.2.7.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "f5d56722d3091946734194ffcfe8ef074f63e8acdd1ff51dfcfc87c2c194fd3f", - "md5": "1b8f76e91c27564708649671f329551f", - "sha256": "7aeee5c80002ab81d4ebf5416292949ff46e1448d183a183fe05ff6344771c83" - }, - "downloads": -1, - "filename": "Django-1.3.tar.gz", - "has_sig": false, - "md5_digest": "1b8f76e91c27564708649671f329551f", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6504003, - "upload_time": "2011-03-23T06:09:12", - "upload_time_iso_8601": "2011-03-23T06:09:12.606484Z", - "url": "https://files.pythonhosted.org/packages/f5/d5/6722d3091946734194ffcfe8ef074f63e8acdd1ff51dfcfc87c2c194fd3f/Django-1.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.3.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "a0cde0ebeb3752bf14f0b21cc3e16223dbfe88f26a7c582dc63346fa2ce6bc3b", - "md5": "62d8642fd06b9a0bf8544178f8500767", - "sha256": "af9118c4e8a063deb0b8cda901fcff2b805e7cf496c93fd43507163f3cde156b" - }, - "downloads": -1, - "filename": "Django-1.3.1.tar.gz", - "has_sig": false, - "md5_digest": "62d8642fd06b9a0bf8544178f8500767", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6514564, - "upload_time": "2011-09-10T03:36:21", - "upload_time_iso_8601": "2011-09-10T03:36:21.376323Z", - "url": "https://files.pythonhosted.org/packages/a0/cd/e0ebeb3752bf14f0b21cc3e16223dbfe88f26a7c582dc63346fa2ce6bc3b/Django-1.3.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.3.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "a35f6af1cd40fe029e3502c621ee51a504c2636bde68419385a5f5ff8d24a62e", - "md5": "b8409b8f061e6c7a7dcfbb24403cb863", - "sha256": "72c4080fe30863c4581521ef6f7bfcc12d01e7c55b9ad1935acaa43e466dd764" - }, - "downloads": -1, - "filename": "Django-1.3.2.tar.gz", - "has_sig": false, - "md5_digest": "b8409b8f061e6c7a7dcfbb24403cb863", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6507042, - "upload_time": "2012-07-30T23:02:55", - "upload_time_iso_8601": "2012-07-30T23:02:55.545349Z", - "url": "https://files.pythonhosted.org/packages/a3/5f/6af1cd40fe029e3502c621ee51a504c2636bde68419385a5f5ff8d24a62e/Django-1.3.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.3.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "a7c3bc93caca99846c73bb76a555f84788bdb3f9b1bb622e2bffcf1aa61f36e1", - "md5": "cbdd86f553b26459352e26ae643fd7c1", - "sha256": "8ef44cfd89dee0331018ec56a2ed27dc14ae8d65feb664c10e128b3437cbd46a" - }, - "downloads": -1, - "filename": "Django-1.3.3.tar.gz", - "has_sig": false, - "md5_digest": "cbdd86f553b26459352e26ae643fd7c1", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6507280, - "upload_time": "2012-08-01T22:08:20", - "upload_time_iso_8601": "2012-08-01T22:08:20.092282Z", - "url": "https://files.pythonhosted.org/packages/a7/c3/bc93caca99846c73bb76a555f84788bdb3f9b1bb622e2bffcf1aa61f36e1/Django-1.3.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.3.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "ff15119ccb1a1f0392443f78140c31b981cb2d081afff82fcf368d86619d8cee", - "md5": "9a610a40ee5fcc4ca283fb499e265936", - "sha256": "2626e6b216e1bdef887bd923f00d94d94b4d4e75fc2e336c6f156d842d10a607" - }, - "downloads": -1, - "filename": "Django-1.3.4.tar.gz", - "has_sig": false, - "md5_digest": "9a610a40ee5fcc4ca283fb499e265936", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6507771, - "upload_time": "2013-03-05T22:33:47", - "upload_time_iso_8601": "2013-03-05T22:33:47.875562Z", - "url": "https://files.pythonhosted.org/packages/ff/15/119ccb1a1f0392443f78140c31b981cb2d081afff82fcf368d86619d8cee/Django-1.3.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.3.5": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "5023d05af45d01732968bef5af7d823835e8e4cbc01f687b345b276d124e3025", - "md5": "ec0ae9edb2ed6f9ffa65007110232637", - "sha256": "8e2c00f51f62a59e047e27cbbc03ef1b29aa15ccdca8062fcbca6f5d5ca85ded" - }, - "downloads": -1, - "filename": "Django-1.3.5.tar.gz", - "has_sig": false, - "md5_digest": "ec0ae9edb2ed6f9ffa65007110232637", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6508570, - "upload_time": "2012-12-10T21:39:30", - "upload_time_iso_8601": "2012-12-10T21:39:30.210465Z", - "url": "https://files.pythonhosted.org/packages/50/23/d05af45d01732968bef5af7d823835e8e4cbc01f687b345b276d124e3025/Django-1.3.5.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.3.6": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "c2242164463152fafd6cbb60cb3b42110418deeeaeef504bf81c3fec46f781fe", - "md5": "357dbedf41ba6db990fd4be7c86cd80d", - "sha256": "df0121a4f90795e1b2a374b4df50219df1db203e7960de5e33a6ce31af17878a" - }, - "downloads": -1, - "filename": "Django-1.3.6.tar.gz", - "has_sig": false, - "md5_digest": "357dbedf41ba6db990fd4be7c86cd80d", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6517083, - "upload_time": "2013-02-19T20:32:04", - "upload_time_iso_8601": "2013-02-19T20:32:04.436478Z", - "url": "https://files.pythonhosted.org/packages/c2/24/2164463152fafd6cbb60cb3b42110418deeeaeef504bf81c3fec46f781fe/Django-1.3.6.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.3.7": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "0df1fe4cf23cea3322dd8883a0510538fe916ab2023c648997976798142603ff", - "md5": "f6720daa392d73d4df8847b41909fd43", - "sha256": "ee50f44744e7238cb45429e4121d643c9e9201f9a63aaf646619bad18547fb8a" - }, - "downloads": -1, - "filename": "Django-1.3.7.tar.gz", - "has_sig": false, - "md5_digest": "f6720daa392d73d4df8847b41909fd43", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6514846, - "upload_time": "2013-02-20T20:03:48", - "upload_time_iso_8601": "2013-02-20T20:03:48.025529Z", - "url": "https://files.pythonhosted.org/packages/0d/f1/fe4cf23cea3322dd8883a0510538fe916ab2023c648997976798142603ff/Django-1.3.7.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "9af82a7b9922817be07d53c664a7702fab70e99899466b5956131a70c08606b6", - "md5": "ba8e86198a93c196015df0b363ab1109", - "sha256": "c096bafbea10e7d359bc15eb00a9bf11dbf5201a16d62acfa2de61d5a35488e9" - }, - "downloads": -1, - "filename": "Django-1.4.tar.gz", - "has_sig": false, - "md5_digest": "ba8e86198a93c196015df0b363ab1109", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7632772, - "upload_time": "2012-03-23T18:00:23", - "upload_time_iso_8601": "2012-03-23T18:00:23.693465Z", - "url": "https://files.pythonhosted.org/packages/9a/f8/2a7b9922817be07d53c664a7702fab70e99899466b5956131a70c08606b6/Django-1.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "e63ff3e67d9c2572765ffe4268fc7f9997ce3b02e78fd144733f337d72dabb12", - "md5": "e345268dacff12876ae4e45de0a61b7d", - "sha256": "4d8d20eba350d3d29613cc5a6302d5c23730c7f9e150985bc58b3175b755409b" - }, - "downloads": -1, - "filename": "Django-1.4.1.tar.gz", - "has_sig": false, - "md5_digest": "e345268dacff12876ae4e45de0a61b7d", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7656756, - "upload_time": "2012-07-30T22:48:27", - "upload_time_iso_8601": "2012-07-30T22:48:27.374116Z", - "url": "https://files.pythonhosted.org/packages/e6/3f/f3e67d9c2572765ffe4268fc7f9997ce3b02e78fd144733f337d72dabb12/Django-1.4.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.10": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "bad1523aed5b49f94be8c526fc1fdbffe2edac0bf0579e14d1fc34a5c7c3f0a4", - "md5": "d324aecc37ce5430f548653b8b1509b6", - "sha256": "3d1f083c039fdab1400c32b5406a60891c9dd16f880999c4a53d054742ac29de" - }, - "downloads": -1, - "filename": "Django-1.4.10.tar.gz", - "has_sig": false, - "md5_digest": "d324aecc37ce5430f548653b8b1509b6", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7745002, - "upload_time": "2013-11-06T14:21:25", - "upload_time_iso_8601": "2013-11-06T14:21:25.765558Z", - "url": "https://files.pythonhosted.org/packages/ba/d1/523aed5b49f94be8c526fc1fdbffe2edac0bf0579e14d1fc34a5c7c3f0a4/Django-1.4.10.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.11": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "74056af26eccffa61427ebf44975046369f8c95b21c419ad6ca98308db26e1af", - "md5": "9cd5913b038ebc9582903b2fccbbb54b", - "sha256": "4819d8b37405b33f4f0d156f60918094d566249f52137c5e6e0dbaa12995c201" - }, - "downloads": -1, - "filename": "Django-1.4.11.tar.gz", - "has_sig": false, - "md5_digest": "9cd5913b038ebc9582903b2fccbbb54b", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7752172, - "upload_time": "2014-04-21T22:40:17", - "upload_time_iso_8601": "2014-04-21T22:40:17.318319Z", - "url": "https://files.pythonhosted.org/packages/74/05/6af26eccffa61427ebf44975046369f8c95b21c419ad6ca98308db26e1af/Django-1.4.11.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.12": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "ae2893c30b241a468ebb895ecbc01f45ac1fa42bb04f30d33338696ceb2f22b1", - "md5": "9dc17c3f5409f9a4e662b5550e1c6505", - "sha256": "2b9164dc3b26e077590c6ebb95996aab0e66fe3298113fafe960c4ff7fb53e25" - }, - "downloads": -1, - "filename": "Django-1.4.12.tar.gz", - "has_sig": false, - "md5_digest": "9dc17c3f5409f9a4e662b5550e1c6505", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7752752, - "upload_time": "2014-04-28T20:30:21", - "upload_time_iso_8601": "2014-04-28T20:30:21.125270Z", - "url": "https://files.pythonhosted.org/packages/ae/28/93c30b241a468ebb895ecbc01f45ac1fa42bb04f30d33338696ceb2f22b1/Django-1.4.12.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.13": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "3829044333b3caf5ff1d2394b0e6c6d87df79973f5652ab7deddc46c7eb9d935", - "md5": "9e28e33680f28b027ad67a026a785ea5", - "sha256": "a8fede657378b6862744b19012e7071279b952ecd208fd83227723866068f2c0" - }, - "downloads": -1, - "filename": "Django-1.4.13.tar.gz", - "has_sig": false, - "md5_digest": "9e28e33680f28b027ad67a026a785ea5", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7753532, - "upload_time": "2014-05-14T18:27:42", - "upload_time_iso_8601": "2014-05-14T18:27:42.778726Z", - "url": "https://files.pythonhosted.org/packages/38/29/044333b3caf5ff1d2394b0e6c6d87df79973f5652ab7deddc46c7eb9d935/Django-1.4.13.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.14": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "79fa6f02aa9b46f12701d21ed3cfdd40e7bc40724405ba29ad690cb5b96c85b6", - "md5": "80dc1b9866487afc2ab3f774e29181bc", - "sha256": "81edad81211fd515677a35ab2d40833557649dd650f150baf8416f416b8a6c9c" - }, - "downloads": -1, - "filename": "Django-1.4.14.tar.gz", - "has_sig": false, - "md5_digest": "80dc1b9866487afc2ab3f774e29181bc", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7754876, - "upload_time": "2014-08-20T20:01:35", - "upload_time_iso_8601": "2014-08-20T20:01:35.076618Z", - "url": "https://files.pythonhosted.org/packages/79/fa/6f02aa9b46f12701d21ed3cfdd40e7bc40724405ba29ad690cb5b96c85b6/Django-1.4.14.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.15": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "7acdd862e13993e6bdf0915b1b39eee5df7949c5071caa2dff9d0d3e454e019c", - "md5": "84837da82df11d0e04b7458af8777dc0", - "sha256": "aa57ceb345091c25648b41c98a6f46fffd7884695fa884c7039291177ded14e9" - }, - "downloads": -1, - "filename": "Django-1.4.15.tar.gz", - "has_sig": false, - "md5_digest": "84837da82df11d0e04b7458af8777dc0", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7754429, - "upload_time": "2014-09-02T20:44:06", - "upload_time_iso_8601": "2014-09-02T20:44:06.366428Z", - "url": "https://files.pythonhosted.org/packages/7a/cd/d862e13993e6bdf0915b1b39eee5df7949c5071caa2dff9d0d3e454e019c/Django-1.4.15.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.16": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "995aabac765a76d384fe31d5378bb697c8e8fd1742283790c47107692d1b0f0a", - "md5": "132d088d9e2cbcf43a661a9f05d6e63a", - "sha256": "482315cf32c65ed4a4ee2de257d453430d48ffca9a01b17d984ee0d67354ad12" - }, - "downloads": -1, - "filename": "Django-1.4.16.tar.gz", - "has_sig": false, - "md5_digest": "132d088d9e2cbcf43a661a9f05d6e63a", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7755970, - "upload_time": "2014-10-22T16:37:17", - "upload_time_iso_8601": "2014-10-22T16:37:17.950676Z", - "url": "https://files.pythonhosted.org/packages/99/5a/abac765a76d384fe31d5378bb697c8e8fd1742283790c47107692d1b0f0a/Django-1.4.16.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.17": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "6858a3bf326b57234bbfceb4a84bdddc65d0bedc805b229858503b23faeefc5c", - "md5": "8dd1133b718ce23a0eed3df20d6619c2", - "sha256": "f195879586df5c53b6c964df5fad4e7b675e5fcd36a032d886192ffbdfb41988" - }, - "downloads": -1, - "filename": "Django-1.4.17.tar.gz", - "has_sig": false, - "md5_digest": "8dd1133b718ce23a0eed3df20d6619c2", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7875448, - "upload_time": "2015-01-03T02:20:41", - "upload_time_iso_8601": "2015-01-03T02:20:41.434868Z", - "url": "https://files.pythonhosted.org/packages/68/58/a3bf326b57234bbfceb4a84bdddc65d0bedc805b229858503b23faeefc5c/Django-1.4.17.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.18": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "82b322f00becd75d181db40af8fac2f19add8e4fce441892b39d1245e572c77d", - "md5": "d82b2219052bb47ba0838c2ebd3832ae", - "sha256": "bfd326fe490d03a2a86466fcb1ac335e7d8d58bc498cfe2311b1d751b515521f" - }, - "downloads": -1, - "filename": "Django-1.4.18.tar.gz", - "has_sig": false, - "md5_digest": "d82b2219052bb47ba0838c2ebd3832ae", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7876896, - "upload_time": "2015-01-13T18:54:01", - "upload_time_iso_8601": "2015-01-13T18:54:01.979098Z", - "url": "https://files.pythonhosted.org/packages/82/b3/22f00becd75d181db40af8fac2f19add8e4fce441892b39d1245e572c77d/Django-1.4.18.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.19": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "aff49ff951020928ce0e0acbf61543d9d9ab2de32050639be3b70a0f9e3d7091", - "md5": "8ebda674e81c3886a67eecc72e3f62df", - "sha256": "d75d605e574305e1c2864c392e1454963ead4552477ce14e67e64b9ef9faa1a6" - }, - "downloads": -1, - "filename": "Django-1.4.19.tar.gz", - "has_sig": false, - "md5_digest": "8ebda674e81c3886a67eecc72e3f62df", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7877522, - "upload_time": "2015-01-27T17:11:15", - "upload_time_iso_8601": "2015-01-27T17:11:15.646882Z", - "url": "https://files.pythonhosted.org/packages/af/f4/9ff951020928ce0e0acbf61543d9d9ab2de32050639be3b70a0f9e3d7091/Django-1.4.19.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "ed5611e0c6ce22d7e7048dbabc37202a536a35fbe45bcae661afd16fdefb8903", - "md5": "6ffecdc01ad360e1abdca1015ae0893a", - "sha256": "edfd8733f45bbaa524cee25bcac3080ce28c21242c27227464eae3fa6b3d80e7" - }, - "downloads": -1, - "filename": "Django-1.4.2.tar.gz", - "has_sig": false, - "md5_digest": "6ffecdc01ad360e1abdca1015ae0893a", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7722026, - "upload_time": "2012-10-17T22:18:35", - "upload_time_iso_8601": "2012-10-17T22:18:35.480417Z", - "url": "https://files.pythonhosted.org/packages/ed/56/11e0c6ce22d7e7048dbabc37202a536a35fbe45bcae661afd16fdefb8903/Django-1.4.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.20": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "17a0b4ba4aa5ec80c2c6b351e68d51d29b7b1cb71dc8bc663862e1ea6f5297ad", - "md5": "4a3710921fb51422c9083e52e012ca33", - "sha256": "58ac719464c4c8b13d664ded6770450602528bf4c36f9fd5daabdae8d410ebb1" - }, - "downloads": -1, - "filename": "Django-1.4.20.tar.gz", - "has_sig": false, - "md5_digest": "4a3710921fb51422c9083e52e012ca33", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7877794, - "upload_time": "2015-03-19T00:03:58", - "upload_time_iso_8601": "2015-03-19T00:03:58.032976Z", - "url": "https://files.pythonhosted.org/packages/17/a0/b4ba4aa5ec80c2c6b351e68d51d29b7b1cb71dc8bc663862e1ea6f5297ad/Django-1.4.20.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.21": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "d96d79a3f4e6b05056356922d1df44fc648b2c1857db0f793b8a835759c11a8d", - "md5": "87f69cbc6e189895be48f02357d10990", - "sha256": "934f1975218680d51c4da9d63a39bc5fb1ddaac48476fd34b9ab7903fd98bcf4" - }, - "downloads": -1, - "filename": "Django-1.4.21.tar.gz", - "has_sig": false, - "md5_digest": "87f69cbc6e189895be48f02357d10990", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7878015, - "upload_time": "2015-07-08T19:56:26", - "upload_time_iso_8601": "2015-07-08T19:56:26.333839Z", - "url": "https://files.pythonhosted.org/packages/d9/6d/79a3f4e6b05056356922d1df44fc648b2c1857db0f793b8a835759c11a8d/Django-1.4.21.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.22": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "467fcead60a10b0208451c42e80db2cc90564cf810148ee46631699ec691cbea", - "md5": "12dc09e5909ce4da93a9d4338db0a43d", - "sha256": "d0e2c9d772fcab2cf9c09e1c05e711cf5fe5eb93225762b29f0739d65e0d1784" - }, - "downloads": -1, - "filename": "Django-1.4.22.tar.gz", - "has_sig": false, - "md5_digest": "12dc09e5909ce4da93a9d4338db0a43d", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7802249, - "upload_time": "2015-08-18T17:22:09", - "upload_time_iso_8601": "2015-08-18T17:22:09.242532Z", - "url": "https://files.pythonhosted.org/packages/46/7f/cead60a10b0208451c42e80db2cc90564cf810148ee46631699ec691cbea/Django-1.4.22.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "18c0fba52cb5af8aa21d2c438ad4067a7c088f74931577c447b49642fc4a65a2", - "md5": "0b134c44b6dc8eb36822677ef506c9ab", - "sha256": "dcadb4b612e5d14f62078869617a26a79b3da719573801d351c4a0a7f4181c4e" - }, - "downloads": -1, - "filename": "Django-1.4.3.tar.gz", - "has_sig": false, - "md5_digest": "0b134c44b6dc8eb36822677ef506c9ab", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7729808, - "upload_time": "2012-12-10T21:46:28", - "upload_time_iso_8601": "2012-12-10T21:46:28.825133Z", - "url": "https://files.pythonhosted.org/packages/18/c0/fba52cb5af8aa21d2c438ad4067a7c088f74931577c447b49642fc4a65a2/Django-1.4.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "b0ab25aba75284beae715089a808b27f0fcf23f818f5216c6f186e55f0ec9774", - "md5": "833f531479948201f0f0a3b5b5972565", - "sha256": "0dd9fa4f0dfc4f64eedecc82bde8dfe15a0a420ceeb11ca1ed050f1742b57077" - }, - "downloads": -1, - "filename": "Django-1.4.4.tar.gz", - "has_sig": false, - "md5_digest": "833f531479948201f0f0a3b5b5972565", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7740176, - "upload_time": "2013-02-19T20:27:55", - "upload_time_iso_8601": "2013-02-19T20:27:55.134186Z", - "url": "https://files.pythonhosted.org/packages/b0/ab/25aba75284beae715089a808b27f0fcf23f818f5216c6f186e55f0ec9774/Django-1.4.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.5": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "5e44879326efd8368a96ecde95500ac02e1421697b3c856d1365ecd03464b9d7", - "md5": "851d00905eb70e4aa6384b3b8b111fb7", - "sha256": "0e1e8c4217299672bbf9404994717fca2d8d4b7a4f7b8b3b74d413e1fda81428" - }, - "downloads": -1, - "filename": "Django-1.4.5.tar.gz", - "has_sig": false, - "md5_digest": "851d00905eb70e4aa6384b3b8b111fb7", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7735582, - "upload_time": "2013-02-20T19:54:40", - "upload_time_iso_8601": "2013-02-20T19:54:40.773224Z", - "url": "https://files.pythonhosted.org/packages/5e/44/879326efd8368a96ecde95500ac02e1421697b3c856d1365ecd03464b9d7/Django-1.4.5.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.6": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "142328d8dc369d13febf55b987aca848f8692ab7e0a70284139c045c940a65c4", - "md5": "5c222ba388f8729151f2fda6be20af90", - "sha256": "cbd3dcc13448fb26d00ec06cd922a593a197ff462984c14b64a6b25be1d703bb" - }, - "downloads": -1, - "filename": "Django-1.4.6.tar.gz", - "has_sig": false, - "md5_digest": "5c222ba388f8729151f2fda6be20af90", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7744137, - "upload_time": "2013-08-13T16:52:54", - "upload_time_iso_8601": "2013-08-13T16:52:54.160398Z", - "url": "https://files.pythonhosted.org/packages/14/23/28d8dc369d13febf55b987aca848f8692ab7e0a70284139c045c940a65c4/Django-1.4.6.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.7": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "1484d2bdc262c35fffa54f85399295018c51d02fa5f0929748a701a08a421c01", - "md5": "28da2e8111ff951adbfce0651f945326", - "sha256": "5846c9a1a1a59eb8f802b3d694971a030f8a30bfb6d17ed4b29dc40768539ce6" - }, - "downloads": -1, - "filename": "Django-1.4.7.tar.gz", - "has_sig": false, - "md5_digest": "28da2e8111ff951adbfce0651f945326", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7743506, - "upload_time": "2013-09-11T01:18:42", - "upload_time_iso_8601": "2013-09-11T01:18:42.411587Z", - "url": "https://files.pythonhosted.org/packages/14/84/d2bdc262c35fffa54f85399295018c51d02fa5f0929748a701a08a421c01/Django-1.4.7.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.8": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "597cc02441c384e2287b8398e85a8cb25a9b2602e4b05d699e8e64c7b982a64c", - "md5": "7075e08ef06155e07002189b837cde85", - "sha256": "b9c356411af17dd9017081c884065976745659b3ab0e80493d0656911f920a2d" - }, - "downloads": -1, - "filename": "Django-1.4.8.tar.gz", - "has_sig": false, - "md5_digest": "7075e08ef06155e07002189b837cde85", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7743397, - "upload_time": "2013-09-15T06:22:11", - "upload_time_iso_8601": "2013-09-15T06:22:11.681231Z", - "url": "https://files.pythonhosted.org/packages/59/7c/c02441c384e2287b8398e85a8cb25a9b2602e4b05d699e8e64c7b982a64c/Django-1.4.8.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.4.9": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "cde65a62c92604fa044fa025679448435200a20fd8a751411dee0b264c4659ca", - "md5": "cc0c9752b46de362bd2114a65871330f", - "sha256": "87c6bf92517e686a7c76af56e309aeef6ad93a83b27b4b41e4b95c673acf9ade" - }, - "downloads": -1, - "filename": "Django-1.4.9.tar.gz", - "has_sig": false, - "md5_digest": "cc0c9752b46de362bd2114a65871330f", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7745043, - "upload_time": "2013-10-25T04:38:13", - "upload_time_iso_8601": "2013-10-25T04:38:13.629577Z", - "url": "https://files.pythonhosted.org/packages/cd/e6/5a62c92604fa044fa025679448435200a20fd8a751411dee0b264c4659ca/Django-1.4.9.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.5": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "0d5bd190b641cbf77e6eeebba93c75cf1a585d7275fa6c87450ecc1714eef93c", - "md5": "fac09e1e0f11bb83bb187d652a9be967", - "sha256": "078bf8f8ab025ed79e41ed5cee145a64dffea638eb5c2928c8cd106720824416" - }, - "downloads": -1, - "filename": "Django-1.5.tar.gz", - "has_sig": false, - "md5_digest": "fac09e1e0f11bb83bb187d652a9be967", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 8007045, - "upload_time": "2013-02-26T19:30:37", - "upload_time_iso_8601": "2013-02-26T19:30:37.100371Z", - "url": "https://files.pythonhosted.org/packages/0d/5b/d190b641cbf77e6eeebba93c75cf1a585d7275fa6c87450ecc1714eef93c/Django-1.5.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.5.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "7fc37a38a4985447e2812c99e66501abf09c6edd55ccfe070bbc82a58054dd1e", - "md5": "7465f6383264ba167a9a031d6b058bff", - "sha256": "885fadcbb8963c0ccda5d9d2cca792970b0289b4e662406b2de2b736ff46123d" - }, - "downloads": -1, - "filename": "Django-1.5.1.tar.gz", - "has_sig": false, - "md5_digest": "7465f6383264ba167a9a031d6b058bff", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 8028963, - "upload_time": "2013-03-28T20:57:18", - "upload_time_iso_8601": "2013-03-28T20:57:18.760489Z", - "url": "https://files.pythonhosted.org/packages/7f/c3/7a38a4985447e2812c99e66501abf09c6edd55ccfe070bbc82a58054dd1e/Django-1.5.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.5.10": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "ff2f7412428c52976797c5229a613326cf41d678a0f2fcfafb427ada6c8a561a", - "md5": "b055361f04c0b8e862f8e8ffbb44e464", - "sha256": "7cb4217e740f7d5d6d74617dbb9d960f9c09e8269c6762fe68c6e762219f4018" - }, - "downloads": -1, - "filename": "Django-1.5.10.tar.gz", - "has_sig": false, - "md5_digest": "b055361f04c0b8e862f8e8ffbb44e464", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 8074324, - "upload_time": "2014-09-02T20:51:16", - "upload_time_iso_8601": "2014-09-02T20:51:16.978621Z", - "url": "https://files.pythonhosted.org/packages/ff/2f/7412428c52976797c5229a613326cf41d678a0f2fcfafb427ada6c8a561a/Django-1.5.10.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.5.11": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "814cbecfb2589793a5052528ced88863427f1d8b09600bcefd6dd52c7eda29bf", - "md5": "6e88cab476e5149812accc143d313a22", - "sha256": "bf7d9bb21f24a67badd751bafbda85cb1003f6274ad43ba5984a0868182bf26c" - }, - "downloads": -1, - "filename": "Django-1.5.11.tar.gz", - "has_sig": false, - "md5_digest": "6e88cab476e5149812accc143d313a22", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 8075434, - "upload_time": "2014-10-22T16:45:00", - "upload_time_iso_8601": "2014-10-22T16:45:00.838415Z", - "url": "https://files.pythonhosted.org/packages/81/4c/becfb2589793a5052528ced88863427f1d8b09600bcefd6dd52c7eda29bf/Django-1.5.11.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.5.12": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "e2192b5fba33b2c6548fcd9ae4961f9d89bafeedf232fe3367c28a85280e9e13", - "md5": "c35cb78bbf20a8ef60d37207d75a0f34", - "sha256": "3574000cdf223c0dd860e807f1156f858f993318720deeb3c7ea70797c78cbf1" - }, - "downloads": -1, - "filename": "Django-1.5.12-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "c35cb78bbf20a8ef60d37207d75a0f34", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 8317210, - "upload_time": "2015-01-03T02:09:00", - "upload_time_iso_8601": "2015-01-03T02:09:00.443838Z", - "url": "https://files.pythonhosted.org/packages/e2/19/2b5fba33b2c6548fcd9ae4961f9d89bafeedf232fe3367c28a85280e9e13/Django-1.5.12-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "7ba8e4ffd9cc175c0b6b41cc33463449e6515eff28a27f7102818e24d9fd2517", - "md5": "0e0b48cd0bb59cbc5499dcbb4fe1fb90", - "sha256": "b3de77beb6e59b72071ca66f20c2ad34e1b90d39b0241e62c1f03c668ddd6ced" - }, - "downloads": -1, - "filename": "Django-1.5.12.tar.gz", - "has_sig": false, - "md5_digest": "0e0b48cd0bb59cbc5499dcbb4fe1fb90", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 8202839, - "upload_time": "2015-01-03T02:09:17", - "upload_time_iso_8601": "2015-01-03T02:09:17.471850Z", - "url": "https://files.pythonhosted.org/packages/7b/a8/e4ffd9cc175c0b6b41cc33463449e6515eff28a27f7102818e24d9fd2517/Django-1.5.12.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.5.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "ad3962b220a0efda308abeccb66653b8b950c6d1c51c673bd7978fc13f53eaf5", - "md5": "07f0d2d42162945d0ad031fc9737847d", - "sha256": "0bab8f96f57f28e3676087d07301b6ba519f946ab9257b4ea072a5ff13c8e3f9" - }, - "downloads": -1, - "filename": "Django-1.5.2-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "07f0d2d42162945d0ad031fc9737847d", - "packagetype": "bdist_wheel", - "python_version": "any", - "requires_python": null, - "size": 8311743, - "upload_time": "2013-08-13T16:54:03", - "upload_time_iso_8601": "2013-08-13T16:54:03.787628Z", - "url": "https://files.pythonhosted.org/packages/ad/39/62b220a0efda308abeccb66653b8b950c6d1c51c673bd7978fc13f53eaf5/Django-1.5.2-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "8c3df57d35c5fe63e298fe33f9c5fea5b3920385f6fc5b562ae6e3b9eff7a0ef", - "md5": "26e83e6394a15a86212777d5f61eae86", - "sha256": "9a4b19adaaa096843425d426ffbeb928e85d861ff9c106527cb747dc67b434da" - }, - "downloads": -1, - "filename": "Django-1.5.2.tar.gz", - "has_sig": false, - "md5_digest": "26e83e6394a15a86212777d5f61eae86", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 8044778, - "upload_time": "2013-08-13T16:52:40", - "upload_time_iso_8601": "2013-08-13T16:52:40.312336Z", - "url": "https://files.pythonhosted.org/packages/8c/3d/f57d35c5fe63e298fe33f9c5fea5b3920385f6fc5b562ae6e3b9eff7a0ef/Django-1.5.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.5.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "9532da1aaef056d4652aa6a43d60647b55488aaaa14aa36515fceb52836cd571", - "md5": "1581e28b4aeb269c34a9b0417e103aaa", - "sha256": "e0fd8dec0497ed98e8e03bd5297064f276f203f4c192f252231a3f25187b59b5" - }, - "downloads": -1, - "filename": "Django-1.5.3.tar.gz", - "has_sig": false, - "md5_digest": "1581e28b4aeb269c34a9b0417e103aaa", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 8049029, - "upload_time": "2013-09-11T01:26:50", - "upload_time_iso_8601": "2013-09-11T01:26:50.200599Z", - "url": "https://files.pythonhosted.org/packages/95/32/da1aaef056d4652aa6a43d60647b55488aaaa14aa36515fceb52836cd571/Django-1.5.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.5.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "b2efd37cd67c9eccd7329ce421382f517bb6f9a431ded3c6fd60cba8c966712a", - "md5": "b2685469bb4d1fbb091316e21f4108de", - "sha256": "428defe3fd515dfc8613039bb0a80622a13fb4b988c5be48db07ec098ea1704e" - }, - "downloads": -1, - "filename": "Django-1.5.4.tar.gz", - "has_sig": false, - "md5_digest": "b2685469bb4d1fbb091316e21f4108de", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 8050758, - "upload_time": "2013-09-15T06:30:37", - "upload_time_iso_8601": "2013-09-15T06:30:37.726078Z", - "url": "https://files.pythonhosted.org/packages/b2/ef/d37cd67c9eccd7329ce421382f517bb6f9a431ded3c6fd60cba8c966712a/Django-1.5.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.5.5": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "384993511c5d3367b6b21fc2995a0e53399721afc15e4cd6eb57be879ae13ad4", - "md5": "e33355ee4bb2cbb4ab3954d3dff5eddd", - "sha256": "6ae69c1dfbfc9d0c44ae80e2fbe48e59bbbbb70e8df66ad2b7029bd39947d71d" - }, - "downloads": -1, - "filename": "Django-1.5.5.tar.gz", - "has_sig": false, - "md5_digest": "e33355ee4bb2cbb4ab3954d3dff5eddd", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 8060441, - "upload_time": "2013-10-25T04:32:41", - "upload_time_iso_8601": "2013-10-25T04:32:41.565014Z", - "url": "https://files.pythonhosted.org/packages/38/49/93511c5d3367b6b21fc2995a0e53399721afc15e4cd6eb57be879ae13ad4/Django-1.5.5.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.5.6": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "b2e578f1e96e8d3ae38cd9b3a4690a448a4716be73f5f5408d9f0da84576e36e", - "md5": "b46fe29c7d26310d19aec6d8666f08c6", - "sha256": "9b7fcb99d20289189ec0f1e06d1d2bed3b4772e3a393fddbfb006ea7c3f9bfaf" - }, - "downloads": -1, - "filename": "Django-1.5.6.tar.gz", - "has_sig": false, - "md5_digest": "b46fe29c7d26310d19aec6d8666f08c6", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 8068359, - "upload_time": "2014-04-21T22:53:10", - "upload_time_iso_8601": "2014-04-21T22:53:10.449788Z", - "url": "https://files.pythonhosted.org/packages/b2/e5/78f1e96e8d3ae38cd9b3a4690a448a4716be73f5f5408d9f0da84576e36e/Django-1.5.6.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.5.7": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "561332b4ad8b2bc3b9d820467dc30d28a53823314bcfcdfff55c0218dd4e5b08", - "md5": "a2c127e85a34c2eb6c74db4f7e02d4e4", - "sha256": "08a41c2a37451b8cc1136823b802dd6f17ad6ec0c8d2cadb4c9a219ff4c08593" - }, - "downloads": -1, - "filename": "Django-1.5.7.tar.gz", - "has_sig": false, - "md5_digest": "a2c127e85a34c2eb6c74db4f7e02d4e4", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 8069177, - "upload_time": "2014-04-28T20:35:50", - "upload_time_iso_8601": "2014-04-28T20:35:50.148784Z", - "url": "https://files.pythonhosted.org/packages/56/13/32b4ad8b2bc3b9d820467dc30d28a53823314bcfcdfff55c0218dd4e5b08/Django-1.5.7.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.5.8": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "8a28b99f0e8977b7227fffc43aeefe87d7561dec06cc6bdc586f4c97ec60c853", - "md5": "1e3418bd1d6f9725a3d1264c9352f2a1", - "sha256": "65009f8060e1c246c04d25c3b4b7f7bfaf6c8f9f4ef4db0c34fb18e061118b31" - }, - "downloads": -1, - "filename": "Django-1.5.8-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "1e3418bd1d6f9725a3d1264c9352f2a1", - "packagetype": "bdist_wheel", - "python_version": "any", - "requires_python": null, - "size": 8315409, - "upload_time": "2014-05-14T18:35:13", - "upload_time_iso_8601": "2014-05-14T18:35:13.385578Z", - "url": "https://files.pythonhosted.org/packages/8a/28/b99f0e8977b7227fffc43aeefe87d7561dec06cc6bdc586f4c97ec60c853/Django-1.5.8-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "a0ea8ea5e891fe5c6a7e9dd0da9449bbe680cc43d21a78f9a62132b8cfdf690f", - "md5": "675fc736e2c29090f005e217ccf90b5b", - "sha256": "01db30f38a081241a9cbc7bef12cb599506b80727613350e427547bed12aaaa3" - }, - "downloads": -1, - "filename": "Django-1.5.8.tar.gz", - "has_sig": false, - "md5_digest": "675fc736e2c29090f005e217ccf90b5b", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 8071329, - "upload_time": "2014-05-14T18:36:01", - "upload_time_iso_8601": "2014-05-14T18:36:01.585251Z", - "url": "https://files.pythonhosted.org/packages/a0/ea/8ea5e891fe5c6a7e9dd0da9449bbe680cc43d21a78f9a62132b8cfdf690f/Django-1.5.8.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.5.9": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "e3eabe67f2d55fd583b7ff6e7cc0764da2b1d6437317354beedabd0ff15cad3a", - "md5": "4c6f03748043a32059d905033e0dc770", - "sha256": "47ce505c5046c38817828bee253b7256872f86c4340db1af698cb8548dbaa0d2" - }, - "downloads": -1, - "filename": "Django-1.5.9.tar.gz", - "has_sig": false, - "md5_digest": "4c6f03748043a32059d905033e0dc770", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 8074400, - "upload_time": "2014-08-20T20:10:42", - "upload_time_iso_8601": "2014-08-20T20:10:42.568983Z", - "url": "https://files.pythonhosted.org/packages/e3/ea/be67f2d55fd583b7ff6e7cc0764da2b1d6437317354beedabd0ff15cad3a/Django-1.5.9.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.6": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "f1dd271a9fa17b95a980ac66c44848fef72d29d904d3e141b219f6e91d1904ec", - "md5": "1078059a13d83a091e952917d22da9af", - "sha256": "9365e04db9ad92524350247906ce6edc8b98fc95d146f697540edbfddb23ba13" - }, - "downloads": -1, - "filename": "Django-1.6-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "1078059a13d83a091e952917d22da9af", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6667901, - "upload_time": "2013-11-06T15:01:02", - "upload_time_iso_8601": "2013-11-06T15:01:02.246764Z", - "url": "https://files.pythonhosted.org/packages/f1/dd/271a9fa17b95a980ac66c44848fef72d29d904d3e141b219f6e91d1904ec/Django-1.6-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "471cc18fbec3ea87bfa23808fb6527e1a48e5bf59db1ff24ff7d8d6cb0a2bccb", - "md5": "65db1bc313124c3754c89073942e38a8", - "sha256": "d3d9fdc8f313e5a33a6dc7ebdeca19147c11029822b462064c56895d7969ab98" - }, - "downloads": -1, - "filename": "Django-1.6.tar.gz", - "has_sig": false, - "md5_digest": "65db1bc313124c3754c89073942e38a8", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6597874, - "upload_time": "2013-11-06T15:01:29", - "upload_time_iso_8601": "2013-11-06T15:01:29.487525Z", - "url": "https://files.pythonhosted.org/packages/47/1c/c18fbec3ea87bfa23808fb6527e1a48e5bf59db1ff24ff7d8d6cb0a2bccb/Django-1.6.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.6.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "53c428cc8a55aa9bf9579bd496f88505f3a14ff0ed4b1c6954a8ba5ce649a685", - "md5": "c7b7a4437b36400f1c23953e9700fd29", - "sha256": "989d42289663ac88169ac2abe8d50b82b29b2fe135307badf588a3d2235c1eef" - }, - "downloads": -1, - "filename": "Django-1.6.1-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "c7b7a4437b36400f1c23953e9700fd29", - "packagetype": "bdist_wheel", - "python_version": "any", - "requires_python": null, - "size": 6680676, - "upload_time": "2013-12-12T20:04:35", - "upload_time_iso_8601": "2013-12-12T20:04:35.467654Z", - "url": "https://files.pythonhosted.org/packages/53/c4/28cc8a55aa9bf9579bd496f88505f3a14ff0ed4b1c6954a8ba5ce649a685/Django-1.6.1-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "4e1a06610645c7be99f342b10d9aa29bf833800778532cc1738406e1805322c1", - "md5": "3ea7a00ea9e7a014e8a4067dd6466a1b", - "sha256": "cf011874f54a16e7452e0fe1e7f4ec144b95b47ecf31766c9f1f8cf438f09c06" - }, - "downloads": -1, - "filename": "Django-1.6.1.tar.gz", - "has_sig": false, - "md5_digest": "3ea7a00ea9e7a014e8a4067dd6466a1b", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6608178, - "upload_time": "2013-12-12T20:04:13", - "upload_time_iso_8601": "2013-12-12T20:04:13.572132Z", - "url": "https://files.pythonhosted.org/packages/4e/1a/06610645c7be99f342b10d9aa29bf833800778532cc1738406e1805322c1/Django-1.6.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.6.10": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "5c50f3eb71c2eee5abbfa4f9368ac0ccdf4d58dcc1d391d60cf9d8eb7070d52d", - "md5": "f83dcaec9e3b7d956a4d29e9401b0b97", - "sha256": "ceee83ff4c4fa1461289fe07a5879e8440089fadde150d40753691cdd8c942c1" - }, - "downloads": -1, - "filename": "Django-1.6.10-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "f83dcaec9e3b7d956a4d29e9401b0b97", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6688308, - "upload_time": "2015-01-13T18:48:42", - "upload_time_iso_8601": "2015-01-13T18:48:42.343729Z", - "url": "https://files.pythonhosted.org/packages/5c/50/f3eb71c2eee5abbfa4f9368ac0ccdf4d58dcc1d391d60cf9d8eb7070d52d/Django-1.6.10-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "ff83fad3cc3ac64f8096ae5e56fa1de303ff947fb3fab69cd96d3df9ccf353fb", - "md5": "d7123f14ac19ae001be02ed841937b91", - "sha256": "54eb59ce785401c7d1fdeed245efce597e90f811d6a20f6b5c6931c0049d63a6" - }, - "downloads": -1, - "filename": "Django-1.6.10.tar.gz", - "has_sig": false, - "md5_digest": "d7123f14ac19ae001be02ed841937b91", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6760152, - "upload_time": "2015-01-13T18:48:53", - "upload_time_iso_8601": "2015-01-13T18:48:53.435836Z", - "url": "https://files.pythonhosted.org/packages/ff/83/fad3cc3ac64f8096ae5e56fa1de303ff947fb3fab69cd96d3df9ccf353fb/Django-1.6.10.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.6.11": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "8086f52eec28e96fb211122424a3db696e7676ad3555c11027afd9fee9bb0d23", - "md5": "e43d8c4147a821205b554aebc1da62d5", - "sha256": "6f2cc848b2b72adf53a6f3f21be049c477e82c408bce7cedb57efeb0984bde24" - }, - "downloads": -1, - "filename": "Django-1.6.11-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "e43d8c4147a821205b554aebc1da62d5", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6688479, - "upload_time": "2015-03-18T23:57:46", - "upload_time_iso_8601": "2015-03-18T23:57:46.298309Z", - "url": "https://files.pythonhosted.org/packages/80/86/f52eec28e96fb211122424a3db696e7676ad3555c11027afd9fee9bb0d23/Django-1.6.11-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "691aa47b2efd22bf642d19d3dde71f82b29c5608bb96e140a38208bf813f9c3b", - "md5": "3bf6086c3b923876d283dc3404e32fdd", - "sha256": "7e50e573e484435873b3515d7982d80093b2695aba17fd0ff024307454dc3a56" - }, - "downloads": -1, - "filename": "Django-1.6.11.tar.gz", - "has_sig": false, - "md5_digest": "3bf6086c3b923876d283dc3404e32fdd", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6764000, - "upload_time": "2015-03-18T23:58:02", - "upload_time_iso_8601": "2015-03-18T23:58:02.690545Z", - "url": "https://files.pythonhosted.org/packages/69/1a/a47b2efd22bf642d19d3dde71f82b29c5608bb96e140a38208bf813f9c3b/Django-1.6.11.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.6.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "66f4e695878103eb7371e1077374e09dec23ce95302f2e55d7b1ad908613cac9", - "md5": "3bd014923e85df771b34d12c0ab3c9e1", - "sha256": "b81091fd41e952e9d7150b8bc2055b140c2c1132485f78e4ea075013708000d5" - }, - "downloads": -1, - "filename": "Django-1.6.2-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "3bd014923e85df771b34d12c0ab3c9e1", - "packagetype": "bdist_wheel", - "python_version": "any", - "requires_python": null, - "size": 6682217, - "upload_time": "2014-02-06T21:51:23", - "upload_time_iso_8601": "2014-02-06T21:51:23.493703Z", - "url": "https://files.pythonhosted.org/packages/66/f4/e695878103eb7371e1077374e09dec23ce95302f2e55d7b1ad908613cac9/Django-1.6.2-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "21780c2958697c30b35de1ac691f9002c486da052e3fc64b4cf60a9d28ef1d51", - "md5": "45d974c623b3bfbf9976f3d808fe1ee9", - "sha256": "d1b3f8460e936f47846e7c4f80af951eda82a41c253c3a51ff3389863ff1c03a" - }, - "downloads": -1, - "filename": "Django-1.6.2.tar.gz", - "has_sig": false, - "md5_digest": "45d974c623b3bfbf9976f3d808fe1ee9", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6615116, - "upload_time": "2014-02-06T21:50:52", - "upload_time_iso_8601": "2014-02-06T21:50:52.673014Z", - "url": "https://files.pythonhosted.org/packages/21/78/0c2958697c30b35de1ac691f9002c486da052e3fc64b4cf60a9d28ef1d51/Django-1.6.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.6.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "4a6fbd944fae447c329a9c316966e2736cda68f5319bd0dba34e4cb92c695457", - "md5": "e5937a962ce1298ac67d1aa7484883a6", - "sha256": "876629e55678f890186671c426084ba20e1ff9f87a45b516923379c543976a5e" - }, - "downloads": -1, - "filename": "Django-1.6.3-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "e5937a962ce1298ac67d1aa7484883a6", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6683433, - "upload_time": "2014-04-21T23:12:03", - "upload_time_iso_8601": "2014-04-21T23:12:03.573997Z", - "url": "https://files.pythonhosted.org/packages/4a/6f/bd944fae447c329a9c316966e2736cda68f5319bd0dba34e4cb92c695457/Django-1.6.3-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "9ea9a49cfb89bbec60c7720435795c96d315703f1bd0a2fc5b88a9a322534436", - "md5": "727fec03f15db8f80a7231696b79adf7", - "sha256": "6d9d3c468f9a09470d00e85fe492ba35edfc72cee7fb65ad0281010eba58b8f1" - }, - "downloads": -1, - "filename": "Django-1.6.3.tar.gz", - "has_sig": false, - "md5_digest": "727fec03f15db8f80a7231696b79adf7", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6628812, - "upload_time": "2014-04-21T23:12:25", - "upload_time_iso_8601": "2014-04-21T23:12:25.701466Z", - "url": "https://files.pythonhosted.org/packages/9e/a9/a49cfb89bbec60c7720435795c96d315703f1bd0a2fc5b88a9a322534436/Django-1.6.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.6.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "01d3aa92215ab08064294337bcc51a141201e798a7dd80ca9fc1384397b717de", - "md5": "8093262e13535869720d7100aed72fd6", - "sha256": "bd90127488621b040a49b7a12e29ba20bebcf2148f6431569f52804b1beb508d" - }, - "downloads": -1, - "filename": "Django-1.6.4-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "8093262e13535869720d7100aed72fd6", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6683808, - "upload_time": "2014-04-28T20:40:32", - "upload_time_iso_8601": "2014-04-28T20:40:32.164010Z", - "url": "https://files.pythonhosted.org/packages/01/d3/aa92215ab08064294337bcc51a141201e798a7dd80ca9fc1384397b717de/Django-1.6.4-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "71581482ebcd3ac40c6f5a8d0d2504c62323e3ebb94b6775d62011b73178d796", - "md5": "0d23bf836d3a52d93aee9411eccaa609", - "sha256": "ceee0beea79b1926c767aaa837e1b9e621e5f6b7d27138d90474b3917ca5527b" - }, - "downloads": -1, - "filename": "Django-1.6.4.tar.gz", - "has_sig": false, - "md5_digest": "0d23bf836d3a52d93aee9411eccaa609", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6630474, - "upload_time": "2014-04-28T20:40:52", - "upload_time_iso_8601": "2014-04-28T20:40:52.661964Z", - "url": "https://files.pythonhosted.org/packages/71/58/1482ebcd3ac40c6f5a8d0d2504c62323e3ebb94b6775d62011b73178d796/Django-1.6.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.6.5": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "436c7fffbe73fe5703125aa1c9e3279ad3a5e542128ee55a4aa83669db1985cd", - "md5": "2bcdb4729f9f358b0925b532eef0a8ff", - "sha256": "4eda29d8eb0c8b4a836660b5eff78a0d3e0dc6c191e998a14194f2ff51130da3" - }, - "downloads": -1, - "filename": "Django-1.6.5-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "2bcdb4729f9f358b0925b532eef0a8ff", - "packagetype": "bdist_wheel", - "python_version": "any", - "requires_python": null, - "size": 6683598, - "upload_time": "2014-05-14T18:33:13", - "upload_time_iso_8601": "2014-05-14T18:33:13.760622Z", - "url": "https://files.pythonhosted.org/packages/43/6c/7fffbe73fe5703125aa1c9e3279ad3a5e542128ee55a4aa83669db1985cd/Django-1.6.5-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "cc2cfc2495af4dcc3c81834f9c390cfccfe6b037fdfb5dfdce6c30413170c20b", - "md5": "e4c5b2d35ecb3807317713afa70a0c77", - "sha256": "36940268c087fede32d3f5887cce9af9e5d27962a0c405aacafc2a3cc1f755c5" - }, - "downloads": -1, - "filename": "Django-1.6.5.tar.gz", - "has_sig": false, - "md5_digest": "e4c5b2d35ecb3807317713afa70a0c77", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6633768, - "upload_time": "2014-05-14T18:34:11", - "upload_time_iso_8601": "2014-05-14T18:34:11.991240Z", - "url": "https://files.pythonhosted.org/packages/cc/2c/fc2495af4dcc3c81834f9c390cfccfe6b037fdfb5dfdce6c30413170c20b/Django-1.6.5.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.6.6": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "9ec819c7a7e76d9777366af0de6db934780b2d276a23054aa8109a661689f09c", - "md5": "74ffe011439efffcefbda6fac294c6f6", - "sha256": "75e3429cac94389d4b0f58501ef1f7a0d500db0b35367157ce0d09caefec3372" - }, - "downloads": -1, - "filename": "Django-1.6.6-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "74ffe011439efffcefbda6fac294c6f6", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6684772, - "upload_time": "2014-08-20T20:17:23", - "upload_time_iso_8601": "2014-08-20T20:17:23.703605Z", - "url": "https://files.pythonhosted.org/packages/9e/c8/19c7a7e76d9777366af0de6db934780b2d276a23054aa8109a661689f09c/Django-1.6.6-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "0e960e6b38a7d5dca4662fdeb23d021ce47df1fa388f8f34816ba3cbaed61ec6", - "md5": "d14fd332f31799fff39acc0c79e8421c", - "sha256": "536cbd54e533ba3563d205f0c91988b24e7d74b8b253d7825e42214b50ba7e90" - }, - "downloads": -1, - "filename": "Django-1.6.6.tar.gz", - "has_sig": false, - "md5_digest": "d14fd332f31799fff39acc0c79e8421c", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6645456, - "upload_time": "2014-08-20T20:17:01", - "upload_time_iso_8601": "2014-08-20T20:17:01.372683Z", - "url": "https://files.pythonhosted.org/packages/0e/96/0e6b38a7d5dca4662fdeb23d021ce47df1fa388f8f34816ba3cbaed61ec6/Django-1.6.6.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.6.7": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "9e51e6059dba3f8fa2adbcb05ba3271e68182c81210055666abd6a01e15ff515", - "md5": "72a2df8d67a976208420eec2fe2129fe", - "sha256": "b65dc7f98c5a729314d001ebbac9228befd625705a9e3ae039a5d160c3976fe1" - }, - "downloads": -1, - "filename": "Django-1.6.7-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "72a2df8d67a976208420eec2fe2129fe", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6684835, - "upload_time": "2014-09-02T20:55:35", - "upload_time_iso_8601": "2014-09-02T20:55:35.591619Z", - "url": "https://files.pythonhosted.org/packages/9e/51/e6059dba3f8fa2adbcb05ba3271e68182c81210055666abd6a01e15ff515/Django-1.6.7-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "c26b44ae1265b3f3b4f215ce23720f66bb75b0019b5ecd78d720ecfbe36fad0d", - "md5": "f31e2f953feb258e3569e962790630b6", - "sha256": "9a64211c96a3262bb2545acc82af5d8f3da0175299f7c7e901e4ed455be965fb" - }, - "downloads": -1, - "filename": "Django-1.6.7.tar.gz", - "has_sig": false, - "md5_digest": "f31e2f953feb258e3569e962790630b6", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6647301, - "upload_time": "2014-09-02T20:56:17", - "upload_time_iso_8601": "2014-09-02T20:56:17.794128Z", - "url": "https://files.pythonhosted.org/packages/c2/6b/44ae1265b3f3b4f215ce23720f66bb75b0019b5ecd78d720ecfbe36fad0d/Django-1.6.7.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.6.8": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "1e01e985b325b49ab53b4bad6e22ae7959474b34a8053eb6846f28286a4611fd", - "md5": "1cb695150210433200929623f800edcb", - "sha256": "8a68c0e8beef56560e01eb0c734d8b969e64d38801e6d98bcae70bb836aa4c3f" - }, - "downloads": -1, - "filename": "Django-1.6.8-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "1cb695150210433200929623f800edcb", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6685693, - "upload_time": "2014-10-22T16:50:19", - "upload_time_iso_8601": "2014-10-22T16:50:19.920524Z", - "url": "https://files.pythonhosted.org/packages/1e/01/e985b325b49ab53b4bad6e22ae7959474b34a8053eb6846f28286a4611fd/Django-1.6.8-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "8e389e47ac747026608f284813ee41d49f0b1770bb78ca75aefb2fb268abe16f", - "md5": "b00f9f73535db7c9ce52a6f707d61ab6", - "sha256": "a8685c1fb5b0bcad9007d941c81493668f9613578add631f406d3f95b84cf6d0" - }, - "downloads": -1, - "filename": "Django-1.6.8.tar.gz", - "has_sig": false, - "md5_digest": "b00f9f73535db7c9ce52a6f707d61ab6", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6650671, - "upload_time": "2014-10-22T16:50:38", - "upload_time_iso_8601": "2014-10-22T16:50:38.114205Z", - "url": "https://files.pythonhosted.org/packages/8e/38/9e47ac747026608f284813ee41d49f0b1770bb78ca75aefb2fb268abe16f/Django-1.6.8.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.6.9": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "c7cb1b2eb81a62f81e862c90675915dd07375d459d789ef372198abec9fbdc7a", - "md5": "279bc0f844c644939dbe2a77c2249124", - "sha256": "20dd4ae31564df143fe8ca6daf507b68e32f3ef70987049fdb978023a843431c" - }, - "downloads": -1, - "filename": "Django-1.6.9-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "279bc0f844c644939dbe2a77c2249124", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6687845, - "upload_time": "2015-01-03T01:52:23", - "upload_time_iso_8601": "2015-01-03T01:52:23.765940Z", - "url": "https://files.pythonhosted.org/packages/c7/cb/1b2eb81a62f81e862c90675915dd07375d459d789ef372198abec9fbdc7a/Django-1.6.9-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "dec0b9e7fbf78afb964a7ff7e9609ea7d4c65fb67c7c4d456638dfa45c3f8d0c", - "md5": "03893cd1232f6cf75f0523e1b2c91ed2", - "sha256": "d8c182e9ac88f6ef7e5f89e71282793d9682e76a8da39a0c4bfd452e611a06a8" - }, - "downloads": -1, - "filename": "Django-1.6.9.tar.gz", - "has_sig": false, - "md5_digest": "03893cd1232f6cf75f0523e1b2c91ed2", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 6753466, - "upload_time": "2015-01-03T01:52:35", - "upload_time_iso_8601": "2015-01-03T01:52:35.812923Z", - "url": "https://files.pythonhosted.org/packages/de/c0/b9e7fbf78afb964a7ff7e9609ea7d4c65fb67c7c4d456638dfa45c3f8d0c/Django-1.6.9.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.7": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "d55ea1328223c382024b2b184dacc713086463ce1a60fe1471d500b3b66f840a", - "md5": "15efe093b40d058acf24682c31e7b24c", - "sha256": "009ddda445c5750c1a8392979fbd28f3e55de6e43310cd316199837065dff559" - }, - "downloads": -1, - "filename": "Django-1.7-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "15efe093b40d058acf24682c31e7b24c", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 7384572, - "upload_time": "2014-09-02T21:09:27", - "upload_time_iso_8601": "2014-09-02T21:09:27.945216Z", - "url": "https://files.pythonhosted.org/packages/d5/5e/a1328223c382024b2b184dacc713086463ce1a60fe1471d500b3b66f840a/Django-1.7-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "df07e87b36c24a742deebbffafb31504bbaffe4a20db799b43404f655413052a", - "md5": "03edab6828119aa9b32b2252d25eb38d", - "sha256": "33f781f17f145f79ee8e0b8d753498e0e0188f0b53b2accad4045d623422d5e1" - }, - "downloads": -1, - "filename": "Django-1.7.tar.gz", - "has_sig": false, - "md5_digest": "03edab6828119aa9b32b2252d25eb38d", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7486550, - "upload_time": "2014-09-02T21:10:05", - "upload_time_iso_8601": "2014-09-02T21:10:05.229716Z", - "url": "https://files.pythonhosted.org/packages/df/07/e87b36c24a742deebbffafb31504bbaffe4a20db799b43404f655413052a/Django-1.7.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.7.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "d2291935a5825b8820d1e398ab83f0730d483ec731fae34745ddac8318cf6ac8", - "md5": "83bd3e5cfba6d6d2bee5a37efb34771c", - "sha256": "679fc24b3e85bf5a07ca2f6d5c4cdf3d4477bbb02f43a6548335952cc75b5d23" - }, - "downloads": -1, - "filename": "Django-1.7.1-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "83bd3e5cfba6d6d2bee5a37efb34771c", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 7414830, - "upload_time": "2014-10-22T16:56:56", - "upload_time_iso_8601": "2014-10-22T16:56:56.139953Z", - "url": "https://files.pythonhosted.org/packages/d2/29/1935a5825b8820d1e398ab83f0730d483ec731fae34745ddac8318cf6ac8/Django-1.7.1-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "8ca39876f73205ed8e1a2e345245e7e66fc02a056f29503590eca392fcb3e7c0", - "md5": "81dae89f21647b9aa5c46c6b7dbfa349", - "sha256": "3de62e71ce2cfbcdecb6e344cad04948506c8410ea5c6eab15c8f3b31b8ac1c0" - }, - "downloads": -1, - "filename": "Django-1.7.1.tar.gz", - "has_sig": false, - "md5_digest": "81dae89f21647b9aa5c46c6b7dbfa349", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7527499, - "upload_time": "2014-10-22T16:57:16", - "upload_time_iso_8601": "2014-10-22T16:57:16.294800Z", - "url": "https://files.pythonhosted.org/packages/8c/a3/9876f73205ed8e1a2e345245e7e66fc02a056f29503590eca392fcb3e7c0/Django-1.7.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.7.10": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "2373c0440d75121e69292115dd9d5e9f547113a86b73309652700b02fcb13d08", - "md5": "ba5adcf08d8b60bf82284a09a4d384a3", - "sha256": "1db5dafe4fe8302f34449283864baa74d0d64013613aa200918187bf76902d5f" - }, - "downloads": -1, - "filename": "Django-1.7.10-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "ba5adcf08d8b60bf82284a09a4d384a3", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 7424722, - "upload_time": "2015-08-18T17:15:28", - "upload_time_iso_8601": "2015-08-18T17:15:28.710738Z", - "url": "https://files.pythonhosted.org/packages/23/73/c0440d75121e69292115dd9d5e9f547113a86b73309652700b02fcb13d08/Django-1.7.10-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "b18c57e58359f201c8b3b46706bf06a5cef32d0a8176cf287f45ad7b2eb10891", - "md5": "90315a9bec9b073a91beeb3f60994600", - "sha256": "b9357d2cebe61997055d417d607f9c650e817cd1a383b9a1b88bf1edad797c75" - }, - "downloads": -1, - "filename": "Django-1.7.10.tar.gz", - "has_sig": false, - "md5_digest": "90315a9bec9b073a91beeb3f60994600", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7584312, - "upload_time": "2015-08-18T17:15:47", - "upload_time_iso_8601": "2015-08-18T17:15:47.478514Z", - "url": "https://files.pythonhosted.org/packages/b1/8c/57e58359f201c8b3b46706bf06a5cef32d0a8176cf287f45ad7b2eb10891/Django-1.7.10.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.7.11": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "cd1a9797706779fc77317887bcf4b12632c24aed8404b694ed1b8d1f7053c92b", - "md5": "9a1e3e2767c5800a9d4e700c4b3aa514", - "sha256": "100164556897c1219f33706e63a656b8848d33d09b0161e2deefcc50978cf62d" - }, - "downloads": -1, - "filename": "Django-1.7.11-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "9a1e3e2767c5800a9d4e700c4b3aa514", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 7424228, - "upload_time": "2015-11-24T17:19:47", - "upload_time_iso_8601": "2015-11-24T17:19:47.417001Z", - "url": "https://files.pythonhosted.org/packages/cd/1a/9797706779fc77317887bcf4b12632c24aed8404b694ed1b8d1f7053c92b/Django-1.7.11-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "532acca393c0d65813f9f8e4d41e4a4d7d64fe8a6bd6dd7276583e2b6653af61", - "md5": "030b2f9c99a6e4e0418eadf7dba9e235", - "sha256": "2039144fce8f1b603d03fa5a5643578df1ad007c4ed41a617f02a3943f7059a1" - }, - "downloads": -1, - "filename": "Django-1.7.11.tar.gz", - "has_sig": false, - "md5_digest": "030b2f9c99a6e4e0418eadf7dba9e235", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7586798, - "upload_time": "2015-11-24T17:20:22", - "upload_time_iso_8601": "2015-11-24T17:20:22.836539Z", - "url": "https://files.pythonhosted.org/packages/53/2a/cca393c0d65813f9f8e4d41e4a4d7d64fe8a6bd6dd7276583e2b6653af61/Django-1.7.11.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.7.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "bbe397cfd29ab86059c1df12a5a2593d214d52b9c22f6f40e2f0dc891b1899fe", - "md5": "b57f9a2dec214b60e338aa80fb902936", - "sha256": "b22871edc9ddf3e57b18989c3c7e9174b4c168dc7b8dbe3f31d4101a73bf2006" - }, - "downloads": -1, - "filename": "Django-1.7.2-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "b57f9a2dec214b60e338aa80fb902936", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 7420120, - "upload_time": "2015-01-03T01:37:27", - "upload_time_iso_8601": "2015-01-03T01:37:27.712134Z", - "url": "https://files.pythonhosted.org/packages/bb/e3/97cfd29ab86059c1df12a5a2593d214d52b9c22f6f40e2f0dc891b1899fe/Django-1.7.2-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "32283e17392530e8bc38db501a9d85d2013e260657a785fc0b568f2c1bcf2a9e", - "md5": "855a53a9a5581c62b6031c9b3bd80ec5", - "sha256": "31c6c3c229f8c04b3be87e6afc3492903b57ec8f1188a47b6ae160d90cf653c8" - }, - "downloads": -1, - "filename": "Django-1.7.2.tar.gz", - "has_sig": false, - "md5_digest": "855a53a9a5581c62b6031c9b3bd80ec5", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7577911, - "upload_time": "2015-01-03T01:37:40", - "upload_time_iso_8601": "2015-01-03T01:37:40.931378Z", - "url": "https://files.pythonhosted.org/packages/32/28/3e17392530e8bc38db501a9d85d2013e260657a785fc0b568f2c1bcf2a9e/Django-1.7.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.7.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "4db9b972b87a27330e6647abb9766c1d32e5726430cfcc60b2b9c2f2d0e28bcb", - "md5": "bd24beec81e161d30ad925aef9d23e57", - "sha256": "72edd47b55ae748d29f1a71d5ca4b86e785c9fb974407cf242b3168e6f1b177e" - }, - "downloads": -1, - "filename": "Django-1.7.3-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "bd24beec81e161d30ad925aef9d23e57", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 7421196, - "upload_time": "2015-01-13T18:39:44", - "upload_time_iso_8601": "2015-01-13T18:39:44.032666Z", - "url": "https://files.pythonhosted.org/packages/4d/b9/b972b87a27330e6647abb9766c1d32e5726430cfcc60b2b9c2f2d0e28bcb/Django-1.7.3-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "4957e57dc454702a565815160fec24cff7a833331c6b603afb6e30f2102ca29c", - "md5": "ea9a3fe7eca2280b233938a98c4a35a0", - "sha256": "f226fb8aa438456968d403f6739de1cf2dad128db86f66ee2b41dfebe3645c5b" - }, - "downloads": -1, - "filename": "Django-1.7.3.tar.gz", - "has_sig": false, - "md5_digest": "ea9a3fe7eca2280b233938a98c4a35a0", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7589559, - "upload_time": "2015-01-13T18:39:56", - "upload_time_iso_8601": "2015-01-13T18:39:56.760858Z", - "url": "https://files.pythonhosted.org/packages/49/57/e57dc454702a565815160fec24cff7a833331c6b603afb6e30f2102ca29c/Django-1.7.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.7.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "c91e66f185ca0d4d0ca11b94caeac96a33a13954963a8b563b67d11f50bfeee7", - "md5": "f465b25daeaa559ffc329f1e5daaa520", - "sha256": "42002065cc98bc99b5bae0084ddb13a0ad611a171dcdaf5ba96731935da744a4" - }, - "downloads": -1, - "filename": "Django-1.7.4-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "f465b25daeaa559ffc329f1e5daaa520", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 7422181, - "upload_time": "2015-01-27T17:22:19", - "upload_time_iso_8601": "2015-01-27T17:22:19.680252Z", - "url": "https://files.pythonhosted.org/packages/c9/1e/66f185ca0d4d0ca11b94caeac96a33a13954963a8b563b67d11f50bfeee7/Django-1.7.4-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "f3af91bf47f38841c1b64875cb6ff40c6466e71ed8c0864ef837e1db5c4fbac8", - "md5": "f8db10520f0268747d402a47a1a4b191", - "sha256": "f33255afbb9ee0977d9095ab0b50fde1f8ddff4220b57e8d19c6620b3e316170" - }, - "downloads": -1, - "filename": "Django-1.7.4.tar.gz", - "has_sig": false, - "md5_digest": "f8db10520f0268747d402a47a1a4b191", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7592584, - "upload_time": "2015-01-27T17:22:47", - "upload_time_iso_8601": "2015-01-27T17:22:47.486790Z", - "url": "https://files.pythonhosted.org/packages/f3/af/91bf47f38841c1b64875cb6ff40c6466e71ed8c0864ef837e1db5c4fbac8/Django-1.7.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.7.5": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "4ddc7b29fa98915a25a443eb5091aff1a0ef007d27c9848365a1fd0a108b5f39", - "md5": "d6b529414f3093c848a69996979a1bea", - "sha256": "ac85d95150ad013632bef5e7ff9ea6decce8100713945d3c61a1b671e6c11d2c" - }, - "downloads": -1, - "filename": "Django-1.7.5-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "d6b529414f3093c848a69996979a1bea", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 7422703, - "upload_time": "2015-02-25T13:58:22", - "upload_time_iso_8601": "2015-02-25T13:58:22.048763Z", - "url": "https://files.pythonhosted.org/packages/4d/dc/7b29fa98915a25a443eb5091aff1a0ef007d27c9848365a1fd0a108b5f39/Django-1.7.5-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "d1ab8d0fe798860f897e388dbb5a2df468db47c52645cbbd378963eff508cf91", - "md5": "e76c70a5dd7d56a511974b28ab38df20", - "sha256": "1c391f9349c97df503dac3461599f24235e4d04393498e6060e74dd2721460bc" - }, - "downloads": -1, - "filename": "Django-1.7.5.tar.gz", - "has_sig": false, - "md5_digest": "e76c70a5dd7d56a511974b28ab38df20", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7599017, - "upload_time": "2015-02-25T13:58:38", - "upload_time_iso_8601": "2015-02-25T13:58:38.544144Z", - "url": "https://files.pythonhosted.org/packages/d1/ab/8d0fe798860f897e388dbb5a2df468db47c52645cbbd378963eff508cf91/Django-1.7.5.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.7.6": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "f71df2928f2be4da63a6304d56dcfde893fcd68cae49a70d35d2fda625f66525", - "md5": "9ede80f1a893835ffecce15b23534d82", - "sha256": "f82fb42e84ca28dcbb1149a933ec9c2958546f338f657dd59307db4da0a53c3e" - }, - "downloads": -1, - "filename": "Django-1.7.6-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "9ede80f1a893835ffecce15b23534d82", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 7422729, - "upload_time": "2015-03-09T15:30:40", - "upload_time_iso_8601": "2015-03-09T15:30:40.158699Z", - "url": "https://files.pythonhosted.org/packages/f7/1d/f2928f2be4da63a6304d56dcfde893fcd68cae49a70d35d2fda625f66525/Django-1.7.6-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "f313f7ac67daff05d52bc4db7163040cc88aaf9dcb9cb65349056276f08bf536", - "md5": "e73ec0ba059a5f24563d785763cae37d", - "sha256": "b0f15e0ffe59a2f37cbaf53543f05d2f40c5a755390df03ec0655b5e4a8d4c90" - }, - "downloads": -1, - "filename": "Django-1.7.6.tar.gz", - "has_sig": false, - "md5_digest": "e73ec0ba059a5f24563d785763cae37d", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7601179, - "upload_time": "2015-03-09T15:30:50", - "upload_time_iso_8601": "2015-03-09T15:30:50.859703Z", - "url": "https://files.pythonhosted.org/packages/f3/13/f7ac67daff05d52bc4db7163040cc88aaf9dcb9cb65349056276f08bf536/Django-1.7.6.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.7.7": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "b56b810aad2c715fa19ffcf30b592cc90b931152268e59c3c526563be0b462e0", - "md5": "52919e0b00a228535a3c4b5a72ac5df5", - "sha256": "eec57d3219501ec6e685646826f2eb8e77687a93fe374a7e6994490520db093e" - }, - "downloads": -1, - "filename": "Django-1.7.7-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "52919e0b00a228535a3c4b5a72ac5df5", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 7422997, - "upload_time": "2015-03-18T23:49:08", - "upload_time_iso_8601": "2015-03-18T23:49:08.712508Z", - "url": "https://files.pythonhosted.org/packages/b5/6b/810aad2c715fa19ffcf30b592cc90b931152268e59c3c526563be0b462e0/Django-1.7.7-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "61522e39a916cb6697fa89cc9143134ec9ef470f4014a16d432044334b7204ff", - "md5": "a62d6598966947d150525ad2ab20fb0c", - "sha256": "4816f892063569ca9a77584fa23cb4995c1b3b954ef875102a8219229cbd2e33" - }, - "downloads": -1, - "filename": "Django-1.7.7.tar.gz", - "has_sig": false, - "md5_digest": "a62d6598966947d150525ad2ab20fb0c", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7603286, - "upload_time": "2015-03-18T23:49:18", - "upload_time_iso_8601": "2015-03-18T23:49:18.866382Z", - "url": "https://files.pythonhosted.org/packages/61/52/2e39a916cb6697fa89cc9143134ec9ef470f4014a16d432044334b7204ff/Django-1.7.7.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.7.8": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "871201d937731539a729cb82ecec6763f0f893782287a66ca5768c44b507a2e7", - "md5": "709db47978acf982f87f73035e9f87a9", - "sha256": "3c63d4e8a6bab7d4eccf41240ce1bcdc59cf92a09499eca75bcb701e342306aa" - }, - "downloads": -1, - "filename": "Django-1.7.8-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "709db47978acf982f87f73035e9f87a9", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 7423353, - "upload_time": "2015-05-01T20:42:44", - "upload_time_iso_8601": "2015-05-01T20:42:44.291666Z", - "url": "https://files.pythonhosted.org/packages/87/12/01d937731539a729cb82ecec6763f0f893782287a66ca5768c44b507a2e7/Django-1.7.8-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "b2a896ffe1a3755e4d7763f014dda866077697ee949a63ac7980c4432b77aa8a", - "md5": "ceed714b0d69983da2b585ff4f82c4a8", - "sha256": "f0ab12c7c88a033681e44e2e4bf4a93d93c85d5b1e8e9c875b4b917abb246921" - }, - "downloads": -1, - "filename": "Django-1.7.8.tar.gz", - "has_sig": false, - "md5_digest": "ceed714b0d69983da2b585ff4f82c4a8", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7604685, - "upload_time": "2015-05-01T20:42:55", - "upload_time_iso_8601": "2015-05-01T20:42:55.907784Z", - "url": "https://files.pythonhosted.org/packages/b2/a8/96ffe1a3755e4d7763f014dda866077697ee949a63ac7980c4432b77aa8a/Django-1.7.8.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.7.9": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "cd69f41554b2d76bddd1de18b6c2f7729c18fa823f24f2140f995599e2b9f43f", - "md5": "2df76c90453477a9d0bb6a1608f0dfa5", - "sha256": "4d58e452744b76876d8199ca44144a252c254e8914939240ae3527c4eaa1882f" - }, - "downloads": -1, - "filename": "Django-1.7.9-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "2df76c90453477a9d0bb6a1608f0dfa5", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 7423771, - "upload_time": "2015-07-08T21:33:04", - "upload_time_iso_8601": "2015-07-08T21:33:04.180643Z", - "url": "https://files.pythonhosted.org/packages/cd/69/f41554b2d76bddd1de18b6c2f7729c18fa823f24f2140f995599e2b9f43f/Django-1.7.9-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "e6f3544641b38e4ac212562b0c9722e8ee5e3ec07e6e2ba644a41d5044f2a9ad", - "md5": "6ea69f3ebb73755bd2a4c9e3743f17c8", - "sha256": "4f3f9fe4e5d20ff8ed6a90b5d2f2df2d8fc054e478cdcc3db81c6b29bd217860" - }, - "downloads": -1, - "filename": "Django-1.7.9.tar.gz", - "has_sig": false, - "md5_digest": "6ea69f3ebb73755bd2a4c9e3743f17c8", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7605194, - "upload_time": "2015-07-08T19:50:47", - "upload_time_iso_8601": "2015-07-08T19:50:47.243246Z", - "url": "https://files.pythonhosted.org/packages/e6/f3/544641b38e4ac212562b0c9722e8ee5e3ec07e6e2ba644a41d5044f2a9ad/Django-1.7.9.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "4e1c17a429cfb79c1814d1ec31939fc5cf4a8ac68fe934279e095fb6160123a9", - "md5": "6f3df764d5826f922eab2d8131c81e60", - "sha256": "6a03ce2feafdd193a0ba8a26dbd9773e757d2e5d5e7933a62eac129813bd381a" - }, - "downloads": -1, - "filename": "Django-1.8-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "6f3df764d5826f922eab2d8131c81e60", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6154536, - "upload_time": "2015-04-01T20:12:37", - "upload_time_iso_8601": "2015-04-01T20:12:37.288238Z", - "url": "https://files.pythonhosted.org/packages/4e/1c/17a429cfb79c1814d1ec31939fc5cf4a8ac68fe934279e095fb6160123a9/Django-1.8-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "7e8d505c0673f65d5aec80012d27ba7924054b665ff88e2babab93e1e7c77663", - "md5": "9a811faf67ca0f3e0d43e670a1cc503d", - "sha256": "066bad42cb4c66944e7efcf7304d3d17f7b0eb222e53958cdd866420d2e8b412" - }, - "downloads": -1, - "filename": "Django-1.8.tar.gz", - "has_sig": false, - "md5_digest": "9a811faf67ca0f3e0d43e670a1cc503d", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7258871, - "upload_time": "2015-04-01T20:12:48", - "upload_time_iso_8601": "2015-04-01T20:12:48.080635Z", - "url": "https://files.pythonhosted.org/packages/7e/8d/505c0673f65d5aec80012d27ba7924054b665ff88e2babab93e1e7c77663/Django-1.8.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "0aa94cc0b81cb74da397e0c6374322442f43e54da0af4da1e870b660c99ba58a", - "md5": "0a16060bf7e3e9ad1354b3b5d9a6e6c4", - "sha256": "d92ad85a8684d86d078312acb0860824861a3cfcb482428f40878421f2253398" - }, - "downloads": -1, - "filename": "Django-1.8.1-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "0a16060bf7e3e9ad1354b3b5d9a6e6c4", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6163056, - "upload_time": "2015-05-01T20:36:34", - "upload_time_iso_8601": "2015-05-01T20:36:34.829219Z", - "url": "https://files.pythonhosted.org/packages/0a/a9/4cc0b81cb74da397e0c6374322442f43e54da0af4da1e870b660c99ba58a/Django-1.8.1-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "42fee0e7d13a0cb47f0af9f36e2523438d85b8d30ed640b72ace2629bd75b181", - "md5": "0f0a677a2cd56b9ab7ccb1c562d70f53", - "sha256": "c6c7e7a961e2847d050d214ca96dc3167bb5f2b25cd5c6cb2eea96e1717f4ade" - }, - "downloads": -1, - "filename": "Django-1.8.1.tar.gz", - "has_sig": false, - "md5_digest": "0f0a677a2cd56b9ab7ccb1c562d70f53", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7270084, - "upload_time": "2015-05-01T20:36:45", - "upload_time_iso_8601": "2015-05-01T20:36:45.639030Z", - "url": "https://files.pythonhosted.org/packages/42/fe/e0e7d13a0cb47f0af9f36e2523438d85b8d30ed640b72ace2629bd75b181/Django-1.8.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8.10": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "4d67af6637a50da0a9aee3c032b7333c670734ecc0df6f7c7562fd8b3fdcac22", - "md5": "589ccb8fe8d75c1c0472473f7c11be5b", - "sha256": "471b41cb53d675138475b488c429424ed143e57ad755a2c8ab1206ac30490284" - }, - "downloads": -1, - "filename": "Django-1.8.10-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "589ccb8fe8d75c1c0472473f7c11be5b", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6171409, - "upload_time": "2016-03-01T17:10:07", - "upload_time_iso_8601": "2016-03-01T17:10:07.758955Z", - "url": "https://files.pythonhosted.org/packages/4d/67/af6637a50da0a9aee3c032b7333c670734ecc0df6f7c7562fd8b3fdcac22/Django-1.8.10-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "5252a0454fb9a5d0b4c866e5ab917484b7e3e79f0ae9fbbc1639835a78186df1", - "md5": "8be1858dfee4878768ce686165e29c89", - "sha256": "d2e5b11eb694984957378419f809b7205598326373d509d0262f9f84b17d1a23" - }, - "downloads": -1, - "filename": "Django-1.8.10.tar.gz", - "has_sig": false, - "md5_digest": "8be1858dfee4878768ce686165e29c89", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7291016, - "upload_time": "2016-03-01T17:10:34", - "upload_time_iso_8601": "2016-03-01T17:10:34.083677Z", - "url": "https://files.pythonhosted.org/packages/52/52/a0454fb9a5d0b4c866e5ab917484b7e3e79f0ae9fbbc1639835a78186df1/Django-1.8.10.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8.11": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "b8c2e162862650c30999ab39e05c7b12794d3f80e6b5a99c786419eddbb28fa3", - "md5": "7d8f0eeb83c497a9d6a4820bfe8f0d36", - "sha256": "54be9d6eab6cc0e2da558c12aea6cff7d5a0124c8a470e1ff61134ba9ed37f20" - }, - "downloads": -1, - "filename": "Django-1.8.11-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "7d8f0eeb83c497a9d6a4820bfe8f0d36", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6171434, - "upload_time": "2016-03-05T18:36:08", - "upload_time_iso_8601": "2016-03-05T18:36:08.828432Z", - "url": "https://files.pythonhosted.org/packages/b8/c2/e162862650c30999ab39e05c7b12794d3f80e6b5a99c786419eddbb28fa3/Django-1.8.11-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "924146db17c36b6eaab3dbf600f094b2e3643f96c78d2cfcbeb4fe97d27ce03e", - "md5": "1c85d0d582dc211adc6851dd2dc86228", - "sha256": "ec148be73548da090dd76c2e8c57c98e8b1e84f2cb87500b9be5420187a435fb" - }, - "downloads": -1, - "filename": "Django-1.8.11.tar.gz", - "has_sig": false, - "md5_digest": "1c85d0d582dc211adc6851dd2dc86228", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7292193, - "upload_time": "2016-03-05T18:36:25", - "upload_time_iso_8601": "2016-03-05T18:36:25.916378Z", - "url": "https://files.pythonhosted.org/packages/92/41/46db17c36b6eaab3dbf600f094b2e3643f96c78d2cfcbeb4fe97d27ce03e/Django-1.8.11.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8.12": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "d50d445186a82bbcc75166a507eff586df683c73641e7d6bb7424a44426dca71", - "md5": "c8f867d0aaf25e5031b6a6e1137dfea1", - "sha256": "9c60c4af02faffb6f1c2bc7c7a09169c59230fa06c30a552414b816ee79f0c2a" - }, - "downloads": -1, - "filename": "Django-1.8.12-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "c8f867d0aaf25e5031b6a6e1137dfea1", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6171522, - "upload_time": "2016-04-01T17:54:27", - "upload_time_iso_8601": "2016-04-01T17:54:27.137333Z", - "url": "https://files.pythonhosted.org/packages/d5/0d/445186a82bbcc75166a507eff586df683c73641e7d6bb7424a44426dca71/Django-1.8.12-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "c6a23e3870966c64a46e0d974d4a40eb60f921a7cbb195d38ff00328899dcc7a", - "md5": "905c4b8eeb180c1710331f6fd65fa378", - "sha256": "b68fa73d537f8362d73fec1aa2b7a1e8572349b12942ac756ec1041b6b0e7113" - }, - "downloads": -1, - "filename": "Django-1.8.12.tar.gz", - "has_sig": false, - "md5_digest": "905c4b8eeb180c1710331f6fd65fa378", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7294654, - "upload_time": "2016-04-01T17:54:40", - "upload_time_iso_8601": "2016-04-01T17:54:40.693793Z", - "url": "https://files.pythonhosted.org/packages/c6/a2/3e3870966c64a46e0d974d4a40eb60f921a7cbb195d38ff00328899dcc7a/Django-1.8.12.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8.13": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "2208904238c9c53654ff92024424ed545755873b127d73fe6393fa8ff0433815", - "md5": "da9f8cec98670fc2e22b54b6b95ddee9", - "sha256": "dad5da0cd7f3cca7da3ac42a19abba30f5cc10fae4976e474051e7085b4e95d1" - }, - "downloads": -1, - "filename": "Django-1.8.13-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "da9f8cec98670fc2e22b54b6b95ddee9", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6171517, - "upload_time": "2016-05-02T22:49:58", - "upload_time_iso_8601": "2016-05-02T22:49:58.426172Z", - "url": "https://files.pythonhosted.org/packages/22/08/904238c9c53654ff92024424ed545755873b127d73fe6393fa8ff0433815/Django-1.8.13-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "850e0200c2c3792a4fb8bda6a8597d267abaeb00d862e3a110f99cb101398d4d", - "md5": "a77e1ba9991f762de20bf03de57e39eb", - "sha256": "128e8bdc11c69ea90f778435d38126453d3bf283dbd28cf15a33aa8e52245df4" - }, - "downloads": -1, - "filename": "Django-1.8.13.tar.gz", - "has_sig": false, - "md5_digest": "a77e1ba9991f762de20bf03de57e39eb", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7293834, - "upload_time": "2016-05-02T22:51:42", - "upload_time_iso_8601": "2016-05-02T22:51:42.144726Z", - "url": "https://files.pythonhosted.org/packages/85/0e/0200c2c3792a4fb8bda6a8597d267abaeb00d862e3a110f99cb101398d4d/Django-1.8.13.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8.14": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "121366eeba22d40f86d6cecc5a12784ae84b53f2ba171c448b1646ede25a99cd", - "md5": "a7cc0445cd937d1f9e3c1a47c783af73", - "sha256": "1716747f7e0fbee6e2c1c0bcdb74307139d441a79eb4dcc97d206c615e1ded15" - }, - "downloads": -1, - "filename": "Django-1.8.14-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "a7cc0445cd937d1f9e3c1a47c783af73", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6171120, - "upload_time": "2016-07-18T18:38:12", - "upload_time_iso_8601": "2016-07-18T18:38:12.976852Z", - "url": "https://files.pythonhosted.org/packages/12/13/66eeba22d40f86d6cecc5a12784ae84b53f2ba171c448b1646ede25a99cd/Django-1.8.14-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "4fa50fb863c3b83f8a15b50731d02d835cb15fef93193c7cbbaacbb2b6adf1e0", - "md5": "e275863f0d629831d99c58be1a7ed268", - "sha256": "5282c48b90fbb29507299f592215378e849a2808f485e4144626e66715d5464d" - }, - "downloads": -1, - "filename": "Django-1.8.14.tar.gz", - "has_sig": false, - "md5_digest": "e275863f0d629831d99c58be1a7ed268", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7293650, - "upload_time": "2016-07-18T18:38:36", - "upload_time_iso_8601": "2016-07-18T18:38:36.642604Z", - "url": "https://files.pythonhosted.org/packages/4f/a5/0fb863c3b83f8a15b50731d02d835cb15fef93193c7cbbaacbb2b6adf1e0/Django-1.8.14.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8.15": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "f3bb865626b78c2e87acd4dc1bd8f8cb0a49582ec65e40968f3d24f594bba600", - "md5": "88e80fdc30e8a8a8a2e538d4e49d28a3", - "sha256": "e2e41aeb4fb757575021621dc28fceb9ad137879ae0b854067f1726d9a772807" - }, - "downloads": -1, - "filename": "Django-1.8.15-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "88e80fdc30e8a8a8a2e538d4e49d28a3", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6171252, - "upload_time": "2016-09-26T18:30:16", - "upload_time_iso_8601": "2016-09-26T18:30:16.752726Z", - "url": "https://files.pythonhosted.org/packages/f3/bb/865626b78c2e87acd4dc1bd8f8cb0a49582ec65e40968f3d24f594bba600/Django-1.8.15-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "903b659733fa8eec56780e5f268efc33d4c57c5f062a2d659fbb95467ba9da46", - "md5": "d24c3c5fc6d784296693659b05efa70f", - "sha256": "863e543ac985d5cfbce09213fa30bc7c802cbdf60d6db8b5f9dab41e1341eacd" - }, - "downloads": -1, - "filename": "Django-1.8.15.tar.gz", - "has_sig": false, - "md5_digest": "d24c3c5fc6d784296693659b05efa70f", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7295362, - "upload_time": "2016-09-26T18:30:30", - "upload_time_iso_8601": "2016-09-26T18:30:30.224525Z", - "url": "https://files.pythonhosted.org/packages/90/3b/659733fa8eec56780e5f268efc33d4c57c5f062a2d659fbb95467ba9da46/Django-1.8.15.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8.16": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "5c387324791951945b7fa0ba5c6dd57cff12ac544a260e110b2509cda6ced2f8", - "md5": "af3c3489295bfbc443eebcfd9454376f", - "sha256": "cc3a95187788627dfdc94b41de908aadfc4241fabb3ceaef19f4bd3b89c0fdf7" - }, - "downloads": -1, - "filename": "Django-1.8.16-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "af3c3489295bfbc443eebcfd9454376f", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6173515, - "upload_time": "2016-11-01T14:09:28", - "upload_time_iso_8601": "2016-11-01T14:09:28.452420Z", - "url": "https://files.pythonhosted.org/packages/5c/38/7324791951945b7fa0ba5c6dd57cff12ac544a260e110b2509cda6ced2f8/Django-1.8.16-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "1a83daf33e1e897236d1a27cde3cbb49da03b7eec57fcbd658a3b05696b156d4", - "md5": "34cad45c7d372313d5fdb6c8fd1a96ee", - "sha256": "224aaf17a28609707d942deafe6d0a5b382baf22a6f33e4e61c56c62f09081dd" - }, - "downloads": -1, - "filename": "Django-1.8.16.tar.gz", - "has_sig": false, - "md5_digest": "34cad45c7d372313d5fdb6c8fd1a96ee", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7299872, - "upload_time": "2016-11-01T14:09:38", - "upload_time_iso_8601": "2016-11-01T14:09:38.084055Z", - "url": "https://files.pythonhosted.org/packages/1a/83/daf33e1e897236d1a27cde3cbb49da03b7eec57fcbd658a3b05696b156d4/Django-1.8.16.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8.17": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "0adf5ac011953ca547a1fb602e3e5c96a4cb94f7333e8f27b13d51dca880a110", - "md5": "a9dc59b63114424009c429d48884413b", - "sha256": "87618c1011712faf7400e2a73315f4f4c3a6e68ab6309c3e642d5fef73d66d9e" - }, - "downloads": -1, - "filename": "Django-1.8.17-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "a9dc59b63114424009c429d48884413b", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6171460, - "upload_time": "2016-12-01T23:03:26", - "upload_time_iso_8601": "2016-12-01T23:03:26.958750Z", - "url": "https://files.pythonhosted.org/packages/0a/df/5ac011953ca547a1fb602e3e5c96a4cb94f7333e8f27b13d51dca880a110/Django-1.8.17-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "4441bf93934082e9897a56a591a67bacbd9fb74e71244f3f42253432a9e627e6", - "md5": "e76842cdfbcb31286bd44f51e087a04c", - "sha256": "021bd648fcf454027063187e63a1ab4136c6929430ef5dfbe36235f60015eb07" - }, - "downloads": -1, - "filename": "Django-1.8.17.tar.gz", - "has_sig": false, - "md5_digest": "e76842cdfbcb31286bd44f51e087a04c", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7298057, - "upload_time": "2016-12-01T23:03:42", - "upload_time_iso_8601": "2016-12-01T23:03:42.965764Z", - "url": "https://files.pythonhosted.org/packages/44/41/bf93934082e9897a56a591a67bacbd9fb74e71244f3f42253432a9e627e6/Django-1.8.17.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8.18": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "4bfc13e6c9279a5245be456da4cc73146c7bc76d26484b1474bf4a29d7cd2e93", - "md5": "33ddd1967cf1d8be6959baaf5def09e2", - "sha256": "d8e2fd119756ab10b43a31052c3c8efbc262064b81eecb7871372de4d37b1a94" - }, - "downloads": -1, - "filename": "Django-1.8.18-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "33ddd1967cf1d8be6959baaf5def09e2", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6171967, - "upload_time": "2017-04-04T14:07:38", - "upload_time_iso_8601": "2017-04-04T14:07:38.852519Z", - "url": "https://files.pythonhosted.org/packages/4b/fc/13e6c9279a5245be456da4cc73146c7bc76d26484b1474bf4a29d7cd2e93/Django-1.8.18-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "58df55c6e7761a8c144f5bf9629ed3f066dc9fbe39df3f7a1ca4af093a0911ba", - "md5": "ffc3767f5a06c346fd2d07a18c0ebc54", - "sha256": "c7611cdd5e2539a443b7960c7cafd867d986c2720a1b44808deaa60ce3da50c7" - }, - "downloads": -1, - "filename": "Django-1.8.18.tar.gz", - "has_sig": false, - "md5_digest": "ffc3767f5a06c346fd2d07a18c0ebc54", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7297986, - "upload_time": "2017-04-04T14:07:49", - "upload_time_iso_8601": "2017-04-04T14:07:49.913438Z", - "url": "https://files.pythonhosted.org/packages/58/df/55c6e7761a8c144f5bf9629ed3f066dc9fbe39df3f7a1ca4af093a0911ba/Django-1.8.18.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8.19": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "96b9b4108da1275dc2ac1bba1e87739cb31b3d44339affb83b0e949fb09c2bef", - "md5": "2541acc9a25b258affcd5e9bd3575f81", - "sha256": "674c525d3aa90ed683313b64aa27490c31874e16155e6b44772d84e76c83c46c" - }, - "downloads": -1, - "filename": "Django-1.8.19-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "2541acc9a25b258affcd5e9bd3575f81", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6172137, - "upload_time": "2018-03-06T14:22:46", - "upload_time_iso_8601": "2018-03-06T14:22:46.480552Z", - "url": "https://files.pythonhosted.org/packages/96/b9/b4108da1275dc2ac1bba1e87739cb31b3d44339affb83b0e949fb09c2bef/Django-1.8.19-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "adda980dbd68970fefbdf9c62faeed5da1d8ed49214ff3ea3991c2d233719b51", - "md5": "25eafb0a2708ff9990c02f5ce0a11db4", - "sha256": "33d44a5cf9d333247a9a374ae1478b78b83c9b78eb316fc04adde62053b4c047" - }, - "downloads": -1, - "filename": "Django-1.8.19.tar.gz", - "has_sig": false, - "md5_digest": "25eafb0a2708ff9990c02f5ce0a11db4", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7359244, - "upload_time": "2018-03-06T14:23:15", - "upload_time_iso_8601": "2018-03-06T14:23:15.360659Z", - "url": "https://files.pythonhosted.org/packages/ad/da/980dbd68970fefbdf9c62faeed5da1d8ed49214ff3ea3991c2d233719b51/Django-1.8.19.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "4e9d2a1835ccbf8e1f0d6755d0e938ffd855f23886d055a7a18cc00a5224a99b", - "md5": "ef4e1c047ec900ae321126b22c7659f2", - "sha256": "bd57d950778db81f55f89efcbcb905ee839a778ba790ae4308b8a316835eb7ce" - }, - "downloads": -1, - "filename": "Django-1.8.2-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "ef4e1c047ec900ae321126b22c7659f2", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6164181, - "upload_time": "2015-05-20T18:02:14", - "upload_time_iso_8601": "2015-05-20T18:02:14.621999Z", - "url": "https://files.pythonhosted.org/packages/4e/9d/2a1835ccbf8e1f0d6755d0e938ffd855f23886d055a7a18cc00a5224a99b/Django-1.8.2-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "b3c01ee1c6a2e3ebf6c98fd401e2a039e515308595ddff7525357eeadafb4cb3", - "md5": "ec4330cd275dd6ce64230feebcb449c4", - "sha256": "3bb60536b2fb2084612fc9486634295e7208790029081842524916b5a66d206f" - }, - "downloads": -1, - "filename": "Django-1.8.2.tar.gz", - "has_sig": false, - "md5_digest": "ec4330cd275dd6ce64230feebcb449c4", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7275112, - "upload_time": "2015-05-20T18:02:30", - "upload_time_iso_8601": "2015-05-20T18:02:30.286562Z", - "url": "https://files.pythonhosted.org/packages/b3/c0/1ee1c6a2e3ebf6c98fd401e2a039e515308595ddff7525357eeadafb4cb3/Django-1.8.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "a3e10f3c17b1caa559ba69513ff72e250377c268d5bd3e8ad2b22809c7e2e907", - "md5": "a5d397c65a880228c58a443070cc18a8", - "sha256": "047d0f4c93262b33801049a2dcddaef09c29e741c03a947a3556ea4748eed2e2" - }, - "downloads": -1, - "filename": "Django-1.8.3-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "a5d397c65a880228c58a443070cc18a8", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6165665, - "upload_time": "2015-07-08T19:43:24", - "upload_time_iso_8601": "2015-07-08T19:43:24.197330Z", - "url": "https://files.pythonhosted.org/packages/a3/e1/0f3c17b1caa559ba69513ff72e250377c268d5bd3e8ad2b22809c7e2e907/Django-1.8.3-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "9eed40a000533fd6e5147f326203163650bff9cd4e4432c5664a9799b95bddd1", - "md5": "31760322115c3ae51fbd8ac85c9ac428", - "sha256": "2bb654fcc05fd53017c88caf2bc38b5c5ea23c91f8ac7f0a28b290daf2305bba" - }, - "downloads": -1, - "filename": "Django-1.8.3.tar.gz", - "has_sig": false, - "md5_digest": "31760322115c3ae51fbd8ac85c9ac428", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7284327, - "upload_time": "2015-07-08T19:43:35", - "upload_time_iso_8601": "2015-07-08T19:43:35.397608Z", - "url": "https://files.pythonhosted.org/packages/9e/ed/40a000533fd6e5147f326203163650bff9cd4e4432c5664a9799b95bddd1/Django-1.8.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "7ec32574a681af30d99b8fd60008c3e56f4ab0acad2af70fea6c135b8bff60a2", - "md5": "3a586ed7da0715ea65ec94a9b1ff2d87", - "sha256": "2376c3d8c5f495b302b112d7232c84761130c430e1840c05a2a02b28f17dd596" - }, - "downloads": -1, - "filename": "Django-1.8.4-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "3a586ed7da0715ea65ec94a9b1ff2d87", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6167091, - "upload_time": "2015-08-18T17:06:21", - "upload_time_iso_8601": "2015-08-18T17:06:21.380689Z", - "url": "https://files.pythonhosted.org/packages/7e/c3/2574a681af30d99b8fd60008c3e56f4ab0acad2af70fea6c135b8bff60a2/Django-1.8.4-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "da31cc1ab8109387d178bfa0d1f85a6a6175b07c72b63db92f42c27aab83df3c", - "md5": "8eb569a5b9d984d9f3366fda67fb0bb8", - "sha256": "826996c81e1cc773500124d5c19212e4a7681a55ee169fab9085f2b3015a70d8" - }, - "downloads": -1, - "filename": "Django-1.8.4.tar.gz", - "has_sig": false, - "md5_digest": "8eb569a5b9d984d9f3366fda67fb0bb8", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7265101, - "upload_time": "2015-08-18T17:06:55", - "upload_time_iso_8601": "2015-08-18T17:06:55.158421Z", - "url": "https://files.pythonhosted.org/packages/da/31/cc1ab8109387d178bfa0d1f85a6a6175b07c72b63db92f42c27aab83df3c/Django-1.8.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8.5": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "69cc9aa13faa16849cdf0b27e5ad9b1a9f82d1c1136c88382f24fe07b4290e35", - "md5": "3c182cf9de00382ecf27fdc65fcfbe70", - "sha256": "644c5cc2d064a38d439627549e6382a8bc28b80a243fe5cbba40b1efd2babdeb" - }, - "downloads": -1, - "filename": "Django-1.8.5-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "3c182cf9de00382ecf27fdc65fcfbe70", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6167038, - "upload_time": "2015-10-04T00:06:01", - "upload_time_iso_8601": "2015-10-04T00:06:01.281154Z", - "url": "https://files.pythonhosted.org/packages/69/cc/9aa13faa16849cdf0b27e5ad9b1a9f82d1c1136c88382f24fe07b4290e35/Django-1.8.5-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "d99ae6ca54fe351085b907551adda560a5494811a6aa1521ddddf72fe657d8e4", - "md5": "02426a28fb356e52006e053503d66490", - "sha256": "2d174e4a3f54708d0d5b6ff1bf54ae71652e83bb06d7576b3b20d916b29c3653" - }, - "downloads": -1, - "filename": "Django-1.8.5.tar.gz", - "has_sig": false, - "md5_digest": "02426a28fb356e52006e053503d66490", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7270297, - "upload_time": "2015-10-04T00:06:18", - "upload_time_iso_8601": "2015-10-04T00:06:18.616258Z", - "url": "https://files.pythonhosted.org/packages/d9/9a/e6ca54fe351085b907551adda560a5494811a6aa1521ddddf72fe657d8e4/Django-1.8.5.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8.6": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "d28b115baec29ab28754da734b89e4a6aa8e5f0588d1b8f0db03d8b9f87bb88c", - "md5": "8d3e35eb13674c1ff77dfcc4accb668b", - "sha256": "c3283f41f25334bbc0279d535218c949da847c8f2ea1dc03e02981f2e813ad31" - }, - "downloads": -1, - "filename": "Django-1.8.6-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "8d3e35eb13674c1ff77dfcc4accb668b", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6169165, - "upload_time": "2015-11-04T17:03:39", - "upload_time_iso_8601": "2015-11-04T17:03:39.883061Z", - "url": "https://files.pythonhosted.org/packages/d2/8b/115baec29ab28754da734b89e4a6aa8e5f0588d1b8f0db03d8b9f87bb88c/Django-1.8.6-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "b124e169968d525c4d5ce20c85cb3a6c094543cbc109aa270dcac0759c6a644f", - "md5": "12ba7b57a1f5268f6e8ba555628c0657", - "sha256": "359d56f55a033a92831eab1f7ec47db3f9ad8e07f28ead9035d961886d54459a" - }, - "downloads": -1, - "filename": "Django-1.8.6.tar.gz", - "has_sig": false, - "md5_digest": "12ba7b57a1f5268f6e8ba555628c0657", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7341303, - "upload_time": "2015-11-04T17:03:52", - "upload_time_iso_8601": "2015-11-04T17:03:52.915842Z", - "url": "https://files.pythonhosted.org/packages/b1/24/e169968d525c4d5ce20c85cb3a6c094543cbc109aa270dcac0759c6a644f/Django-1.8.6.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8.7": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "aac8227d39d4014cdc999a480d341ee2119a74ecc5192d6ce91883e6661d51d4", - "md5": "b0ad232d645c0183de89fae29d12ce4e", - "sha256": "8fb693ecfe4cd6ff9ae3136ff0a1eaa4acae01af227bb81e646dc2bad3295ccf" - }, - "downloads": -1, - "filename": "Django-1.8.7-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "b0ad232d645c0183de89fae29d12ce4e", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6169517, - "upload_time": "2015-11-24T17:28:47", - "upload_time_iso_8601": "2015-11-24T17:28:47.348560Z", - "url": "https://files.pythonhosted.org/packages/aa/c8/227d39d4014cdc999a480d341ee2119a74ecc5192d6ce91883e6661d51d4/Django-1.8.7-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "e15d6922fc7c382be5157b723f7bfe78035ccfd3e53fa21e9bffb44381153be5", - "md5": "44c01355b5efa01938a89b8bd798b1ed", - "sha256": "17a66de5cf59b5ee81c3dc57609b145bb45adddc0dc06937b998597d6e7b4523" - }, - "downloads": -1, - "filename": "Django-1.8.7.tar.gz", - "has_sig": false, - "md5_digest": "44c01355b5efa01938a89b8bd798b1ed", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7276831, - "upload_time": "2015-11-24T17:29:07", - "upload_time_iso_8601": "2015-11-24T17:29:07.982583Z", - "url": "https://files.pythonhosted.org/packages/e1/5d/6922fc7c382be5157b723f7bfe78035ccfd3e53fa21e9bffb44381153be5/Django-1.8.7.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8.8": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "30d6acf4ab0a1e0d4409533853c6b2b008d44b38b8bbae367cc3b1838b49299f", - "md5": "97334c82efbac0f93f8b6dd4ee4b516f", - "sha256": "05816963cbbadd131bcee8fb5069d8695796e25d081ec24eff62bc1fc8ed891a" - }, - "downloads": -1, - "filename": "Django-1.8.8-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "97334c82efbac0f93f8b6dd4ee4b516f", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6170205, - "upload_time": "2016-01-02T14:28:21", - "upload_time_iso_8601": "2016-01-02T14:28:21.103773Z", - "url": "https://files.pythonhosted.org/packages/30/d6/acf4ab0a1e0d4409533853c6b2b008d44b38b8bbae367cc3b1838b49299f/Django-1.8.8-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "1c1c02308e66ee4ce3ad5a2b8f1e7b538b8a8637e0b48e0ad905bfc4105ee585", - "md5": "08ecf83b7e9d064ed7e3981ddc3a8a15", - "sha256": "8255242fa0d9e0bf331259a6bdb81364933acbe8863291661558ffdb2fc9ed70" - }, - "downloads": -1, - "filename": "Django-1.8.8.tar.gz", - "has_sig": false, - "md5_digest": "08ecf83b7e9d064ed7e3981ddc3a8a15", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7286780, - "upload_time": "2016-01-02T14:28:34", - "upload_time_iso_8601": "2016-01-02T14:28:34.130412Z", - "url": "https://files.pythonhosted.org/packages/1c/1c/02308e66ee4ce3ad5a2b8f1e7b538b8a8637e0b48e0ad905bfc4105ee585/Django-1.8.8.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8.9": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "a7a2f4e7926062f5c61d98e87b9d0365ca1a0fab6be599dbea90ce2fab27f395", - "md5": "24fa1e0a2aa82d9e2ebe23c8729fe306", - "sha256": "6481bbb3d410d8a7923e9c9ed23b799f269a6526b2fa83a04b8ba3069f61dd71" - }, - "downloads": -1, - "filename": "Django-1.8.9-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "24fa1e0a2aa82d9e2ebe23c8729fe306", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6170645, - "upload_time": "2016-02-01T17:24:16", - "upload_time_iso_8601": "2016-02-01T17:24:16.379699Z", - "url": "https://files.pythonhosted.org/packages/a7/a2/f4e7926062f5c61d98e87b9d0365ca1a0fab6be599dbea90ce2fab27f395/Django-1.8.9-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "a48c82cfd400e620cedc36ac5677d4c1fdebea688a4b3f5193e45b8ff15169ce", - "md5": "49f6863b1c83825fb2f473c141c28e15", - "sha256": "fc012d8507201a628e877202bb7800799152285f69aa0d42a7c506a96fbbd2e3" - }, - "downloads": -1, - "filename": "Django-1.8.9.tar.gz", - "has_sig": false, - "md5_digest": "49f6863b1c83825fb2f473c141c28e15", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7288701, - "upload_time": "2016-02-01T17:24:29", - "upload_time_iso_8601": "2016-02-01T17:24:29.386552Z", - "url": "https://files.pythonhosted.org/packages/a4/8c/82cfd400e620cedc36ac5677d4c1fdebea688a4b3f5193e45b8ff15169ce/Django-1.8.9.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.8a1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "92e795566ef630009a2b21fb1f0815287aa291be8756eb822f6f29a80cdd6084", - "md5": "f7619792a8d8028c5be10f7d06a444ca", - "sha256": "ecca07bdf863444f955160a822fe3c979e3c64e1eb1d9553a4ba03c2af8864f2" - }, - "downloads": -1, - "filename": "Django-1.8a1-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "f7619792a8d8028c5be10f7d06a444ca", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6899082, - "upload_time": "2015-01-16T22:25:13", - "upload_time_iso_8601": "2015-01-16T22:25:13.083005Z", - "url": "https://files.pythonhosted.org/packages/92/e7/95566ef630009a2b21fb1f0815287aa291be8756eb822f6f29a80cdd6084/Django-1.8a1-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - } - ], - "1.8b1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "76d4a489819b66706406c07819844bae3f7d9613104c1581af2a4ef567e67c48", - "md5": "5239ff2f5a6901e1222b03d45d9c8936", - "sha256": "2cc94282a29ecea856acde4363d2f2f87170f0c30d96c7f953c5c46e2bca367a" - }, - "downloads": -1, - "filename": "Django-1.8b1-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "5239ff2f5a6901e1222b03d45d9c8936", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6521989, - "upload_time": "2015-02-25T13:42:42", - "upload_time_iso_8601": "2015-02-25T13:42:42.788820Z", - "url": "https://files.pythonhosted.org/packages/76/d4/a489819b66706406c07819844bae3f7d9613104c1581af2a4ef567e67c48/Django-1.8b1-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - } - ], - "1.8b2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "abd697affc9b5435234b906009ad11f6ca6e9c10d6c17752a25e82a92f44acca", - "md5": "f585d674396cde011b9cf5878ba0852f", - "sha256": "4f03f3e14c43cdb4002e705ca4631d0623706d2ba2ce51385c61742f31268fc6" - }, - "downloads": -1, - "filename": "Django-1.8b2-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "f585d674396cde011b9cf5878ba0852f", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6522351, - "upload_time": "2015-03-09T15:55:16", - "upload_time_iso_8601": "2015-03-09T15:55:16.864430Z", - "url": "https://files.pythonhosted.org/packages/ab/d6/97affc9b5435234b906009ad11f6ca6e9c10d6c17752a25e82a92f44acca/Django-1.8b2-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - } - ], - "1.8c1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "6806d854d6b1b3ad307571860cc6b1a65ed01325cc7702efdce514695dd6b8af", - "md5": "6288c878e8d643033026f50f456cb05c", - "sha256": "eb11796aa8e0e083409d513bd144002a986cce7bb22041b8df3f2c0734387898" - }, - "downloads": -1, - "filename": "Django-1.8c1-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "6288c878e8d643033026f50f456cb05c", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6410414, - "upload_time": "2015-03-18T23:39:34", - "upload_time_iso_8601": "2015-03-18T23:39:34.451697Z", - "url": "https://files.pythonhosted.org/packages/68/06/d854d6b1b3ad307571860cc6b1a65ed01325cc7702efdce514695dd6b8af/Django-1.8c1-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - } - ], - "1.9": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "ea9bb5a6360b3dfcd88d4bad70f59da26cbde4bdec395a31bb26dc840e806a50", - "md5": "f98b94b9911b397ea3794a05079cbc78", - "sha256": "e66d58bfeed3a5eb44f2af6d5f1b6a85d656c4180ebba63b692e58d29db2a716" - }, - "downloads": -1, - "filename": "Django-1.9-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "f98b94b9911b397ea3794a05079cbc78", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6563557, - "upload_time": "2015-12-01T23:55:38", - "upload_time_iso_8601": "2015-12-01T23:55:38.706647Z", - "url": "https://files.pythonhosted.org/packages/ea/9b/b5a6360b3dfcd88d4bad70f59da26cbde4bdec395a31bb26dc840e806a50/Django-1.9-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "c214e282ae720c21b48316b66126d7295ace0790438b27482b7a3dd9a6e3c3e1", - "md5": "110389cf89196334182295165852e082", - "sha256": "05fe4b19a8778d4b48bbf1f4dfca3106881fea7982664553e7f7f861606f7c66" - }, - "downloads": -1, - "filename": "Django-1.9.tar.gz", - "has_sig": false, - "md5_digest": "110389cf89196334182295165852e082", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7392116, - "upload_time": "2015-12-01T23:55:55", - "upload_time_iso_8601": "2015-12-01T23:55:55.779479Z", - "url": "https://files.pythonhosted.org/packages/c2/14/e282ae720c21b48316b66126d7295ace0790438b27482b7a3dd9a6e3c3e1/Django-1.9.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.9.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "03175e30516ed4b18a4698d457349e17f42605a3786efbef87e45eaa592c033b", - "md5": "fe01a2b006be4a4f1b17b43b9ee3ba4e", - "sha256": "9f7ca04c6dbcf08b794f2ea5283c60156a37ebf2b8316d1027f594f34ff61101" - }, - "downloads": -1, - "filename": "Django-1.9.1-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "fe01a2b006be4a4f1b17b43b9ee3ba4e", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6579261, - "upload_time": "2016-01-02T13:50:19", - "upload_time_iso_8601": "2016-01-02T13:50:19.247096Z", - "url": "https://files.pythonhosted.org/packages/03/17/5e30516ed4b18a4698d457349e17f42605a3786efbef87e45eaa592c033b/Django-1.9.1-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "4d25af35d7a8b623edeac79cbc2fd24b8496eb2f063542e2905f0c564a052303", - "md5": "02754aa2d5c9c171dfc3f9422b20e12c", - "sha256": "a29aac46a686cade6da87ce7e7287d5d53cddabc41d777c6230a583c36244a18" - }, - "downloads": -1, - "filename": "Django-1.9.1.tar.gz", - "has_sig": false, - "md5_digest": "02754aa2d5c9c171dfc3f9422b20e12c", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7411671, - "upload_time": "2016-01-02T13:50:42", - "upload_time_iso_8601": "2016-01-02T13:50:42.992870Z", - "url": "https://files.pythonhosted.org/packages/4d/25/af35d7a8b623edeac79cbc2fd24b8496eb2f063542e2905f0c564a052303/Django-1.9.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.9.10": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "296311d40144c1514959052a3ba2102d2bb22f9dd336d6aaaa3c949b952cdf2d", - "md5": "f7134ee0e05b1d97ee070eb639f76fc9", - "sha256": "bb9c4ddc128b3a7a7c61534de9c0de5c1bf0ccc67c5e7a07533ff5e4d4f11c4b" - }, - "downloads": -1, - "filename": "Django-1.9.10-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "f7134ee0e05b1d97ee070eb639f76fc9", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6622276, - "upload_time": "2016-09-26T18:32:46", - "upload_time_iso_8601": "2016-09-26T18:32:46.350902Z", - "url": "https://files.pythonhosted.org/packages/29/63/11d40144c1514959052a3ba2102d2bb22f9dd336d6aaaa3c949b952cdf2d/Django-1.9.10-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "4acf43ae852033f739fdef007807a3871c733f21060d342295c3f4ae47c5e0a1", - "md5": "2dd882995873ecef366696d2793f993c", - "sha256": "5b3a7c9e5a21b487f16b7d52d12aab8806f0b93be1a0fbdc333acc05f515fc00" - }, - "downloads": -1, - "filename": "Django-1.9.10.tar.gz", - "has_sig": false, - "md5_digest": "2dd882995873ecef366696d2793f993c", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7494610, - "upload_time": "2016-09-26T18:34:45", - "upload_time_iso_8601": "2016-09-26T18:34:45.917564Z", - "url": "https://files.pythonhosted.org/packages/4a/cf/43ae852033f739fdef007807a3871c733f21060d342295c3f4ae47c5e0a1/Django-1.9.10.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.9.11": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "90b3a00f97b60355df2983e71ccf06e08db2acd70c2a2b62c2c0c5bffda70cda", - "md5": "1e5bc4d09a1801636bcaf3ebbc49598c", - "sha256": "bec3e58d9458e3121180adf9c33dedae0091ef6e73f79b2f9a0f8c0a34925429" - }, - "downloads": -1, - "filename": "Django-1.9.11-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "1e5bc4d09a1801636bcaf3ebbc49598c", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6622478, - "upload_time": "2016-11-01T14:02:12", - "upload_time_iso_8601": "2016-11-01T14:02:12.102715Z", - "url": "https://files.pythonhosted.org/packages/90/b3/a00f97b60355df2983e71ccf06e08db2acd70c2a2b62c2c0c5bffda70cda/Django-1.9.11-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "75a455081011adc16aa5c90b19f205e612458afa64e20992fd37ab8b159c11fb", - "md5": "ea8c3c7ecd02e377a01064a94f3def91", - "sha256": "dadcfd8b03dfbf472c2d88c12202d9d015af68fb6561099992bc2d91aeab7d9d" - }, - "downloads": -1, - "filename": "Django-1.9.11.tar.gz", - "has_sig": false, - "md5_digest": "ea8c3c7ecd02e377a01064a94f3def91", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7498558, - "upload_time": "2016-11-01T14:02:27", - "upload_time_iso_8601": "2016-11-01T14:02:27.855507Z", - "url": "https://files.pythonhosted.org/packages/75/a4/55081011adc16aa5c90b19f205e612458afa64e20992fd37ab8b159c11fb/Django-1.9.11.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.9.12": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "30001e440d60cb4764757324a31e4a0248988b7fa0673d31fb2710b51f6b515e", - "md5": "18e3c04b58c0609b34a0a7ddd0309521", - "sha256": "a59f85a2b007145006915f6134ec3b9c09e68e4377e0a6fd3529d6c56d6aeb04" - }, - "downloads": -1, - "filename": "Django-1.9.12-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "18e3c04b58c0609b34a0a7ddd0309521", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6622479, - "upload_time": "2016-12-01T23:16:22", - "upload_time_iso_8601": "2016-12-01T23:16:22.406064Z", - "url": "https://files.pythonhosted.org/packages/30/00/1e440d60cb4764757324a31e4a0248988b7fa0673d31fb2710b51f6b515e/Django-1.9.12-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "9f0ca11439e91e3ca43b9ed1d406558f2d7280cb3f105cc1e6cc9f1bb76a98f5", - "md5": "bb9d13c690a5baf2270822dc57c4633c", - "sha256": "bd20830d403689efa46ea2a7cd3ef689af17de633c095aeabf9ce770b3f84a35" - }, - "downloads": -1, - "filename": "Django-1.9.12.tar.gz", - "has_sig": false, - "md5_digest": "bb9d13c690a5baf2270822dc57c4633c", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7497785, - "upload_time": "2016-12-01T23:16:35", - "upload_time_iso_8601": "2016-12-01T23:16:35.055179Z", - "url": "https://files.pythonhosted.org/packages/9f/0c/a11439e91e3ca43b9ed1d406558f2d7280cb3f105cc1e6cc9f1bb76a98f5/Django-1.9.12.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.9.13": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "a6ee6dfd9e8e0d61ab22ed09c6c8f6ae5a4ad0d3bf07bca9124fb5a087f876df", - "md5": "72e89cb92638f4000b6213fc3a6bd3d0", - "sha256": "c213590aa8599b17a9a914ab017f7dc6fc16c9c69603ecf071100346b8d9d777" - }, - "downloads": -1, - "filename": "Django-1.9.13-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "72e89cb92638f4000b6213fc3a6bd3d0", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6623016, - "upload_time": "2017-04-04T14:14:50", - "upload_time_iso_8601": "2017-04-04T14:14:50.895823Z", - "url": "https://files.pythonhosted.org/packages/a6/ee/6dfd9e8e0d61ab22ed09c6c8f6ae5a4ad0d3bf07bca9124fb5a087f876df/Django-1.9.13-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "6cbf4ac9eaa04581797a68649828b88a147f641d745e0c7b3696db5c6caedc41", - "md5": "75e461515dcfe6e8aaea3fb569e8d8fe", - "sha256": "c007dba5086061f7d0f4d88a3bc4016d881a7eede86d6c1c4fdbbaadddd53f1d" - }, - "downloads": -1, - "filename": "Django-1.9.13.tar.gz", - "has_sig": false, - "md5_digest": "75e461515dcfe6e8aaea3fb569e8d8fe", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7498364, - "upload_time": "2017-04-04T14:15:10", - "upload_time_iso_8601": "2017-04-04T14:15:10.543830Z", - "url": "https://files.pythonhosted.org/packages/6c/bf/4ac9eaa04581797a68649828b88a147f641d745e0c7b3696db5c6caedc41/Django-1.9.13.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.9.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "2649c51f94a3940b3dee47f5a4a423fe50df7b9120bccda424f092269c083898", - "md5": "72317fd693fe1c95b6192d25d8fcd323", - "sha256": "cc2ee91769af012654ae4904b6704f2fa0cc6b283675869c2f2ed879eaba11e8" - }, - "downloads": -1, - "filename": "Django-1.9.2-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "72317fd693fe1c95b6192d25d8fcd323", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6580359, - "upload_time": "2016-02-01T17:17:18", - "upload_time_iso_8601": "2016-02-01T17:17:18.693011Z", - "url": "https://files.pythonhosted.org/packages/26/49/c51f94a3940b3dee47f5a4a423fe50df7b9120bccda424f092269c083898/Django-1.9.2-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "7fa64ab15169e9075635c0a64bb0a569d1e02c7bd9b5ba839f05fe5d1fbbe3fb", - "md5": "ee90280973d435a1a6aa01b453b50cd1", - "sha256": "7a233322eeb35da5fd8315f9e5dd48f2171de43ca2cfb11b138607daa4bf8a2f" - }, - "downloads": -1, - "filename": "Django-1.9.2.tar.gz", - "has_sig": false, - "md5_digest": "ee90280973d435a1a6aa01b453b50cd1", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7419200, - "upload_time": "2016-02-01T17:17:34", - "upload_time_iso_8601": "2016-02-01T17:17:34.977864Z", - "url": "https://files.pythonhosted.org/packages/7f/a6/4ab15169e9075635c0a64bb0a569d1e02c7bd9b5ba839f05fe5d1fbbe3fb/Django-1.9.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.9.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "f9222ecdc946394c2490212691bd44fc8b54aa6b1c49e8cb505d853dc0716646", - "md5": "dff96b62225f29aa0203ae2eff2d36e2", - "sha256": "0d4754d674be02b6e520633b4068d7917abedeabf7481994d4fb9870a5c8781c" - }, - "downloads": -1, - "filename": "Django-1.9.3-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "dff96b62225f29aa0203ae2eff2d36e2", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6581198, - "upload_time": "2016-03-01T17:00:44", - "upload_time_iso_8601": "2016-03-01T17:00:44.803585Z", - "url": "https://files.pythonhosted.org/packages/f9/22/2ecdc946394c2490212691bd44fc8b54aa6b1c49e8cb505d853dc0716646/Django-1.9.3-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "316d853c0006643daedfd4d1bd15ad3afcf3311ee3c0c55bc19572ae1cec52d5", - "md5": "4751d0ebf323083e74b38b02bbbe5b29", - "sha256": "05191a2487de2726d3a964176cb0ffd7cbc071ad117fe06263b7932b96243b56" - }, - "downloads": -1, - "filename": "Django-1.9.3.tar.gz", - "has_sig": false, - "md5_digest": "4751d0ebf323083e74b38b02bbbe5b29", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7425368, - "upload_time": "2016-03-01T17:01:04", - "upload_time_iso_8601": "2016-03-01T17:01:04.656339Z", - "url": "https://files.pythonhosted.org/packages/31/6d/853c0006643daedfd4d1bd15ad3afcf3311ee3c0c55bc19572ae1cec52d5/Django-1.9.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.9.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "361387a668592e133d402ac6f9d07da7fea82e8b217e350a70c7ff53408cd93f", - "md5": "89481f08178f7d28a943fea1bd41de44", - "sha256": "af6264550f8d1cc468db6bbd38151e539b0468ecc5d7d39598af918eae2428b2" - }, - "downloads": -1, - "filename": "Django-1.9.4-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "89481f08178f7d28a943fea1bd41de44", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6581228, - "upload_time": "2016-03-05T14:31:47", - "upload_time_iso_8601": "2016-03-05T14:31:47.633376Z", - "url": "https://files.pythonhosted.org/packages/36/13/87a668592e133d402ac6f9d07da7fea82e8b217e350a70c7ff53408cd93f/Django-1.9.4-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "b09bfbe58870d98321b08043cae62b6062552c5dc7515a2b43958c7330090ef3", - "md5": "e8d389532e248174a9859f2987be6a04", - "sha256": "ada8e7aa697e47c94b5660291cc0a14bb555385e0898da0a119d8f4b648fbde9" - }, - "downloads": -1, - "filename": "Django-1.9.4.tar.gz", - "has_sig": false, - "md5_digest": "e8d389532e248174a9859f2987be6a04", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7426995, - "upload_time": "2016-03-05T14:32:11", - "upload_time_iso_8601": "2016-03-05T14:32:11.209251Z", - "url": "https://files.pythonhosted.org/packages/b0/9b/fbe58870d98321b08043cae62b6062552c5dc7515a2b43958c7330090ef3/Django-1.9.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.9.5": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "6e10208551440e06dee3c6d2cdb36d73b586bfb0ae50a6f91127f1e1492dfb99", - "md5": "849713c2687bbd5739e3cbf738af1a71", - "sha256": "e4aa221fc52f32074f295fc81486e54d35fe176284d6a19c8bf9acaeddd058b4" - }, - "downloads": -1, - "filename": "Django-1.9.5-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "849713c2687bbd5739e3cbf738af1a71", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6582330, - "upload_time": "2016-04-01T17:47:06", - "upload_time_iso_8601": "2016-04-01T17:47:06.038168Z", - "url": "https://files.pythonhosted.org/packages/6e/10/208551440e06dee3c6d2cdb36d73b586bfb0ae50a6f91127f1e1492dfb99/Django-1.9.5-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "3b0c8dc093164d96dcbeb70a4d83e69655b3bec22599828e530f921f53aba08b", - "md5": "419835cef8d42a1a0a3fd6e1eaa24475", - "sha256": "e54667ad305a29f5895f14108127cc79dabebce2e80c7c6cf852a6495de26aa6" - }, - "downloads": -1, - "filename": "Django-1.9.5.tar.gz", - "has_sig": false, - "md5_digest": "419835cef8d42a1a0a3fd6e1eaa24475", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7430219, - "upload_time": "2016-04-01T17:47:24", - "upload_time_iso_8601": "2016-04-01T17:47:24.975699Z", - "url": "https://files.pythonhosted.org/packages/3b/0c/8dc093164d96dcbeb70a4d83e69655b3bec22599828e530f921f53aba08b/Django-1.9.5.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.9.6": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "cb97081df31f2a3850988b92ad4464e95f9e4b257aa5a34e120bca89c260de96", - "md5": "70687b2d4a2dfa51f70407a3ce2abcbb", - "sha256": "83f234f52a86eb983de3616b8ff3b0368e4da1d975fa5449b08588b749e7946c" - }, - "downloads": -1, - "filename": "Django-1.9.6-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "70687b2d4a2dfa51f70407a3ce2abcbb", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6582592, - "upload_time": "2016-05-02T22:33:30", - "upload_time_iso_8601": "2016-05-02T22:33:30.439399Z", - "url": "https://files.pythonhosted.org/packages/cb/97/081df31f2a3850988b92ad4464e95f9e4b257aa5a34e120bca89c260de96/Django-1.9.6-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "3f92701cc1c7d209c830375b884a657ec0c9777f1bed1b5b241f4b70f95e2d55", - "md5": "07d9b3883198caa847b917c94b665452", - "sha256": "e5a1d1a831c06475c708126c6a7e0ae0e67adac9d8a7c39cf0695ad79030b9d9" - }, - "downloads": -1, - "filename": "Django-1.9.6.tar.gz", - "has_sig": false, - "md5_digest": "07d9b3883198caa847b917c94b665452", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7441794, - "upload_time": "2016-05-02T22:33:50", - "upload_time_iso_8601": "2016-05-02T22:33:50.459797Z", - "url": "https://files.pythonhosted.org/packages/3f/92/701cc1c7d209c830375b884a657ec0c9777f1bed1b5b241f4b70f95e2d55/Django-1.9.6.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.9.7": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "e6f9154e1460c4a95c90ab28ead50314161ea2c4016f3561033b41f687f0a76d", - "md5": "5224b6f237a9e46a84fc0f9921f678ae", - "sha256": "74676acbe69964f3aabfe5eb8697a2c82e6095d3fab57589eaab962a98575f58" - }, - "downloads": -1, - "filename": "Django-1.9.7-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "5224b6f237a9e46a84fc0f9921f678ae", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6582730, - "upload_time": "2016-06-04T23:43:47", - "upload_time_iso_8601": "2016-06-04T23:43:47.386292Z", - "url": "https://files.pythonhosted.org/packages/e6/f9/154e1460c4a95c90ab28ead50314161ea2c4016f3561033b41f687f0a76d/Django-1.9.7-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "5076aeb1bdde528b23e76df5964003e3e4e734c57c74e7358c3b2224987617dd", - "md5": "7de9ba83bfe01f4b7d45645c1b259c83", - "sha256": "2b29e81c8c32b3c0d9a0119217416887c480d927ae2630bada2da83078c93bf6" - }, - "downloads": -1, - "filename": "Django-1.9.7.tar.gz", - "has_sig": false, - "md5_digest": "7de9ba83bfe01f4b7d45645c1b259c83", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7442680, - "upload_time": "2016-06-04T23:44:08", - "upload_time_iso_8601": "2016-06-04T23:44:08.110124Z", - "url": "https://files.pythonhosted.org/packages/50/76/aeb1bdde528b23e76df5964003e3e4e734c57c74e7358c3b2224987617dd/Django-1.9.7.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.9.8": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "f5b8f0fbab57bca8d88ac4d41f816a80d24f7dcfc66616984c255b1974df702f", - "md5": "768fe4c89d40f412dfcb151306c5bae5", - "sha256": "e50c8d4653433413d26ecf25a84ccdcb9c7eac7cd1bb36eed4cd89c71f43064e" - }, - "downloads": -1, - "filename": "Django-1.9.8-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "768fe4c89d40f412dfcb151306c5bae5", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6622120, - "upload_time": "2016-07-18T18:19:55", - "upload_time_iso_8601": "2016-07-18T18:19:55.636304Z", - "url": "https://files.pythonhosted.org/packages/f5/b8/f0fbab57bca8d88ac4d41f816a80d24f7dcfc66616984c255b1974df702f/Django-1.9.8-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "cc36cc342dd7a9921b0da8b0e2e7ec25b7bdec66701196ba20cca36f79906d34", - "md5": "486d18b73d38313058ec8f81b765421c", - "sha256": "b9806e0d598fd04b29b8ef35aea8c9308b3803c3ce8adab4d342db9cdfd42dfb" - }, - "downloads": -1, - "filename": "Django-1.9.8.tar.gz", - "has_sig": false, - "md5_digest": "486d18b73d38313058ec8f81b765421c", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7494012, - "upload_time": "2016-07-18T18:20:13", - "upload_time_iso_8601": "2016-07-18T18:20:13.055798Z", - "url": "https://files.pythonhosted.org/packages/cc/36/cc342dd7a9921b0da8b0e2e7ec25b7bdec66701196ba20cca36f79906d34/Django-1.9.8.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.9.9": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "46e4c21618dd809bb31c7a78cb93b66294c38275baed7c9c84a55692ef16274a", - "md5": "22a7b6ebd216b3bfe2f7c25525ecd444", - "sha256": "3236d2292d6ef6afbda0226c255a7901d271ae86d14768bb20e1b6864c66f7e6" - }, - "downloads": -1, - "filename": "Django-1.9.9-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "22a7b6ebd216b3bfe2f7c25525ecd444", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6622115, - "upload_time": "2016-08-01T18:10:04", - "upload_time_iso_8601": "2016-08-01T18:10:04.220963Z", - "url": "https://files.pythonhosted.org/packages/46/e4/c21618dd809bb31c7a78cb93b66294c38275baed7c9c84a55692ef16274a/Django-1.9.9-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "21c03c13bd233cb65932ec8f7d25ebc2942c40fa5c424f33314ac35839d50631", - "md5": "af2e5f02b72400ddaef7f923a7fda6a9", - "sha256": "e340fb9d534aeb543280e46c3b85e6e1049029a4e6ba5571375b11c914bfde8c" - }, - "downloads": -1, - "filename": "Django-1.9.9.tar.gz", - "has_sig": false, - "md5_digest": "af2e5f02b72400ddaef7f923a7fda6a9", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7493187, - "upload_time": "2016-08-01T18:10:24", - "upload_time_iso_8601": "2016-08-01T18:10:24.753259Z", - "url": "https://files.pythonhosted.org/packages/21/c0/3c13bd233cb65932ec8f7d25ebc2942c40fa5c424f33314ac35839d50631/Django-1.9.9.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.9a1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "cfbcb2f7aad1b5fdde56efb6a9d19598ad01124a2690d6054a6b9900496a6e11", - "md5": "631e2bc72c996cba2c18594fc4fe2c48", - "sha256": "267c4a419b04bb4b7bea2e15f360120e41684c0812628d06c434a496d7bf2c70" - }, - "downloads": -1, - "filename": "Django-1.9a1-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "631e2bc72c996cba2c18594fc4fe2c48", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6410150, - "upload_time": "2015-09-24T00:20:01", - "upload_time_iso_8601": "2015-09-24T00:20:01.735219Z", - "url": "https://files.pythonhosted.org/packages/cf/bc/b2f7aad1b5fdde56efb6a9d19598ad01124a2690d6054a6b9900496a6e11/Django-1.9a1-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - } - ], - "1.9b1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "67360d8dcae06869a84e07106332a5b0ebd6c7b759d4b66ffe75585581d48866", - "md5": "d3109c26215da8ec7ece0b37c06692c5", - "sha256": "aaccd01ddc78c87b8fe6fab4c2f53a85df34fb16f7016678755f43d674483bc0" - }, - "downloads": -1, - "filename": "Django-1.9b1-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "d3109c26215da8ec7ece0b37c06692c5", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6410460, - "upload_time": "2015-10-20T01:17:13", - "upload_time_iso_8601": "2015-10-20T01:17:13.619306Z", - "url": "https://files.pythonhosted.org/packages/67/36/0d8dcae06869a84e07106332a5b0ebd6c7b759d4b66ffe75585581d48866/Django-1.9b1-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - } - ], - "1.9rc1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "61df714f75a6e8172821ee5c541a6fcff7e0f5abfe1a357821c11a859e9f86e0", - "md5": "b971686521ea09b4bf82aec3e794fcbc", - "sha256": "6f0a13411c13fb200c18f636279075682bc520734ff52beeb2d9a2c98381dede" - }, - "downloads": -1, - "filename": "Django-1.9rc1.tar.gz", - "has_sig": false, - "md5_digest": "b971686521ea09b4bf82aec3e794fcbc", - "packagetype": "sdist", - "python_version": "source", - "requires_python": null, - "size": 7282447, - "upload_time": "2015-11-16T21:10:10", - "upload_time_iso_8601": "2015-11-16T21:10:10.276918Z", - "url": "https://files.pythonhosted.org/packages/61/df/714f75a6e8172821ee5c541a6fcff7e0f5abfe1a357821c11a859e9f86e0/Django-1.9rc1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "1.9rc2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "b8afe0378ae58310dae25155f136c1addad211860f156aa82ed8bc5728cb8e1e", - "md5": "9bf8fba95db2ad30018f50cb42304d54", - "sha256": "57001f6b74aafd0fcd347c41a17e3c83d6434edc9cddbc449fb47e481863aed3" - }, - "downloads": -1, - "filename": "Django-1.9rc2-py2.py3-none-any.whl", - "has_sig": false, - "md5_digest": "9bf8fba95db2ad30018f50cb42304d54", - "packagetype": "bdist_wheel", - "python_version": "py2.py3", - "requires_python": null, - "size": 6411011, - "upload_time": "2015-11-24T17:35:35", - "upload_time_iso_8601": "2015-11-24T17:35:35.276135Z", - "url": "https://files.pythonhosted.org/packages/b8/af/e0378ae58310dae25155f136c1addad211860f156aa82ed8bc5728cb8e1e/Django-1.9rc2-py2.py3-none-any.whl", - "yanked": false, - "yanked_reason": null - } - ], - "2.0": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "449835b935a98a17e9a188efc2d53fc51ae0c8bf498a77bc224f9321ae5d111c", - "md5": "da2fdc3901e8751aa7835f49fb6246b2", - "sha256": "af18618ce3291be5092893d8522fe3919661bf3a1fb60e3858ae74865a4f07c2" - }, - "downloads": -1, - "filename": "Django-2.0-py3-none-any.whl", - "has_sig": false, - "md5_digest": "da2fdc3901e8751aa7835f49fb6246b2", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.4", - "size": 7101093, - "upload_time": "2017-12-02T15:11:49", - "upload_time_iso_8601": "2017-12-02T15:11:49.751284Z", - "url": "https://files.pythonhosted.org/packages/44/98/35b935a98a17e9a188efc2d53fc51ae0c8bf498a77bc224f9321ae5d111c/Django-2.0-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "879f4ec8b197d83666fddd2398842024c5341ee7d40bbec6aee9705d1ad22f13", - "md5": "d1afff8f277842a915852b2113671938", - "sha256": "9614851d4a7ff8cbd32b73c6076441f377c45a5bbff7e771798fb02c43c31f47" - }, - "downloads": -1, - "filename": "Django-2.0.tar.gz", - "has_sig": false, - "md5_digest": "d1afff8f277842a915852b2113671938", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.4", - "size": 7997472, - "upload_time": "2017-12-02T15:12:05", - "upload_time_iso_8601": "2017-12-02T15:12:05.947934Z", - "url": "https://files.pythonhosted.org/packages/87/9f/4ec8b197d83666fddd2398842024c5341ee7d40bbec6aee9705d1ad22f13/Django-2.0.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.0.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "212a3a0ec97b18d6e8d295142228f03604ac78ea6de05cf9bc3773a74f0b58bb", - "md5": "e61da2fb97bed3cb9dda3f228832f82b", - "sha256": "52475f607c92035d4ac8fee284f56213065a4a6b25ed43f7e39df0e576e69e9f" - }, - "downloads": -1, - "filename": "Django-2.0.1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "e61da2fb97bed3cb9dda3f228832f82b", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.4", - "size": 7101551, - "upload_time": "2018-01-02T00:50:49", - "upload_time_iso_8601": "2018-01-02T00:50:49.281283Z", - "url": "https://files.pythonhosted.org/packages/21/2a/3a0ec97b18d6e8d295142228f03604ac78ea6de05cf9bc3773a74f0b58bb/Django-2.0.1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "546cb751b614e5a5693169e456d109ff26c5d090341d4d2e1b860c7a7c0f3788", - "md5": "29be9ee303a5a592b4ac1ad9316c620a", - "sha256": "d96b804be412a5125a594023ec524a2010a6ffa4d408e5482ab6ff3cb97ec12f" - }, - "downloads": -1, - "filename": "Django-2.0.1.tar.gz", - "has_sig": false, - "md5_digest": "29be9ee303a5a592b4ac1ad9316c620a", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.4", - "size": 8000807, - "upload_time": "2018-01-02T00:51:04", - "upload_time_iso_8601": "2018-01-02T00:51:04.317791Z", - "url": "https://files.pythonhosted.org/packages/54/6c/b751b614e5a5693169e456d109ff26c5d090341d4d2e1b860c7a7c0f3788/Django-2.0.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.0.10": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "f192d58198bdcb00e8ed79fd3978f4af61a50009bb1d0ee8bbb799fa3476e5ff", - "md5": "65fec2f96aabb791f19a6c7fd70c0d0b", - "sha256": "e89f613e3c1f7ff245ffee3560472f9fa9c07060b11f65e1de3cb763f8dcd4b9" - }, - "downloads": -1, - "filename": "Django-2.0.10-py3-none-any.whl", - "has_sig": false, - "md5_digest": "65fec2f96aabb791f19a6c7fd70c0d0b", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.4", - "size": 7117592, - "upload_time": "2019-01-04T14:03:05", - "upload_time_iso_8601": "2019-01-04T14:03:05.360117Z", - "url": "https://files.pythonhosted.org/packages/f1/92/d58198bdcb00e8ed79fd3978f4af61a50009bb1d0ee8bbb799fa3476e5ff/Django-2.0.10-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "021f3766ba3a7f2356ba9816f8e4eb82168fd0233f1b63f1ce52cd57b15acd6a", - "md5": "1d90088c3f66e1ccbb03c1ad1aabc63a", - "sha256": "0292a7ad7d8ffc9cfc6a77f043d2e81f5bbc360c0c4a1686e130ef3432437d23" - }, - "downloads": -1, - "filename": "Django-2.0.10.tar.gz", - "has_sig": false, - "md5_digest": "1d90088c3f66e1ccbb03c1ad1aabc63a", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.4", - "size": 7992290, - "upload_time": "2019-01-04T14:03:19", - "upload_time_iso_8601": "2019-01-04T14:03:19.511618Z", - "url": "https://files.pythonhosted.org/packages/02/1f/3766ba3a7f2356ba9816f8e4eb82168fd0233f1b63f1ce52cd57b15acd6a/Django-2.0.10.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.0.12": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "08d5eda89c77bffafef4fc3eafb1a334ffffadebece16845aa7afff19a2c29e8", - "md5": "0218bf85d73b7795025656cd418fac70", - "sha256": "9493f9c4417ab6c15221a7ab964f73db0632bc6337ade33810078423b800dbd9" - }, - "downloads": -1, - "filename": "Django-2.0.12-py3-none-any.whl", - "has_sig": false, - "md5_digest": "0218bf85d73b7795025656cd418fac70", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.4", - "size": 7117804, - "upload_time": "2019-02-11T15:10:42", - "upload_time_iso_8601": "2019-02-11T15:10:42.905456Z", - "url": "https://files.pythonhosted.org/packages/08/d5/eda89c77bffafef4fc3eafb1a334ffffadebece16845aa7afff19a2c29e8/Django-2.0.12-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "e020e5f922be212959b08db3f8b42b8504b13ab915e17770fb4e42d835e8e6bb", - "md5": "0bca61f6ef40a7f62aa436cdfa033627", - "sha256": "e2f3ada50cf2106af7b2631b7b0b6f0d983ae30aa218f06d5ec93d3049f94897" - }, - "downloads": -1, - "filename": "Django-2.0.12.tar.gz", - "has_sig": false, - "md5_digest": "0bca61f6ef40a7f62aa436cdfa033627", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.4", - "size": 7989632, - "upload_time": "2019-02-11T15:11:02", - "upload_time_iso_8601": "2019-02-11T15:11:02.427442Z", - "url": "https://files.pythonhosted.org/packages/e0/20/e5f922be212959b08db3f8b42b8504b13ab915e17770fb4e42d835e8e6bb/Django-2.0.12.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.0.13": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "67b064645bd6c5cdabb07d361e568eecfa9e64027ae4cb4778bb00be8c4bde00", - "md5": "7bb85dc71318f74cf2adaeeb245d664f", - "sha256": "665457d4146bbd34ae9d2970fa3b37082d7b225b0671bfd24c337458f229db78" - }, - "downloads": -1, - "filename": "Django-2.0.13-py3-none-any.whl", - "has_sig": false, - "md5_digest": "7bb85dc71318f74cf2adaeeb245d664f", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.4", - "size": 7117798, - "upload_time": "2019-02-12T10:50:02", - "upload_time_iso_8601": "2019-02-12T10:50:02.942164Z", - "url": "https://files.pythonhosted.org/packages/67/b0/64645bd6c5cdabb07d361e568eecfa9e64027ae4cb4778bb00be8c4bde00/Django-2.0.13-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "57001717ea3aabb17b2c4981561c7d00ac2d0eb5706647bcfd4a0ca2d09e2413", - "md5": "38b8958369c4396408d248a26ad2b057", - "sha256": "bde46d4dbc410678e89bc95ea5d312dd6eb4c37d0fa0e19c9415cad94addf22f" - }, - "downloads": -1, - "filename": "Django-2.0.13.tar.gz", - "has_sig": false, - "md5_digest": "38b8958369c4396408d248a26ad2b057", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.4", - "size": 7990057, - "upload_time": "2019-02-12T10:50:09", - "upload_time_iso_8601": "2019-02-12T10:50:09.191763Z", - "url": "https://files.pythonhosted.org/packages/57/00/1717ea3aabb17b2c4981561c7d00ac2d0eb5706647bcfd4a0ca2d09e2413/Django-2.0.13.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.0.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "48b14e4875dabd9603c89ca753313fd13f429426e4a80a0264603a30d56e0338", - "md5": "36d3f5a7f230296c040b11344f48ffce", - "sha256": "7c8ff92285406fb349e765e9ade685eec7271d6f5c3f918e495a74768b765c99" - }, - "downloads": -1, - "filename": "Django-2.0.2-py3-none-any.whl", - "has_sig": false, - "md5_digest": "36d3f5a7f230296c040b11344f48ffce", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.4", - "size": 7101554, - "upload_time": "2018-02-01T14:30:13", - "upload_time_iso_8601": "2018-02-01T14:30:13.345780Z", - "url": "https://files.pythonhosted.org/packages/48/b1/4e4875dabd9603c89ca753313fd13f429426e4a80a0264603a30d56e0338/Django-2.0.2-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "21eb534ac46e63c51eabbfc768d8c11cc851275f9047c8eaaefc17c41845987f", - "md5": "9d4ae0d4193bad0c6af751e54f3a4690", - "sha256": "dc3b61d054f1bced64628c62025d480f655303aea9f408e5996c339a543b45f0" - }, - "downloads": -1, - "filename": "Django-2.0.2.tar.gz", - "has_sig": false, - "md5_digest": "9d4ae0d4193bad0c6af751e54f3a4690", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.4", - "size": 8002374, - "upload_time": "2018-02-01T14:30:22", - "upload_time_iso_8601": "2018-02-01T14:30:22.407202Z", - "url": "https://files.pythonhosted.org/packages/21/eb/534ac46e63c51eabbfc768d8c11cc851275f9047c8eaaefc17c41845987f/Django-2.0.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.0.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "3d817e6cf5cb6f0f333946b5d3ee22e17c3c3f329d3bfeb86943a2a3cd861092", - "md5": "84742d3266ae1a2c02326b1e3036c7b5", - "sha256": "3d9916515599f757043c690ae2b5ea28666afa09779636351da505396cbb2f19" - }, - "downloads": -1, - "filename": "Django-2.0.3-py3-none-any.whl", - "has_sig": false, - "md5_digest": "84742d3266ae1a2c02326b1e3036c7b5", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.4", - "size": 7116284, - "upload_time": "2018-03-06T14:05:32", - "upload_time_iso_8601": "2018-03-06T14:05:32.353199Z", - "url": "https://files.pythonhosted.org/packages/3d/81/7e6cf5cb6f0f333946b5d3ee22e17c3c3f329d3bfeb86943a2a3cd861092/Django-2.0.3-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "54594987ae4a4a8be8507af1b213e75a449c05939ab1e0f62b5e90ccea2b51c3", - "md5": "ef1a31d36aaaa7cfe0a26af351c7ebbe", - "sha256": "769f212ffd5762f72c764fa648fca3b7f7dd4ec27407198b68e7c4abf4609fd0" - }, - "downloads": -1, - "filename": "Django-2.0.3.tar.gz", - "has_sig": false, - "md5_digest": "ef1a31d36aaaa7cfe0a26af351c7ebbe", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.4", - "size": 8114604, - "upload_time": "2018-03-06T14:06:37", - "upload_time_iso_8601": "2018-03-06T14:06:37.383785Z", - "url": "https://files.pythonhosted.org/packages/54/59/4987ae4a4a8be8507af1b213e75a449c05939ab1e0f62b5e90ccea2b51c3/Django-2.0.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.0.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "89f994c20658f0cdecc2b6607811e2c0bb042408a51f589e5ad0cb0eac3236a1", - "md5": "a646aec55e5182051ca8f6f9f87dc286", - "sha256": "2d8b9eed8815f172a8e898678ae4289a5e9176bc08295676eff4228dd638ea61" - }, - "downloads": -1, - "filename": "Django-2.0.4-py3-none-any.whl", - "has_sig": false, - "md5_digest": "a646aec55e5182051ca8f6f9f87dc286", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.4", - "size": 7115298, - "upload_time": "2018-04-03T02:39:33", - "upload_time_iso_8601": "2018-04-03T02:39:33.419093Z", - "url": "https://files.pythonhosted.org/packages/89/f9/94c20658f0cdecc2b6607811e2c0bb042408a51f589e5ad0cb0eac3236a1/Django-2.0.4-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "d473227f6efdb7dd3b74b9785795dffc0f28c4db9c59e2a05e3547fcf003635b", - "md5": "9d4c555b798406361521dcf282f6638a", - "sha256": "d81a1652963c81488e709729a80b510394050e312f386037f26b54912a3a10d0" - }, - "downloads": -1, - "filename": "Django-2.0.4.tar.gz", - "has_sig": false, - "md5_digest": "9d4c555b798406361521dcf282f6638a", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.4", - "size": 8017145, - "upload_time": "2018-04-03T02:39:41", - "upload_time_iso_8601": "2018-04-03T02:39:41.734614Z", - "url": "https://files.pythonhosted.org/packages/d4/73/227f6efdb7dd3b74b9785795dffc0f28c4db9c59e2a05e3547fcf003635b/Django-2.0.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.0.5": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "23912245462e57798e9251de87c88b2b8f996d10ddcb68206a8a020561ef7bd3", - "md5": "e6edc1c129ef54cc2a73c73d58f78ae4", - "sha256": "26b34f4417aa38d895b6b5307177b51bc3f4d53179d8696a5c19dcb50582523c" - }, - "downloads": -1, - "filename": "Django-2.0.5-py3-none-any.whl", - "has_sig": false, - "md5_digest": "e6edc1c129ef54cc2a73c73d58f78ae4", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.4", - "size": 7115507, - "upload_time": "2018-05-02T01:34:20", - "upload_time_iso_8601": "2018-05-02T01:34:20.212616Z", - "url": "https://files.pythonhosted.org/packages/23/91/2245462e57798e9251de87c88b2b8f996d10ddcb68206a8a020561ef7bd3/Django-2.0.5-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "0279e0cf8263d85f15987f34710e325438f8ac6c93961714916ac7ea343e6a08", - "md5": "75bf9f01fef8ae9601f9f30a3f5a44b7", - "sha256": "71d1a584bb4ad2b4f933d07d02c716755c1394feaac1ce61ce37843ac5401092" - }, - "downloads": -1, - "filename": "Django-2.0.5.tar.gz", - "has_sig": false, - "md5_digest": "75bf9f01fef8ae9601f9f30a3f5a44b7", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.4", - "size": 8016448, - "upload_time": "2018-05-02T01:34:52", - "upload_time_iso_8601": "2018-05-02T01:34:52.880781Z", - "url": "https://files.pythonhosted.org/packages/02/79/e0cf8263d85f15987f34710e325438f8ac6c93961714916ac7ea343e6a08/Django-2.0.5.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.0.6": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "560eafdacb47503b805f3ed213fe732bff05254c8befaa034bbada580be8a0ac", - "md5": "51da719305494d838ab65585065ccf7a", - "sha256": "69ff89fa3c3a8337015478a1a0744f52a9fef5d12c1efa01a01f99bcce9bf10c" - }, - "downloads": -1, - "filename": "Django-2.0.6-py3-none-any.whl", - "has_sig": false, - "md5_digest": "51da719305494d838ab65585065ccf7a", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.4", - "size": 7118853, - "upload_time": "2018-06-01T15:32:04", - "upload_time_iso_8601": "2018-06-01T15:32:04.039149Z", - "url": "https://files.pythonhosted.org/packages/56/0e/afdacb47503b805f3ed213fe732bff05254c8befaa034bbada580be8a0ac/Django-2.0.6-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "082ba6a12fa67a9d52a8f5ef929337165dcc5842a4461a87fec94eb6345fed57", - "md5": "b33f1c1f86867787aade905fbfb9e7c8", - "sha256": "3eb25c99df1523446ec2dc1b00e25eb2ecbdf42c9d8b0b8b32a204a8db9011f8" - }, - "downloads": -1, - "filename": "Django-2.0.6.tar.gz", - "has_sig": false, - "md5_digest": "b33f1c1f86867787aade905fbfb9e7c8", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.4", - "size": 7989435, - "upload_time": "2018-06-01T15:32:11", - "upload_time_iso_8601": "2018-06-01T15:32:11.407175Z", - "url": "https://files.pythonhosted.org/packages/08/2b/a6a12fa67a9d52a8f5ef929337165dcc5842a4461a87fec94eb6345fed57/Django-2.0.6.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.0.7": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "ab15cfde97943f0db45e4f999c60b696fbb4df59e82bbccc686770f4e44c9094", - "md5": "edc4fefc12f35893a52ea6548224c466", - "sha256": "e900b73beee8977c7b887d90c6c57d68af10066b9dac898e1eaf0f82313de334" - }, - "downloads": -1, - "filename": "Django-2.0.7-py3-none-any.whl", - "has_sig": false, - "md5_digest": "edc4fefc12f35893a52ea6548224c466", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.4", - "size": 7119179, - "upload_time": "2018-07-02T09:02:02", - "upload_time_iso_8601": "2018-07-02T09:02:02.793300Z", - "url": "https://files.pythonhosted.org/packages/ab/15/cfde97943f0db45e4f999c60b696fbb4df59e82bbccc686770f4e44c9094/Django-2.0.7-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "584a26f99e2b094a2edefb2cffbcdbaca9207835d4b2765dd8afa553a1714ea6", - "md5": "a5bc9a999972f821a73a5d00aa864e18", - "sha256": "97886b8a13bbc33bfeba2ff133035d3eca014e2309dff2b6da0bdfc0b8656613" - }, - "downloads": -1, - "filename": "Django-2.0.7.tar.gz", - "has_sig": false, - "md5_digest": "a5bc9a999972f821a73a5d00aa864e18", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.4", - "size": 7988568, - "upload_time": "2018-07-02T09:02:19", - "upload_time_iso_8601": "2018-07-02T09:02:19.534930Z", - "url": "https://files.pythonhosted.org/packages/58/4a/26f99e2b094a2edefb2cffbcdbaca9207835d4b2765dd8afa553a1714ea6/Django-2.0.7.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.0.8": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "2b85337bfa37c4b82f59ee9b8deca55a8daa7ef16c8cdfa86d273625bc6ed887", - "md5": "af2b9d548c64f7150f89e8a6b612cdf2", - "sha256": "0c5b65847d00845ee404bbc0b4a85686f15eb3001ffddda3db4e9baa265bf136" - }, - "downloads": -1, - "filename": "Django-2.0.8-py3-none-any.whl", - "has_sig": false, - "md5_digest": "af2b9d548c64f7150f89e8a6b612cdf2", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.4", - "size": 7117321, - "upload_time": "2018-08-01T13:51:58", - "upload_time_iso_8601": "2018-08-01T13:51:58.926049Z", - "url": "https://files.pythonhosted.org/packages/2b/85/337bfa37c4b82f59ee9b8deca55a8daa7ef16c8cdfa86d273625bc6ed887/Django-2.0.8-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "19be23674a0140e79345ba72da70bc646a4db982f2ff39f14037d5cd53807c78", - "md5": "5dd7b6d2c462bec94cff68d9de0b0e3e", - "sha256": "68aeea369a8130259354b6ba1fa9babe0c5ee6bced505dea4afcd00f765ae38b" - }, - "downloads": -1, - "filename": "Django-2.0.8.tar.gz", - "has_sig": false, - "md5_digest": "5dd7b6d2c462bec94cff68d9de0b0e3e", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.4", - "size": 7987343, - "upload_time": "2018-08-01T13:52:08", - "upload_time_iso_8601": "2018-08-01T13:52:08.556325Z", - "url": "https://files.pythonhosted.org/packages/19/be/23674a0140e79345ba72da70bc646a4db982f2ff39f14037d5cd53807c78/Django-2.0.8.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.0.9": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "6c9dc0feec696b815708354a2fd06ae0f51330a15043822a29bc8be2f185d9fe", - "md5": "d1c46a6a58e3002eb04ef10407ebe6c6", - "sha256": "25df265e1fdb74f7e7305a1de620a84681bcc9c05e84a3ed97e4a1a63024f18d" - }, - "downloads": -1, - "filename": "Django-2.0.9-py3-none-any.whl", - "has_sig": false, - "md5_digest": "d1c46a6a58e3002eb04ef10407ebe6c6", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.4", - "size": 7119483, - "upload_time": "2018-10-01T09:22:07", - "upload_time_iso_8601": "2018-10-01T09:22:07.871505Z", - "url": "https://files.pythonhosted.org/packages/6c/9d/c0feec696b815708354a2fd06ae0f51330a15043822a29bc8be2f185d9fe/Django-2.0.9-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "fb9fcd3fe00fdaecc34a5ca1d667b8436de604543f77c9e16bc1396a40b720bb", - "md5": "577f668bc9031578583eef14c25da0b4", - "sha256": "d6d94554abc82ca37e447c3d28958f5ac39bd7d4adaa285543ae97fb1129fd69" - }, - "downloads": -1, - "filename": "Django-2.0.9.tar.gz", - "has_sig": false, - "md5_digest": "577f668bc9031578583eef14c25da0b4", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.4", - "size": 7992507, - "upload_time": "2018-10-01T09:22:23", - "upload_time_iso_8601": "2018-10-01T09:22:23.082215Z", - "url": "https://files.pythonhosted.org/packages/fb/9f/cd3fe00fdaecc34a5ca1d667b8436de604543f77c9e16bc1396a40b720bb/Django-2.0.9.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.0a1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "9a3d5050d001444d0078d6c5a2453ffb0fc4b871de41d67af8e7f190a276fcf0", - "md5": "4d83cd13a20ae81c2bb81c124616a027", - "sha256": "a8a695977109830cd3f1bef65bce43d78d361751a304c5cf215d876fbdcf6416" - }, - "downloads": -1, - "filename": "Django-2.0a1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "4d83cd13a20ae81c2bb81c124616a027", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.4", - "size": 7002800, - "upload_time": "2017-09-22T18:09:22", - "upload_time_iso_8601": "2017-09-22T18:09:22.828675Z", - "url": "https://files.pythonhosted.org/packages/9a/3d/5050d001444d0078d6c5a2453ffb0fc4b871de41d67af8e7f190a276fcf0/Django-2.0a1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - } - ], - "2.0b1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "6c37e3fccf0f86d016cd6b851a80e07aae20c750dbccba1098d98563fa7e498e", - "md5": "4d834100601f91e885f4a806be8756c6", - "sha256": "5e7f0d33a1908070d81ffea029c7c69f7b8aa9c95fa2415885ea3c970c2cf8e0" - }, - "downloads": -1, - "filename": "Django-2.0b1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "4d834100601f91e885f4a806be8756c6", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.4", - "size": 7003741, - "upload_time": "2017-10-17T02:00:54", - "upload_time_iso_8601": "2017-10-17T02:00:54.980990Z", - "url": "https://files.pythonhosted.org/packages/6c/37/e3fccf0f86d016cd6b851a80e07aae20c750dbccba1098d98563fa7e498e/Django-2.0b1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - } - ], - "2.0rc1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "49656cc2315700549de13daf0f6dc5728bdcbcb19195876a7534e54649b0a132", - "md5": "7e24467bac63025333f76881f8f7d889", - "sha256": "bbf9aedead4417142d78e2135089737d0205fd2d3641c6cec65583109ce676bf" - }, - "downloads": -1, - "filename": "Django-2.0rc1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "7e24467bac63025333f76881f8f7d889", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.4", - "size": 7004428, - "upload_time": "2017-11-15T23:51:54", - "upload_time_iso_8601": "2017-11-15T23:51:54.051974Z", - "url": "https://files.pythonhosted.org/packages/49/65/6cc2315700549de13daf0f6dc5728bdcbcb19195876a7534e54649b0a132/Django-2.0rc1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - } - ], - "2.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "511ae0ac7886c7123a03814178d7517dc822af0fe51a72e1a6bff26153103322", - "md5": "b915941cbfea97580c1c928a5c11bc72", - "sha256": "ea50d85709708621d956187c6b61d9f9ce155007b496dd914fdb35db8d790aec" - }, - "downloads": -1, - "filename": "Django-2.1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "b915941cbfea97580c1c928a5c11bc72", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7263150, - "upload_time": "2018-08-01T14:11:27", - "upload_time_iso_8601": "2018-08-01T14:11:27.382635Z", - "url": "https://files.pythonhosted.org/packages/51/1a/e0ac7886c7123a03814178d7517dc822af0fe51a72e1a6bff26153103322/Django-2.1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "b6cf8cbe9bd4bb83ce2dd277564b43435edb7b151a099458e63706d10ec9e4fa", - "md5": "4a01d9325ac60e8d329762ecb9c9d2ea", - "sha256": "7f246078d5a546f63c28fc03ce71f4d7a23677ce42109219c24c9ffb28416137" - }, - "downloads": -1, - "filename": "Django-2.1.tar.gz", - "has_sig": false, - "md5_digest": "4a01d9325ac60e8d329762ecb9c9d2ea", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8583964, - "upload_time": "2018-08-01T14:11:34", - "upload_time_iso_8601": "2018-08-01T14:11:34.715690Z", - "url": "https://files.pythonhosted.org/packages/b6/cf/8cbe9bd4bb83ce2dd277564b43435edb7b151a099458e63706d10ec9e4fa/Django-2.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.1.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "ca7efc068d164b32552ae3a8f8d5d0280c083f2e8d553e71ecacc21927564561", - "md5": "4ef1290007f8ccb865e27c68dd1aa6cb", - "sha256": "04f2e423f2e60943c02bd2959174b844f7d1bcd19eabb7f8e4282999958021fd" - }, - "downloads": -1, - "filename": "Django-2.1.1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "4ef1290007f8ccb865e27c68dd1aa6cb", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7265628, - "upload_time": "2018-08-31T08:42:13", - "upload_time_iso_8601": "2018-08-31T08:42:13.116397Z", - "url": "https://files.pythonhosted.org/packages/ca/7e/fc068d164b32552ae3a8f8d5d0280c083f2e8d553e71ecacc21927564561/Django-2.1.1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "14c8b6f5c67cf34ae7586258af110e53657da671325b146fcc67ac64a4daace5", - "md5": "06e7c47864e0a38cd99ddabb152b9e9b", - "sha256": "e1cc1cd6b658aa4e052f5f2b148bfda08091d7c3558529708342e37e4e33f72c" - }, - "downloads": -1, - "filename": "Django-2.1.1.tar.gz", - "has_sig": false, - "md5_digest": "06e7c47864e0a38cd99ddabb152b9e9b", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8595422, - "upload_time": "2018-08-31T08:42:20", - "upload_time_iso_8601": "2018-08-31T08:42:20.929702Z", - "url": "https://files.pythonhosted.org/packages/14/c8/b6f5c67cf34ae7586258af110e53657da671325b146fcc67ac64a4daace5/Django-2.1.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.1.10": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "d466dbf6b7fabb0882eac49a33baac096396f98e5ba0cdea7c02e0a3bf02e8fb", - "md5": "210ab4e6a1f7bc82dd19f6514d3f8963", - "sha256": "d4244e2c9653e1349b37ab0cdf108607ba9d1e41f71d0711ac71360e4026ce94" - }, - "downloads": -1, - "filename": "Django-2.1.10-py3-none-any.whl", - "has_sig": false, - "md5_digest": "210ab4e6a1f7bc82dd19f6514d3f8963", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7281394, - "upload_time": "2019-07-01T07:19:18", - "upload_time_iso_8601": "2019-07-01T07:19:18.135804Z", - "url": "https://files.pythonhosted.org/packages/d4/66/dbf6b7fabb0882eac49a33baac096396f98e5ba0cdea7c02e0a3bf02e8fb/Django-2.1.10-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "be1b009ec818adf51c7641f3bd9dae778e8b28291b3ceedb352317b0eeafd7ff", - "md5": "2162aed4111da837433f41a9eed5c8bd", - "sha256": "65e2a548a52fca560cdd4e35f4fa1a79140f405af48950e59702a37e4227e958" - }, - "downloads": -1, - "filename": "Django-2.1.10.tar.gz", - "has_sig": false, - "md5_digest": "2162aed4111da837433f41a9eed5c8bd", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8758324, - "upload_time": "2019-07-01T07:19:36", - "upload_time_iso_8601": "2019-07-01T07:19:36.568033Z", - "url": "https://files.pythonhosted.org/packages/be/1b/009ec818adf51c7641f3bd9dae778e8b28291b3ceedb352317b0eeafd7ff/Django-2.1.10.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.1.11": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "8c49d5038239995594281478bf209f8d93524ad342d500009a697b27f884668a", - "md5": "5419065b487d22f584ca9d42df971092", - "sha256": "305b6c4fce9e03bb746e35780c2c4d52f29ea1669f15633cfd41bc8821c74c76" - }, - "downloads": -1, - "filename": "Django-2.1.11-py3-none-any.whl", - "has_sig": false, - "md5_digest": "5419065b487d22f584ca9d42df971092", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7281462, - "upload_time": "2019-08-01T09:04:30", - "upload_time_iso_8601": "2019-08-01T09:04:30.982323Z", - "url": "https://files.pythonhosted.org/packages/8c/49/d5038239995594281478bf209f8d93524ad342d500009a697b27f884668a/Django-2.1.11-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "e0e97e6008abee3eb2a40704c95a5cfc8a9627012df1580289d3df0f34c99766", - "md5": "42f0d3ccdcd89c566a30765ee0e25d42", - "sha256": "1a41831eace203fd1939edf899e07d7abd12ce9bafc3d9a5a63a24a8d1d12bd5" - }, - "downloads": -1, - "filename": "Django-2.1.11.tar.gz", - "has_sig": false, - "md5_digest": "42f0d3ccdcd89c566a30765ee0e25d42", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8612487, - "upload_time": "2019-08-01T09:04:51", - "upload_time_iso_8601": "2019-08-01T09:04:51.690400Z", - "url": "https://files.pythonhosted.org/packages/e0/e9/7e6008abee3eb2a40704c95a5cfc8a9627012df1580289d3df0f34c99766/Django-2.1.11.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.1.12": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "1ba7e6fcb3a4b4728b532a72f13f3b9999376e5e9487dd5777dae75e8f392f97", - "md5": "31e4237678ef875d872bf698d39c7cf9", - "sha256": "33e4bb3f3268ecc362a7bd982c7f19fff30597c749fedbc1fd3773ebb82ec38c" - }, - "downloads": -1, - "filename": "Django-2.1.12-py3-none-any.whl", - "has_sig": false, - "md5_digest": "31e4237678ef875d872bf698d39c7cf9", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7281462, - "upload_time": "2019-09-02T07:18:34", - "upload_time_iso_8601": "2019-09-02T07:18:34.162779Z", - "url": "https://files.pythonhosted.org/packages/1b/a7/e6fcb3a4b4728b532a72f13f3b9999376e5e9487dd5777dae75e8f392f97/Django-2.1.12-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "a3e4a7399d5c9044fc2cf1c27884865ee0e9ee3e02a775628f9f29f43f657baa", - "md5": "b975a7880067279938b0115d922455d4", - "sha256": "f4351f1f921bb6c3de03e24cdba823365fb9a79a44f607ba2560e9e3b7f16ff3" - }, - "downloads": -1, - "filename": "Django-2.1.12.tar.gz", - "has_sig": false, - "md5_digest": "b975a7880067279938b0115d922455d4", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8761663, - "upload_time": "2019-09-02T07:18:54", - "upload_time_iso_8601": "2019-09-02T07:18:54.146400Z", - "url": "https://files.pythonhosted.org/packages/a3/e4/a7399d5c9044fc2cf1c27884865ee0e9ee3e02a775628f9f29f43f657baa/Django-2.1.12.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.1.13": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "c4e19c31da650533111ec044c2e540be72287d87c7b0ddb076640f943d900f98", - "md5": "93cbf3969da78db04702e065ba76860e", - "sha256": "79c6cfbc76a4612efc8a80d0094b96f50cb2aa654b2b2b530a4e707fa2985b4b" - }, - "downloads": -1, - "filename": "Django-2.1.13-py3-none-any.whl", - "has_sig": false, - "md5_digest": "93cbf3969da78db04702e065ba76860e", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7281478, - "upload_time": "2019-10-01T08:36:38", - "upload_time_iso_8601": "2019-10-01T08:36:38.282724Z", - "url": "https://files.pythonhosted.org/packages/c4/e1/9c31da650533111ec044c2e540be72287d87c7b0ddb076640f943d900f98/Django-2.1.13-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "cad6940c06d14c5b60e32cce5f47143d4c944e93690dedbaa52b48a2f4f4ece9", - "md5": "ea31f78a203dde4e4df428fad9d2b7b5", - "sha256": "7a28a4eb0167eba491ccfafd7006843b5cdd26d8c93b955a74c2ea74f94efc2c" - }, - "downloads": -1, - "filename": "Django-2.1.13.tar.gz", - "has_sig": false, - "md5_digest": "ea31f78a203dde4e4df428fad9d2b7b5", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8611325, - "upload_time": "2019-10-01T08:36:59", - "upload_time_iso_8601": "2019-10-01T08:36:59.308397Z", - "url": "https://files.pythonhosted.org/packages/ca/d6/940c06d14c5b60e32cce5f47143d4c944e93690dedbaa52b48a2f4f4ece9/Django-2.1.13.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.1.14": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "cd535b415cf77c091e0011b0c2681c27fb4f63ed6d8cc4f087189090042d4b6d", - "md5": "3163790bad4a546ed4cc7792f916d83f", - "sha256": "660e7cc240231b2fe702fada5fe952d29720fd6153fa6b59ea5dbf3161ab5cdd" - }, - "downloads": -1, - "filename": "Django-2.1.14-py3-none-any.whl", - "has_sig": false, - "md5_digest": "3163790bad4a546ed4cc7792f916d83f", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7281484, - "upload_time": "2019-11-04T08:33:13", - "upload_time_iso_8601": "2019-11-04T08:33:13.835019Z", - "url": "https://files.pythonhosted.org/packages/cd/53/5b415cf77c091e0011b0c2681c27fb4f63ed6d8cc4f087189090042d4b6d/Django-2.1.14-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "0e0dcd8d35031bb97db9f6a7172dd8b3dd8a899cac151f7fbf3096b2cc45a51a", - "md5": "b5c9338d7f7a4da25d77bbdb088b01bb", - "sha256": "d9159141fc354c4c28cc2b2586d55ba6d5e1531f5470218bb56a75be03d67398" - }, - "downloads": -1, - "filename": "Django-2.1.14.tar.gz", - "has_sig": false, - "md5_digest": "b5c9338d7f7a4da25d77bbdb088b01bb", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8763129, - "upload_time": "2019-11-04T08:33:33", - "upload_time_iso_8601": "2019-11-04T08:33:33.296744Z", - "url": "https://files.pythonhosted.org/packages/0e/0d/cd8d35031bb97db9f6a7172dd8b3dd8a899cac151f7fbf3096b2cc45a51a/Django-2.1.14.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.1.15": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "ff8255a696532518aa47666b45480b579a221638ab29d60d33ce71fcbd3cef9a", - "md5": "dc0ae37c86505e64a58b5f28d4fe768c", - "sha256": "48522428f4a285cf265af969f4744c5ebb027c7f41958ba48b639ace2068ffe7" - }, - "downloads": -1, - "filename": "Django-2.1.15-py3-none-any.whl", - "has_sig": false, - "md5_digest": "dc0ae37c86505e64a58b5f28d4fe768c", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7281877, - "upload_time": "2019-12-02T08:57:47", - "upload_time_iso_8601": "2019-12-02T08:57:47.175049Z", - "url": "https://files.pythonhosted.org/packages/ff/82/55a696532518aa47666b45480b579a221638ab29d60d33ce71fcbd3cef9a/Django-2.1.15-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "a5eaa3424e68851acb44a1f8f823dc32ee3eb10b7fda474b03d527f7e666b443", - "md5": "a9d02735cb5722608c08fb2d79350523", - "sha256": "a794f7a2f4b7c928eecfbc4ebad03712ff27fb545abe269bf01aa8500781eb1c" - }, - "downloads": -1, - "filename": "Django-2.1.15.tar.gz", - "has_sig": false, - "md5_digest": "a9d02735cb5722608c08fb2d79350523", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8621769, - "upload_time": "2019-12-02T08:57:57", - "upload_time_iso_8601": "2019-12-02T08:57:57.603087Z", - "url": "https://files.pythonhosted.org/packages/a5/ea/a3424e68851acb44a1f8f823dc32ee3eb10b7fda474b03d527f7e666b443/Django-2.1.15.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.1.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "32ab22530cc1b2114e6067eece94a333d6c749fa1c56a009f0721e51c181ea53", - "md5": "1c49272e55d2676455b7d03ec1374170", - "sha256": "acdcc1f61fdb0a0c82a1d3bf1879a414e7732ea894a7632af7f6d66ec7ab5bb3" - }, - "downloads": -1, - "filename": "Django-2.1.2-py3-none-any.whl", - "has_sig": false, - "md5_digest": "1c49272e55d2676455b7d03ec1374170", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7282853, - "upload_time": "2018-10-01T09:22:12", - "upload_time_iso_8601": "2018-10-01T09:22:12.971407Z", - "url": "https://files.pythonhosted.org/packages/32/ab/22530cc1b2114e6067eece94a333d6c749fa1c56a009f0721e51c181ea53/Django-2.1.2-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "8b034c74d3712919613f2c611e6689522df507a2753a92049009661a81b4b72f", - "md5": "383ca4e98ad5d0aa9d71378fe743bdef", - "sha256": "efbcad7ebb47daafbcead109b38a5bd519a3c3cd92c6ed0f691ff97fcdd16b45" - }, - "downloads": -1, - "filename": "Django-2.1.2.tar.gz", - "has_sig": false, - "md5_digest": "383ca4e98ad5d0aa9d71378fe743bdef", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8611286, - "upload_time": "2018-10-01T09:22:28", - "upload_time_iso_8601": "2018-10-01T09:22:28.985017Z", - "url": "https://files.pythonhosted.org/packages/8b/03/4c74d3712919613f2c611e6689522df507a2753a92049009661a81b4b72f/Django-2.1.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.1.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "d1e52676be45ea49cfd09a663f289376b3888accd57ff06c953297bfdee1fb08", - "md5": "1b7ed80cb0c0a62d12fd3ec4982d4d4b", - "sha256": "dd46d87af4c1bf54f4c926c3cfa41dc2b5c15782f15e4329752ce65f5dad1c37" - }, - "downloads": -1, - "filename": "Django-2.1.3-py3-none-any.whl", - "has_sig": false, - "md5_digest": "1b7ed80cb0c0a62d12fd3ec4982d4d4b", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7283286, - "upload_time": "2018-11-01T14:36:34", - "upload_time_iso_8601": "2018-11-01T14:36:34.965003Z", - "url": "https://files.pythonhosted.org/packages/d1/e5/2676be45ea49cfd09a663f289376b3888accd57ff06c953297bfdee1fb08/Django-2.1.3-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "93b10d6febb88712c39aced7df232d432fa22f5613c4bff246a1f4841248a60d", - "md5": "1a447e95c81e1c9b2a7ad61f4681f022", - "sha256": "1ffab268ada3d5684c05ba7ce776eaeedef360712358d6a6b340ae9f16486916" - }, - "downloads": -1, - "filename": "Django-2.1.3.tar.gz", - "has_sig": false, - "md5_digest": "1a447e95c81e1c9b2a7ad61f4681f022", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8611851, - "upload_time": "2018-11-01T14:36:54", - "upload_time_iso_8601": "2018-11-01T14:36:54.164080Z", - "url": "https://files.pythonhosted.org/packages/93/b1/0d6febb88712c39aced7df232d432fa22f5613c4bff246a1f4841248a60d/Django-2.1.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.1.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "fd9a0c028ea0fe4f5803dda1a7afabeed958d0c8b79b0fe762ffbf728db3b90d", - "md5": "96ce7a0bfe0237df2e16f3a6f82d9ea7", - "sha256": "55409a056b27e6d1246f19ede41c6c610e4cab549c005b62cbeefabc6433356b" - }, - "downloads": -1, - "filename": "Django-2.1.4-py3-none-any.whl", - "has_sig": false, - "md5_digest": "96ce7a0bfe0237df2e16f3a6f82d9ea7", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7282586, - "upload_time": "2018-12-03T17:02:55", - "upload_time_iso_8601": "2018-12-03T17:02:55.993312Z", - "url": "https://files.pythonhosted.org/packages/fd/9a/0c028ea0fe4f5803dda1a7afabeed958d0c8b79b0fe762ffbf728db3b90d/Django-2.1.4-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "83f74939b60c4127d5f49ccb570e34f4c59ecc222949220234a88e4f363f1456", - "md5": "3afc8bcec941e37221287f1a5323b1f1", - "sha256": "068d51054083d06ceb32ce02b7203f1854256047a0d58682677dd4f81bceabd7" - }, - "downloads": -1, - "filename": "Django-2.1.4.tar.gz", - "has_sig": false, - "md5_digest": "3afc8bcec941e37221287f1a5323b1f1", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8611886, - "upload_time": "2018-12-03T17:03:23", - "upload_time_iso_8601": "2018-12-03T17:03:23.910206Z", - "url": "https://files.pythonhosted.org/packages/83/f7/4939b60c4127d5f49ccb570e34f4c59ecc222949220234a88e4f363f1456/Django-2.1.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.1.5": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "3650078a42b4e9bedb94efd3e0278c0eb71650ed9672cdc91bd5542953bec17f", - "md5": "90ac057753cff4d5b154ef4ca3d0e1e6", - "sha256": "a32c22af23634e1d11425574dce756098e015a165be02e4690179889b207c7a8" - }, - "downloads": -1, - "filename": "Django-2.1.5-py3-none-any.whl", - "has_sig": false, - "md5_digest": "90ac057753cff4d5b154ef4ca3d0e1e6", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7280910, - "upload_time": "2019-01-04T13:52:50", - "upload_time_iso_8601": "2019-01-04T13:52:50.146659Z", - "url": "https://files.pythonhosted.org/packages/36/50/078a42b4e9bedb94efd3e0278c0eb71650ed9672cdc91bd5542953bec17f/Django-2.1.5-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "5c7f4c750e09b246621e5e90fa08f93dec1b991f5c203b0ff615d62a891c8f41", - "md5": "9309c48c8b92503b8969a7603a97e2a1", - "sha256": "d6393918da830530a9516bbbcbf7f1214c3d733738779f06b0f649f49cc698c3" - }, - "downloads": -1, - "filename": "Django-2.1.5.tar.gz", - "has_sig": false, - "md5_digest": "9309c48c8b92503b8969a7603a97e2a1", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8612384, - "upload_time": "2019-01-04T13:53:03", - "upload_time_iso_8601": "2019-01-04T13:53:03.556241Z", - "url": "https://files.pythonhosted.org/packages/5c/7f/4c750e09b246621e5e90fa08f93dec1b991f5c203b0ff615d62a891c8f41/Django-2.1.5.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.1.7": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "c787fbd666c4f87591ae25b7bb374298e8629816e87193c4099d3608ef11fab9", - "md5": "9b2efcc20342cb780630c02734553c1a", - "sha256": "275bec66fd2588dd517ada59b8bfb23d4a9abc5a362349139ddda3c7ff6f5ade" - }, - "downloads": -1, - "filename": "Django-2.1.7-py3-none-any.whl", - "has_sig": false, - "md5_digest": "9b2efcc20342cb780630c02734553c1a", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7281127, - "upload_time": "2019-02-11T15:10:47", - "upload_time_iso_8601": "2019-02-11T15:10:47.735733Z", - "url": "https://files.pythonhosted.org/packages/c7/87/fbd666c4f87591ae25b7bb374298e8629816e87193c4099d3608ef11fab9/Django-2.1.7-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "7eae29c28f6afddae0e305326078f31372f03d7f2e6d6210c9963843196ce67e", - "md5": "a042e6ba117d2e01950d842cceb5eee0", - "sha256": "939652e9d34d7d53d74d5d8ef82a19e5f8bb2de75618f7e5360691b6e9667963" - }, - "downloads": -1, - "filename": "Django-2.1.7.tar.gz", - "has_sig": false, - "md5_digest": "a042e6ba117d2e01950d842cceb5eee0", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8608548, - "upload_time": "2019-02-11T15:11:10", - "upload_time_iso_8601": "2019-02-11T15:11:10.020675Z", - "url": "https://files.pythonhosted.org/packages/7e/ae/29c28f6afddae0e305326078f31372f03d7f2e6d6210c9963843196ce67e/Django-2.1.7.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.1.8": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "a9e4fb8f473fe8ee659859cb712e25222243bbd55ece7c319301eeb60ccddc46", - "md5": "22ad9bd7cba46295a3d88473e37418d9", - "sha256": "0fd54e4f27bc3e0b7054a11e6b3a18fa53f2373f6b2df8a22e8eadfe018970a5" - }, - "downloads": -1, - "filename": "Django-2.1.8-py3-none-any.whl", - "has_sig": false, - "md5_digest": "22ad9bd7cba46295a3d88473e37418d9", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7281258, - "upload_time": "2019-04-01T09:18:55", - "upload_time_iso_8601": "2019-04-01T09:18:55.801525Z", - "url": "https://files.pythonhosted.org/packages/a9/e4/fb8f473fe8ee659859cb712e25222243bbd55ece7c319301eeb60ccddc46/Django-2.1.8-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "78a0c82585ddba004f99d19eeee277bc1b9e3b5e55d883e60ed713c0a4c9e3e8", - "md5": "e4492251be65fda35c6bc0c6ea03fd1a", - "sha256": "f3b28084101d516f56104856761bc247f85a2a5bbd9da39d9f6197ff461b3ee4" - }, - "downloads": -1, - "filename": "Django-2.1.8.tar.gz", - "has_sig": false, - "md5_digest": "e4492251be65fda35c6bc0c6ea03fd1a", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8613572, - "upload_time": "2019-04-01T09:19:04", - "upload_time_iso_8601": "2019-04-01T09:19:04.186926Z", - "url": "https://files.pythonhosted.org/packages/78/a0/c82585ddba004f99d19eeee277bc1b9e3b5e55d883e60ed713c0a4c9e3e8/Django-2.1.8.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.1.9": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "bb47a4fdf24409656dc624a802571c3d6bb809e396ebbe6d668b16cb8ae431fa", - "md5": "55f8903d2e34bc6ab9c76a2e0a12a14b", - "sha256": "bb72b5f8b53f8156280eaea520b548ac128a53f80cebc856c5e0fb555d44d529" - }, - "downloads": -1, - "filename": "Django-2.1.9-py3-none-any.whl", - "has_sig": false, - "md5_digest": "55f8903d2e34bc6ab9c76a2e0a12a14b", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7281385, - "upload_time": "2019-06-03T10:11:04", - "upload_time_iso_8601": "2019-06-03T10:11:04.565407Z", - "url": "https://files.pythonhosted.org/packages/bb/47/a4fdf24409656dc624a802571c3d6bb809e396ebbe6d668b16cb8ae431fa/Django-2.1.9-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "c1b33cdc60dc2e3c11236539f9470e42c5075a2e9c9f4885f5b4b912e9f19992", - "md5": "909c2e7761893a922dcf721521d9239e", - "sha256": "5052def4ff0a84bdf669827fdbd7b7cc1ac058f10232be6b21f37c6824f578da" - }, - "downloads": -1, - "filename": "Django-2.1.9.tar.gz", - "has_sig": false, - "md5_digest": "909c2e7761893a922dcf721521d9239e", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8608747, - "upload_time": "2019-06-03T10:11:23", - "upload_time_iso_8601": "2019-06-03T10:11:23.821054Z", - "url": "https://files.pythonhosted.org/packages/c1/b3/3cdc60dc2e3c11236539f9470e42c5075a2e9c9f4885f5b4b912e9f19992/Django-2.1.9.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.1a1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "6372d779f069acda9f5ce7aedca90a1410bfce03721e60c156472c95222cc1b7", - "md5": "813c556c785c948d5f10dc3a3f9383e6", - "sha256": "89ac9fce84db947da7e5d1e7a24eabd21216c04cc4386c93197f0aa2487e9945" - }, - "downloads": -1, - "filename": "Django-2.1a1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "813c556c785c948d5f10dc3a3f9383e6", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7202943, - "upload_time": "2018-05-18T01:01:19", - "upload_time_iso_8601": "2018-05-18T01:01:19.030204Z", - "url": "https://files.pythonhosted.org/packages/63/72/d779f069acda9f5ce7aedca90a1410bfce03721e60c156472c95222cc1b7/Django-2.1a1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - } - ], - "2.1b1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "91190e6f1e9e78d91cdb6fa6e326c24f7d765640f59888c982efdfcbb410146d", - "md5": "5c1ea8f1f0a1afdad997c16123bc8b1b", - "sha256": "21b35f96f3d84eae6a6f582156bacee1fd05badfa692da90cef48d9512e39e16" - }, - "downloads": -1, - "filename": "Django-2.1b1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "5c1ea8f1f0a1afdad997c16123bc8b1b", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7204079, - "upload_time": "2018-06-18T23:55:36", - "upload_time_iso_8601": "2018-06-18T23:55:36.588389Z", - "url": "https://files.pythonhosted.org/packages/91/19/0e6f1e9e78d91cdb6fa6e326c24f7d765640f59888c982efdfcbb410146d/Django-2.1b1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - } - ], - "2.1rc1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "503dbafe9b26f71cac7fa52621cdd43ce78ceca1798fd3c4aa8bfd4f680ec653", - "md5": "ce2d8596a0a8f83352c55b3802447468", - "sha256": "486f8c08fd8fcb44c66e5d6d4a845a28831b3f5be1bbd4dcc245938ed55cdff1" - }, - "downloads": -1, - "filename": "Django-2.1rc1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "ce2d8596a0a8f83352c55b3802447468", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7204422, - "upload_time": "2018-07-18T17:35:02", - "upload_time_iso_8601": "2018-07-18T17:35:02.726810Z", - "url": "https://files.pythonhosted.org/packages/50/3d/bafe9b26f71cac7fa52621cdd43ce78ceca1798fd3c4aa8bfd4f680ec653/Django-2.1rc1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - } - ], - "2.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "54850bef63668fb170888c1a2970ec897d4528d6072f32dee27653381a332642", - "md5": "a7a8bfd5fcd4cba08ca683944e375843", - "sha256": "a2814bffd1f007805b19194eb0b9a331933b82bd5da1c3ba3d7b7ba16e06dc4b" - }, - "downloads": -1, - "filename": "Django-2.2-py3-none-any.whl", - "has_sig": false, - "md5_digest": "a7a8bfd5fcd4cba08ca683944e375843", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7447163, - "upload_time": "2019-04-01T12:47:35", - "upload_time_iso_8601": "2019-04-01T12:47:35.288179Z", - "url": "https://files.pythonhosted.org/packages/54/85/0bef63668fb170888c1a2970ec897d4528d6072f32dee27653381a332642/Django-2.2-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "1789bb1b5b75d2ee7b7946d02c9284c067c827dc2b34031180a9442a774da8bf", - "md5": "41c4b4ec55b1cb373e5128156e9dcbd2", - "sha256": "7c3543e4fb070d14e10926189a7fcf42ba919263b7473dceaefce34d54e8a119" - }, - "downloads": -1, - "filename": "Django-2.2.tar.gz", - "has_sig": false, - "md5_digest": "41c4b4ec55b1cb373e5128156e9dcbd2", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8843237, - "upload_time": "2019-04-01T12:47:42", - "upload_time_iso_8601": "2019-04-01T12:47:42.240453Z", - "url": "https://files.pythonhosted.org/packages/17/89/bb1b5b75d2ee7b7946d02c9284c067c827dc2b34031180a9442a774da8bf/Django-2.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "b11d2476110614367adfb079a9bc718621f9fc8351e9214e1750cae1832d4090", - "md5": "8a2f51f779351edcbceda98719e07254", - "sha256": "bb407d0bb46395ca1241f829f5bd03f7e482f97f7d1936e26e98dacb201ed4ec" - }, - "downloads": -1, - "filename": "Django-2.2.1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "8a2f51f779351edcbceda98719e07254", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7447510, - "upload_time": "2019-05-01T06:57:39", - "upload_time_iso_8601": "2019-05-01T06:57:39.789016Z", - "url": "https://files.pythonhosted.org/packages/b1/1d/2476110614367adfb079a9bc718621f9fc8351e9214e1750cae1832d4090/Django-2.2.1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "fd7036c08f4c3b22523173b3a5e21fbdaa137bdb1722b76f356e0e2d5d8aa645", - "md5": "3b1721c1b5014316e1af8b10613c7592", - "sha256": "6fcc3cbd55b16f9a01f37de8bcbe286e0ea22e87096557f1511051780338eaea" - }, - "downloads": -1, - "filename": "Django-2.2.1.tar.gz", - "has_sig": false, - "md5_digest": "3b1721c1b5014316e1af8b10613c7592", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8973889, - "upload_time": "2019-05-01T06:57:47", - "upload_time_iso_8601": "2019-05-01T06:57:47.651991Z", - "url": "https://files.pythonhosted.org/packages/fd/70/36c08f4c3b22523173b3a5e21fbdaa137bdb1722b76f356e0e2d5d8aa645/Django-2.2.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.10": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "2bb2eb6230a30a5cc3a71b4df733de95c1a888e098e60b5e233703936f9c4dad", - "md5": "d24676ee3a4e112abc46f5363a608cd6", - "sha256": "9a4635813e2d498a3c01b10c701fe4a515d76dd290aaa792ccb65ca4ccb6b038" - }, - "downloads": -1, - "filename": "Django-2.2.10-py3-none-any.whl", - "has_sig": false, - "md5_digest": "d24676ee3a4e112abc46f5363a608cd6", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7460410, - "upload_time": "2020-02-03T09:50:41", - "upload_time_iso_8601": "2020-02-03T09:50:41.861960Z", - "url": "https://files.pythonhosted.org/packages/2b/b2/eb6230a30a5cc3a71b4df733de95c1a888e098e60b5e233703936f9c4dad/Django-2.2.10-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "b769abb0cfd9ee83209535e7021b34db6b19f4094a43a70a6e6f77da3c0ba606", - "md5": "10f192f8565ab137aea2dda4a4cb3d26", - "sha256": "1226168be1b1c7efd0e66ee79b0e0b58b2caa7ed87717909cd8a57bb13a7079a" - }, - "downloads": -1, - "filename": "Django-2.2.10.tar.gz", - "has_sig": false, - "md5_digest": "10f192f8565ab137aea2dda4a4cb3d26", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8865888, - "upload_time": "2020-02-03T09:50:57", - "upload_time_iso_8601": "2020-02-03T09:50:57.805388Z", - "url": "https://files.pythonhosted.org/packages/b7/69/abb0cfd9ee83209535e7021b34db6b19f4094a43a70a6e6f77da3c0ba606/Django-2.2.10.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.11": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "be767ccbcf52366590ca76997ce7860308b257b79962a4e4fada5353f72d7be5", - "md5": "c56b564c33b2803c00bb3087d1e316c2", - "sha256": "b51c9c548d5c3b3ccbb133d0bebc992e8ec3f14899bce8936e6fdda6b23a1881" - }, - "downloads": -1, - "filename": "Django-2.2.11-py3-none-any.whl", - "has_sig": false, - "md5_digest": "c56b564c33b2803c00bb3087d1e316c2", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7460553, - "upload_time": "2020-03-04T09:31:51", - "upload_time_iso_8601": "2020-03-04T09:31:51.267840Z", - "url": "https://files.pythonhosted.org/packages/be/76/7ccbcf52366590ca76997ce7860308b257b79962a4e4fada5353f72d7be5/Django-2.2.11-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "f05b428db83c37c2cf69e077ed327ada3511837371356204befc654b5b4bd444", - "md5": "3d8cc4ec1329c742d848c418932e488a", - "sha256": "65e2387e6bde531d3bb803244a2b74e0253550a9612c64a60c8c5be267b30f50" - }, - "downloads": -1, - "filename": "Django-2.2.11.tar.gz", - "has_sig": false, - "md5_digest": "3d8cc4ec1329c742d848c418932e488a", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 9010479, - "upload_time": "2020-03-04T09:32:08", - "upload_time_iso_8601": "2020-03-04T09:32:08.970734Z", - "url": "https://files.pythonhosted.org/packages/f0/5b/428db83c37c2cf69e077ed327ada3511837371356204befc654b5b4bd444/Django-2.2.11.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.12": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "afd1903cdbda68cd6ee74bf8ac7c86ffa04b2baf0254dfd6edeeafe4426c9c8b", - "md5": "9c6ffef77bb3d9eaf4abf76574b37656", - "sha256": "6ecd229e1815d4fc5240fc98f1cca78c41e7a8cd3e3f2eefadc4735031077916" - }, - "downloads": -1, - "filename": "Django-2.2.12-py3-none-any.whl", - "has_sig": false, - "md5_digest": "9c6ffef77bb3d9eaf4abf76574b37656", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7461035, - "upload_time": "2020-04-01T07:59:11", - "upload_time_iso_8601": "2020-04-01T07:59:11.345640Z", - "url": "https://files.pythonhosted.org/packages/af/d1/903cdbda68cd6ee74bf8ac7c86ffa04b2baf0254dfd6edeeafe4426c9c8b/Django-2.2.12-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "872582b01f8d6b154ea3023ab7d1412ca60875c105e8209350ec6cfee35b0f8d", - "md5": "6dc321ea9b123320702f678cc7287352", - "sha256": "69897097095f336d5aeef45b4103dceae51c00afa6d3ae198a2a18e519791b7a" - }, - "downloads": -1, - "filename": "Django-2.2.12.tar.gz", - "has_sig": false, - "md5_digest": "6dc321ea9b123320702f678cc7287352", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8877061, - "upload_time": "2020-04-01T07:59:21", - "upload_time_iso_8601": "2020-04-01T07:59:21.455385Z", - "url": "https://files.pythonhosted.org/packages/87/25/82b01f8d6b154ea3023ab7d1412ca60875c105e8209350ec6cfee35b0f8d/Django-2.2.12.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.13": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "fbe1c5520a00ae75060b0c03eea0115b272d6dc5dbd2fd3b75d0c0fbc9d262bc", - "md5": "d4a2170813154df2a6f976fb4ab027f5", - "sha256": "e8fe3c2b2212dce6126becab7a693157f1a441a07b62ec994c046c76af5bb66d" - }, - "downloads": -1, - "filename": "Django-2.2.13-py3-none-any.whl", - "has_sig": false, - "md5_digest": "d4a2170813154df2a6f976fb4ab027f5", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7465753, - "upload_time": "2020-06-03T09:36:32", - "upload_time_iso_8601": "2020-06-03T09:36:32.337142Z", - "url": "https://files.pythonhosted.org/packages/fb/e1/c5520a00ae75060b0c03eea0115b272d6dc5dbd2fd3b75d0c0fbc9d262bc/Django-2.2.13-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "da0c5ef899626da946b00322b70c2b28ac9d04f95cbedf525187ab4dad5db22d", - "md5": "30c688af9b63c4800ef9b044e0dd4145", - "sha256": "84f370f6acedbe1f3c41e1a02de44ac206efda3355e427139ecb785b5f596d80" - }, - "downloads": -1, - "filename": "Django-2.2.13.tar.gz", - "has_sig": false, - "md5_digest": "30c688af9b63c4800ef9b044e0dd4145", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8879757, - "upload_time": "2020-06-03T09:36:39", - "upload_time_iso_8601": "2020-06-03T09:36:39.587482Z", - "url": "https://files.pythonhosted.org/packages/da/0c/5ef899626da946b00322b70c2b28ac9d04f95cbedf525187ab4dad5db22d/Django-2.2.13.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.14": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "f26cf7e0ed3d07952742439be43e7fb5a8b07b065ab927c6493be2a6cea59f33", - "md5": "b3a2cbc20e6e591687ff83821f400a46", - "sha256": "f2250bd35d0f6c23e930c544629934144e5dd39a4c06092e1050c731c1712ba8" - }, - "downloads": -1, - "filename": "Django-2.2.14-py3-none-any.whl", - "has_sig": false, - "md5_digest": "b3a2cbc20e6e591687ff83821f400a46", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7465761, - "upload_time": "2020-07-01T04:49:29", - "upload_time_iso_8601": "2020-07-01T04:49:29.459267Z", - "url": "https://files.pythonhosted.org/packages/f2/6c/f7e0ed3d07952742439be43e7fb5a8b07b065ab927c6493be2a6cea59f33/Django-2.2.14-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "7846a3af6eb3037044eaad9518082eff5ed473ad343dee165eb14b1f1a4a3d87", - "md5": "d60c50b5bfed1dc11398abb7e41a6e55", - "sha256": "edf0ecf6657713b0435b6757e6069466925cae70d634a3283c96b80c01e06191" - }, - "downloads": -1, - "filename": "Django-2.2.14.tar.gz", - "has_sig": false, - "md5_digest": "d60c50b5bfed1dc11398abb7e41a6e55", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 9022051, - "upload_time": "2020-07-01T04:50:13", - "upload_time_iso_8601": "2020-07-01T04:50:13.815347Z", - "url": "https://files.pythonhosted.org/packages/78/46/a3af6eb3037044eaad9518082eff5ed473ad343dee165eb14b1f1a4a3d87/Django-2.2.14.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.15": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "3a5a9e3cffed37fa9a277c04377b71b9038c270c2677c86cf406f803df89038f", - "md5": "c90b81bd50ea18fe6291094b2ba1000d", - "sha256": "91f540000227eace0504a24f508de26daa756353aa7376c6972d7920bc339a3a" - }, - "downloads": -1, - "filename": "Django-2.2.15-py3-none-any.whl", - "has_sig": false, - "md5_digest": "c90b81bd50ea18fe6291094b2ba1000d", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7465895, - "upload_time": "2020-08-03T07:23:23", - "upload_time_iso_8601": "2020-08-03T07:23:23.362449Z", - "url": "https://files.pythonhosted.org/packages/3a/5a/9e3cffed37fa9a277c04377b71b9038c270c2677c86cf406f803df89038f/Django-2.2.15-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "51b000dadff17778c3204bbd658dcbc9dda1d5e29fff0862ac9dd10880d8b265", - "md5": "a0885a388b6d9247ec1c305b0ef3fa27", - "sha256": "3e2f5d172215862abf2bac3138d8a04229d34dbd2d0dab42c6bf33876cc22323" - }, - "downloads": -1, - "filename": "Django-2.2.15.tar.gz", - "has_sig": false, - "md5_digest": "a0885a388b6d9247ec1c305b0ef3fa27", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 9023679, - "upload_time": "2020-08-03T07:23:37", - "upload_time_iso_8601": "2020-08-03T07:23:37.474778Z", - "url": "https://files.pythonhosted.org/packages/51/b0/00dadff17778c3204bbd658dcbc9dda1d5e29fff0862ac9dd10880d8b265/Django-2.2.15.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.16": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "06303b7e846c1a2f5435f52c408442e09c0c35ca6fa1c4abbad0ccf4d42fb172", - "md5": "d1ed6773fc10ce59492e8e871c32689e", - "sha256": "83ced795a0f239f41d8ecabf51cc5fad4b97462a6008dc12e5af3cb9288724ec" - }, - "downloads": -1, - "filename": "Django-2.2.16-py3-none-any.whl", - "has_sig": false, - "md5_digest": "d1ed6773fc10ce59492e8e871c32689e", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7466055, - "upload_time": "2020-09-01T09:14:25", - "upload_time_iso_8601": "2020-09-01T09:14:25.126958Z", - "url": "https://files.pythonhosted.org/packages/06/30/3b7e846c1a2f5435f52c408442e09c0c35ca6fa1c4abbad0ccf4d42fb172/Django-2.2.16-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "eb9c1beb8860f99acedd3e231c9690e4a9e7103818a1fff9d44ae9553b2e1442", - "md5": "93faf5bbd54a19ea49f4932a813b9758", - "sha256": "62cf45e5ee425c52e411c0742e641a6588b7e8af0d2c274a27940931b2786594" - }, - "downloads": -1, - "filename": "Django-2.2.16.tar.gz", - "has_sig": false, - "md5_digest": "93faf5bbd54a19ea49f4932a813b9758", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8884774, - "upload_time": "2020-09-01T09:14:36", - "upload_time_iso_8601": "2020-09-01T09:14:36.209966Z", - "url": "https://files.pythonhosted.org/packages/eb/9c/1beb8860f99acedd3e231c9690e4a9e7103818a1fff9d44ae9553b2e1442/Django-2.2.16.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.17": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "822b75f2909ba02a3b0e343e560863101aa3d43f58357e7c053596aa29d1cce7", - "md5": "5587ac4b0e4863cbfaa6a8d9158d1e75", - "sha256": "558cb27930defd9a6042133258caf797b2d1dee233959f537e3dc475cb49bd7c" - }, - "downloads": -1, - "filename": "Django-2.2.17-py3-none-any.whl", - "has_sig": false, - "md5_digest": "5587ac4b0e4863cbfaa6a8d9158d1e75", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7466059, - "upload_time": "2020-11-02T08:12:33", - "upload_time_iso_8601": "2020-11-02T08:12:33.238873Z", - "url": "https://files.pythonhosted.org/packages/82/2b/75f2909ba02a3b0e343e560863101aa3d43f58357e7c053596aa29d1cce7/Django-2.2.17-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "9dfd1d27e69b17b3c14c2d6120b17319725461e2732257ac26c64c17fcba53f4", - "md5": "832805a3fdf817d4546609df1ed2a174", - "sha256": "cf5370a4d7765a9dd6d42a7b96b53c74f9446cd38209211304b210fe0404b861" - }, - "downloads": -1, - "filename": "Django-2.2.17.tar.gz", - "has_sig": false, - "md5_digest": "832805a3fdf817d4546609df1ed2a174", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8885492, - "upload_time": "2020-11-02T08:12:44", - "upload_time_iso_8601": "2020-11-02T08:12:44.286790Z", - "url": "https://files.pythonhosted.org/packages/9d/fd/1d27e69b17b3c14c2d6120b17319725461e2732257ac26c64c17fcba53f4/Django-2.2.17.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.18": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "6a4c4072f2f11c523898fb994874bbd1346d2e6c0b6968a2e1e87bee1c7a1729", - "md5": "3629a741684c0734df8c87964be5da0f", - "sha256": "0eaca08f236bf502a9773e53623f766cc3ceee6453cc41e6de1c8b80f07d2364" - }, - "downloads": -1, - "filename": "Django-2.2.18-py3-none-any.whl", - "has_sig": false, - "md5_digest": "3629a741684c0734df8c87964be5da0f", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7466191, - "upload_time": "2021-02-01T09:28:14", - "upload_time_iso_8601": "2021-02-01T09:28:14.241915Z", - "url": "https://files.pythonhosted.org/packages/6a/4c/4072f2f11c523898fb994874bbd1346d2e6c0b6968a2e1e87bee1c7a1729/Django-2.2.18-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "f04247a9ba406657df4e7244697fcd4c7916b1ee4e89f9bbc5d850d08751f0f3", - "md5": "c6cf78dae9c0be5833d37be73ab63962", - "sha256": "c9c994f5e0a032cbd45089798b52e4080f4dea7241c58e3e0636c54146480bb4" - }, - "downloads": -1, - "filename": "Django-2.2.18.tar.gz", - "has_sig": false, - "md5_digest": "c6cf78dae9c0be5833d37be73ab63962", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 9180844, - "upload_time": "2021-02-01T09:28:29", - "upload_time_iso_8601": "2021-02-01T09:28:29.173011Z", - "url": "https://files.pythonhosted.org/packages/f0/42/47a9ba406657df4e7244697fcd4c7916b1ee4e89f9bbc5d850d08751f0f3/Django-2.2.18.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.19": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "1ace846a9dbf536991be0004f5ae414520c3a64eaa167d09e51d75ab410c45e8", - "md5": "6f579975354c5c63912dc130092cc35f", - "sha256": "e319a7164d6d30cb177b3fd74d02c52f1185c37304057bb76d74047889c605d9" - }, - "downloads": -1, - "filename": "Django-2.2.19-py3-none-any.whl", - "has_sig": false, - "md5_digest": "6f579975354c5c63912dc130092cc35f", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7466191, - "upload_time": "2021-02-19T09:07:57", - "upload_time_iso_8601": "2021-02-19T09:07:57.568044Z", - "url": "https://files.pythonhosted.org/packages/1a/ce/846a9dbf536991be0004f5ae414520c3a64eaa167d09e51d75ab410c45e8/Django-2.2.19-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "dc35ad92bf082f7faa39bd949911b9df09ec1843167d657df6eb3720d2212427", - "md5": "adecf675c2af9dab8ed65246963718d4", - "sha256": "30c235dec87e05667597e339f194c9fed6c855bda637266ceee891bf9093da43" - }, - "downloads": -1, - "filename": "Django-2.2.19.tar.gz", - "has_sig": false, - "md5_digest": "adecf675c2af9dab8ed65246963718d4", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 9209434, - "upload_time": "2021-02-19T09:08:15", - "upload_time_iso_8601": "2021-02-19T09:08:15.073175Z", - "url": "https://files.pythonhosted.org/packages/dc/35/ad92bf082f7faa39bd949911b9df09ec1843167d657df6eb3720d2212427/Django-2.2.19.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "eb4b743d5008fc7432c714d753e1fc7ee56c6a776dc566cc6cfb4136d46cdcbb", - "md5": "41fdd9254fcbce92001c6881ba5af68d", - "sha256": "7cb67e8b934fab23b6daed7144da52e8a25a47eba7f360ca43d2b448506b01ad" - }, - "downloads": -1, - "filename": "Django-2.2.2-py3-none-any.whl", - "has_sig": false, - "md5_digest": "41fdd9254fcbce92001c6881ba5af68d", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7447968, - "upload_time": "2019-06-03T10:11:10", - "upload_time_iso_8601": "2019-06-03T10:11:10.358783Z", - "url": "https://files.pythonhosted.org/packages/eb/4b/743d5008fc7432c714d753e1fc7ee56c6a776dc566cc6cfb4136d46cdcbb/Django-2.2.2-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "562ec0495314c0bdcf19d9b888a98ff16a4c58a90dd77ed741f4dbab2cbf7efe", - "md5": "c52b05c2bc4898bd68dc0359347fff69", - "sha256": "753d30d3eb078064d2ddadfea65083c9848074a7f93d7b4dc7fa6b1380d278f5" - }, - "downloads": -1, - "filename": "Django-2.2.2.tar.gz", - "has_sig": false, - "md5_digest": "c52b05c2bc4898bd68dc0359347fff69", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8841523, - "upload_time": "2019-06-03T10:11:32", - "upload_time_iso_8601": "2019-06-03T10:11:32.109900Z", - "url": "https://files.pythonhosted.org/packages/56/2e/c0495314c0bdcf19d9b888a98ff16a4c58a90dd77ed741f4dbab2cbf7efe/Django-2.2.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.20": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "b188bd220ea3e4e12c42f2215285cb101d925f95e42b093f390c823e61c94f00", - "md5": "97add963ed8a45386a704b494a8c59a2", - "sha256": "2484f115891ab1a0e9ae153602a641fbc15d7894c036d79fb78662c0965d7954" - }, - "downloads": -1, - "filename": "Django-2.2.20-py3-none-any.whl", - "has_sig": false, - "md5_digest": "97add963ed8a45386a704b494a8c59a2", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7466207, - "upload_time": "2021-04-06T07:34:52", - "upload_time_iso_8601": "2021-04-06T07:34:52.201974Z", - "url": "https://files.pythonhosted.org/packages/b1/88/bd220ea3e4e12c42f2215285cb101d925f95e42b093f390c823e61c94f00/Django-2.2.20-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "3ce02ab562729727dc25d800134f811f8e15e56fb28c44dfab5222a51b9d2b20", - "md5": "947060d96ccc0a05e8049d839e541b25", - "sha256": "2569f9dc5f8e458a5e988b03d6b7a02bda59b006d6782f4ea0fd590ed7336a64" - }, - "downloads": -1, - "filename": "Django-2.2.20.tar.gz", - "has_sig": false, - "md5_digest": "947060d96ccc0a05e8049d839e541b25", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 9182853, - "upload_time": "2021-04-06T07:35:02", - "upload_time_iso_8601": "2021-04-06T07:35:02.866945Z", - "url": "https://files.pythonhosted.org/packages/3c/e0/2ab562729727dc25d800134f811f8e15e56fb28c44dfab5222a51b9d2b20/Django-2.2.20.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.21": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "8193d6dc44d73dfa9e773f97c90edccbb0c359f7bb27aadc90f5c0da48ab7138", - "md5": "bb7b593f29908cc5eac710467ca59630", - "sha256": "7d8b61b6e7031a3c1a4aef29e2701849af0837f4f2062bc203077a0b46d51c2c" - }, - "downloads": -1, - "filename": "Django-2.2.21-py3-none-any.whl", - "has_sig": false, - "md5_digest": "bb7b593f29908cc5eac710467ca59630", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7466829, - "upload_time": "2021-05-04T08:47:20", - "upload_time_iso_8601": "2021-05-04T08:47:20.595323Z", - "url": "https://files.pythonhosted.org/packages/81/93/d6dc44d73dfa9e773f97c90edccbb0c359f7bb27aadc90f5c0da48ab7138/Django-2.2.21-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "868c95243cf026de8c006766f2f583cd582f51fec28158dec2a0684ac77ccca3", - "md5": "fa2da272f5103dfe56c4ddc6d43037ca", - "sha256": "7460cfe3781d36d1625230267dad255deb33e9229e41f21e32b33b9d536d20cd" - }, - "downloads": -1, - "filename": "Django-2.2.21.tar.gz", - "has_sig": false, - "md5_digest": "fa2da272f5103dfe56c4ddc6d43037ca", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 9209871, - "upload_time": "2021-05-04T08:47:51", - "upload_time_iso_8601": "2021-05-04T08:47:51.750991Z", - "url": "https://files.pythonhosted.org/packages/86/8c/95243cf026de8c006766f2f583cd582f51fec28158dec2a0684ac77ccca3/Django-2.2.21.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.22": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "6cccd8ace1c2de7a149934cae343eccc75fd07d261c8f3bee05008fc630186d5", - "md5": "89b61e5257362f4ed46ebbea9ae18dcd", - "sha256": "e831105edb153af1324de44d06091ca75520a227456387dda4a47d2f1cc2731a" - }, - "downloads": -1, - "filename": "Django-2.2.22-py3-none-any.whl", - "has_sig": false, - "md5_digest": "89b61e5257362f4ed46ebbea9ae18dcd", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7466871, - "upload_time": "2021-05-06T07:40:38", - "upload_time_iso_8601": "2021-05-06T07:40:38.479356Z", - "url": "https://files.pythonhosted.org/packages/6c/cc/d8ace1c2de7a149934cae343eccc75fd07d261c8f3bee05008fc630186d5/Django-2.2.22-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "7d597c19d4a11f976fac155b0d60b6d4f41dc986209296de2af971b11568bac6", - "md5": "dca447b605dcabd924ac7ba17680cf73", - "sha256": "db2214db1c99017cbd971e58824e6f424375154fe358afc30e976f5b99fc6060" - }, - "downloads": -1, - "filename": "Django-2.2.22.tar.gz", - "has_sig": false, - "md5_digest": "dca447b605dcabd924ac7ba17680cf73", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 9182392, - "upload_time": "2021-05-06T07:40:41", - "upload_time_iso_8601": "2021-05-06T07:40:41.663188Z", - "url": "https://files.pythonhosted.org/packages/7d/59/7c19d4a11f976fac155b0d60b6d4f41dc986209296de2af971b11568bac6/Django-2.2.22.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.23": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "fa355cf6c94b1cc7ac847fde1d36257e06533ce853c2f92875b7abac89a65bfc", - "md5": "7d1c3f821f8dc4e5ee1018b16f6b9157", - "sha256": "2710bff9dd480cf886e38947ee00aea3d6b9b04b77a748e352e3ce447b0fe17f" - }, - "downloads": -1, - "filename": "Django-2.2.23-py3-none-any.whl", - "has_sig": false, - "md5_digest": "7d1c3f821f8dc4e5ee1018b16f6b9157", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7467076, - "upload_time": "2021-05-13T07:36:40", - "upload_time_iso_8601": "2021-05-13T07:36:40.797557Z", - "url": "https://files.pythonhosted.org/packages/fa/35/5cf6c94b1cc7ac847fde1d36257e06533ce853c2f92875b7abac89a65bfc/Django-2.2.23-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "51fb6ad422c1a18fcb21f014091316c56031b739aef6ad98d65880b870bb311c", - "md5": "d72405637143e201b745714e300bb546", - "sha256": "12cfc045a4ccb2348719aaaa77b17e66a26bff9fc238b4c765a3e825ef92e414" - }, - "downloads": -1, - "filename": "Django-2.2.23.tar.gz", - "has_sig": false, - "md5_digest": "d72405637143e201b745714e300bb546", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 9182567, - "upload_time": "2021-05-13T07:36:51", - "upload_time_iso_8601": "2021-05-13T07:36:51.046462Z", - "url": "https://files.pythonhosted.org/packages/51/fb/6ad422c1a18fcb21f014091316c56031b739aef6ad98d65880b870bb311c/Django-2.2.23.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.24": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "654a3d3f05ff8cb284cca565adb8b704b3094b0810ae2b5925c6e30e604400f7", - "md5": "8c2101fdb7f16c71c4751cb12d57922d", - "sha256": "f2084ceecff86b1e631c2cd4107d435daf4e12f1efcdf11061a73bf0b5e95f92" - }, - "downloads": -1, - "filename": "Django-2.2.24-py3-none-any.whl", - "has_sig": false, - "md5_digest": "8c2101fdb7f16c71c4751cb12d57922d", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7467267, - "upload_time": "2021-06-02T08:53:39", - "upload_time_iso_8601": "2021-06-02T08:53:39.899614Z", - "url": "https://files.pythonhosted.org/packages/65/4a/3d3f05ff8cb284cca565adb8b704b3094b0810ae2b5925c6e30e604400f7/Django-2.2.24-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "f9a3bc91152d021c340f35b3c6d463602704e6590b0273df7e27fa80ed87cb14", - "md5": "ebf3bbb7716a7b11029e860475b9a122", - "sha256": "3339ff0e03dee13045aef6ae7b523edff75b6d726adf7a7a48f53d5a501f7db7" - }, - "downloads": -1, - "filename": "Django-2.2.24.tar.gz", - "has_sig": false, - "md5_digest": "ebf3bbb7716a7b11029e860475b9a122", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 9211396, - "upload_time": "2021-06-02T08:54:12", - "upload_time_iso_8601": "2021-06-02T08:54:12.637481Z", - "url": "https://files.pythonhosted.org/packages/f9/a3/bc91152d021c340f35b3c6d463602704e6590b0273df7e27fa80ed87cb14/Django-2.2.24.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.25": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "5e211ee492ab7d6b10fd7d5f519161837f9bfae85b375dcf9dfbddfd4d9f3f4c", - "md5": "5cef2f8bbca79918c65425c1b9f5dd96", - "sha256": "08bad7ef7e90286b438dbe1412c3e633fbc7b96db04735f0c7baadeed52f3fad" - }, - "downloads": -1, - "filename": "Django-2.2.25-py3-none-any.whl", - "has_sig": false, - "md5_digest": "5cef2f8bbca79918c65425c1b9f5dd96", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7467310, - "upload_time": "2021-12-07T07:34:42", - "upload_time_iso_8601": "2021-12-07T07:34:42.416731Z", - "url": "https://files.pythonhosted.org/packages/5e/21/1ee492ab7d6b10fd7d5f519161837f9bfae85b375dcf9dfbddfd4d9f3f4c/Django-2.2.25-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "07f0b14af22a49ed975b94f23e70470bbb64b3ce4b0459f5f6ec8b7a2b6edd09", - "md5": "2f9c16ab2a38e330897d0adb84ce0268", - "sha256": "b1e65eaf371347d4b13eb7e061b09786c973061de95390c327c85c1e2aa2349c" - }, - "downloads": -1, - "filename": "Django-2.2.25.tar.gz", - "has_sig": false, - "md5_digest": "2f9c16ab2a38e330897d0adb84ce0268", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 9185326, - "upload_time": "2021-12-07T07:34:54", - "upload_time_iso_8601": "2021-12-07T07:34:54.595264Z", - "url": "https://files.pythonhosted.org/packages/07/f0/b14af22a49ed975b94f23e70470bbb64b3ce4b0459f5f6ec8b7a2b6edd09/Django-2.2.25.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.26": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "8a3d40f62f004b9a763a3726432d4c820526e880f7f2902de7469e5c0d119590", - "md5": "85532c47020dc8a35c6a3476a90e8458", - "sha256": "85e62019366692f1d5afed946ca32fef34c8693edf342ac9d067d75d64faf0ac" - }, - "downloads": -1, - "filename": "Django-2.2.26-py3-none-any.whl", - "has_sig": false, - "md5_digest": "85532c47020dc8a35c6a3476a90e8458", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7468014, - "upload_time": "2022-01-04T09:53:18", - "upload_time_iso_8601": "2022-01-04T09:53:18.329006Z", - "url": "https://files.pythonhosted.org/packages/8a/3d/40f62f004b9a763a3726432d4c820526e880f7f2902de7469e5c0d119590/Django-2.2.26-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "6490e1557256c4e4113953d7ce2aa02e70a024d09658ce1b55e0f7ea9b61f17a", - "md5": "1ab7932747f3e473d0a2fcb7ebc5d700", - "sha256": "dfa537267d52c6243a62b32855a744ca83c37c70600aacffbfd98bc5d6d8518f" - }, - "downloads": -1, - "filename": "Django-2.2.26.tar.gz", - "has_sig": false, - "md5_digest": "1ab7932747f3e473d0a2fcb7ebc5d700", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 9208028, - "upload_time": "2022-01-04T09:53:29", - "upload_time_iso_8601": "2022-01-04T09:53:29.324996Z", - "url": "https://files.pythonhosted.org/packages/64/90/e1557256c4e4113953d7ce2aa02e70a024d09658ce1b55e0f7ea9b61f17a/Django-2.2.26.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.27": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "cf2b229178fdeea21210768620d0dab8ca40ed2f823bbd57b93c39dbd644ef90", - "md5": "0e77b076deeb3e86341f5fe03fa5abb7", - "sha256": "90763c764738586b11d7e1f44828032c153366e43ad7f782908193a1bb2d6d92" - }, - "downloads": -1, - "filename": "Django-2.2.27-py3-none-any.whl", - "has_sig": false, - "md5_digest": "0e77b076deeb3e86341f5fe03fa5abb7", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7468028, - "upload_time": "2022-02-01T07:56:15", - "upload_time_iso_8601": "2022-02-01T07:56:15.889914Z", - "url": "https://files.pythonhosted.org/packages/cf/2b/229178fdeea21210768620d0dab8ca40ed2f823bbd57b93c39dbd644ef90/Django-2.2.27-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "38c59f8bd15f2193c72f8f7e22b938380a2f043e7e6f5cbdaa371d969ae50424", - "md5": "4af3aeed9e515ccde107ae6a9804c31f", - "sha256": "1ee37046b0bf2b61e83b3a01d067323516ec3b6f2b17cd49b1326dd4ba9dc913" - }, - "downloads": -1, - "filename": "Django-2.2.27.tar.gz", - "has_sig": false, - "md5_digest": "4af3aeed9e515ccde107ae6a9804c31f", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 9185716, - "upload_time": "2022-02-01T07:56:27", - "upload_time_iso_8601": "2022-02-01T07:56:27.211309Z", - "url": "https://files.pythonhosted.org/packages/38/c5/9f8bd15f2193c72f8f7e22b938380a2f043e7e6f5cbdaa371d969ae50424/Django-2.2.27.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.28": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "4369eaac9a827335898d626066989865974fa48bfc23efd1b34628d64543b2d9", - "md5": "092eb87671abf03f688d85ee2cd9396d", - "sha256": "365429d07c1336eb42ba15aa79f45e1c13a0b04d5c21569e7d596696418a6a45" - }, - "downloads": -1, - "filename": "Django-2.2.28-py3-none-any.whl", - "has_sig": false, - "md5_digest": "092eb87671abf03f688d85ee2cd9396d", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7468548, - "upload_time": "2022-04-11T07:52:52", - "upload_time_iso_8601": "2022-04-11T07:52:52.782601Z", - "url": "https://files.pythonhosted.org/packages/43/69/eaac9a827335898d626066989865974fa48bfc23efd1b34628d64543b2d9/Django-2.2.28-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "172724d66baeb05a53e38637a4b793df069f2bebacf75bc6e854ff32068a7ef3", - "md5": "62550f105ef66ac7d08e0126f457578a", - "sha256": "0200b657afbf1bc08003845ddda053c7641b9b24951e52acd51f6abda33a7413" - }, - "downloads": -1, - "filename": "Django-2.2.28.tar.gz", - "has_sig": false, - "md5_digest": "62550f105ef66ac7d08e0126f457578a", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 9187543, - "upload_time": "2022-04-11T07:53:05", - "upload_time_iso_8601": "2022-04-11T07:53:05.463179Z", - "url": "https://files.pythonhosted.org/packages/17/27/24d66baeb05a53e38637a4b793df069f2bebacf75bc6e854ff32068a7ef3/Django-2.2.28.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "39b02138c31bf13e17afc32277239da53e9dfcce27bac8cb68cf1c0123f1fdf5", - "md5": "32c2feb280afee531389ec8fa38f49d8", - "sha256": "6e974d4b57e3b29e4882b244d40171d6a75202ab8d2402b8e8adbd182e25cf0c" - }, - "downloads": -1, - "filename": "Django-2.2.3-py3-none-any.whl", - "has_sig": false, - "md5_digest": "32c2feb280afee531389ec8fa38f49d8", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7459212, - "upload_time": "2019-07-01T07:19:23", - "upload_time_iso_8601": "2019-07-01T07:19:23.256282Z", - "url": "https://files.pythonhosted.org/packages/39/b0/2138c31bf13e17afc32277239da53e9dfcce27bac8cb68cf1c0123f1fdf5/Django-2.2.3-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "d320b447eb3d820e0d05fe83cbfb016842bd8da35f4b0f83498dca43d02aebc3", - "md5": "f152164e77d38460ee06c42c210d2f57", - "sha256": "4d23f61b26892bac785f07401bc38cbf8fa4cec993f400e9cd9ddf28fd51c0ea" - }, - "downloads": -1, - "filename": "Django-2.2.3.tar.gz", - "has_sig": false, - "md5_digest": "f152164e77d38460ee06c42c210d2f57", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8992109, - "upload_time": "2019-07-01T07:19:43", - "upload_time_iso_8601": "2019-07-01T07:19:43.693346Z", - "url": "https://files.pythonhosted.org/packages/d3/20/b447eb3d820e0d05fe83cbfb016842bd8da35f4b0f83498dca43d02aebc3/Django-2.2.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "d65766997ca6ef17d2d0f0ebcd860bc6778095ffee04077ca8985928175da358", - "md5": "0b4efcaafec4ef999513c9f40c7e3746", - "sha256": "9a2f98211ab474c710fcdad29c82f30fc14ce9917c7a70c3682162a624de4035" - }, - "downloads": -1, - "filename": "Django-2.2.4-py3-none-any.whl", - "has_sig": false, - "md5_digest": "0b4efcaafec4ef999513c9f40c7e3746", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7459472, - "upload_time": "2019-08-01T09:04:37", - "upload_time_iso_8601": "2019-08-01T09:04:37.287191Z", - "url": "https://files.pythonhosted.org/packages/d6/57/66997ca6ef17d2d0f0ebcd860bc6778095ffee04077ca8985928175da358/Django-2.2.4-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "19113449a2071df9427e7a5c4dddee2462e88840dd968a9b0c161097154fcb0c", - "md5": "b32e396c354880742d85a7628a0bdd5a", - "sha256": "16a5d54411599780ac9dfe3b9b38f90f785c51259a584e0b24b6f14a7f69aae8" - }, - "downloads": -1, - "filename": "Django-2.2.4.tar.gz", - "has_sig": false, - "md5_digest": "b32e396c354880742d85a7628a0bdd5a", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8856979, - "upload_time": "2019-08-01T09:05:00", - "upload_time_iso_8601": "2019-08-01T09:05:00.463982Z", - "url": "https://files.pythonhosted.org/packages/19/11/3449a2071df9427e7a5c4dddee2462e88840dd968a9b0c161097154fcb0c/Django-2.2.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.5": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "949fa56f7893b1280e5019482260e246ab944d54a9a633a01ed04683d9ce5078", - "md5": "a3b581f61effe58a8dfd765e4d960ce0", - "sha256": "148a4a2d1a85b23883b0a4e99ab7718f518a83675e4485e44dc0c1d36988c5fa" - }, - "downloads": -1, - "filename": "Django-2.2.5-py3-none-any.whl", - "has_sig": false, - "md5_digest": "a3b581f61effe58a8dfd765e4d960ce0", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7459805, - "upload_time": "2019-09-02T07:18:39", - "upload_time_iso_8601": "2019-09-02T07:18:39.462663Z", - "url": "https://files.pythonhosted.org/packages/94/9f/a56f7893b1280e5019482260e246ab944d54a9a633a01ed04683d9ce5078/Django-2.2.5-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "1d0679ddea0bfd4e7cd1f9fa4700c8e524820a5263c6fd8bb91db14f1812c17d", - "md5": "a48efbcf951c1320feefb2841500e997", - "sha256": "deb70aa038e59b58593673b15e9a711d1e5ccd941b5973b30750d5d026abfd56" - }, - "downloads": -1, - "filename": "Django-2.2.5.tar.gz", - "has_sig": false, - "md5_digest": "a48efbcf951c1320feefb2841500e997", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8995543, - "upload_time": "2019-09-02T07:19:02", - "upload_time_iso_8601": "2019-09-02T07:19:02.248056Z", - "url": "https://files.pythonhosted.org/packages/1d/06/79ddea0bfd4e7cd1f9fa4700c8e524820a5263c6fd8bb91db14f1812c17d/Django-2.2.5.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.6": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "b279df0ffea7bf1e02c073c2633702c90f4384645c40a1dd09a308e02ef0c817", - "md5": "b44e5fdbabf8a34b4748b226979561ac", - "sha256": "4025317ca01f75fc79250ff7262a06d8ba97cd4f82e93394b2a0a6a4a925caeb" - }, - "downloads": -1, - "filename": "Django-2.2.6-py3-none-any.whl", - "has_sig": false, - "md5_digest": "b44e5fdbabf8a34b4748b226979561ac", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7459712, - "upload_time": "2019-10-01T08:36:44", - "upload_time_iso_8601": "2019-10-01T08:36:44.466204Z", - "url": "https://files.pythonhosted.org/packages/b2/79/df0ffea7bf1e02c073c2633702c90f4384645c40a1dd09a308e02ef0c817/Django-2.2.6-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "c72cbbd0fddf6a08456c3100b8e8b230f3288d4511985aa4e2368b0d115b5aae", - "md5": "796c175a2f94e938c60d84b4565216af", - "sha256": "a8ca1033acac9f33995eb2209a6bf18a4681c3e5269a878e9a7e0b7384ed1ca3" - }, - "downloads": -1, - "filename": "Django-2.2.6.tar.gz", - "has_sig": false, - "md5_digest": "796c175a2f94e938c60d84b4565216af", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8859044, - "upload_time": "2019-10-01T08:37:07", - "upload_time_iso_8601": "2019-10-01T08:37:07.504139Z", - "url": "https://files.pythonhosted.org/packages/c7/2c/bbd0fddf6a08456c3100b8e8b230f3288d4511985aa4e2368b0d115b5aae/Django-2.2.6.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.7": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "a036463632a2e9161a7e713488d719a280e8cb0c7e3a66ed32a32e801891caae", - "md5": "501704dd5d29b597763a8e9dd7737f6b", - "sha256": "89c2007ca4fa5b351a51a279eccff298520783b713bf28efb89dfb81c80ea49b" - }, - "downloads": -1, - "filename": "Django-2.2.7-py3-none-any.whl", - "has_sig": false, - "md5_digest": "501704dd5d29b597763a8e9dd7737f6b", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7459759, - "upload_time": "2019-11-04T08:33:19", - "upload_time_iso_8601": "2019-11-04T08:33:19.548538Z", - "url": "https://files.pythonhosted.org/packages/a0/36/463632a2e9161a7e713488d719a280e8cb0c7e3a66ed32a32e801891caae/Django-2.2.7-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "0d055de305261e0a6bcd5701e2bfb5237e76303fde36f1f7c5a40ff86480ab5a", - "md5": "b0833024aac4c8240467e4dc91a12e9b", - "sha256": "16040e1288c6c9f68c6da2fe75ebde83c0a158f6f5d54f4c5177b0c1478c5b86" - }, - "downloads": -1, - "filename": "Django-2.2.7.tar.gz", - "has_sig": false, - "md5_digest": "b0833024aac4c8240467e4dc91a12e9b", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8999415, - "upload_time": "2019-11-04T08:33:42", - "upload_time_iso_8601": "2019-11-04T08:33:42.021959Z", - "url": "https://files.pythonhosted.org/packages/0d/05/5de305261e0a6bcd5701e2bfb5237e76303fde36f1f7c5a40ff86480ab5a/Django-2.2.7.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.8": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "d3d0ef75c788627f4218a8d08dccdf4ebc91f5b83c48d09ec8f2a3db9610014b", - "md5": "2dd61e8dfadc3754e35f927d4142fc0f", - "sha256": "fa98ec9cc9bf5d72a08ebf3654a9452e761fbb8566e3f80de199cbc15477e891" - }, - "downloads": -1, - "filename": "Django-2.2.8-py3-none-any.whl", - "has_sig": false, - "md5_digest": "2dd61e8dfadc3754e35f927d4142fc0f", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7460221, - "upload_time": "2019-12-02T08:57:52", - "upload_time_iso_8601": "2019-12-02T08:57:52.178782Z", - "url": "https://files.pythonhosted.org/packages/d3/d0/ef75c788627f4218a8d08dccdf4ebc91f5b83c48d09ec8f2a3db9610014b/Django-2.2.8-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "1caaf618f346b895123be44739b276099a2b418b45b2b7afb5e1071403e8d2e9", - "md5": "57d965818410a4e00e2267eef66aa9c9", - "sha256": "a4ad4f6f9c6a4b7af7e2deec8d0cbff28501852e5010d6c2dc695d3d1fae7ca0" - }, - "downloads": -1, - "filename": "Django-2.2.8.tar.gz", - "has_sig": false, - "md5_digest": "57d965818410a4e00e2267eef66aa9c9", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8870662, - "upload_time": "2019-12-02T08:58:05", - "upload_time_iso_8601": "2019-12-02T08:58:05.035772Z", - "url": "https://files.pythonhosted.org/packages/1c/aa/f618f346b895123be44739b276099a2b418b45b2b7afb5e1071403e8d2e9/Django-2.2.8.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2.9": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "cbc9ef1e25bdd092749dae74c95c2707dff892fde36e4053c4a2354b2303be10", - "md5": "2bdad7b5e9a0012f916b14f68df8084b", - "sha256": "687c37153486cf26c3fdcbdd177ef16de38dc3463f094b5f9c9955d91f277b14" - }, - "downloads": -1, - "filename": "Django-2.2.9-py3-none-any.whl", - "has_sig": false, - "md5_digest": "2bdad7b5e9a0012f916b14f68df8084b", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7460387, - "upload_time": "2019-12-18T08:59:07", - "upload_time_iso_8601": "2019-12-18T08:59:07.891032Z", - "url": "https://files.pythonhosted.org/packages/cb/c9/ef1e25bdd092749dae74c95c2707dff892fde36e4053c4a2354b2303be10/Django-2.2.9-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "2c0d2aa8e58c791d2aa65658fa26f2b035a9da13a6a34d1b2d991912c8a33729", - "md5": "a9a6555d166196e502b69715341f7ad4", - "sha256": "662a1ff78792e3fd77f16f71b1f31149489434de4b62a74895bd5d6534e635a5" - }, - "downloads": -1, - "filename": "Django-2.2.9.tar.gz", - "has_sig": false, - "md5_digest": "a9a6555d166196e502b69715341f7ad4", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 9006404, - "upload_time": "2019-12-18T08:59:27", - "upload_time_iso_8601": "2019-12-18T08:59:27.308468Z", - "url": "https://files.pythonhosted.org/packages/2c/0d/2aa8e58c791d2aa65658fa26f2b035a9da13a6a34d1b2d991912c8a33729/Django-2.2.9.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2a1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "b196020a0f9b83334080896e3ec47d0fb8936e9c948c6495ff45a8a67f1eefaa", - "md5": "7b88f3a420ed38af6df50945d46c6f87", - "sha256": "737bd3d5f70cb8bde3c660e69b077a0221b47caf499d4d3759fb086376002d4a" - }, - "downloads": -1, - "filename": "Django-2.2a1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "7b88f3a420ed38af6df50945d46c6f87", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7381403, - "upload_time": "2019-01-17T15:35:52", - "upload_time_iso_8601": "2019-01-17T15:35:52.375114Z", - "url": "https://files.pythonhosted.org/packages/b1/96/020a0f9b83334080896e3ec47d0fb8936e9c948c6495ff45a8a67f1eefaa/Django-2.2a1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "37a2bd8b1615c85d802c56ac137eb102976eba6bf844628e6228a6b8ecec77ff", - "md5": "e64af8f0f94cef419c3bb99b07aeb731", - "sha256": "146aa583364553c9ecbed55613a6c27bc23226048f86ed183350c80df3f2a844" - }, - "downloads": -1, - "filename": "Django-2.2a1.tar.gz", - "has_sig": false, - "md5_digest": "e64af8f0f94cef419c3bb99b07aeb731", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8777005, - "upload_time": "2019-01-17T15:35:59", - "upload_time_iso_8601": "2019-01-17T15:35:59.171725Z", - "url": "https://files.pythonhosted.org/packages/37/a2/bd8b1615c85d802c56ac137eb102976eba6bf844628e6228a6b8ecec77ff/Django-2.2a1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2b1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "f0eaaab3ef6a2f375a62c21d6e82d463ff2edd4f9b8f01a94fe85cd568d231e1", - "md5": "35d1b1faa579831cd59fb6a6280f416e", - "sha256": "58819ca72a13b963c16383687421261657abe5754aad9ad66166a921dd17559f" - }, - "downloads": -1, - "filename": "Django-2.2b1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "35d1b1faa579831cd59fb6a6280f416e", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7382001, - "upload_time": "2019-02-11T10:33:53", - "upload_time_iso_8601": "2019-02-11T10:33:53.482248Z", - "url": "https://files.pythonhosted.org/packages/f0/ea/aab3ef6a2f375a62c21d6e82d463ff2edd4f9b8f01a94fe85cd568d231e1/Django-2.2b1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "af47de47a1c047dd6ae180fff698f8d8ce785aabacaee5279efb4db3200dcca9", - "md5": "bbcc507ec3ad41daf3232f15c3f3ca47", - "sha256": "62644444551e8e6fd36600e741a4d24dd2b4b58acf7bae8847a8da952468d771" - }, - "downloads": -1, - "filename": "Django-2.2b1.tar.gz", - "has_sig": false, - "md5_digest": "bbcc507ec3ad41daf3232f15c3f3ca47", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8769147, - "upload_time": "2019-02-11T10:34:19", - "upload_time_iso_8601": "2019-02-11T10:34:19.912823Z", - "url": "https://files.pythonhosted.org/packages/af/47/de47a1c047dd6ae180fff698f8d8ce785aabacaee5279efb4db3200dcca9/Django-2.2b1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "2.2rc1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "820cabf6d3d97513de4188d7c2f92d4641fdfd12e8cf740781f09714db15c179", - "md5": "b5bc49235f960556558979abb3ec7edf", - "sha256": "6cdb98d464e5ffc9cbe3d3ca5ca74d61e69ce1784d5e15b85ceb19dee939f650" - }, - "downloads": -1, - "filename": "Django-2.2rc1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "b5bc49235f960556558979abb3ec7edf", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.5", - "size": 7382583, - "upload_time": "2019-03-18T08:57:29", - "upload_time_iso_8601": "2019-03-18T08:57:29.505366Z", - "url": "https://files.pythonhosted.org/packages/82/0c/abf6d3d97513de4188d7c2f92d4641fdfd12e8cf740781f09714db15c179/Django-2.2rc1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "a27f8b66660782ec1f226c601bfba8fcdf9363eb3daf104324e2a3008ea76592", - "md5": "76a60cab4301600ae5de574af19feac5", - "sha256": "b74420065c5a3f7bcb07c0811b1308239c05989e29963ff39f7b5e9c1cacc195" - }, - "downloads": -1, - "filename": "Django-2.2rc1.tar.gz", - "has_sig": false, - "md5_digest": "76a60cab4301600ae5de574af19feac5", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.5", - "size": 8787055, - "upload_time": "2019-03-18T08:57:35", - "upload_time_iso_8601": "2019-03-18T08:57:35.243662Z", - "url": "https://files.pythonhosted.org/packages/a2/7f/8b66660782ec1f226c601bfba8fcdf9363eb3daf104324e2a3008ea76592/Django-2.2rc1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.0": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "43d60aed0b12c66527748ce5a007da4618a65dfbe1f8fca82eccedf57d60295f", - "md5": "62020205feeac36093077863fd1fdd38", - "sha256": "6f857bd4e574442ba35a7172f1397b303167dae964cf18e53db5e85fe248d000" - }, - "downloads": -1, - "filename": "Django-3.0-py3-none-any.whl", - "has_sig": false, - "md5_digest": "62020205feeac36093077863fd1fdd38", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7427980, - "upload_time": "2019-12-02T11:13:11", - "upload_time_iso_8601": "2019-12-02T11:13:11.252980Z", - "url": "https://files.pythonhosted.org/packages/43/d6/0aed0b12c66527748ce5a007da4618a65dfbe1f8fca82eccedf57d60295f/Django-3.0-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "f846b3b8c61f867827fff2305db40659495dcd64fb35c399e75c53f23c113871", - "md5": "bd2aebfa7c1106755544f7f217d2acde", - "sha256": "d98c9b6e5eed147bc51f47c014ff6826bd1ab50b166956776ee13db5a58804ae" - }, - "downloads": -1, - "filename": "Django-3.0.tar.gz", - "has_sig": false, - "md5_digest": "bd2aebfa7c1106755544f7f217d2acde", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 8909597, - "upload_time": "2019-12-02T11:13:17", - "upload_time_iso_8601": "2019-12-02T11:13:17.709458Z", - "url": "https://files.pythonhosted.org/packages/f8/46/b3b8c61f867827fff2305db40659495dcd64fb35c399e75c53f23c113871/Django-3.0.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.0.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "6a2308f7fd7afdd24184a400fcaebf921bd09b5b5235cbd62ffa02308a7d35d6", - "md5": "1fdb13c754a81e5b478ce5eb5fb5bfda", - "sha256": "b61295749be7e1c42467c55bcabdaee9fbe9496fdf9ed2e22cef44d9de2ff953" - }, - "downloads": -1, - "filename": "Django-3.0.1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "1fdb13c754a81e5b478ce5eb5fb5bfda", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7428297, - "upload_time": "2019-12-18T08:59:13", - "upload_time_iso_8601": "2019-12-18T08:59:13.338777Z", - "url": "https://files.pythonhosted.org/packages/6a/23/08f7fd7afdd24184a400fcaebf921bd09b5b5235cbd62ffa02308a7d35d6/Django-3.0.1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "44e84ae9ef3d455f4ce5aa22259cb6e40c69b29ef6b02d49c5cdfa265f7fc821", - "md5": "12f434ed7ccd6ee57be6f05a45e20e97", - "sha256": "315b11ea265dd15348d47f2cbb044ef71da2018f6e582fed875c889758e6f844" - }, - "downloads": -1, - "filename": "Django-3.0.1.tar.gz", - "has_sig": false, - "md5_digest": "12f434ed7ccd6ee57be6f05a45e20e97", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9022787, - "upload_time": "2019-12-18T08:59:35", - "upload_time_iso_8601": "2019-12-18T08:59:35.114312Z", - "url": "https://files.pythonhosted.org/packages/44/e8/4ae9ef3d455f4ce5aa22259cb6e40c69b29ef6b02d49c5cdfa265f7fc821/Django-3.0.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.0.10": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "388642288adf01882693e00981524ee0c0a692c3358b81a9f835a0d7812b2190", - "md5": "8428ca321b1536662c53989b8fa85053", - "sha256": "313d0b8f96685e99327785cc600a5178ca855f8e6f4ed162e671e8c3cf749739" - }, - "downloads": -1, - "filename": "Django-3.0.10-py3-none-any.whl", - "has_sig": false, - "md5_digest": "8428ca321b1536662c53989b8fa85053", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7461438, - "upload_time": "2020-09-01T09:14:28", - "upload_time_iso_8601": "2020-09-01T09:14:28.893793Z", - "url": "https://files.pythonhosted.org/packages/38/86/42288adf01882693e00981524ee0c0a692c3358b81a9f835a0d7812b2190/Django-3.0.10-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "f409d7c995b128bec61233cfea0e5fa40e442cae54c127b4b2b0881e1fdd0023", - "md5": "deec48e8713727e443a7cee6b54baaeb", - "sha256": "2d14be521c3ae24960e5e83d4575e156a8c479a75c935224b671b1c6e66eddaf" - }, - "downloads": -1, - "filename": "Django-3.0.10.tar.gz", - "has_sig": false, - "md5_digest": "deec48e8713727e443a7cee6b54baaeb", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 8958332, - "upload_time": "2020-09-01T09:14:40", - "upload_time_iso_8601": "2020-09-01T09:14:40.538593Z", - "url": "https://files.pythonhosted.org/packages/f4/09/d7c995b128bec61233cfea0e5fa40e442cae54c127b4b2b0881e1fdd0023/Django-3.0.10.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.0.11": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "2065c3bec6de30da7184a3632670a3ca9db305859a4db17e8bda42dee56f3245", - "md5": "8452bf7ea9ffe582cb1c0219a1f572a6", - "sha256": "8c334df4160f7c89f6a8a359dd4e95c688ec5ac0db5db75fcc6fec8f590dc8cf" - }, - "downloads": -1, - "filename": "Django-3.0.11-py3-none-any.whl", - "has_sig": false, - "md5_digest": "8452bf7ea9ffe582cb1c0219a1f572a6", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7461465, - "upload_time": "2020-11-02T08:12:36", - "upload_time_iso_8601": "2020-11-02T08:12:36.684860Z", - "url": "https://files.pythonhosted.org/packages/20/65/c3bec6de30da7184a3632670a3ca9db305859a4db17e8bda42dee56f3245/Django-3.0.11-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "24942bfe789a7ff881e4730a1a3d20cb28adacdcec1e8415bf7874c312cb3a03", - "md5": "ea1b8817d3c936b07baa45bb26917bbf", - "sha256": "96436d3d2f744d26e193bfb5a1cff3e01b349f835bb0ea16f71743accf9c6fa9" - }, - "downloads": -1, - "filename": "Django-3.0.11.tar.gz", - "has_sig": false, - "md5_digest": "ea1b8817d3c936b07baa45bb26917bbf", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 8958879, - "upload_time": "2020-11-02T08:12:49", - "upload_time_iso_8601": "2020-11-02T08:12:49.355431Z", - "url": "https://files.pythonhosted.org/packages/24/94/2bfe789a7ff881e4730a1a3d20cb28adacdcec1e8415bf7874c312cb3a03/Django-3.0.11.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.0.12": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "bd42712421cdd7f39e8b26e32021df9cecb94533b1892b2668c941a98e89af3c", - "md5": "40fdd385d31b88edb5223c2e22b97950", - "sha256": "30c9ad3413805c0d4f2d619c2d06cfa66d070e812ca524562a80ba4291d1661f" - }, - "downloads": -1, - "filename": "Django-3.0.12-py3-none-any.whl", - "has_sig": false, - "md5_digest": "40fdd385d31b88edb5223c2e22b97950", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7461604, - "upload_time": "2021-02-01T09:28:19", - "upload_time_iso_8601": "2021-02-01T09:28:19.270471Z", - "url": "https://files.pythonhosted.org/packages/bd/42/712421cdd7f39e8b26e32021df9cecb94533b1892b2668c941a98e89af3c/Django-3.0.12-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "32e3e7e9a9378321fdfc3eb55de151911dce968fa245d1f16d8c480c63ea4ed1", - "md5": "55291777e25bd9e0a286c6f64751246a", - "sha256": "fd63e2c7acca5f2e7ad93dfb53d566e040d871404fc0f684a3e720006d221f9a" - }, - "downloads": -1, - "filename": "Django-3.0.12.tar.gz", - "has_sig": false, - "md5_digest": "55291777e25bd9e0a286c6f64751246a", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9255277, - "upload_time": "2021-02-01T09:28:35", - "upload_time_iso_8601": "2021-02-01T09:28:35.717844Z", - "url": "https://files.pythonhosted.org/packages/32/e3/e7e9a9378321fdfc3eb55de151911dce968fa245d1f16d8c480c63ea4ed1/Django-3.0.12.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.0.13": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "37b88f1a4cc2a427108784cbd9202a33f08d5ef7dd2985fa4a0e8ed7213c72f5", - "md5": "af81d35821046893dd00ea9b34ef392a", - "sha256": "2afe4900667bcceac792fa34b4fb25448c4fd950d8b32c5508b3442c4b10442a" - }, - "downloads": -1, - "filename": "Django-3.0.13-py3-none-any.whl", - "has_sig": false, - "md5_digest": "af81d35821046893dd00ea9b34ef392a", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7461602, - "upload_time": "2021-02-19T09:08:04", - "upload_time_iso_8601": "2021-02-19T09:08:04.026276Z", - "url": "https://files.pythonhosted.org/packages/37/b8/8f1a4cc2a427108784cbd9202a33f08d5ef7dd2985fa4a0e8ed7213c72f5/Django-3.0.13-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "3bfe11ec9b4cbae447e7b90d551be035d55c1293973592b491540334452f1f1f", - "md5": "7020810fb65b17e82d22001883b63a12", - "sha256": "6f13c3e8109236129c49d65a42fbf30c928e66b05ca6862246061b9343ecbaf2" - }, - "downloads": -1, - "filename": "Django-3.0.13.tar.gz", - "has_sig": false, - "md5_digest": "7020810fb65b17e82d22001883b63a12", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9285769, - "upload_time": "2021-02-19T09:08:22", - "upload_time_iso_8601": "2021-02-19T09:08:22.462879Z", - "url": "https://files.pythonhosted.org/packages/3b/fe/11ec9b4cbae447e7b90d551be035d55c1293973592b491540334452f1f1f/Django-3.0.13.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.0.14": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "343b393482eb369b3c50608b66feac8eabd4c5c15ae80468fb1eccc0b9c136a2", - "md5": "ac71977a20ca150f1fcd3f88e4a9f3d8", - "sha256": "9bc7aa619ed878fedba62ce139abe663a147dccfd20e907725ec11e02a1ca225" - }, - "downloads": -1, - "filename": "Django-3.0.14-py3-none-any.whl", - "has_sig": false, - "md5_digest": "ac71977a20ca150f1fcd3f88e4a9f3d8", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7461625, - "upload_time": "2021-04-06T07:34:56", - "upload_time_iso_8601": "2021-04-06T07:34:56.099064Z", - "url": "https://files.pythonhosted.org/packages/34/3b/393482eb369b3c50608b66feac8eabd4c5c15ae80468fb1eccc0b9c136a2/Django-3.0.14-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "760e5d847a77b7b42cacd01405b45e4e370124c1d8a15970865df5ab0f09f83a", - "md5": "f444fdd6ff8edec132991cbc343368d4", - "sha256": "d58d8394036db75a81896037d757357e79406e8f68816c3e8a28721c1d9d4c11" - }, - "downloads": -1, - "filename": "Django-3.0.14.tar.gz", - "has_sig": false, - "md5_digest": "f444fdd6ff8edec132991cbc343368d4", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9259569, - "upload_time": "2021-04-06T07:35:07", - "upload_time_iso_8601": "2021-04-06T07:35:07.952242Z", - "url": "https://files.pythonhosted.org/packages/76/0e/5d847a77b7b42cacd01405b45e4e370124c1d8a15970865df5ab0f09f83a/Django-3.0.14.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.0.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "55d18ade70e65fa157e1903fe4078305ca53b6819ab212d9fbbe5755afc8ea2e", - "md5": "219b8ac5c00c9e0f608bc5b748cd378e", - "sha256": "4f2c913303be4f874015993420bf0bd8fd2097a9c88e6b49c6a92f9bdd3fb13a" - }, - "downloads": -1, - "filename": "Django-3.0.2-py3-none-any.whl", - "has_sig": false, - "md5_digest": "219b8ac5c00c9e0f608bc5b748cd378e", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7428596, - "upload_time": "2020-01-02T07:22:14", - "upload_time_iso_8601": "2020-01-02T07:22:14.549127Z", - "url": "https://files.pythonhosted.org/packages/55/d1/8ade70e65fa157e1903fe4078305ca53b6819ab212d9fbbe5755afc8ea2e/Django-3.0.2-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "c5c15b901e21114b5dd9233726c2975c0aa7e9f48f63e41ec95d8777721d8aff", - "md5": "24d5364af6b04c4dd173111a3207459a", - "sha256": "8c3575f81e11390893860d97e1e0154c47512f180ea55bd84ce8fa69ba8051ca" - }, - "downloads": -1, - "filename": "Django-3.0.2.tar.gz", - "has_sig": false, - "md5_digest": "24d5364af6b04c4dd173111a3207459a", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9028261, - "upload_time": "2020-01-02T07:22:21", - "upload_time_iso_8601": "2020-01-02T07:22:21.750853Z", - "url": "https://files.pythonhosted.org/packages/c5/c1/5b901e21114b5dd9233726c2975c0aa7e9f48f63e41ec95d8777721d8aff/Django-3.0.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.0.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "c6b763d23df1e311ca0d90f41352a9efe7389ba353df95deea5676652e615420", - "md5": "023d4e362c2760e65f3c532bee48fa57", - "sha256": "c91c91a7ad6ef67a874a4f76f58ba534f9208412692a840e1d125eb5c279cb0a" - }, - "downloads": -1, - "filename": "Django-3.0.3-py3-none-any.whl", - "has_sig": false, - "md5_digest": "023d4e362c2760e65f3c532bee48fa57", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7457940, - "upload_time": "2020-02-03T09:50:46", - "upload_time_iso_8601": "2020-02-03T09:50:46.338782Z", - "url": "https://files.pythonhosted.org/packages/c6/b7/63d23df1e311ca0d90f41352a9efe7389ba353df95deea5676652e615420/Django-3.0.3-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "3d21316d435bf8bd6f355be6b5765da91394fb38f405e5bea6680e411e4d470c", - "md5": "37ec335a56234c0ad56c383b810afc7f", - "sha256": "2f1ba1db8648484dd5c238fb62504777b7ad090c81c5f1fd8d5eb5ec21b5f283" - }, - "downloads": -1, - "filename": "Django-3.0.3.tar.gz", - "has_sig": false, - "md5_digest": "37ec335a56234c0ad56c383b810afc7f", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 8932015, - "upload_time": "2020-02-03T09:51:04", - "upload_time_iso_8601": "2020-02-03T09:51:04.774633Z", - "url": "https://files.pythonhosted.org/packages/3d/21/316d435bf8bd6f355be6b5765da91394fb38f405e5bea6680e411e4d470c/Django-3.0.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.0.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "12688c125da33aaf0942add5095a7a2a8e064b3812d598e9fb5aca9957872d71", - "md5": "5102a4aed93ad570f44bf9cb60fe881a", - "sha256": "89e451bfbb815280b137e33e454ddd56481fdaa6334054e6e031041ee1eda360" - }, - "downloads": -1, - "filename": "Django-3.0.4-py3-none-any.whl", - "has_sig": false, - "md5_digest": "5102a4aed93ad570f44bf9cb60fe881a", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7458172, - "upload_time": "2020-03-04T09:31:56", - "upload_time_iso_8601": "2020-03-04T09:31:56.305550Z", - "url": "https://files.pythonhosted.org/packages/12/68/8c125da33aaf0942add5095a7a2a8e064b3812d598e9fb5aca9957872d71/Django-3.0.4-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "1d3889ea18b5aeb9b56fff7430388946e8e9dfd7a451f3e6ddb8a9b637f442c1", - "md5": "0b0299419770eaff86ff3a4af519cd6a", - "sha256": "50b781f6cbeb98f673aa76ed8e572a019a45e52bdd4ad09001072dfd91ab07c8" - }, - "downloads": -1, - "filename": "Django-3.0.4.tar.gz", - "has_sig": false, - "md5_digest": "0b0299419770eaff86ff3a4af519cd6a", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9060331, - "upload_time": "2020-03-04T09:32:15", - "upload_time_iso_8601": "2020-03-04T09:32:15.777733Z", - "url": "https://files.pythonhosted.org/packages/1d/38/89ea18b5aeb9b56fff7430388946e8e9dfd7a451f3e6ddb8a9b637f442c1/Django-3.0.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.0.5": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "a94f8a247eee2958529a6a805d38fbacd9764fd566462fa0016aa2a2947ab2a6", - "md5": "d9ef2b8c88dae4b65b789c6821c36114", - "sha256": "642d8eceab321ca743ae71e0f985ff8fdca59f07aab3a9fb362c617d23e33a76" - }, - "downloads": -1, - "filename": "Django-3.0.5-py3-none-any.whl", - "has_sig": false, - "md5_digest": "d9ef2b8c88dae4b65b789c6821c36114", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7458815, - "upload_time": "2020-04-01T07:59:15", - "upload_time_iso_8601": "2020-04-01T07:59:15.811931Z", - "url": "https://files.pythonhosted.org/packages/a9/4f/8a247eee2958529a6a805d38fbacd9764fd566462fa0016aa2a2947ab2a6/Django-3.0.5-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "ed521f281f39fe38d10c6c73e1c1d26a0aad5406be1108bf5f50423751ea8aa3", - "md5": "592912b4d708ef45e6cc85b44a24fcc2", - "sha256": "d4666c2edefa38c5ede0ec1655424c56dc47ceb04b6d8d62a7eac09db89545c1" - }, - "downloads": -1, - "filename": "Django-3.0.5.tar.gz", - "has_sig": false, - "md5_digest": "592912b4d708ef45e6cc85b44a24fcc2", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 8943850, - "upload_time": "2020-04-01T07:59:27", - "upload_time_iso_8601": "2020-04-01T07:59:27.770337Z", - "url": "https://files.pythonhosted.org/packages/ed/52/1f281f39fe38d10c6c73e1c1d26a0aad5406be1108bf5f50423751ea8aa3/Django-3.0.5.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.0.6": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "9d0404abb097c84c770180eeebe7ed920ce42f9917ab5ad4de01ff8ed11bc25b", - "md5": "b4a8610f563fc8f9a974eaa683cdbb3c", - "sha256": "051ba55d42daa3eeda3944a8e4df2bc96d4c62f94316dea217248a22563c3621" - }, - "downloads": -1, - "filename": "Django-3.0.6-py3-none-any.whl", - "has_sig": false, - "md5_digest": "b4a8610f563fc8f9a974eaa683cdbb3c", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7458833, - "upload_time": "2020-05-04T05:26:36", - "upload_time_iso_8601": "2020-05-04T05:26:36.090306Z", - "url": "https://files.pythonhosted.org/packages/9d/04/04abb097c84c770180eeebe7ed920ce42f9917ab5ad4de01ff8ed11bc25b/Django-3.0.6-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "1e9b75ef2ff5482f2970ba59093a67b9a8371626af8559caca7fcdb16ea4bd1c", - "md5": "48185be66d29f5552911ddfaad3e957c", - "sha256": "9aaa6a09678e1b8f0d98a948c56482eac3e3dd2ddbfb8de70a868135ef3b5e01" - }, - "downloads": -1, - "filename": "Django-3.0.6.tar.gz", - "has_sig": false, - "md5_digest": "48185be66d29f5552911ddfaad3e957c", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9070990, - "upload_time": "2020-05-04T05:26:41", - "upload_time_iso_8601": "2020-05-04T05:26:41.532361Z", - "url": "https://files.pythonhosted.org/packages/1e/9b/75ef2ff5482f2970ba59093a67b9a8371626af8559caca7fcdb16ea4bd1c/Django-3.0.6.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.0.7": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "5c636d7efecbf3f06db8c6577950a24a191e55cadf7cda4d7fe6976206c886dd", - "md5": "fbe615d79cebdd75bb057b729e6f1224", - "sha256": "e1630333248c9b3d4e38f02093a26f1e07b271ca896d73097457996e0fae12e8" - }, - "downloads": -1, - "filename": "Django-3.0.7-py3-none-any.whl", - "has_sig": false, - "md5_digest": "fbe615d79cebdd75bb057b729e6f1224", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7460889, - "upload_time": "2020-06-03T09:36:35", - "upload_time_iso_8601": "2020-06-03T09:36:35.471004Z", - "url": "https://files.pythonhosted.org/packages/5c/63/6d7efecbf3f06db8c6577950a24a191e55cadf7cda4d7fe6976206c886dd/Django-3.0.7-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "74ad8a1bc5e0f8b740792c99c7bef5ecc043018e2b605a2fe1e2513fde586b72", - "md5": "c3ac98d5503c671d316cf78ded3c9809", - "sha256": "5052b34b34b3425233c682e0e11d658fd6efd587d11335a0203d827224ada8f2" - }, - "downloads": -1, - "filename": "Django-3.0.7.tar.gz", - "has_sig": false, - "md5_digest": "c3ac98d5503c671d316cf78ded3c9809", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 8947502, - "upload_time": "2020-06-03T09:36:44", - "upload_time_iso_8601": "2020-06-03T09:36:44.234306Z", - "url": "https://files.pythonhosted.org/packages/74/ad/8a1bc5e0f8b740792c99c7bef5ecc043018e2b605a2fe1e2513fde586b72/Django-3.0.7.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.0.8": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "caab5e004afa025a6fb640c6e983d4983e6507421ff01be224da79ab7de7a21f", - "md5": "f7739227c05ae04ee7df444c92d6989a", - "sha256": "5457fc953ec560c5521b41fad9e6734a4668b7ba205832191bbdff40ec61073c" - }, - "downloads": -1, - "filename": "Django-3.0.8-py3-none-any.whl", - "has_sig": false, - "md5_digest": "f7739227c05ae04ee7df444c92d6989a", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7461178, - "upload_time": "2020-07-01T04:49:49", - "upload_time_iso_8601": "2020-07-01T04:49:49.522725Z", - "url": "https://files.pythonhosted.org/packages/ca/ab/5e004afa025a6fb640c6e983d4983e6507421ff01be224da79ab7de7a21f/Django-3.0.8-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "c6e019b529ca9c55fa0ee095edffa7135a8eff354490159d2d64d006928beb84", - "md5": "bc53ba68c6a322dc1f3c03f76eadbf41", - "sha256": "31a5fbbea5fc71c99e288ec0b2f00302a0a92c44b13ede80b73a6a4d6d205582" - }, - "downloads": -1, - "filename": "Django-3.0.8.tar.gz", - "has_sig": false, - "md5_digest": "bc53ba68c6a322dc1f3c03f76eadbf41", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9080731, - "upload_time": "2020-07-01T04:50:38", - "upload_time_iso_8601": "2020-07-01T04:50:38.186817Z", - "url": "https://files.pythonhosted.org/packages/c6/e0/19b529ca9c55fa0ee095edffa7135a8eff354490159d2d64d006928beb84/Django-3.0.8.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.0.9": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "87dc3f96196718beab06849f94f54ff20ffec78af8424c866c61cdced30fb1ce", - "md5": "1a95cfff1e7d419c16a94991de126cf1", - "sha256": "96fbe04e8ba0df289171e7f6970e0ff8b472bf4f909ed9e0e5beccbac7e1dbbe" - }, - "downloads": -1, - "filename": "Django-3.0.9-py3-none-any.whl", - "has_sig": false, - "md5_digest": "1a95cfff1e7d419c16a94991de126cf1", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7461267, - "upload_time": "2020-08-03T07:23:29", - "upload_time_iso_8601": "2020-08-03T07:23:29.043044Z", - "url": "https://files.pythonhosted.org/packages/87/dc/3f96196718beab06849f94f54ff20ffec78af8424c866c61cdced30fb1ce/Django-3.0.9-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "2d2519176dee332d7aabd0109b4909a296468f87fdc2b7ff2f41a3f8c571041e", - "md5": "b2129e2208419655709f2d22dbfa98cf", - "sha256": "c22b4cd8e388f8219dc121f091e53a8701f9f5bca9aa132b5254263cab516215" - }, - "downloads": -1, - "filename": "Django-3.0.9.tar.gz", - "has_sig": false, - "md5_digest": "b2129e2208419655709f2d22dbfa98cf", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9081099, - "upload_time": "2020-08-03T07:23:44", - "upload_time_iso_8601": "2020-08-03T07:23:44.286778Z", - "url": "https://files.pythonhosted.org/packages/2d/25/19176dee332d7aabd0109b4909a296468f87fdc2b7ff2f41a3f8c571041e/Django-3.0.9.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.0a1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "d8a602cce92ec00e4d0dc72f130a2066b5f30b16d3b27b8423cd75a6b591ebfd", - "md5": "173db82e0fb48a083f0cdf79df71cab1", - "sha256": "ccf5d58cfe53601afcf38d53cb5a3ed6a915afcfdb728593ce8594f96ca21cd8" - }, - "downloads": -1, - "filename": "Django-3.0a1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "173db82e0fb48a083f0cdf79df71cab1", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7499892, - "upload_time": "2019-09-10T09:19:32", - "upload_time_iso_8601": "2019-09-10T09:19:32.778684Z", - "url": "https://files.pythonhosted.org/packages/d8/a6/02cce92ec00e4d0dc72f130a2066b5f30b16d3b27b8423cd75a6b591ebfd/Django-3.0a1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "b6c0ffc5c11eccfd5111491595753981a259bd709994d83648aebcfcb288131a", - "md5": "585129261f754989a833aa2f1f1da5ad", - "sha256": "ccfaf8c7e72cecbf97de52633b30995659a6ad14a0e4a8c4ebc78b353e8ed78d" - }, - "downloads": -1, - "filename": "Django-3.0a1.tar.gz", - "has_sig": false, - "md5_digest": "585129261f754989a833aa2f1f1da5ad", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 8936672, - "upload_time": "2019-09-10T09:19:39", - "upload_time_iso_8601": "2019-09-10T09:19:39.974331Z", - "url": "https://files.pythonhosted.org/packages/b6/c0/ffc5c11eccfd5111491595753981a259bd709994d83648aebcfcb288131a/Django-3.0a1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.0b1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "d1149e339e2fa2befa6ea629cb5cbc2c84a201b03ac04ad15dc1f196c8675562", - "md5": "510cba559c1f897c651b4e4206125fd9", - "sha256": "3c43d11a9d04a860b4ab658f6cbee11a33a87755667b0a9f42e0fc327d015d39" - }, - "downloads": -1, - "filename": "Django-3.0b1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "510cba559c1f897c651b4e4206125fd9", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7500629, - "upload_time": "2019-10-14T10:21:37", - "upload_time_iso_8601": "2019-10-14T10:21:37.792635Z", - "url": "https://files.pythonhosted.org/packages/d1/14/9e339e2fa2befa6ea629cb5cbc2c84a201b03ac04ad15dc1f196c8675562/Django-3.0b1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "3ab153c75411a3e2e82429ab1506d7565973e193b92eb030eab847971a2b46e2", - "md5": "1a51f1ebc60d3bdfc8b70ecd4d399e40", - "sha256": "e0336c2bc6bb99b87788d79e4c5dd88db255bcd097240e0f7f9843e222e43672" - }, - "downloads": -1, - "filename": "Django-3.0b1.tar.gz", - "has_sig": false, - "md5_digest": "1a51f1ebc60d3bdfc8b70ecd4d399e40", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9076737, - "upload_time": "2019-10-14T10:21:46", - "upload_time_iso_8601": "2019-10-14T10:21:46.008282Z", - "url": "https://files.pythonhosted.org/packages/3a/b1/53c75411a3e2e82429ab1506d7565973e193b92eb030eab847971a2b46e2/Django-3.0b1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.0rc1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "6265957dc58af5094fc3c2d5d18233866a01eefbf71b1ad17741da098fc774d1", - "md5": "8cb9d8481ca92b0111e224574fa7daca", - "sha256": "bd7e7dd0c3065e356f150fb03d1620e98e947d8c0ce913826d43bd468e020bed" - }, - "downloads": -1, - "filename": "Django-3.0rc1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "8cb9d8481ca92b0111e224574fa7daca", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7501670, - "upload_time": "2019-11-18T08:51:12", - "upload_time_iso_8601": "2019-11-18T08:51:12.063347Z", - "url": "https://files.pythonhosted.org/packages/62/65/957dc58af5094fc3c2d5d18233866a01eefbf71b1ad17741da098fc774d1/Django-3.0rc1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "38bd9324965c7a900fba18db907eada900df754b448ce4b76972cbc8a55ae7e2", - "md5": "697ef507f57b5140fb112fc653a27af8", - "sha256": "0a1efde1b685a6c30999ba00902f23613cf5db864c5a1532d2edf3eda7896a37" - }, - "downloads": -1, - "filename": "Django-3.0rc1.tar.gz", - "has_sig": false, - "md5_digest": "697ef507f57b5140fb112fc653a27af8", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9083478, - "upload_time": "2019-11-18T08:51:19", - "upload_time_iso_8601": "2019-11-18T08:51:19.037478Z", - "url": "https://files.pythonhosted.org/packages/38/bd/9324965c7a900fba18db907eada900df754b448ce4b76972cbc8a55ae7e2/Django-3.0rc1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "2b5a4bd5624546912082a1bd2709d0edc0685f5c7827a278d806a20cf6adea28", - "md5": "281c2e919cb60fd09a64fd068cf152fb", - "sha256": "1a63f5bb6ff4d7c42f62a519edc2adbb37f9b78068a5a862beff858b68e3dc8b" - }, - "downloads": -1, - "filename": "Django-3.1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "281c2e919cb60fd09a64fd068cf152fb", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7835282, - "upload_time": "2020-08-04T08:07:00", - "upload_time_iso_8601": "2020-08-04T08:07:00.217377Z", - "url": "https://files.pythonhosted.org/packages/2b/5a/4bd5624546912082a1bd2709d0edc0685f5c7827a278d806a20cf6adea28/Django-3.1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "ccbfb9f5e4c4707bcabcd4202d78a4d23459d5de3083a7e7efde4dd215b997ac", - "md5": "2001ba40467d61a2b90570a68c657e35", - "sha256": "2d390268a13c655c97e0e2ede9d117007996db692c1bb93eabebd4fb7ea7012b" - }, - "downloads": -1, - "filename": "Django-3.1.tar.gz", - "has_sig": false, - "md5_digest": "2001ba40467d61a2b90570a68c657e35", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9382872, - "upload_time": "2020-08-04T08:07:06", - "upload_time_iso_8601": "2020-08-04T08:07:06.894968Z", - "url": "https://files.pythonhosted.org/packages/cc/bf/b9f5e4c4707bcabcd4202d78a4d23459d5de3083a7e7efde4dd215b997ac/Django-3.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.1.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "01a5fb3dad18422fcd4241d18460a1fe17542bfdeadcf74e3861d1a2dfc9e459", - "md5": "f4eb53dd67fc64f9b62514fb21a95949", - "sha256": "b5fbb818e751f660fa2d576d9f40c34a4c615c8b48dd383f5216e609f383371f" - }, - "downloads": -1, - "filename": "Django-3.1.1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "f4eb53dd67fc64f9b62514fb21a95949", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7835662, - "upload_time": "2020-09-01T09:14:32", - "upload_time_iso_8601": "2020-09-01T09:14:32.579007Z", - "url": "https://files.pythonhosted.org/packages/01/a5/fb3dad18422fcd4241d18460a1fe17542bfdeadcf74e3861d1a2dfc9e459/Django-3.1.1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "bb156df0530f86ffc1efdadd98b749a5d330fa719a07a6c839eedbccd74e27a2", - "md5": "d5e894fb3c46064e84e9dc68a08a46d0", - "sha256": "59c8125ca873ed3bdae9c12b146fbbd6ed8d0f743e4cf5f5817af50c51f1fc2f" - }, - "downloads": -1, - "filename": "Django-3.1.1.tar.gz", - "has_sig": false, - "md5_digest": "d5e894fb3c46064e84e9dc68a08a46d0", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9250616, - "upload_time": "2020-09-01T09:14:45", - "upload_time_iso_8601": "2020-09-01T09:14:45.550781Z", - "url": "https://files.pythonhosted.org/packages/bb/15/6df0530f86ffc1efdadd98b749a5d330fa719a07a6c839eedbccd74e27a2/Django-3.1.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.1.10": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "6f7877e58e30f8e8528a244663ea8551f5a741c8b9fcbc6993935bd802c676d1", - "md5": "b14b26d6fb2f04318639b1ea296e32ac", - "sha256": "973c968e63518859732f018975364785dd96f0581b1e4b12e2a4b749415ac43a" - }, - "downloads": -1, - "filename": "Django-3.1.10-py3-none-any.whl", - "has_sig": false, - "md5_digest": "b14b26d6fb2f04318639b1ea296e32ac", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7834910, - "upload_time": "2021-05-06T07:40:15", - "upload_time_iso_8601": "2021-05-06T07:40:15.707259Z", - "url": "https://files.pythonhosted.org/packages/6f/78/77e58e30f8e8528a244663ea8551f5a741c8b9fcbc6993935bd802c676d1/Django-3.1.10-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "d7fbc5b08041e02e256edc36f75a79f86d852504cee9b428a572236615899688", - "md5": "14e7761f37ca390a6bd2efa45b84ec02", - "sha256": "cd6ec37db950a384dba3341b135394fdc776ede4d149fc7abde1e45a21ec4f22" - }, - "downloads": -1, - "filename": "Django-3.1.10.tar.gz", - "has_sig": false, - "md5_digest": "14e7761f37ca390a6bd2efa45b84ec02", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9654073, - "upload_time": "2021-05-06T07:40:18", - "upload_time_iso_8601": "2021-05-06T07:40:18.974792Z", - "url": "https://files.pythonhosted.org/packages/d7/fb/c5b08041e02e256edc36f75a79f86d852504cee9b428a572236615899688/Django-3.1.10.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.1.11": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "587d0bedb56dfe0e104b212d04f66176f7d2b315d1f929ce86899b4c5c16e467", - "md5": "c9f9c13acf08c2a06270974399175cf0", - "sha256": "c79245c488411d1ae300b8f7a08ac18a496380204cf3035aff97ad917a8de999" - }, - "downloads": -1, - "filename": "Django-3.1.11-py3-none-any.whl", - "has_sig": false, - "md5_digest": "c9f9c13acf08c2a06270974399175cf0", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7835125, - "upload_time": "2021-05-13T07:36:44", - "upload_time_iso_8601": "2021-05-13T07:36:44.240638Z", - "url": "https://files.pythonhosted.org/packages/58/7d/0bedb56dfe0e104b212d04f66176f7d2b315d1f929ce86899b4c5c16e467/Django-3.1.11-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "35ca4f17af043f43a2a969bdd029c1fa97d902a03f9913bb259912de080942b4", - "md5": "7aa79a1d643a624aed09172de1a8c17b", - "sha256": "9a0a2f3d34c53032578b54db7ec55929b87dda6fec27a06cc2587afbea1965e5" - }, - "downloads": -1, - "filename": "Django-3.1.11.tar.gz", - "has_sig": false, - "md5_digest": "7aa79a1d643a624aed09172de1a8c17b", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9654094, - "upload_time": "2021-05-13T07:36:56", - "upload_time_iso_8601": "2021-05-13T07:36:56.726950Z", - "url": "https://files.pythonhosted.org/packages/35/ca/4f17af043f43a2a969bdd029c1fa97d902a03f9913bb259912de080942b4/Django-3.1.11.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.1.12": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "4e344fe787d20c5997a5ee246903791fd8ec2c4672f37ee7cbed1d75f7cf4065", - "md5": "bfe1e1adbc8d3ffc68a2bccdfafe2d07", - "sha256": "a523d62b7ab2908f551dabc32b99017a86aa7784e32b761708e52be3dce6d35d" - }, - "downloads": -1, - "filename": "Django-3.1.12-py3-none-any.whl", - "has_sig": false, - "md5_digest": "bfe1e1adbc8d3ffc68a2bccdfafe2d07", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7835286, - "upload_time": "2021-06-02T08:53:50", - "upload_time_iso_8601": "2021-06-02T08:53:50.036333Z", - "url": "https://files.pythonhosted.org/packages/4e/34/4fe787d20c5997a5ee246903791fd8ec2c4672f37ee7cbed1d75f7cf4065/Django-3.1.12-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "4f1089d18af7fe301f3fc18e3553d1c6b9b99d1c88d46c65fdd44aaae66c91c5", - "md5": "51c1e49e93f198aa6d76e29cbb5e5d1d", - "sha256": "dc41bf07357f1f4810c1c555b685cb51f780b41e37892d6cc92b89789f2847e1" - }, - "downloads": -1, - "filename": "Django-3.1.12.tar.gz", - "has_sig": false, - "md5_digest": "51c1e49e93f198aa6d76e29cbb5e5d1d", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9676259, - "upload_time": "2021-06-02T08:54:32", - "upload_time_iso_8601": "2021-06-02T08:54:32.877136Z", - "url": "https://files.pythonhosted.org/packages/4f/10/89d18af7fe301f3fc18e3553d1c6b9b99d1c88d46c65fdd44aaae66c91c5/Django-3.1.12.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.1.13": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "ba526d9ff9141e2eca84c94e4939c2a7666096ee7c64ce57d8bc65aa0c95fa7a", - "md5": "04296a18dd76b564c3c5f1af57ebfeed", - "sha256": "a6e0d1ff11095b7394c079ade7094c73b2dc3df4a7a373c9b58ed73b77a97feb" - }, - "downloads": -1, - "filename": "Django-3.1.13-py3-none-any.whl", - "has_sig": false, - "md5_digest": "04296a18dd76b564c3c5f1af57ebfeed", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7835379, - "upload_time": "2021-07-01T07:39:56", - "upload_time_iso_8601": "2021-07-01T07:39:56.137639Z", - "url": "https://files.pythonhosted.org/packages/ba/52/6d9ff9141e2eca84c94e4939c2a7666096ee7c64ce57d8bc65aa0c95fa7a/Django-3.1.13-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "0b0de9b2e0af13dc6a3694f7bf2518e6642db08ccde83a726da25ddd2da2608e", - "md5": "da97ac7e5ebba4681b1aedeed040cac7", - "sha256": "9f8be75646f62204320b195062b1d696ba28aa3d45ee72fb7c888ffaebc5bdb2" - }, - "downloads": -1, - "filename": "Django-3.1.13.tar.gz", - "has_sig": false, - "md5_digest": "da97ac7e5ebba4681b1aedeed040cac7", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9656683, - "upload_time": "2021-07-01T07:40:04", - "upload_time_iso_8601": "2021-07-01T07:40:04.542907Z", - "url": "https://files.pythonhosted.org/packages/0b/0d/e9b2e0af13dc6a3694f7bf2518e6642db08ccde83a726da25ddd2da2608e/Django-3.1.13.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.1.14": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "6b2eb031dc0789840ebc108d4654c134e86a9f6593ce9c58ce39328c142b24b8", - "md5": "f312d72a1c66385fdf3eba1da2935348", - "sha256": "0fabc786489af16ad87a8c170ba9d42bfd23f7b699bd5ef05675864e8d012859" - }, - "downloads": -1, - "filename": "Django-3.1.14-py3-none-any.whl", - "has_sig": false, - "md5_digest": "f312d72a1c66385fdf3eba1da2935348", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7835424, - "upload_time": "2021-12-07T07:34:46", - "upload_time_iso_8601": "2021-12-07T07:34:46.584704Z", - "url": "https://files.pythonhosted.org/packages/6b/2e/b031dc0789840ebc108d4654c134e86a9f6593ce9c58ce39328c142b24b8/Django-3.1.14-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "7a59b774fbaf743e675e9a808406a39814effb723c7c0b1d4a5b3b2e794ce077", - "md5": "f9be4bfe4a4bf3609d0905cfa88d8b02", - "sha256": "72a4a5a136a214c39cf016ccdd6b69e2aa08c7479c66d93f3a9b5e4bb9d8a347" - }, - "downloads": -1, - "filename": "Django-3.1.14.tar.gz", - "has_sig": false, - "md5_digest": "f9be4bfe4a4bf3609d0905cfa88d8b02", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9659386, - "upload_time": "2021-12-07T07:35:00", - "upload_time_iso_8601": "2021-12-07T07:35:00.760884Z", - "url": "https://files.pythonhosted.org/packages/7a/59/b774fbaf743e675e9a808406a39814effb723c7c0b1d4a5b3b2e794ce077/Django-3.1.14.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.1.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "50224c91847beceadbb54b5a518909ed5000bb1777168c7d6b087e8f79e5e05b", - "md5": "92cbac59a94e96a6f732c0ba493e7bc3", - "sha256": "c93c28ccf1d094cbd00d860e83128a39e45d2c571d3b54361713aaaf9a94cac4" - }, - "downloads": -1, - "filename": "Django-3.1.2-py3-none-any.whl", - "has_sig": false, - "md5_digest": "92cbac59a94e96a6f732c0ba493e7bc3", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7833339, - "upload_time": "2020-10-01T05:38:28", - "upload_time_iso_8601": "2020-10-01T05:38:28.704502Z", - "url": "https://files.pythonhosted.org/packages/50/22/4c91847beceadbb54b5a518909ed5000bb1777168c7d6b087e8f79e5e05b/Django-3.1.2-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "93565c79283e5addc81931788bf9ca96063c01bf8c5e34c0d160f4e102a4d9d9", - "md5": "5fd4b5bd4f474f59fbd70137f4a053ed", - "sha256": "a2127ad0150ec6966655bedf15dbbff9697cc86d61653db2da1afa506c0b04cc" - }, - "downloads": -1, - "filename": "Django-3.1.2.tar.gz", - "has_sig": false, - "md5_digest": "5fd4b5bd4f474f59fbd70137f4a053ed", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9387482, - "upload_time": "2020-10-01T05:38:34", - "upload_time_iso_8601": "2020-10-01T05:38:34.306154Z", - "url": "https://files.pythonhosted.org/packages/93/56/5c79283e5addc81931788bf9ca96063c01bf8c5e34c0d160f4e102a4d9d9/Django-3.1.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.1.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "7f1716267e782a30ea2ce08a9a452c1db285afb0ff226cfe3753f484d3d65662", - "md5": "3a220982452cde94fb4a544fb7cc6e63", - "sha256": "14a4b7cd77297fba516fc0d92444cc2e2e388aa9de32d7a68d4a83d58f5a4927" - }, - "downloads": -1, - "filename": "Django-3.1.3-py3-none-any.whl", - "has_sig": false, - "md5_digest": "3a220982452cde94fb4a544fb7cc6e63", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7833691, - "upload_time": "2020-11-02T08:12:40", - "upload_time_iso_8601": "2020-11-02T08:12:40.297190Z", - "url": "https://files.pythonhosted.org/packages/7f/17/16267e782a30ea2ce08a9a452c1db285afb0ff226cfe3753f484d3d65662/Django-3.1.3-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "98d0d55c4beb0fa8a48031777ee16c8d94c5ec2e66d43163a43b88f4d3e7a520", - "md5": "959f489ab9163691f4f90a41448573ba", - "sha256": "14b87775ffedab2ef6299b73343d1b4b41e5d4e2aa58c6581f114dbec01e3f8f" - }, - "downloads": -1, - "filename": "Django-3.1.3.tar.gz", - "has_sig": false, - "md5_digest": "959f489ab9163691f4f90a41448573ba", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9253273, - "upload_time": "2020-11-02T08:12:54", - "upload_time_iso_8601": "2020-11-02T08:12:54.518784Z", - "url": "https://files.pythonhosted.org/packages/98/d0/d55c4beb0fa8a48031777ee16c8d94c5ec2e66d43163a43b88f4d3e7a520/Django-3.1.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.1.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "08c77ce40e5a5cb47ede081b9fa8a3dd93d101c884882ae34927967b0792f5fb", - "md5": "1478c3a6d6d821b6ab6abdccfbf0217d", - "sha256": "5c866205f15e7a7123f1eec6ab939d22d5bde1416635cab259684af66d8e48a2" - }, - "downloads": -1, - "filename": "Django-3.1.4-py3-none-any.whl", - "has_sig": false, - "md5_digest": "1478c3a6d6d821b6ab6abdccfbf0217d", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7833976, - "upload_time": "2020-12-01T06:03:32", - "upload_time_iso_8601": "2020-12-01T06:03:32.688772Z", - "url": "https://files.pythonhosted.org/packages/08/c7/7ce40e5a5cb47ede081b9fa8a3dd93d101c884882ae34927967b0792f5fb/Django-3.1.4-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "19c9d51f1ddf9afc15f0567e7493e741b0ba283d5240ef1d2f1a1a76e0839ed3", - "md5": "6c8d45beab6c8e41b112db094c608a4c", - "sha256": "edb10b5c45e7e9c0fb1dc00b76ec7449aca258a39ffd613dbd078c51d19c9f03" - }, - "downloads": -1, - "filename": "Django-3.1.4.tar.gz", - "has_sig": false, - "md5_digest": "6c8d45beab6c8e41b112db094c608a4c", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9392125, - "upload_time": "2020-12-01T06:03:38", - "upload_time_iso_8601": "2020-12-01T06:03:38.760397Z", - "url": "https://files.pythonhosted.org/packages/19/c9/d51f1ddf9afc15f0567e7493e741b0ba283d5240ef1d2f1a1a76e0839ed3/Django-3.1.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.1.5": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "b28fd27f35f0639103271231bc81a96ad9188e6b5bc878e140ccb0dc610ccef0", - "md5": "e0ddf20ecf1400b831ecbe721769c3eb", - "sha256": "efa2ab96b33b20c2182db93147a0c3cd7769d418926f9e9f140a60dca7c64ca9" - }, - "downloads": -1, - "filename": "Django-3.1.5-py3-none-any.whl", - "has_sig": false, - "md5_digest": "e0ddf20ecf1400b831ecbe721769c3eb", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7834051, - "upload_time": "2021-01-04T07:54:31", - "upload_time_iso_8601": "2021-01-04T07:54:31.195043Z", - "url": "https://files.pythonhosted.org/packages/b2/8f/d27f35f0639103271231bc81a96ad9188e6b5bc878e140ccb0dc610ccef0/Django-3.1.5-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "4a6fe6ae2d599028c63f3a6a6454b86cacb0106070ffb6d06e09142c0e6ff31d", - "md5": "b1f925747cc6867ead474de68212a42b", - "sha256": "2d78425ba74c7a1a74b196058b261b9733a8570782f4e2828974777ccca7edf7" - }, - "downloads": -1, - "filename": "Django-3.1.5.tar.gz", - "has_sig": false, - "md5_digest": "b1f925747cc6867ead474de68212a42b", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9257571, - "upload_time": "2021-01-04T07:54:40", - "upload_time_iso_8601": "2021-01-04T07:54:40.039501Z", - "url": "https://files.pythonhosted.org/packages/4a/6f/e6ae2d599028c63f3a6a6454b86cacb0106070ffb6d06e09142c0e6ff31d/Django-3.1.5.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.1.6": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "7542f59a8ebf14be6d17438f13042c775f53d3dfa71fff973e4aef64ca89582c", - "md5": "6ea90e9a768405bcfd599028a2c5eadf", - "sha256": "169e2e7b4839a7910b393eec127fd7cbae62e80fa55f89c6510426abf673fe5f" - }, - "downloads": -1, - "filename": "Django-3.1.6-py3-none-any.whl", - "has_sig": false, - "md5_digest": "6ea90e9a768405bcfd599028a2c5eadf", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7834188, - "upload_time": "2021-02-01T09:28:24", - "upload_time_iso_8601": "2021-02-01T09:28:24.161399Z", - "url": "https://files.pythonhosted.org/packages/75/42/f59a8ebf14be6d17438f13042c775f53d3dfa71fff973e4aef64ca89582c/Django-3.1.6-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "3b846ed065944dd7ab438e1d19d7bb1b1911b27629946a5e0f8c007a692cc1a7", - "md5": "989ab24e42328bceefebcc2ad8b00db9", - "sha256": "c6c0462b8b361f8691171af1fb87eceb4442da28477e12200c40420176206ba7" - }, - "downloads": -1, - "filename": "Django-3.1.6.tar.gz", - "has_sig": false, - "md5_digest": "989ab24e42328bceefebcc2ad8b00db9", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9645871, - "upload_time": "2021-02-01T09:28:42", - "upload_time_iso_8601": "2021-02-01T09:28:42.846948Z", - "url": "https://files.pythonhosted.org/packages/3b/84/6ed065944dd7ab438e1d19d7bb1b1911b27629946a5e0f8c007a692cc1a7/Django-3.1.6.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.1.7": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "b86f9a4415cc4fe9228e26ea53cf2005961799b2abb8da0411e519fdb74754fa", - "md5": "28fea6fbd0ab6cb4e0c861c7b192daeb", - "sha256": "baf099db36ad31f970775d0be5587cc58a6256a6771a44eb795b554d45f211b8" - }, - "downloads": -1, - "filename": "Django-3.1.7-py3-none-any.whl", - "has_sig": false, - "md5_digest": "28fea6fbd0ab6cb4e0c861c7b192daeb", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7834206, - "upload_time": "2021-02-19T09:08:09", - "upload_time_iso_8601": "2021-02-19T09:08:09.213984Z", - "url": "https://files.pythonhosted.org/packages/b8/6f/9a4415cc4fe9228e26ea53cf2005961799b2abb8da0411e519fdb74754fa/Django-3.1.7-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "c2b85fcb2fdfe015fbea62c728255b09887dd71ba09e90cfe996461f3bf63cfe", - "md5": "80c868540848598b73bb12f43accc2b5", - "sha256": "32ce792ee9b6a0cbbec340123e229ac9f765dff8c2a4ae9247a14b2ba3a365a7" - }, - "downloads": -1, - "filename": "Django-3.1.7.tar.gz", - "has_sig": false, - "md5_digest": "80c868540848598b73bb12f43accc2b5", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9673009, - "upload_time": "2021-02-19T09:08:29", - "upload_time_iso_8601": "2021-02-19T09:08:29.394194Z", - "url": "https://files.pythonhosted.org/packages/c2/b8/5fcb2fdfe015fbea62c728255b09887dd71ba09e90cfe996461f3bf63cfe/Django-3.1.7.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.1.8": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "6790f5dbec986deb4706e2baa957aaf9d547fe80d1f8f9a80f2878d22204b76e", - "md5": "a0009705bf1c09a21194209bf9ac1348", - "sha256": "c348b3ddc452bf4b62361f0752f71a339140c777ebea3cdaaaa8fdb7f417a862" - }, - "downloads": -1, - "filename": "Django-3.1.8-py3-none-any.whl", - "has_sig": false, - "md5_digest": "a0009705bf1c09a21194209bf9ac1348", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7834224, - "upload_time": "2021-04-06T07:34:59", - "upload_time_iso_8601": "2021-04-06T07:34:59.785006Z", - "url": "https://files.pythonhosted.org/packages/67/90/f5dbec986deb4706e2baa957aaf9d547fe80d1f8f9a80f2878d22204b76e/Django-3.1.8-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "a1e6e68c5acd0a74c6daccfd109c3dc5c8003afaaa669db7eacd15863b0035b5", - "md5": "70775d8139fe1f04585214328d0a9dd5", - "sha256": "f8393103e15ec2d2d313ccbb95a3f1da092f9f58d74ac1c61ca2ac0436ae1eac" - }, - "downloads": -1, - "filename": "Django-3.1.8.tar.gz", - "has_sig": false, - "md5_digest": "70775d8139fe1f04585214328d0a9dd5", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9651582, - "upload_time": "2021-04-06T07:35:13", - "upload_time_iso_8601": "2021-04-06T07:35:13.000931Z", - "url": "https://files.pythonhosted.org/packages/a1/e6/e68c5acd0a74c6daccfd109c3dc5c8003afaaa669db7eacd15863b0035b5/Django-3.1.8.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.1.9": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "7a4ec6ec8e544eea36804c8795078df33aaca4932b7257188bd72124287372c4", - "md5": "aaa882ce4f6cbe47c55c908bae15d873", - "sha256": "f5c3f1ce2473abf2f35d992c46d5534e53bd3397867689bde28fd29e856b46b7" - }, - "downloads": -1, - "filename": "Django-3.1.9-py3-none-any.whl", - "has_sig": false, - "md5_digest": "aaa882ce4f6cbe47c55c908bae15d873", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7834850, - "upload_time": "2021-05-04T08:47:31", - "upload_time_iso_8601": "2021-05-04T08:47:31.247471Z", - "url": "https://files.pythonhosted.org/packages/7a/4e/c6ec8e544eea36804c8795078df33aaca4932b7257188bd72124287372c4/Django-3.1.9-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "0b80e5d65132f062e48094b274b6ec18715262c3c609acc49747f284509851fd", - "md5": "1d5d47ccc3468f653247526252346301", - "sha256": "296d10092561a1cf16100885b935dc56cf09c8ccb114e64004554941677c6342" - }, - "downloads": -1, - "filename": "Django-3.1.9.tar.gz", - "has_sig": false, - "md5_digest": "1d5d47ccc3468f653247526252346301", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9673018, - "upload_time": "2021-05-04T08:48:12", - "upload_time_iso_8601": "2021-05-04T08:48:12.197387Z", - "url": "https://files.pythonhosted.org/packages/0b/80/e5d65132f062e48094b274b6ec18715262c3c609acc49747f284509851fd/Django-3.1.9.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.1a1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "d864fe1af1b31d7c0f5aa7c5ccaf8d263c63652002853fa323cd80b5410788bc", - "md5": "8da21aabff06bd967f345fa84515897c", - "sha256": "db4c9b29615d17f808f2b1914d5cd73cd457c9fd90581195172c0888c210d944" - }, - "downloads": -1, - "filename": "Django-3.1a1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "8da21aabff06bd967f345fa84515897c", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7520707, - "upload_time": "2020-05-14T09:41:05", - "upload_time_iso_8601": "2020-05-14T09:41:05.996512Z", - "url": "https://files.pythonhosted.org/packages/d8/64/fe1af1b31d7c0f5aa7c5ccaf8d263c63652002853fa323cd80b5410788bc/Django-3.1a1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "05f5ccf81420fe3b3231a61c2fae1ab905bf2e66ba3cab9e4f5d982906bcf709", - "md5": "e4583f6a9981390b92dc5d26d28ded5e", - "sha256": "dd96f98ec1c3e60877d45cea7350215f16de409848d23cced8443db1b188bd9b" - }, - "downloads": -1, - "filename": "Django-3.1a1.tar.gz", - "has_sig": false, - "md5_digest": "e4583f6a9981390b92dc5d26d28ded5e", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9164027, - "upload_time": "2020-05-14T09:41:11", - "upload_time_iso_8601": "2020-05-14T09:41:11.988119Z", - "url": "https://files.pythonhosted.org/packages/05/f5/ccf81420fe3b3231a61c2fae1ab905bf2e66ba3cab9e4f5d982906bcf709/Django-3.1a1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.1b1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "0cbd49bbafb42531f77c483fc725863eb1ce76a180dd2771cef354378b191cec", - "md5": "16d9668b8eeaba7621930f5920ba29d7", - "sha256": "ccf6c208424c0e1b0eaffd36efe12618a9ab4d0037e26f6ffceaa5277af985d7" - }, - "downloads": -1, - "filename": "Django-3.1b1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "16d9668b8eeaba7621930f5920ba29d7", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7523548, - "upload_time": "2020-06-15T08:15:27", - "upload_time_iso_8601": "2020-06-15T08:15:27.667675Z", - "url": "https://files.pythonhosted.org/packages/0c/bd/49bbafb42531f77c483fc725863eb1ce76a180dd2771cef354378b191cec/Django-3.1b1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "ac92261266a01e8ff03c71973a3c8ab1604dc3147b8ef35363697cbbe5c2563a", - "md5": "0ce3271b1a12abb8aab48823e54334b6", - "sha256": "045be31d68dfed684831e39ab1d9e77a595f1a393935cb43b6c5451d2e78c8a4" - }, - "downloads": -1, - "filename": "Django-3.1b1.tar.gz", - "has_sig": false, - "md5_digest": "0ce3271b1a12abb8aab48823e54334b6", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9171734, - "upload_time": "2020-06-15T08:15:33", - "upload_time_iso_8601": "2020-06-15T08:15:33.357243Z", - "url": "https://files.pythonhosted.org/packages/ac/92/261266a01e8ff03c71973a3c8ab1604dc3147b8ef35363697cbbe5c2563a/Django-3.1b1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.1rc1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "976ff4f7fed53ae14d6dc3a934e4e4c6e8349d86b8ef8c5dd281fe80da438a41", - "md5": "381eaae505ac141ceebb0eaa7b4855e4", - "sha256": "e5893668d8981a35a5bed8287c56363f83181566037e5beabe09d2a29c8bf4ce" - }, - "downloads": -1, - "filename": "Django-3.1rc1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "381eaae505ac141ceebb0eaa7b4855e4", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7525899, - "upload_time": "2020-07-20T06:38:23", - "upload_time_iso_8601": "2020-07-20T06:38:23.709756Z", - "url": "https://files.pythonhosted.org/packages/97/6f/f4f7fed53ae14d6dc3a934e4e4c6e8349d86b8ef8c5dd281fe80da438a41/Django-3.1rc1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "8a11a2ac7a31950d517518bfbffce011a74acdb65393f8adb85d49571a464456", - "md5": "49dfe691e65a6f9b6cf4d7b9f2cbf834", - "sha256": "a805872093720658e2b858a51c1b43a634790df073b732832d6b8080c38d6263" - }, - "downloads": -1, - "filename": "Django-3.1rc1.tar.gz", - "has_sig": false, - "md5_digest": "49dfe691e65a6f9b6cf4d7b9f2cbf834", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9184160, - "upload_time": "2020-07-20T06:38:29", - "upload_time_iso_8601": "2020-07-20T06:38:29.308366Z", - "url": "https://files.pythonhosted.org/packages/8a/11/a2ac7a31950d517518bfbffce011a74acdb65393f8adb85d49571a464456/Django-3.1rc1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "a89bfe94c509e514f6c227308e81076506eb9d67f2bfb8061ce5cdfbde0432e3", - "md5": "e2cfd14ad74a389429bec15cd8b7391b", - "sha256": "0604e84c4fb698a5e53e5857b5aea945b2f19a18f25f10b8748dbdf935788927" - }, - "downloads": -1, - "filename": "Django-3.2-py3-none-any.whl", - "has_sig": false, - "md5_digest": "e2cfd14ad74a389429bec15cd8b7391b", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7881999, - "upload_time": "2021-04-06T09:33:15", - "upload_time_iso_8601": "2021-04-06T09:33:15.157770Z", - "url": "https://files.pythonhosted.org/packages/a8/9b/fe94c509e514f6c227308e81076506eb9d67f2bfb8061ce5cdfbde0432e3/Django-3.2-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "0e4d5137309d6c83bcbd2966c604fdc633e22ce6fef1292b44c38ae22b4ad90d", - "md5": "0db580470a6a1dc20ccb805f94479ffa", - "sha256": "21f0f9643722675976004eb683c55d33c05486f94506672df3d6a141546f389d" - }, - "downloads": -1, - "filename": "Django-3.2.tar.gz", - "has_sig": false, - "md5_digest": "0db580470a6a1dc20ccb805f94479ffa", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9819119, - "upload_time": "2021-04-06T09:33:26", - "upload_time_iso_8601": "2021-04-06T09:33:26.902785Z", - "url": "https://files.pythonhosted.org/packages/0e/4d/5137309d6c83bcbd2966c604fdc633e22ce6fef1292b44c38ae22b4ad90d/Django-3.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "12cf3eab33ab45e2ce93d5a7500c523f8a407ddcb71177759741c7bf514b7bbe", - "md5": "dd5ba0f289ab783e2359a078b569e054", - "sha256": "e2f73790c60188d3f94f08f644de249d956b3789161e7604509d128a13fb2fcc" - }, - "downloads": -1, - "filename": "Django-3.2.1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "dd5ba0f289ab783e2359a078b569e054", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7883285, - "upload_time": "2021-05-04T08:47:40", - "upload_time_iso_8601": "2021-05-04T08:47:40.847587Z", - "url": "https://files.pythonhosted.org/packages/12/cf/3eab33ab45e2ce93d5a7500c523f8a407ddcb71177759741c7bf514b7bbe/Django-3.2.1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "7502a1cbf429bbbe6c7aa0c37124e3dc55bf932770351a8dcde0b27cba7f3c8c", - "md5": "0ded0d3408c38f4a5cff2128f5a9c4ba", - "sha256": "95c13c750f1f214abadec92b82c2768a5e795e6c2ebd0b4126f895ce9efffcdd" - }, - "downloads": -1, - "filename": "Django-3.2.1.tar.gz", - "has_sig": false, - "md5_digest": "0ded0d3408c38f4a5cff2128f5a9c4ba", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9820723, - "upload_time": "2021-05-04T08:48:26", - "upload_time_iso_8601": "2021-05-04T08:48:26.422878Z", - "url": "https://files.pythonhosted.org/packages/75/02/a1cbf429bbbe6c7aa0c37124e3dc55bf932770351a8dcde0b27cba7f3c8c/Django-3.2.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.10": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "8f9c1ba2005b24cdb443eee5a43ebfa427f7915d6ed68f41484244ae21cd558f", - "md5": "a7c62b14a2123f22e6b3cbdc591f2ce0", - "sha256": "df6f5eb3c797b27c096d61494507b7634526d4ce8d7c8ca1e57a4fb19c0738a3" - }, - "downloads": -1, - "filename": "Django-3.2.10-py3-none-any.whl", - "has_sig": false, - "md5_digest": "a7c62b14a2123f22e6b3cbdc591f2ce0", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7886736, - "upload_time": "2021-12-07T07:34:50", - "upload_time_iso_8601": "2021-12-07T07:34:50.571943Z", - "url": "https://files.pythonhosted.org/packages/8f/9c/1ba2005b24cdb443eee5a43ebfa427f7915d6ed68f41484244ae21cd558f/Django-3.2.10-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "a58ec6dfc718d572e4b33b56824b9e71e5ab9be8072e6747fc6184d206c3fdb3", - "md5": "eaf0c3b4ac6b22cae9068360e6fd2d1b", - "sha256": "074e8818b4b40acdc2369e67dcd6555d558329785408dcd25340ee98f1f1d5c4" - }, - "downloads": -1, - "filename": "Django-3.2.10.tar.gz", - "has_sig": false, - "md5_digest": "eaf0c3b4ac6b22cae9068360e6fd2d1b", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9811341, - "upload_time": "2021-12-07T07:35:06", - "upload_time_iso_8601": "2021-12-07T07:35:06.857668Z", - "url": "https://files.pythonhosted.org/packages/a5/8e/c6dfc718d572e4b33b56824b9e71e5ab9be8072e6747fc6184d206c3fdb3/Django-3.2.10.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.11": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "03401ec2b4abb0c91f0c6195692a9f7a3709f1c0fe95258f3e4d8aa7d8dab92b", - "md5": "12b3d38359e61c4843ab50c62dc43552", - "sha256": "0a0a37f0b93aef30c4bf3a839c187e1175bcdeb7e177341da0cb7b8194416891" - }, - "downloads": -1, - "filename": "Django-3.2.11-py3-none-any.whl", - "has_sig": false, - "md5_digest": "12b3d38359e61c4843ab50c62dc43552", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7887450, - "upload_time": "2022-01-04T09:53:21", - "upload_time_iso_8601": "2022-01-04T09:53:21.952080Z", - "url": "https://files.pythonhosted.org/packages/03/40/1ec2b4abb0c91f0c6195692a9f7a3709f1c0fe95258f3e4d8aa7d8dab92b/Django-3.2.11-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "2086e4348aac45bc83fc8e9dda2cfd81004b007c65b68c1499a4233acabdaa3b", - "md5": "6c4a53d2ccb464bc3dd772c6f2f07df9", - "sha256": "69c94abe5d6b1b088bf475e09b7b74403f943e34da107e798465d2045da27e75" - }, - "downloads": -1, - "filename": "Django-3.2.11.tar.gz", - "has_sig": false, - "md5_digest": "6c4a53d2ccb464bc3dd772c6f2f07df9", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9821958, - "upload_time": "2022-01-04T09:53:34", - "upload_time_iso_8601": "2022-01-04T09:53:34.580658Z", - "url": "https://files.pythonhosted.org/packages/20/86/e4348aac45bc83fc8e9dda2cfd81004b007c65b68c1499a4233acabdaa3b/Django-3.2.11.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.12": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "9c0e02b7eff8fac2c25ede489933d4e899f6e6f283ae8eaf5189431057c8d406", - "md5": "d8628e805946c4440b2ee701951dd255", - "sha256": "9b06c289f9ba3a8abea16c9c9505f25107809fb933676f6c891ded270039d965" - }, - "downloads": -1, - "filename": "Django-3.2.12-py3-none-any.whl", - "has_sig": false, - "md5_digest": "d8628e805946c4440b2ee701951dd255", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7887444, - "upload_time": "2022-02-01T07:56:19", - "upload_time_iso_8601": "2022-02-01T07:56:19.625795Z", - "url": "https://files.pythonhosted.org/packages/9c/0e/02b7eff8fac2c25ede489933d4e899f6e6f283ae8eaf5189431057c8d406/Django-3.2.12-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "83d97f28811ff78ce1903dc9a32ac439e4e6c98298cd2e99cb01f528e51dd796", - "md5": "1847b2f286930a9d84e820a757e3a7ec", - "sha256": "9772e6935703e59e993960832d66a614cf0233a1c5123bc6224ecc6ad69e41e2" - }, - "downloads": -1, - "filename": "Django-3.2.12.tar.gz", - "has_sig": false, - "md5_digest": "1847b2f286930a9d84e820a757e3a7ec", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9812448, - "upload_time": "2022-02-01T07:56:32", - "upload_time_iso_8601": "2022-02-01T07:56:32.000097Z", - "url": "https://files.pythonhosted.org/packages/83/d9/7f28811ff78ce1903dc9a32ac439e4e6c98298cd2e99cb01f528e51dd796/Django-3.2.12.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.13": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "c368b63abc009adee25c1bd266b3740e173eab656608cf21641594a37e02cf57", - "md5": "f0326d2e15e10872c77248a10f24273a", - "sha256": "b896ca61edc079eb6bbaa15cf6071eb69d6aac08cce5211583cfb41515644fdf" - }, - "downloads": -1, - "filename": "Django-3.2.13-py3-none-any.whl", - "has_sig": false, - "md5_digest": "f0326d2e15e10872c77248a10f24273a", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7887998, - "upload_time": "2022-04-11T07:52:57", - "upload_time_iso_8601": "2022-04-11T07:52:57.318239Z", - "url": "https://files.pythonhosted.org/packages/c3/68/b63abc009adee25c1bd266b3740e173eab656608cf21641594a37e02cf57/Django-3.2.13-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "926952760ffd860e995b390a4e897d38ae0669ce633c0fbb76b15f71cb0bb930", - "md5": "fc8b0799ebe689fac24f13384b450c00", - "sha256": "6d93497a0a9bf6ba0e0b1a29cccdc40efbfc76297255b1309b3a884a688ec4b6" - }, - "downloads": -1, - "filename": "Django-3.2.13.tar.gz", - "has_sig": false, - "md5_digest": "fc8b0799ebe689fac24f13384b450c00", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9813985, - "upload_time": "2022-04-11T07:53:10", - "upload_time_iso_8601": "2022-04-11T07:53:10.596399Z", - "url": "https://files.pythonhosted.org/packages/92/69/52760ffd860e995b390a4e897d38ae0669ce633c0fbb76b15f71cb0bb930/Django-3.2.13.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.14": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "c0a84afc7742ad1e506909ce8c5056761f22c8a2db0a8b4a46cfa290b1cd6685", - "md5": "95a74996e5e1fcd8b0c5c4f67c13d363", - "sha256": "a8681e098fa60f7c33a4b628d6fcd3fe983a0939ff1301ecacac21d0b38bad56" - }, - "downloads": -1, - "filename": "Django-3.2.14-py3-none-any.whl", - "has_sig": false, - "md5_digest": "95a74996e5e1fcd8b0c5c4f67c13d363", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7888106, - "upload_time": "2022-07-04T07:57:18", - "upload_time_iso_8601": "2022-07-04T07:57:18.239960Z", - "url": "https://files.pythonhosted.org/packages/c0/a8/4afc7742ad1e506909ce8c5056761f22c8a2db0a8b4a46cfa290b1cd6685/Django-3.2.14-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "8e50649a8e31a5f9e09df32250103c9b34e4af7445bc0a177a1f4d1ed0d1b4eb", - "md5": "8e8872a592e4f85b670f1d9cf21b09ab", - "sha256": "677182ba8b5b285a4e072f3ac17ceee6aff1b5ce77fd173cc5b6a2d3dc022fcf" - }, - "downloads": -1, - "filename": "Django-3.2.14.tar.gz", - "has_sig": false, - "md5_digest": "8e8872a592e4f85b670f1d9cf21b09ab", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9814965, - "upload_time": "2022-07-04T07:57:28", - "upload_time_iso_8601": "2022-07-04T07:57:28.551095Z", - "url": "https://files.pythonhosted.org/packages/8e/50/649a8e31a5f9e09df32250103c9b34e4af7445bc0a177a1f4d1ed0d1b4eb/Django-3.2.14.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.15": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "dbf99ddc8444397ed7e72c52f63b48ecc2849ae1ca4d621776399a81e501ee3c", - "md5": "ccc439198ed761387ed2c9d076253322", - "sha256": "115baf5049d5cf4163e43492cdc7139c306ed6d451e7d3571fe9612903903713" - }, - "downloads": -1, - "filename": "Django-3.2.15-py3-none-any.whl", - "has_sig": false, - "md5_digest": "ccc439198ed761387ed2c9d076253322", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7888258, - "upload_time": "2022-08-03T07:38:15", - "upload_time_iso_8601": "2022-08-03T07:38:15.296674Z", - "url": "https://files.pythonhosted.org/packages/db/f9/9ddc8444397ed7e72c52f63b48ecc2849ae1ca4d621776399a81e501ee3c/Django-3.2.15-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "0245371717cbed904aa95864cf2b28beece8e73f6314f1d2b26cba4c2890dbeb", - "md5": "0bc509df6fd459fd7258b3ffe78b1d99", - "sha256": "f71934b1a822f14a86c9ac9634053689279cd04ae69cb6ade4a59471b886582b" - }, - "downloads": -1, - "filename": "Django-3.2.15.tar.gz", - "has_sig": false, - "md5_digest": "0bc509df6fd459fd7258b3ffe78b1d99", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9833828, - "upload_time": "2022-08-03T07:38:21", - "upload_time_iso_8601": "2022-08-03T07:38:21.696991Z", - "url": "https://files.pythonhosted.org/packages/02/45/371717cbed904aa95864cf2b28beece8e73f6314f1d2b26cba4c2890dbeb/Django-3.2.15.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.16": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "8ac4f946a6b02fcbba84e56074f2fc36866433b009bea2528b09fe0bac4fe1aa", - "md5": "72e20ce30031ad244dcaf79257ed0830", - "sha256": "18ba8efa36b69cfcd4b670d0fa187c6fe7506596f0ababe580e16909bcdec121" - }, - "downloads": -1, - "filename": "Django-3.2.16-py3-none-any.whl", - "has_sig": false, - "md5_digest": "72e20ce30031ad244dcaf79257ed0830", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7888258, - "upload_time": "2022-10-04T07:54:16", - "upload_time_iso_8601": "2022-10-04T07:54:16.950129Z", - "url": "https://files.pythonhosted.org/packages/8a/c4/f946a6b02fcbba84e56074f2fc36866433b009bea2528b09fe0bac4fe1aa/Django-3.2.16-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "f1d78ddb9d8fbbffd33329cf6dfc33e8367a8969c05cf91bc63138ca116301ba", - "md5": "247d5fc679fa295fff99e9617eff827d", - "sha256": "3adc285124244724a394fa9b9839cc8cd116faf7d159554c43ecdaa8cdf0b94d" - }, - "downloads": -1, - "filename": "Django-3.2.16.tar.gz", - "has_sig": false, - "md5_digest": "247d5fc679fa295fff99e9617eff827d", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9847052, - "upload_time": "2022-10-04T07:54:27", - "upload_time_iso_8601": "2022-10-04T07:54:27.079681Z", - "url": "https://files.pythonhosted.org/packages/f1/d7/8ddb9d8fbbffd33329cf6dfc33e8367a8969c05cf91bc63138ca116301ba/Django-3.2.16.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.17": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "1f2eefaa4b1c7db4292715ecedd95b2221438d7a897132f652708d84e4587d59", - "md5": "c4f42d28a6f7abf7dc20c2215a34f9f8", - "sha256": "59c39fc342b242fb42b6b040ad8b1b4c15df438706c1d970d416d63cdd73e7fd" - }, - "downloads": -1, - "filename": "Django-3.2.17-py3-none-any.whl", - "has_sig": false, - "md5_digest": "c4f42d28a6f7abf7dc20c2215a34f9f8", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7888712, - "upload_time": "2023-02-01T09:55:48", - "upload_time_iso_8601": "2023-02-01T09:55:48.915222Z", - "url": "https://files.pythonhosted.org/packages/1f/2e/efaa4b1c7db4292715ecedd95b2221438d7a897132f652708d84e4587d59/Django-3.2.17-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "5a765ae8a875cdb1d132408c5e7c0c6c2832e1d795837c2355abb2b0de9990ec", - "md5": "ef4c165db99f7f6e32b62846b9f7a36e", - "sha256": "644288341f06ebe4938eec6801b6bd59a6534a78e4aedde2a153075d11143894" - }, - "downloads": -1, - "filename": "Django-3.2.17.tar.gz", - "has_sig": false, - "md5_digest": "ef4c165db99f7f6e32b62846b9f7a36e", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9830188, - "upload_time": "2023-02-01T09:56:01", - "upload_time_iso_8601": "2023-02-01T09:56:01.031856Z", - "url": "https://files.pythonhosted.org/packages/5a/76/5ae8a875cdb1d132408c5e7c0c6c2832e1d795837c2355abb2b0de9990ec/Django-3.2.17.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.18": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "5712da22535f809b8c06c8d58eaf236ec8683ffd4e1dc4eced175b174e6446fa", - "md5": "b596a293579c30238d2c486f59f67da6", - "sha256": "4d492d9024c7b3dfababf49f94511ab6a58e2c9c3c7207786f1ba4eb77750706" - }, - "downloads": -1, - "filename": "Django-3.2.18-py3-none-any.whl", - "has_sig": false, - "md5_digest": "b596a293579c30238d2c486f59f67da6", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7889264, - "upload_time": "2023-02-14T08:25:28", - "upload_time_iso_8601": "2023-02-14T08:25:28.389548Z", - "url": "https://files.pythonhosted.org/packages/57/12/da22535f809b8c06c8d58eaf236ec8683ffd4e1dc4eced175b174e6446fa/Django-3.2.18-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "66e6e06dd4fceb09cbb50598cbaecca0dc1a2ac3bc4b8cf3e073e394603c3d85", - "md5": "03831fdb086d0efb7ba0b4e1c521427e", - "sha256": "08208dfe892eb64fff073ca743b3b952311104f939e7f6dae954fe72dcc533ba" - }, - "downloads": -1, - "filename": "Django-3.2.18.tar.gz", - "has_sig": false, - "md5_digest": "03831fdb086d0efb7ba0b4e1c521427e", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9848949, - "upload_time": "2023-02-14T08:25:40", - "upload_time_iso_8601": "2023-02-14T08:25:40.252336Z", - "url": "https://files.pythonhosted.org/packages/66/e6/e06dd4fceb09cbb50598cbaecca0dc1a2ac3bc4b8cf3e073e394603c3d85/Django-3.2.18.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.19": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "22af979a4c610e727cc936c3db3d48cfcb3c270e106ff919f23fc1a27870ba00", - "md5": "8b5d5a06e2c288ff6be0a5e8e556a6dd", - "sha256": "21cc991466245d659ab79cb01204f9515690f8dae00e5eabde307f14d24d4d7d" - }, - "downloads": -1, - "filename": "Django-3.2.19-py3-none-any.whl", - "has_sig": false, - "md5_digest": "8b5d5a06e2c288ff6be0a5e8e556a6dd", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7889399, - "upload_time": "2023-05-03T12:58:19", - "upload_time_iso_8601": "2023-05-03T12:58:19.066030Z", - "url": "https://files.pythonhosted.org/packages/22/af/979a4c610e727cc936c3db3d48cfcb3c270e106ff919f23fc1a27870ba00/Django-3.2.19-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "e5a7cc0b6f97d18a70fd9e026e00a885a9fb995fe2166a56e7784fb361abb027", - "md5": "d84f0b8669678fea14579d7400a521e2", - "sha256": "031365bae96814da19c10706218c44dff3b654cc4de20a98bd2d29b9bde469f0" - }, - "downloads": -1, - "filename": "Django-3.2.19.tar.gz", - "has_sig": false, - "md5_digest": "d84f0b8669678fea14579d7400a521e2", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9832772, - "upload_time": "2023-05-03T12:58:31", - "upload_time_iso_8601": "2023-05-03T12:58:31.141156Z", - "url": "https://files.pythonhosted.org/packages/e5/a7/cc0b6f97d18a70fd9e026e00a885a9fb995fe2166a56e7784fb361abb027/Django-3.2.19.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "cf91e23103dd21fa1b5c1fefb65c4d403107b10bf450ee6955621169fcc86db9", - "md5": "abd67e107427fb9b5f68863bf0b384d5", - "sha256": "18dd3145ddbd04bf189ff79b9954d08fda5171ea7b57bf705789fea766a07d50" - }, - "downloads": -1, - "filename": "Django-3.2.2-py3-none-any.whl", - "has_sig": false, - "md5_digest": "abd67e107427fb9b5f68863bf0b384d5", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7883316, - "upload_time": "2021-05-06T07:40:00", - "upload_time_iso_8601": "2021-05-06T07:40:00.067668Z", - "url": "https://files.pythonhosted.org/packages/cf/91/e23103dd21fa1b5c1fefb65c4d403107b10bf450ee6955621169fcc86db9/Django-3.2.2-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "da24e2e6e534464f8e0bd010401f06d2cfc773141776d2952d6418d01c97f12c", - "md5": "43784c090a8805605e3d0b768cd21cb2", - "sha256": "0a1d195ad65c52bf275b8277b3d49680bd1137a5f55039a806f25f6b9752ce3d" - }, - "downloads": -1, - "filename": "Django-3.2.2.tar.gz", - "has_sig": false, - "md5_digest": "43784c090a8805605e3d0b768cd21cb2", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9796920, - "upload_time": "2021-05-06T07:40:03", - "upload_time_iso_8601": "2021-05-06T07:40:03.835147Z", - "url": "https://files.pythonhosted.org/packages/da/24/e2e6e534464f8e0bd010401f06d2cfc773141776d2952d6418d01c97f12c/Django-3.2.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.20": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "84eb5329ae72bf26b91844985d0de74e4edf876e3ca409d085820f230eea2eba", - "md5": "f573fccdc95866fa1256e10b251ce257", - "sha256": "a477ab326ae7d8807dc25c186b951ab8c7648a3a23f9497763c37307a2b5ef87" - }, - "downloads": -1, - "filename": "Django-3.2.20-py3-none-any.whl", - "has_sig": false, - "md5_digest": "f573fccdc95866fa1256e10b251ce257", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7889501, - "upload_time": "2023-07-03T07:57:12", - "upload_time_iso_8601": "2023-07-03T07:57:12.203517Z", - "url": "https://files.pythonhosted.org/packages/84/eb/5329ae72bf26b91844985d0de74e4edf876e3ca409d085820f230eea2eba/Django-3.2.20-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "52b62a36af602c8edcaa74ab9c046d8e950771bd503c3261d95466f9d5faac89", - "md5": "effe6b4ccf2606818578b9e9a94d01e0", - "sha256": "dec2a116787b8e14962014bf78e120bba454135108e1af9e9b91ade7b2964c40" - }, - "downloads": -1, - "filename": "Django-3.2.20.tar.gz", - "has_sig": false, - "md5_digest": "effe6b4ccf2606818578b9e9a94d01e0", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9831078, - "upload_time": "2023-07-03T07:57:23", - "upload_time_iso_8601": "2023-07-03T07:57:23.346933Z", - "url": "https://files.pythonhosted.org/packages/52/b6/2a36af602c8edcaa74ab9c046d8e950771bd503c3261d95466f9d5faac89/Django-3.2.20.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.21": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "978d03ac2f7a3f751f597be12c03a09147f9ca80906264044bb05758bbcc2b32", - "md5": "f31d1bdaf205742d64a0830484d4f708", - "sha256": "d31b06c58aa2cd73998ca5966bc3001243d3c4e77ee2d0c479bced124765fd99" - }, - "downloads": -1, - "filename": "Django-3.2.21-py3-none-any.whl", - "has_sig": false, - "md5_digest": "f31d1bdaf205742d64a0830484d4f708", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7889527, - "upload_time": "2023-09-04T10:58:15", - "upload_time_iso_8601": "2023-09-04T10:58:15.304456Z", - "url": "https://files.pythonhosted.org/packages/97/8d/03ac2f7a3f751f597be12c03a09147f9ca80906264044bb05758bbcc2b32/Django-3.2.21-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "9f4a2b57890fd9a1511ed3c580ef3636f018d06b1f9ae42b3033e4fff3ff7720", - "md5": "38c4eba2d11374a9c1dd73300df7771d", - "sha256": "a5de4c484e7b7418e6d3e52a5b8794f0e6b9f9e4ce3c037018cf1c489fa87f3c" - }, - "downloads": -1, - "filename": "Django-3.2.21.tar.gz", - "has_sig": false, - "md5_digest": "38c4eba2d11374a9c1dd73300df7771d", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9836824, - "upload_time": "2023-09-04T10:58:25", - "upload_time_iso_8601": "2023-09-04T10:58:25.702642Z", - "url": "https://files.pythonhosted.org/packages/9f/4a/2b57890fd9a1511ed3c580ef3636f018d06b1f9ae42b3033e4fff3ff7720/Django-3.2.21.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.22": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "1b2913d5f2a5de460b66a92e4e99b0e2ceb9f3e619da70e51772981268524e0d", - "md5": "eba26762f6788e9d9ad2e707ec956610", - "sha256": "c5e7b668025a6e06cad9ba6d4de1fd1a21212acebb51ea34abb400c6e4d33430" - }, - "downloads": -1, - "filename": "Django-3.2.22-py3-none-any.whl", - "has_sig": false, - "md5_digest": "eba26762f6788e9d9ad2e707ec956610", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7889667, - "upload_time": "2023-10-04T14:59:53", - "upload_time_iso_8601": "2023-10-04T14:59:53.611484Z", - "url": "https://files.pythonhosted.org/packages/1b/29/13d5f2a5de460b66a92e4e99b0e2ceb9f3e619da70e51772981268524e0d/Django-3.2.22-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "46fa6ff52b6d92e52eec8d5037f79337ff70321eb7831dda1d13da3ad58f53ab", - "md5": "cf9fb498a109855822326fe7cf2a1a28", - "sha256": "83b6d66b06e484807d778263fdc7f9186d4dc1862fcfa6507830446ac6b060ba" - }, - "downloads": -1, - "filename": "Django-3.2.22.tar.gz", - "has_sig": false, - "md5_digest": "cf9fb498a109855822326fe7cf2a1a28", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9827193, - "upload_time": "2023-10-04T15:00:00", - "upload_time_iso_8601": "2023-10-04T15:00:00.548503Z", - "url": "https://files.pythonhosted.org/packages/46/fa/6ff52b6d92e52eec8d5037f79337ff70321eb7831dda1d13da3ad58f53ab/Django-3.2.22.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.23": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "c96972663af69f1a70fa186813220c8ec5ce7f9056b6d987077e2c38ed794bee", - "md5": "f0a76afe885b290578337eedca37eecc", - "sha256": "d48608d5f62f2c1e260986835db089fa3b79d6f58510881d316b8d88345ae6e1" - }, - "downloads": -1, - "filename": "Django-3.2.23-py3-none-any.whl", - "has_sig": false, - "md5_digest": "f0a76afe885b290578337eedca37eecc", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7889872, - "upload_time": "2023-11-01T06:59:05", - "upload_time_iso_8601": "2023-11-01T06:59:05.105405Z", - "url": "https://files.pythonhosted.org/packages/c9/69/72663af69f1a70fa186813220c8ec5ce7f9056b6d987077e2c38ed794bee/Django-3.2.23-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "dd6e7c97a4c54c1a4d3d700fb0d687323c96dc5e9f5e7d2bcac163e2f0c17cb7", - "md5": "eaca9b9fba9106fbb52c26c9b2ec8cfa", - "sha256": "82968f3640e29ef4a773af2c28448f5f7a08d001c6ac05b32d02aeee6509508b" - }, - "downloads": -1, - "filename": "Django-3.2.23.tar.gz", - "has_sig": false, - "md5_digest": "eaca9b9fba9106fbb52c26c9b2ec8cfa", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9834665, - "upload_time": "2023-11-01T06:59:19", - "upload_time_iso_8601": "2023-11-01T06:59:19.806570Z", - "url": "https://files.pythonhosted.org/packages/dd/6e/7c97a4c54c1a4d3d700fb0d687323c96dc5e9f5e7d2bcac163e2f0c17cb7/Django-3.2.23.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "8969c556b5b3e7a6701724485fc07c8349791e585b784dc70c9c0683d98ef0db", - "md5": "5a33b1123433c5df329de05d92148730", - "sha256": "7e0a1393d18c16b503663752a8b6790880c5084412618990ce8a81cc908b4962" - }, - "downloads": -1, - "filename": "Django-3.2.3-py3-none-any.whl", - "has_sig": false, - "md5_digest": "5a33b1123433c5df329de05d92148730", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7883562, - "upload_time": "2021-05-13T07:36:47", - "upload_time_iso_8601": "2021-05-13T07:36:47.334786Z", - "url": "https://files.pythonhosted.org/packages/89/69/c556b5b3e7a6701724485fc07c8349791e585b784dc70c9c0683d98ef0db/Django-3.2.3-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "f2a5905c45599a599d8b66f2a572a384ffd29bf5482233ec701fd53d5f52a513", - "md5": "ec5fc12eabe33d0ccacc2f12ee43d1fe", - "sha256": "13ac78dbfd189532cad8f383a27e58e18b3d33f80009ceb476d7fcbfc5dcebd8" - }, - "downloads": -1, - "filename": "Django-3.2.3.tar.gz", - "has_sig": false, - "md5_digest": "ec5fc12eabe33d0ccacc2f12ee43d1fe", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9798957, - "upload_time": "2021-05-13T07:37:01", - "upload_time_iso_8601": "2021-05-13T07:37:01.485953Z", - "url": "https://files.pythonhosted.org/packages/f2/a5/905c45599a599d8b66f2a572a384ffd29bf5482233ec701fd53d5f52a513/Django-3.2.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "7022ed1943c0ef2be99ade872f49a20aa4cfc70eb4ffc866fc61b128211f3e5d", - "md5": "d2975d6084b5740de5838ccf7db3e823", - "sha256": "ea735cbbbb3b2fba6d4da4784a0043d84c67c92f1fdf15ad6db69900e792c10f" - }, - "downloads": -1, - "filename": "Django-3.2.4-py3-none-any.whl", - "has_sig": false, - "md5_digest": "d2975d6084b5740de5838ccf7db3e823", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7883920, - "upload_time": "2021-06-02T08:54:01", - "upload_time_iso_8601": "2021-06-02T08:54:01.716712Z", - "url": "https://files.pythonhosted.org/packages/70/22/ed1943c0ef2be99ade872f49a20aa4cfc70eb4ffc866fc61b128211f3e5d/Django-3.2.4-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "2794123b3a95e9965819a3d30d36da6fc12ddff83bcfb0099f3e15437347480a", - "md5": "2f30db9154efb8c9ed891781d29fae2a", - "sha256": "66c9d8db8cc6fe938a28b7887c1596e42d522e27618562517cc8929eb7e7f296" - }, - "downloads": -1, - "filename": "Django-3.2.4.tar.gz", - "has_sig": false, - "md5_digest": "2f30db9154efb8c9ed891781d29fae2a", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9824343, - "upload_time": "2021-06-02T08:54:46", - "upload_time_iso_8601": "2021-06-02T08:54:46.382181Z", - "url": "https://files.pythonhosted.org/packages/27/94/123b3a95e9965819a3d30d36da6fc12ddff83bcfb0099f3e15437347480a/Django-3.2.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.5": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "37a1790e01bf4348dd68090d47108052a8130954d473d54b4ea7924f5bb154de", - "md5": "f6eb0de91e191b6e6114ded04a22ee89", - "sha256": "c58b5f19c5ae0afe6d75cbdd7df561e6eb929339985dbbda2565e1cabb19a62e" - }, - "downloads": -1, - "filename": "Django-3.2.5-py3-none-any.whl", - "has_sig": false, - "md5_digest": "f6eb0de91e191b6e6114ded04a22ee89", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7886308, - "upload_time": "2021-07-01T07:40:00", - "upload_time_iso_8601": "2021-07-01T07:40:00.999209Z", - "url": "https://files.pythonhosted.org/packages/37/a1/790e01bf4348dd68090d47108052a8130954d473d54b4ea7924f5bb154de/Django-3.2.5-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "953b468fa33908feefac03c0a773bd73bb8a1ab1fb4ee06e9dd62d24981f4603", - "md5": "46e306a5a775cace03a03d5a158ff767", - "sha256": "3da05fea54fdec2315b54a563d5b59f3b4e2b1e69c3a5841dda35019c01855cd" - }, - "downloads": -1, - "filename": "Django-3.2.5.tar.gz", - "has_sig": false, - "md5_digest": "46e306a5a775cace03a03d5a158ff767", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9806547, - "upload_time": "2021-07-01T07:40:10", - "upload_time_iso_8601": "2021-07-01T07:40:10.126225Z", - "url": "https://files.pythonhosted.org/packages/95/3b/468fa33908feefac03c0a773bd73bb8a1ab1fb4ee06e9dd62d24981f4603/Django-3.2.5.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.6": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "d59b3514fae1e9d0c71044739dca5ed55f50443bd1309548b63603712365e6e9", - "md5": "5d57618d2af3ea3cdd1c076212adc317", - "sha256": "7f92413529aa0e291f3be78ab19be31aefb1e1c9a52cd59e130f505f27a51f13" - }, - "downloads": -1, - "filename": "Django-3.2.6-py3-none-any.whl", - "has_sig": false, - "md5_digest": "5d57618d2af3ea3cdd1c076212adc317", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7886351, - "upload_time": "2021-08-02T06:28:33", - "upload_time_iso_8601": "2021-08-02T06:28:33.280855Z", - "url": "https://files.pythonhosted.org/packages/d5/9b/3514fae1e9d0c71044739dca5ed55f50443bd1309548b63603712365e6e9/Django-3.2.6-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "683e068446a8bf87199d0fec94f38569a8884a49fdf5811fe652f653218975f0", - "md5": "5956dbbbc9e93bde3fdf47aada293adc", - "sha256": "f27f8544c9d4c383bbe007c57e3235918e258364577373d4920e9162837be022" - }, - "downloads": -1, - "filename": "Django-3.2.6.tar.gz", - "has_sig": false, - "md5_digest": "5956dbbbc9e93bde3fdf47aada293adc", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9821499, - "upload_time": "2021-08-02T06:28:42", - "upload_time_iso_8601": "2021-08-02T06:28:42.678399Z", - "url": "https://files.pythonhosted.org/packages/68/3e/068446a8bf87199d0fec94f38569a8884a49fdf5811fe652f653218975f0/Django-3.2.6.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.7": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "271c6fe40cdfdbbc8a0d7c211dde68777f1d435bde7879697d0bc20c73d136ac", - "md5": "5edb6aad7336d2d1fafb49c6d3989cb0", - "sha256": "e93c93565005b37ddebf2396b4dc4b6913c1838baa82efdfb79acedd5816c240" - }, - "downloads": -1, - "filename": "Django-3.2.7-py3-none-any.whl", - "has_sig": false, - "md5_digest": "5edb6aad7336d2d1fafb49c6d3989cb0", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7886385, - "upload_time": "2021-09-01T05:57:16", - "upload_time_iso_8601": "2021-09-01T05:57:16.373371Z", - "url": "https://files.pythonhosted.org/packages/27/1c/6fe40cdfdbbc8a0d7c211dde68777f1d435bde7879697d0bc20c73d136ac/Django-3.2.7-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "5945c6fbb3a206df0b7dc3e6e8fae738e042c63d4ddf828c6e1ba10d7417a1d9", - "md5": "2ade1eecca77640abbde6c4589da27dd", - "sha256": "95b318319d6997bac3595517101ad9cc83fe5672ac498ba48d1a410f47afecd2" - }, - "downloads": -1, - "filename": "Django-3.2.7.tar.gz", - "has_sig": false, - "md5_digest": "2ade1eecca77640abbde6c4589da27dd", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9808777, - "upload_time": "2021-09-01T05:57:20", - "upload_time_iso_8601": "2021-09-01T05:57:20.280283Z", - "url": "https://files.pythonhosted.org/packages/59/45/c6fbb3a206df0b7dc3e6e8fae738e042c63d4ddf828c6e1ba10d7417a1d9/Django-3.2.7.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.8": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "22e67c6f9e682578550f748cc33dff04bf812fe0fb7d819fa094a4bb6d2dc2a1", - "md5": "afeaa167836ef02bb9891d9fbfe4f4ab", - "sha256": "42573831292029639b798fe4d3812996bfe4ff3275f04566da90764daec011a5" - }, - "downloads": -1, - "filename": "Django-3.2.8-py3-none-any.whl", - "has_sig": false, - "md5_digest": "afeaa167836ef02bb9891d9fbfe4f4ab", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7886424, - "upload_time": "2021-10-05T07:46:17", - "upload_time_iso_8601": "2021-10-05T07:46:17.668882Z", - "url": "https://files.pythonhosted.org/packages/22/e6/7c6f9e682578550f748cc33dff04bf812fe0fb7d819fa094a4bb6d2dc2a1/Django-3.2.8-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "99f78d62b9881bc8923da227f4bbfa01be97d2b7c89b07237c944fe60bcad047", - "md5": "58b3847abe6908d681b72f0941076e7d", - "sha256": "f6d2c4069c9b9bfac03bedff927ea1f9e0d29e34525cec8a68fd28eb2a8df7af" - }, - "downloads": -1, - "filename": "Django-3.2.8.tar.gz", - "has_sig": false, - "md5_digest": "58b3847abe6908d681b72f0941076e7d", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9820955, - "upload_time": "2021-10-05T07:46:25", - "upload_time_iso_8601": "2021-10-05T07:46:25.737120Z", - "url": "https://files.pythonhosted.org/packages/99/f7/8d62b9881bc8923da227f4bbfa01be97d2b7c89b07237c944fe60bcad047/Django-3.2.8.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2.9": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "a8cae88eb097959c48cd313dfc4bc394699a48fe5c158ed3a64c13e4fa46c1fd", - "md5": "dbd4ed5f73c2a7976189ea7a504b1911", - "sha256": "e22c9266da3eec7827737cde57694d7db801fedac938d252bf27377cec06ed1b" - }, - "downloads": -1, - "filename": "Django-3.2.9-py3-none-any.whl", - "has_sig": false, - "md5_digest": "dbd4ed5f73c2a7976189ea7a504b1911", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7886534, - "upload_time": "2021-11-01T09:31:58", - "upload_time_iso_8601": "2021-11-01T09:31:58.387758Z", - "url": "https://files.pythonhosted.org/packages/a8/ca/e88eb097959c48cd313dfc4bc394699a48fe5c158ed3a64c13e4fa46c1fd/Django-3.2.9-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "b676eed8788c87da8890cdeeeb6893b5559ae9ee18ed25923dd8663ae80af64c", - "md5": "86b100c1b2fd4ddf1a35ba394e4ad2d1", - "sha256": "51284300f1522ffcdb07ccbdf676a307c6678659e1284f0618e5a774127a6a08" - }, - "downloads": -1, - "filename": "Django-3.2.9.tar.gz", - "has_sig": false, - "md5_digest": "86b100c1b2fd4ddf1a35ba394e4ad2d1", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9809157, - "upload_time": "2021-11-01T09:32:25", - "upload_time_iso_8601": "2021-11-01T09:32:25.457689Z", - "url": "https://files.pythonhosted.org/packages/b6/76/eed8788c87da8890cdeeeb6893b5559ae9ee18ed25923dd8663ae80af64c/Django-3.2.9.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2a1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "98a3a78b8d425ae30ff02e62703f826ff78e82985da27d08c46e807475bd022b", - "md5": "99fd0019dde851cf1d92779b9c7119cb", - "sha256": "ace5a4cb96e9344537d71a72607764486994d23099166b70c32bb9e8242b9ef1" - }, - "downloads": -1, - "filename": "Django-3.2a1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "99fd0019dde851cf1d92779b9c7119cb", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7862533, - "upload_time": "2021-01-19T13:04:20", - "upload_time_iso_8601": "2021-01-19T13:04:20.758172Z", - "url": "https://files.pythonhosted.org/packages/98/a3/a78b8d425ae30ff02e62703f826ff78e82985da27d08c46e807475bd022b/Django-3.2a1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "29912092ec479d5070a83b0d4c5d43b28a6d0ccb2bcc6ec0d44d2c9c62feb553", - "md5": "009735595126781647664a408193f2b2", - "sha256": "f6b99ef95aa0c5bf51ef0e469f19c21486966c9dce5b2ab5037763dc89a127f0" - }, - "downloads": -1, - "filename": "Django-3.2a1.tar.gz", - "has_sig": false, - "md5_digest": "009735595126781647664a408193f2b2", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9369355, - "upload_time": "2021-01-19T13:04:37", - "upload_time_iso_8601": "2021-01-19T13:04:37.298129Z", - "url": "https://files.pythonhosted.org/packages/29/91/2092ec479d5070a83b0d4c5d43b28a6d0ccb2bcc6ec0d44d2c9c62feb553/Django-3.2a1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2b1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "76f21a32796484a9761852102dcd99a3aad0c73bd41bda5dc927a94fba0bd45d", - "md5": "8aec4c6219b47f79356f329ef5fb21f5", - "sha256": "b6349c182a1985b7488365ec6bb0042446153601b2d7af02e4e768f618ba25b6" - }, - "downloads": -1, - "filename": "Django-3.2b1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "8aec4c6219b47f79356f329ef5fb21f5", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7863052, - "upload_time": "2021-02-19T09:35:36", - "upload_time_iso_8601": "2021-02-19T09:35:36.603900Z", - "url": "https://files.pythonhosted.org/packages/76/f2/1a32796484a9761852102dcd99a3aad0c73bd41bda5dc927a94fba0bd45d/Django-3.2b1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "a7415a78e507a7be6869c7d36e4f04af32c406a6722756ad516bf3e8b4ea48d1", - "md5": "d1ceb874d529de9d808dab3a65ce2dd1", - "sha256": "b3f1afcab9bd5a6250be9caff4cdc505f4f4313cb6c0b301e92739cb80820b3f" - }, - "downloads": -1, - "filename": "Django-3.2b1.tar.gz", - "has_sig": false, - "md5_digest": "d1ceb874d529de9d808dab3a65ce2dd1", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9783658, - "upload_time": "2021-02-19T09:35:43", - "upload_time_iso_8601": "2021-02-19T09:35:43.063787Z", - "url": "https://files.pythonhosted.org/packages/a7/41/5a78e507a7be6869c7d36e4f04af32c406a6722756ad516bf3e8b4ea48d1/Django-3.2b1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "3.2rc1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "1eb9d8c444c88bead7b3a4e5a06c828b693b7bb82ef85a70063e0629db7d7c9f", - "md5": "bbb2a6410e9669a68c435b467e7ecb54", - "sha256": "8a3e3758f7e531e78d473f526ae5d20a3e58551fe4441cd6610b448b3b6f7044" - }, - "downloads": -1, - "filename": "Django-3.2rc1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "bbb2a6410e9669a68c435b467e7ecb54", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.6", - "size": 7863237, - "upload_time": "2021-03-18T13:55:59", - "upload_time_iso_8601": "2021-03-18T13:55:59.358147Z", - "url": "https://files.pythonhosted.org/packages/1e/b9/d8c444c88bead7b3a4e5a06c828b693b7bb82ef85a70063e0629db7d7c9f/Django-3.2rc1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "0a3f3d5d114e4f7705d85a6a8c818f9f70d867c8e45c4ff682fc6a5e5f863995", - "md5": "9d145fcb05f9b7fe96c5636bafae339c", - "sha256": "28c24ae710d71ed65641b83ba3f06cffdb3d620475890cd2bb72ace92cb97a21" - }, - "downloads": -1, - "filename": "Django-3.2rc1.tar.gz", - "has_sig": false, - "md5_digest": "9d145fcb05f9b7fe96c5636bafae339c", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.6", - "size": 9791598, - "upload_time": "2021-03-18T13:56:15", - "upload_time_iso_8601": "2021-03-18T13:56:15.907546Z", - "url": "https://files.pythonhosted.org/packages/0a/3f/3d5d114e4f7705d85a6a8c818f9f70d867c8e45c4ff682fc6a5e5f863995/Django-3.2rc1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.0": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "807052fce3520b7a1421685828b04f76d5a26aabc7603fdb7af26c4ca7bb0512", - "md5": "be6fc2824c0c920728baa7d1911fc425", - "sha256": "59304646ebc6a77b9b6a59adc67d51ecb03c5e3d63ed1f14c909cdfda84e8010" - }, - "downloads": -1, - "filename": "Django-4.0-py3-none-any.whl", - "has_sig": false, - "md5_digest": "be6fc2824c0c920728baa7d1911fc425", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8014008, - "upload_time": "2021-12-07T09:19:58", - "upload_time_iso_8601": "2021-12-07T09:19:58.185240Z", - "url": "https://files.pythonhosted.org/packages/80/70/52fce3520b7a1421685828b04f76d5a26aabc7603fdb7af26c4ca7bb0512/Django-4.0-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "4c32390205123250b54438a6cb3e0c9b5fd0347c318df744f32b50d8e10260d0", - "md5": "0d2c393916334cbd80fdc08fa2b60932", - "sha256": "d5a8a14da819a8b9237ee4d8c78dfe056ff6e8a7511987be627192225113ee75" - }, - "downloads": -1, - "filename": "Django-4.0.tar.gz", - "has_sig": false, - "md5_digest": "0d2c393916334cbd80fdc08fa2b60932", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 9980403, - "upload_time": "2021-12-07T09:20:02", - "upload_time_iso_8601": "2021-12-07T09:20:02.897592Z", - "url": "https://files.pythonhosted.org/packages/4c/32/390205123250b54438a6cb3e0c9b5fd0347c318df744f32b50d8e10260d0/Django-4.0.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.0.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "67ec20e5e4a29a7aef4ba7d2b3c24aac875e3bdf6c7d4de7d73db12282f15e22", - "md5": "ef2e75d202a13038256312523d825443", - "sha256": "7cd8e8a3ed2bc0dfda05ce1e53a9c81b30eefd7aa350e538a18884475e4d4ce2" - }, - "downloads": -1, - "filename": "Django-4.0.1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "ef2e75d202a13038256312523d825443", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8015001, - "upload_time": "2022-01-04T09:53:25", - "upload_time_iso_8601": "2022-01-04T09:53:25.637001Z", - "url": "https://files.pythonhosted.org/packages/67/ec/20e5e4a29a7aef4ba7d2b3c24aac875e3bdf6c7d4de7d73db12282f15e22/Django-4.0.1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "ffa587fcd0ac916824c2e7ede565ea522bf9cc2428b06b8e7070f92fb6a4ad77", - "md5": "6d0fba754d678f69b573dd9fbf5e6fa6", - "sha256": "2485eea3cc4c3bae13080dee866ebf90ba9f98d1afe8fda89bfb0eb2e218ef86" - }, - "downloads": -1, - "filename": "Django-4.0.1.tar.gz", - "has_sig": false, - "md5_digest": "6d0fba754d678f69b573dd9fbf5e6fa6", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 9995484, - "upload_time": "2022-01-04T09:53:39", - "upload_time_iso_8601": "2022-01-04T09:53:39.880362Z", - "url": "https://files.pythonhosted.org/packages/ff/a5/87fcd0ac916824c2e7ede565ea522bf9cc2428b06b8e7070f92fb6a4ad77/Django-4.0.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.0.10": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "60f0b7322879f851bcf939c97310e4e42f223105a7cb0f123c62553a05f132c4", - "md5": "cde3317edb4d4ac926182c5ef03da59f", - "sha256": "4496eb4f65071578b703fdc6e6f29302553c7440e3f77baf4cefa4a4e091fc3d" - }, - "downloads": -1, - "filename": "Django-4.0.10-py3-none-any.whl", - "has_sig": false, - "md5_digest": "cde3317edb4d4ac926182c5ef03da59f", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8042923, - "upload_time": "2023-02-14T08:25:32", - "upload_time_iso_8601": "2023-02-14T08:25:32.863462Z", - "url": "https://files.pythonhosted.org/packages/60/f0/b7322879f851bcf939c97310e4e42f223105a7cb0f123c62553a05f132c4/Django-4.0.10-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "4492b0ceee230f9252460abf8c0a3bcc003a914af85ad63006596c3fb3669fbf", - "md5": "1ff999292535f0c9fd729e60e3365c49", - "sha256": "2c2f73c16b11cb272c6d5e3b063f0d1be06f378d8dc6005fbe8542565db659cc" - }, - "downloads": -1, - "filename": "Django-4.0.10.tar.gz", - "has_sig": false, - "md5_digest": "1ff999292535f0c9fd729e60e3365c49", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10430363, - "upload_time": "2023-02-14T08:25:45", - "upload_time_iso_8601": "2023-02-14T08:25:45.089209Z", - "url": "https://files.pythonhosted.org/packages/44/92/b0ceee230f9252460abf8c0a3bcc003a914af85ad63006596c3fb3669fbf/Django-4.0.10.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.0.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "5d2f4705e8f2848f4523bb1b94b0dd9b425431aa1875c0921fb86777e62b4335", - "md5": "81bb12a26b4c2081ca491c4902bddef9", - "sha256": "996495c58bff749232426c88726d8cd38d24c94d7c1d80835aafffa9bc52985a" - }, - "downloads": -1, - "filename": "Django-4.0.2-py3-none-any.whl", - "has_sig": false, - "md5_digest": "81bb12a26b4c2081ca491c4902bddef9", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8028128, - "upload_time": "2022-02-01T07:56:23", - "upload_time_iso_8601": "2022-02-01T07:56:23.036066Z", - "url": "https://files.pythonhosted.org/packages/5d/2f/4705e8f2848f4523bb1b94b0dd9b425431aa1875c0921fb86777e62b4335/Django-4.0.2-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "6184676c840e8f1188a6c836e3224b97aa8be4c2e6857c690d6c564eb23a4975", - "md5": "a86339c0e87241597afa8744704d9965", - "sha256": "110fb58fb12eca59e072ad59fc42d771cd642dd7a2f2416582aa9da7a8ef954a" - }, - "downloads": -1, - "filename": "Django-4.0.2.tar.gz", - "has_sig": false, - "md5_digest": "a86339c0e87241597afa8744704d9965", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 9996300, - "upload_time": "2022-02-01T07:56:37", - "upload_time_iso_8601": "2022-02-01T07:56:37.646212Z", - "url": "https://files.pythonhosted.org/packages/61/84/676c840e8f1188a6c836e3224b97aa8be4c2e6857c690d6c564eb23a4975/Django-4.0.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.0.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "1a80b06ce333aabba7ab1b6a41ea3c4e46970ceb396e705733480a2d47a7f74b", - "md5": "064fed57c27fcaa6f8dd8c53579b4fd5", - "sha256": "1239218849e922033a35d2a2f777cb8bee18bd725416744074f455f34ff50d0c" - }, - "downloads": -1, - "filename": "Django-4.0.3-py3-none-any.whl", - "has_sig": false, - "md5_digest": "064fed57c27fcaa6f8dd8c53579b4fd5", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8040983, - "upload_time": "2022-03-01T08:47:23", - "upload_time_iso_8601": "2022-03-01T08:47:23.930184Z", - "url": "https://files.pythonhosted.org/packages/1a/80/b06ce333aabba7ab1b6a41ea3c4e46970ceb396e705733480a2d47a7f74b/Django-4.0.3-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "7e3db403c3b82e20da491cc629504665260660c9e7efe12b134aee8f13da29a7", - "md5": "dfafa6178fd2af5a7bfb6f0d1945f0a4", - "sha256": "77ff2e7050e3324c9b67e29b6707754566f58514112a9ac73310f60cd5261930" - }, - "downloads": -1, - "filename": "Django-4.0.3.tar.gz", - "has_sig": false, - "md5_digest": "dfafa6178fd2af5a7bfb6f0d1945f0a4", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10061007, - "upload_time": "2022-03-01T08:47:28", - "upload_time_iso_8601": "2022-03-01T08:47:28.425336Z", - "url": "https://files.pythonhosted.org/packages/7e/3d/b403c3b82e20da491cc629504665260660c9e7efe12b134aee8f13da29a7/Django-4.0.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.0.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "6690bce00eb942fbc47b0774ac78910ee4e6f719572aad56dc238823e5d0ee54", - "md5": "7f80b2bc7114e4fe7006e3574b1f30e5", - "sha256": "07c8638e7a7f548dc0acaaa7825d84b7bd42b10e8d22268b3d572946f1e9b687" - }, - "downloads": -1, - "filename": "Django-4.0.4-py3-none-any.whl", - "has_sig": false, - "md5_digest": "7f80b2bc7114e4fe7006e3574b1f30e5", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8041700, - "upload_time": "2022-04-11T07:53:01", - "upload_time_iso_8601": "2022-04-11T07:53:01.287027Z", - "url": "https://files.pythonhosted.org/packages/66/90/bce00eb942fbc47b0774ac78910ee4e6f719572aad56dc238823e5d0ee54/Django-4.0.4-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "d29593d1f75da95624bf89e373d079fa72debf77f9b10acc31998cc52a5ff3f9", - "md5": "153fcb5dd7360b7ad219d65cb53e2d57", - "sha256": "4e8177858524417563cc0430f29ea249946d831eacb0068a1455686587df40b5" - }, - "downloads": -1, - "filename": "Django-4.0.4.tar.gz", - "has_sig": false, - "md5_digest": "153fcb5dd7360b7ad219d65cb53e2d57", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10388499, - "upload_time": "2022-04-11T07:53:16", - "upload_time_iso_8601": "2022-04-11T07:53:16.406304Z", - "url": "https://files.pythonhosted.org/packages/d2/95/93d1f75da95624bf89e373d079fa72debf77f9b10acc31998cc52a5ff3f9/Django-4.0.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.0.5": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "ec71950794da9635865d27c92a6955083264eabb004c2c12c077036194620823", - "md5": "cfcf0c2efce4fa2f2701dc66248b7b35", - "sha256": "502ae42b6ab1b612c933fb50d5ff850facf858a4c212f76946ecd8ea5b3bf2d9" - }, - "downloads": -1, - "filename": "Django-4.0.5-py3-none-any.whl", - "has_sig": false, - "md5_digest": "cfcf0c2efce4fa2f2701dc66248b7b35", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8041794, - "upload_time": "2022-06-01T12:22:12", - "upload_time_iso_8601": "2022-06-01T12:22:12.865544Z", - "url": "https://files.pythonhosted.org/packages/ec/71/950794da9635865d27c92a6955083264eabb004c2c12c077036194620823/Django-4.0.5-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "5a4c93120fb4275d06c3645c94f2e26fe87f3f52258b61c183b25c630ba34d3a", - "md5": "4a9cc538c9d3383a9c4bd84785d7da16", - "sha256": "f7431a5de7277966f3785557c3928433347d998c1e6459324501378a291e5aab" - }, - "downloads": -1, - "filename": "Django-4.0.5.tar.gz", - "has_sig": false, - "md5_digest": "4a9cc538c9d3383a9c4bd84785d7da16", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10410720, - "upload_time": "2022-06-01T12:22:18", - "upload_time_iso_8601": "2022-06-01T12:22:18.618899Z", - "url": "https://files.pythonhosted.org/packages/5a/4c/93120fb4275d06c3645c94f2e26fe87f3f52258b61c183b25c630ba34d3a/Django-4.0.5.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.0.6": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "1297e341b13d605a9d0732a37975ce6bfae643c296721b8644f42d5d720c99bf", - "md5": "8f9ef276f0c1b6f3f560d179c7e76d6a", - "sha256": "ca54ebedfcbc60d191391efbf02ba68fb52165b8bf6ccd6fe71f098cac1fe59e" - }, - "downloads": -1, - "filename": "Django-4.0.6-py3-none-any.whl", - "has_sig": false, - "md5_digest": "8f9ef276f0c1b6f3f560d179c7e76d6a", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8041873, - "upload_time": "2022-07-04T07:57:23", - "upload_time_iso_8601": "2022-07-04T07:57:23.824880Z", - "url": "https://files.pythonhosted.org/packages/12/97/e341b13d605a9d0732a37975ce6bfae643c296721b8644f42d5d720c99bf/Django-4.0.6-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "a417b10aa26d7a566a3c19e9d29fac39c8643cbceb6cd7649a378d676839b5db", - "md5": "ad4e850c7110a45a6c7778d5bd01b85e", - "sha256": "a67a793ff6827fd373555537dca0da293a63a316fe34cb7f367f898ccca3c3ae" - }, - "downloads": -1, - "filename": "Django-4.0.6.tar.gz", - "has_sig": false, - "md5_digest": "ad4e850c7110a45a6c7778d5bd01b85e", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10389543, - "upload_time": "2022-07-04T07:57:34", - "upload_time_iso_8601": "2022-07-04T07:57:34.578441Z", - "url": "https://files.pythonhosted.org/packages/a4/17/b10aa26d7a566a3c19e9d29fac39c8643cbceb6cd7649a378d676839b5db/Django-4.0.6.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.0.7": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "e3419c6d5fe4e4ee6204ecf0cb96d843cca06be59b702d66afe5cf810508575e", - "md5": "9e6f021f50b1c40ea7071bbd37950485", - "sha256": "41bd65a9e5f8a89cdbfa7a7bba45cd7431ae89e750af82dea8a35fd1a7ecbe66" - }, - "downloads": -1, - "filename": "Django-4.0.7-py3-none-any.whl", - "has_sig": false, - "md5_digest": "9e6f021f50b1c40ea7071bbd37950485", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8041977, - "upload_time": "2022-08-03T07:38:18", - "upload_time_iso_8601": "2022-08-03T07:38:18.509869Z", - "url": "https://files.pythonhosted.org/packages/e3/41/9c6d5fe4e4ee6204ecf0cb96d843cca06be59b702d66afe5cf810508575e/Django-4.0.7-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "5c12a368d0caf91f1dab8154f1c579c5025c6645def5f3eee0d297347c7a1def", - "md5": "68abb56603fcc5c961083859bda8b04b", - "sha256": "9c6d5ad36be798e562ddcaa6b17b1c3ff2d3c4f529a47432b69fb9a30f847461" - }, - "downloads": -1, - "filename": "Django-4.0.7.tar.gz", - "has_sig": false, - "md5_digest": "68abb56603fcc5c961083859bda8b04b", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10407810, - "upload_time": "2022-08-03T07:38:26", - "upload_time_iso_8601": "2022-08-03T07:38:26.373780Z", - "url": "https://files.pythonhosted.org/packages/5c/12/a368d0caf91f1dab8154f1c579c5025c6645def5f3eee0d297347c7a1def/Django-4.0.7.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.0.8": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "e1d0d90528978da16288d470bb423abad307ed7ae724090132ff6bf67d6a5579", - "md5": "386349753c386e574dceca5067e2788a", - "sha256": "27cb08fa6458c1eff8b97c4c2d03774646fb26feeaa4587dca10c49e6d4fc6a3" - }, - "downloads": -1, - "filename": "Django-4.0.8-py3-none-any.whl", - "has_sig": false, - "md5_digest": "386349753c386e574dceca5067e2788a", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8041982, - "upload_time": "2022-10-04T07:54:20", - "upload_time_iso_8601": "2022-10-04T07:54:20.471830Z", - "url": "https://files.pythonhosted.org/packages/e1/d0/d90528978da16288d470bb423abad307ed7ae724090132ff6bf67d6a5579/Django-4.0.8-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "1ade08d8a349ed0e3e1999eb86ae0347cc9eaf634cd65f1eb80b9387ac1dbe3c", - "md5": "75ec07b3e00c79fd6e67fbee53786b7a", - "sha256": "07e6433f263c3839939cfabeb6d7557841e0419e47759a7b7d37f6d44d40adcb" - }, - "downloads": -1, - "filename": "Django-4.0.8.tar.gz", - "has_sig": false, - "md5_digest": "75ec07b3e00c79fd6e67fbee53786b7a", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10427857, - "upload_time": "2022-10-04T07:54:33", - "upload_time_iso_8601": "2022-10-04T07:54:33.062795Z", - "url": "https://files.pythonhosted.org/packages/1a/de/08d8a349ed0e3e1999eb86ae0347cc9eaf634cd65f1eb80b9387ac1dbe3c/Django-4.0.8.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.0.9": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "0f8a65b639fcf107cb77ef759ae540a53dc5c38496673c0b2884635c726ad23f", - "md5": "6cf14cc8da799c4d8236a8e310bbe41a", - "sha256": "6ddab9f9ac9211e2dc544819b87343b49c53f99d882e68727b38fd44204f752e" - }, - "downloads": -1, - "filename": "Django-4.0.9-py3-none-any.whl", - "has_sig": false, - "md5_digest": "6cf14cc8da799c4d8236a8e310bbe41a", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8042429, - "upload_time": "2023-02-01T09:55:52", - "upload_time_iso_8601": "2023-02-01T09:55:52.888766Z", - "url": "https://files.pythonhosted.org/packages/0f/8a/65b639fcf107cb77ef759ae540a53dc5c38496673c0b2884635c726ad23f/Django-4.0.9-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "69f0bf6010636003ac2bd1d222f1c5a29c4975318fc1931b132abd000993f4d6", - "md5": "c7676df3f34100403ea01cb368a4a4cd", - "sha256": "7cc034a72f99536c5feb39d1575f2952b86ffa409f109efeb2c8b398985f78a7" - }, - "downloads": -1, - "filename": "Django-4.0.9.tar.gz", - "has_sig": false, - "md5_digest": "c7676df3f34100403ea01cb368a4a4cd", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10410176, - "upload_time": "2023-02-01T09:56:06", - "upload_time_iso_8601": "2023-02-01T09:56:06.045214Z", - "url": "https://files.pythonhosted.org/packages/69/f0/bf6010636003ac2bd1d222f1c5a29c4975318fc1931b132abd000993f4d6/Django-4.0.9.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.0a1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "4f0dc346909f167ef9eb245ec189f37a993d7aeabf184d6e07920ca99ea74f1d", - "md5": "d116ffc74af4f4c6ce4b0c9d765efe14", - "sha256": "86b1cdeec72768defb39e27ca8c998e9e465888fd7663269a82cfe182b680c94" - }, - "downloads": -1, - "filename": "Django-4.0a1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "d116ffc74af4f4c6ce4b0c9d765efe14", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 7913692, - "upload_time": "2021-09-21T19:08:46", - "upload_time_iso_8601": "2021-09-21T19:08:46.532552Z", - "url": "https://files.pythonhosted.org/packages/4f/0d/c346909f167ef9eb245ec189f37a993d7aeabf184d6e07920ca99ea74f1d/Django-4.0a1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "a2621e6030f9ff01b7f2d36e11f1181beb51f4e181ce38eb8cd240e8bcb8201c", - "md5": "c7f86283e029d7953d00d7a1fb17501c", - "sha256": "010388b20c94e2ab421c6c4660ab835611be40e758679b93eb1f703a00cefaf3" - }, - "downloads": -1, - "filename": "Django-4.0a1.tar.gz", - "has_sig": false, - "md5_digest": "c7f86283e029d7953d00d7a1fb17501c", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 9896896, - "upload_time": "2021-09-21T19:09:14", - "upload_time_iso_8601": "2021-09-21T19:09:14.295332Z", - "url": "https://files.pythonhosted.org/packages/a2/62/1e6030f9ff01b7f2d36e11f1181beb51f4e181ce38eb8cd240e8bcb8201c/Django-4.0a1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.0b1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "e993a5e3bbd3a6311b4d224da3ab9c7ee08f888a2a977d63aed0c13403610f20", - "md5": "79fa02994b7c19801aabcebb6a56b38b", - "sha256": "12220029292e1c7dce4ea0ad9189942d3204f98f07b6fcbfd0a64c3cc49c0f01" - }, - "downloads": -1, - "filename": "Django-4.0b1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "79fa02994b7c19801aabcebb6a56b38b", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 7914396, - "upload_time": "2021-10-25T09:23:16", - "upload_time_iso_8601": "2021-10-25T09:23:16.638876Z", - "url": "https://files.pythonhosted.org/packages/e9/93/a5e3bbd3a6311b4d224da3ab9c7ee08f888a2a977d63aed0c13403610f20/Django-4.0b1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "21dd5b010156289b277f8f6adda3a18bfcc0a1842ad9a2f9951190d1a3b7bd71", - "md5": "80d2453c7754304c2fb89ce6f9305e17", - "sha256": "135d854f678d6692b4899cbe1b90dd905f60c8fafb13730ce0ac76d4a8c7e2e4" - }, - "downloads": -1, - "filename": "Django-4.0b1.tar.gz", - "has_sig": false, - "md5_digest": "80d2453c7754304c2fb89ce6f9305e17", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 9902026, - "upload_time": "2021-10-25T09:23:22", - "upload_time_iso_8601": "2021-10-25T09:23:22.644895Z", - "url": "https://files.pythonhosted.org/packages/21/dd/5b010156289b277f8f6adda3a18bfcc0a1842ad9a2f9951190d1a3b7bd71/Django-4.0b1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.0rc1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "6869d3f879ec1164568dac7f336eb9d1bfb6e343299bbcf1778fb45737050733", - "md5": "6cc32a61a53ae01e6ba2763f86680732", - "sha256": "d4f007daf8a3834a75026653dee196f0ccbfd1a5b1bc1dafcf10b0727a76c825" - }, - "downloads": -1, - "filename": "Django-4.0rc1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "6cc32a61a53ae01e6ba2763f86680732", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 7915672, - "upload_time": "2021-11-22T06:37:57", - "upload_time_iso_8601": "2021-11-22T06:37:57.726504Z", - "url": "https://files.pythonhosted.org/packages/68/69/d3f879ec1164568dac7f336eb9d1bfb6e343299bbcf1778fb45737050733/Django-4.0rc1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "bbd974b87456553afbf7098c251514736d3575679cc32a9b4430dffaaf0a5d20", - "md5": "2668bcd211127307de3b809b3dba8be1", - "sha256": "fedcdff5ad460169b61cf9597269f265273ae83da349c79b80995ca5c2a3cd92" - }, - "downloads": -1, - "filename": "Django-4.0rc1.tar.gz", - "has_sig": false, - "md5_digest": "2668bcd211127307de3b809b3dba8be1", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 9905788, - "upload_time": "2021-11-22T06:38:01", - "upload_time_iso_8601": "2021-11-22T06:38:01.843947Z", - "url": "https://files.pythonhosted.org/packages/bb/d9/74b87456553afbf7098c251514736d3575679cc32a9b4430dffaaf0a5d20/Django-4.0rc1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "9b41e1e7d6ecc3bc76681dfdc6b373566822bc2aab96fa3eceaaed70accc28b6", - "md5": "940ab96fc1f61c12430d27867d72fccc", - "sha256": "031ccb717782f6af83a0063a1957686e87cb4581ea61b47b3e9addf60687989a" - }, - "downloads": -1, - "filename": "Django-4.1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "940ab96fc1f61c12430d27867d72fccc", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8095228, - "upload_time": "2022-08-03T08:40:20", - "upload_time_iso_8601": "2022-08-03T08:40:20.936536Z", - "url": "https://files.pythonhosted.org/packages/9b/41/e1e7d6ecc3bc76681dfdc6b373566822bc2aab96fa3eceaaed70accc28b6/Django-4.1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "1366757b8f4f4df4b1c1becad346c3a6401ae9fddcc78aa3741d231be657cf94", - "md5": "9ef83433b44e798a3bdc0b15a56e0f80", - "sha256": "032f8a6fc7cf05ccd1214e4a2e21dfcd6a23b9d575c6573cacc8c67828dbe642" - }, - "downloads": -1, - "filename": "Django-4.1.tar.gz", - "has_sig": false, - "md5_digest": "9ef83433b44e798a3bdc0b15a56e0f80", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10484595, - "upload_time": "2022-08-03T08:40:25", - "upload_time_iso_8601": "2022-08-03T08:40:25.070462Z", - "url": "https://files.pythonhosted.org/packages/13/66/757b8f4f4df4b1c1becad346c3a6401ae9fddcc78aa3741d231be657cf94/Django-4.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.1.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "805c884f9fed747679a1f70321465e664f3dc1602e8b7a6e96a21163894018b3", - "md5": "536c81a84a4a2bccc9a00f3d4640f0c4", - "sha256": "acb21fac9275f9972d81c7caf5761a89ec3ea25fe74545dd26b8a48cb3a0203e" - }, - "downloads": -1, - "filename": "Django-4.1.1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "536c81a84a4a2bccc9a00f3d4640f0c4", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8094775, - "upload_time": "2022-09-05T05:02:26", - "upload_time_iso_8601": "2022-09-05T05:02:26.425285Z", - "url": "https://files.pythonhosted.org/packages/80/5c/884f9fed747679a1f70321465e664f3dc1602e8b7a6e96a21163894018b3/Django-4.1.1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "ca7849ac0899f74c485acbf96bc8a5ea6a87575edb06faa9911057cacd505ba9", - "md5": "f2b2516279eacdfe6098b7b9d905a25e", - "sha256": "a153ffd5143bf26a877bfae2f4ec736ebd8924a46600ca089ad96b54a1d4e28e" - }, - "downloads": -1, - "filename": "Django-4.1.1.tar.gz", - "has_sig": false, - "md5_digest": "f2b2516279eacdfe6098b7b9d905a25e", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10474304, - "upload_time": "2022-09-05T05:02:34", - "upload_time_iso_8601": "2022-09-05T05:02:34.094711Z", - "url": "https://files.pythonhosted.org/packages/ca/78/49ac0899f74c485acbf96bc8a5ea6a87575edb06faa9911057cacd505ba9/Django-4.1.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.1.10": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "34258a218de57fc9853297a1a8e4927688eff8107d5bc6dcf6c964c59801f036", - "md5": "32e79f91cc1db9baac92dc3d5faf72d3", - "sha256": "26d0260c2fb8121009e62ffc548b2398dea2522b6454208a852fb0ef264c206c" - }, - "downloads": -1, - "filename": "Django-4.1.10-py3-none-any.whl", - "has_sig": false, - "md5_digest": "32e79f91cc1db9baac92dc3d5faf72d3", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8103933, - "upload_time": "2023-07-03T07:57:16", - "upload_time_iso_8601": "2023-07-03T07:57:16.588462Z", - "url": "https://files.pythonhosted.org/packages/34/25/8a218de57fc9853297a1a8e4927688eff8107d5bc6dcf6c964c59801f036/Django-4.1.10-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "70d4eded564fa5928f68771d082ec0eef4d023f9d19dfa1d2923305bc3e62afe", - "md5": "3720c85a8c25cacbce2f95d345d0f5ad", - "sha256": "56343019a9fd839e2e5bf203daf45f25af79d5bffa4c71d56eae4f4404d82ade" - }, - "downloads": -1, - "filename": "Django-4.1.10.tar.gz", - "has_sig": false, - "md5_digest": "3720c85a8c25cacbce2f95d345d0f5ad", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10513572, - "upload_time": "2023-07-03T07:57:30", - "upload_time_iso_8601": "2023-07-03T07:57:30.301266Z", - "url": "https://files.pythonhosted.org/packages/70/d4/eded564fa5928f68771d082ec0eef4d023f9d19dfa1d2923305bc3e62afe/Django-4.1.10.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.1.11": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "389f144a9293f86886f2e85c9bd981c8c28597fd15b1a59431db1921e7c1938b", - "md5": "1ddfb0397a98cccb1b1ff48d35543a42", - "sha256": "cac9df0ba87b4f439e1a311ef22f75c938fc874bebf1fbabaed58d0e6d559a25" - }, - "downloads": -1, - "filename": "Django-4.1.11-py3-none-any.whl", - "has_sig": false, - "md5_digest": "1ddfb0397a98cccb1b1ff48d35543a42", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8103954, - "upload_time": "2023-09-04T10:58:19", - "upload_time_iso_8601": "2023-09-04T10:58:19.024013Z", - "url": "https://files.pythonhosted.org/packages/38/9f/144a9293f86886f2e85c9bd981c8c28597fd15b1a59431db1921e7c1938b/Django-4.1.11-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "64cbc0570752faba4c8a85996e78eb7699b9d29ee32cb7aeb5e262582c53bad0", - "md5": "2e2deedadcf752e05f2d60a32e5da943", - "sha256": "7b134688965dd331ca4d11ed38e5ce594caed0e906689a9b95251c29c2c03990" - }, - "downloads": -1, - "filename": "Django-4.1.11.tar.gz", - "has_sig": false, - "md5_digest": "2e2deedadcf752e05f2d60a32e5da943", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10511539, - "upload_time": "2023-09-04T10:58:30", - "upload_time_iso_8601": "2023-09-04T10:58:30.124274Z", - "url": "https://files.pythonhosted.org/packages/64/cb/c0570752faba4c8a85996e78eb7699b9d29ee32cb7aeb5e262582c53bad0/Django-4.1.11.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.1.12": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "0dc7cfdf590422a0f7a6a6ae6a4cbb2721456141662af4b203f583286f0dcd62", - "md5": "ba3538987deb35e7843bf334b6c1cfce", - "sha256": "e92ce8f240a856615e96d8b955707f824c29ea0f51dff4f76777caa5e113ec72" - }, - "downloads": -1, - "filename": "Django-4.1.12-py3-none-any.whl", - "has_sig": false, - "md5_digest": "ba3538987deb35e7843bf334b6c1cfce", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8104099, - "upload_time": "2023-10-04T14:59:03", - "upload_time_iso_8601": "2023-10-04T14:59:03.784687Z", - "url": "https://files.pythonhosted.org/packages/0d/c7/cfdf590422a0f7a6a6ae6a4cbb2721456141662af4b203f583286f0dcd62/Django-4.1.12-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "6876b1cf87eda3bf5b2d7dd9a23bb4df70ea5511e426a54a04ed2c4ca2ff67cd", - "md5": "6f95e92e5e8964c4a1e00e2fcaadf437", - "sha256": "d02483ad49872238fa59875c1269293fe4f17ecee13c121893607cc0b284696b" - }, - "downloads": -1, - "filename": "Django-4.1.12.tar.gz", - "has_sig": false, - "md5_digest": "6f95e92e5e8964c4a1e00e2fcaadf437", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10502350, - "upload_time": "2023-10-04T14:59:09", - "upload_time_iso_8601": "2023-10-04T14:59:09.851212Z", - "url": "https://files.pythonhosted.org/packages/68/76/b1cf87eda3bf5b2d7dd9a23bb4df70ea5511e426a54a04ed2c4ca2ff67cd/Django-4.1.12.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.1.13": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "adac203ce7fd2e03fb0392e0a6160f4b27d0d1efe1614cb65f1d29e6bbf2ef69", - "md5": "70624148d45a821b760811a7e8eb45bc", - "sha256": "04ab3f6f46d084a0bba5a2c9a93a3a2eb3fe81589512367a75f79ee8acf790ce" - }, - "downloads": -1, - "filename": "Django-4.1.13-py3-none-any.whl", - "has_sig": false, - "md5_digest": "70624148d45a821b760811a7e8eb45bc", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8104299, - "upload_time": "2023-11-01T06:59:10", - "upload_time_iso_8601": "2023-11-01T06:59:10.387459Z", - "url": "https://files.pythonhosted.org/packages/ad/ac/203ce7fd2e03fb0392e0a6160f4b27d0d1efe1614cb65f1d29e6bbf2ef69/Django-4.1.13-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "3409300498ca0a5f37e0621b90a94c71fc8a4227be5488a3bd5550869c6e0049", - "md5": "a16208af2aa54cbe97ff79ec4426da84", - "sha256": "94a3f471e833c8f124ee7a2de11e92f633991d975e3fa5bdd91e8abd66426318" - }, - "downloads": -1, - "filename": "Django-4.1.13.tar.gz", - "has_sig": false, - "md5_digest": "a16208af2aa54cbe97ff79ec4426da84", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10515104, - "upload_time": "2023-11-01T06:59:24", - "upload_time_iso_8601": "2023-11-01T06:59:24.955376Z", - "url": "https://files.pythonhosted.org/packages/34/09/300498ca0a5f37e0621b90a94c71fc8a4227be5488a3bd5550869c6e0049/Django-4.1.13.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.1.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "9102786ced9f0b9980670f46f7563f9c5494a24e3dd1920cc7be6cc7ac377389", - "md5": "c1f6ff2e6a2c0ca5a1d4eb49e476222f", - "sha256": "26dc24f99c8956374a054bcbf58aab8dc0cad2e6ac82b0fe036b752c00eee793" - }, - "downloads": -1, - "filename": "Django-4.1.2-py3-none-any.whl", - "has_sig": false, - "md5_digest": "c1f6ff2e6a2c0ca5a1d4eb49e476222f", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8094720, - "upload_time": "2022-10-04T07:54:23", - "upload_time_iso_8601": "2022-10-04T07:54:23.273959Z", - "url": "https://files.pythonhosted.org/packages/91/02/786ced9f0b9980670f46f7563f9c5494a24e3dd1920cc7be6cc7ac377389/Django-4.1.2-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "0928fe1484dfc2e984543a07d5444006122f6cc451fe352bce1d85edcf81a05f", - "md5": "92371e5a9e99f70629b52973563f106a", - "sha256": "b8d843714810ab88d59344507d4447be8b2cf12a49031363b6eed9f1b9b2280f" - }, - "downloads": -1, - "filename": "Django-4.1.2.tar.gz", - "has_sig": false, - "md5_digest": "92371e5a9e99f70629b52973563f106a", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10509111, - "upload_time": "2022-10-04T07:54:38", - "upload_time_iso_8601": "2022-10-04T07:54:38.403977Z", - "url": "https://files.pythonhosted.org/packages/09/28/fe1484dfc2e984543a07d5444006122f6cc451fe352bce1d85edcf81a05f/Django-4.1.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.1.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "4fbecd28514516e66c40b87e487be68cf61a1637bd3ec60a2db90bb5075a2df5", - "md5": "7e85acf8072ad476e3d07c9039af1189", - "sha256": "6b1de6886cae14c7c44d188f580f8ba8da05750f544c80ae5ad43375ab293cd5" - }, - "downloads": -1, - "filename": "Django-4.1.3-py3-none-any.whl", - "has_sig": false, - "md5_digest": "7e85acf8072ad476e3d07c9039af1189", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8094697, - "upload_time": "2022-11-01T06:18:15", - "upload_time_iso_8601": "2022-11-01T06:18:15.550215Z", - "url": "https://files.pythonhosted.org/packages/4f/be/cd28514516e66c40b87e487be68cf61a1637bd3ec60a2db90bb5075a2df5/Django-4.1.3-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "d0ab33f9c2bd4cf7ff2d319131eca7231247f962f781ebc8a012ebe02582244a", - "md5": "60c54c37ecf9e5e635ee6bb07a13a9fa", - "sha256": "678bbfc8604eb246ed54e2063f0765f13b321a50526bdc8cb1f943eda7fa31f1" - }, - "downloads": -1, - "filename": "Django-4.1.3.tar.gz", - "has_sig": false, - "md5_digest": "60c54c37ecf9e5e635ee6bb07a13a9fa", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10476258, - "upload_time": "2022-11-01T06:18:21", - "upload_time_iso_8601": "2022-11-01T06:18:21.116130Z", - "url": "https://files.pythonhosted.org/packages/d0/ab/33f9c2bd4cf7ff2d319131eca7231247f962f781ebc8a012ebe02582244a/Django-4.1.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.1.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "212628895838228c46ece278d764720995d5a51e4bce7d02d2a54e70f59108e1", - "md5": "0f8cb5b0877ad67c0324c335273ab4b1", - "sha256": "0b223bfa55511f950ff741983d408d78d772351284c75e9f77d2b830b6b4d148" - }, - "downloads": -1, - "filename": "Django-4.1.4-py3-none-any.whl", - "has_sig": false, - "md5_digest": "0f8cb5b0877ad67c0324c335273ab4b1", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8094989, - "upload_time": "2022-12-06T09:16:48", - "upload_time_iso_8601": "2022-12-06T09:16:48.876036Z", - "url": "https://files.pythonhosted.org/packages/21/26/28895838228c46ece278d764720995d5a51e4bce7d02d2a54e70f59108e1/Django-4.1.4-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "de7dca5d224f7d345b0defd0a57c6deb0c6115fbe5304e4d9571f728b617f1a3", - "md5": "02e2222d064d800f37c6a0ba78ce87f3", - "sha256": "d38a4e108d2386cb9637da66a82dc8d0733caede4c83c4afdbda78af4214211b" - }, - "downloads": -1, - "filename": "Django-4.1.4.tar.gz", - "has_sig": false, - "md5_digest": "02e2222d064d800f37c6a0ba78ce87f3", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10517427, - "upload_time": "2022-12-06T09:16:52", - "upload_time_iso_8601": "2022-12-06T09:16:52.386734Z", - "url": "https://files.pythonhosted.org/packages/de/7d/ca5d224f7d345b0defd0a57c6deb0c6115fbe5304e4d9571f728b617f1a3/Django-4.1.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.1.5": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "2dac9f013a51e6008ba94a282c15778a3ea51a0953f6711a77f9baa471fd1b1d", - "md5": "6003ee8ac3f8b6b30e6fe16bb9b3c8c1", - "sha256": "4b214a05fe4c99476e99e2445c8b978c8369c18d4dea8e22ec412862715ad763" - }, - "downloads": -1, - "filename": "Django-4.1.5-py3-none-any.whl", - "has_sig": false, - "md5_digest": "6003ee8ac3f8b6b30e6fe16bb9b3c8c1", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8102437, - "upload_time": "2023-01-02T07:34:45", - "upload_time_iso_8601": "2023-01-02T07:34:45.541744Z", - "url": "https://files.pythonhosted.org/packages/2d/ac/9f013a51e6008ba94a282c15778a3ea51a0953f6711a77f9baa471fd1b1d/Django-4.1.5-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "41c8b3c469353f9d1b7f0e99b45520582b891da02cd87408bc867affa6e039a3", - "md5": "6b6a47d7478ead15b0a0835841c433dc", - "sha256": "ff56ebd7ead0fd5dbe06fe157b0024a7aaea2e0593bb3785fb594cf94dad58ef" - }, - "downloads": -1, - "filename": "Django-4.1.5.tar.gz", - "has_sig": false, - "md5_digest": "6b6a47d7478ead15b0a0835841c433dc", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10507020, - "upload_time": "2023-01-02T07:34:49", - "upload_time_iso_8601": "2023-01-02T07:34:49.562776Z", - "url": "https://files.pythonhosted.org/packages/41/c8/b3c469353f9d1b7f0e99b45520582b891da02cd87408bc867affa6e039a3/Django-4.1.5.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.1.6": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "5cbe8800ef92dbe66017e78fb9589e16333e0a87f67068485dbfb8114a883b67", - "md5": "ceb0fce2e511c6a6be83d6569ddfcbb5", - "sha256": "c6fe7ebe7c017fe59f1029821dae0acb5a2ddcd6c9a0138fd20a8bfefac914bc" - }, - "downloads": -1, - "filename": "Django-4.1.6-py3-none-any.whl", - "has_sig": false, - "md5_digest": "ceb0fce2e511c6a6be83d6569ddfcbb5", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8102942, - "upload_time": "2023-02-01T09:55:56", - "upload_time_iso_8601": "2023-02-01T09:55:56.855590Z", - "url": "https://files.pythonhosted.org/packages/5c/be/8800ef92dbe66017e78fb9589e16333e0a87f67068485dbfb8114a883b67/Django-4.1.6-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "4fb01378fefb4b0da51e27120e0069e51fa7b3df2078f1250393eccefc274376", - "md5": "4aed2866d732727590e1008dd871e3a3", - "sha256": "bceb0fe1a386781af0788cae4108622756cd05e7775448deec04a71ddf87685d" - }, - "downloads": -1, - "filename": "Django-4.1.6.tar.gz", - "has_sig": false, - "md5_digest": "4aed2866d732727590e1008dd871e3a3", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10496749, - "upload_time": "2023-02-01T09:56:11", - "upload_time_iso_8601": "2023-02-01T09:56:11.799732Z", - "url": "https://files.pythonhosted.org/packages/4f/b0/1378fefb4b0da51e27120e0069e51fa7b3df2078f1250393eccefc274376/Django-4.1.6.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.1.7": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "898659e237f7176cfc1544446914fa329fd560bb8fce46be52dd7af5dc7c54f9", - "md5": "cee5f52df055e8022f1897d4c7eb5665", - "sha256": "f2f431e75adc40039ace496ad3b9f17227022e8b11566f4b363da44c7e44761e" - }, - "downloads": -1, - "filename": "Django-4.1.7-py3-none-any.whl", - "has_sig": false, - "md5_digest": "cee5f52df055e8022f1897d4c7eb5665", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8103455, - "upload_time": "2023-02-14T08:25:36", - "upload_time_iso_8601": "2023-02-14T08:25:36.578599Z", - "url": "https://files.pythonhosted.org/packages/89/86/59e237f7176cfc1544446914fa329fd560bb8fce46be52dd7af5dc7c54f9/Django-4.1.7-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "9fa707939866241b7e8f8d3bf164b7d6ad428163723e29dd472700f8ab0e5fd5", - "md5": "626f96c63ddfab24bab90d80c87a7aad", - "sha256": "44f714b81c5f190d9d2ddad01a532fe502fa01c4cb8faf1d081f4264ed15dcd8" - }, - "downloads": -1, - "filename": "Django-4.1.7.tar.gz", - "has_sig": false, - "md5_digest": "626f96c63ddfab24bab90d80c87a7aad", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10520415, - "upload_time": "2023-02-14T08:25:50", - "upload_time_iso_8601": "2023-02-14T08:25:50.105773Z", - "url": "https://files.pythonhosted.org/packages/9f/a7/07939866241b7e8f8d3bf164b7d6ad428163723e29dd472700f8ab0e5fd5/Django-4.1.7.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.1.8": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "bb3280134d57b584ed8b03e9a52da5860b8ac7d1210431ac25d8d83389006edc", - "md5": "6ce82d8f699b73598aee2608afa48c52", - "sha256": "dba9dd0bf8b748aa9c98d6c679f8b1c0bfa43124844bea9425ad9ba0cd5e65c3" - }, - "downloads": -1, - "filename": "Django-4.1.8-py3-none-any.whl", - "has_sig": false, - "md5_digest": "6ce82d8f699b73598aee2608afa48c52", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8103668, - "upload_time": "2023-04-05T06:11:04", - "upload_time_iso_8601": "2023-04-05T06:11:04.460008Z", - "url": "https://files.pythonhosted.org/packages/bb/32/80134d57b584ed8b03e9a52da5860b8ac7d1210431ac25d8d83389006edc/Django-4.1.8-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "63bc03656c61283badfbef3be182bbfdd85dbfd72c188377d73074c9164c49cf", - "md5": "c0d042a9baae4c42a449af8d2f9b3e7a", - "sha256": "43253f4b4a2c3a1694cd52d3517847271934c84dbd060ca0e82a9f90e569dab3" - }, - "downloads": -1, - "filename": "Django-4.1.8.tar.gz", - "has_sig": false, - "md5_digest": "c0d042a9baae4c42a449af8d2f9b3e7a", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10511976, - "upload_time": "2023-04-05T06:11:09", - "upload_time_iso_8601": "2023-04-05T06:11:09.369362Z", - "url": "https://files.pythonhosted.org/packages/63/bc/03656c61283badfbef3be182bbfdd85dbfd72c188377d73074c9164c49cf/Django-4.1.8.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.1.9": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "7e4ccbdaa1488ae5207e6eab912ff0d23f5d786676efbb55f21e7f3cfe98055a", - "md5": "2d9d8e704a67112e9179730fe81ea56d", - "sha256": "adae3a952fd86800094ae6f64aa558572e8b4ba8dfe21f0ed8175147e75a72a1" - }, - "downloads": -1, - "filename": "Django-4.1.9-py3-none-any.whl", - "has_sig": false, - "md5_digest": "2d9d8e704a67112e9179730fe81ea56d", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8103809, - "upload_time": "2023-05-03T12:58:23", - "upload_time_iso_8601": "2023-05-03T12:58:23.438294Z", - "url": "https://files.pythonhosted.org/packages/7e/4c/cbdaa1488ae5207e6eab912ff0d23f5d786676efbb55f21e7f3cfe98055a/Django-4.1.9-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "11ea8b514434c57c3bef89a475b75f74d768471d8e1bc61f4e5c79daeae9b5ef", - "md5": "1a6f4e5318e3272deaa9cfd61e252fab", - "sha256": "e9f074a84930662104871bfcea55c3c180c50a0a47739db82435deae6cbaf032" - }, - "downloads": -1, - "filename": "Django-4.1.9.tar.gz", - "has_sig": false, - "md5_digest": "1a6f4e5318e3272deaa9cfd61e252fab", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10514944, - "upload_time": "2023-05-03T12:58:36", - "upload_time_iso_8601": "2023-05-03T12:58:36.244311Z", - "url": "https://files.pythonhosted.org/packages/11/ea/8b514434c57c3bef89a475b75f74d768471d8e1bc61f4e5c79daeae9b5ef/Django-4.1.9.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.1a1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "a49f7cd1364cddd43621e5bb95bc7926e036d567a8a6aec507e9fda175702499", - "md5": "11c720f9043eaa841112871e81c1e3e0", - "sha256": "815b1c3e1c1379cca1c54cc7202ee86df006f52b94560b46ea0b19f162c9474a" - }, - "downloads": -1, - "filename": "Django-4.1a1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "11c720f9043eaa841112871e81c1e3e0", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8070898, - "upload_time": "2022-05-18T05:54:32", - "upload_time_iso_8601": "2022-05-18T05:54:32.427903Z", - "url": "https://files.pythonhosted.org/packages/a4/9f/7cd1364cddd43621e5bb95bc7926e036d567a8a6aec507e9fda175702499/Django-4.1a1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "12ef6c7373ba910e09de8a28a124c772314e49dd2faf76be51eba36e5878f636", - "md5": "81c44d4f90305b6255d8a2f61e25d8d2", - "sha256": "91980d3b327e38e44c2bef0af86d624a57adbd5f1c96f43acf150396770caf01" - }, - "downloads": -1, - "filename": "Django-4.1a1.tar.gz", - "has_sig": false, - "md5_digest": "81c44d4f90305b6255d8a2f61e25d8d2", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10465163, - "upload_time": "2022-05-18T05:54:38", - "upload_time_iso_8601": "2022-05-18T05:54:38.575881Z", - "url": "https://files.pythonhosted.org/packages/12/ef/6c7373ba910e09de8a28a124c772314e49dd2faf76be51eba36e5878f636/Django-4.1a1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.1b1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "ffa5e78d8b9d3dda424cc793fe3e265f2f0411bdd59ede7eaf3f2aa732db265c", - "md5": "928fb36c328a4db7e8f8384eaf42188e", - "sha256": "55ea3e07df2eb3994b7261d743c111e4c27a56ca1285eb762dba07432ebe4039" - }, - "downloads": -1, - "filename": "Django-4.1b1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "928fb36c328a4db7e8f8384eaf42188e", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8071810, - "upload_time": "2022-06-21T09:20:00", - "upload_time_iso_8601": "2022-06-21T09:20:00.319542Z", - "url": "https://files.pythonhosted.org/packages/ff/a5/e78d8b9d3dda424cc793fe3e265f2f0411bdd59ede7eaf3f2aa732db265c/Django-4.1b1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "d4b803307e023bbae4184dfb1ed6f264b91b579a3099133c219992f7e7757a2b", - "md5": "47caaf146c9abf90c1026e6bd4f7c7a7", - "sha256": "ec22c3e909af402281ef1dea69cbbc42b43b9ac9379df2914e0a7b85597242d6" - }, - "downloads": -1, - "filename": "Django-4.1b1.tar.gz", - "has_sig": false, - "md5_digest": "47caaf146c9abf90c1026e6bd4f7c7a7", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10466194, - "upload_time": "2022-06-21T09:20:06", - "upload_time_iso_8601": "2022-06-21T09:20:06.847874Z", - "url": "https://files.pythonhosted.org/packages/d4/b8/03307e023bbae4184dfb1ed6f264b91b579a3099133c219992f7e7757a2b/Django-4.1b1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.1rc1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "e7c7e6020d17a451c9be06615ec4d9f0136bfc5a0811aa07363591a6460f1018", - "md5": "21d46263078c5b8f066b49457a0c4a6e", - "sha256": "04182903242268f0dbacb0175bf47a74c477761502f9f7427142f459c006217d" - }, - "downloads": -1, - "filename": "Django-4.1rc1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "21d46263078c5b8f066b49457a0c4a6e", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 8072348, - "upload_time": "2022-07-19T09:02:03", - "upload_time_iso_8601": "2022-07-19T09:02:03.575740Z", - "url": "https://files.pythonhosted.org/packages/e7/c7/e6020d17a451c9be06615ec4d9f0136bfc5a0811aa07363591a6460f1018/Django-4.1rc1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "b889cf535c221e477545bc9f6fe00f644bb0bd51948dc22b0c7cc066ba9cd899", - "md5": "2b89f0b38819210fb189fcef99dc43ea", - "sha256": "3c4204490ad3a7252873bc24c6aa2a32375baa06b4c8beb9b9cfda1f986741de" - }, - "downloads": -1, - "filename": "Django-4.1rc1.tar.gz", - "has_sig": false, - "md5_digest": "2b89f0b38819210fb189fcef99dc43ea", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10472149, - "upload_time": "2022-07-19T09:02:07", - "upload_time_iso_8601": "2022-07-19T09:02:07.093043Z", - "url": "https://files.pythonhosted.org/packages/b8/89/cf535c221e477545bc9f6fe00f644bb0bd51948dc22b0c7cc066ba9cd899/Django-4.1rc1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "d9406012f98b14b64b4d3dc47b0c2f116fccbd0795ab35515d0c40dac73b81b8", - "md5": "4b079efb729678cb02e6203302c52a74", - "sha256": "ad33ed68db9398f5dfb33282704925bce044bef4261cd4fb59e4e7f9ae505a78" - }, - "downloads": -1, - "filename": "Django-4.2-py3-none-any.whl", - "has_sig": false, - "md5_digest": "4b079efb729678cb02e6203302c52a74", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 7988617, - "upload_time": "2023-04-03T08:36:11", - "upload_time_iso_8601": "2023-04-03T08:36:11.797877Z", - "url": "https://files.pythonhosted.org/packages/d9/40/6012f98b14b64b4d3dc47b0c2f116fccbd0795ab35515d0c40dac73b81b8/Django-4.2-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "9abb48aa3e0850923096dff2766d21a6004d6e1a3317f0bd400ba81f586754e1", - "md5": "20a7a89579493c5f264900a5dbb848aa", - "sha256": "c36e2ab12824e2ac36afa8b2515a70c53c7742f0d6eaefa7311ec379558db997" - }, - "downloads": -1, - "filename": "Django-4.2.tar.gz", - "has_sig": false, - "md5_digest": "20a7a89579493c5f264900a5dbb848aa", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10415665, - "upload_time": "2023-04-03T08:36:16", - "upload_time_iso_8601": "2023-04-03T08:36:16.829178Z", - "url": "https://files.pythonhosted.org/packages/9a/bb/48aa3e0850923096dff2766d21a6004d6e1a3317f0bd400ba81f586754e1/Django-4.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.2.1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "121378e8622180f101e95297965045ff1325ea7301c1b80f756debbeaa84c3be", - "md5": "a20ab3eae12ee9113b96dd889343548c", - "sha256": "066b6debb5ac335458d2a713ed995570536c8b59a580005acb0732378d5eb1ee" - }, - "downloads": -1, - "filename": "Django-4.2.1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "a20ab3eae12ee9113b96dd889343548c", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 7988496, - "upload_time": "2023-05-03T12:58:27", - "upload_time_iso_8601": "2023-05-03T12:58:27.208245Z", - "url": "https://files.pythonhosted.org/packages/12/13/78e8622180f101e95297965045ff1325ea7301c1b80f756debbeaa84c3be/Django-4.2.1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "897623ee9b9d2bd4119e930eb19164732b79c0a4f6259ca198209b0fe36551ea", - "md5": "8a047b5d96d7a2b7a173f56ca9869e8a", - "sha256": "7efa6b1f781a6119a10ac94b4794ded90db8accbe7802281cd26f8664ffed59c" - }, - "downloads": -1, - "filename": "Django-4.2.1.tar.gz", - "has_sig": false, - "md5_digest": "8a047b5d96d7a2b7a173f56ca9869e8a", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10420051, - "upload_time": "2023-05-03T12:58:41", - "upload_time_iso_8601": "2023-05-03T12:58:41.313440Z", - "url": "https://files.pythonhosted.org/packages/89/76/23ee9b9d2bd4119e930eb19164732b79c0a4f6259ca198209b0fe36551ea/Django-4.2.1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.2.2": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "baf37a66888bda4017198a692887bd566123732783073c1fd424db15358525fd", - "md5": "5bff9e117350e5d80fd10209a1c38efd", - "sha256": "672b3fa81e1f853bb58be1b51754108ab4ffa12a77c06db86aa8df9ed0c46fe5" - }, - "downloads": -1, - "filename": "Django-4.2.2-py3-none-any.whl", - "has_sig": false, - "md5_digest": "5bff9e117350e5d80fd10209a1c38efd", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 7989179, - "upload_time": "2023-06-05T14:09:31", - "upload_time_iso_8601": "2023-06-05T14:09:31.615719Z", - "url": "https://files.pythonhosted.org/packages/ba/f3/7a66888bda4017198a692887bd566123732783073c1fd424db15358525fd/Django-4.2.2-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "5cfa427fbcac5633f4b88fda7953efb04db903ca7e6d9486afdbda525c4006cc", - "md5": "17feee0fb1ddfcb0144c6362fd53ffca", - "sha256": "2a6b6fbff5b59dd07bef10bcb019bee2ea97a30b2a656d51346596724324badf" - }, - "downloads": -1, - "filename": "Django-4.2.2.tar.gz", - "has_sig": false, - "md5_digest": "17feee0fb1ddfcb0144c6362fd53ffca", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10390518, - "upload_time": "2023-06-05T14:09:38", - "upload_time_iso_8601": "2023-06-05T14:09:38.470129Z", - "url": "https://files.pythonhosted.org/packages/5c/fa/427fbcac5633f4b88fda7953efb04db903ca7e6d9486afdbda525c4006cc/Django-4.2.2.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.2.3": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "d483227ebf197e413f3599cea96dddc7d6b8ff220310cc5b40dd0f1a15e5a9d1", - "md5": "d2a21cc0e28c11e25c852de08c9bbb24", - "sha256": "f7c7852a5ac5a3da5a8d5b35cc6168f31b605971441798dac845f17ca8028039" - }, - "downloads": -1, - "filename": "Django-4.2.3-py3-none-any.whl", - "has_sig": false, - "md5_digest": "d2a21cc0e28c11e25c852de08c9bbb24", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 7989953, - "upload_time": "2023-07-03T07:57:20", - "upload_time_iso_8601": "2023-07-03T07:57:20.221899Z", - "url": "https://files.pythonhosted.org/packages/d4/83/227ebf197e413f3599cea96dddc7d6b8ff220310cc5b40dd0f1a15e5a9d1/Django-4.2.3-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "3624d0e78e667f98efcca76c8b670ef247583349a8f5241cdb3c98eeb92726ff", - "md5": "eaa70abe96b6e6b50ef297531c365265", - "sha256": "45a747e1c5b3d6df1b141b1481e193b033fd1fdbda3ff52677dc81afdaacbaed" - }, - "downloads": -1, - "filename": "Django-4.2.3.tar.gz", - "has_sig": false, - "md5_digest": "eaa70abe96b6e6b50ef297531c365265", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10419003, - "upload_time": "2023-07-03T07:57:37", - "upload_time_iso_8601": "2023-07-03T07:57:37.448508Z", - "url": "https://files.pythonhosted.org/packages/36/24/d0e78e667f98efcca76c8b670ef247583349a8f5241cdb3c98eeb92726ff/Django-4.2.3.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.2.4": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "7f9efc6bab255ae10bc57fa2f65646eace3d5405fbb7f5678b90140052d1db0f", - "md5": "e74396ac1359a3f42eac1568807572c3", - "sha256": "860ae6a138a238fc4f22c99b52f3ead982bb4b1aad8c0122bcd8c8a3a02e409d" - }, - "downloads": -1, - "filename": "Django-4.2.4-py3-none-any.whl", - "has_sig": false, - "md5_digest": "e74396ac1359a3f42eac1568807572c3", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 7990073, - "upload_time": "2023-08-01T17:30:17", - "upload_time_iso_8601": "2023-08-01T17:30:17.903829Z", - "url": "https://files.pythonhosted.org/packages/7f/9e/fc6bab255ae10bc57fa2f65646eace3d5405fbb7f5678b90140052d1db0f/Django-4.2.4-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "8dd2a4f823741a12089e9ea0f68e1e7d3b6ad5b64e1665fbe9377cd9e1dc09cc", - "md5": "5f0f52383b61a2e74166ad89cebae065", - "sha256": "7e4225ec065e0f354ccf7349a22d209de09cc1c074832be9eb84c51c1799c432" - }, - "downloads": -1, - "filename": "Django-4.2.4.tar.gz", - "has_sig": false, - "md5_digest": "5f0f52383b61a2e74166ad89cebae065", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10393648, - "upload_time": "2023-08-01T17:30:23", - "upload_time_iso_8601": "2023-08-01T17:30:23.968800Z", - "url": "https://files.pythonhosted.org/packages/8d/d2/a4f823741a12089e9ea0f68e1e7d3b6ad5b64e1665fbe9377cd9e1dc09cc/Django-4.2.4.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.2.5": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "bf8bc38f2354b6093d9ba310a14b43a830fdf776edd60c2e25c7c5f4d23cc243", - "md5": "a57dde066c29d8dcc418b8f92d56664d", - "sha256": "b6b2b5cae821077f137dc4dade696a1c2aa292f892eca28fa8d7bfdf2608ddd4" - }, - "downloads": -1, - "filename": "Django-4.2.5-py3-none-any.whl", - "has_sig": false, - "md5_digest": "a57dde066c29d8dcc418b8f92d56664d", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 7990309, - "upload_time": "2023-09-04T10:58:22", - "upload_time_iso_8601": "2023-09-04T10:58:22.309202Z", - "url": "https://files.pythonhosted.org/packages/bf/8b/c38f2354b6093d9ba310a14b43a830fdf776edd60c2e25c7c5f4d23cc243/Django-4.2.5-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "20eab0969834e5d79365731303be8b82423e6b1c293aa92c28335532ab542f83", - "md5": "63486f64f91bdc14a2edb84aa3001577", - "sha256": "5e5c1c9548ffb7796b4a8a4782e9a2e5a3df3615259fc1bfd3ebc73b646146c1" - }, - "downloads": -1, - "filename": "Django-4.2.5.tar.gz", - "has_sig": false, - "md5_digest": "63486f64f91bdc14a2edb84aa3001577", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10418606, - "upload_time": "2023-09-04T10:58:34", - "upload_time_iso_8601": "2023-09-04T10:58:34.288156Z", - "url": "https://files.pythonhosted.org/packages/20/ea/b0969834e5d79365731303be8b82423e6b1c293aa92c28335532ab542f83/Django-4.2.5.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.2.6": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "b945707dfc56f381222c1c798503546cb390934ab246fc45b5051ef66e31099c", - "md5": "db83d48600d6afff838e53f42f9ebebb", - "sha256": "a64d2487cdb00ad7461434320ccc38e60af9c404773a2f95ab0093b4453a3215" - }, - "downloads": -1, - "filename": "Django-4.2.6-py3-none-any.whl", - "has_sig": false, - "md5_digest": "db83d48600d6afff838e53f42f9ebebb", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 7990607, - "upload_time": "2023-10-04T14:58:34", - "upload_time_iso_8601": "2023-10-04T14:58:34.647748Z", - "url": "https://files.pythonhosted.org/packages/b9/45/707dfc56f381222c1c798503546cb390934ab246fc45b5051ef66e31099c/Django-4.2.6-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "237bf47d10d870fabfcaa1fba403460a4e482ab7dbba4d715d43981d1f8c8d85", - "md5": "ad84c2b9bbebaa26427a2a656fe5ceea", - "sha256": "08f41f468b63335aea0d904c5729e0250300f6a1907bf293a65499496cdbc68f" - }, - "downloads": -1, - "filename": "Django-4.2.6.tar.gz", - "has_sig": false, - "md5_digest": "ad84c2b9bbebaa26427a2a656fe5ceea", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10407018, - "upload_time": "2023-10-04T14:58:41", - "upload_time_iso_8601": "2023-10-04T14:58:41.808770Z", - "url": "https://files.pythonhosted.org/packages/23/7b/f47d10d870fabfcaa1fba403460a4e482ab7dbba4d715d43981d1f8c8d85/Django-4.2.6.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.2.7": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "2d6de87236e3c7b2f5911d132034177aebb605f3953910cc429df8061b13bf10", - "md5": "3d2bb75adc0c3c3497328175b301e244", - "sha256": "e1d37c51ad26186de355cbcec16613ebdabfa9689bbade9c538835205a8abbe9" - }, - "downloads": -1, - "filename": "Django-4.2.7-py3-none-any.whl", - "has_sig": false, - "md5_digest": "3d2bb75adc0c3c3497328175b301e244", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 7990980, - "upload_time": "2023-11-01T06:59:15", - "upload_time_iso_8601": "2023-11-01T06:59:15.299512Z", - "url": "https://files.pythonhosted.org/packages/2d/6d/e87236e3c7b2f5911d132034177aebb605f3953910cc429df8061b13bf10/Django-4.2.7-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "5c620c6ab2f3ac9a242b4562b6be1c418685fa7d1ccb8ca302cdb97e0b23cf4b", - "md5": "d7afe6a68b631725a1dac116a7832b10", - "sha256": "8e0f1c2c2786b5c0e39fe1afce24c926040fad47c8ea8ad30aaf1188df29fc41" - }, - "downloads": -1, - "filename": "Django-4.2.7.tar.gz", - "has_sig": false, - "md5_digest": "d7afe6a68b631725a1dac116a7832b10", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10425073, - "upload_time": "2023-11-01T06:59:30", - "upload_time_iso_8601": "2023-11-01T06:59:30.228988Z", - "url": "https://files.pythonhosted.org/packages/5c/62/0c6ab2f3ac9a242b4562b6be1c418685fa7d1ccb8ca302cdb97e0b23cf4b/Django-4.2.7.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.2.8": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "8a797f45e9c129c3cd8e4d54806649efeb1db9c223c54a1c54b30511d246bc60", - "md5": "6fd694f0d0519e13138f78caaf1a1c22", - "sha256": "6cb5dcea9e3d12c47834d32156b8841f533a4493c688e2718cafd51aa430ba6d" - }, - "downloads": -1, - "filename": "Django-4.2.8-py3-none-any.whl", - "has_sig": false, - "md5_digest": "6fd694f0d0519e13138f78caaf1a1c22", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 7991169, - "upload_time": "2023-12-04T08:34:42", - "upload_time_iso_8601": "2023-12-04T08:34:42.842122Z", - "url": "https://files.pythonhosted.org/packages/8a/79/7f45e9c129c3cd8e4d54806649efeb1db9c223c54a1c54b30511d246bc60/Django-4.2.8-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "a0bfff46fc1bd0f7d86390134e3869dd093be1d2bbaa9dafcb5389e87ef5bcb5", - "md5": "0ed422916f3766e82382d07235ff2ea8", - "sha256": "d69d5e36cc5d9f4eb4872be36c622878afcdce94062716cf3e25bcedcb168b62" - }, - "downloads": -1, - "filename": "Django-4.2.8.tar.gz", - "has_sig": false, - "md5_digest": "0ed422916f3766e82382d07235ff2ea8", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10425581, - "upload_time": "2023-12-04T08:34:50", - "upload_time_iso_8601": "2023-12-04T08:34:50.201677Z", - "url": "https://files.pythonhosted.org/packages/a0/bf/ff46fc1bd0f7d86390134e3869dd093be1d2bbaa9dafcb5389e87ef5bcb5/Django-4.2.8.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.2a1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "711af3b9a43f9f9118f248e9346f6eeff2bf24902f98f0030ea328941e9286c8", - "md5": "ed1bf2e61bebf6c4323cd79d8e27a19f", - "sha256": "e913e60991a0c2a5f099a80ea1e2ec90b1636e6deb60bad16b89086819b48d4b" - }, - "downloads": -1, - "filename": "Django-4.2a1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "ed1bf2e61bebf6c4323cd79d8e27a19f", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 7871271, - "upload_time": "2023-01-17T09:39:20", - "upload_time_iso_8601": "2023-01-17T09:39:20.186688Z", - "url": "https://files.pythonhosted.org/packages/71/1a/f3b9a43f9f9118f248e9346f6eeff2bf24902f98f0030ea328941e9286c8/Django-4.2a1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "25afcef5e406baaccb37f346662cddf58e9693a6b96f14e36341ef3817810fc5", - "md5": "531a657a21ded8dd22aff2f593d13f5f", - "sha256": "3a29fc014f46fd6a552296fd1cfc77f616c4120b68cc117eee4a8237aa163f7a" - }, - "downloads": -1, - "filename": "Django-4.2a1.tar.gz", - "has_sig": false, - "md5_digest": "531a657a21ded8dd22aff2f593d13f5f", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10318967, - "upload_time": "2023-01-17T09:39:25", - "upload_time_iso_8601": "2023-01-17T09:39:25.092445Z", - "url": "https://files.pythonhosted.org/packages/25/af/cef5e406baaccb37f346662cddf58e9693a6b96f14e36341ef3817810fc5/Django-4.2a1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.2b1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "995bc3cca0daea9f3fa5946d14c98bc1cd3c42c026ec6c41d57100e1d646e23a", - "md5": "306242df7de2e0ffe4da0f60cca734a7", - "sha256": "9bf13063a882a9b0f7028c4cdc32ea36fe104491cd7720859117990933f9c589" - }, - "downloads": -1, - "filename": "Django-4.2b1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "306242df7de2e0ffe4da0f60cca734a7", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 7873536, - "upload_time": "2023-02-20T08:49:43", - "upload_time_iso_8601": "2023-02-20T08:49:43.333662Z", - "url": "https://files.pythonhosted.org/packages/99/5b/c3cca0daea9f3fa5946d14c98bc1cd3c42c026ec6c41d57100e1d646e23a/Django-4.2b1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "4612b69773f84da1d0d5ab6f61c91a7257afa37f69413a5d019f7d2759eda28d", - "md5": "b7cd2f5420af2a204a099fdad63f7239", - "sha256": "33e3b3b80924dae3e6d4b5e697eaee724d5a35c1a430df44b1d72c802657992f" - }, - "downloads": -1, - "filename": "Django-4.2b1.tar.gz", - "has_sig": false, - "md5_digest": "b7cd2f5420af2a204a099fdad63f7239", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10328598, - "upload_time": "2023-02-20T08:49:47", - "upload_time_iso_8601": "2023-02-20T08:49:47.262950Z", - "url": "https://files.pythonhosted.org/packages/46/12/b69773f84da1d0d5ab6f61c91a7257afa37f69413a5d019f7d2759eda28d/Django-4.2b1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "4.2rc1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "8854bebcf756adceeadd39720c2ed589a4ca843c4caf469ff448d06050dba7e2", - "md5": "2ca79a600299d3be6eb84e14e8d91144", - "sha256": "fe3e5616eda5ac54c32e4bf01e3c063e0e20d2ff879a150a17621bf370b5cb2b" - }, - "downloads": -1, - "filename": "Django-4.2rc1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "2ca79a600299d3be6eb84e14e8d91144", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.8", - "size": 7873493, - "upload_time": "2023-03-20T07:31:55", - "upload_time_iso_8601": "2023-03-20T07:31:55.430411Z", - "url": "https://files.pythonhosted.org/packages/88/54/bebcf756adceeadd39720c2ed589a4ca843c4caf469ff448d06050dba7e2/Django-4.2rc1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "a44084ad7cbb7e3a4a94b95a783c62e5862a201126f33d03c103eb70a2b53fff", - "md5": "93eff55251133cdb9b4662dcbe7930ab", - "sha256": "6bf1d4675d3320f0024a77d058449393b9a574fd0d2da8d3feb6e06bf3a307e2" - }, - "downloads": -1, - "filename": "Django-4.2rc1.tar.gz", - "has_sig": false, - "md5_digest": "93eff55251133cdb9b4662dcbe7930ab", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.8", - "size": 10330770, - "upload_time": "2023-03-20T07:32:00", - "upload_time_iso_8601": "2023-03-20T07:32:00.502267Z", - "url": "https://files.pythonhosted.org/packages/a4/40/84ad7cbb7e3a4a94b95a783c62e5862a201126f33d03c103eb70a2b53fff/Django-4.2rc1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "5.0": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "bac761b02c0ef9e129080a8c2bffefb3cb2b9ddddece4c44dc473c1c4f0647c1", - "md5": "3fa7115ed069c54684302278b48b7fd6", - "sha256": "3a9fd52b8dbeae335ddf4a9dfa6c6a0853a1122f1fb071a8d5eca979f73a05c8" - }, - "downloads": -1, - "filename": "Django-5.0-py3-none-any.whl", - "has_sig": false, - "md5_digest": "3fa7115ed069c54684302278b48b7fd6", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.10", - "size": 8136382, - "upload_time": "2023-12-04T13:12:41", - "upload_time_iso_8601": "2023-12-04T13:12:41.502020Z", - "url": "https://files.pythonhosted.org/packages/ba/c7/61b02c0ef9e129080a8c2bffefb3cb2b9ddddece4c44dc473c1c4f0647c1/Django-5.0-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "bea646e250737d46e955e048f6bbc2948fb22f0de3f3ab828d3803070dc1260e", - "md5": "5de06916549d6f81fe6eb102722a1b28", - "sha256": "7d29e14dfbc19cb6a95a4bd669edbde11f5d4c6a71fdaa42c2d40b6846e807f7" - }, - "downloads": -1, - "filename": "Django-5.0.tar.gz", - "has_sig": false, - "md5_digest": "5de06916549d6f81fe6eb102722a1b28", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.10", - "size": 10585390, - "upload_time": "2023-12-04T13:12:50", - "upload_time_iso_8601": "2023-12-04T13:12:50.251293Z", - "url": "https://files.pythonhosted.org/packages/be/a6/46e250737d46e955e048f6bbc2948fb22f0de3f3ab828d3803070dc1260e/Django-5.0.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "5.0a1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "e004edccafd2573a97258a6fc7817a4e59a4e61a37c07b092c7899217087d991", - "md5": "be6c6bd9752c4577d9f8c4619e9b028d", - "sha256": "4b7ee536e8b2bf1e68f6c9bda192300ec94844e32e730076c2e520bc63dac64e" - }, - "downloads": -1, - "filename": "Django-5.0a1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "be6c6bd9752c4577d9f8c4619e9b028d", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.10", - "size": 8014408, - "upload_time": "2023-09-18T22:48:31", - "upload_time_iso_8601": "2023-09-18T22:48:31.458640Z", - "url": "https://files.pythonhosted.org/packages/e0/04/edccafd2573a97258a6fc7817a4e59a4e61a37c07b092c7899217087d991/Django-5.0a1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "7fa20380a5fd74dd435524705e5bbe5e00e491a6afef44f39339e9c2c24f8106", - "md5": "bea378daf7cb1334ad37c5fbf319f88b", - "sha256": "3a52052fbc5e01f2202c020a8241aa40dc030b16c8fb49dc480cc46ab8735158" - }, - "downloads": -1, - "filename": "Django-5.0a1.tar.gz", - "has_sig": false, - "md5_digest": "bea378daf7cb1334ad37c5fbf319f88b", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.10", - "size": 10466505, - "upload_time": "2023-09-18T22:48:42", - "upload_time_iso_8601": "2023-09-18T22:48:42.066135Z", - "url": "https://files.pythonhosted.org/packages/7f/a2/0380a5fd74dd435524705e5bbe5e00e491a6afef44f39339e9c2c24f8106/Django-5.0a1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "5.0b1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "3a0f33ccd4b8975c89c06392a7d6bbcd6ec492ce94cda39c4f53c08b826a80b3", - "md5": "186929b21b30c6525b0a0fe240869420", - "sha256": "fc0fb85721c984d7c7bcbfc59cbd5eb7bcc5dd678f394b3b2b311f0c43cfb943" - }, - "downloads": -1, - "filename": "Django-5.0b1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "186929b21b30c6525b0a0fe240869420", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.10", - "size": 8015514, - "upload_time": "2023-10-23T18:57:51", - "upload_time_iso_8601": "2023-10-23T18:57:51.926093Z", - "url": "https://files.pythonhosted.org/packages/3a/0f/33ccd4b8975c89c06392a7d6bbcd6ec492ce94cda39c4f53c08b826a80b3/Django-5.0b1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "a3d35f81213107d2add6a5527fe65ce457f099c677ea8b4b8cd31930d6afeb32", - "md5": "2b7020e3aa3c03ab4c46b98d624293a7", - "sha256": "c88635e733f0a0ef46c219635e21c8f5978eb2a14c3918224654541bc5d57030" - }, - "downloads": -1, - "filename": "Django-5.0b1.tar.gz", - "has_sig": false, - "md5_digest": "2b7020e3aa3c03ab4c46b98d624293a7", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.10", - "size": 10493582, - "upload_time": "2023-10-23T18:57:57", - "upload_time_iso_8601": "2023-10-23T18:57:57.673233Z", - "url": "https://files.pythonhosted.org/packages/a3/d3/5f81213107d2add6a5527fe65ce457f099c677ea8b4b8cd31930d6afeb32/Django-5.0b1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ], - "5.0rc1": [ - { - "comment_text": "", - "digests": { - "blake2b_256": "187ded709d1fbedebf949ff34fbeb3607de15a639b7d6b20e92db24494a1d125", - "md5": "c6d451064eccf90d4bfec8ff172532bc", - "sha256": "04244960e238ef773db07890a9936351c8509c9a04812d38f5ead6d7acee58d6" - }, - "downloads": -1, - "filename": "Django-5.0rc1-py3-none-any.whl", - "has_sig": false, - "md5_digest": "c6d451064eccf90d4bfec8ff172532bc", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.10", - "size": 8016166, - "upload_time": "2023-11-20T12:40:22", - "upload_time_iso_8601": "2023-11-20T12:40:22.660572Z", - "url": "https://files.pythonhosted.org/packages/18/7d/ed709d1fbedebf949ff34fbeb3607de15a639b7d6b20e92db24494a1d125/Django-5.0rc1-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "c3ebb7cc29146660057225de0d176c0b71538f39b101aa7af3caa9ece4c4f2cb", - "md5": "116720a171f1748be6f4631d24622566", - "sha256": "a4bb77a659da032b741a95dcb957864f3695275d1d581ef7637214300faaaf30" - }, - "downloads": -1, - "filename": "Django-5.0rc1.tar.gz", - "has_sig": false, - "md5_digest": "116720a171f1748be6f4631d24622566", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.10", - "size": 10499875, - "upload_time": "2023-11-20T12:40:32", - "upload_time_iso_8601": "2023-11-20T12:40:32.210081Z", - "url": "https://files.pythonhosted.org/packages/c3/eb/b7cc29146660057225de0d176c0b71538f39b101aa7af3caa9ece4c4f2cb/Django-5.0rc1.tar.gz", - "yanked": false, - "yanked_reason": null - } - ] - }, - "urls": [ + "0.5": [], + "0.5.1": [], + "0.5.2": [], + "0.5.3": [], + "0.5.4": [], + "0.5.5": [], + "0.5.6": [], + "0.5.7": [], + "0.5.8": [], + "0.5.9": [], + "0.5rc2": [], + "0.6": [], + "0.6.1": [], + "0.6.10": [], + "0.6.11": [], + "0.6.12": [], + "0.6.13": [], + "0.6.14": [], + "0.6.15": [], + "0.6.16": [], + "0.6.17": [], + "0.6.18": [], + "0.6.19": [], + "0.6.2": [], + "0.6.20": [], + "0.6.21": [], + "0.6.22": [], + "0.6.23": [], + "0.6.24": [], + "0.6.25": [], + "0.6.26": [], + "0.6.27": [], + "0.6.28": [], + "0.6.29": [], + "0.6.3": [], + "0.6.30": [], + "0.6.31": [], + "0.6.4": [], + "0.6.5": [], + "0.6.6": [], + "0.6.7": [], + "0.6.8": [], + "0.6.9": [], + "0.7": [], + "0.7.1": [], + "0.7.10": [], + "0.7.11": [], + "0.7.12": [], + "0.7.13": [], + "0.7.14": [], + "0.7.15": [], + "0.7.16": [], + "0.7.17": [], + "0.7.18": [], + "0.7.19": [], + "0.7.2": [], + "0.7.20": [], + "0.7.21": [], + "0.7.22": [], + "0.7.23": [], + "0.7.24": [], + "0.7.25": [], + "0.7.26": [], + "0.7.27": [], + "0.7.28": [], + "0.7.29": [], + "0.7.3": [], + "0.7.30": [], + "0.7.31": [], + "0.7.32": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "c4b85408e0dc12788adc2e4653a42dc8a5d6c6558e7649aa3b7dd13f73ec2756", + "md5": "4b9a8c7b901331754255744ca2bcc3b2", + "sha256": "515a36123adad6cfa646a4a417d47f68737a62328d1c5918313c75fb8552f390" + }, + "downloads": -1, + "filename": "Djblets-0.7.32-py2.5.egg", + "has_sig": false, + "md5_digest": "4b9a8c7b901331754255744ca2bcc3b2", + "packagetype": "bdist_egg", + "python_version": "2.5", + "requires_python": null, + "size": 712730, + "upload_time": "2016-02-22T06:12:47", + "upload_time_iso_8601": "2016-02-22T06:12:47.666536Z", + "url": "https://files.pythonhosted.org/packages/c4/b8/5408e0dc12788adc2e4653a42dc8a5d6c6558e7649aa3b7dd13f73ec2756/Djblets-0.7.32-py2.5.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "f308f1a5c1f2b4d0cbf30b16722bc19dc9d6da1800f7a85465a59348c47baf06", + "md5": "cc2ba536993435045dbed27d3a55a83e", + "sha256": "10c8af36ed2908bbe20a75a5aa2383e3c0887a50200873c99063025dfbef5382" + }, + "downloads": -1, + "filename": "Djblets-0.7.32-py2.6.egg", + "has_sig": false, + "md5_digest": "cc2ba536993435045dbed27d3a55a83e", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 711954, + "upload_time": "2016-02-22T06:12:54", + "upload_time_iso_8601": "2016-02-22T06:12:54.132696Z", + "url": "https://files.pythonhosted.org/packages/f3/08/f1a5c1f2b4d0cbf30b16722bc19dc9d6da1800f7a85465a59348c47baf06/Djblets-0.7.32-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "4a442b6f595e2605d300141d21c816ee7649a1c6637493ecf5cf8a19bba4aa35", + "md5": "5fd08e09ca06476da1ea8ada43a93bc7", + "sha256": "f0733962f30cab838e2635d9b30b90b04df5b4ec8d5c8c52af687767a5823c5c" + }, + "downloads": -1, + "filename": "Djblets-0.7.32-py2.7.egg", + "has_sig": false, + "md5_digest": "5fd08e09ca06476da1ea8ada43a93bc7", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 710894, + "upload_time": "2016-02-22T06:13:00", + "upload_time_iso_8601": "2016-02-22T06:13:00.351936Z", + "url": "https://files.pythonhosted.org/packages/4a/44/2b6f595e2605d300141d21c816ee7649a1c6637493ecf5cf8a19bba4aa35/Djblets-0.7.32-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "47d9ecba621fe64fcc09b9a1c02c9200048e15195707029f2b29cfd1ebd0c5a9", + "md5": "8b854a2dc3fc5ff9b7872e6e249b2a86", + "sha256": "8df8b2d7928a7e61111bd18700969b821129186b746abe2c02ec5aec298a1ee5" + }, + "downloads": -1, + "filename": "Djblets-0.7.32-py2-none-any.whl", + "has_sig": false, + "md5_digest": "8b854a2dc3fc5ff9b7872e6e249b2a86", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 521080, + "upload_time": "2016-02-22T06:12:37", + "upload_time_iso_8601": "2016-02-22T06:12:37.989257Z", + "url": "https://files.pythonhosted.org/packages/47/d9/ecba621fe64fcc09b9a1c02c9200048e15195707029f2b29cfd1ebd0c5a9/Djblets-0.7.32-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "03a49b76ea24ed6795a28fe8194d260d2e026d49deb43fd66e688193a02783e9", + "md5": "fd8579ec48895ddcade53eaeb6358fe0", + "sha256": "a06cbd6638edb975569bb2d588da83caa725eb2fa0632619ae0b06813c02c018" + }, + "downloads": -1, + "filename": "Djblets-0.7.32.tar.gz", + "has_sig": false, + "md5_digest": "fd8579ec48895ddcade53eaeb6358fe0", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 430408, + "upload_time": "2016-02-22T06:13:06", + "upload_time_iso_8601": "2016-02-22T06:13:06.774750Z", + "url": "https://files.pythonhosted.org/packages/03/a4/9b76ea24ed6795a28fe8194d260d2e026d49deb43fd66e688193a02783e9/Djblets-0.7.32.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.7.33": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "a811611c6b603577b49beecd26e4cf52ae5320e82427839c2c16cafec7301829", + "md5": "83307ddc5b285439de53826436a3c74d", + "sha256": "893467f5043fd65a25bad41a296ac097674b624fef20e47393cd1ebc71b6b038" + }, + "downloads": -1, + "filename": "Djblets-0.7.33-py2.5.egg", + "has_sig": false, + "md5_digest": "83307ddc5b285439de53826436a3c74d", + "packagetype": "bdist_egg", + "python_version": "2.5", + "requires_python": null, + "size": 712738, + "upload_time": "2016-03-01T18:55:22", + "upload_time_iso_8601": "2016-03-01T18:55:22.724445Z", + "url": "https://files.pythonhosted.org/packages/a8/11/611c6b603577b49beecd26e4cf52ae5320e82427839c2c16cafec7301829/Djblets-0.7.33-py2.5.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "aafb003e1c6c2c2d600c9dbffcb3b5fd0451d5c3d3534d1752de5c7631221751", + "md5": "468e5a325e75789b12111bafa3cdf0d6", + "sha256": "82320f781c3d8d979cf10a7beaa14dbacf14fb05fe9d203820679ad5fcd1ee1d" + }, + "downloads": -1, + "filename": "Djblets-0.7.33-py2.6.egg", + "has_sig": false, + "md5_digest": "468e5a325e75789b12111bafa3cdf0d6", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 711974, + "upload_time": "2016-03-01T18:55:32", + "upload_time_iso_8601": "2016-03-01T18:55:32.102191Z", + "url": "https://files.pythonhosted.org/packages/aa/fb/003e1c6c2c2d600c9dbffcb3b5fd0451d5c3d3534d1752de5c7631221751/Djblets-0.7.33-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "bdb1dfd826bc798785238044ebe04bd34903759b6bc24e918affb7b59d781ed9", + "md5": "8aceae4818372532647684b685ee3f8a", + "sha256": "8f3544923284a307a36a9ea323974fd3810ae6fad45b92fb2a6eb76bd82e2779" + }, + "downloads": -1, + "filename": "Djblets-0.7.33-py2.7.egg", + "has_sig": false, + "md5_digest": "8aceae4818372532647684b685ee3f8a", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 710921, + "upload_time": "2016-03-01T18:55:39", + "upload_time_iso_8601": "2016-03-01T18:55:39.357272Z", + "url": "https://files.pythonhosted.org/packages/bd/b1/dfd826bc798785238044ebe04bd34903759b6bc24e918affb7b59d781ed9/Djblets-0.7.33-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "9cec5e75570fb1dec110938ceb05233f8f71b8803bbcbc5be512e71992d9d9d5", + "md5": "50630fe29d62682b68a63aed0acdaa00", + "sha256": "9d17bcb26b9fe8a763bdbbe56d7db5f41a9d511427e9febdb849267125971e5a" + }, + "downloads": -1, + "filename": "Djblets-0.7.33-py2-none-any.whl", + "has_sig": false, + "md5_digest": "50630fe29d62682b68a63aed0acdaa00", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 521100, + "upload_time": "2016-03-01T18:55:13", + "upload_time_iso_8601": "2016-03-01T18:55:13.729434Z", + "url": "https://files.pythonhosted.org/packages/9c/ec/5e75570fb1dec110938ceb05233f8f71b8803bbcbc5be512e71992d9d9d5/Djblets-0.7.33-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "a1a3a3e7c022f87fab64bfcbf2163f3767418759fc3fb20e0dedc516a39cd4f8", + "md5": "ec06c1a15b09653ad043266d9f4a861e", + "sha256": "48873ef0f0d1bf71dc3dbf5e1790b68b5dc399fbcefb0c3f4140baf6d91d2850" + }, + "downloads": -1, + "filename": "Djblets-0.7.33.tar.gz", + "has_sig": false, + "md5_digest": "ec06c1a15b09653ad043266d9f4a861e", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 430413, + "upload_time": "2016-03-01T18:55:48", + "upload_time_iso_8601": "2016-03-01T18:55:48.013515Z", + "url": "https://files.pythonhosted.org/packages/a1/a3/a3e7c022f87fab64bfcbf2163f3767418759fc3fb20e0dedc516a39cd4f8/Djblets-0.7.33.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.7.4": [], + "0.7.5": [], + "0.7.6": [], + "0.7.7": [], + "0.7.8": [], + "0.7.9": [], + "0.8": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "20befe0e45e2f2b7a3d32b48c27f8c0409cc3c53cbdcf1ccff8d35a6acff86c3", + "md5": "774116450a8ed652ce3e8854d401bd2a", + "sha256": "c9a5a81b877d71d1fd3561d85df5dd35b865151c18d26ca1d10c34be04b9746e" + }, + "downloads": -1, + "filename": "Djblets-0.8-py2.6.egg", + "has_sig": false, + "md5_digest": "774116450a8ed652ce3e8854d401bd2a", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 623706, + "upload_time": "2015-10-29T21:05:45", + "upload_time_iso_8601": "2015-10-29T21:05:45.155440Z", + "url": "https://files.pythonhosted.org/packages/20/be/fe0e45e2f2b7a3d32b48c27f8c0409cc3c53cbdcf1ccff8d35a6acff86c3/Djblets-0.8-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "0140fb24a625bf8f7c1b6163b64af59fd06afeca27d626216da15b060c11639c", + "md5": "c7215e6e593d8d18ceb60b4a56085180", + "sha256": "8249e80bd41ee646311482f3eb6ebfc98eb5bd0210c9f6981930a3be9262ee0b" + }, + "downloads": -1, + "filename": "Djblets-0.8-py2.7.egg", + "has_sig": false, + "md5_digest": "c7215e6e593d8d18ceb60b4a56085180", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 632828, + "upload_time": "2015-10-29T21:05:52", + "upload_time_iso_8601": "2015-10-29T21:05:52.767039Z", + "url": "https://files.pythonhosted.org/packages/01/40/fb24a625bf8f7c1b6163b64af59fd06afeca27d626216da15b060c11639c/Djblets-0.8-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "cebd7745156b8777957eab4dc75916139cacde81ea3aafbdd7a07f0732ae7f74", + "md5": "bfbdb98916550304b90a31426dd7fe49", + "sha256": "4574ba0a50ae9a7be1d0124cda2e4256a53e12901dc02bb6d0e5f69fc48a41aa" + }, + "downloads": -1, + "filename": "Djblets-0.8.tar.gz", + "has_sig": false, + "md5_digest": "bfbdb98916550304b90a31426dd7fe49", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 258340, + "upload_time": "2015-10-29T21:14:49", + "upload_time_iso_8601": "2015-10-29T21:14:49.182641Z", + "url": "https://files.pythonhosted.org/packages/ce/bd/7745156b8777957eab4dc75916139cacde81ea3aafbdd7a07f0732ae7f74/Djblets-0.8.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.1": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "9b8088b49115e2060876d1ccdb7cbdad635232e1958afb0c78edd5d2be172809", + "md5": "570a40f6eb3e3272115b379bc52094e9", + "sha256": "da8127df0da1e61bcb9422d88933fcee93933621fbad0a20801bce52e93046d0" + }, + "downloads": -1, + "filename": "Djblets-0.8.1-py2.6.egg", + "has_sig": false, + "md5_digest": "570a40f6eb3e3272115b379bc52094e9", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 626460, + "upload_time": "2015-10-29T21:05:58", + "upload_time_iso_8601": "2015-10-29T21:05:58.158385Z", + "url": "https://files.pythonhosted.org/packages/9b/80/88b49115e2060876d1ccdb7cbdad635232e1958afb0c78edd5d2be172809/Djblets-0.8.1-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "e494b845ca030b0c54c41114845f2f53df472fb26bd7622a5eb86fe2e6be1108", + "md5": "686f35aa193753de6bc9020322ebb370", + "sha256": "d16b10151bda0b76539cca7b90273b5ebc745005106c47d09b09e5c0c3143f43" + }, + "downloads": -1, + "filename": "Djblets-0.8.1-py2.7.egg", + "has_sig": false, + "md5_digest": "686f35aa193753de6bc9020322ebb370", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 635608, + "upload_time": "2015-10-29T21:06:11", + "upload_time_iso_8601": "2015-10-29T21:06:11.693414Z", + "url": "https://files.pythonhosted.org/packages/e4/94/b845ca030b0c54c41114845f2f53df472fb26bd7622a5eb86fe2e6be1108/Djblets-0.8.1-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "a48a540a849795d951469e91c229cac260b5e82a521c6366d2ccaef2067d34e7", + "md5": "63672a8263c46b107cb06a7ba59112a4", + "sha256": "cbb000a4279c855f03d1f5ef6e70f095cf05005cac08c9b924b1bbb3034f8659" + }, + "downloads": -1, + "filename": "Djblets-0.8.1.tar.gz", + "has_sig": false, + "md5_digest": "63672a8263c46b107cb06a7ba59112a4", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 260872, + "upload_time": "2015-10-29T21:06:19", + "upload_time_iso_8601": "2015-10-29T21:06:19.206843Z", + "url": "https://files.pythonhosted.org/packages/a4/8a/540a849795d951469e91c229cac260b5e82a521c6366d2ccaef2067d34e7/Djblets-0.8.1.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.10": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "0c434f48b4547009e0d3e9b037d56103c066b898a163dbe2b266b5114344e250", + "md5": "b97971b21d694931827a952597441fed", + "sha256": "a9ebf2de1f3080e8ac2148c118a3bc0d403678aa27486b99ad315a260d6191f2" + }, + "downloads": -1, + "filename": "Djblets-0.8.10-py2.6.egg", + "has_sig": false, + "md5_digest": "b97971b21d694931827a952597441fed", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 721179, + "upload_time": "2015-10-29T21:06:26", + "upload_time_iso_8601": "2015-10-29T21:06:26.184173Z", + "url": "https://files.pythonhosted.org/packages/0c/43/4f48b4547009e0d3e9b037d56103c066b898a163dbe2b266b5114344e250/Djblets-0.8.10-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "85928a3a57f0b075cdeb63f60ea60f1eea4c7c7cd67b158c2a131c7065672a7d", + "md5": "15464937cca7973a1e407c93b1c1ab75", + "sha256": "3018c8c6ac6e81ed5eec8532f1172964e16247dd746a6f323061987bb8f6b27d" + }, + "downloads": -1, + "filename": "Djblets-0.8.10-py2.7.egg", + "has_sig": false, + "md5_digest": "15464937cca7973a1e407c93b1c1ab75", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 730145, + "upload_time": "2015-10-29T21:06:32", + "upload_time_iso_8601": "2015-10-29T21:06:32.770064Z", + "url": "https://files.pythonhosted.org/packages/85/92/8a3a57f0b075cdeb63f60ea60f1eea4c7c7cd67b158c2a131c7065672a7d/Djblets-0.8.10-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "08c9c7736d6e5f5c659693a0489d7d1634d0f4dd143d22bef7999070d788eb56", + "md5": "4dfa1110d0a2e41917a5bb953e4910fb", + "sha256": "4e5035ec36aee3cfe8d9c58f9b2052b93716c0f78caec2c396b99c69daa5b97e" + }, + "downloads": -1, + "filename": "Djblets-0.8.10.tar.gz", + "has_sig": false, + "md5_digest": "4dfa1110d0a2e41917a5bb953e4910fb", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 268679, + "upload_time": "2015-10-29T21:06:38", + "upload_time_iso_8601": "2015-10-29T21:06:38.680140Z", + "url": "https://files.pythonhosted.org/packages/08/c9/c7736d6e5f5c659693a0489d7d1634d0f4dd143d22bef7999070d788eb56/Djblets-0.8.10.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.11": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "479d7bd680b41dc35086167e9fc75b8e1b518b85aa2793e7060eb7c2f8a81660", + "md5": "9939b0038a9eac76f98370a82689c370", + "sha256": "070584a637034eb13fa07ce71ee4753d5c334d402a835e93a0a2d58372b68cde" + }, + "downloads": -1, + "filename": "Djblets-0.8.11-py2.6.egg", + "has_sig": false, + "md5_digest": "9939b0038a9eac76f98370a82689c370", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 722010, + "upload_time": "2015-10-29T21:06:45", + "upload_time_iso_8601": "2015-10-29T21:06:45.792831Z", + "url": "https://files.pythonhosted.org/packages/47/9d/7bd680b41dc35086167e9fc75b8e1b518b85aa2793e7060eb7c2f8a81660/Djblets-0.8.11-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "acd80c8bf5b19b77bc857bd193f6bb0ba1ffd3dcc262d5bc5261ba9d8c1b1703", + "md5": "5ca36de44b8165418ba939bdd48e7a35", + "sha256": "f1bc2d354b07c423a5459412af45cd0ab831d6e2c93bfda95e6edc4b8e81231b" + }, + "downloads": -1, + "filename": "Djblets-0.8.11-py2.7.egg", + "has_sig": false, + "md5_digest": "5ca36de44b8165418ba939bdd48e7a35", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 731015, + "upload_time": "2015-10-29T21:06:51", + "upload_time_iso_8601": "2015-10-29T21:06:51.928868Z", + "url": "https://files.pythonhosted.org/packages/ac/d8/0c8bf5b19b77bc857bd193f6bb0ba1ffd3dcc262d5bc5261ba9d8c1b1703/Djblets-0.8.11-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "cf18a3eda2d8daad3328a4e03549fee91c844a92d2f1a19557cb3499e798e79b", + "md5": "add01e5f2949ad8774a90bbfc7fca6a6", + "sha256": "0160b7c2a1a13cebb08898ffec41a2aad239c1c9c12d578a27d0816eb86d8d7f" + }, + "downloads": -1, + "filename": "Djblets-0.8.11.tar.gz", + "has_sig": false, + "md5_digest": "add01e5f2949ad8774a90bbfc7fca6a6", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 268842, + "upload_time": "2015-10-29T21:06:57", + "upload_time_iso_8601": "2015-10-29T21:06:57.940593Z", + "url": "https://files.pythonhosted.org/packages/cf/18/a3eda2d8daad3328a4e03549fee91c844a92d2f1a19557cb3499e798e79b/Djblets-0.8.11.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.12": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "c8146e3a21fe8989339c53577ca211befc03e72e82adb1acf006bf8006cc2aeb", + "md5": "97cccbb6584ad4b375e25ba8236a39f1", + "sha256": "ecbda14a5531bf2d4202efdaaf71a2ae933c09e5bdc7546798b6235007e93b71" + }, + "downloads": -1, + "filename": "Djblets-0.8.12-py2.6.egg", + "has_sig": false, + "md5_digest": "97cccbb6584ad4b375e25ba8236a39f1", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 726465, + "upload_time": "2015-10-29T21:07:07", + "upload_time_iso_8601": "2015-10-29T21:07:07.448662Z", + "url": "https://files.pythonhosted.org/packages/c8/14/6e3a21fe8989339c53577ca211befc03e72e82adb1acf006bf8006cc2aeb/Djblets-0.8.12-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "97b2a9477439d89fdf11afd0dae43cf0003760e6af3f637a8d9512d37001715e", + "md5": "6a68111e1f1e14ff70082df684a093c1", + "sha256": "3d5e3f91164fafc94b8035c391b6e9a0468c71aa154cd46d9cea018e76dfb539" + }, + "downloads": -1, + "filename": "Djblets-0.8.12-py2.7.egg", + "has_sig": false, + "md5_digest": "6a68111e1f1e14ff70082df684a093c1", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 735402, + "upload_time": "2015-10-29T21:07:14", + "upload_time_iso_8601": "2015-10-29T21:07:14.671571Z", + "url": "https://files.pythonhosted.org/packages/97/b2/a9477439d89fdf11afd0dae43cf0003760e6af3f637a8d9512d37001715e/Djblets-0.8.12-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "ca79a0a99f331df016a54545c1bb777dd757a9ad6a9fcdd759222f75649d9ebc", + "md5": "f1bb0d9e32ff9233f5c139b7abf65880", + "sha256": "4580591b0c99bb3d36615af643ff0b098463dbfaaa603e0903b9e1be27671d45" + }, + "downloads": -1, + "filename": "Djblets-0.8.12.tar.gz", + "has_sig": false, + "md5_digest": "f1bb0d9e32ff9233f5c139b7abf65880", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 270559, + "upload_time": "2015-10-29T21:07:20", + "upload_time_iso_8601": "2015-10-29T21:07:20.546698Z", + "url": "https://files.pythonhosted.org/packages/ca/79/a0a99f331df016a54545c1bb777dd757a9ad6a9fcdd759222f75649d9ebc/Djblets-0.8.12.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.13": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "58afa9cfe4cf1d1e177299da25d301b89392f3ec6e4384e6ccc7cdf11991c350", + "md5": "56449486995ee3114f4c6c99bab32219", + "sha256": "34d8ef44b7ab00456ab5a3b674fb3da39f6c442cb37f97d42f8426ed61941abf" + }, + "downloads": -1, + "filename": "Djblets-0.8.13-py2.6.egg", + "has_sig": false, + "md5_digest": "56449486995ee3114f4c6c99bab32219", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 726829, + "upload_time": "2015-10-29T21:07:26", + "upload_time_iso_8601": "2015-10-29T21:07:26.289079Z", + "url": "https://files.pythonhosted.org/packages/58/af/a9cfe4cf1d1e177299da25d301b89392f3ec6e4384e6ccc7cdf11991c350/Djblets-0.8.13-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "9a0828371265514a3071c25d6c31df3971038d9292e4f9c79c909a2dd59b275f", + "md5": "2ad275c804f735da9506c22631df4294", + "sha256": "08929163a49aed9b69aac04f0087476bd90fc9b3c6247c805ea8b3e8bce6224b" + }, + "downloads": -1, + "filename": "Djblets-0.8.13-py2.7.egg", + "has_sig": false, + "md5_digest": "2ad275c804f735da9506c22631df4294", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 735792, + "upload_time": "2015-10-29T21:07:33", + "upload_time_iso_8601": "2015-10-29T21:07:33.234260Z", + "url": "https://files.pythonhosted.org/packages/9a/08/28371265514a3071c25d6c31df3971038d9292e4f9c79c909a2dd59b275f/Djblets-0.8.13-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "fc97007ff49a48223e768c77c68bff5c11a664f3c134b1b10d77a4a265272e08", + "md5": "8ac4ec4c5fbd645e87a26faab9d5e7d1", + "sha256": "8ade1d0507b5db861c0b4d7bdecc348ffff728e445c6381f951e1b070305f7e6" + }, + "downloads": -1, + "filename": "Djblets-0.8.13.tar.gz", + "has_sig": false, + "md5_digest": "8ac4ec4c5fbd645e87a26faab9d5e7d1", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 270558, + "upload_time": "2015-10-29T21:07:39", + "upload_time_iso_8601": "2015-10-29T21:07:39.207404Z", + "url": "https://files.pythonhosted.org/packages/fc/97/007ff49a48223e768c77c68bff5c11a664f3c134b1b10d77a4a265272e08/Djblets-0.8.13.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.14": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "3f9c27d20294661a90cb7a06813288e617ac668ada9307159e159c112852a18b", + "md5": "211297230611d1fccb8069b9c32b1f43", + "sha256": "017c7e482310eed4419e860bb52ec7962d1ccfbd0530fe14536c952181531636" + }, + "downloads": -1, + "filename": "Djblets-0.8.14-py2.6.egg", + "has_sig": false, + "md5_digest": "211297230611d1fccb8069b9c32b1f43", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 734395, + "upload_time": "2015-10-29T21:07:46", + "upload_time_iso_8601": "2015-10-29T21:07:46.341967Z", + "url": "https://files.pythonhosted.org/packages/3f/9c/27d20294661a90cb7a06813288e617ac668ada9307159e159c112852a18b/Djblets-0.8.14-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "63604f948b697958895b5721ec2cc1e85687ae8f0b65148e92ee0c23f1e2ec26", + "md5": "3e274d8e63eaa48fbe433701a358ed85", + "sha256": "c0f94cfea19e097575bc882aaec65501690d5e764df953b6a4a0f47525a72259" + }, + "downloads": -1, + "filename": "Djblets-0.8.14-py2.7.egg", + "has_sig": false, + "md5_digest": "3e274d8e63eaa48fbe433701a358ed85", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 743450, + "upload_time": "2015-10-29T21:07:52", + "upload_time_iso_8601": "2015-10-29T21:07:52.833535Z", + "url": "https://files.pythonhosted.org/packages/63/60/4f948b697958895b5721ec2cc1e85687ae8f0b65148e92ee0c23f1e2ec26/Djblets-0.8.14-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "dbf461e575d782760fc13bd6aa581d033f063f8bbf2c8ac718b585f278913848", + "md5": "52b0f4e5abe8186e6e86849cdf6d19c9", + "sha256": "63b3da8fe64e11891e55399cc890c5d7da36098eaf6ebb8581ef8e335647df0b" + }, + "downloads": -1, + "filename": "Djblets-0.8.14.tar.gz", + "has_sig": false, + "md5_digest": "52b0f4e5abe8186e6e86849cdf6d19c9", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 273185, + "upload_time": "2015-10-29T21:07:58", + "upload_time_iso_8601": "2015-10-29T21:07:58.530827Z", + "url": "https://files.pythonhosted.org/packages/db/f4/61e575d782760fc13bd6aa581d033f063f8bbf2c8ac718b585f278913848/Djblets-0.8.14.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.15": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "4bf46870c4a3edfe31c684524fdc55d48ee6bc4263e513601c39c021e3c3d752", + "md5": "e3381e092b4638b2a4aad26cbfed3727", + "sha256": "2390737021a9318b998063b9902d5103d103e2a60770a9903615533d64750b72" + }, + "downloads": -1, + "filename": "Djblets-0.8.15-py2.6.egg", + "has_sig": false, + "md5_digest": "e3381e092b4638b2a4aad26cbfed3727", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 738364, + "upload_time": "2015-10-29T21:08:12", + "upload_time_iso_8601": "2015-10-29T21:08:12.103732Z", + "url": "https://files.pythonhosted.org/packages/4b/f4/6870c4a3edfe31c684524fdc55d48ee6bc4263e513601c39c021e3c3d752/Djblets-0.8.15-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "eb5f84c111ab171c0eaeeff5be036f33197b466ec2501aff1b4bc9a6fd545b99", + "md5": "524d0964951fd726c243fdab9ccd1739", + "sha256": "523406c217666a92cb16a425ec180063df3b1261c5074e3cbdd3ba9404077c46" + }, + "downloads": -1, + "filename": "Djblets-0.8.15-py2.7.egg", + "has_sig": false, + "md5_digest": "524d0964951fd726c243fdab9ccd1739", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 747295, + "upload_time": "2015-10-29T21:08:17", + "upload_time_iso_8601": "2015-10-29T21:08:17.834033Z", + "url": "https://files.pythonhosted.org/packages/eb/5f/84c111ab171c0eaeeff5be036f33197b466ec2501aff1b4bc9a6fd545b99/Djblets-0.8.15-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "89325b3709c6c58981ef8fee4d6b281d687ffec759b2a84f8a47ea2591f43539", + "md5": "86dd73681ead4557a762ba9aac82aa3d", + "sha256": "def6974edb5fef3e817142c832ed99829087e2ffb99bf2e21f6d5c2e32592318" + }, + "downloads": -1, + "filename": "Djblets-0.8.15.tar.gz", + "has_sig": false, + "md5_digest": "86dd73681ead4557a762ba9aac82aa3d", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 275029, + "upload_time": "2015-10-29T21:08:22", + "upload_time_iso_8601": "2015-10-29T21:08:22.566012Z", + "url": "https://files.pythonhosted.org/packages/89/32/5b3709c6c58981ef8fee4d6b281d687ffec759b2a84f8a47ea2591f43539/Djblets-0.8.15.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.16": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "3e963387ba8a8cab437fa69cec13bc43e093fb2d9dfb7e31f9cc71336fd2f359", + "md5": "fb6fbfb1c2edb321eddb577a8a2c35b3", + "sha256": "d0edd00228a21bd21a7783da55290aad58c82319e0f892815e410e625bc0b92d" + }, + "downloads": -1, + "filename": "Djblets-0.8.16-py2.6.egg", + "has_sig": false, + "md5_digest": "fb6fbfb1c2edb321eddb577a8a2c35b3", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 738314, + "upload_time": "2015-10-29T21:08:28", + "upload_time_iso_8601": "2015-10-29T21:08:28.617754Z", + "url": "https://files.pythonhosted.org/packages/3e/96/3387ba8a8cab437fa69cec13bc43e093fb2d9dfb7e31f9cc71336fd2f359/Djblets-0.8.16-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "294e591f9fca8042bfd959f7348dc9cfe7ec46b6d8f551c0c5c0218e0d3483ee", + "md5": "6d7d0675fcacc4513dd92dd939d1990d", + "sha256": "e68caad065e7062959af730c65b7c281b2b2fb8d2a26ce6ee15211b3d78f641e" + }, + "downloads": -1, + "filename": "Djblets-0.8.16-py2.7.egg", + "has_sig": false, + "md5_digest": "6d7d0675fcacc4513dd92dd939d1990d", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 747199, + "upload_time": "2015-10-29T21:08:34", + "upload_time_iso_8601": "2015-10-29T21:08:34.415616Z", + "url": "https://files.pythonhosted.org/packages/29/4e/591f9fca8042bfd959f7348dc9cfe7ec46b6d8f551c0c5c0218e0d3483ee/Djblets-0.8.16-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "b0444206f70bed1458e008296df642ec58b78b3462363dea42dca93449667e5c", + "md5": "de79182f7ca30e1d11cddf228bf054ef", + "sha256": "2de6f70d808548c274903d6c12ed825d22901d8cffb204d257b73f524d9cb7dc" + }, + "downloads": -1, + "filename": "Djblets-0.8.16.tar.gz", + "has_sig": false, + "md5_digest": "de79182f7ca30e1d11cddf228bf054ef", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 275109, + "upload_time": "2015-10-29T21:08:39", + "upload_time_iso_8601": "2015-10-29T21:08:39.373749Z", + "url": "https://files.pythonhosted.org/packages/b0/44/4206f70bed1458e008296df642ec58b78b3462363dea42dca93449667e5c/Djblets-0.8.16.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.17": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "6e230a4021c0abe7a53c37ecbf8599db03e9aed1b9b0a71394042fba659410ac", + "md5": "a137b1eab3d42bfda3559bcb5bed3a3c", + "sha256": "3fa6f59a3d3cdf04cb93d39c4b3170db046a7ab4a78429e85d8254d08ce976a9" + }, + "downloads": -1, + "filename": "Djblets-0.8.17-py2.6.egg", + "has_sig": false, + "md5_digest": "a137b1eab3d42bfda3559bcb5bed3a3c", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 738226, + "upload_time": "2015-10-29T21:08:45", + "upload_time_iso_8601": "2015-10-29T21:08:45.446685Z", + "url": "https://files.pythonhosted.org/packages/6e/23/0a4021c0abe7a53c37ecbf8599db03e9aed1b9b0a71394042fba659410ac/Djblets-0.8.17-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "df9abf80563210f608506f5140dbc1a0b6af09bacb018643e45feea06773cea1", + "md5": "f74596cdf6377ea099a38f6fb0150867", + "sha256": "23f84013ea84e5f58589714d8dcc3dcefdeaa18d2d64198ef6e895755fb4f326" + }, + "downloads": -1, + "filename": "Djblets-0.8.17-py2.7.egg", + "has_sig": false, + "md5_digest": "f74596cdf6377ea099a38f6fb0150867", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 747201, + "upload_time": "2015-10-29T21:08:51", + "upload_time_iso_8601": "2015-10-29T21:08:51.682055Z", + "url": "https://files.pythonhosted.org/packages/df/9a/bf80563210f608506f5140dbc1a0b6af09bacb018643e45feea06773cea1/Djblets-0.8.17-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "9b932a0f9355177ba6002014f1aa3fc627dee2019489382833d96262ee7a6199", + "md5": "a72dbb2f4cf5d0248b4fcc9173105bb7", + "sha256": "9e65d016beab530f22ac0c04db2a5a2a15fa773df5818441d6085c68670a6281" + }, + "downloads": -1, + "filename": "Djblets-0.8.17.tar.gz", + "has_sig": false, + "md5_digest": "a72dbb2f4cf5d0248b4fcc9173105bb7", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 275135, + "upload_time": "2015-10-29T21:08:57", + "upload_time_iso_8601": "2015-10-29T21:08:57.975915Z", + "url": "https://files.pythonhosted.org/packages/9b/93/2a0f9355177ba6002014f1aa3fc627dee2019489382833d96262ee7a6199/Djblets-0.8.17.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.18": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "9dec5cc2398e447d8b8c731de3ed40935da19ffda4574887d6100e03eaf85607", + "md5": "ab1ba15e3013ea77705a65cc691a5923", + "sha256": "f810d280cbb6c9231a66590271d444c7ef085a1eb8b95e2a275af79f26f6f5f1" + }, + "downloads": -1, + "filename": "Djblets-0.8.18-py2.6.egg", + "has_sig": false, + "md5_digest": "ab1ba15e3013ea77705a65cc691a5923", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 739039, + "upload_time": "2015-10-29T21:09:07", + "upload_time_iso_8601": "2015-10-29T21:09:07.971241Z", + "url": "https://files.pythonhosted.org/packages/9d/ec/5cc2398e447d8b8c731de3ed40935da19ffda4574887d6100e03eaf85607/Djblets-0.8.18-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "0e320ba95c7f80edd35fc1ef3ff16195a9ff9a14ae035252faa2bbff46d2f24a", + "md5": "18670f0c6750a4e72d162133e72adf95", + "sha256": "7aa60ddb6c2a21d87ac4f4b8fc83e4be2e1e2ffa8bf9cf81a013da5386466c8d" + }, + "downloads": -1, + "filename": "Djblets-0.8.18-py2.7.egg", + "has_sig": false, + "md5_digest": "18670f0c6750a4e72d162133e72adf95", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 747953, + "upload_time": "2015-10-29T21:09:14", + "upload_time_iso_8601": "2015-10-29T21:09:14.232162Z", + "url": "https://files.pythonhosted.org/packages/0e/32/0ba95c7f80edd35fc1ef3ff16195a9ff9a14ae035252faa2bbff46d2f24a/Djblets-0.8.18-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "67f50502f4d6e945b598029a369b47cef1cd1cdde30e1ebe3bbbc43981de73bd", + "md5": "1cdcb6d02fb2066f4b01b74c8cde550a", + "sha256": "e227b2c644d2dc7c2d2b92007bf99beaf6ea7e1bb42e1e66ad5b1a058f56eeae" + }, + "downloads": -1, + "filename": "Djblets-0.8.18.tar.gz", + "has_sig": false, + "md5_digest": "1cdcb6d02fb2066f4b01b74c8cde550a", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 275427, + "upload_time": "2015-10-29T21:09:20", + "upload_time_iso_8601": "2015-10-29T21:09:20.067292Z", + "url": "https://files.pythonhosted.org/packages/67/f5/0502f4d6e945b598029a369b47cef1cd1cdde30e1ebe3bbbc43981de73bd/Djblets-0.8.18.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.19": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "4a3a1c2d0bc04a73287c78aa609b886379b7b00e05ac925b76efba8d228db070", + "md5": "f4dd2f6e481af3e88224f5bd180ab0c5", + "sha256": "c553e50a7df7604b330011b802927b00b7d89bcb93ca3fd637b2847f44852f3b" + }, + "downloads": -1, + "filename": "Djblets-0.8.19-py2.6.egg", + "has_sig": false, + "md5_digest": "f4dd2f6e481af3e88224f5bd180ab0c5", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 743046, + "upload_time": "2015-10-29T21:09:26", + "upload_time_iso_8601": "2015-10-29T21:09:26.635046Z", + "url": "https://files.pythonhosted.org/packages/4a/3a/1c2d0bc04a73287c78aa609b886379b7b00e05ac925b76efba8d228db070/Djblets-0.8.19-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "e3df3bbc1a9c6f739d9529f90da6ec3e85b53444571bb786f7086e70dd3161c5", + "md5": "37695bf11f38bd2c18e899812ea7a5df", + "sha256": "8a247640db8652152635274e6b785df179d03a6ec9d038e168c0666128fa286e" + }, + "downloads": -1, + "filename": "Djblets-0.8.19-py2.7.egg", + "has_sig": false, + "md5_digest": "37695bf11f38bd2c18e899812ea7a5df", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 751924, + "upload_time": "2015-10-29T21:09:33", + "upload_time_iso_8601": "2015-10-29T21:09:33.204979Z", + "url": "https://files.pythonhosted.org/packages/e3/df/3bbc1a9c6f739d9529f90da6ec3e85b53444571bb786f7086e70dd3161c5/Djblets-0.8.19-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "14b3bb532c52d9a7824f6926815eb81ae9795ec17cef5801068b809dd59f77dc", + "md5": "65661d4bc1c66741c67ec433ba4f0411", + "sha256": "a9f74c7739cce597ee74373864fd4864a6959aaf2ebc85d40c91fa2d2d735988" + }, + "downloads": -1, + "filename": "Djblets-0.8.19-py2-none-any.whl", + "has_sig": false, + "md5_digest": "65661d4bc1c66741c67ec433ba4f0411", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 505927, + "upload_time": "2015-10-29T21:05:06", + "upload_time_iso_8601": "2015-10-29T21:05:06.959715Z", + "url": "https://files.pythonhosted.org/packages/14/b3/bb532c52d9a7824f6926815eb81ae9795ec17cef5801068b809dd59f77dc/Djblets-0.8.19-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "d19d514b938c76795a80d4c23e6fa855740540f662d83f8329d929895df1aeb3", + "md5": "8009baa85e18551586acd5daef6ead85", + "sha256": "04121639af82e8e86eacf2dc312ea6f2100baaea0f6b35a4b88e188f83ab1837" + }, + "downloads": -1, + "filename": "Djblets-0.8.19.tar.gz", + "has_sig": false, + "md5_digest": "8009baa85e18551586acd5daef6ead85", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 277608, + "upload_time": "2015-10-29T21:09:38", + "upload_time_iso_8601": "2015-10-29T21:09:38.790874Z", + "url": "https://files.pythonhosted.org/packages/d1/9d/514b938c76795a80d4c23e6fa855740540f662d83f8329d929895df1aeb3/Djblets-0.8.19.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.2": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "8b701526e7ec70e9120dfee6fe041607addaafd07e9eb6fdc30873b4e597eefa", + "md5": "1fb44561bd46e63520aad82541f6e39c", + "sha256": "a66b94a4745fa4cf4ba86fcf8adc0ab2decebe97a44e922b37d582bfc30bdd79" + }, + "downloads": -1, + "filename": "Djblets-0.8.2-py2.6.egg", + "has_sig": false, + "md5_digest": "1fb44561bd46e63520aad82541f6e39c", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 644606, + "upload_time": "2015-10-29T21:09:45", + "upload_time_iso_8601": "2015-10-29T21:09:45.113780Z", + "url": "https://files.pythonhosted.org/packages/8b/70/1526e7ec70e9120dfee6fe041607addaafd07e9eb6fdc30873b4e597eefa/Djblets-0.8.2-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "a3fcc01b2a47bda4eb324d92ed3abd3f5303d2bb9e070872a7272ddf41b11d32", + "md5": "be01ade2d2d766559ba53513f7ce69ee", + "sha256": "11bcc384086192864539b6d71c2fc56308f3a048ac0a6a1cb47255569106f8b9" + }, + "downloads": -1, + "filename": "Djblets-0.8.2-py2.7.egg", + "has_sig": false, + "md5_digest": "be01ade2d2d766559ba53513f7ce69ee", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 653728, + "upload_time": "2015-10-29T21:09:51", + "upload_time_iso_8601": "2015-10-29T21:09:51.243409Z", + "url": "https://files.pythonhosted.org/packages/a3/fc/c01b2a47bda4eb324d92ed3abd3f5303d2bb9e070872a7272ddf41b11d32/Djblets-0.8.2-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "c913206260f99c38eceeca876a1ff1e45c6880d5c344980502f031e23532c55c", + "md5": "efce164fa8daaa934cb24f47efc153c0", + "sha256": "c2f66c3f74673b3b9c0f533b29076f8ef6ad08ddbe28f2d74b37eade2fa54d36" + }, + "downloads": -1, + "filename": "Djblets-0.8.2.tar.gz", + "has_sig": false, + "md5_digest": "efce164fa8daaa934cb24f47efc153c0", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 267202, + "upload_time": "2015-10-29T21:09:57", + "upload_time_iso_8601": "2015-10-29T21:09:57.412044Z", + "url": "https://files.pythonhosted.org/packages/c9/13/206260f99c38eceeca876a1ff1e45c6880d5c344980502f031e23532c55c/Djblets-0.8.2.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.20": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "b3def89440d6e05c62a07735139712c491c17e6e780cece8b1749dd23bd2906d", + "md5": "d0d6c04c316ed38c1c08141a5b5c6e4d", + "sha256": "d95ac5a97e6d79310e67f90e47e3ec088408cd7dee5c6863488c0a722df4a0a4" + }, + "downloads": -1, + "filename": "Djblets-0.8.20-py2.6.egg", + "has_sig": false, + "md5_digest": "d0d6c04c316ed38c1c08141a5b5c6e4d", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 743572, + "upload_time": "2015-10-29T21:10:20", + "upload_time_iso_8601": "2015-10-29T21:10:20.688860Z", + "url": "https://files.pythonhosted.org/packages/b3/de/f89440d6e05c62a07735139712c491c17e6e780cece8b1749dd23bd2906d/Djblets-0.8.20-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "03928e723c16406c78ebb361afc524eb5791353be41473ea5385c320464983a1", + "md5": "05168003fc984aaf692b4d672cecf419", + "sha256": "ba27060c006d3df04f06c252e3964d043483a6180cc54b1a519b60971d2993eb" + }, + "downloads": -1, + "filename": "Djblets-0.8.20-py2.7.egg", + "has_sig": false, + "md5_digest": "05168003fc984aaf692b4d672cecf419", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 752550, + "upload_time": "2015-10-29T21:10:31", + "upload_time_iso_8601": "2015-10-29T21:10:31.118458Z", + "url": "https://files.pythonhosted.org/packages/03/92/8e723c16406c78ebb361afc524eb5791353be41473ea5385c320464983a1/Djblets-0.8.20-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "8386394ac5fc3c1cbc5612038de5de6181cbcff95500d67c6f0ec94190860295", + "md5": "3354070b07aa306157f29891efdadf2c", + "sha256": "8b4301c6cea227b2c6789b64de94b7aa77558a4c9760e368812cd8fc8c2bcfb4" + }, + "downloads": -1, + "filename": "Djblets-0.8.20-py2-none-any.whl", + "has_sig": false, + "md5_digest": "3354070b07aa306157f29891efdadf2c", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 506154, + "upload_time": "2015-10-29T21:05:14", + "upload_time_iso_8601": "2015-10-29T21:05:14.489556Z", + "url": "https://files.pythonhosted.org/packages/83/86/394ac5fc3c1cbc5612038de5de6181cbcff95500d67c6f0ec94190860295/Djblets-0.8.20-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "c57fa71011ba76d1dd22770887abfee11d523cbbd458f345464b185c10f298f5", + "md5": "f5f79d6502e1977188553ae56e93719d", + "sha256": "ba814cf0600a7486f09301f7103933b04acd2aeb978e61ba18114e88a966df57" + }, + "downloads": -1, + "filename": "Djblets-0.8.20.tar.gz", + "has_sig": false, + "md5_digest": "f5f79d6502e1977188553ae56e93719d", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 277880, + "upload_time": "2015-10-29T21:10:39", + "upload_time_iso_8601": "2015-10-29T21:10:39.796073Z", + "url": "https://files.pythonhosted.org/packages/c5/7f/a71011ba76d1dd22770887abfee11d523cbbd458f345464b185c10f298f5/Djblets-0.8.20.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.21": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "73413aad7d55eda9910bcd18b7de8f53193d32648ec21d68665dda1b7ecc6576", + "md5": "3035e32b5bcd124cccd02b9857d926d6", + "sha256": "28f3027350b5ee7c1636d2ac1a5a90d90160f88b6e0bfcafddfcb9e5d0c66143" + }, + "downloads": -1, + "filename": "Djblets-0.8.21-py2.6.egg", + "has_sig": false, + "md5_digest": "3035e32b5bcd124cccd02b9857d926d6", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 745593, + "upload_time": "2015-10-29T21:10:48", + "upload_time_iso_8601": "2015-10-29T21:10:48.417361Z", + "url": "https://files.pythonhosted.org/packages/73/41/3aad7d55eda9910bcd18b7de8f53193d32648ec21d68665dda1b7ecc6576/Djblets-0.8.21-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "76a619aa80503ef01364c33f9a9c9a05858b64501c34ef954b21a1bc79a19dcb", + "md5": "370b2da10d12f83e2dfcb0d1d587e75e", + "sha256": "1eb6a1ee3afc605da37051a936b8d987522c9674187d56150dfcc3e2c31375c9" + }, + "downloads": -1, + "filename": "Djblets-0.8.21-py2.7.egg", + "has_sig": false, + "md5_digest": "370b2da10d12f83e2dfcb0d1d587e75e", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 754345, + "upload_time": "2015-10-29T21:10:56", + "upload_time_iso_8601": "2015-10-29T21:10:56.610547Z", + "url": "https://files.pythonhosted.org/packages/76/a6/19aa80503ef01364c33f9a9c9a05858b64501c34ef954b21a1bc79a19dcb/Djblets-0.8.21-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "75fc98293c127c32ab695c31a1b211c62e6ffc16ffb241209328e7e63c5d2766", + "md5": "8ac6f2189b7c955c9692b9bf12d7ccfe", + "sha256": "eec390e094745317bd1cf9508f3645158fe44088081eb8b5f71dd1288f2b43c8" + }, + "downloads": -1, + "filename": "Djblets-0.8.21-py2-none-any.whl", + "has_sig": false, + "md5_digest": "8ac6f2189b7c955c9692b9bf12d7ccfe", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 507164, + "upload_time": "2015-10-29T21:05:21", + "upload_time_iso_8601": "2015-10-29T21:05:21.789931Z", + "url": "https://files.pythonhosted.org/packages/75/fc/98293c127c32ab695c31a1b211c62e6ffc16ffb241209328e7e63c5d2766/Djblets-0.8.21-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "884a8ff7b5382047882745be3dd874dc9ceaed12ab8b4d411b0d5567f291fc46", + "md5": "a25d76d815ce4f258294e4a18a93626c", + "sha256": "30eeb641c5293cd2511e8712f55f4c5976f58a8f76404dd98b2103588abbbc56" + }, + "downloads": -1, + "filename": "Djblets-0.8.21.tar.gz", + "has_sig": false, + "md5_digest": "a25d76d815ce4f258294e4a18a93626c", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 278774, + "upload_time": "2015-10-29T21:11:02", + "upload_time_iso_8601": "2015-10-29T21:11:02.588393Z", + "url": "https://files.pythonhosted.org/packages/88/4a/8ff7b5382047882745be3dd874dc9ceaed12ab8b4d411b0d5567f291fc46/Djblets-0.8.21.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.22": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "8c518f4a2bebe1ee6b06020f972d1d99a5d9fd7cc63cc0c19b5f6cdd9fb264ad", + "md5": "18b58fbd545508d12e83bf6ec26022f6", + "sha256": "fb7eda9158b27955338d7ac41e7cb7d4c1b7dbf34c638d33184ced89d442a900" + }, + "downloads": -1, + "filename": "Djblets-0.8.22-py2.6.egg", + "has_sig": false, + "md5_digest": "18b58fbd545508d12e83bf6ec26022f6", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 753112, + "upload_time": "2015-10-29T21:11:09", + "upload_time_iso_8601": "2015-10-29T21:11:09.664530Z", + "url": "https://files.pythonhosted.org/packages/8c/51/8f4a2bebe1ee6b06020f972d1d99a5d9fd7cc63cc0c19b5f6cdd9fb264ad/Djblets-0.8.22-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "9ccaa5c5d61404f6d193e14d273b090ca8ef099fbf9a2a6ca4156379766746ab", + "md5": "840821cb56ea30467a59dd08d4e6eaa6", + "sha256": "20ff580b0aaed94af5b28b363e6c9d23f283cba669696eb660c5cba7159cb5c6" + }, + "downloads": -1, + "filename": "Djblets-0.8.22-py2.7.egg", + "has_sig": false, + "md5_digest": "840821cb56ea30467a59dd08d4e6eaa6", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 761831, + "upload_time": "2015-10-29T21:11:16", + "upload_time_iso_8601": "2015-10-29T21:11:16.215481Z", + "url": "https://files.pythonhosted.org/packages/9c/ca/a5c5d61404f6d193e14d273b090ca8ef099fbf9a2a6ca4156379766746ab/Djblets-0.8.22-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "107a9f759de02b11d13cf08e797f2cfb6a6f5cd497808a03a9b35a5dbda5267e", + "md5": "0e9978614152c4e90c2fea9f710bd36f", + "sha256": "dece755310fe28b90f899a50ee2489e42aa8b4b84119285f07a4fc216ec56c43" + }, + "downloads": -1, + "filename": "Djblets-0.8.22-py2-none-any.whl", + "has_sig": false, + "md5_digest": "0e9978614152c4e90c2fea9f710bd36f", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 512269, + "upload_time": "2015-10-29T21:05:33", + "upload_time_iso_8601": "2015-10-29T21:05:33.193282Z", + "url": "https://files.pythonhosted.org/packages/10/7a/9f759de02b11d13cf08e797f2cfb6a6f5cd497808a03a9b35a5dbda5267e/Djblets-0.8.22-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "c753b2785b378911494d1216c41f758f9637bc75318e58051d8abae821e5b3d1", + "md5": "5090bca9dcd6fe5d41a312db7c6849bf", + "sha256": "3f80420e8ac4406ab083b0eb72675279e0b2cdedf5fc03972520f637422a8af6" + }, + "downloads": -1, + "filename": "Djblets-0.8.22.tar.gz", + "has_sig": false, + "md5_digest": "5090bca9dcd6fe5d41a312db7c6849bf", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 282117, + "upload_time": "2015-10-29T21:11:21", + "upload_time_iso_8601": "2015-10-29T21:11:21.857612Z", + "url": "https://files.pythonhosted.org/packages/c7/53/b2785b378911494d1216c41f758f9637bc75318e58051d8abae821e5b3d1/Djblets-0.8.22.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.23": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "d2528f045ccd3622305cc0265ab2939c5b5d1362b5c83f882f15c7ffe7dc224d", + "md5": "a19faa04c02792c652b54a439bd61715", + "sha256": "ef29e03176385a761f3e50113a0b6ab37e9f9475b8f0924ed649fadb53803fd5" + }, + "downloads": -1, + "filename": "Djblets-0.8.23-py2.6.egg", + "has_sig": false, + "md5_digest": "a19faa04c02792c652b54a439bd61715", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 753399, + "upload_time": "2015-10-29T21:11:27", + "upload_time_iso_8601": "2015-10-29T21:11:27.696254Z", + "url": "https://files.pythonhosted.org/packages/d2/52/8f045ccd3622305cc0265ab2939c5b5d1362b5c83f882f15c7ffe7dc224d/Djblets-0.8.23-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "749bd7da8a055e8198bf565a8c25bb45476ba0b31f29a305d427e37d29f7171b", + "md5": "367095e248a0233ecec4536cd2247575", + "sha256": "8a84dda2805bb2600acf72607355b5ffc21bb0a8e75b980d14385ed564e8f720" + }, + "downloads": -1, + "filename": "Djblets-0.8.23-py2.7.egg", + "has_sig": false, + "md5_digest": "367095e248a0233ecec4536cd2247575", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 762811, + "upload_time": "2015-10-29T21:11:34", + "upload_time_iso_8601": "2015-10-29T21:11:34.666483Z", + "url": "https://files.pythonhosted.org/packages/74/9b/d7da8a055e8198bf565a8c25bb45476ba0b31f29a305d427e37d29f7171b/Djblets-0.8.23-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "223b5013e16b37b3bde7158536ff3fa38a3686ff06877f51c0da1e08c8965248", + "md5": "57bb972ed08f76884c40d69d116fd90c", + "sha256": "6593b08904a9e956dd9fccf5dbe01c6b59c0e110d2994590d8a5b3626c2fb967" + }, + "downloads": -1, + "filename": "Djblets-0.8.23-py2-none-any.whl", + "has_sig": false, + "md5_digest": "57bb972ed08f76884c40d69d116fd90c", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 512398, + "upload_time": "2015-10-29T21:05:39", + "upload_time_iso_8601": "2015-10-29T21:05:39.310849Z", + "url": "https://files.pythonhosted.org/packages/22/3b/5013e16b37b3bde7158536ff3fa38a3686ff06877f51c0da1e08c8965248/Djblets-0.8.23-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "b05b81bed54b849b7d2b8435570ebfd92ce7aba80abe3850755aa81c3a29cbf3", + "md5": "3d2984df3a11b658f254f89971d2c83e", + "sha256": "6bfbdb8180e5ba39f898a3449433d414263c36da0860d95e184661fedadfda4a" + }, + "downloads": -1, + "filename": "Djblets-0.8.23.tar.gz", + "has_sig": false, + "md5_digest": "3d2984df3a11b658f254f89971d2c83e", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 282227, + "upload_time": "2015-10-29T21:11:40", + "upload_time_iso_8601": "2015-10-29T21:11:40.771833Z", + "url": "https://files.pythonhosted.org/packages/b0/5b/81bed54b849b7d2b8435570ebfd92ce7aba80abe3850755aa81c3a29cbf3/Djblets-0.8.23.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.24": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "66b269a5519610b70fb4692a762b48d21b165f98e4333c90f768228d2ac99678", + "md5": "19c218add09c0cfaa6d017cee3e75e93", + "sha256": "059311bcee107f744eb58b93d44f7c5c3ff1eaa255b38df4ab5b5542fc190f90" + }, + "downloads": -1, + "filename": "Djblets-0.8.24-py2.6.egg", + "has_sig": false, + "md5_digest": "19c218add09c0cfaa6d017cee3e75e93", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 755849, + "upload_time": "2016-02-22T06:41:30", + "upload_time_iso_8601": "2016-02-22T06:41:30.654001Z", + "url": "https://files.pythonhosted.org/packages/66/b2/69a5519610b70fb4692a762b48d21b165f98e4333c90f768228d2ac99678/Djblets-0.8.24-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "9c01a62df242ae0d8eab708d4e829b1eb56107f202fee9b4aaa9f27017dbcdba", + "md5": "84a0d24b9b93aaddc7287d46ac8c3793", + "sha256": "33e3fa84ecddaa6b05280f2ce3f6676dd8153890bf5e65f9643d261826b5a72f" + }, + "downloads": -1, + "filename": "Djblets-0.8.24-py2.7.egg", + "has_sig": false, + "md5_digest": "84a0d24b9b93aaddc7287d46ac8c3793", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 781003, + "upload_time": "2016-02-22T06:41:40", + "upload_time_iso_8601": "2016-02-22T06:41:40.381843Z", + "url": "https://files.pythonhosted.org/packages/9c/01/a62df242ae0d8eab708d4e829b1eb56107f202fee9b4aaa9f27017dbcdba/Djblets-0.8.24-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "845582fd56463be4babb4d15e50ac09f15bbaa1f0b8091cc44afc61b6a46c4fa", + "md5": "2518761bc8d2897b5c631bd2e7533f3f", + "sha256": "7e20b1df81f11ffb7419f8fd93dee8f16dd9947428d4c2d80546d12e6433cd80" + }, + "downloads": -1, + "filename": "Djblets-0.8.24-py2-none-any.whl", + "has_sig": false, + "md5_digest": "2518761bc8d2897b5c631bd2e7533f3f", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 529969, + "upload_time": "2016-02-22T06:41:23", + "upload_time_iso_8601": "2016-02-22T06:41:23.704360Z", + "url": "https://files.pythonhosted.org/packages/84/55/82fd56463be4babb4d15e50ac09f15bbaa1f0b8091cc44afc61b6a46c4fa/Djblets-0.8.24-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "4b5c9291d219c39943a762d0e3451c03007459f8a816acdc982a0e9934611d40", + "md5": "c49ed3db08a7c47a48e4faeaa46bd5e4", + "sha256": "f22c87904054b83f70d164f7d4ff9c7de6ebcad2be743ddf76e54578d1084d7a" + }, + "downloads": -1, + "filename": "Djblets-0.8.24.tar.gz", + "has_sig": false, + "md5_digest": "c49ed3db08a7c47a48e4faeaa46bd5e4", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 290411, + "upload_time": "2016-02-22T06:41:51", + "upload_time_iso_8601": "2016-02-22T06:41:51.694397Z", + "url": "https://files.pythonhosted.org/packages/4b/5c/9291d219c39943a762d0e3451c03007459f8a816acdc982a0e9934611d40/Djblets-0.8.24.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.25": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "0647948b65a91e241513cf18b563bace6cfb67bb3f3d70b21ec935b0cfc0a1af", + "md5": "4186165bf0f63bceabfd73b689c5bee4", + "sha256": "d8a89cf00f6bd232b95fd0faa8e545286a7745da33cb67ef25f2b91b47ff8c1b" + }, + "downloads": -1, + "filename": "Djblets-0.8.25-py2.6.egg", + "has_sig": false, + "md5_digest": "4186165bf0f63bceabfd73b689c5bee4", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 756053, + "upload_time": "2016-03-01T19:01:39", + "upload_time_iso_8601": "2016-03-01T19:01:39.344026Z", + "url": "https://files.pythonhosted.org/packages/06/47/948b65a91e241513cf18b563bace6cfb67bb3f3d70b21ec935b0cfc0a1af/Djblets-0.8.25-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "e100e4e8c79962062c77875672c8795e290ff284dd9b2fd82e52fe5e4300f705", + "md5": "62e1aeaba8fa13fd7149346c8031f657", + "sha256": "875fdb82684890fe1e14ba953291acb86ab13123c272784a5d4874491c22bd61" + }, + "downloads": -1, + "filename": "Djblets-0.8.25-py2.7.egg", + "has_sig": false, + "md5_digest": "62e1aeaba8fa13fd7149346c8031f657", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 764791, + "upload_time": "2016-03-01T19:01:46", + "upload_time_iso_8601": "2016-03-01T19:01:46.070982Z", + "url": "https://files.pythonhosted.org/packages/e1/00/e4e8c79962062c77875672c8795e290ff284dd9b2fd82e52fe5e4300f705/Djblets-0.8.25-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "b058843ecec3665a888652a631378208549236e9004cce51f28e97817de8b7b5", + "md5": "7329d58419ac99178eacfebe0750d391", + "sha256": "d9ec4fe0c0798a5fb5463fb047e124689ed336f6684806184677b8bb74a98fad" + }, + "downloads": -1, + "filename": "Djblets-0.8.25-py2-none-any.whl", + "has_sig": false, + "md5_digest": "7329d58419ac99178eacfebe0750d391", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 513368, + "upload_time": "2016-03-01T19:01:32", + "upload_time_iso_8601": "2016-03-01T19:01:32.512174Z", + "url": "https://files.pythonhosted.org/packages/b0/58/843ecec3665a888652a631378208549236e9004cce51f28e97817de8b7b5/Djblets-0.8.25-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "081f6abac707c0ae841eaf21d574a2fa05e121053f35c32b36841de616f9b619", + "md5": "29999f0bc97c24ccc6b90c807f8b8afa", + "sha256": "23ad79868b28afbf08be1fb21198b958c9ffafc8820802e75c74188721249b2c" + }, + "downloads": -1, + "filename": "Djblets-0.8.25.tar.gz", + "has_sig": false, + "md5_digest": "29999f0bc97c24ccc6b90c807f8b8afa", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 283201, + "upload_time": "2016-03-01T19:01:52", + "upload_time_iso_8601": "2016-03-01T19:01:52.507584Z", + "url": "https://files.pythonhosted.org/packages/08/1f/6abac707c0ae841eaf21d574a2fa05e121053f35c32b36841de616f9b619/Djblets-0.8.25.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.26": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "e18c39ec48c6819eca353f217b96dfd1a9783382b64fce21432c66b9b0df8496", + "md5": "0e92c9852a482fb138511391b7bb8acf", + "sha256": "c973bed37bec942bd838c8e4e7a0a9d3c67c03186a77dfb38bea4000f0515063" + }, + "downloads": -1, + "filename": "Djblets-0.8.26-py2.6.egg", + "has_sig": false, + "md5_digest": "0e92c9852a482fb138511391b7bb8acf", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 757511, + "upload_time": "2016-04-13T21:12:11", + "upload_time_iso_8601": "2016-04-13T21:12:11.147802Z", + "url": "https://files.pythonhosted.org/packages/e1/8c/39ec48c6819eca353f217b96dfd1a9783382b64fce21432c66b9b0df8496/Djblets-0.8.26-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "40ff9f1c5bb9c5b62adb6552ae18624dd0c3be166ddb1d2cf093c3da7c902087", + "md5": "93f8cbf2a01358e8fb05c6ea75b2d269", + "sha256": "f5e8975a5682d2da620d8ec3de29c591a74d62b586b64b4a92af7b1b74072ae3" + }, + "downloads": -1, + "filename": "Djblets-0.8.26-py2.7.egg", + "has_sig": false, + "md5_digest": "93f8cbf2a01358e8fb05c6ea75b2d269", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 766226, + "upload_time": "2016-04-13T21:12:34", + "upload_time_iso_8601": "2016-04-13T21:12:34.463556Z", + "url": "https://files.pythonhosted.org/packages/40/ff/9f1c5bb9c5b62adb6552ae18624dd0c3be166ddb1d2cf093c3da7c902087/Djblets-0.8.26-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "b53dcdb562c00d8be489578d33ec34e747265edd5262ffbcdaec51414b33f65c", + "md5": "119849d2c07c41a3d8c03d899e5ddd78", + "sha256": "6adb252947178b504244c9ed2610eebd97c6b09e8ba3bdcf9becdbca47c650bc" + }, + "downloads": -1, + "filename": "Djblets-0.8.26-py2-none-any.whl", + "has_sig": false, + "md5_digest": "119849d2c07c41a3d8c03d899e5ddd78", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 514164, + "upload_time": "2016-04-13T21:11:57", + "upload_time_iso_8601": "2016-04-13T21:11:57.025631Z", + "url": "https://files.pythonhosted.org/packages/b5/3d/cdb562c00d8be489578d33ec34e747265edd5262ffbcdaec51414b33f65c/Djblets-0.8.26-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "63af056861fdc42bc43a303d197b1e7c784db368b4f9e7b99acf5efafe537bde", + "md5": "8e92b35fd000c3e3f1c65b9f04081a03", + "sha256": "3ee289e87b759a038f2c1ba8e7f1039f6e021d61ab629113cf09624af65280c7" + }, + "downloads": -1, + "filename": "Djblets-0.8.26.tar.gz", + "has_sig": false, + "md5_digest": "8e92b35fd000c3e3f1c65b9f04081a03", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 283969, + "upload_time": "2016-04-13T21:12:52", + "upload_time_iso_8601": "2016-04-13T21:12:52.594112Z", + "url": "https://files.pythonhosted.org/packages/63/af/056861fdc42bc43a303d197b1e7c784db368b4f9e7b99acf5efafe537bde/Djblets-0.8.26.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.27": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "6b72cbee4d570637f8b4089d1d3b7b947251b7bbfe68dae20fd92d5f11bb9f70", + "md5": "e84c088845d034ee984204fc775c80ca", + "sha256": "f4707dbe7b5bbbdb819e7fa05181f322e72f0e8dfe298f393aff2a56a0f06fd9" + }, + "downloads": -1, + "filename": "Djblets-0.8.27-py2.6.egg", + "has_sig": false, + "md5_digest": "e84c088845d034ee984204fc775c80ca", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 772534, + "upload_time": "2016-10-18T01:59:43", + "upload_time_iso_8601": "2016-10-18T01:59:43.661524Z", + "url": "https://files.pythonhosted.org/packages/6b/72/cbee4d570637f8b4089d1d3b7b947251b7bbfe68dae20fd92d5f11bb9f70/Djblets-0.8.27-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "7464becc366408c0d9c035d9bdbab8af5864505bdb8d146fc2445a26cc2ba24f", + "md5": "cfac3af3cf2015bb7728c5d64312add5", + "sha256": "fa0d13892ab737d91bbc89ee9c3c9d9af65c473d16faf8122cf087574502b2d6" + }, + "downloads": -1, + "filename": "Djblets-0.8.27-py2.7.egg", + "has_sig": false, + "md5_digest": "cfac3af3cf2015bb7728c5d64312add5", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 781231, + "upload_time": "2016-10-18T01:59:48", + "upload_time_iso_8601": "2016-10-18T01:59:48.319277Z", + "url": "https://files.pythonhosted.org/packages/74/64/becc366408c0d9c035d9bdbab8af5864505bdb8d146fc2445a26cc2ba24f/Djblets-0.8.27-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "94bbff1f359addb5ea1522ecb866b426f92dab096a523b8976f22870f989272b", + "md5": "2ffcb91eb8ebc7a9b8f2554b3a22497e", + "sha256": "849769be7d1ae5344de55b74dd16bbc884bf164124a038a79072a7e9eaca5648" + }, + "downloads": -1, + "filename": "Djblets-0.8.27-py2-none-any.whl", + "has_sig": false, + "md5_digest": "2ffcb91eb8ebc7a9b8f2554b3a22497e", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 521332, + "upload_time": "2016-10-18T01:59:39", + "upload_time_iso_8601": "2016-10-18T01:59:39.668555Z", + "url": "https://files.pythonhosted.org/packages/94/bb/ff1f359addb5ea1522ecb866b426f92dab096a523b8976f22870f989272b/Djblets-0.8.27-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "6b23b407d2b768890f690852151966e2f8ab4f7011088c32c31763262ef135b3", + "md5": "24eb5f955af6e343e59b79cee2c9662b", + "sha256": "04c27c5f174b5bc87bf72ec2801b453753ba4a85a6f9fa817ef3aa8cb56b55d7" + }, + "downloads": -1, + "filename": "Djblets-0.8.27.tar.gz", + "has_sig": false, + "md5_digest": "24eb5f955af6e343e59b79cee2c9662b", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 290665, + "upload_time": "2016-10-18T01:59:51", + "upload_time_iso_8601": "2016-10-18T01:59:51.726624Z", + "url": "https://files.pythonhosted.org/packages/6b/23/b407d2b768890f690852151966e2f8ab4f7011088c32c31763262ef135b3/Djblets-0.8.27.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.28": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "177778f72a5eda600ac42cf84361e0c4542cdedb66e90583585ac1c20ee5d3a3", + "md5": "8594823091ba6a3658aa56cbe167b747", + "sha256": "f095517b57961092d91f100eda3eb191bddec8d3fe910e27a50503f61c72a76e" + }, + "downloads": -1, + "filename": "Djblets-0.8.28-py2.6.egg", + "has_sig": false, + "md5_digest": "8594823091ba6a3658aa56cbe167b747", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 780325, + "upload_time": "2017-02-09T23:05:59", + "upload_time_iso_8601": "2017-02-09T23:05:59.452791Z", + "url": "https://files.pythonhosted.org/packages/17/77/78f72a5eda600ac42cf84361e0c4542cdedb66e90583585ac1c20ee5d3a3/Djblets-0.8.28-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "19d8776ec2f3d542c42da2ab2a7547339e3ecad9f9337451c76d035868c80f5f", + "md5": "c77653fd829ab46853543bef52215fe8", + "sha256": "a791766e7381c9a2b3c8f07c18f3d67d0f2ca3ca562bf6046a2f29dc835b8d7f" + }, + "downloads": -1, + "filename": "Djblets-0.8.28-py2.7.egg", + "has_sig": false, + "md5_digest": "c77653fd829ab46853543bef52215fe8", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 788884, + "upload_time": "2017-02-09T23:06:03", + "upload_time_iso_8601": "2017-02-09T23:06:03.196827Z", + "url": "https://files.pythonhosted.org/packages/19/d8/776ec2f3d542c42da2ab2a7547339e3ecad9f9337451c76d035868c80f5f/Djblets-0.8.28-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "6eb751f72f12ff38ea37959da64a2e6724ea4399e429ea37193517c218cf6625", + "md5": "7d5fa2440bd4b05b71489a523e814853", + "sha256": "7363cdd4f1c41671e146f955bf2c1be4a69bbf86aa5a3f9514350db6e10db073" + }, + "downloads": -1, + "filename": "Djblets-0.8.28-py2-none-any.whl", + "has_sig": false, + "md5_digest": "7d5fa2440bd4b05b71489a523e814853", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 524358, + "upload_time": "2017-02-09T23:05:55", + "upload_time_iso_8601": "2017-02-09T23:05:55.862095Z", + "url": "https://files.pythonhosted.org/packages/6e/b7/51f72f12ff38ea37959da64a2e6724ea4399e429ea37193517c218cf6625/Djblets-0.8.28-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "684d15e6c1cd9e47b4b34eb8d6c18bd1d9dd220535b7fa926a903b457e68005d", + "md5": "1074f5570c831186a6853dcc1a173add", + "sha256": "85a0c7bcf0248d2a32deefcc2f70a254b3ef67b014aeb2d62b01e077f6a1f5fe" + }, + "downloads": -1, + "filename": "Djblets-0.8.28.tar.gz", + "has_sig": false, + "md5_digest": "1074f5570c831186a6853dcc1a173add", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 290481, + "upload_time": "2017-02-09T23:06:06", + "upload_time_iso_8601": "2017-02-09T23:06:06.737009Z", + "url": "https://files.pythonhosted.org/packages/68/4d/15e6c1cd9e47b4b34eb8d6c18bd1d9dd220535b7fa926a903b457e68005d/Djblets-0.8.28.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.29": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "636d5a76d0e388b3caf387437c2f77109effefe620d99c1f463fdc10bbb3d5dc", + "md5": "ef0c0d105e84d2bd1f872a3cd5e26505", + "sha256": "c2f85157d1f1c16bca7ab01c0507076f02f78d7e49158bd88bcb5717b68f99f0" + }, + "downloads": -1, + "filename": "Djblets-0.8.29-py2.6.egg", + "has_sig": false, + "md5_digest": "ef0c0d105e84d2bd1f872a3cd5e26505", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 780628, + "upload_time": "2017-02-16T02:03:03", + "upload_time_iso_8601": "2017-02-16T02:03:03.640010Z", + "url": "https://files.pythonhosted.org/packages/63/6d/5a76d0e388b3caf387437c2f77109effefe620d99c1f463fdc10bbb3d5dc/Djblets-0.8.29-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "c8433def730ea8783275a79eb4ed84822010657fd96bcc5cf152391668fc2e6a", + "md5": "895651228d83769d6db0d6fd24418185", + "sha256": "641f1755276dea786f4f0fd1dcea56c71faf2db881fef3864d395d3e0ed3c416" + }, + "downloads": -1, + "filename": "Djblets-0.8.29-py2.7.egg", + "has_sig": false, + "md5_digest": "895651228d83769d6db0d6fd24418185", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 789214, + "upload_time": "2017-02-16T02:03:07", + "upload_time_iso_8601": "2017-02-16T02:03:07.234660Z", + "url": "https://files.pythonhosted.org/packages/c8/43/3def730ea8783275a79eb4ed84822010657fd96bcc5cf152391668fc2e6a/Djblets-0.8.29-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "b3ae5bcd78f3a01fa33865b81e931eb1d7773034bf4a4682b0eb8a75806ed319", + "md5": "a4c590bea87550f87245d7b489928591", + "sha256": "77b5c693c12e29c8f7633a393568d7aa6036d32b06cfde7288aabeaa2bc021e8" + }, + "downloads": -1, + "filename": "Djblets-0.8.29-py2-none-any.whl", + "has_sig": false, + "md5_digest": "a4c590bea87550f87245d7b489928591", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 524681, + "upload_time": "2017-02-16T02:03:00", + "upload_time_iso_8601": "2017-02-16T02:03:00.043309Z", + "url": "https://files.pythonhosted.org/packages/b3/ae/5bcd78f3a01fa33865b81e931eb1d7773034bf4a4682b0eb8a75806ed319/Djblets-0.8.29-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "9eab649be4d31b581c61b930aafe8c09b431883431bcdda63ea200de7cab74f3", + "md5": "e8ee0bd1c3d84d096977ac351d8d3ffd", + "sha256": "ee11dd82c904b0f26a5edc633ab5fa6fbe6e36bdebc97e1acc52ae44f4f0ce42" + }, + "downloads": -1, + "filename": "Djblets-0.8.29.tar.gz", + "has_sig": false, + "md5_digest": "e8ee0bd1c3d84d096977ac351d8d3ffd", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 289892, + "upload_time": "2017-02-16T02:03:10", + "upload_time_iso_8601": "2017-02-16T02:03:10.400432Z", + "url": "https://files.pythonhosted.org/packages/9e/ab/649be4d31b581c61b930aafe8c09b431883431bcdda63ea200de7cab74f3/Djblets-0.8.29.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.3": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "ba2bee35a7a4459290e6be1b79df29ef20e338e2c046733f75afacace08f896f", + "md5": "8de5512f5ce12deb77edf8c9d218f673", + "sha256": "ec67edca44e0bc2f1e3465175bab18b4795cf5c375a15993d7ca3446978d1d23" + }, + "downloads": -1, + "filename": "Djblets-0.8.3-py2.6.egg", + "has_sig": false, + "md5_digest": "8de5512f5ce12deb77edf8c9d218f673", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 647673, + "upload_time": "2015-10-29T21:11:48", + "upload_time_iso_8601": "2015-10-29T21:11:48.368765Z", + "url": "https://files.pythonhosted.org/packages/ba/2b/ee35a7a4459290e6be1b79df29ef20e338e2c046733f75afacace08f896f/Djblets-0.8.3-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "d082e1d8d0c67c244e53a5e48cf8c27e41bbea40af4d7f6a139a06ce4c359513", + "md5": "89a98597b92738b2d364465fe50f24db", + "sha256": "624782dca4c004e58a31d8db26b3530b5d49d8cec833b9177858acf0cd5eb737" + }, + "downloads": -1, + "filename": "Djblets-0.8.3-py2.7.egg", + "has_sig": false, + "md5_digest": "89a98597b92738b2d364465fe50f24db", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 656740, + "upload_time": "2015-10-29T21:11:54", + "upload_time_iso_8601": "2015-10-29T21:11:54.285701Z", + "url": "https://files.pythonhosted.org/packages/d0/82/e1d8d0c67c244e53a5e48cf8c27e41bbea40af4d7f6a139a06ce4c359513/Djblets-0.8.3-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "78e125a361a0eac03f948d2081cf4fe49b5aa2470ead0618fee69c055e98a386", + "md5": "09a64de0581fb4bfe3b39a6b94d7f613", + "sha256": "c1cd50cf86d64fed64b81b021746026c6c8cf5df1dc947d0acfbfd96e84a482a" + }, + "downloads": -1, + "filename": "Djblets-0.8.3.tar.gz", + "has_sig": false, + "md5_digest": "09a64de0581fb4bfe3b39a6b94d7f613", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 267970, + "upload_time": "2015-10-29T21:12:00", + "upload_time_iso_8601": "2015-10-29T21:12:00.978974Z", + "url": "https://files.pythonhosted.org/packages/78/e1/25a361a0eac03f948d2081cf4fe49b5aa2470ead0618fee69c055e98a386/Djblets-0.8.3.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.4": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "2d12466d18dd11badd026d3caa361248c95612c2c34693e78c922d77d637948d", + "md5": "92b869ae50795342901c17b7a12483b2", + "sha256": "6fd31eafbdf03561dbbebd80077c60a79f5ca59c4e8951b3e8c355407ba49ba9" + }, + "downloads": -1, + "filename": "Djblets-0.8.4-py2.6.egg", + "has_sig": false, + "md5_digest": "92b869ae50795342901c17b7a12483b2", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 658990, + "upload_time": "2015-10-29T21:12:15", + "upload_time_iso_8601": "2015-10-29T21:12:15.618653Z", + "url": "https://files.pythonhosted.org/packages/2d/12/466d18dd11badd026d3caa361248c95612c2c34693e78c922d77d637948d/Djblets-0.8.4-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "e889c818849357033da8206e038d3e81c7c9e713ef9239dcc3aacd6e9b744387", + "md5": "1745e4896da90d97ef54ba120db8b9d4", + "sha256": "f19bb06291501cf7eccb2f839d15c675bca014d1c3ab89800bcba7375989950f" + }, + "downloads": -1, + "filename": "Djblets-0.8.4-py2.7.egg", + "has_sig": false, + "md5_digest": "1745e4896da90d97ef54ba120db8b9d4", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 668032, + "upload_time": "2015-10-29T21:12:28", + "upload_time_iso_8601": "2015-10-29T21:12:28.748535Z", + "url": "https://files.pythonhosted.org/packages/e8/89/c818849357033da8206e038d3e81c7c9e713ef9239dcc3aacd6e9b744387/Djblets-0.8.4-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "d718c6c0387de890415f1a3c1fa1181c133d0415aee51c8a1584595db706d781", + "md5": "91c12273ef64f33306bebc908eaed0f7", + "sha256": "7d0f38bdbfab45b4f1cd5a9a882a4ea083a3a1e464092554a17a3a2fcfee54ca" + }, + "downloads": -1, + "filename": "Djblets-0.8.4.tar.gz", + "has_sig": false, + "md5_digest": "91c12273ef64f33306bebc908eaed0f7", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 272462, + "upload_time": "2015-10-29T21:12:35", + "upload_time_iso_8601": "2015-10-29T21:12:35.979475Z", + "url": "https://files.pythonhosted.org/packages/d7/18/c6c0387de890415f1a3c1fa1181c133d0415aee51c8a1584595db706d781/Djblets-0.8.4.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.5": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "7598d27dd20d792949bc81bafad86517c0387c9cdf6236be711e99aaf4fc9414", + "md5": "f1b593e6cb537657ae4fac65e644e458", + "sha256": "3c785de9ef9ea10566d9a7065be6ddc78326bc237f87a242e25d68212febf6f6" + }, + "downloads": -1, + "filename": "Djblets-0.8.5-py2.6.egg", + "has_sig": false, + "md5_digest": "f1b593e6cb537657ae4fac65e644e458", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 658918, + "upload_time": "2015-10-29T21:12:42", + "upload_time_iso_8601": "2015-10-29T21:12:42.041375Z", + "url": "https://files.pythonhosted.org/packages/75/98/d27dd20d792949bc81bafad86517c0387c9cdf6236be711e99aaf4fc9414/Djblets-0.8.5-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "75b22a43b8aa60de9f65c7a7dec2f47a32698a69c4fbaaed3ad739f5c4749760", + "md5": "f87249d936fffb3d73628e1f4084bc61", + "sha256": "7306399ce3db0064d1d8c1163e81719667ee91c60b70f60968f593bb88e75998" + }, + "downloads": -1, + "filename": "Djblets-0.8.5-py2.7.egg", + "has_sig": false, + "md5_digest": "f87249d936fffb3d73628e1f4084bc61", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 667963, + "upload_time": "2015-10-29T21:12:47", + "upload_time_iso_8601": "2015-10-29T21:12:47.825441Z", + "url": "https://files.pythonhosted.org/packages/75/b2/2a43b8aa60de9f65c7a7dec2f47a32698a69c4fbaaed3ad739f5c4749760/Djblets-0.8.5-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "c164bd8ab5e9c438a84633810e7b374d5ae8c7e4b59ff0f713d9dac56b3d7eea", + "md5": "70a0f4ee69dfd42224eb396fffba069f", + "sha256": "ce6a569c30c39838b55642554aa628adbbb3983878885acd9a9c1f1bb628d725" + }, + "downloads": -1, + "filename": "Djblets-0.8.5.tar.gz", + "has_sig": false, + "md5_digest": "70a0f4ee69dfd42224eb396fffba069f", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 272524, + "upload_time": "2015-10-29T21:12:52", + "upload_time_iso_8601": "2015-10-29T21:12:52.525366Z", + "url": "https://files.pythonhosted.org/packages/c1/64/bd8ab5e9c438a84633810e7b374d5ae8c7e4b59ff0f713d9dac56b3d7eea/Djblets-0.8.5.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.6": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "b8bf86bc5b355d0025189500ba7c25c8e605fd7f8c759f40ec1296ecbb5577f1", + "md5": "18d088fb3cc4f20142d1316d6c6fc4e8", + "sha256": "cc70664869046b834efa367d5ba5aa8fe5e91b491dd43de006eece084c2944c6" + }, + "downloads": -1, + "filename": "Djblets-0.8.6-py2.6.egg", + "has_sig": false, + "md5_digest": "18d088fb3cc4f20142d1316d6c6fc4e8", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 672290, + "upload_time": "2015-10-29T21:13:01", + "upload_time_iso_8601": "2015-10-29T21:13:01.194683Z", + "url": "https://files.pythonhosted.org/packages/b8/bf/86bc5b355d0025189500ba7c25c8e605fd7f8c759f40ec1296ecbb5577f1/Djblets-0.8.6-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "59d32827d9a70960b3adc28675950b8af02e31c9198b4113cf521c9df538e8f9", + "md5": "1f3cece0ff45e1cf51bc4ffbba431fb3", + "sha256": "eb6b8a6c89f33dd8e49f93a9162a9767c4ea83c46809fe3ac0cc7e1a93e69f4a" + }, + "downloads": -1, + "filename": "Djblets-0.8.6-py2.7.egg", + "has_sig": false, + "md5_digest": "1f3cece0ff45e1cf51bc4ffbba431fb3", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 681404, + "upload_time": "2015-10-29T21:13:14", + "upload_time_iso_8601": "2015-10-29T21:13:14.789563Z", + "url": "https://files.pythonhosted.org/packages/59/d3/2827d9a70960b3adc28675950b8af02e31c9198b4113cf521c9df538e8f9/Djblets-0.8.6-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "86ea48e1728bdaecbc015d80715ee9a8f3e0a9e3b4aca232fe98a7ec569dca0e", + "md5": "add5ebaa54b89a1b9f2385e1808cb9f2", + "sha256": "e5295ba9b2acf3c198595c4b0927ef63822e20f409ed239d552fbfe73192303a" + }, + "downloads": -1, + "filename": "Djblets-0.8.6.tar.gz", + "has_sig": false, + "md5_digest": "add5ebaa54b89a1b9f2385e1808cb9f2", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 246311, + "upload_time": "2015-10-29T21:13:20", + "upload_time_iso_8601": "2015-10-29T21:13:20.969507Z", + "url": "https://files.pythonhosted.org/packages/86/ea/48e1728bdaecbc015d80715ee9a8f3e0a9e3b4aca232fe98a7ec569dca0e/Djblets-0.8.6.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.6.dev": [], + "0.8.7": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "c1acabd05b6e936bfd4aa4d461e91016daf8330eb2e84481fcd3b451fa9be3a5", + "md5": "e105628e5ae1cc8306dfe6a1ef83c8e4", + "sha256": "124f20f744f797a89085b18d448ae3cc216961fe242e85be250f145ec9a9a9a9" + }, + "downloads": -1, + "filename": "Djblets-0.8.7-py2.6.egg", + "has_sig": false, + "md5_digest": "e105628e5ae1cc8306dfe6a1ef83c8e4", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 675869, + "upload_time": "2015-10-29T21:13:28", + "upload_time_iso_8601": "2015-10-29T21:13:28.162979Z", + "url": "https://files.pythonhosted.org/packages/c1/ac/abd05b6e936bfd4aa4d461e91016daf8330eb2e84481fcd3b451fa9be3a5/Djblets-0.8.7-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "82a04a26c5bdd9c8a604d51d520f413d7bccc5d353e7621e3b2acd14cd11a274", + "md5": "c3c64dd4024dac66bdf190e92c915d52", + "sha256": "2e7a47a6a663ad415bc1ba43cb86b47e073e13f0cbca9359176019f2483c8a9d" + }, + "downloads": -1, + "filename": "Djblets-0.8.7-py2.7.egg", + "has_sig": false, + "md5_digest": "c3c64dd4024dac66bdf190e92c915d52", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 685048, + "upload_time": "2015-10-29T21:13:34", + "upload_time_iso_8601": "2015-10-29T21:13:34.461617Z", + "url": "https://files.pythonhosted.org/packages/82/a0/4a26c5bdd9c8a604d51d520f413d7bccc5d353e7621e3b2acd14cd11a274/Djblets-0.8.7-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "c4084cb6828bd3947cd21686062f56125abd94a6390307e524c96933bfa9aef6", + "md5": "b13e41288915734011d7f2fb09528e4d", + "sha256": "72449642100ef81e319fc961f2c64ac117a40d728803fa52753d107c339acf32" + }, + "downloads": -1, + "filename": "Djblets-0.8.7.tar.gz", + "has_sig": false, + "md5_digest": "b13e41288915734011d7f2fb09528e4d", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 248266, + "upload_time": "2015-10-29T21:13:40", + "upload_time_iso_8601": "2015-10-29T21:13:40.939660Z", + "url": "https://files.pythonhosted.org/packages/c4/08/4cb6828bd3947cd21686062f56125abd94a6390307e524c96933bfa9aef6/Djblets-0.8.7.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.8": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "514e33a7df235730b2edbcd1176ee89fed23ccd36dd3039ba55641f1f518b790", + "md5": "6b366993b40aa9bf2d3855f7978012be", + "sha256": "5431f2b4dab2246037c86e7426317775fdfd5b15b3e8361464370e37dd1599b1" + }, + "downloads": -1, + "filename": "Djblets-0.8.8-py2.6.egg", + "has_sig": false, + "md5_digest": "6b366993b40aa9bf2d3855f7978012be", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 675703, + "upload_time": "2015-10-29T21:13:46", + "upload_time_iso_8601": "2015-10-29T21:13:46.573878Z", + "url": "https://files.pythonhosted.org/packages/51/4e/33a7df235730b2edbcd1176ee89fed23ccd36dd3039ba55641f1f518b790/Djblets-0.8.8-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "25d5882da1d19e2c70bf5cbdf00e643ccf4416b4352b72d578d2f1378f1927a1", + "md5": "dd8ef91838f1b7b6af5dbc4a5dc40728", + "sha256": "6a3695310563c3ae76b49e5a6ac7e3ba2e4b66cf7d13db3ea4432d59935a317f" + }, + "downloads": -1, + "filename": "Djblets-0.8.8-py2.7.egg", + "has_sig": false, + "md5_digest": "dd8ef91838f1b7b6af5dbc4a5dc40728", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 684940, + "upload_time": "2015-10-29T21:13:52", + "upload_time_iso_8601": "2015-10-29T21:13:52.907120Z", + "url": "https://files.pythonhosted.org/packages/25/d5/882da1d19e2c70bf5cbdf00e643ccf4416b4352b72d578d2f1378f1927a1/Djblets-0.8.8-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "be66a7d44ebe84daeaaad1cf0925161cac217440af7193980dcbbcdbebce708c", + "md5": "612a0c5980a5fd5de173345cfa839c5c", + "sha256": "6a9303f962ed5e2d2b1aa1ac6b1e5bab22d27a039b8ac72eae93948eac1ce1bd" + }, + "downloads": -1, + "filename": "Djblets-0.8.8.tar.gz", + "has_sig": false, + "md5_digest": "612a0c5980a5fd5de173345cfa839c5c", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 248306, + "upload_time": "2015-10-29T21:14:00", + "upload_time_iso_8601": "2015-10-29T21:14:00.169901Z", + "url": "https://files.pythonhosted.org/packages/be/66/a7d44ebe84daeaaad1cf0925161cac217440af7193980dcbbcdbebce708c/Djblets-0.8.8.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.8.9": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "5a8ccb593a25ff2c972bd2739bd849076fac4e63be477545ef10eedb2c585016", + "md5": "ad90aea1e1da07741661e8571941c7fa", + "sha256": "0d13fc279ae9ce52311e199471955913b7f612616ce5917c2832122efb09b40d" + }, + "downloads": -1, + "filename": "Djblets-0.8.9-py2.6.egg", + "has_sig": false, + "md5_digest": "ad90aea1e1da07741661e8571941c7fa", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 696170, + "upload_time": "2015-10-29T21:14:18", + "upload_time_iso_8601": "2015-10-29T21:14:18.533387Z", + "url": "https://files.pythonhosted.org/packages/5a/8c/cb593a25ff2c972bd2739bd849076fac4e63be477545ef10eedb2c585016/Djblets-0.8.9-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "11dc1ed88a87641f201cba131a91dcbecfc224276bce26afed5bc6c4e8529cd7", + "md5": "9def900fe3afc325de411a45e2293ae3", + "sha256": "31f9d908e3727e782945c54aa41bf19caa05b47e643251a074ac0e93dc63f6f0" + }, + "downloads": -1, + "filename": "Djblets-0.8.9-py2.7.egg", + "has_sig": false, + "md5_digest": "9def900fe3afc325de411a45e2293ae3", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 705216, + "upload_time": "2015-10-29T21:14:35", + "upload_time_iso_8601": "2015-10-29T21:14:35.805701Z", + "url": "https://files.pythonhosted.org/packages/11/dc/1ed88a87641f201cba131a91dcbecfc224276bce26afed5bc6c4e8529cd7/Djblets-0.8.9-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "100361bcba8b5e8cd4c8feec8b25757df99ad503504db5d3a5d281bf3ed30d04", + "md5": "253f77717a7466037dd0a0f2ac997fb9", + "sha256": "cf37856260badee000bb05f3b249af3f9472e3df1499e9f43069f9cc3c95c592" + }, + "downloads": -1, + "filename": "Djblets-0.8.9.tar.gz", + "has_sig": false, + "md5_digest": "253f77717a7466037dd0a0f2ac997fb9", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 258496, + "upload_time": "2015-10-29T21:14:42", + "upload_time_iso_8601": "2015-10-29T21:14:42.996197Z", + "url": "https://files.pythonhosted.org/packages/10/03/61bcba8b5e8cd4c8feec8b25757df99ad503504db5d3a5d281bf3ed30d04/Djblets-0.8.9.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.9": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "a554c175fc75d42e0a4492c45523f4e445560b02691b5246c71fd3c9084f6ed0", + "md5": "f52f68dca31c6de7afe34c98a6795389", + "sha256": "e2ec58e305d2b61d210116cc64e6ebcf09aafd3cd886fa05526614cbc94eb5ba" + }, + "downloads": -1, + "filename": "Djblets-0.9-py2.6.egg", + "has_sig": false, + "md5_digest": "f52f68dca31c6de7afe34c98a6795389", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 900559, + "upload_time": "2015-10-29T21:03:44", + "upload_time_iso_8601": "2015-10-29T21:03:44.747820Z", + "url": "https://files.pythonhosted.org/packages/a5/54/c175fc75d42e0a4492c45523f4e445560b02691b5246c71fd3c9084f6ed0/Djblets-0.9-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "b2d9b27ec150ee94a26ae47cd8d5af449c984984cedd372c5c8053e93c7de218", + "md5": "279600f4cc09a3bb768443f9afd024d1", + "sha256": "78d09d34209ab5a6efdace43621cfbc3f05fa172dc69efee1b1c0dd4d0c707b1" + }, + "downloads": -1, + "filename": "Djblets-0.9-py2.7.egg", + "has_sig": false, + "md5_digest": "279600f4cc09a3bb768443f9afd024d1", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 909271, + "upload_time": "2015-10-29T21:03:53", + "upload_time_iso_8601": "2015-10-29T21:03:53.540904Z", + "url": "https://files.pythonhosted.org/packages/b2/d9/b27ec150ee94a26ae47cd8d5af449c984984cedd372c5c8053e93c7de218/Djblets-0.9-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "169fec48ff5f43f8aef34fc43047488a4e32f886a597a111306194f3638637c1", + "md5": "a70674996ff402661c9e408af6790c5e", + "sha256": "981d5a1fae8ea9545e16af7a1080de986af9948c6375ff80fdecfc9ba26ce163" + }, + "downloads": -1, + "filename": "Djblets-0.9-py2-none-any.whl", + "has_sig": false, + "md5_digest": "a70674996ff402661c9e408af6790c5e", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 581275, + "upload_time": "2015-10-29T21:02:34", + "upload_time_iso_8601": "2015-10-29T21:02:34.342399Z", + "url": "https://files.pythonhosted.org/packages/16/9f/ec48ff5f43f8aef34fc43047488a4e32f886a597a111306194f3638637c1/Djblets-0.9-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "b546f520e9673a95575764c2c32187886578a09d9c15e675f1774aa60f849aa7", + "md5": "bd13058da818b1649ec78032a6d9fcf3", + "sha256": "f4b6bb2e3597efa67e7cbdccd5d0403480e3a88e5cd934098c73f018b9dc25e7" + }, + "downloads": -1, + "filename": "Djblets-0.9.tar.gz", + "has_sig": false, + "md5_digest": "bd13058da818b1649ec78032a6d9fcf3", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 330999, + "upload_time": "2015-10-29T21:04:02", + "upload_time_iso_8601": "2015-10-29T21:04:02.806960Z", + "url": "https://files.pythonhosted.org/packages/b5/46/f520e9673a95575764c2c32187886578a09d9c15e675f1774aa60f849aa7/Djblets-0.9.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.9.1": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "5dc0189d937cba1f9f14015f283b2299ca195b699f54a4a6d41d88869fee65fa", + "md5": "00fd2c9694ee9d4f7c2491ffc3d5067b", + "sha256": "ceedae6779f80a584c0a44dac9b5acf3fd4d364187edecec049e9f3ab1fc8138" + }, + "downloads": -1, + "filename": "Djblets-0.9.1-py2.6.egg", + "has_sig": false, + "md5_digest": "00fd2c9694ee9d4f7c2491ffc3d5067b", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 904487, + "upload_time": "2016-02-22T06:48:38", + "upload_time_iso_8601": "2016-02-22T06:48:38.542586Z", + "url": "https://files.pythonhosted.org/packages/5d/c0/189d937cba1f9f14015f283b2299ca195b699f54a4a6d41d88869fee65fa/Djblets-0.9.1-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "a92ff80c42bfb9afe85fd28e2053fba30a25a01221e71c45664465633cb4cf62", + "md5": "37e128107c0ddf3ab62ae62bc20ea7cd", + "sha256": "7bb41c230f15cd1bd4c92d6fbeed54a589877830be81112b8b38ccf98f3bf982" + }, + "downloads": -1, + "filename": "Djblets-0.9.1-py2.7.egg", + "has_sig": false, + "md5_digest": "37e128107c0ddf3ab62ae62bc20ea7cd", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 913282, + "upload_time": "2016-02-22T06:48:58", + "upload_time_iso_8601": "2016-02-22T06:48:58.135680Z", + "url": "https://files.pythonhosted.org/packages/a9/2f/f80c42bfb9afe85fd28e2053fba30a25a01221e71c45664465633cb4cf62/Djblets-0.9.1-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "6a894123c328b076b26824ead1554f32bc86d76f40f9d28571858cfe20fccf84", + "md5": "a5c572a3c4b17497d25aca1b8b0d8d6e", + "sha256": "eb8ed3f0cabdb8ba359df237b1e4ae507beb49e443a1fbb29d67a15500316824" + }, + "downloads": -1, + "filename": "Djblets-0.9.1-py2-none-any.whl", + "has_sig": false, + "md5_digest": "a5c572a3c4b17497d25aca1b8b0d8d6e", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 582917, + "upload_time": "2016-02-22T06:48:21", + "upload_time_iso_8601": "2016-02-22T06:48:21.184557Z", + "url": "https://files.pythonhosted.org/packages/6a/89/4123c328b076b26824ead1554f32bc86d76f40f9d28571858cfe20fccf84/Djblets-0.9.1-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "16e56567fae7e0583a9e9f3e9d2fda1ef67b119a9bcf322abf2854ab81f9228c", + "md5": "a4d275f5e8ac2ae017350a53a0eedc2e", + "sha256": "f0801b3b9b48b493ed70a389e917747fcca9e827a2a31ff7c7213ec72ad66b5d" + }, + "downloads": -1, + "filename": "Djblets-0.9.1.tar.gz", + "has_sig": false, + "md5_digest": "a4d275f5e8ac2ae017350a53a0eedc2e", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 332720, + "upload_time": "2016-02-22T06:49:11", + "upload_time_iso_8601": "2016-02-22T06:49:11.661244Z", + "url": "https://files.pythonhosted.org/packages/16/e5/6567fae7e0583a9e9f3e9d2fda1ef67b119a9bcf322abf2854ab81f9228c/Djblets-0.9.1.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.9.2": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "c6f7ab46f6ad196e82640fd0ca8e5736855bffe373b9d321de47049908420ab7", + "md5": "09f202e4bff3c0ca3036ca295033db48", + "sha256": "c37bcd05c208bb88cb3de0d56b75e1407374b80e1ba15b03b149d5ad603b1327" + }, + "downloads": -1, + "filename": "Djblets-0.9.2-py2.6.egg", + "has_sig": false, + "md5_digest": "09f202e4bff3c0ca3036ca295033db48", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 904595, + "upload_time": "2016-03-01T19:04:45", + "upload_time_iso_8601": "2016-03-01T19:04:45.593584Z", + "url": "https://files.pythonhosted.org/packages/c6/f7/ab46f6ad196e82640fd0ca8e5736855bffe373b9d321de47049908420ab7/Djblets-0.9.2-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "5d3cce1c5075ace92acc284e9dea35edcab64adfa8322e4b6ad5f645e5389a79", + "md5": "2756227f77f4cf9807686ea92ca6b0da", + "sha256": "38faf155789f4018e48920bba22d7ee83fd607de7655d98a0e0cfa72bf1ba0cf" + }, + "downloads": -1, + "filename": "Djblets-0.9.2-py2.7.egg", + "has_sig": false, + "md5_digest": "2756227f77f4cf9807686ea92ca6b0da", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 913481, + "upload_time": "2016-03-01T19:04:56", + "upload_time_iso_8601": "2016-03-01T19:04:56.472651Z", + "url": "https://files.pythonhosted.org/packages/5d/3c/ce1c5075ace92acc284e9dea35edcab64adfa8322e4b6ad5f645e5389a79/Djblets-0.9.2-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "70b392a0e3e0bfd940916a593277c2a28207f518d53ed8a171d3daaf69522da2", + "md5": "4b9b070a100ca9750f7d363a7e2f4ab3", + "sha256": "35282652ab3a68709a118cf1228c3c8d1022dc7fa9d74d4a1251bf434178e9d3" + }, + "downloads": -1, + "filename": "Djblets-0.9.2-py2-none-any.whl", + "has_sig": false, + "md5_digest": "4b9b070a100ca9750f7d363a7e2f4ab3", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 582934, + "upload_time": "2016-03-01T19:04:36", + "upload_time_iso_8601": "2016-03-01T19:04:36.829796Z", + "url": "https://files.pythonhosted.org/packages/70/b3/92a0e3e0bfd940916a593277c2a28207f518d53ed8a171d3daaf69522da2/Djblets-0.9.2-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "f3ef7b64d32c95756b8247b7786bc8fa753d27c620f69f14c0ec2607465239a3", + "md5": "00356c7a9c1855ab9999e511d90a61b8", + "sha256": "9df3db467ccc427d85f8a2f929557a884f9149fd32a96765c8854b1463a193f6" + }, + "downloads": -1, + "filename": "Djblets-0.9.2.tar.gz", + "has_sig": false, + "md5_digest": "00356c7a9c1855ab9999e511d90a61b8", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 332675, + "upload_time": "2016-03-01T19:05:05", + "upload_time_iso_8601": "2016-03-01T19:05:05.497377Z", + "url": "https://files.pythonhosted.org/packages/f3/ef/7b64d32c95756b8247b7786bc8fa753d27c620f69f14c0ec2607465239a3/Djblets-0.9.2.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.9.3": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "6c1af13c28cfff77999358ba3d2ee93d1df6610eab36a8e43f91dd3179b5de68", + "md5": "1e61497a7d3d100cc7fd5304167f5953", + "sha256": "e19189f65b16e215754c023bc5ffb9954b1aeda6e24cc27520b0a3ce9b2454c3" + }, + "downloads": -1, + "filename": "Djblets-0.9.3-py2.6.egg", + "has_sig": false, + "md5_digest": "1e61497a7d3d100cc7fd5304167f5953", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 906489, + "upload_time": "2016-04-13T21:14:32", + "upload_time_iso_8601": "2016-04-13T21:14:32.954046Z", + "url": "https://files.pythonhosted.org/packages/6c/1a/f13c28cfff77999358ba3d2ee93d1df6610eab36a8e43f91dd3179b5de68/Djblets-0.9.3-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "a469dad63181f464f53f25ac3e02635b6c0ba98ee2b9697fcbed91548acd72fa", + "md5": "f68cfc0d87385e2da487da195bcdf864", + "sha256": "2810dd9c1535bfb2c9974aa4f80f377709b6647004927cd62bbf39316e19e79c" + }, + "downloads": -1, + "filename": "Djblets-0.9.3-py2.7.egg", + "has_sig": false, + "md5_digest": "f68cfc0d87385e2da487da195bcdf864", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 915227, + "upload_time": "2016-04-13T21:14:40", + "upload_time_iso_8601": "2016-04-13T21:14:40.496754Z", + "url": "https://files.pythonhosted.org/packages/a4/69/dad63181f464f53f25ac3e02635b6c0ba98ee2b9697fcbed91548acd72fa/Djblets-0.9.3-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "55d850959bf1e7816e5d7b1a7669018480bc3f0349e56b15d1efdd13e6072ec9", + "md5": "40f511a697aff98702443320812ee27e", + "sha256": "be1f5ca56681f6feee6ec02cab9d74ffb9e1875f64e4125659b690a8f40d4e16" + }, + "downloads": -1, + "filename": "Djblets-0.9.3-py2-none-any.whl", + "has_sig": false, + "md5_digest": "40f511a697aff98702443320812ee27e", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 584292, + "upload_time": "2016-04-13T21:14:22", + "upload_time_iso_8601": "2016-04-13T21:14:22.559647Z", + "url": "https://files.pythonhosted.org/packages/55/d8/50959bf1e7816e5d7b1a7669018480bc3f0349e56b15d1efdd13e6072ec9/Djblets-0.9.3-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, { - "comment_text": "", - "digests": { - "blake2b_256": "bac761b02c0ef9e129080a8c2bffefb3cb2b9ddddece4c44dc473c1c4f0647c1", - "md5": "3fa7115ed069c54684302278b48b7fd6", - "sha256": "3a9fd52b8dbeae335ddf4a9dfa6c6a0853a1122f1fb071a8d5eca979f73a05c8" - }, - "downloads": -1, - "filename": "Django-5.0-py3-none-any.whl", - "has_sig": false, - "md5_digest": "3fa7115ed069c54684302278b48b7fd6", - "packagetype": "bdist_wheel", - "python_version": "py3", - "requires_python": ">=3.10", - "size": 8136382, - "upload_time": "2023-12-04T13:12:41", - "upload_time_iso_8601": "2023-12-04T13:12:41.502020Z", - "url": "https://files.pythonhosted.org/packages/ba/c7/61b02c0ef9e129080a8c2bffefb3cb2b9ddddece4c44dc473c1c4f0647c1/Django-5.0-py3-none-any.whl", - "yanked": false, - "yanked_reason": null - }, - { - "comment_text": "", - "digests": { - "blake2b_256": "bea646e250737d46e955e048f6bbc2948fb22f0de3f3ab828d3803070dc1260e", - "md5": "5de06916549d6f81fe6eb102722a1b28", - "sha256": "7d29e14dfbc19cb6a95a4bd669edbde11f5d4c6a71fdaa42c2d40b6846e807f7" - }, - "downloads": -1, - "filename": "Django-5.0.tar.gz", - "has_sig": false, - "md5_digest": "5de06916549d6f81fe6eb102722a1b28", - "packagetype": "sdist", - "python_version": "source", - "requires_python": ">=3.10", - "size": 10585390, - "upload_time": "2023-12-04T13:12:50", - "upload_time_iso_8601": "2023-12-04T13:12:50.251293Z", - "url": "https://files.pythonhosted.org/packages/be/a6/46e250737d46e955e048f6bbc2948fb22f0de3f3ab828d3803070dc1260e/Django-5.0.tar.gz", - "yanked": false, - "yanked_reason": null + "comment_text": "", + "digests": { + "blake2b_256": "9c7fa8d958eb894916b50736996723f333fdf319d0db50226ec96efb9cea3662", + "md5": "97face4b271c4a7047da2512d414ceb4", + "sha256": "327440d2dbd97930438b30d349423f5a33aeb85d3a50142ef379d3a7552c1549" + }, + "downloads": -1, + "filename": "Djblets-0.9.3.tar.gz", + "has_sig": false, + "md5_digest": "97face4b271c4a7047da2512d414ceb4", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 333765, + "upload_time": "2016-04-13T21:14:46", + "upload_time_iso_8601": "2016-04-13T21:14:46.224818Z", + "url": "https://files.pythonhosted.org/packages/9c/7f/a8d958eb894916b50736996723f333fdf319d0db50226ec96efb9cea3662/Djblets-0.9.3.tar.gz", + "yanked": false, + "yanked_reason": null } + ], + "0.9.4": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "e6ddfaec69076e0dca47b618e2a25bc405e8bea7655682455310ef06b66bdfe2", + "md5": "0fe682a356bd9219370b907df6ae3160", + "sha256": "0c5e3a54939e9f6ad926d8a51b90859ca4f806c0df99ed1ba23925adacd34178" + }, + "downloads": -1, + "filename": "Djblets-0.9.4-py2.6.egg", + "has_sig": false, + "md5_digest": "0fe682a356bd9219370b907df6ae3160", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 1011689, + "upload_time": "2016-10-18T02:16:56", + "upload_time_iso_8601": "2016-10-18T02:16:56.098675Z", + "url": "https://files.pythonhosted.org/packages/e6/dd/faec69076e0dca47b618e2a25bc405e8bea7655682455310ef06b66bdfe2/Djblets-0.9.4-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "bc4d04aee84c5d3c777bb58f28ee71138cba1af513f6887c8a1d4c916f199092", + "md5": "612fdd72fed9c56630a24e5b751312d2", + "sha256": "933cc5ab943c03742aae78992aa5a3ec904afe36395a63abb7c54f1b59f101cf" + }, + "downloads": -1, + "filename": "Djblets-0.9.4-py2.7.egg", + "has_sig": false, + "md5_digest": "612fdd72fed9c56630a24e5b751312d2", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 1020390, + "upload_time": "2016-10-18T02:17:00", + "upload_time_iso_8601": "2016-10-18T02:17:00.424639Z", + "url": "https://files.pythonhosted.org/packages/bc/4d/04aee84c5d3c777bb58f28ee71138cba1af513f6887c8a1d4c916f199092/Djblets-0.9.4-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "e83a1931b302a28f984477f4a5f58ea05c53a53b6371fff3198c0b63f9515863", + "md5": "37dc0a8ff1ffcfe7a00523da08ba64c4", + "sha256": "ffbad91b7c5f29f41db9b486b5fc98b93f7036a0ce8795de82a35d58a44ff03b" + }, + "downloads": -1, + "filename": "Djblets-0.9.4-py2-none-any.whl", + "has_sig": false, + "md5_digest": "37dc0a8ff1ffcfe7a00523da08ba64c4", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 666488, + "upload_time": "2016-10-18T02:16:51", + "upload_time_iso_8601": "2016-10-18T02:16:51.813505Z", + "url": "https://files.pythonhosted.org/packages/e8/3a/1931b302a28f984477f4a5f58ea05c53a53b6371fff3198c0b63f9515863/Djblets-0.9.4-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "1126c5e4285d487ac1c07d71785403ce035f8e87b64328178a327fb3a47be24b", + "md5": "ee240791b2fae83bb71e6b4702b06196", + "sha256": "f2b34295fe1dbe224ae973ec1bb6ca423637112cdd4b6063767e06b90341f264" + }, + "downloads": -1, + "filename": "Djblets-0.9.4.tar.gz", + "has_sig": false, + "md5_digest": "ee240791b2fae83bb71e6b4702b06196", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 412601, + "upload_time": "2016-10-18T02:17:04", + "upload_time_iso_8601": "2016-10-18T02:17:04.972690Z", + "url": "https://files.pythonhosted.org/packages/11/26/c5e4285d487ac1c07d71785403ce035f8e87b64328178a327fb3a47be24b/Djblets-0.9.4.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.9.5": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "e65cad14b223026e53ba67bb9b87f7825128d14d4239efd89dcc21e7639776f3", + "md5": "2c930af4044ca1dae282ca2eb7e442b1", + "sha256": "0d6037d80c75bb1bbda4a31462f447756a07c892413f1e12ca9c764f7859b038" + }, + "downloads": -1, + "filename": "Djblets-0.9.5-py2.6.egg", + "has_sig": false, + "md5_digest": "2c930af4044ca1dae282ca2eb7e442b1", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 1037769, + "upload_time": "2017-02-09T23:14:17", + "upload_time_iso_8601": "2017-02-09T23:14:17.801601Z", + "url": "https://files.pythonhosted.org/packages/e6/5c/ad14b223026e53ba67bb9b87f7825128d14d4239efd89dcc21e7639776f3/Djblets-0.9.5-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "fe6284810d15e6262f3353c2a4742faa8358395b4e5c08b0c263db02f1a2a8b4", + "md5": "09b863ed9cb714c261f2282cfc36917e", + "sha256": "f4d8ed8c0b921462422137bafee505778e2e2a0373d68741845c1b1124c0cde3" + }, + "downloads": -1, + "filename": "Djblets-0.9.5-py2.7.egg", + "has_sig": false, + "md5_digest": "09b863ed9cb714c261f2282cfc36917e", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 1046282, + "upload_time": "2017-02-09T23:14:21", + "upload_time_iso_8601": "2017-02-09T23:14:21.779351Z", + "url": "https://files.pythonhosted.org/packages/fe/62/84810d15e6262f3353c2a4742faa8358395b4e5c08b0c263db02f1a2a8b4/Djblets-0.9.5-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "fbae37dd541c8338babe3d3efe2d70b3cd320da5ace65b5c1544568264000fd4", + "md5": "dc6c537eeca83ee8d9b6e2faf3e9bab6", + "sha256": "466558751ccd11a6c8721f6b27e5aeba501e27dbf551e4a01972cce6c1c3bccf" + }, + "downloads": -1, + "filename": "Djblets-0.9.5-py2-none-any.whl", + "has_sig": false, + "md5_digest": "dc6c537eeca83ee8d9b6e2faf3e9bab6", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 677553, + "upload_time": "2017-02-09T23:14:14", + "upload_time_iso_8601": "2017-02-09T23:14:14.451774Z", + "url": "https://files.pythonhosted.org/packages/fb/ae/37dd541c8338babe3d3efe2d70b3cd320da5ace65b5c1544568264000fd4/Djblets-0.9.5-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "1e451915d50663de263a91638572513b201c18105b143abd6be3d5a432ea11e2", + "md5": "e2ee081d0c440083ebdb1c16413dfff3", + "sha256": "7e58e225aad32aefa4b2ee87b7746fee4d8f58f9efbe3c65d3327f5381bb259e" + }, + "downloads": -1, + "filename": "Djblets-0.9.5.tar.gz", + "has_sig": false, + "md5_digest": "e2ee081d0c440083ebdb1c16413dfff3", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 419184, + "upload_time": "2017-02-09T23:14:25", + "upload_time_iso_8601": "2017-02-09T23:14:25.588925Z", + "url": "https://files.pythonhosted.org/packages/1e/45/1915d50663de263a91638572513b201c18105b143abd6be3d5a432ea11e2/Djblets-0.9.5.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.9.6": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "e57248f919f4692d02fabd9aa7fc6245fcd5a3db281003ac2328293055c23f76", + "md5": "8ec4815c037cfda8c7941f2069a217b0", + "sha256": "b73219a87179454781a072ad3efcf987d65919e4588581667282f993234a84f0" + }, + "downloads": -1, + "filename": "Djblets-0.9.6-py2.6.egg", + "has_sig": false, + "md5_digest": "8ec4815c037cfda8c7941f2069a217b0", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 1038029, + "upload_time": "2017-02-16T02:07:23", + "upload_time_iso_8601": "2017-02-16T02:07:23.808178Z", + "url": "https://files.pythonhosted.org/packages/e5/72/48f919f4692d02fabd9aa7fc6245fcd5a3db281003ac2328293055c23f76/Djblets-0.9.6-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "ec9743c7f3e1648ed523152ac32713800ace431d99df51fe5d214e84c1eb76f6", + "md5": "458a9e431605bc6fec1ead9024945ae1", + "sha256": "44da60e9b0801db0e2f5f9e89ac8eeead2096e1e8d8cbb27aa8ba995760249fa" + }, + "downloads": -1, + "filename": "Djblets-0.9.6-py2.7.egg", + "has_sig": false, + "md5_digest": "458a9e431605bc6fec1ead9024945ae1", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 1046535, + "upload_time": "2017-02-16T02:07:28", + "upload_time_iso_8601": "2017-02-16T02:07:28.335154Z", + "url": "https://files.pythonhosted.org/packages/ec/97/43c7f3e1648ed523152ac32713800ace431d99df51fe5d214e84c1eb76f6/Djblets-0.9.6-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "74207d5a293ce3a4a97ba575a149625c4cb61f44d60fcad9ab6f2f78985db0a8", + "md5": "ea582dbb907de0e1580376e65a84bf70", + "sha256": "9b825540d1e33e549783e4411d75080fc6e4621a6a722f59d3177fb664d32971" + }, + "downloads": -1, + "filename": "Djblets-0.9.6-py2-none-any.whl", + "has_sig": false, + "md5_digest": "ea582dbb907de0e1580376e65a84bf70", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 677883, + "upload_time": "2017-02-16T02:07:20", + "upload_time_iso_8601": "2017-02-16T02:07:20.410162Z", + "url": "https://files.pythonhosted.org/packages/74/20/7d5a293ce3a4a97ba575a149625c4cb61f44d60fcad9ab6f2f78985db0a8/Djblets-0.9.6-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "579da335600896e8726dfb51d81c8e2f89f51f1af5e70fc5b86391e2fb6b492e", + "md5": "03cb467a4aaa12b2b7b937343e7f00e5", + "sha256": "4ec554d4f494f10f01299fee47d55dea598a085280cde589123627e258d59a12" + }, + "downloads": -1, + "filename": "Djblets-0.9.6.tar.gz", + "has_sig": false, + "md5_digest": "03cb467a4aaa12b2b7b937343e7f00e5", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 418551, + "upload_time": "2017-02-16T02:07:31", + "upload_time_iso_8601": "2017-02-16T02:07:31.869735Z", + "url": "https://files.pythonhosted.org/packages/57/9d/a335600896e8726dfb51d81c8e2f89f51f1af5e70fc5b86391e2fb6b492e/Djblets-0.9.6.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.9.7": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "f93a533d8d7cfdcc13eb7e0d7a92f420d6a91f6ac0255d0f412cbe08083f6547", + "md5": "e0dff346bf64620e78a30c05b35ce9a0", + "sha256": "c941550a9f55eab9830c850327eb8295ba35be0c97364a51d441a70d666c59a6" + }, + "downloads": -1, + "filename": "Djblets-0.9.7-py2.6.egg", + "has_sig": false, + "md5_digest": "e0dff346bf64620e78a30c05b35ce9a0", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 1045268, + "upload_time": "2017-05-26T04:49:22", + "upload_time_iso_8601": "2017-05-26T04:49:22.434273Z", + "url": "https://files.pythonhosted.org/packages/f9/3a/533d8d7cfdcc13eb7e0d7a92f420d6a91f6ac0255d0f412cbe08083f6547/Djblets-0.9.7-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "701a80d57655cd6885e9b834b3449202cd265789443a19f008b04ba523f8a96f", + "md5": "1711709dfe62ca1be1136f140f103408", + "sha256": "3975877e130f78b853de77f1f3260c7709de0138b7998110ae8d8b84cb692258" + }, + "downloads": -1, + "filename": "Djblets-0.9.7-py2.7.egg", + "has_sig": false, + "md5_digest": "1711709dfe62ca1be1136f140f103408", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 1053779, + "upload_time": "2017-05-26T04:49:26", + "upload_time_iso_8601": "2017-05-26T04:49:26.835068Z", + "url": "https://files.pythonhosted.org/packages/70/1a/80d57655cd6885e9b834b3449202cd265789443a19f008b04ba523f8a96f/Djblets-0.9.7-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "d7c63c447864cb3b3f882ec0ee007eb56811dd457e8271b8aa0c190d8f87d68d", + "md5": "b41de274a2117e4c1b7af959ef06b226", + "sha256": "3083219e694d970508d244d98350891e554d0ddf46b299c4d1a8a61fd6339ccc" + }, + "downloads": -1, + "filename": "Djblets-0.9.7-py2-none-any.whl", + "has_sig": false, + "md5_digest": "b41de274a2117e4c1b7af959ef06b226", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 681329, + "upload_time": "2017-05-26T04:49:17", + "upload_time_iso_8601": "2017-05-26T04:49:17.499829Z", + "url": "https://files.pythonhosted.org/packages/d7/c6/3c447864cb3b3f882ec0ee007eb56811dd457e8271b8aa0c190d8f87d68d/Djblets-0.9.7-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "5bd7f8dad1add4be55182f19114e6439b074afb11addc15af78d4c07872cde12", + "md5": "561d23f0b3b8d70c55bcdc8e2a5b602a", + "sha256": "439234c8bbdd2b5a55bb257fb2db4d4c08d5a761e4a6f6ba5e45c03e692b59a3" + }, + "downloads": -1, + "filename": "Djblets-0.9.7.tar.gz", + "has_sig": false, + "md5_digest": "561d23f0b3b8d70c55bcdc8e2a5b602a", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 421204, + "upload_time": "2017-05-26T04:49:30", + "upload_time_iso_8601": "2017-05-26T04:49:30.085130Z", + "url": "https://files.pythonhosted.org/packages/5b/d7/f8dad1add4be55182f19114e6439b074afb11addc15af78d4c07872cde12/Djblets-0.9.7.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.9.8": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "d45a78f89f3a655f351c6a0d929b8e1eb7144513fc3eaf1c23d8374f7faadb8f", + "md5": "11af708c1163f2bee33dee2b4fd9aa71", + "sha256": "35638fa0b3d32862cae0d52280bef86aff9f47871310cf60c0b3534c2679b6f2" + }, + "downloads": -1, + "filename": "Djblets-0.9.8-py2.6.egg", + "has_sig": false, + "md5_digest": "11af708c1163f2bee33dee2b4fd9aa71", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 1049258, + "upload_time": "2017-06-19T23:10:58", + "upload_time_iso_8601": "2017-06-19T23:10:58.435516Z", + "url": "https://files.pythonhosted.org/packages/d4/5a/78f89f3a655f351c6a0d929b8e1eb7144513fc3eaf1c23d8374f7faadb8f/Djblets-0.9.8-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "f9cdbf376970e1befcc8314e99e68a203cd172a00b9857672bfe6f67fe34bfe1", + "md5": "063e88b4955259a3a3a91b17d2bedae6", + "sha256": "54b0f79b572cb682dacba1a4c9039ea43069cd8b554b00cbafc73e5f85c9e6d6" + }, + "downloads": -1, + "filename": "Djblets-0.9.8-py2.7.egg", + "has_sig": false, + "md5_digest": "063e88b4955259a3a3a91b17d2bedae6", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 1057850, + "upload_time": "2017-06-19T23:11:03", + "upload_time_iso_8601": "2017-06-19T23:11:03.438443Z", + "url": "https://files.pythonhosted.org/packages/f9/cd/bf376970e1befcc8314e99e68a203cd172a00b9857672bfe6f67fe34bfe1/Djblets-0.9.8-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "b5502bd6757be71a1e9a48d0edb6e67c3400bee1aa00d31898a2cdc05301dbc8", + "md5": "750e3ef3563acbf38673af153a05905e", + "sha256": "fc692efa9caca703d873901e1540fe197b89e709bc861dce2d23295696359fce" + }, + "downloads": -1, + "filename": "Djblets-0.9.8-py2-none-any.whl", + "has_sig": false, + "md5_digest": "750e3ef3563acbf38673af153a05905e", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 683956, + "upload_time": "2017-06-19T23:10:52", + "upload_time_iso_8601": "2017-06-19T23:10:52.692028Z", + "url": "https://files.pythonhosted.org/packages/b5/50/2bd6757be71a1e9a48d0edb6e67c3400bee1aa00d31898a2cdc05301dbc8/Djblets-0.9.8-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "14253d614e6cb781828e4567d8cd211e6a0b77dbe9126847502fc36e86e0a824", + "md5": "5c38bef8bee300560afa1a36c90251c7", + "sha256": "68969914e8334622ccfd3458ac4e4fc37eab7449058dddcb85ed7afab139cd41" + }, + "downloads": -1, + "filename": "Djblets-0.9.8.tar.gz", + "has_sig": false, + "md5_digest": "5c38bef8bee300560afa1a36c90251c7", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 422661, + "upload_time": "2017-06-19T23:11:07", + "upload_time_iso_8601": "2017-06-19T23:11:07.980310Z", + "url": "https://files.pythonhosted.org/packages/14/25/3d614e6cb781828e4567d8cd211e6a0b77dbe9126847502fc36e86e0a824/Djblets-0.9.8.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "0.9.9": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "d20cc289b2cf06b0a149e9855808849641f41dde616461ed79fd5f56bd491311", + "md5": "74006cb06baae4daa8b8343dab5adc92", + "sha256": "e9250d17346f8b932ff6cb0d15b09abeecf0d259b34a29a662b6b3ffdbbe56c0" + }, + "downloads": -1, + "filename": "Djblets-0.9.9-py2.6.egg", + "has_sig": false, + "md5_digest": "74006cb06baae4daa8b8343dab5adc92", + "packagetype": "bdist_egg", + "python_version": "2.6", + "requires_python": null, + "size": 1051546, + "upload_time": "2017-08-01T19:55:46", + "upload_time_iso_8601": "2017-08-01T19:55:46.082229Z", + "url": "https://files.pythonhosted.org/packages/d2/0c/c289b2cf06b0a149e9855808849641f41dde616461ed79fd5f56bd491311/Djblets-0.9.9-py2.6.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "cf61d5c5229969e5f1ad996ec15beb56ec459b2baeee4f32a01ef70534884605", + "md5": "b6a0d2a95a5e3c49c50a8ce2b7197c7c", + "sha256": "32e42d0a8094fadd82b2533ffd6c3056f17ef76dfa04aa622490954e664ab041" + }, + "downloads": -1, + "filename": "Djblets-0.9.9-py2.7.egg", + "has_sig": false, + "md5_digest": "b6a0d2a95a5e3c49c50a8ce2b7197c7c", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 1060141, + "upload_time": "2017-08-01T19:55:49", + "upload_time_iso_8601": "2017-08-01T19:55:49.757744Z", + "url": "https://files.pythonhosted.org/packages/cf/61/d5c5229969e5f1ad996ec15beb56ec459b2baeee4f32a01ef70534884605/Djblets-0.9.9-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "965d1269df519a62c2e6e2ee18eec27cacab83c333c18eb6fdae4ec3ed80d8f6", + "md5": "4a78a11f75806d461ea2e5021915fb5d", + "sha256": "1f6592856eba3be3c18aab62fc20800c4140d097bb6c0d54aeea54b7b640f8a3" + }, + "downloads": -1, + "filename": "Djblets-0.9.9-py2-none-any.whl", + "has_sig": false, + "md5_digest": "4a78a11f75806d461ea2e5021915fb5d", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 685418, + "upload_time": "2017-08-01T19:55:42", + "upload_time_iso_8601": "2017-08-01T19:55:42.245930Z", + "url": "https://files.pythonhosted.org/packages/96/5d/1269df519a62c2e6e2ee18eec27cacab83c333c18eb6fdae4ec3ed80d8f6/Djblets-0.9.9-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "e854f83ff22f47fdd1f51e659a53ca4535c373f95da0831eefa2e65f6ff37c9b", + "md5": "9adba232f8bd62c450ad0931a87df519", + "sha256": "53070308fb738b438c6832ac0ee3267018bcd1a30a5f2c5764ed4915c50d5f22" + }, + "downloads": -1, + "filename": "Djblets-0.9.9.tar.gz", + "has_sig": false, + "md5_digest": "9adba232f8bd62c450ad0931a87df519", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 423445, + "upload_time": "2017-08-01T19:55:53", + "upload_time_iso_8601": "2017-08-01T19:55:53.085622Z", + "url": "https://files.pythonhosted.org/packages/e8/54/f83ff22f47fdd1f51e659a53ca4535c373f95da0831eefa2e65f6ff37c9b/Djblets-0.9.9.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "1.0": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "599cbb31a1b611b9f2e8babefa458b59cee274dfe674a5f3ea4e6ed51d43108e", + "md5": "c4197e854aeb6eb681f05635375fea38", + "sha256": "8641683deb22f686bf467dbf772ca923eb5b475bfe215efe64281c600362a139" + }, + "downloads": -1, + "filename": "Djblets-1.0-py2.7.egg", + "has_sig": false, + "md5_digest": "c4197e854aeb6eb681f05635375fea38", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 1690691, + "upload_time": "2017-11-21T07:29:49", + "upload_time_iso_8601": "2017-11-21T07:29:49.554120Z", + "url": "https://files.pythonhosted.org/packages/59/9c/bb31a1b611b9f2e8babefa458b59cee274dfe674a5f3ea4e6ed51d43108e/Djblets-1.0-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "81db56735e28a9505a7d31996b33b00069356a378826ced743471421315ad946", + "md5": "aa06cb43fb1353a4c220c1cb09922a8c", + "sha256": "62f66972389e536e568588a295d4fc8f5b8479137719a1015217fe4e88158649" + }, + "downloads": -1, + "filename": "Djblets-1.0-py2-none-any.whl", + "has_sig": false, + "md5_digest": "aa06cb43fb1353a4c220c1cb09922a8c", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 1044973, + "upload_time": "2017-11-21T07:29:46", + "upload_time_iso_8601": "2017-11-21T07:29:46.165249Z", + "url": "https://files.pythonhosted.org/packages/81/db/56735e28a9505a7d31996b33b00069356a378826ced743471421315ad946/Djblets-1.0-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "a99242114cca43e47b79ea031bde3a3ea1ed6c4b2c377fc271e48f61eed1e747", + "md5": "e5773a9ad59ce4c904721e2aa49f83b6", + "sha256": "536b4ec7871c03c1a15362dc4e3e07e0d70ab035d2e1e55042dbfa1028e6b2a5" + }, + "downloads": -1, + "filename": "Djblets-1.0.tar.gz", + "has_sig": false, + "md5_digest": "e5773a9ad59ce4c904721e2aa49f83b6", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 638875, + "upload_time": "2017-11-21T07:29:52", + "upload_time_iso_8601": "2017-11-21T07:29:52.294894Z", + "url": "https://files.pythonhosted.org/packages/a9/92/42114cca43e47b79ea031bde3a3ea1ed6c4b2c377fc271e48f61eed1e747/Djblets-1.0.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "1.0.1": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "6447904c944a0f11b6d4a51a321391156566f0276ffc3edcc339992dc87709bd", + "md5": "451d2fe3f6d9e8b865f37d9e5fa4a970", + "sha256": "0dff9f56832dd74f05263933dd95fc6973cfea179f3b93e9b5269f88c27f8494" + }, + "downloads": -1, + "filename": "Djblets-1.0.1-py2.7.egg", + "has_sig": false, + "md5_digest": "451d2fe3f6d9e8b865f37d9e5fa4a970", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 1691961, + "upload_time": "2017-12-05T23:40:02", + "upload_time_iso_8601": "2017-12-05T23:40:02.516872Z", + "url": "https://files.pythonhosted.org/packages/64/47/904c944a0f11b6d4a51a321391156566f0276ffc3edcc339992dc87709bd/Djblets-1.0.1-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "3115aa84994294a20b4b7ee2b1816eb20493c1d460eb76455a24058fa683700b", + "md5": "05f32fac2ebc501d360fe4ee33df3d7a", + "sha256": "71e74add58589805780d6213a65459150776aa07e2a0d93480d7ec9f3b885155" + }, + "downloads": -1, + "filename": "Djblets-1.0.1-py2-none-any.whl", + "has_sig": false, + "md5_digest": "05f32fac2ebc501d360fe4ee33df3d7a", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 1045684, + "upload_time": "2017-12-05T23:39:55", + "upload_time_iso_8601": "2017-12-05T23:39:55.913336Z", + "url": "https://files.pythonhosted.org/packages/31/15/aa84994294a20b4b7ee2b1816eb20493c1d460eb76455a24058fa683700b/Djblets-1.0.1-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "363cd4b9b10df9362ff523ca67879fa0981b7b7faa88dab220a029f1fa369434", + "md5": "4d86d17de746149051193be177d46b95", + "sha256": "c9347007e7a50dde897013ffd0c30f49364eff733164a82c5ceaf3e9a90871d4" + }, + "downloads": -1, + "filename": "Djblets-1.0.1.tar.gz", + "has_sig": false, + "md5_digest": "4d86d17de746149051193be177d46b95", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 639791, + "upload_time": "2017-12-05T23:40:07", + "upload_time_iso_8601": "2017-12-05T23:40:07.861672Z", + "url": "https://files.pythonhosted.org/packages/36/3c/d4b9b10df9362ff523ca67879fa0981b7b7faa88dab220a029f1fa369434/Djblets-1.0.1.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "1.0.10": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "2e0422a9d99233b69a74208233ecdcf1034427702c75a186a5ba2f4e9f401e22", + "md5": "ff23b9a2c277232f30f7c1122977753f", + "sha256": "82bbcf99fad35fdebba4a60b1b5ae2833f52c9e54224c57e998b274bb42ee838" + }, + "downloads": -1, + "filename": "Djblets-1.0.10-py2.7.egg", + "has_sig": false, + "md5_digest": "ff23b9a2c277232f30f7c1122977753f", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 2024461, + "upload_time": "2019-02-19T10:36:01", + "upload_time_iso_8601": "2019-02-19T10:36:01.018640Z", + "url": "https://files.pythonhosted.org/packages/2e/04/22a9d99233b69a74208233ecdcf1034427702c75a186a5ba2f4e9f401e22/Djblets-1.0.10-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "0ddf7b0b13eb6ae0011d9fb8e1019cf54509b36572eae0fc99faa2df2b343a86", + "md5": "e2e6a59712c7f9eeb5ed10ae6aba3634", + "sha256": "a49fa21d487b9f985304d478362a5e2dfbf9e61149ecdad8cf1b951fdfe38fb8" + }, + "downloads": -1, + "filename": "Djblets-1.0.10-py2-none-any.whl", + "has_sig": false, + "md5_digest": "e2e6a59712c7f9eeb5ed10ae6aba3634", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 1259731, + "upload_time": "2019-02-19T10:35:56", + "upload_time_iso_8601": "2019-02-19T10:35:56.657597Z", + "url": "https://files.pythonhosted.org/packages/0d/df/7b0b13eb6ae0011d9fb8e1019cf54509b36572eae0fc99faa2df2b343a86/Djblets-1.0.10-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "086c20cf538c6e1ce8a46dd97eb089035ad28059d46af5368fc89ff710a720c5", + "md5": "660aa521360ce0fc1b1af31f9be7ed01", + "sha256": "e14910161fddcc368198fff29ba9174607d9b0eb0d7b41e6650b77fc1ecfa0e9" + }, + "downloads": -1, + "filename": "Djblets-1.0.10.tar.gz", + "has_sig": false, + "md5_digest": "660aa521360ce0fc1b1af31f9be7ed01", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 968574, + "upload_time": "2019-02-19T10:36:04", + "upload_time_iso_8601": "2019-02-19T10:36:04.556612Z", + "url": "https://files.pythonhosted.org/packages/08/6c/20cf538c6e1ce8a46dd97eb089035ad28059d46af5368fc89ff710a720c5/Djblets-1.0.10.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "1.0.11": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "a9e1047fa7cf0ccd32840a47b5ca0eaf2d0144d6f8e377fbed6882c8ce4a8aaf", + "md5": "6cb735b917c6b3906613fd7166f09a15", + "sha256": "906ed82236f75c6eb01b9936f4aef0e5cd0f97b2af2e7dcbc859467e67324515" + }, + "downloads": -1, + "filename": "Djblets-1.0.11-py2.7.egg", + "has_sig": false, + "md5_digest": "6cb735b917c6b3906613fd7166f09a15", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 2109818, + "upload_time": "2019-03-19T07:53:02", + "upload_time_iso_8601": "2019-03-19T07:53:02.941155Z", + "url": "https://files.pythonhosted.org/packages/a9/e1/047fa7cf0ccd32840a47b5ca0eaf2d0144d6f8e377fbed6882c8ce4a8aaf/Djblets-1.0.11-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "0d889347b59d123f8110f9dfe9f02b8d1b6c768d7c60ef5b76afa85984eec4a2", + "md5": "0de7f3177d3817ced182fd75249a08d5", + "sha256": "6fde1a98657966d195a808c167dbe11103af61fda9f121900b3e8621f3061333" + }, + "downloads": -1, + "filename": "Djblets-1.0.11-py2-none-any.whl", + "has_sig": false, + "md5_digest": "0de7f3177d3817ced182fd75249a08d5", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 1341575, + "upload_time": "2019-03-19T07:52:58", + "upload_time_iso_8601": "2019-03-19T07:52:58.347704Z", + "url": "https://files.pythonhosted.org/packages/0d/88/9347b59d123f8110f9dfe9f02b8d1b6c768d7c60ef5b76afa85984eec4a2/Djblets-1.0.11-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "2baae84e50cacaed2aaae821ab5143535c62730c0b2febbc1ec1f499fdf7e89d", + "md5": "e49505459fcd8dec774ef3b84d77f0d6", + "sha256": "20d0a027e2358e1df2c2a14a2f3aff26a822beb2b0139007b9550eb0bba71440" + }, + "downloads": -1, + "filename": "Djblets-1.0.11.tar.gz", + "has_sig": false, + "md5_digest": "e49505459fcd8dec774ef3b84d77f0d6", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 1033622, + "upload_time": "2019-03-19T07:53:06", + "upload_time_iso_8601": "2019-03-19T07:53:06.333147Z", + "url": "https://files.pythonhosted.org/packages/2b/aa/e84e50cacaed2aaae821ab5143535c62730c0b2febbc1ec1f499fdf7e89d/Djblets-1.0.11.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "1.0.12": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "93c10691f04da6f9b0ff24b16ebf4282e5a6c952782372dc79c053660625f960", + "md5": "ee994af3980f52e5945054475d9e572c", + "sha256": "acec7407176d347b7c76a6911c2ed341361cc83e7adc54a81027b0acb9e9758b" + }, + "downloads": -1, + "filename": "Djblets-1.0.12-py2.7.egg", + "has_sig": false, + "md5_digest": "ee994af3980f52e5945054475d9e572c", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 2110846, + "upload_time": "2019-06-12T05:21:37", + "upload_time_iso_8601": "2019-06-12T05:21:37.826733Z", + "url": "https://files.pythonhosted.org/packages/93/c1/0691f04da6f9b0ff24b16ebf4282e5a6c952782372dc79c053660625f960/Djblets-1.0.12-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "f0eece6f004af43807d7b3d3dfacce61dc4b0c2468c1d4e6027ca3a576903b93", + "md5": "b2ba0b219a48112f111e0f9bcef018e4", + "sha256": "e4bd7d3bf28d272f4905fbdc1eb64ad1712f12a9eb464f07af3d58820550439f" + }, + "downloads": -1, + "filename": "Djblets-1.0.12-py2-none-any.whl", + "has_sig": false, + "md5_digest": "b2ba0b219a48112f111e0f9bcef018e4", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 1342072, + "upload_time": "2019-06-12T05:21:26", + "upload_time_iso_8601": "2019-06-12T05:21:26.826989Z", + "url": "https://files.pythonhosted.org/packages/f0/ee/ce6f004af43807d7b3d3dfacce61dc4b0c2468c1d4e6027ca3a576903b93/Djblets-1.0.12-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "728eae6dadff544fb0499f9491589d0621ee2f5a9705880c4d0784c22c72233a", + "md5": "cd77e549c9dde2b49c970d625d68384d", + "sha256": "95ff3c53916bf2bd39cdeb480db433fe49e604f58ddc447635e76bf85f2628f3" + }, + "downloads": -1, + "filename": "Djblets-1.0.12.tar.gz", + "has_sig": false, + "md5_digest": "cd77e549c9dde2b49c970d625d68384d", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 1034311, + "upload_time": "2019-06-12T05:21:43", + "upload_time_iso_8601": "2019-06-12T05:21:43.341447Z", + "url": "https://files.pythonhosted.org/packages/72/8e/ae6dadff544fb0499f9491589d0621ee2f5a9705880c4d0784c22c72233a/Djblets-1.0.12.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "1.0.13": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "e72e1a3a193a43b5e6935fd24412e54502e5d9a697832f8ff217907fe24c3b06", + "md5": "2b63da6cbb49687245649122158bb904", + "sha256": "97a86df080d51a1c6a5ccaaefd7be81ea7dd20d467bd504d9e8c94d7240b4435" + }, + "downloads": -1, + "filename": "Djblets-1.0.13-py2.7.egg", + "has_sig": false, + "md5_digest": "2b63da6cbb49687245649122158bb904", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 2145040, + "upload_time": "2020-06-30T06:35:44", + "upload_time_iso_8601": "2020-06-30T06:35:44.833936Z", + "url": "https://files.pythonhosted.org/packages/e7/2e/1a3a193a43b5e6935fd24412e54502e5d9a697832f8ff217907fe24c3b06/Djblets-1.0.13-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "2bf2255747b6489444dfae0dd71c9a18ca9df89064004b981f1d0e002547e513", + "md5": "d95747a9ee1b62f145c64ef3f8f41413", + "sha256": "19794924dc4fa5328da83eaea76631b86ecd7fd2db5a61b6540d9ed140512d7d" + }, + "downloads": -1, + "filename": "Djblets-1.0.13-py2-none-any.whl", + "has_sig": false, + "md5_digest": "d95747a9ee1b62f145c64ef3f8f41413", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 1355421, + "upload_time": "2020-06-30T06:35:41", + "upload_time_iso_8601": "2020-06-30T06:35:41.534685Z", + "url": "https://files.pythonhosted.org/packages/2b/f2/255747b6489444dfae0dd71c9a18ca9df89064004b981f1d0e002547e513/Djblets-1.0.13-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "986a52a804bbf015b561f601c769629bfa16e9c8178fb1e38f45c5c303b22094", + "md5": "77832b36ed4a0860f82397170f87ccec", + "sha256": "43c526bfa60576ba61fbb711f8717814fd08cdc769b0649b48b413f7929c9804" + }, + "downloads": -1, + "filename": "Djblets-1.0.13.tar.gz", + "has_sig": false, + "md5_digest": "77832b36ed4a0860f82397170f87ccec", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 1040962, + "upload_time": "2020-06-30T06:35:47", + "upload_time_iso_8601": "2020-06-30T06:35:47.512638Z", + "url": "https://files.pythonhosted.org/packages/98/6a/52a804bbf015b561f601c769629bfa16e9c8178fb1e38f45c5c303b22094/Djblets-1.0.13.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "1.0.14": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "2c6a7930b2cb84eea5a56675134d67b2073d0537ddaf1a38360e43d262414ce1", + "md5": "714d9b1924aa2e14aade829460221dcb", + "sha256": "0f68d7dc8f006cf5603f447a0011afbf40a5fba479d8c164284e36f85e5520a5" + }, + "downloads": -1, + "filename": "Djblets-1.0.14-py2.7.egg", + "has_sig": false, + "md5_digest": "714d9b1924aa2e14aade829460221dcb", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 2146132, + "upload_time": "2020-12-10T09:50:52", + "upload_time_iso_8601": "2020-12-10T09:50:52.580765Z", + "url": "https://files.pythonhosted.org/packages/2c/6a/7930b2cb84eea5a56675134d67b2073d0537ddaf1a38360e43d262414ce1/Djblets-1.0.14-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "dbed25ee96761e3f78fd3a3002f77cd6d4727b3f3a2158bf884c0984ca599871", + "md5": "0202771703077f40479e4dc7b8f3017d", + "sha256": "0cf1cb9512fbfea0a99ba85dcdc64e4f351a4479a0b1afc20eb1e105f731f44b" + }, + "downloads": -1, + "filename": "Djblets-1.0.14-py2-none-any.whl", + "has_sig": false, + "md5_digest": "0202771703077f40479e4dc7b8f3017d", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 1355897, + "upload_time": "2020-12-10T09:50:50", + "upload_time_iso_8601": "2020-12-10T09:50:50.256687Z", + "url": "https://files.pythonhosted.org/packages/db/ed/25ee96761e3f78fd3a3002f77cd6d4727b3f3a2158bf884c0984ca599871/Djblets-1.0.14-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "5353a6c824ef39fc665b74a12e4373c658b821a8ef1d88611e700169c9e94e84", + "md5": "ba9d4544dbe37017996935456f7477cd", + "sha256": "9f2a2174dcb20e75339642e6fce360ddcf3daf1c65a6e561ed26a65a4df56557" + }, + "downloads": -1, + "filename": "Djblets-1.0.14.tar.gz", + "has_sig": false, + "md5_digest": "ba9d4544dbe37017996935456f7477cd", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 1041973, + "upload_time": "2020-12-10T09:50:55", + "upload_time_iso_8601": "2020-12-10T09:50:55.012440Z", + "url": "https://files.pythonhosted.org/packages/53/53/a6c824ef39fc665b74a12e4373c658b821a8ef1d88611e700169c9e94e84/Djblets-1.0.14.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "1.0.15": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "09e9131d7cdb4d167bf0e4b5b3884c2ba0ceca0c20a109c73c7efe25f1775bb9", + "md5": "cbc7e5e182b3f9e2e7e7afacf5d1f018", + "sha256": "b09a8152d003db14ac0764c9b5bc78d5fbb8c5578f71bb56a3c93b262e230b34" + }, + "downloads": -1, + "filename": "Djblets-1.0.15-py2.7.egg", + "has_sig": false, + "md5_digest": "cbc7e5e182b3f9e2e7e7afacf5d1f018", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 2144630, + "upload_time": "2021-01-20T09:11:12", + "upload_time_iso_8601": "2021-01-20T09:11:12.023409Z", + "url": "https://files.pythonhosted.org/packages/09/e9/131d7cdb4d167bf0e4b5b3884c2ba0ceca0c20a109c73c7efe25f1775bb9/Djblets-1.0.15-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "5562a54ea6dd2d83be09dbeac29a2f894b73dd98c0a8212ba60543f0498cedb3", + "md5": "6de0a14de9f351464615a0aacfecb8f0", + "sha256": "5bacc2957d990da1de17e84880b2c29bcb4eb958a8e0cf679a4920f0a367bfd7" + }, + "downloads": -1, + "filename": "Djblets-1.0.15-py2-none-any.whl", + "has_sig": false, + "md5_digest": "6de0a14de9f351464615a0aacfecb8f0", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 1355934, + "upload_time": "2021-01-20T09:11:06", + "upload_time_iso_8601": "2021-01-20T09:11:06.777550Z", + "url": "https://files.pythonhosted.org/packages/55/62/a54ea6dd2d83be09dbeac29a2f894b73dd98c0a8212ba60543f0498cedb3/Djblets-1.0.15-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "22bc15c8bf51691528f4e65b67e526d4349368d82bb4650b37107eb9c24873dd", + "md5": "06247f2a33ad6b5192c00b50a3f010e9", + "sha256": "ae885cee9560902ff608656ecc90f1927d8a4dfdaf8fc900e17b250f4f3c7fb7" + }, + "downloads": -1, + "filename": "Djblets-1.0.15.tar.gz", + "has_sig": false, + "md5_digest": "06247f2a33ad6b5192c00b50a3f010e9", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 1042277, + "upload_time": "2021-01-20T09:11:15", + "upload_time_iso_8601": "2021-01-20T09:11:15.176142Z", + "url": "https://files.pythonhosted.org/packages/22/bc/15c8bf51691528f4e65b67e526d4349368d82bb4650b37107eb9c24873dd/Djblets-1.0.15.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "1.0.16": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "f9c3ebf0cc4fdec6e6c9d4c9037606fcf47a3fea3232238b60b023ed2a3cccce", + "md5": "33ba45f6b62fd5de916ebdc059230aa9", + "sha256": "a3e4f0c0edca047532e0634fe879065b9b12d9cc2f99556b6feeac36b4164b28" + }, + "downloads": -1, + "filename": "Djblets-1.0.16-py2.7.egg", + "has_sig": false, + "md5_digest": "33ba45f6b62fd5de916ebdc059230aa9", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 2144841, + "upload_time": "2021-02-12T07:22:47", + "upload_time_iso_8601": "2021-02-12T07:22:47.307781Z", + "url": "https://files.pythonhosted.org/packages/f9/c3/ebf0cc4fdec6e6c9d4c9037606fcf47a3fea3232238b60b023ed2a3cccce/Djblets-1.0.16-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "37a660b4663a19a460646ac11c62be7d2e5c334d299cb815f49a64290e893133", + "md5": "c1237d4bb7ccddb67cd4d3bd195e5b02", + "sha256": "d37d67a26d718af8c1eca7f49ec1f3833f1eeec9b86c2b5874840775c288250a" + }, + "downloads": -1, + "filename": "Djblets-1.0.16-py2-none-any.whl", + "has_sig": false, + "md5_digest": "c1237d4bb7ccddb67cd4d3bd195e5b02", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 1355935, + "upload_time": "2021-02-12T07:22:45", + "upload_time_iso_8601": "2021-02-12T07:22:45.046900Z", + "url": "https://files.pythonhosted.org/packages/37/a6/60b4663a19a460646ac11c62be7d2e5c334d299cb815f49a64290e893133/Djblets-1.0.16-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "0b4b33d9c2ee327400ce2be06a5c6ad8f442e629c89fa3af180c3e0f228cba21", + "md5": "cd2ba24b365905d62ffa63f08018da3f", + "sha256": "94d6c08c255128d6bc9c85272eae0ab02757c03897293013bf92f6a3ed9cdd5a" + }, + "downloads": -1, + "filename": "Djblets-1.0.16.tar.gz", + "has_sig": false, + "md5_digest": "cd2ba24b365905d62ffa63f08018da3f", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 1042411, + "upload_time": "2021-02-12T07:22:49", + "upload_time_iso_8601": "2021-02-12T07:22:49.844813Z", + "url": "https://files.pythonhosted.org/packages/0b/4b/33d9c2ee327400ce2be06a5c6ad8f442e629c89fa3af180c3e0f228cba21/Djblets-1.0.16.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "1.0.17": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "bc07dd0a2d3b09e077f2c9516cb489b995b61a689c0b7bcef8bf70e37c836f22", + "md5": "4d0c403bf80dac96995fb2d8af76f8c5", + "sha256": "91d1d3576286e94b5fea29688968a08b6bac1ba13ebdc15c73c3f891b7bde64f" + }, + "downloads": -1, + "filename": "Djblets-1.0.17-py2.7.egg", + "has_sig": false, + "md5_digest": "4d0c403bf80dac96995fb2d8af76f8c5", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 2145362, + "upload_time": "2021-04-14T02:11:17", + "upload_time_iso_8601": "2021-04-14T02:11:17.916009Z", + "url": "https://files.pythonhosted.org/packages/bc/07/dd0a2d3b09e077f2c9516cb489b995b61a689c0b7bcef8bf70e37c836f22/Djblets-1.0.17-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "5c4141576b247d58220b98d18ef61e9152e17ca60d44cb087f44e2c041c41ec9", + "md5": "3e26130ed097a7313464159cb87e7b7e", + "sha256": "535f7af3f9a268b0b481790176922810d00e4f7cd4bd39248f62fa45c9cc96dc" + }, + "downloads": -1, + "filename": "Djblets-1.0.17-py2-none-any.whl", + "has_sig": false, + "md5_digest": "3e26130ed097a7313464159cb87e7b7e", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 1356090, + "upload_time": "2021-04-14T02:11:15", + "upload_time_iso_8601": "2021-04-14T02:11:15.882376Z", + "url": "https://files.pythonhosted.org/packages/5c/41/41576b247d58220b98d18ef61e9152e17ca60d44cb087f44e2c041c41ec9/Djblets-1.0.17-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "bbfd9c2577d35d7f74801fb6e53106e8e5c005c988dc345f8287a0b86f3a47e1", + "md5": "95331163ece2f2c887fe1c24a0346583", + "sha256": "3a4bfd982e1ae11f391b3907592ae721dd549be03b16e1f0c20d9182337e5725" + }, + "downloads": -1, + "filename": "Djblets-1.0.17.tar.gz", + "has_sig": false, + "md5_digest": "95331163ece2f2c887fe1c24a0346583", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 1042797, + "upload_time": "2021-04-14T02:11:20", + "upload_time_iso_8601": "2021-04-14T02:11:20.188424Z", + "url": "https://files.pythonhosted.org/packages/bb/fd/9c2577d35d7f74801fb6e53106e8e5c005c988dc345f8287a0b86f3a47e1/Djblets-1.0.17.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "1.0.18": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "8fe2c87c60962babfb3930bf7f8f262712be48a8ecd345937ba02e3e5a4c3bc0", + "md5": "55110fe23665b248f18337e0a9db76e5", + "sha256": "09753418061e82afbc10dc8b585fe9166d48ff41cd337212b4be374406b59c3e" + }, + "downloads": -1, + "filename": "Djblets-1.0.18-py2.7.egg", + "has_sig": false, + "md5_digest": "55110fe23665b248f18337e0a9db76e5", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 2147015, + "upload_time": "2021-06-15T07:17:00", + "upload_time_iso_8601": "2021-06-15T07:17:00.389346Z", + "url": "https://files.pythonhosted.org/packages/8f/e2/c87c60962babfb3930bf7f8f262712be48a8ecd345937ba02e3e5a4c3bc0/Djblets-1.0.18-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "1a98e81de674d3b8f44e34bfc4c65862ce408b5799acbd7ea5008b7ac58ef7ca", + "md5": "8bd81a33d5c19095233dbbe3bb757876", + "sha256": "66b763b61ea2d91ed389f5b7dbd1a1501ae7e91a1da86818d8b6847e2106d9e8" + }, + "downloads": -1, + "filename": "Djblets-1.0.18-py2-none-any.whl", + "has_sig": false, + "md5_digest": "8bd81a33d5c19095233dbbe3bb757876", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 1356749, + "upload_time": "2021-06-15T07:16:57", + "upload_time_iso_8601": "2021-06-15T07:16:57.678824Z", + "url": "https://files.pythonhosted.org/packages/1a/98/e81de674d3b8f44e34bfc4c65862ce408b5799acbd7ea5008b7ac58ef7ca/Djblets-1.0.18-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "71fe3a944513436e3c7d14895a9995b2c2fdbacc76d614a0f8153897e137c7ad", + "md5": "cb79d46c2081d3f9d39b50a2e592d3f5", + "sha256": "1bc87e9dd2b7d648f29cb07742b2fd140208ee9978d85d7972c635ddcbd9b5e7" + }, + "downloads": -1, + "filename": "Djblets-1.0.18.tar.gz", + "has_sig": false, + "md5_digest": "cb79d46c2081d3f9d39b50a2e592d3f5", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 1043603, + "upload_time": "2021-06-15T07:17:02", + "upload_time_iso_8601": "2021-06-15T07:17:02.882111Z", + "url": "https://files.pythonhosted.org/packages/71/fe/3a944513436e3c7d14895a9995b2c2fdbacc76d614a0f8153897e137c7ad/Djblets-1.0.18.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "1.0.2": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "e00fad162aa1cfcdbe7f0b152565ebd3eba86ed9f9e74c8ea03287982afe2635", + "md5": "c7d09633715abb7ecd2f4afa1e17cd44", + "sha256": "c5a5963a6ee825c5b29029d477f37f2c584ae5af23164da69b33e5c34bb775b0" + }, + "downloads": -1, + "filename": "Djblets-1.0.2-py2.7.egg", + "has_sig": false, + "md5_digest": "c7d09633715abb7ecd2f4afa1e17cd44", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 1707426, + "upload_time": "2018-01-09T08:40:34", + "upload_time_iso_8601": "2018-01-09T08:40:34.097391Z", + "url": "https://files.pythonhosted.org/packages/e0/0f/ad162aa1cfcdbe7f0b152565ebd3eba86ed9f9e74c8ea03287982afe2635/Djblets-1.0.2-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "6c52f06d8318a574306fecae0f0e2d4493ba61c91ad3569fe566cff2fb575858", + "md5": "8ba446b456b6fe1f8494d9f6650ebc3c", + "sha256": "aded3a4e1d62eb0e59c54ea13005c7b85c82d800b4fee982e2c957190af0c038" + }, + "downloads": -1, + "filename": "Djblets-1.0.2-py2-none-any.whl", + "has_sig": false, + "md5_digest": "8ba446b456b6fe1f8494d9f6650ebc3c", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 1052093, + "upload_time": "2018-01-09T08:40:30", + "upload_time_iso_8601": "2018-01-09T08:40:30.728441Z", + "url": "https://files.pythonhosted.org/packages/6c/52/f06d8318a574306fecae0f0e2d4493ba61c91ad3569fe566cff2fb575858/Djblets-1.0.2-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "2af2027766d25dc3e6b4df45504da511b8e92969ae2b9eaabd5d1ece25bbd852", + "md5": "ac9b60067c19a46869655ad3f2197b4e", + "sha256": "78b4b4251135019f5db7aba9b840020139d1c59df3dcb89925a8e341feb60624" + }, + "downloads": -1, + "filename": "Djblets-1.0.2.tar.gz", + "has_sig": false, + "md5_digest": "ac9b60067c19a46869655ad3f2197b4e", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 645035, + "upload_time": "2018-01-09T08:40:36", + "upload_time_iso_8601": "2018-01-09T08:40:36.936214Z", + "url": "https://files.pythonhosted.org/packages/2a/f2/027766d25dc3e6b4df45504da511b8e92969ae2b9eaabd5d1ece25bbd852/Djblets-1.0.2.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "1.0.3": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "e9144804994d4e2953a66b4c251341b7b2d158af65bdc9169d8e8b721c47d539", + "md5": "36b8f02d2bfbbd3e7bf86a4fd2fc974f", + "sha256": "c68f83041e65474398e9195467bbf4bbe294dee3b851803297804ea30c59f0e7" + }, + "downloads": -1, + "filename": "Djblets-1.0.3-py2.7.egg", + "has_sig": false, + "md5_digest": "36b8f02d2bfbbd3e7bf86a4fd2fc974f", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 1807797, + "upload_time": "2018-02-06T07:52:08", + "upload_time_iso_8601": "2018-02-06T07:52:08.419984Z", + "url": "https://files.pythonhosted.org/packages/e9/14/4804994d4e2953a66b4c251341b7b2d158af65bdc9169d8e8b721c47d539/Djblets-1.0.3-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "ee319369524aa75f6c9916f0b766bf47f01cd26a0c303c3944580b3bcdac1ab1", + "md5": "3c9145bf0667f47524157749e5266a07", + "sha256": "ec1350b89ea6a6ed7f7e26a13efb6c0b6651efaffe71b7b9fa2286bd150ba5a5" + }, + "downloads": -1, + "filename": "Djblets-1.0.3-py2-none-any.whl", + "has_sig": false, + "md5_digest": "3c9145bf0667f47524157749e5266a07", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 1147485, + "upload_time": "2018-02-06T07:52:05", + "upload_time_iso_8601": "2018-02-06T07:52:05.198139Z", + "url": "https://files.pythonhosted.org/packages/ee/31/9369524aa75f6c9916f0b766bf47f01cd26a0c303c3944580b3bcdac1ab1/Djblets-1.0.3-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "838c9ed857fe9897d864df268177c3f2922160f6d717ea063e6db40846f6aa00", + "md5": "b29def116381c12b2e64bc8e9f40c3c1", + "sha256": "b80c58edecc376cf47db76916485a0e6cacfc7b58b120300e7d75b49aec66966" + }, + "downloads": -1, + "filename": "Djblets-1.0.3.tar.gz", + "has_sig": false, + "md5_digest": "b29def116381c12b2e64bc8e9f40c3c1", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 821880, + "upload_time": "2018-02-06T07:52:11", + "upload_time_iso_8601": "2018-02-06T07:52:11.555307Z", + "url": "https://files.pythonhosted.org/packages/83/8c/9ed857fe9897d864df268177c3f2922160f6d717ea063e6db40846f6aa00/Djblets-1.0.3.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "1.0.4": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "4cf4d5419dd49121438c18f6c25a091d84a4e54a81125692436ac6aba0230723", + "md5": "471f01dc3179268f08e8e5392b600192", + "sha256": "3a1d33d1776c2c3664f39ae70a5858b0b784704a2171ef4f8032acb205e2afea" + }, + "downloads": -1, + "filename": "Djblets-1.0.4-py2.7.egg", + "has_sig": false, + "md5_digest": "471f01dc3179268f08e8e5392b600192", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 1878111, + "upload_time": "2018-03-20T06:40:16", + "upload_time_iso_8601": "2018-03-20T06:40:16.084700Z", + "url": "https://files.pythonhosted.org/packages/4c/f4/d5419dd49121438c18f6c25a091d84a4e54a81125692436ac6aba0230723/Djblets-1.0.4-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "5711b7d12a4acfc4208f88ba7b869a750144abb8c1bf4248a939b1b9ee55a08f", + "md5": "e82835886add03011ebb45f4703e6839", + "sha256": "f3e5017387abfefd25dd17ec9b3f13d13adfb69737d4fd35c38f01ea3844cb12" + }, + "downloads": -1, + "filename": "Djblets-1.0.4-py2-none-any.whl", + "has_sig": false, + "md5_digest": "e82835886add03011ebb45f4703e6839", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 1185084, + "upload_time": "2018-03-20T06:40:09", + "upload_time_iso_8601": "2018-03-20T06:40:09.475980Z", + "url": "https://files.pythonhosted.org/packages/57/11/b7d12a4acfc4208f88ba7b869a750144abb8c1bf4248a939b1b9ee55a08f/Djblets-1.0.4-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "f5904218ee9bc3812c909edd30ca01468941842111be18c358502c218c57485d", + "md5": "7de112c9fabbb6faf0f15d0522dd9b33", + "sha256": "7eacb0d0b85e3b6bf44182b380c7745af2a77003ff252cd8bdee61195707ed03" + }, + "downloads": -1, + "filename": "Djblets-1.0.4.tar.gz", + "has_sig": false, + "md5_digest": "7de112c9fabbb6faf0f15d0522dd9b33", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 910273, + "upload_time": "2018-03-20T06:40:21", + "upload_time_iso_8601": "2018-03-20T06:40:21.420724Z", + "url": "https://files.pythonhosted.org/packages/f5/90/4218ee9bc3812c909edd30ca01468941842111be18c358502c218c57485d/Djblets-1.0.4.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "1.0.5": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "65243fc4f1a54ba6bffcb7624ac25abd664a7bde00e6e21626828a4fad06a343", + "md5": "f24f4b0ef166c9f69ba32eeb1ae063cd", + "sha256": "695036c6d0132e9cccff79be531ac3c120eecf063b6f2b3a53cb6826dda57fa5" + }, + "downloads": -1, + "filename": "Djblets-1.0.5-py2.7.egg", + "has_sig": false, + "md5_digest": "f24f4b0ef166c9f69ba32eeb1ae063cd", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 1878190, + "upload_time": "2018-03-23T07:12:19", + "upload_time_iso_8601": "2018-03-23T07:12:19.168592Z", + "url": "https://files.pythonhosted.org/packages/65/24/3fc4f1a54ba6bffcb7624ac25abd664a7bde00e6e21626828a4fad06a343/Djblets-1.0.5-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "89f9617242cb97a2349b2e65ae4b763c696bb4057f5685576ad174cf9560f395", + "md5": "e3875964b8b0ef026d8aca52c3143df5", + "sha256": "146dcf32f37e9728498e1d9c61321cfe1a668c421499ae5d85f5a1d8ee92b296" + }, + "downloads": -1, + "filename": "Djblets-1.0.5-py2-none-any.whl", + "has_sig": false, + "md5_digest": "e3875964b8b0ef026d8aca52c3143df5", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 1185064, + "upload_time": "2018-03-23T07:12:14", + "upload_time_iso_8601": "2018-03-23T07:12:14.882193Z", + "url": "https://files.pythonhosted.org/packages/89/f9/617242cb97a2349b2e65ae4b763c696bb4057f5685576ad174cf9560f395/Djblets-1.0.5-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "f0b1ca0a489867e70dcfbf8b3308ca785a2783524ac977ee38211806a7d0c07e", + "md5": "a126bf8daf8817c8d7229c9723f63f4d", + "sha256": "d4c29d8b955463f836ac700fa1c185c2bf712eef91b74ac74d918e95afe5c58d" + }, + "downloads": -1, + "filename": "Djblets-1.0.5.tar.gz", + "has_sig": false, + "md5_digest": "a126bf8daf8817c8d7229c9723f63f4d", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 910135, + "upload_time": "2018-03-23T07:12:22", + "upload_time_iso_8601": "2018-03-23T07:12:22.151189Z", + "url": "https://files.pythonhosted.org/packages/f0/b1/ca0a489867e70dcfbf8b3308ca785a2783524ac977ee38211806a7d0c07e/Djblets-1.0.5.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "1.0.6": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "6ae3ebfde1c0414cccbbedfd74992c10ceddee505c6377881e9505dd0513d299", + "md5": "3573ecbc2797c5517d63d74ada975e11", + "sha256": "ff2a9cb6590b4334a8d3b870b96e20e3ab87d461275b2b034f7a8b3c5350c735" + }, + "downloads": -1, + "filename": "Djblets-1.0.6-py2.7.egg", + "has_sig": false, + "md5_digest": "3573ecbc2797c5517d63d74ada975e11", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 1991312, + "upload_time": "2018-06-05T07:54:06", + "upload_time_iso_8601": "2018-06-05T07:54:06.210317Z", + "url": "https://files.pythonhosted.org/packages/6a/e3/ebfde1c0414cccbbedfd74992c10ceddee505c6377881e9505dd0513d299/Djblets-1.0.6-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "6b28ff50c1b1a7292f18b6850157fce7b5bb45d77e27247fac166726d0ba89cd", + "md5": "ba2c5c50c81b2d72a2bdd04056caa8db", + "sha256": "d4d1b9396084a24527f26f593f6b0c9f4dd4b1bfe1eed3869df3fdab7e1cb27e" + }, + "downloads": -1, + "filename": "Djblets-1.0.6-py2-none-any.whl", + "has_sig": false, + "md5_digest": "ba2c5c50c81b2d72a2bdd04056caa8db", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 1241010, + "upload_time": "2018-06-05T07:54:02", + "upload_time_iso_8601": "2018-06-05T07:54:02.698839Z", + "url": "https://files.pythonhosted.org/packages/6b/28/ff50c1b1a7292f18b6850157fce7b5bb45d77e27247fac166726d0ba89cd/Djblets-1.0.6-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "f1e5a693954c3104c2f0906a295f0a3ca1cc4d0493132b5d7464730f0e9fe35d", + "md5": "b17755896b878b78b31030a37bf026e2", + "sha256": "228d69ee1d08344eaa3b923b32f2da4d4e023578d4eed7f748189a58a66e7095" + }, + "downloads": -1, + "filename": "Djblets-1.0.6.tar.gz", + "has_sig": false, + "md5_digest": "b17755896b878b78b31030a37bf026e2", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 946887, + "upload_time": "2018-06-05T07:54:08", + "upload_time_iso_8601": "2018-06-05T07:54:08.985947Z", + "url": "https://files.pythonhosted.org/packages/f1/e5/a693954c3104c2f0906a295f0a3ca1cc4d0493132b5d7464730f0e9fe35d/Djblets-1.0.6.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "1.0.7": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "c932fcafebac6e8cd4d3ffeeabdb04f3d6e00377bc6710efc653bc82ee3adb6e", + "md5": "22f8fe44e77fbcc4b479da5ba4c9d6c4", + "sha256": "165255494033ceffcb53b83aeb283b705c8bab87bede95fb286ab5b5d883029f" + }, + "downloads": -1, + "filename": "Djblets-1.0.7-py2.7.egg", + "has_sig": false, + "md5_digest": "22f8fe44e77fbcc4b479da5ba4c9d6c4", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 1999455, + "upload_time": "2018-10-12T07:40:14", + "upload_time_iso_8601": "2018-10-12T07:40:14.727094Z", + "url": "https://files.pythonhosted.org/packages/c9/32/fcafebac6e8cd4d3ffeeabdb04f3d6e00377bc6710efc653bc82ee3adb6e/Djblets-1.0.7-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "72fd7ce24a25def58f564a50270c8e8c0a693086cc08ba6e55fc8aaa78911f4c", + "md5": "5d6f686ae7b93b565763b600cc62de07", + "sha256": "2994992159b1b4adc9eb45715f69e4b0ab6c5575181ea6c89b494601958e435a" + }, + "downloads": -1, + "filename": "Djblets-1.0.7-py2-none-any.whl", + "has_sig": false, + "md5_digest": "5d6f686ae7b93b565763b600cc62de07", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 1247978, + "upload_time": "2018-10-12T07:40:10", + "upload_time_iso_8601": "2018-10-12T07:40:10.045588Z", + "url": "https://files.pythonhosted.org/packages/72/fd/7ce24a25def58f564a50270c8e8c0a693086cc08ba6e55fc8aaa78911f4c/Djblets-1.0.7-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "363055a431d8b19460e0cd48659e7380030edcc927cf0d92582b1b192847ce54", + "md5": "5712aba848ec588a02a85b0024102713", + "sha256": "8c9335189bee628e1df62be9cc5a85f6545132e09f8250c6768982ae6637aed7" + }, + "downloads": -1, + "filename": "Djblets-1.0.7.tar.gz", + "has_sig": false, + "md5_digest": "5712aba848ec588a02a85b0024102713", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 953410, + "upload_time": "2018-10-12T07:40:17", + "upload_time_iso_8601": "2018-10-12T07:40:17.609873Z", + "url": "https://files.pythonhosted.org/packages/36/30/55a431d8b19460e0cd48659e7380030edcc927cf0d92582b1b192847ce54/Djblets-1.0.7.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "1.0.8": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "4810ddbf2719b6daaad3c7c887b663f5ece7312122a837fe5e82e643d7a52964", + "md5": "08a4c62c6927f6a0190498f4e3c967a7", + "sha256": "05f3340e2e686e504698a6260a79e09bda3361ced01324565cfcc247c73d262c" + }, + "downloads": -1, + "filename": "Djblets-1.0.8-py2.7.egg", + "has_sig": false, + "md5_digest": "08a4c62c6927f6a0190498f4e3c967a7", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 2005462, + "upload_time": "2018-11-06T11:04:03", + "upload_time_iso_8601": "2018-11-06T11:04:03.913726Z", + "url": "https://files.pythonhosted.org/packages/48/10/ddbf2719b6daaad3c7c887b663f5ece7312122a837fe5e82e643d7a52964/Djblets-1.0.8-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "7ec1b80a6b240b47d24e9aa8eb9091775e9b508bf7312ebabc2c736b49393b97", + "md5": "a6feee9815e4b50a6a24096fab0f5f3b", + "sha256": "1b99df175a3478eb0991952e540f3b3d7efa911f958566103a16105f3439256b" + }, + "downloads": -1, + "filename": "Djblets-1.0.8-py2-none-any.whl", + "has_sig": false, + "md5_digest": "a6feee9815e4b50a6a24096fab0f5f3b", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 1250531, + "upload_time": "2018-11-06T11:02:54", + "upload_time_iso_8601": "2018-11-06T11:02:54.742053Z", + "url": "https://files.pythonhosted.org/packages/7e/c1/b80a6b240b47d24e9aa8eb9091775e9b508bf7312ebabc2c736b49393b97/Djblets-1.0.8-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "4604c6a134efef06b3dfe64bce2c77232d06a8b2eb9e54385778bc683259562a", + "md5": "677ff5c0a6bd5ceebdb9a4104f3055a5", + "sha256": "ea47c2416ccb7df1867d8e6b01a7a309092663e60e1a5c3799068205b46a2876" + }, + "downloads": -1, + "filename": "Djblets-1.0.8.tar.gz", + "has_sig": false, + "md5_digest": "677ff5c0a6bd5ceebdb9a4104f3055a5", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 955978, + "upload_time": "2018-11-06T11:05:03", + "upload_time_iso_8601": "2018-11-06T11:05:03.005040Z", + "url": "https://files.pythonhosted.org/packages/46/04/c6a134efef06b3dfe64bce2c77232d06a8b2eb9e54385778bc683259562a/Djblets-1.0.8.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "1.0.9": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "f849e527389c661d65e939c40213ce47ac412baa3c89f8aadbed2239a63b8ee5", + "md5": "5bfd2d2c03b3556c75a29d7d4520ee62", + "sha256": "ef0a753cb91e763e0888cdfcf12bf77b4e124b9c710685b4d4bec22616905589" + }, + "downloads": -1, + "filename": "Djblets-1.0.9-py2.7.egg", + "has_sig": false, + "md5_digest": "5bfd2d2c03b3556c75a29d7d4520ee62", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 2018356, + "upload_time": "2019-01-29T09:23:46", + "upload_time_iso_8601": "2019-01-29T09:23:46.768031Z", + "url": "https://files.pythonhosted.org/packages/f8/49/e527389c661d65e939c40213ce47ac412baa3c89f8aadbed2239a63b8ee5/Djblets-1.0.9-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "a5e8648b0c15954d3899c7e8dfb6c6b0411a769eb9e9315fc3003498b194d7b7", + "md5": "e411430282fd345b5c8967d8644e88f7", + "sha256": "416633d139f1d5d38f079483dee48b1366ed17d26cac683956ab99f0a2cedf17" + }, + "downloads": -1, + "filename": "Djblets-1.0.9-py2-none-any.whl", + "has_sig": false, + "md5_digest": "e411430282fd345b5c8967d8644e88f7", + "packagetype": "bdist_wheel", + "python_version": "py2", + "requires_python": null, + "size": 1256063, + "upload_time": "2019-01-29T09:23:42", + "upload_time_iso_8601": "2019-01-29T09:23:42.315378Z", + "url": "https://files.pythonhosted.org/packages/a5/e8/648b0c15954d3899c7e8dfb6c6b0411a769eb9e9315fc3003498b194d7b7/Djblets-1.0.9-py2-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "6aa6e2af59b7f4541d65b2db856752ac382b443ecce4fcc950b176ad43028eec", + "md5": "7e4d08d7e3e598a67a8e0a2f1c66483f", + "sha256": "8562fd9b05807afd000cf80b213d70b35a2b6dd01b0ed830fafa6e5c625f2818" + }, + "downloads": -1, + "filename": "Djblets-1.0.9.tar.gz", + "has_sig": false, + "md5_digest": "7e4d08d7e3e598a67a8e0a2f1c66483f", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 961437, + "upload_time": "2019-01-29T09:23:50", + "upload_time_iso_8601": "2019-01-29T09:23:50.140504Z", + "url": "https://files.pythonhosted.org/packages/6a/a6/e2af59b7f4541d65b2db856752ac382b443ecce4fcc950b176ad43028eec/Djblets-1.0.9.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "2.0": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "a20d5ef392ba6714a7dc3d3189283542212ba5584a834579db5bbad708af060f", + "md5": "343cd1e88bc08f32069447f9b7459929", + "sha256": "e957abc65d985175167af9e7251392d71086da570c7e780cbd53916b1309c98c" + }, + "downloads": -1, + "filename": "Djblets-2.0-py2.7.egg", + "has_sig": false, + "md5_digest": "343cd1e88bc08f32069447f9b7459929", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 2436732, + "upload_time": "2020-12-17T04:59:23", + "upload_time_iso_8601": "2020-12-17T04:59:23.111871Z", + "url": "https://files.pythonhosted.org/packages/a2/0d/5ef392ba6714a7dc3d3189283542212ba5584a834579db5bbad708af060f/Djblets-2.0-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "dd69ff8bd280970809a6e2c19bc5f6f6f7c3c65f3fb38045453aebaef44c47c9", + "md5": "31cf070f19d877f1d2243bd71a1c7454", + "sha256": "f5a255f310c17d9240f943893aebe0e653d1980ceda7b7ae3d2743167ef6e0a9" + }, + "downloads": -1, + "filename": "Djblets-2.0-py2.py3-none-any.whl", + "has_sig": false, + "md5_digest": "31cf070f19d877f1d2243bd71a1c7454", + "packagetype": "bdist_wheel", + "python_version": "py2.py3", + "requires_python": null, + "size": 1605633, + "upload_time": "2020-12-17T04:59:20", + "upload_time_iso_8601": "2020-12-17T04:59:20.917003Z", + "url": "https://files.pythonhosted.org/packages/dd/69/ff8bd280970809a6e2c19bc5f6f6f7c3c65f3fb38045453aebaef44c47c9/Djblets-2.0-py2.py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "7ab53c344b8ab757e63cbe93020800c7fddcedb27035b17ef2abe95a7cf4f8d9", + "md5": "f632e40ecc872d4d0a6699570fa77fab", + "sha256": "6ecb325581fc744a663a587512b5b6fdb959fb2d5bda4527c625a442f4a92ef1" + }, + "downloads": -1, + "filename": "Djblets-2.0.tar.gz", + "has_sig": false, + "md5_digest": "f632e40ecc872d4d0a6699570fa77fab", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 1237578, + "upload_time": "2020-12-17T04:59:25", + "upload_time_iso_8601": "2020-12-17T04:59:25.614106Z", + "url": "https://files.pythonhosted.org/packages/7a/b5/3c344b8ab757e63cbe93020800c7fddcedb27035b17ef2abe95a7cf4f8d9/Djblets-2.0.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "2.1": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "e737697b6888ac5a56e581299543629544ca2ee47304fb21dd5d1af5a1aa10f5", + "md5": "91a2cdb217ee77171f00864323c36da0", + "sha256": "855ca0d0a9b8f80d3ec096e95afee78e0f197bb017d49d3f34f1a5d641d3580b" + }, + "downloads": -1, + "filename": "Djblets-2.1-py2.7.egg", + "has_sig": false, + "md5_digest": "91a2cdb217ee77171f00864323c36da0", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 2441535, + "upload_time": "2021-01-20T09:53:52", + "upload_time_iso_8601": "2021-01-20T09:53:52.654994Z", + "url": "https://files.pythonhosted.org/packages/e7/37/697b6888ac5a56e581299543629544ca2ee47304fb21dd5d1af5a1aa10f5/Djblets-2.1-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "e32f4ce2d57c502fa946b94eea94ec1fd9a8f551d31f666582be2c4d369675f3", + "md5": "ae2d2c1e69e78c115e3853c6e0bb36e8", + "sha256": "176cb6b9866e44fe915ad1209c1d230bd86c3ea19bbde775d32dd4730f606408" + }, + "downloads": -1, + "filename": "Djblets-2.1-py2.py3-none-any.whl", + "has_sig": false, + "md5_digest": "ae2d2c1e69e78c115e3853c6e0bb36e8", + "packagetype": "bdist_wheel", + "python_version": "py2.py3", + "requires_python": null, + "size": 1608317, + "upload_time": "2021-01-20T09:53:50", + "upload_time_iso_8601": "2021-01-20T09:53:50.314271Z", + "url": "https://files.pythonhosted.org/packages/e3/2f/4ce2d57c502fa946b94eea94ec1fd9a8f551d31f666582be2c4d369675f3/Djblets-2.1-py2.py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "7744ae1213e5815da56be6cde6740e6c5019d326da63afefc8ed05bc26b37a82", + "md5": "af4fc65a5f2728c50ce16128470c0cf0", + "sha256": "a3940705bb31aaff5c963fd6431d0d5e80379e923fcb223613c16a47fa33d5a7" + }, + "downloads": -1, + "filename": "Djblets-2.1.tar.gz", + "has_sig": false, + "md5_digest": "af4fc65a5f2728c50ce16128470c0cf0", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 1240156, + "upload_time": "2021-01-20T09:53:55", + "upload_time_iso_8601": "2021-01-20T09:53:55.441808Z", + "url": "https://files.pythonhosted.org/packages/77/44/ae1213e5815da56be6cde6740e6c5019d326da63afefc8ed05bc26b37a82/Djblets-2.1.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "2.1.1": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "8bbeb2e3e641b1291d8ba740cc5bd0d08cb6ea95ef4115cbc96b89d5228e05a3", + "md5": "0196246390e267d40d61143eb55dd951", + "sha256": "af6006841a3387b045e1ed9958f2e20fb60995abffac0d8e1c1628f6de981589" + }, + "downloads": -1, + "filename": "Djblets-2.1.1-py2.7.egg", + "has_sig": false, + "md5_digest": "0196246390e267d40d61143eb55dd951", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 2441543, + "upload_time": "2021-01-20T10:50:01", + "upload_time_iso_8601": "2021-01-20T10:50:01.830717Z", + "url": "https://files.pythonhosted.org/packages/8b/be/b2e3e641b1291d8ba740cc5bd0d08cb6ea95ef4115cbc96b89d5228e05a3/Djblets-2.1.1-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "43a3165331629345c7872bc2a6de08eb3ec5333283215651af50a7f41b973171", + "md5": "b8f833d09768512b63ea897869f75161", + "sha256": "687be6fbd698ed8a743cfb48dd312adb999ef7ec98b037fa07949d863ca5588a" + }, + "downloads": -1, + "filename": "Djblets-2.1.1-py2.py3-none-any.whl", + "has_sig": false, + "md5_digest": "b8f833d09768512b63ea897869f75161", + "packagetype": "bdist_wheel", + "python_version": "py2.py3", + "requires_python": null, + "size": 1608333, + "upload_time": "2021-01-20T10:49:59", + "upload_time_iso_8601": "2021-01-20T10:49:59.528455Z", + "url": "https://files.pythonhosted.org/packages/43/a3/165331629345c7872bc2a6de08eb3ec5333283215651af50a7f41b973171/Djblets-2.1.1-py2.py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "233de1c95d3eed1381a61aa917efed5385938578fcacd78f8c2ff8298367af09", + "md5": "e413b6f511e3461770a89ea3637c881d", + "sha256": "9249ba93bdc653dd3badd19d25b413fafb93725f040daa313b38243d692bee4f" + }, + "downloads": -1, + "filename": "Djblets-2.1.1.tar.gz", + "has_sig": false, + "md5_digest": "e413b6f511e3461770a89ea3637c881d", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 1240732, + "upload_time": "2021-01-20T10:50:04", + "upload_time_iso_8601": "2021-01-20T10:50:04.217950Z", + "url": "https://files.pythonhosted.org/packages/23/3d/e1c95d3eed1381a61aa917efed5385938578fcacd78f8c2ff8298367af09/Djblets-2.1.1.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "2.2": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "4bcb3b94778840eb1417341b5b35cae64d739ef239d0e92f3949aa58f7823ee8", + "md5": "78636980fa7256328a6b952a85dd4bbb", + "sha256": "b9f5a9933ab481a719853b98caeba0bab38fe920774c8d1c74af1fce4c1f8427" + }, + "downloads": -1, + "filename": "Djblets-2.2-py2.7.egg", + "has_sig": false, + "md5_digest": "78636980fa7256328a6b952a85dd4bbb", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 2449322, + "upload_time": "2021-03-02T09:27:25", + "upload_time_iso_8601": "2021-03-02T09:27:25.478715Z", + "url": "https://files.pythonhosted.org/packages/4b/cb/3b94778840eb1417341b5b35cae64d739ef239d0e92f3949aa58f7823ee8/Djblets-2.2-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "7d527345f3246618f9374831b3a517f4be41114f1a3978938662d12f5896720d", + "md5": "042e77c0f1a6456a39f65435a53031ab", + "sha256": "addb56eef40f2dff71dab595b824ac079e9a0b7f9514a8b25118cef441a8b787" + }, + "downloads": -1, + "filename": "Djblets-2.2-py2.py3-none-any.whl", + "has_sig": false, + "md5_digest": "042e77c0f1a6456a39f65435a53031ab", + "packagetype": "bdist_wheel", + "python_version": "py2.py3", + "requires_python": null, + "size": 1611619, + "upload_time": "2021-03-02T09:27:23", + "upload_time_iso_8601": "2021-03-02T09:27:23.363183Z", + "url": "https://files.pythonhosted.org/packages/7d/52/7345f3246618f9374831b3a517f4be41114f1a3978938662d12f5896720d/Djblets-2.2-py2.py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "34cfc7c74d263317675078ddc3403c6e3ede6aa8d5aeb0a0199516ea970f8caa", + "md5": "51aa85da074f3f6a24a96103caed66ba", + "sha256": "6880380322a54605285abe4559cab185f021f55ae9d84149091a0dc697754b13" + }, + "downloads": -1, + "filename": "Djblets-2.2.tar.gz", + "has_sig": false, + "md5_digest": "51aa85da074f3f6a24a96103caed66ba", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 1243824, + "upload_time": "2021-03-02T09:27:28", + "upload_time_iso_8601": "2021-03-02T09:27:28.223673Z", + "url": "https://files.pythonhosted.org/packages/34/cf/c7c74d263317675078ddc3403c6e3ede6aa8d5aeb0a0199516ea970f8caa/Djblets-2.2.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "2.2.1": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "0cb0342ed606dd27fc01fa5bdcbc587019205ef88f309fce4983b769fef8af0c", + "md5": "bf76fa93d6eda68b392473015c688668", + "sha256": "a1bd9247700e893e4f25e9c17890547c5ebc682fc1f567fa982b44e83ccfab1c" + }, + "downloads": -1, + "filename": "Djblets-2.2.1-py2.7.egg", + "has_sig": false, + "md5_digest": "bf76fa93d6eda68b392473015c688668", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 2450291, + "upload_time": "2021-05-20T22:40:44", + "upload_time_iso_8601": "2021-05-20T22:40:44.557865Z", + "url": "https://files.pythonhosted.org/packages/0c/b0/342ed606dd27fc01fa5bdcbc587019205ef88f309fce4983b769fef8af0c/Djblets-2.2.1-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "9998c3bd5c4e813e7d14f892e884893e6fe5561392acb0ebdc127a98bec71ed5", + "md5": "c7e0c259c4dcffc68bd43d0e7fb91673", + "sha256": "b2a328424bef93d49bf02c9e86c559fdc9e7e3e128f30d2f03ff74c5751dd01d" + }, + "downloads": -1, + "filename": "Djblets-2.2.1-py2.py3-none-any.whl", + "has_sig": false, + "md5_digest": "c7e0c259c4dcffc68bd43d0e7fb91673", + "packagetype": "bdist_wheel", + "python_version": "py2.py3", + "requires_python": null, + "size": 1611859, + "upload_time": "2021-05-20T22:40:39", + "upload_time_iso_8601": "2021-05-20T22:40:39.783157Z", + "url": "https://files.pythonhosted.org/packages/99/98/c3bd5c4e813e7d14f892e884893e6fe5561392acb0ebdc127a98bec71ed5/Djblets-2.2.1-py2.py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "4f2653450e03ba4c1c24b59703f6ac2b8a114a22604e49c7c5194b85518035b5", + "md5": "8694a30da6415dca7ca3a95c710efa36", + "sha256": "6ae945c390a69d0320a1a9e4f3ed316148bbf4a67638b4c76a93475975e3149b" + }, + "downloads": -1, + "filename": "Djblets-2.2.1.tar.gz", + "has_sig": false, + "md5_digest": "8694a30da6415dca7ca3a95c710efa36", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 1244834, + "upload_time": "2021-05-20T22:40:46", + "upload_time_iso_8601": "2021-05-20T22:40:46.954786Z", + "url": "https://files.pythonhosted.org/packages/4f/26/53450e03ba4c1c24b59703f6ac2b8a114a22604e49c7c5194b85518035b5/Djblets-2.2.1.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "2.2.2": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "db362185b70cd5a1bfa7b88f4dc67b809485765a9fdfc6c609cf49a2bef4a5ef", + "md5": "5663fc141675be54cb5f6cad1fa58a22", + "sha256": "c44bb5620d7c790dbb34178d1f7a7e26b781b069e4e296741f329757ce109941" + }, + "downloads": -1, + "filename": "Djblets-2.2.2-py2.7.egg", + "has_sig": false, + "md5_digest": "5663fc141675be54cb5f6cad1fa58a22", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 2450724, + "upload_time": "2021-06-15T07:22:27", + "upload_time_iso_8601": "2021-06-15T07:22:27.256915Z", + "url": "https://files.pythonhosted.org/packages/db/36/2185b70cd5a1bfa7b88f4dc67b809485765a9fdfc6c609cf49a2bef4a5ef/Djblets-2.2.2-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "d6ffcc5466ca6357a88bd06e74553d36f2c7ca0e8ed900fcebf0318083efadba", + "md5": "9d8456b70558a2b7be51127645f8e92d", + "sha256": "6f43735fdc9b8b411830f625168291f864e145c9a665d9d73f5f0b7237aece0b" + }, + "downloads": -1, + "filename": "Djblets-2.2.2-py2.py3-none-any.whl", + "has_sig": false, + "md5_digest": "9d8456b70558a2b7be51127645f8e92d", + "packagetype": "bdist_wheel", + "python_version": "py2.py3", + "requires_python": null, + "size": 1612060, + "upload_time": "2021-06-15T07:22:24", + "upload_time_iso_8601": "2021-06-15T07:22:24.865091Z", + "url": "https://files.pythonhosted.org/packages/d6/ff/cc5466ca6357a88bd06e74553d36f2c7ca0e8ed900fcebf0318083efadba/Djblets-2.2.2-py2.py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "2bfe7aceb4c660f7e18bb4ea2cb6fc9ecb9d08a87a3056bffc96e3e877f32e3e", + "md5": "56404d07e5fcf7b4b30da86dcfd4aa9d", + "sha256": "c78950a7478104826d22cff7d2a27945ea3bc4f9225e5fd6aa9fc2cee031a7cc" + }, + "downloads": -1, + "filename": "Djblets-2.2.2.tar.gz", + "has_sig": false, + "md5_digest": "56404d07e5fcf7b4b30da86dcfd4aa9d", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 1245418, + "upload_time": "2021-06-15T07:22:30", + "upload_time_iso_8601": "2021-06-15T07:22:30.602407Z", + "url": "https://files.pythonhosted.org/packages/2b/fe/7aceb4c660f7e18bb4ea2cb6fc9ecb9d08a87a3056bffc96e3e877f32e3e/Djblets-2.2.2.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "2.2.3": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "40b7227f31f851267a02e4d123bf5bdcc91a8403ed8ad014c3fd4f36dbf30569", + "md5": "9579fa049ae21579212ac43d8a6d75af", + "sha256": "d00268b137eb4ff2641baf7fb4c5ab36f2959223e0338a767b46323addd9b27e" + }, + "downloads": -1, + "filename": "Djblets-2.2.3-py2.7.egg", + "has_sig": false, + "md5_digest": "9579fa049ae21579212ac43d8a6d75af", + "packagetype": "bdist_egg", + "python_version": "2.7", + "requires_python": null, + "size": 2454322, + "upload_time": "2021-06-29T06:39:47", + "upload_time_iso_8601": "2021-06-29T06:39:47.267412Z", + "url": "https://files.pythonhosted.org/packages/40/b7/227f31f851267a02e4d123bf5bdcc91a8403ed8ad014c3fd4f36dbf30569/Djblets-2.2.3-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "0cbd8b4fb26f42e79fcd23d29c9832a61bd3ad4c3d84ff47c9ebe84c092e94c4", + "md5": "7400c14ee7b9e24cd41024e329cea421", + "sha256": "ee07f9a86bc2cc92835fa564a7b5c28b4a9a5176d5b1982322994486a49aad6e" + }, + "downloads": -1, + "filename": "Djblets-2.2.3-py2.py3-none-any.whl", + "has_sig": false, + "md5_digest": "7400c14ee7b9e24cd41024e329cea421", + "packagetype": "bdist_wheel", + "python_version": "py2.py3", + "requires_python": null, + "size": 1613384, + "upload_time": "2021-06-29T06:39:44", + "upload_time_iso_8601": "2021-06-29T06:39:44.801504Z", + "url": "https://files.pythonhosted.org/packages/0c/bd/8b4fb26f42e79fcd23d29c9832a61bd3ad4c3d84ff47c9ebe84c092e94c4/Djblets-2.2.3-py2.py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "f2fa2827d0507bad7886053bd0a23799d67472adcd4986222ff0ff8f26e08e34", + "md5": "dcef8cd6d13bde8c6ba0e7c0f4788da2", + "sha256": "9b43ceb8a0b2400f58d8cdbbe5b719d5569bec71e4e84d94de59d8f333f61d43" + }, + "downloads": -1, + "filename": "Djblets-2.2.3.tar.gz", + "has_sig": false, + "md5_digest": "dcef8cd6d13bde8c6ba0e7c0f4788da2", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 1247372, + "upload_time": "2021-06-29T06:39:50", + "upload_time_iso_8601": "2021-06-29T06:39:50.246408Z", + "url": "https://files.pythonhosted.org/packages/f2/fa/2827d0507bad7886053bd0a23799d67472adcd4986222ff0ff8f26e08e34/Djblets-2.2.3.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "2.3": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "f00512273a882b83a1bb509e8008e9edc3218814eeb90691743c3fa828e4f87d", + "md5": "738ea2596d32ad78562e44d961b7e2b4", + "sha256": "9da363a84cf67afafb9ca03ad8d94bf0e3289e1738c1df88bfe524e21b7e885b" + }, + "downloads": -1, + "filename": "Djblets-2.3-py2.7.egg", + "has_sig": false, + "md5_digest": "738ea2596d32ad78562e44d961b7e2b4", + "packagetype": "bdist_egg", + "python_version": "2.3", + "requires_python": null, + "size": 2474680, + "upload_time": "2022-02-08T11:23:52", + "upload_time_iso_8601": "2022-02-08T11:23:52.873877Z", + "url": "https://files.pythonhosted.org/packages/f0/05/12273a882b83a1bb509e8008e9edc3218814eeb90691743c3fa828e4f87d/Djblets-2.3-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "193fa96d10cc831a4dcd1939f3e3cb9e2f6f5eba7ab1030dcad19846ae9ecb4e", + "md5": "fbc2be802a74d26bde04fc59307226ab", + "sha256": "ea1304f2ec8235ee3775b0a14954d81177a2b89f19280574e7ed585373830b01" + }, + "downloads": -1, + "filename": "Djblets-2.3-py2.py3-none-any.whl", + "has_sig": false, + "md5_digest": "fbc2be802a74d26bde04fc59307226ab", + "packagetype": "bdist_wheel", + "python_version": "py2.py3", + "requires_python": null, + "size": 1625989, + "upload_time": "2022-02-08T11:23:50", + "upload_time_iso_8601": "2022-02-08T11:23:50.321056Z", + "url": "https://files.pythonhosted.org/packages/19/3f/a96d10cc831a4dcd1939f3e3cb9e2f6f5eba7ab1030dcad19846ae9ecb4e/Djblets-2.3-py2.py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "f3f44ae500569bbbadd60b513d4ac949d37a54be22f94f0726a602da904ec1e0", + "md5": "f520508fa4c387a3212ac63e0609e9db", + "sha256": "b959ff576eb29fb8c3e6b156cf4af2130075b5e1b050382a9edfe973e9494f74" + }, + "downloads": -1, + "filename": "Djblets-2.3.tar.gz", + "has_sig": false, + "md5_digest": "f520508fa4c387a3212ac63e0609e9db", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 1261108, + "upload_time": "2022-02-08T11:23:56", + "upload_time_iso_8601": "2022-02-08T11:23:56.597596Z", + "url": "https://files.pythonhosted.org/packages/f3/f4/4ae500569bbbadd60b513d4ac949d37a54be22f94f0726a602da904ec1e0/Djblets-2.3.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "2.3.1": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "48cd554730e0835ac6facfd76bad224865c78334accd7f3cd29d539fcb43713b", + "md5": "fc848adec99b1819fe49efdaf1a4effd", + "sha256": "83288429bde921314fec702f357fd5af16fbb46ac3f378e5dd818162f245c822" + }, + "downloads": -1, + "filename": "Djblets-2.3.1-py2.7.egg", + "has_sig": false, + "md5_digest": "fc848adec99b1819fe49efdaf1a4effd", + "packagetype": "bdist_egg", + "python_version": "2.3.1", + "requires_python": null, + "size": 2485251, + "upload_time": "2022-06-14T06:32:53", + "upload_time_iso_8601": "2022-06-14T06:32:53.959739Z", + "url": "https://files.pythonhosted.org/packages/48/cd/554730e0835ac6facfd76bad224865c78334accd7f3cd29d539fcb43713b/Djblets-2.3.1-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "505697a383ba6e7db19a53e26573e4bd78acba5eda0766fdf469265ce64d3cb2", + "md5": "868f42a7319e04d29b0c30f4f05b2725", + "sha256": "b82baec2a8491ea33f45a6915b331194abf6dd273e880b47c3f09554d92326bd" + }, + "downloads": -1, + "filename": "Djblets-2.3.1-py2.py3-none-any.whl", + "has_sig": false, + "md5_digest": "868f42a7319e04d29b0c30f4f05b2725", + "packagetype": "bdist_wheel", + "python_version": "py2.py3", + "requires_python": null, + "size": 1635980, + "upload_time": "2022-06-14T06:32:50", + "upload_time_iso_8601": "2022-06-14T06:32:50.684135Z", + "url": "https://files.pythonhosted.org/packages/50/56/97a383ba6e7db19a53e26573e4bd78acba5eda0766fdf469265ce64d3cb2/Djblets-2.3.1-py2.py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "b657678506bc7fa3571233447030e7f5caea69a0b7cf10b5e99cbb3f1a680c2e", + "md5": "637792b30e7c5cc9a47e2255e884af7d", + "sha256": "32b1bd708c0d0a92a4e74ad27e443635bb5977351ff5ed93c7475bbf9b49a9ba" + }, + "downloads": -1, + "filename": "Djblets-2.3.1.tar.gz", + "has_sig": false, + "md5_digest": "637792b30e7c5cc9a47e2255e884af7d", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 1241531, + "upload_time": "2022-06-14T06:32:56", + "upload_time_iso_8601": "2022-06-14T06:32:56.736553Z", + "url": "https://files.pythonhosted.org/packages/b6/57/678506bc7fa3571233447030e7f5caea69a0b7cf10b5e99cbb3f1a680c2e/Djblets-2.3.1.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "2.3.2": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "d88b5a4e1795b8dfc3999984a0d26e0a17f00129119f96be42c7ff82ffc66377", + "md5": "38819b7c086c668fb02073691b6a42ba", + "sha256": "b925551e2fe785d069b94dabced2b443140d5e4314e828373ff93c5b7d166563" + }, + "downloads": -1, + "filename": "Djblets-2.3.2-py2.7.egg", + "has_sig": false, + "md5_digest": "38819b7c086c668fb02073691b6a42ba", + "packagetype": "bdist_egg", + "python_version": "2.3.2", + "requires_python": null, + "size": 2476942, + "upload_time": "2022-07-27T04:51:42", + "upload_time_iso_8601": "2022-07-27T04:51:42.316963Z", + "url": "https://files.pythonhosted.org/packages/d8/8b/5a4e1795b8dfc3999984a0d26e0a17f00129119f96be42c7ff82ffc66377/Djblets-2.3.2-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "b4558e8d1b5c79f5bf5a65eebf049b75ab8ee5b25611521415432ee62dc9cb1e", + "md5": "e1ff5f9c8f30f26fdc828a7040455859", + "sha256": "6ea2f8524cffd91440477de4fd1fa13538e2e8befed853e1376e4d7741865b83" + }, + "downloads": -1, + "filename": "Djblets-2.3.2-py2.py3-none-any.whl", + "has_sig": false, + "md5_digest": "e1ff5f9c8f30f26fdc828a7040455859", + "packagetype": "bdist_wheel", + "python_version": "py2.py3", + "requires_python": null, + "size": 1627014, + "upload_time": "2022-07-27T04:51:39", + "upload_time_iso_8601": "2022-07-27T04:51:39.956126Z", + "url": "https://files.pythonhosted.org/packages/b4/55/8e8d1b5c79f5bf5a65eebf049b75ab8ee5b25611521415432ee62dc9cb1e/Djblets-2.3.2-py2.py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "85d3e2ef0f44cab73134830f83d5d03318046c1e3dd13ac12c3b0cac6be2178b", + "md5": "e98e4c2f2cc42c749f22aaedcdd19c1c", + "sha256": "437dd392dab399d4035277283aab65e5fc8df3e9fe04d7238d641e14262e3747" + }, + "downloads": -1, + "filename": "Djblets-2.3.2.tar.gz", + "has_sig": false, + "md5_digest": "e98e4c2f2cc42c749f22aaedcdd19c1c", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 1233628, + "upload_time": "2022-07-27T04:51:45", + "upload_time_iso_8601": "2022-07-27T04:51:45.416890Z", + "url": "https://files.pythonhosted.org/packages/85/d3/e2ef0f44cab73134830f83d5d03318046c1e3dd13ac12c3b0cac6be2178b/Djblets-2.3.2.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "2.3.3": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "5d5f87e4d0665fefbd3623e0c33410bc9992281de54649a62364c705f0a1261f", + "md5": "413bf7511afde1a616d2c4bbe0ffee15", + "sha256": "545171664b4d1dcec800af31a29aa2b00fca408a38248477da25c2a02a77987b" + }, + "downloads": -1, + "filename": "Djblets-2.3.3-py2.7.egg", + "has_sig": false, + "md5_digest": "413bf7511afde1a616d2c4bbe0ffee15", + "packagetype": "bdist_egg", + "python_version": "2.3.3", + "requires_python": null, + "size": 2477334, + "upload_time": "2022-07-27T05:09:01", + "upload_time_iso_8601": "2022-07-27T05:09:01.233278Z", + "url": "https://files.pythonhosted.org/packages/5d/5f/87e4d0665fefbd3623e0c33410bc9992281de54649a62364c705f0a1261f/Djblets-2.3.3-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "80d805239236971e5d5ac8b5cd0cd31b3f10028102730422721c8510bf05254c", + "md5": "1999f942686bd936e437667a9795dde6", + "sha256": "27553f9637b0417ecbe708abe772b9ba81c3d8296c4b91d41553ce6e3abe36e2" + }, + "downloads": -1, + "filename": "Djblets-2.3.3-py2.py3-none-any.whl", + "has_sig": false, + "md5_digest": "1999f942686bd936e437667a9795dde6", + "packagetype": "bdist_wheel", + "python_version": "py2.py3", + "requires_python": null, + "size": 1627080, + "upload_time": "2022-07-27T05:08:58", + "upload_time_iso_8601": "2022-07-27T05:08:58.441969Z", + "url": "https://files.pythonhosted.org/packages/80/d8/05239236971e5d5ac8b5cd0cd31b3f10028102730422721c8510bf05254c/Djblets-2.3.3-py2.py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "35e896be319e59300bb9fc6f591f0d33dc60ec3ce4e3b3bfd80e6c454b362931", + "md5": "5a0ee56bdbc19d779121c22bfc2cc649", + "sha256": "bcf99282178d348b6f38750fdb989d4962a64a01424d23e599077ca2a951c389" + }, + "downloads": -1, + "filename": "Djblets-2.3.3.tar.gz", + "has_sig": false, + "md5_digest": "5a0ee56bdbc19d779121c22bfc2cc649", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 1233834, + "upload_time": "2022-07-27T05:09:03", + "upload_time_iso_8601": "2022-07-27T05:09:03.986835Z", + "url": "https://files.pythonhosted.org/packages/35/e8/96be319e59300bb9fc6f591f0d33dc60ec3ce4e3b3bfd80e6c454b362931/Djblets-2.3.3.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "2.3.4": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "dd820862d2178603efb44b29c23b59177e854f780f413cba3e3456fc055b0291", + "md5": "c691479606e8ab16a46e0d339fb579d6", + "sha256": "8b71416202114cc3ca2ec898494b543302e8947ce72afe2fe23bcb9f5e7b3b4e" + }, + "downloads": -1, + "filename": "Djblets-2.3.4-py2.7.egg", + "has_sig": false, + "md5_digest": "c691479606e8ab16a46e0d339fb579d6", + "packagetype": "bdist_egg", + "python_version": "2.3.4", + "requires_python": null, + "size": 2477885, + "upload_time": "2022-09-29T07:10:47", + "upload_time_iso_8601": "2022-09-29T07:10:47.024449Z", + "url": "https://files.pythonhosted.org/packages/dd/82/0862d2178603efb44b29c23b59177e854f780f413cba3e3456fc055b0291/Djblets-2.3.4-py2.7.egg", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "d975ccce839d72feaf373550482bfb75df384b2e45f2aa03f0d14e9c7e3a9a08", + "md5": "ad99462114b90c289dd0df1982c27f6a", + "sha256": "26f6f5f4603765281c282b842b1ee4038ef383002a3e23eaec551a89b76fbe31" + }, + "downloads": -1, + "filename": "Djblets-2.3.4-py2.py3-none-any.whl", + "has_sig": false, + "md5_digest": "ad99462114b90c289dd0df1982c27f6a", + "packagetype": "bdist_wheel", + "python_version": "py2.py3", + "requires_python": null, + "size": 1627630, + "upload_time": "2022-09-29T07:10:43", + "upload_time_iso_8601": "2022-09-29T07:10:43.266446Z", + "url": "https://files.pythonhosted.org/packages/d9/75/ccce839d72feaf373550482bfb75df384b2e45f2aa03f0d14e9c7e3a9a08/Djblets-2.3.4-py2.py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "cc24c7e1f16b3e51ae8860b5b2d204a3d52fc9c2352adbd2c1078dbc6e6fa4e7", + "md5": "95631e8091868874d5736759d9dc8a8f", + "sha256": "8a581050502f9d38d4c1f070b726241cce706212d36932aa39f12d7e8b2e2363" + }, + "downloads": -1, + "filename": "Djblets-2.3.4.tar.gz", + "has_sig": false, + "md5_digest": "95631e8091868874d5736759d9dc8a8f", + "packagetype": "sdist", + "python_version": "source", + "requires_python": null, + "size": 1235224, + "upload_time": "2022-09-29T07:10:49", + "upload_time_iso_8601": "2022-09-29T07:10:49.861732Z", + "url": "https://files.pythonhosted.org/packages/cc/24/c7e1f16b3e51ae8860b5b2d204a3d52fc9c2352adbd2c1078dbc6e6fa4e7/Djblets-2.3.4.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "3.0": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "215870c1310e258bfd5e4e316da5c8f354a139509b2a8139d7c97122fe2ff302", + "md5": "ebd7597fc1981348e5a75aac029d2d3e", + "sha256": "58f19fc92bf2c1ed4f9025355f86f5e2b93f23e451c4bda2d7690c9c1a4c574b" + }, + "downloads": -1, + "filename": "Djblets-3.0-py3-none-any.whl", + "has_sig": false, + "md5_digest": "ebd7597fc1981348e5a75aac029d2d3e", + "packagetype": "bdist_wheel", + "python_version": "py3", + "requires_python": ">=3.7", + "size": 1602400, + "upload_time": "2022-10-04T21:44:08", + "upload_time_iso_8601": "2022-10-04T21:44:08.239460Z", + "url": "https://files.pythonhosted.org/packages/21/58/70c1310e258bfd5e4e316da5c8f354a139509b2a8139d7c97122fe2ff302/Djblets-3.0-py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "0a117c82cab84f9590c73b7c749718078f8c2a2699da38fee666179a4a7817f5", + "md5": "b5d846e4611c6a8a901584db189cf38f", + "sha256": "3a3b049468efe508348f8cb9f098c3158b423cc3c3f5f3c787aa0928b98939eb" + }, + "downloads": -1, + "filename": "Djblets-3.0.tar.gz", + "has_sig": false, + "md5_digest": "b5d846e4611c6a8a901584db189cf38f", + "packagetype": "sdist", + "python_version": "source", + "requires_python": ">=3.7", + "size": 1237005, + "upload_time": "2022-10-04T21:44:10", + "upload_time_iso_8601": "2022-10-04T21:44:10.883330Z", + "url": "https://files.pythonhosted.org/packages/0a/11/7c82cab84f9590c73b7c749718078f8c2a2699da38fee666179a4a7817f5/Djblets-3.0.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "3.1": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "d7f38648c9a31fd887d7c331262b274552dd3942913c1bb488215fc723c7b8cc", + "md5": "9abce2b44ffd8a4ee6d460d1d1f90c50", + "sha256": "bccc05c541f7c42207c3d10a414479defa3d51e7175d42a379f0f9e587b1941a" + }, + "downloads": -1, + "filename": "Djblets-3.1-py3-none-any.whl", + "has_sig": false, + "md5_digest": "9abce2b44ffd8a4ee6d460d1d1f90c50", + "packagetype": "bdist_wheel", + "python_version": "py3", + "requires_python": ">=3.7", + "size": 1601523, + "upload_time": "2022-11-02T06:36:58", + "upload_time_iso_8601": "2022-11-02T06:36:58.201636Z", + "url": "https://files.pythonhosted.org/packages/d7/f3/8648c9a31fd887d7c331262b274552dd3942913c1bb488215fc723c7b8cc/Djblets-3.1-py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "efe20e89da5e09c6b5b02142b194c7766dcb6a079f85edf5d940f97c787f8fa1", + "md5": "602d285d0b420888d486f427680be1b7", + "sha256": "996793e36fd6f64ef699e9c32d133def9b86ed2d14f0109758e764d2b624d272" + }, + "downloads": -1, + "filename": "Djblets-3.1.tar.gz", + "has_sig": false, + "md5_digest": "602d285d0b420888d486f427680be1b7", + "packagetype": "sdist", + "python_version": "source", + "requires_python": ">=3.7", + "size": 1235429, + "upload_time": "2022-11-02T06:37:00", + "upload_time_iso_8601": "2022-11-02T06:37:00.468916Z", + "url": "https://files.pythonhosted.org/packages/ef/e2/0e89da5e09c6b5b02142b194c7766dcb6a079f85edf5d940f97c787f8fa1/Djblets-3.1.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "3.2": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "bbd05363f3f9c00f5bb87419d196d485eedd683c9664f73f207c1b31482d5720", + "md5": "aa6a7dd300b03b5cdeaf233e9bcf59ed", + "sha256": "824249ccbc9e7050da074d8797ff760213bec88d200234897f857ba15d8f6811" + }, + "downloads": -1, + "filename": "Djblets-3.2-py3-none-any.whl", + "has_sig": false, + "md5_digest": "aa6a7dd300b03b5cdeaf233e9bcf59ed", + "packagetype": "bdist_wheel", + "python_version": "py3", + "requires_python": ">=3.7", + "size": 1613410, + "upload_time": "2023-02-07T07:14:53", + "upload_time_iso_8601": "2023-02-07T07:14:53.608941Z", + "url": "https://files.pythonhosted.org/packages/bb/d0/5363f3f9c00f5bb87419d196d485eedd683c9664f73f207c1b31482d5720/Djblets-3.2-py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "2ef9d1fcc842dfaa4f06c912fa12d08ea4a37f8c6cbd2caf9fe7948c9a12a83d", + "md5": "ae6270c21ddb21e314e5977bfe88561f", + "sha256": "58233cfa7eae9f964ac6bdf365601c8f21db14da504ff332f3693d124b6b46d6" + }, + "downloads": -1, + "filename": "Djblets-3.2.tar.gz", + "has_sig": false, + "md5_digest": "ae6270c21ddb21e314e5977bfe88561f", + "packagetype": "sdist", + "python_version": "source", + "requires_python": ">=3.7", + "size": 1246910, + "upload_time": "2023-02-07T07:14:55", + "upload_time_iso_8601": "2023-02-07T07:14:55.780452Z", + "url": "https://files.pythonhosted.org/packages/2e/f9/d1fcc842dfaa4f06c912fa12d08ea4a37f8c6cbd2caf9fe7948c9a12a83d/Djblets-3.2.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "3.3": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "68356a480f4e959b446048215055dc8c9a2aecbdb0bbd5122f528f508f5b3153", + "md5": "e60c029ea6a7fa5aaeeaa137fc20da84", + "sha256": "6cf3f88ca11dc5f2843c3aea4bc34f4802650e22315277f5aebef7a0fc760fb4" + }, + "downloads": -1, + "filename": "Djblets-3.3-py3-none-any.whl", + "has_sig": false, + "md5_digest": "e60c029ea6a7fa5aaeeaa137fc20da84", + "packagetype": "bdist_wheel", + "python_version": "py3", + "requires_python": ">=3.7", + "size": 1622305, + "upload_time": "2023-06-20T05:03:31", + "upload_time_iso_8601": "2023-06-20T05:03:31.083050Z", + "url": "https://files.pythonhosted.org/packages/68/35/6a480f4e959b446048215055dc8c9a2aecbdb0bbd5122f528f508f5b3153/Djblets-3.3-py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "e91e5ba7fa7f19099bab22618e9633735ab92f67fb89cdfeafbaf9ee65a2ce9a", + "md5": "49ce0f238bb98398eca43716aba105a3", + "sha256": "f9f066eddd70339001024e3508890933ec249f5fe68779f6add81298a6f0db2c" + }, + "downloads": -1, + "filename": "Djblets-3.3.tar.gz", + "has_sig": false, + "md5_digest": "49ce0f238bb98398eca43716aba105a3", + "packagetype": "sdist", + "python_version": "source", + "requires_python": ">=3.7", + "size": 1255405, + "upload_time": "2023-06-20T05:03:33", + "upload_time_iso_8601": "2023-06-20T05:03:33.610935Z", + "url": "https://files.pythonhosted.org/packages/e9/1e/5ba7fa7f19099bab22618e9633735ab92f67fb89cdfeafbaf9ee65a2ce9a/Djblets-3.3.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "4.0": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "41afefd3f5497b6bcfc218a14c476191ca816a4baf0683153595bffd7f1f53b6", + "md5": "3784af51ef66cc47d19ec26a487ba7d8", + "sha256": "cc000fe68f6217c032ca3e5c4bf9a693fb91155f28ac7077ae39ff39a919ce57" + }, + "downloads": -1, + "filename": "Djblets-4.0-py3-none-any.whl", + "has_sig": false, + "md5_digest": "3784af51ef66cc47d19ec26a487ba7d8", + "packagetype": "bdist_wheel", + "python_version": "py3", + "requires_python": ">=3.8", + "size": 1671724, + "upload_time": "2023-10-17T07:55:35", + "upload_time_iso_8601": "2023-10-17T07:55:35.396181Z", + "url": "https://files.pythonhosted.org/packages/41/af/efd3f5497b6bcfc218a14c476191ca816a4baf0683153595bffd7f1f53b6/Djblets-4.0-py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "5ef93571a6140a5dce9eaf7d3dfbceaf8c88ad6a0a407ec5e63a6591f5dcbb3e", + "md5": "2df8f8ec08235d9aa08464f826d871ca", + "sha256": "c4bed4c4c91dd07a2f185157a8431bfc4e76c473f1bd5314be430891c6549a55" + }, + "downloads": -1, + "filename": "Djblets-4.0.tar.gz", + "has_sig": false, + "md5_digest": "2df8f8ec08235d9aa08464f826d871ca", + "packagetype": "sdist", + "python_version": "source", + "requires_python": ">=3.8", + "size": 1320215, + "upload_time": "2023-10-17T07:55:37", + "upload_time_iso_8601": "2023-10-17T07:55:37.827462Z", + "url": "https://files.pythonhosted.org/packages/5e/f9/3571a6140a5dce9eaf7d3dfbceaf8c88ad6a0a407ec5e63a6591f5dcbb3e/Djblets-4.0.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "5.0": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "a8ab1c371b9472698a18d5a1f967c6f009c29d1155303e8569cc7644fb210be8", + "md5": "148468fc67bc5cd1646c9f6d4845bd36", + "sha256": "cc9d32f74ec50094eb4f8e766422611d69c6b6c4649596d17a4a8f4fadea4e96" + }, + "downloads": -1, + "filename": "Djblets-5.0-py3-none-any.whl", + "has_sig": false, + "md5_digest": "148468fc67bc5cd1646c9f6d4845bd36", + "packagetype": "bdist_wheel", + "python_version": "py3", + "requires_python": ">=3.8", + "size": 1722709, + "upload_time": "2024-06-06T08:38:31", + "upload_time_iso_8601": "2024-06-06T08:38:31.391966Z", + "url": "https://files.pythonhosted.org/packages/a8/ab/1c371b9472698a18d5a1f967c6f009c29d1155303e8569cc7644fb210be8/Djblets-5.0-py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "e440300ccd3f2da18385b5cc5206d0bd0d209e77388ebe6d1edb347a10f2407f", + "md5": "65488c9ccac32a5251dff03545fc8ad4", + "sha256": "2ad3e1dd58cdaa8ea2bcdc36b4ce42ed025e64d7f1c4854c6b52e26cda1fd045" + }, + "downloads": -1, + "filename": "Djblets-5.0.tar.gz", + "has_sig": false, + "md5_digest": "65488c9ccac32a5251dff03545fc8ad4", + "packagetype": "sdist", + "python_version": "source", + "requires_python": ">=3.8", + "size": 1362554, + "upload_time": "2024-06-06T08:38:33", + "upload_time_iso_8601": "2024-06-06T08:38:33.983937Z", + "url": "https://files.pythonhosted.org/packages/e4/40/300ccd3f2da18385b5cc5206d0bd0d209e77388ebe6d1edb347a10f2407f/Djblets-5.0.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "5.0.1": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "6c33f8adb08ca2be49b206994d3fa7f1fd89bd7a41480f44929f5640fa048f68", + "md5": "1ccc21f92014cf732998cc3e2a55f9bc", + "sha256": "8f90dad57d6b35ea5125e96d9a63be62877594fe922091219fad752ff667c2eb" + }, + "downloads": -1, + "filename": "Djblets-5.0.1-py3-none-any.whl", + "has_sig": false, + "md5_digest": "1ccc21f92014cf732998cc3e2a55f9bc", + "packagetype": "bdist_wheel", + "python_version": "py3", + "requires_python": ">=3.8", + "size": 1722630, + "upload_time": "2024-06-07T18:44:27", + "upload_time_iso_8601": "2024-06-07T18:44:27.573664Z", + "url": "https://files.pythonhosted.org/packages/6c/33/f8adb08ca2be49b206994d3fa7f1fd89bd7a41480f44929f5640fa048f68/Djblets-5.0.1-py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "3323bcd7bd6eba91adf928fde3742be2f135c7977b0ab04cf93c7e7bede41b5e", + "md5": "9ffa6c897881a144add02366968dd453", + "sha256": "75efba870163790db84a4e0f1fc0f228deb4d8a320d6f5ba76dadc88a36dcabc" + }, + "downloads": -1, + "filename": "Djblets-5.0.1.tar.gz", + "has_sig": false, + "md5_digest": "9ffa6c897881a144add02366968dd453", + "packagetype": "sdist", + "python_version": "source", + "requires_python": ">=3.8", + "size": 1362914, + "upload_time": "2024-06-07T18:44:31", + "upload_time_iso_8601": "2024-06-07T18:44:31.459744Z", + "url": "https://files.pythonhosted.org/packages/33/23/bcd7bd6eba91adf928fde3742be2f135c7977b0ab04cf93c7e7bede41b5e/Djblets-5.0.1.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "5.0.2": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "6f38c69fc596b9602206f19152db0f6bbeea886c60c0da4e7083651e9ca01a2e", + "md5": "bb2b69fc7026c43ca9180a546403bd06", + "sha256": "0809f0d04dc8fc131b5d4dd54ccf86854362c689e53910136402b5593b3ca1cb" + }, + "downloads": -1, + "filename": "Djblets-5.0.2-py3-none-any.whl", + "has_sig": false, + "md5_digest": "bb2b69fc7026c43ca9180a546403bd06", + "packagetype": "bdist_wheel", + "python_version": "py3", + "requires_python": ">=3.8", + "size": 1722641, + "upload_time": "2024-06-12T07:07:00", + "upload_time_iso_8601": "2024-06-12T07:07:00.133127Z", + "url": "https://files.pythonhosted.org/packages/6f/38/c69fc596b9602206f19152db0f6bbeea886c60c0da4e7083651e9ca01a2e/Djblets-5.0.2-py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "08ffba6f8c3882018b18d76f0b7bc2e165dbc1ea98a7414266b022233f141e72", + "md5": "c3e4fef5887b161b4730930ac8bf59fa", + "sha256": "4f2f1f7afe9561d506640abe51137d647bc6da4be055a0a278869a9e08654716" + }, + "downloads": -1, + "filename": "Djblets-5.0.2.tar.gz", + "has_sig": false, + "md5_digest": "c3e4fef5887b161b4730930ac8bf59fa", + "packagetype": "sdist", + "python_version": "source", + "requires_python": ">=3.8", + "size": 1363442, + "upload_time": "2024-06-12T07:07:06", + "upload_time_iso_8601": "2024-06-12T07:07:06.679773Z", + "url": "https://files.pythonhosted.org/packages/08/ff/ba6f8c3882018b18d76f0b7bc2e165dbc1ea98a7414266b022233f141e72/Djblets-5.0.2.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "5.1": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "a259ea23385440449e5c4e1209d0e1110ac94cc9dcdf99cca695c99dd554a9a1", + "md5": "3edb679b8e59b2f9fee0eebf539b6e64", + "sha256": "17af591f892c7c302ad7451c678074cec4cf6d897c7c7a7073417e1530393552" + }, + "downloads": -1, + "filename": "Djblets-5.1-py3-none-any.whl", + "has_sig": false, + "md5_digest": "3edb679b8e59b2f9fee0eebf539b6e64", + "packagetype": "bdist_wheel", + "python_version": "py3", + "requires_python": ">=3.8", + "size": 1731030, + "upload_time": "2024-07-02T06:08:51", + "upload_time_iso_8601": "2024-07-02T06:08:51.829415Z", + "url": "https://files.pythonhosted.org/packages/a2/59/ea23385440449e5c4e1209d0e1110ac94cc9dcdf99cca695c99dd554a9a1/Djblets-5.1-py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "c67e3408a2c4b4ce5bd6d19a10a04e7adb1c6647ab115fb486d57cf7a0e5d534", + "md5": "1b91e4dd93b72ad3e2575f0103e51859", + "sha256": "20b95d58c5257bb851c0cd7c8e5fb7d3533ea4c0141f1537c4fa24af0b58b7c6" + }, + "downloads": -1, + "filename": "Djblets-5.1.tar.gz", + "has_sig": false, + "md5_digest": "1b91e4dd93b72ad3e2575f0103e51859", + "packagetype": "sdist", + "python_version": "source", + "requires_python": ">=3.8", + "size": 1374585, + "upload_time": "2024-07-02T06:08:58", + "upload_time_iso_8601": "2024-07-02T06:08:58.290996Z", + "url": "https://files.pythonhosted.org/packages/c6/7e/3408a2c4b4ce5bd6d19a10a04e7adb1c6647ab115fb486d57cf7a0e5d534/Djblets-5.1.tar.gz", + "yanked": false, + "yanked_reason": null + } + ], + "5.1.1": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "4be2a2ce0284a18c116f346d67b4d9d6b4cf6bb45a551a8d4f7ba259b88bcdd5", + "md5": "8c5df0c09b2de290d439667baaf619aa", + "sha256": "6e5be1d860d4e5f95a63b714ee476b0ba26c3b78323da1de164db2b243e39256" + }, + "downloads": -1, + "filename": "Djblets-5.1.1-py3-none-any.whl", + "has_sig": false, + "md5_digest": "8c5df0c09b2de290d439667baaf619aa", + "packagetype": "bdist_wheel", + "python_version": "py3", + "requires_python": ">=3.8", + "size": 1731213, + "upload_time": "2024-08-20T04:56:16", + "upload_time_iso_8601": "2024-08-20T04:56:16.839923Z", + "url": "https://files.pythonhosted.org/packages/4b/e2/a2ce0284a18c116f346d67b4d9d6b4cf6bb45a551a8d4f7ba259b88bcdd5/Djblets-5.1.1-py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "d5ed70afec15854ad82ae13a41ecfa79c824e8103aa982e3d4c21f06bb1ea4e0", + "md5": "3b3604c8014f42fb5158f9eb3d072103", + "sha256": "6efea6c518f2d4e7926bb756081cb0c3429616c76199d1c50376d79348fcb8d3" + }, + "downloads": -1, + "filename": "djblets-5.1.1.tar.gz", + "has_sig": false, + "md5_digest": "3b3604c8014f42fb5158f9eb3d072103", + "packagetype": "sdist", + "python_version": "source", + "requires_python": ">=3.8", + "size": 1374931, + "upload_time": "2024-08-20T04:56:19", + "upload_time_iso_8601": "2024-08-20T04:56:19.606276Z", + "url": "https://files.pythonhosted.org/packages/d5/ed/70afec15854ad82ae13a41ecfa79c824e8103aa982e3d4c21f06bb1ea4e0/djblets-5.1.1.tar.gz", + "yanked": false, + "yanked_reason": null + } + ] + }, + "urls": [ + { + "comment_text": "", + "digests": { + "blake2b_256": "4be2a2ce0284a18c116f346d67b4d9d6b4cf6bb45a551a8d4f7ba259b88bcdd5", + "md5": "8c5df0c09b2de290d439667baaf619aa", + "sha256": "6e5be1d860d4e5f95a63b714ee476b0ba26c3b78323da1de164db2b243e39256" + }, + "downloads": -1, + "filename": "Djblets-5.1.1-py3-none-any.whl", + "has_sig": false, + "md5_digest": "8c5df0c09b2de290d439667baaf619aa", + "packagetype": "bdist_wheel", + "python_version": "py3", + "requires_python": ">=3.8", + "size": 1731213, + "upload_time": "2024-08-20T04:56:16", + "upload_time_iso_8601": "2024-08-20T04:56:16.839923Z", + "url": "https://files.pythonhosted.org/packages/4b/e2/a2ce0284a18c116f346d67b4d9d6b4cf6bb45a551a8d4f7ba259b88bcdd5/Djblets-5.1.1-py3-none-any.whl", + "yanked": false, + "yanked_reason": null + }, + { + "comment_text": "", + "digests": { + "blake2b_256": "d5ed70afec15854ad82ae13a41ecfa79c824e8103aa982e3d4c21f06bb1ea4e0", + "md5": "3b3604c8014f42fb5158f9eb3d072103", + "sha256": "6efea6c518f2d4e7926bb756081cb0c3429616c76199d1c50376d79348fcb8d3" + }, + "downloads": -1, + "filename": "djblets-5.1.1.tar.gz", + "has_sig": false, + "md5_digest": "3b3604c8014f42fb5158f9eb3d072103", + "packagetype": "sdist", + "python_version": "source", + "requires_python": ">=3.8", + "size": 1374931, + "upload_time": "2024-08-20T04:56:19", + "upload_time_iso_8601": "2024-08-20T04:56:19.606276Z", + "url": "https://files.pythonhosted.org/packages/d5/ed/70afec15854ad82ae13a41ecfa79c824e8103aa982e3d4c21f06bb1ea4e0/djblets-5.1.1.tar.gz", + "yanked": false, + "yanked_reason": null + } ], "vulnerabilities": [] -} \ No newline at end of file + } \ No newline at end of file diff --git a/tests/test_package_versions.py b/tests/test_package_versions.py index 6c17c7b3..113cdbfb 100644 --- a/tests/test_package_versions.py +++ b/tests/test_package_versions.py @@ -61,7 +61,7 @@ def test_get_launchpad_versions_from_purl(mock_get_response): @mock.patch("fetchcode.package_versions.get_response") def test_get_pypi_versions_from_purl(mock_get_response): side_effect = [get_json_data(data_location / "pypi_mock_data.json")] - purl = "pkg:pypi/django" + purl = "pkg:pypi/Djblets" expected_file = data_location / "pypi.json" mock_get_response.side_effect = side_effect result = list(versions(purl)) From d984103fa44d6b6bfb02b28332dd719f835f211e Mon Sep 17 00:00:00 2001 From: Keshav Priyadarshi Date: Tue, 3 Sep 2024 18:16:41 +0530 Subject: [PATCH 16/28] Return all versions of a PyPI package Signed-off-by: Keshav Priyadarshi --- src/fetchcode/package_versions.py | 12 +- tests/data/package_versions/pypi.json | 304 ++++++++++++++++++++++++++ 2 files changed, 308 insertions(+), 8 deletions(-) diff --git a/src/fetchcode/package_versions.py b/src/fetchcode/package_versions.py index d2720b3d..e08fc4b9 100644 --- a/src/fetchcode/package_versions.py +++ b/src/fetchcode/package_versions.py @@ -122,10 +122,8 @@ def get_pypi_versions_from_purl(purl): releases = response.get("releases") or {} for version, download_items in releases.items(): - if not download_items: - continue + release_date = get_pypi_latest_date(download_items) if download_items else None - release_date = get_pypi_latest_date(download_items) yield PackageVersion( value=version, release_date=release_date, @@ -317,8 +315,7 @@ def get_golang_versions_from_purl(purl): break if response is None or escaped_pkg is None or trimmed_pkg is None: - logger.error( - f"Error while fetching versions for {package_slug!r} from goproxy") + logger.error(f"Error while fetching versions for {package_slug!r} from goproxy") return for version_info in response.split("\n"): @@ -348,7 +345,7 @@ def trim_go_url_path(url_path: str) -> Optional[str]: # some advisories contains this prefix in package name, e.g. https://github.com/advisories/GHSA-7h6j-2268-fhcm go_url_prefix = "https://pkg.go.dev/" if url_path.startswith(go_url_prefix): - url_path = url_path[len(go_url_prefix):] + url_path = url_path[len(go_url_prefix) :] parsed_url_path = urlparse(url_path) path = parsed_url_path.path @@ -409,8 +406,7 @@ def fetch_version_info(version_info: str, escaped_pkg: str) -> Optional[PackageV f"Error while fetching version info for {escaped_pkg}/{escaped_ver} " f"from goproxy:\n{traceback.format_exc()}" ) - release_date = dateparser.parse( - response.get("Time", "")) if response else None + release_date = dateparser.parse(response.get("Time", "")) if response else None return PackageVersion(value=version, release_date=release_date) diff --git a/tests/data/package_versions/pypi.json b/tests/data/package_versions/pypi.json index 259bf0e3..feaabdef 100644 --- a/tests/data/package_versions/pypi.json +++ b/tests/data/package_versions/pypi.json @@ -1,4 +1,280 @@ [ + { + "value": "0.5", + "release_date": null + }, + { + "value": "0.5.1", + "release_date": null + }, + { + "value": "0.5.2", + "release_date": null + }, + { + "value": "0.5.3", + "release_date": null + }, + { + "value": "0.5.4", + "release_date": null + }, + { + "value": "0.5.5", + "release_date": null + }, + { + "value": "0.5.6", + "release_date": null + }, + { + "value": "0.5.7", + "release_date": null + }, + { + "value": "0.5.8", + "release_date": null + }, + { + "value": "0.5.9", + "release_date": null + }, + { + "value": "0.5rc2", + "release_date": null + }, + { + "value": "0.6", + "release_date": null + }, + { + "value": "0.6.1", + "release_date": null + }, + { + "value": "0.6.10", + "release_date": null + }, + { + "value": "0.6.11", + "release_date": null + }, + { + "value": "0.6.12", + "release_date": null + }, + { + "value": "0.6.13", + "release_date": null + }, + { + "value": "0.6.14", + "release_date": null + }, + { + "value": "0.6.15", + "release_date": null + }, + { + "value": "0.6.16", + "release_date": null + }, + { + "value": "0.6.17", + "release_date": null + }, + { + "value": "0.6.18", + "release_date": null + }, + { + "value": "0.6.19", + "release_date": null + }, + { + "value": "0.6.2", + "release_date": null + }, + { + "value": "0.6.20", + "release_date": null + }, + { + "value": "0.6.21", + "release_date": null + }, + { + "value": "0.6.22", + "release_date": null + }, + { + "value": "0.6.23", + "release_date": null + }, + { + "value": "0.6.24", + "release_date": null + }, + { + "value": "0.6.25", + "release_date": null + }, + { + "value": "0.6.26", + "release_date": null + }, + { + "value": "0.6.27", + "release_date": null + }, + { + "value": "0.6.28", + "release_date": null + }, + { + "value": "0.6.29", + "release_date": null + }, + { + "value": "0.6.3", + "release_date": null + }, + { + "value": "0.6.30", + "release_date": null + }, + { + "value": "0.6.31", + "release_date": null + }, + { + "value": "0.6.4", + "release_date": null + }, + { + "value": "0.6.5", + "release_date": null + }, + { + "value": "0.6.6", + "release_date": null + }, + { + "value": "0.6.7", + "release_date": null + }, + { + "value": "0.6.8", + "release_date": null + }, + { + "value": "0.6.9", + "release_date": null + }, + { + "value": "0.7", + "release_date": null + }, + { + "value": "0.7.1", + "release_date": null + }, + { + "value": "0.7.10", + "release_date": null + }, + { + "value": "0.7.11", + "release_date": null + }, + { + "value": "0.7.12", + "release_date": null + }, + { + "value": "0.7.13", + "release_date": null + }, + { + "value": "0.7.14", + "release_date": null + }, + { + "value": "0.7.15", + "release_date": null + }, + { + "value": "0.7.16", + "release_date": null + }, + { + "value": "0.7.17", + "release_date": null + }, + { + "value": "0.7.18", + "release_date": null + }, + { + "value": "0.7.19", + "release_date": null + }, + { + "value": "0.7.2", + "release_date": null + }, + { + "value": "0.7.20", + "release_date": null + }, + { + "value": "0.7.21", + "release_date": null + }, + { + "value": "0.7.22", + "release_date": null + }, + { + "value": "0.7.23", + "release_date": null + }, + { + "value": "0.7.24", + "release_date": null + }, + { + "value": "0.7.25", + "release_date": null + }, + { + "value": "0.7.26", + "release_date": null + }, + { + "value": "0.7.27", + "release_date": null + }, + { + "value": "0.7.28", + "release_date": null + }, + { + "value": "0.7.29", + "release_date": null + }, + { + "value": "0.7.3", + "release_date": null + }, + { + "value": "0.7.30", + "release_date": null + }, + { + "value": "0.7.31", + "release_date": null + }, { "value": "0.7.32", "release_date": "2016-02-22T06:13:06.774750+00:00" @@ -7,6 +283,30 @@ "value": "0.7.33", "release_date": "2016-03-01T18:55:48.013515+00:00" }, + { + "value": "0.7.4", + "release_date": null + }, + { + "value": "0.7.5", + "release_date": null + }, + { + "value": "0.7.6", + "release_date": null + }, + { + "value": "0.7.7", + "release_date": null + }, + { + "value": "0.7.8", + "release_date": null + }, + { + "value": "0.7.9", + "release_date": null + }, { "value": "0.8", "release_date": "2015-10-29T21:14:49.182641+00:00" @@ -115,6 +415,10 @@ "value": "0.8.6", "release_date": "2015-10-29T21:13:20.969507+00:00" }, + { + "value": "0.8.6.dev", + "release_date": null + }, { "value": "0.8.7", "release_date": "2015-10-29T21:13:40.939660+00:00" From e870e2aedcc9f4833117a29a66459d80c4ed2a33 Mon Sep 17 00:00:00 2001 From: Keshav Priyadarshi Date: Tue, 10 Sep 2024 18:00:07 +0530 Subject: [PATCH 17/28] Remove macos-14 runner Signed-off-by: Keshav Priyadarshi --- azure-pipelines.yml | 8 -------- 1 file changed, 8 deletions(-) diff --git a/azure-pipelines.yml b/azure-pipelines.yml index 69f88637..40ace8bc 100644 --- a/azure-pipelines.yml +++ b/azure-pipelines.yml @@ -45,14 +45,6 @@ jobs: test_suites: all: venv/bin/pytest -n 2 -vvs - - template: etc/ci/azure-posix.yml - parameters: - job_name: macos14_cpython - image_name: macOS-14-large - python_versions: ['3.8', '3.8', '3.9', '3.10', '3.12'] - test_suites: - all: venv/bin/pytest -n 2 -vvs - - template: etc/ci/azure-win.yml parameters: job_name: win2019_cpython From a73c35b8bdd60fe92f1c8a429dddeee60724f81d Mon Sep 17 00:00:00 2001 From: Keshav Priyadarshi Date: Tue, 10 Sep 2024 19:04:19 +0530 Subject: [PATCH 18/28] Upgrade pip to 24.2 Signed-off-by: Keshav Priyadarshi --- src/fetchcode/vcs/git.py | 3 + src/fetchcode/vcs/pip.ABOUT | 6 +- src/fetchcode/vcs/pip/__init__.py | 13 +- src/fetchcode/vcs/pip/__main__.py | 10 +- src/fetchcode/vcs/pip/__pip-runner__.py | 50 + src/fetchcode/vcs/pip/_internal/__init__.py | 15 +- src/fetchcode/vcs/pip/_internal/build_env.py | 306 +- src/fetchcode/vcs/pip/_internal/cache.py | 271 +- .../vcs/pip/_internal/cli/autocompletion.py | 112 +- .../vcs/pip/_internal/cli/base_command.py | 253 +- .../vcs/pip/_internal/cli/cmdoptions.py | 1125 +- .../vcs/pip/_internal/cli/command_context.py | 25 +- .../vcs/pip/_internal/cli/index_command.py | 170 + src/fetchcode/vcs/pip/_internal/cli/main.py | 33 +- .../vcs/pip/_internal/cli/main_parser.py | 107 +- src/fetchcode/vcs/pip/_internal/cli/parser.py | 184 +- .../vcs/pip/_internal/cli/progress_bars.py | 361 +- .../vcs/pip/_internal/cli/req_command.py | 373 +- .../vcs/pip/_internal/cli/spinners.py | 82 +- .../vcs/pip/_internal/cli/status_codes.py | 2 - .../vcs/pip/_internal/commands/__init__.py | 200 +- .../vcs/pip/_internal/commands/cache.py | 190 +- .../vcs/pip/_internal/commands/check.py | 48 +- .../vcs/pip/_internal/commands/completion.py | 129 +- .../pip/_internal/commands/configuration.py | 191 +- .../vcs/pip/_internal/commands/debug.py | 222 +- .../vcs/pip/_internal/commands/download.py | 116 +- .../vcs/pip/_internal/commands/freeze.py | 126 +- .../vcs/pip/_internal/commands/hash.py | 45 +- .../vcs/pip/_internal/commands/help.py | 26 +- .../vcs/pip/_internal/commands/index.py | 139 + .../vcs/pip/_internal/commands/inspect.py | 92 + .../vcs/pip/_internal/commands/install.py | 840 +- .../vcs/pip/_internal/commands/list.py | 356 +- .../vcs/pip/_internal/commands/search.py | 166 +- .../vcs/pip/_internal/commands/show.py | 265 +- .../vcs/pip/_internal/commands/uninstall.py | 93 +- .../vcs/pip/_internal/commands/wheel.py | 182 +- .../vcs/pip/_internal/configuration.py | 333 +- .../pip/_internal/distributions/__init__.py | 19 +- .../vcs/pip/_internal/distributions/base.py | 44 +- .../pip/_internal/distributions/installed.py | 27 +- .../vcs/pip/_internal/distributions/sdist.py | 152 +- .../vcs/pip/_internal/distributions/wheel.py | 48 +- src/fetchcode/vcs/pip/_internal/exceptions.py | 665 +- .../vcs/pip/_internal/index/collector.py | 809 +- .../vcs/pip/_internal/index/package_finder.py | 730 +- .../vcs/pip/_internal/index/sources.py | 285 + src/fetchcode/vcs/pip/_internal/locations.py | 194 - .../vcs/pip/_internal/locations/__init__.py | 456 + .../vcs/pip/_internal/locations/_distutils.py | 172 + .../vcs/pip/_internal/locations/_sysconfig.py | 214 + .../vcs/pip/_internal/locations/base.py | 81 + src/fetchcode/vcs/pip/_internal/main.py | 10 +- .../vcs/pip/_internal/metadata/__init__.py | 128 + .../vcs/pip/_internal/metadata/_json.py | 84 + .../vcs/pip/_internal/metadata/base.py | 688 + .../_internal/metadata/importlib/__init__.py | 6 + .../_internal/metadata/importlib/_compat.py | 85 + .../_internal/metadata/importlib/_dists.py | 221 + .../pip/_internal/metadata/importlib/_envs.py | 189 + .../pip/_internal/metadata/pkg_resources.py | 301 + .../vcs/pip/_internal/models/candidate.py | 61 +- .../vcs/pip/_internal/models/direct_url.py | 195 +- .../pip/_internal/models/format_control.py | 84 +- .../vcs/pip/_internal/models/index.py | 31 +- .../_internal/models/installation_report.py | 56 + .../vcs/pip/_internal/models/link.py | 578 +- .../vcs/pip/_internal/models/scheme.py | 28 +- .../vcs/pip/_internal/models/search_scope.py | 96 +- .../pip/_internal/models/selection_prefs.py | 32 +- .../vcs/pip/_internal/models/target_python.py | 95 +- .../vcs/pip/_internal/models/wheel.py | 75 +- .../vcs/pip/_internal/network/auth.py | 444 +- .../vcs/pip/_internal/network/cache.py | 91 +- .../vcs/pip/_internal/network/download.py | 193 +- .../vcs/pip/_internal/network/lazy_wheel.py | 210 + .../vcs/pip/_internal/network/session.py | 363 +- .../vcs/pip/_internal/network/utils.py | 66 +- .../vcs/pip/_internal/network/xmlrpc.py | 58 +- .../operations/build}/__init__.py | 0 .../operations/build/build_tracker.py | 138 + .../_internal/operations/build/metadata.py | 39 + .../operations/build/metadata_editable.py | 41 + .../operations/build/metadata_legacy.py | 74 + .../pip/_internal/operations/build/wheel.py | 37 + .../operations/build/wheel_editable.py | 46 + .../operations/build/wheel_legacy.py | 102 + .../vcs/pip/_internal/operations/check.py | 150 +- .../vcs/pip/_internal/operations/freeze.py | 298 +- .../operations/install/editable_legacy.py | 43 +- .../_internal/operations/install/legacy.py | 142 - .../pip/_internal/operations/install/wheel.py | 932 +- .../vcs/pip/_internal/operations/prepare.py | 940 +- src/fetchcode/vcs/pip/_internal/pyproject.py | 129 +- .../vcs/pip/_internal/req/__init__.py | 92 +- .../vcs/pip/_internal/req/constructors.py | 490 +- .../vcs/pip/_internal/req/req_file.py | 427 +- .../vcs/pip/_internal/req/req_install.py | 950 +- .../vcs/pip/_internal/req/req_set.py | 192 +- .../vcs/pip/_internal/req/req_tracker.py | 151 - .../vcs/pip/_internal/req/req_uninstall.py | 548 +- .../vcs/pip/_internal/resolution/base.py | 26 +- .../_internal/resolution/legacy/resolver.py | 410 +- .../_internal/resolution/resolvelib/base.py | 139 +- .../resolution/resolvelib/candidates.py | 679 +- .../resolution/resolvelib/factory.py | 876 +- .../resolution/resolvelib/found_candidates.py | 174 + .../resolution/resolvelib/provider.py | 280 +- .../resolution/resolvelib/reporter.py | 81 + .../resolution/resolvelib/requirements.py | 270 +- .../resolution/resolvelib/resolver.py | 355 +- .../vcs/pip/_internal/self_outdated_check.py | 342 +- .../vcs/pip/_internal/utils/_jaraco_text.py | 109 + src/fetchcode/vcs/pip/_internal/utils/_log.py | 38 + .../vcs/pip/_internal/utils/appdirs.py | 54 +- .../vcs/pip/_internal/utils/compat.py | 235 +- .../pip/_internal/utils/compatibility_tags.py | 108 +- .../vcs/pip/_internal/utils/datetime.py | 11 + .../vcs/pip/_internal/utils/deprecation.py | 106 +- .../pip/_internal/utils/direct_url_helpers.py | 85 +- .../vcs/pip/_internal/utils/distutils_args.py | 48 - .../vcs/pip/_internal/utils/egg_link.py | 80 + .../vcs/pip/_internal/utils/encoding.py | 42 +- .../vcs/pip/_internal/utils/entrypoints.py | 65 +- .../vcs/pip/_internal/utils/filesystem.py | 145 +- .../vcs/pip/_internal/utils/filetypes.py | 33 +- .../vcs/pip/_internal/utils/glibc.py | 45 +- .../vcs/pip/_internal/utils/hashes.py | 132 +- .../_internal/utils/inject_securetransport.py | 36 - .../vcs/pip/_internal/utils/logging.py | 460 +- src/fetchcode/vcs/pip/_internal/utils/misc.py | 962 +- .../vcs/pip/_internal/utils/models.py | 42 - .../vcs/pip/_internal/utils/packaging.py | 89 +- .../vcs/pip/_internal/utils/pkg_resources.py | 44 - .../vcs/pip/_internal/utils/retry.py | 42 + .../pip/_internal/utils/setuptools_build.py | 179 +- .../vcs/pip/_internal/utils/subprocess.py | 270 +- .../vcs/pip/_internal/utils/temp_dir.py | 179 +- .../vcs/pip/_internal/utils/typing.py | 38 - .../vcs/pip/_internal/utils/unpacking.py | 335 +- src/fetchcode/vcs/pip/_internal/utils/urls.py | 62 +- .../vcs/pip/_internal/utils/virtualenv.py | 56 +- .../vcs/pip/_internal/utils/wheel.py | 143 +- .../vcs/pip/_internal/vcs/__init__.py | 14 +- src/fetchcode/vcs/pip/_internal/vcs/bazaar.py | 140 +- src/fetchcode/vcs/pip/_internal/vcs/git.py | 427 +- .../vcs/pip/_internal/vcs/mercurial.py | 156 +- .../vcs/pip/_internal/vcs/subversion.py | 244 +- .../vcs/pip/_internal/vcs/versioncontrol.py | 525 +- .../vcs/pip/_internal/wheel_builder.py | 293 +- src/fetchcode/vcs/pip/_vendor/README.rst | 152 +- src/fetchcode/vcs/pip/_vendor/__init__.py | 30 +- .../vcs/pip/_vendor/appdirs.LICENSE.txt | 23 - src/fetchcode/vcs/pip/_vendor/appdirs.py | 633 - .../vcs/pip/_vendor/cachecontrol.pyi | 1 - .../vcs/pip/_vendor/cachecontrol/LICENSE.txt | 6 +- .../vcs/pip/_vendor/cachecontrol/__init__.py | 25 +- .../vcs/pip/_vendor/cachecontrol/_cmd.py | 37 +- .../vcs/pip/_vendor/cachecontrol/adapter.py | 88 +- .../vcs/pip/_vendor/cachecontrol/cache.py | 55 +- .../_vendor/cachecontrol/caches/__init__.py | 10 +- .../_vendor/cachecontrol/caches/file_cache.py | 124 +- .../cachecontrol/caches/redis_cache.py | 39 +- .../vcs/pip/_vendor/cachecontrol/compat.py | 29 - .../pip/_vendor/cachecontrol/controller.py | 255 +- .../pip/_vendor/cachecontrol/filewrapper.py | 67 +- .../pip/_vendor/cachecontrol/heuristics.py | 61 +- .../vcs/pip/_vendor/cachecontrol/py.typed | 0 .../vcs/pip/_vendor/cachecontrol/serialize.py | 216 +- .../vcs/pip/_vendor/cachecontrol/wrapper.py | 38 +- src/fetchcode/vcs/pip/_vendor/certifi.pyi | 1 - src/fetchcode/vcs/pip/_vendor/certifi/LICENSE | 5 +- .../vcs/pip/_vendor/certifi/__init__.py | 3 +- .../vcs/pip/_vendor/certifi/__main__.py | 2 +- .../vcs/pip/_vendor/certifi/cacert.pem | 3149 ++-- src/fetchcode/vcs/pip/_vendor/certifi/core.py | 114 +- .../vcs/pip/_vendor/certifi/py.typed | 0 src/fetchcode/vcs/pip/_vendor/chardet.pyi | 1 - src/fetchcode/vcs/pip/_vendor/chardet/LICENSE | 504 - .../vcs/pip/_vendor/chardet/__init__.py | 39 - .../vcs/pip/_vendor/chardet/big5freq.py | 386 - .../vcs/pip/_vendor/chardet/big5prober.py | 47 - .../pip/_vendor/chardet/chardistribution.py | 233 - .../pip/_vendor/chardet/charsetgroupprober.py | 106 - .../vcs/pip/_vendor/chardet/charsetprober.py | 145 - .../vcs/pip/_vendor/chardet/cli/__init__.py | 1 - .../vcs/pip/_vendor/chardet/cli/chardetect.py | 85 - .../pip/_vendor/chardet/codingstatemachine.py | 88 - .../vcs/pip/_vendor/chardet/compat.py | 34 - .../vcs/pip/_vendor/chardet/cp949prober.py | 49 - .../vcs/pip/_vendor/chardet/enums.py | 76 - .../vcs/pip/_vendor/chardet/escprober.py | 101 - .../vcs/pip/_vendor/chardet/escsm.py | 246 - .../vcs/pip/_vendor/chardet/eucjpprober.py | 92 - .../vcs/pip/_vendor/chardet/euckrfreq.py | 195 - .../vcs/pip/_vendor/chardet/euckrprober.py | 47 - .../vcs/pip/_vendor/chardet/euctwfreq.py | 387 - .../vcs/pip/_vendor/chardet/euctwprober.py | 46 - .../vcs/pip/_vendor/chardet/gb2312freq.py | 283 - .../vcs/pip/_vendor/chardet/gb2312prober.py | 46 - .../vcs/pip/_vendor/chardet/hebrewprober.py | 292 - .../vcs/pip/_vendor/chardet/jisfreq.py | 325 - .../vcs/pip/_vendor/chardet/jpcntx.py | 233 - .../pip/_vendor/chardet/langbulgarianmodel.py | 228 - .../pip/_vendor/chardet/langcyrillicmodel.py | 333 - .../vcs/pip/_vendor/chardet/langgreekmodel.py | 225 - .../pip/_vendor/chardet/langhebrewmodel.py | 200 - .../pip/_vendor/chardet/langhungarianmodel.py | 225 - .../vcs/pip/_vendor/chardet/langthaimodel.py | 199 - .../pip/_vendor/chardet/langturkishmodel.py | 193 - .../vcs/pip/_vendor/chardet/latin1prober.py | 145 - .../pip/_vendor/chardet/mbcharsetprober.py | 91 - .../pip/_vendor/chardet/mbcsgroupprober.py | 54 - .../vcs/pip/_vendor/chardet/mbcssm.py | 572 - .../pip/_vendor/chardet/sbcharsetprober.py | 132 - .../pip/_vendor/chardet/sbcsgroupprober.py | 73 - .../vcs/pip/_vendor/chardet/sjisprober.py | 92 - .../pip/_vendor/chardet/universaldetector.py | 286 - .../vcs/pip/_vendor/chardet/utf8prober.py | 82 - .../vcs/pip/_vendor/chardet/version.py | 9 - src/fetchcode/vcs/pip/_vendor/colorama.pyi | 1 - .../vcs/pip/_vendor/colorama/LICENSE.txt | 27 - .../vcs/pip/_vendor/colorama/__init__.py | 6 - .../vcs/pip/_vendor/colorama/ansi.py | 102 - .../vcs/pip/_vendor/colorama/ansitowin32.py | 257 - .../vcs/pip/_vendor/colorama/initialise.py | 80 - .../vcs/pip/_vendor/colorama/win32.py | 152 - .../vcs/pip/_vendor/colorama/winterm.py | 169 - .../vcs/pip/_vendor/contextlib2.LICENSE.txt | 122 - src/fetchcode/vcs/pip/_vendor/contextlib2.py | 518 - src/fetchcode/vcs/pip/_vendor/distlib.pyi | 1 - .../vcs/pip/_vendor/distlib/__init__.py | 22 +- .../pip/_vendor/distlib/_backport/__init__.py | 6 - .../vcs/pip/_vendor/distlib/_backport/misc.py | 41 - .../pip/_vendor/distlib/_backport/shutil.py | 761 - .../_vendor/distlib/_backport/sysconfig.cfg | 84 - .../_vendor/distlib/_backport/sysconfig.py | 786 - .../pip/_vendor/distlib/_backport/tarfile.py | 2607 ---- .../vcs/pip/_vendor/distlib/compat.py | 208 +- .../vcs/pip/_vendor/distlib/database.py | 182 +- .../vcs/pip/_vendor/distlib/index.py | 48 +- .../vcs/pip/_vendor/distlib/locators.py | 47 +- .../vcs/pip/_vendor/distlib/manifest.py | 25 +- .../vcs/pip/_vendor/distlib/markers.py | 64 +- .../vcs/pip/_vendor/distlib/metadata.py | 186 +- .../vcs/pip/_vendor/distlib/resources.py | 7 +- .../vcs/pip/_vendor/distlib/scripts.py | 158 +- src/fetchcode/vcs/pip/_vendor/distlib/t32.exe | Bin 96768 -> 97792 bytes .../vcs/pip/_vendor/distlib/t64-arm.exe | Bin 0 -> 182784 bytes src/fetchcode/vcs/pip/_vendor/distlib/t64.exe | Bin 105984 -> 108032 bytes src/fetchcode/vcs/pip/_vendor/distlib/util.py | 524 +- .../vcs/pip/_vendor/distlib/version.py | 43 +- src/fetchcode/vcs/pip/_vendor/distlib/w32.exe | Bin 90112 -> 91648 bytes .../vcs/pip/_vendor/distlib/w64-arm.exe | Bin 0 -> 168448 bytes src/fetchcode/vcs/pip/_vendor/distlib/w64.exe | Bin 99840 -> 101888 bytes .../vcs/pip/_vendor/distlib/wheel.py | 223 +- src/fetchcode/vcs/pip/_vendor/distro.pyi | 1 - .../{distro.LICENSE => distro/LICENSE} | 0 .../vcs/pip/_vendor/distro/__init__.py | 54 + .../vcs/pip/_vendor/distro/__main__.py | 4 + .../vcs/pip/_vendor/{ => distro}/distro.py | 741 +- src/fetchcode/vcs/pip/_vendor/distro/py.typed | 0 src/fetchcode/vcs/pip/_vendor/html5lib.pyi | 1 - .../vcs/pip/_vendor/html5lib/LICENSE | 20 - .../vcs/pip/_vendor/html5lib/__init__.py | 35 - .../vcs/pip/_vendor/html5lib/_ihatexml.py | 288 - .../vcs/pip/_vendor/html5lib/_inputstream.py | 923 -- .../vcs/pip/_vendor/html5lib/_tokenizer.py | 1721 --- .../pip/_vendor/html5lib/_trie/__init__.py | 14 - .../vcs/pip/_vendor/html5lib/_trie/_base.py | 40 - .../vcs/pip/_vendor/html5lib/_trie/datrie.py | 44 - .../vcs/pip/_vendor/html5lib/_trie/py.py | 67 - .../vcs/pip/_vendor/html5lib/_utils.py | 124 - .../vcs/pip/_vendor/html5lib/constants.py | 2947 ---- .../filters/alphabeticalattributes.py | 29 - .../vcs/pip/_vendor/html5lib/filters/base.py | 12 - .../html5lib/filters/inject_meta_charset.py | 73 - .../vcs/pip/_vendor/html5lib/filters/lint.py | 93 - .../_vendor/html5lib/filters/optionaltags.py | 207 - .../pip/_vendor/html5lib/filters/sanitizer.py | 896 -- .../_vendor/html5lib/filters/whitespace.py | 38 - .../vcs/pip/_vendor/html5lib/html5parser.py | 2791 ---- .../vcs/pip/_vendor/html5lib/serializer.py | 409 - .../_vendor/html5lib/treeadapters/__init__.py | 30 - .../_vendor/html5lib/treeadapters/genshi.py | 54 - .../pip/_vendor/html5lib/treeadapters/sax.py | 50 - .../_vendor/html5lib/treebuilders/__init__.py | 88 - .../pip/_vendor/html5lib/treebuilders/base.py | 417 - .../pip/_vendor/html5lib/treebuilders/dom.py | 239 - .../_vendor/html5lib/treebuilders/etree.py | 340 - .../html5lib/treebuilders/etree_lxml.py | 366 - .../_vendor/html5lib/treewalkers/__init__.py | 154 - .../pip/_vendor/html5lib/treewalkers/base.py | 252 - .../pip/_vendor/html5lib/treewalkers/dom.py | 43 - .../pip/_vendor/html5lib/treewalkers/etree.py | 130 - .../html5lib/treewalkers/etree_lxml.py | 213 - .../_vendor/html5lib/treewalkers/genshi.py | 69 - src/fetchcode/vcs/pip/_vendor/idna.pyi | 1 - src/fetchcode/vcs/pip/_vendor/idna/LICENSE.md | 31 + .../vcs/pip/_vendor/idna/LICENSE.rst | 34 - .../vcs/pip/_vendor/idna/__init__.py | 44 +- src/fetchcode/vcs/pip/_vendor/idna/codec.py | 74 +- src/fetchcode/vcs/pip/_vendor/idna/compat.py | 9 +- src/fetchcode/vcs/pip/_vendor/idna/core.py | 199 +- .../vcs/pip/_vendor/idna/idnadata.py | 2426 ++- .../vcs/pip/_vendor/idna/intranges.py | 9 +- .../vcs/pip/_vendor/idna/package_data.py | 2 +- src/fetchcode/vcs/pip/_vendor/idna/py.typed | 0 .../vcs/pip/_vendor/idna/uts46data.py | 12409 ++++++++-------- .../vcs/pip/_vendor/ipaddress.LICENSE | 50 - src/fetchcode/vcs/pip/_vendor/ipaddress.py | 2420 --- src/fetchcode/vcs/pip/_vendor/ipaddress.pyi | 1 - src/fetchcode/vcs/pip/_vendor/msgpack.pyi | 1 - .../vcs/pip/_vendor/msgpack/__init__.py | 9 +- .../vcs/pip/_vendor/msgpack/_version.py | 1 - src/fetchcode/vcs/pip/_vendor/msgpack/ext.py | 63 +- .../vcs/pip/_vendor/msgpack/fallback.py | 440 +- src/fetchcode/vcs/pip/_vendor/packaging.pyi | 1 - .../vcs/pip/_vendor/packaging/__about__.py | 27 - .../vcs/pip/_vendor/packaging/__init__.py | 31 +- .../vcs/pip/_vendor/packaging/_compat.py | 38 - .../vcs/pip/_vendor/packaging/_elffile.py | 110 + .../vcs/pip/_vendor/packaging/_manylinux.py | 262 + .../vcs/pip/_vendor/packaging/_musllinux.py | 85 + .../vcs/pip/_vendor/packaging/_parser.py | 354 + .../vcs/pip/_vendor/packaging/_structures.py | 61 +- .../vcs/pip/_vendor/packaging/_tokenizer.py | 194 + .../vcs/pip/_vendor/packaging/_typing.py | 39 - .../vcs/pip/_vendor/packaging/markers.py | 333 +- .../vcs/pip/_vendor/packaging/metadata.py | 804 + .../vcs/pip/_vendor/packaging/requirements.py | 162 +- .../vcs/pip/_vendor/packaging/specifiers.py | 1100 +- .../vcs/pip/_vendor/packaging/tags.py | 705 +- .../vcs/pip/_vendor/packaging/utils.py | 166 +- .../vcs/pip/_vendor/packaging/version.py | 558 +- src/fetchcode/vcs/pip/_vendor/pep517.pyi | 1 - .../vcs/pip/_vendor/pep517/__init__.py | 4 - src/fetchcode/vcs/pip/_vendor/pep517/build.py | 124 - src/fetchcode/vcs/pip/_vendor/pep517/check.py | 203 - .../vcs/pip/_vendor/pep517/colorlog.py | 115 - .../vcs/pip/_vendor/pep517/compat.py | 34 - .../vcs/pip/_vendor/pep517/dirtools.py | 44 - .../vcs/pip/_vendor/pep517/envbuild.py | 167 - src/fetchcode/vcs/pip/_vendor/pep517/meta.py | 92 - .../vcs/pip/_vendor/pep517/wrappers.py | 308 - .../vcs/pip/_vendor/pkg_resources.pyi | 1 - .../vcs/pip/_vendor/pkg_resources/LICENSE | 24 +- .../vcs/pip/_vendor/pkg_resources/__init__.py | 1816 ++- .../pip/_vendor/pkg_resources/py31compat.py | 23 - .../vcs/pip/_vendor/platformdirs/LICENSE | 21 + .../vcs/pip/_vendor/platformdirs/__init__.py | 627 + .../vcs/pip/_vendor/platformdirs/__main__.py | 55 + .../vcs/pip/_vendor/platformdirs/android.py | 249 + .../vcs/pip/_vendor/platformdirs/api.py | 292 + .../vcs/pip/_vendor/platformdirs/macos.py | 130 + .../vcs/pip/_vendor/platformdirs/py.typed | 0 .../vcs/pip/_vendor/platformdirs/unix.py | 275 + .../vcs/pip/_vendor/platformdirs/version.py | 16 + .../vcs/pip/_vendor/platformdirs/windows.py | 272 + src/fetchcode/vcs/pip/_vendor/progress.pyi | 1 - .../vcs/pip/_vendor/progress/LICENSE | 13 - .../vcs/pip/_vendor/progress/__init__.py | 177 - src/fetchcode/vcs/pip/_vendor/progress/bar.py | 91 - .../vcs/pip/_vendor/progress/counter.py | 41 - .../vcs/pip/_vendor/progress/spinner.py | 43 - .../{webencodings => pygments}/LICENSE | 20 +- .../vcs/pip/_vendor/pygments/__init__.py | 82 + .../vcs/pip/_vendor/pygments/__main__.py | 17 + .../vcs/pip/_vendor/pygments/cmdline.py | 668 + .../vcs/pip/_vendor/pygments/console.py | 70 + .../vcs/pip/_vendor/pygments/filter.py | 70 + .../pip/_vendor/pygments/filters/__init__.py | 940 ++ .../vcs/pip/_vendor/pygments/formatter.py | 129 + .../_vendor/pygments/formatters/__init__.py | 157 + .../_vendor/pygments/formatters/_mapping.py | 23 + .../pip/_vendor/pygments/formatters/bbcode.py | 108 + .../pip/_vendor/pygments/formatters/groff.py | 170 + .../pip/_vendor/pygments/formatters/html.py | 987 ++ .../pip/_vendor/pygments/formatters/img.py | 685 + .../pip/_vendor/pygments/formatters/irc.py | 154 + .../pip/_vendor/pygments/formatters/latex.py | 518 + .../pip/_vendor/pygments/formatters/other.py | 160 + .../pygments/formatters/pangomarkup.py | 83 + .../pip/_vendor/pygments/formatters/rtf.py | 349 + .../pip/_vendor/pygments/formatters/svg.py | 185 + .../_vendor/pygments/formatters/terminal.py | 127 + .../pygments/formatters/terminal256.py | 338 + .../vcs/pip/_vendor/pygments/lexer.py | 963 ++ .../pip/_vendor/pygments/lexers/__init__.py | 362 + .../pip/_vendor/pygments/lexers/_mapping.py | 589 + .../vcs/pip/_vendor/pygments/lexers/python.py | 1198 ++ .../vcs/pip/_vendor/pygments/modeline.py | 43 + .../vcs/pip/_vendor/pygments/plugin.py | 72 + .../vcs/pip/_vendor/pygments/regexopt.py | 91 + .../vcs/pip/_vendor/pygments/scanner.py | 104 + .../vcs/pip/_vendor/pygments/sphinxext.py | 247 + .../vcs/pip/_vendor/pygments/style.py | 203 + .../pip/_vendor/pygments/styles/__init__.py | 61 + .../pip/_vendor/pygments/styles/_mapping.py | 54 + .../vcs/pip/_vendor/pygments/token.py | 214 + .../vcs/pip/_vendor/pygments/unistring.py | 153 + .../vcs/pip/_vendor/pygments/util.py | 324 + .../vcs/pip/_vendor/pyparsing.LICENSE | 18 - src/fetchcode/vcs/pip/_vendor/pyparsing.py | 7107 --------- src/fetchcode/vcs/pip/_vendor/pyparsing.pyi | 1 - .../{pep517 => pyproject_hooks}/LICENSE | 0 .../pip/_vendor/pyproject_hooks/__init__.py | 23 + .../pip/_vendor/pyproject_hooks/_compat.py | 8 + .../vcs/pip/_vendor/pyproject_hooks/_impl.py | 330 + .../pyproject_hooks/_in_process/__init__.py | 18 + .../_in_process}/_in_process.py | 131 +- src/fetchcode/vcs/pip/_vendor/requests.pyi | 1 - .../vcs/pip/_vendor/requests/LICENSE | 182 +- .../vcs/pip/_vendor/requests/__init__.py | 150 +- .../vcs/pip/_vendor/requests/__version__.py | 20 +- .../pip/_vendor/requests/_internal_utils.py | 26 +- .../vcs/pip/_vendor/requests/adapters.py | 484 +- src/fetchcode/vcs/pip/_vendor/requests/api.py | 28 +- .../vcs/pip/_vendor/requests/auth.py | 157 +- .../vcs/pip/_vendor/requests/certs.py | 12 +- .../vcs/pip/_vendor/requests/compat.py | 104 +- .../vcs/pip/_vendor/requests/cookies.py | 168 +- .../vcs/pip/_vendor/requests/exceptions.py | 57 +- .../vcs/pip/_vendor/requests/help.py | 108 +- .../vcs/pip/_vendor/requests/hooks.py | 7 +- .../vcs/pip/_vendor/requests/models.py | 399 +- .../vcs/pip/_vendor/requests/packages.py | 13 +- .../vcs/pip/_vendor/requests/sessions.py | 328 +- .../vcs/pip/_vendor/requests/status_codes.py | 169 +- .../vcs/pip/_vendor/requests/structures.py | 12 +- .../vcs/pip/_vendor/requests/utils.py | 482 +- src/fetchcode/vcs/pip/_vendor/resolvelib.pyi | 1 - .../vcs/pip/_vendor/resolvelib/__init__.py | 4 +- .../pip/_vendor/resolvelib/compat/__init__.py | 0 .../resolvelib/compat/collections_abc.py | 6 + .../vcs/pip/_vendor/resolvelib/providers.py | 130 +- .../vcs/pip/_vendor/resolvelib/py.typed | 0 .../vcs/pip/_vendor/resolvelib/reporters.py | 31 +- .../vcs/pip/_vendor/resolvelib/resolvers.py | 403 +- .../vcs/pip/_vendor/resolvelib/structs.py | 118 +- .../vcs/pip/_vendor/retrying.LICENSE | 202 - src/fetchcode/vcs/pip/_vendor/retrying.py | 267 - src/fetchcode/vcs/pip/_vendor/retrying.pyi | 1 - src/fetchcode/vcs/pip/_vendor/rich/LICENSE | 19 + .../vcs/pip/_vendor/rich/__init__.py | 177 + .../vcs/pip/_vendor/rich/__main__.py | 273 + .../vcs/pip/_vendor/rich/_cell_widths.py | 454 + .../vcs/pip/_vendor/rich/_emoji_codes.py | 3610 +++++ .../vcs/pip/_vendor/rich/_emoji_replace.py | 32 + .../vcs/pip/_vendor/rich/_export_format.py | 76 + .../vcs/pip/_vendor/rich/_extension.py | 10 + src/fetchcode/vcs/pip/_vendor/rich/_fileno.py | 24 + .../vcs/pip/_vendor/rich/_inspect.py | 270 + .../vcs/pip/_vendor/rich/_log_render.py | 94 + src/fetchcode/vcs/pip/_vendor/rich/_loop.py | 43 + .../vcs/pip/_vendor/rich/_null_file.py | 69 + .../vcs/pip/_vendor/rich/_palettes.py | 309 + src/fetchcode/vcs/pip/_vendor/rich/_pick.py | 17 + src/fetchcode/vcs/pip/_vendor/rich/_ratio.py | 159 + .../vcs/pip/_vendor/rich/_spinners.py | 482 + src/fetchcode/vcs/pip/_vendor/rich/_stack.py | 16 + src/fetchcode/vcs/pip/_vendor/rich/_timer.py | 19 + .../vcs/pip/_vendor/rich/_win32_console.py | 662 + .../vcs/pip/_vendor/rich/_windows.py | 71 + .../vcs/pip/_vendor/rich/_windows_renderer.py | 56 + src/fetchcode/vcs/pip/_vendor/rich/_wrap.py | 93 + src/fetchcode/vcs/pip/_vendor/rich/abc.py | 33 + src/fetchcode/vcs/pip/_vendor/rich/align.py | 311 + src/fetchcode/vcs/pip/_vendor/rich/ansi.py | 240 + src/fetchcode/vcs/pip/_vendor/rich/bar.py | 93 + src/fetchcode/vcs/pip/_vendor/rich/box.py | 480 + src/fetchcode/vcs/pip/_vendor/rich/cells.py | 167 + src/fetchcode/vcs/pip/_vendor/rich/color.py | 621 + .../vcs/pip/_vendor/rich/color_triplet.py | 38 + src/fetchcode/vcs/pip/_vendor/rich/columns.py | 187 + src/fetchcode/vcs/pip/_vendor/rich/console.py | 2633 ++++ .../vcs/pip/_vendor/rich/constrain.py | 37 + .../vcs/pip/_vendor/rich/containers.py | 167 + src/fetchcode/vcs/pip/_vendor/rich/control.py | 225 + .../vcs/pip/_vendor/rich/default_styles.py | 190 + .../vcs/pip/_vendor/rich/diagnose.py | 37 + src/fetchcode/vcs/pip/_vendor/rich/emoji.py | 96 + src/fetchcode/vcs/pip/_vendor/rich/errors.py | 34 + .../vcs/pip/_vendor/rich/file_proxy.py | 57 + .../vcs/pip/_vendor/rich/filesize.py | 89 + .../vcs/pip/_vendor/rich/highlighter.py | 232 + src/fetchcode/vcs/pip/_vendor/rich/json.py | 139 + src/fetchcode/vcs/pip/_vendor/rich/jupyter.py | 101 + src/fetchcode/vcs/pip/_vendor/rich/layout.py | 442 + src/fetchcode/vcs/pip/_vendor/rich/live.py | 375 + .../vcs/pip/_vendor/rich/live_render.py | 112 + src/fetchcode/vcs/pip/_vendor/rich/logging.py | 289 + src/fetchcode/vcs/pip/_vendor/rich/markup.py | 251 + src/fetchcode/vcs/pip/_vendor/rich/measure.py | 151 + src/fetchcode/vcs/pip/_vendor/rich/padding.py | 141 + src/fetchcode/vcs/pip/_vendor/rich/pager.py | 34 + src/fetchcode/vcs/pip/_vendor/rich/palette.py | 100 + src/fetchcode/vcs/pip/_vendor/rich/panel.py | 312 + src/fetchcode/vcs/pip/_vendor/rich/pretty.py | 995 ++ .../vcs/pip/_vendor/rich/progress.py | 1699 +++ .../vcs/pip/_vendor/rich/progress_bar.py | 223 + src/fetchcode/vcs/pip/_vendor/rich/prompt.py | 375 + .../vcs/pip/_vendor/rich/protocol.py | 42 + src/fetchcode/vcs/pip/_vendor/rich/py.typed | 0 src/fetchcode/vcs/pip/_vendor/rich/region.py | 10 + src/fetchcode/vcs/pip/_vendor/rich/repr.py | 149 + src/fetchcode/vcs/pip/_vendor/rich/rule.py | 130 + src/fetchcode/vcs/pip/_vendor/rich/scope.py | 86 + src/fetchcode/vcs/pip/_vendor/rich/screen.py | 54 + src/fetchcode/vcs/pip/_vendor/rich/segment.py | 738 + src/fetchcode/vcs/pip/_vendor/rich/spinner.py | 137 + src/fetchcode/vcs/pip/_vendor/rich/status.py | 131 + src/fetchcode/vcs/pip/_vendor/rich/style.py | 796 + src/fetchcode/vcs/pip/_vendor/rich/styled.py | 42 + src/fetchcode/vcs/pip/_vendor/rich/syntax.py | 958 ++ src/fetchcode/vcs/pip/_vendor/rich/table.py | 1000 ++ .../vcs/pip/_vendor/rich/terminal_theme.py | 153 + src/fetchcode/vcs/pip/_vendor/rich/text.py | 1357 ++ src/fetchcode/vcs/pip/_vendor/rich/theme.py | 115 + src/fetchcode/vcs/pip/_vendor/rich/themes.py | 5 + .../vcs/pip/_vendor/rich/traceback.py | 753 + src/fetchcode/vcs/pip/_vendor/rich/tree.py | 249 + src/fetchcode/vcs/pip/_vendor/six.LICENSE | 18 - src/fetchcode/vcs/pip/_vendor/six.py | 980 -- .../vcs/pip/_vendor/six/__init__.pyi | 1 - .../vcs/pip/_vendor/six/moves/__init__.pyi | 1 - .../pip/_vendor/six/moves/configparser.pyi | 1 - src/fetchcode/vcs/pip/_vendor/toml.py | 1039 -- src/fetchcode/vcs/pip/_vendor/toml.pyi | 1 - .../vcs/pip/_vendor/toml/__init__.py | 21 - src/fetchcode/vcs/pip/_vendor/toml/decoder.py | 945 -- src/fetchcode/vcs/pip/_vendor/toml/encoder.py | 250 - src/fetchcode/vcs/pip/_vendor/toml/ordered.py | 15 - src/fetchcode/vcs/pip/_vendor/toml/tz.py | 21 - src/fetchcode/vcs/pip/_vendor/tomli/LICENSE | 21 + .../vcs/pip/_vendor/tomli/__init__.py | 11 + .../vcs/pip/_vendor/tomli/_parser.py | 691 + src/fetchcode/vcs/pip/_vendor/tomli/_re.py | 107 + src/fetchcode/vcs/pip/_vendor/tomli/_types.py | 10 + src/fetchcode/vcs/pip/_vendor/tomli/py.typed | 1 + .../pip/_vendor/{toml => truststore}/LICENSE | 11 +- .../vcs/pip/_vendor/truststore/__init__.py | 13 + .../vcs/pip/_vendor/truststore/_api.py | 313 + .../vcs/pip/_vendor/truststore/_macos.py | 499 + .../vcs/pip/_vendor/truststore/_openssl.py | 66 + .../pip/_vendor/truststore/_ssl_constants.py | 31 + .../vcs/pip/_vendor/truststore/_windows.py | 564 + .../vcs/pip/_vendor/truststore/py.typed | 0 .../vcs/pip/_vendor/typing_extensions.LICENSE | 279 + .../vcs/pip/_vendor/typing_extensions.py | 3641 +++++ src/fetchcode/vcs/pip/_vendor/urllib3.pyi | 1 - .../vcs/pip/_vendor/urllib3/LICENSE.txt | 2 +- .../vcs/pip/_vendor/urllib3/__init__.py | 36 +- .../vcs/pip/_vendor/urllib3/_collections.py | 27 +- .../vcs/pip/_vendor/urllib3/_version.py | 2 + .../vcs/pip/_vendor/urllib3/connection.py | 284 +- .../vcs/pip/_vendor/urllib3/connectionpool.py | 286 +- .../contrib/_securetransport/bindings.py | 62 +- .../contrib/_securetransport/low_level.py | 75 +- .../pip/_vendor/urllib3/contrib/appengine.py | 18 +- .../pip/_vendor/urllib3/contrib/ntlmpool.py | 13 +- .../pip/_vendor/urllib3/contrib/pyopenssl.py | 66 +- .../urllib3/contrib/securetransport.py | 91 +- .../vcs/pip/_vendor/urllib3/contrib/socks.py | 22 +- .../vcs/pip/_vendor/urllib3/exceptions.py | 144 +- .../vcs/pip/_vendor/urllib3/fields.py | 5 +- .../vcs/pip/_vendor/urllib3/filepost.py | 4 +- .../pip/_vendor/urllib3/packages/__init__.py | 5 - .../urllib3/packages/backports/makefile.py | 1 - .../packages/backports/weakref_finalize.py | 155 + .../vcs/pip/_vendor/urllib3/packages/six.py | 125 +- .../packages/ssl_match_hostname/__init__.py | 19 - .../vcs/pip/_vendor/urllib3/poolmanager.py | 116 +- .../vcs/pip/_vendor/urllib3/request.py | 30 +- .../vcs/pip/_vendor/urllib3/response.py | 152 +- .../vcs/pip/_vendor/urllib3/util/__init__.py | 17 +- .../pip/_vendor/urllib3/util/connection.py | 19 +- .../vcs/pip/_vendor/urllib3/util/proxy.py | 57 + .../vcs/pip/_vendor/urllib3/util/queue.py | 1 + .../vcs/pip/_vendor/urllib3/util/request.py | 16 +- .../vcs/pip/_vendor/urllib3/util/response.py | 31 +- .../vcs/pip/_vendor/urllib3/util/retry.py | 246 +- .../vcs/pip/_vendor/urllib3/util/ssl_.py | 170 +- .../ssl_match_hostname.py} | 17 +- .../pip/_vendor/urllib3/util/ssltransport.py | 221 + .../vcs/pip/_vendor/urllib3/util/timeout.py | 61 +- .../vcs/pip/_vendor/urllib3/util/url.py | 23 +- .../vcs/pip/_vendor/urllib3/util/wait.py | 7 +- src/fetchcode/vcs/pip/_vendor/vendor.txt | 42 +- .../vcs/pip/_vendor/webencodings.pyi | 1 - .../vcs/pip/_vendor/webencodings/__init__.py | 342 - .../vcs/pip/_vendor/webencodings/labels.py | 231 - .../vcs/pip/_vendor/webencodings/mklabels.py | 59 - .../vcs/pip/_vendor/webencodings/tests.py | 153 - .../_vendor/webencodings/x_user_defined.py | 325 - src/fetchcode/vcs/pip/py.typed | 4 + 597 files changed, 86106 insertions(+), 68963 deletions(-) create mode 100644 src/fetchcode/vcs/pip/__pip-runner__.py create mode 100644 src/fetchcode/vcs/pip/_internal/cli/index_command.py create mode 100644 src/fetchcode/vcs/pip/_internal/commands/index.py create mode 100644 src/fetchcode/vcs/pip/_internal/commands/inspect.py create mode 100644 src/fetchcode/vcs/pip/_internal/index/sources.py delete mode 100644 src/fetchcode/vcs/pip/_internal/locations.py create mode 100644 src/fetchcode/vcs/pip/_internal/locations/__init__.py create mode 100644 src/fetchcode/vcs/pip/_internal/locations/_distutils.py create mode 100644 src/fetchcode/vcs/pip/_internal/locations/_sysconfig.py create mode 100644 src/fetchcode/vcs/pip/_internal/locations/base.py create mode 100644 src/fetchcode/vcs/pip/_internal/metadata/__init__.py create mode 100644 src/fetchcode/vcs/pip/_internal/metadata/_json.py create mode 100644 src/fetchcode/vcs/pip/_internal/metadata/base.py create mode 100644 src/fetchcode/vcs/pip/_internal/metadata/importlib/__init__.py create mode 100644 src/fetchcode/vcs/pip/_internal/metadata/importlib/_compat.py create mode 100644 src/fetchcode/vcs/pip/_internal/metadata/importlib/_dists.py create mode 100644 src/fetchcode/vcs/pip/_internal/metadata/importlib/_envs.py create mode 100644 src/fetchcode/vcs/pip/_internal/metadata/pkg_resources.py create mode 100644 src/fetchcode/vcs/pip/_internal/models/installation_report.py create mode 100644 src/fetchcode/vcs/pip/_internal/network/lazy_wheel.py rename src/fetchcode/vcs/pip/{_vendor/html5lib/filters => _internal/operations/build}/__init__.py (100%) create mode 100644 src/fetchcode/vcs/pip/_internal/operations/build/build_tracker.py create mode 100644 src/fetchcode/vcs/pip/_internal/operations/build/metadata.py create mode 100644 src/fetchcode/vcs/pip/_internal/operations/build/metadata_editable.py create mode 100644 src/fetchcode/vcs/pip/_internal/operations/build/metadata_legacy.py create mode 100644 src/fetchcode/vcs/pip/_internal/operations/build/wheel.py create mode 100644 src/fetchcode/vcs/pip/_internal/operations/build/wheel_editable.py create mode 100644 src/fetchcode/vcs/pip/_internal/operations/build/wheel_legacy.py delete mode 100644 src/fetchcode/vcs/pip/_internal/operations/install/legacy.py delete mode 100644 src/fetchcode/vcs/pip/_internal/req/req_tracker.py create mode 100644 src/fetchcode/vcs/pip/_internal/resolution/resolvelib/found_candidates.py create mode 100644 src/fetchcode/vcs/pip/_internal/resolution/resolvelib/reporter.py create mode 100644 src/fetchcode/vcs/pip/_internal/utils/_jaraco_text.py create mode 100644 src/fetchcode/vcs/pip/_internal/utils/_log.py create mode 100644 src/fetchcode/vcs/pip/_internal/utils/datetime.py delete mode 100644 src/fetchcode/vcs/pip/_internal/utils/distutils_args.py create mode 100644 src/fetchcode/vcs/pip/_internal/utils/egg_link.py delete mode 100644 src/fetchcode/vcs/pip/_internal/utils/inject_securetransport.py delete mode 100644 src/fetchcode/vcs/pip/_internal/utils/models.py delete mode 100644 src/fetchcode/vcs/pip/_internal/utils/pkg_resources.py create mode 100644 src/fetchcode/vcs/pip/_internal/utils/retry.py delete mode 100644 src/fetchcode/vcs/pip/_internal/utils/typing.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/appdirs.LICENSE.txt delete mode 100644 src/fetchcode/vcs/pip/_vendor/appdirs.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/cachecontrol.pyi delete mode 100644 src/fetchcode/vcs/pip/_vendor/cachecontrol/compat.py create mode 100644 src/fetchcode/vcs/pip/_vendor/cachecontrol/py.typed delete mode 100644 src/fetchcode/vcs/pip/_vendor/certifi.pyi create mode 100644 src/fetchcode/vcs/pip/_vendor/certifi/py.typed delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet.pyi delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/LICENSE delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/__init__.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/big5freq.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/big5prober.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/chardistribution.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/charsetgroupprober.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/charsetprober.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/cli/__init__.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/cli/chardetect.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/codingstatemachine.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/compat.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/cp949prober.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/enums.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/escprober.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/escsm.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/eucjpprober.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/euckrfreq.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/euckrprober.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/euctwfreq.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/euctwprober.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/gb2312freq.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/gb2312prober.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/hebrewprober.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/jisfreq.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/jpcntx.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/langbulgarianmodel.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/langcyrillicmodel.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/langgreekmodel.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/langhebrewmodel.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/langhungarianmodel.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/langthaimodel.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/langturkishmodel.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/latin1prober.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/mbcharsetprober.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/mbcsgroupprober.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/mbcssm.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/sbcharsetprober.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/sbcsgroupprober.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/sjisprober.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/universaldetector.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/utf8prober.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/chardet/version.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/colorama.pyi delete mode 100644 src/fetchcode/vcs/pip/_vendor/colorama/LICENSE.txt delete mode 100644 src/fetchcode/vcs/pip/_vendor/colorama/__init__.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/colorama/ansi.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/colorama/ansitowin32.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/colorama/initialise.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/colorama/win32.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/colorama/winterm.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/contextlib2.LICENSE.txt delete mode 100644 src/fetchcode/vcs/pip/_vendor/contextlib2.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/distlib.pyi delete mode 100644 src/fetchcode/vcs/pip/_vendor/distlib/_backport/__init__.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/distlib/_backport/misc.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/distlib/_backport/shutil.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/distlib/_backport/sysconfig.cfg delete mode 100644 src/fetchcode/vcs/pip/_vendor/distlib/_backport/sysconfig.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/distlib/_backport/tarfile.py create mode 100644 src/fetchcode/vcs/pip/_vendor/distlib/t64-arm.exe create mode 100644 src/fetchcode/vcs/pip/_vendor/distlib/w64-arm.exe delete mode 100644 src/fetchcode/vcs/pip/_vendor/distro.pyi rename src/fetchcode/vcs/pip/_vendor/{distro.LICENSE => distro/LICENSE} (100%) create mode 100644 src/fetchcode/vcs/pip/_vendor/distro/__init__.py create mode 100644 src/fetchcode/vcs/pip/_vendor/distro/__main__.py rename src/fetchcode/vcs/pip/_vendor/{ => distro}/distro.py (64%) create mode 100644 src/fetchcode/vcs/pip/_vendor/distro/py.typed delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib.pyi delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/LICENSE delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/__init__.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/_ihatexml.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/_inputstream.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/_tokenizer.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/_trie/__init__.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/_trie/_base.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/_trie/datrie.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/_trie/py.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/_utils.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/constants.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/filters/alphabeticalattributes.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/filters/base.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/filters/inject_meta_charset.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/filters/lint.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/filters/optionaltags.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/filters/sanitizer.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/filters/whitespace.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/html5parser.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/serializer.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/treeadapters/__init__.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/treeadapters/genshi.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/treeadapters/sax.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/treebuilders/__init__.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/treebuilders/base.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/treebuilders/dom.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/treebuilders/etree.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/treebuilders/etree_lxml.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/treewalkers/__init__.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/treewalkers/base.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/treewalkers/dom.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/treewalkers/etree.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/treewalkers/etree_lxml.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/html5lib/treewalkers/genshi.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/idna.pyi create mode 100644 src/fetchcode/vcs/pip/_vendor/idna/LICENSE.md delete mode 100644 src/fetchcode/vcs/pip/_vendor/idna/LICENSE.rst create mode 100644 src/fetchcode/vcs/pip/_vendor/idna/py.typed delete mode 100644 src/fetchcode/vcs/pip/_vendor/ipaddress.LICENSE delete mode 100644 src/fetchcode/vcs/pip/_vendor/ipaddress.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/ipaddress.pyi delete mode 100644 src/fetchcode/vcs/pip/_vendor/msgpack.pyi delete mode 100644 src/fetchcode/vcs/pip/_vendor/msgpack/_version.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/packaging.pyi delete mode 100644 src/fetchcode/vcs/pip/_vendor/packaging/__about__.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/packaging/_compat.py create mode 100644 src/fetchcode/vcs/pip/_vendor/packaging/_elffile.py create mode 100644 src/fetchcode/vcs/pip/_vendor/packaging/_manylinux.py create mode 100644 src/fetchcode/vcs/pip/_vendor/packaging/_musllinux.py create mode 100644 src/fetchcode/vcs/pip/_vendor/packaging/_parser.py create mode 100644 src/fetchcode/vcs/pip/_vendor/packaging/_tokenizer.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/packaging/_typing.py create mode 100644 src/fetchcode/vcs/pip/_vendor/packaging/metadata.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/pep517.pyi delete mode 100644 src/fetchcode/vcs/pip/_vendor/pep517/__init__.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/pep517/build.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/pep517/check.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/pep517/colorlog.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/pep517/compat.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/pep517/dirtools.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/pep517/envbuild.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/pep517/meta.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/pep517/wrappers.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/pkg_resources.pyi delete mode 100644 src/fetchcode/vcs/pip/_vendor/pkg_resources/py31compat.py create mode 100644 src/fetchcode/vcs/pip/_vendor/platformdirs/LICENSE create mode 100644 src/fetchcode/vcs/pip/_vendor/platformdirs/__init__.py create mode 100644 src/fetchcode/vcs/pip/_vendor/platformdirs/__main__.py create mode 100644 src/fetchcode/vcs/pip/_vendor/platformdirs/android.py create mode 100644 src/fetchcode/vcs/pip/_vendor/platformdirs/api.py create mode 100644 src/fetchcode/vcs/pip/_vendor/platformdirs/macos.py create mode 100644 src/fetchcode/vcs/pip/_vendor/platformdirs/py.typed create mode 100644 src/fetchcode/vcs/pip/_vendor/platformdirs/unix.py create mode 100644 src/fetchcode/vcs/pip/_vendor/platformdirs/version.py create mode 100644 src/fetchcode/vcs/pip/_vendor/platformdirs/windows.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/progress.pyi delete mode 100644 src/fetchcode/vcs/pip/_vendor/progress/LICENSE delete mode 100644 src/fetchcode/vcs/pip/_vendor/progress/__init__.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/progress/bar.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/progress/counter.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/progress/spinner.py rename src/fetchcode/vcs/pip/_vendor/{webencodings => pygments}/LICENSE (60%) create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/__init__.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/__main__.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/cmdline.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/console.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/filter.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/filters/__init__.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/formatter.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/formatters/__init__.py create mode 100755 src/fetchcode/vcs/pip/_vendor/pygments/formatters/_mapping.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/formatters/bbcode.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/formatters/groff.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/formatters/html.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/formatters/img.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/formatters/irc.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/formatters/latex.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/formatters/other.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/formatters/pangomarkup.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/formatters/rtf.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/formatters/svg.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/formatters/terminal.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/formatters/terminal256.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/lexer.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/lexers/__init__.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/lexers/_mapping.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/lexers/python.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/modeline.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/plugin.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/regexopt.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/scanner.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/sphinxext.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/style.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/styles/__init__.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/styles/_mapping.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/token.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/unistring.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pygments/util.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/pyparsing.LICENSE delete mode 100644 src/fetchcode/vcs/pip/_vendor/pyparsing.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/pyparsing.pyi rename src/fetchcode/vcs/pip/_vendor/{pep517 => pyproject_hooks}/LICENSE (100%) create mode 100644 src/fetchcode/vcs/pip/_vendor/pyproject_hooks/__init__.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pyproject_hooks/_compat.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pyproject_hooks/_impl.py create mode 100644 src/fetchcode/vcs/pip/_vendor/pyproject_hooks/_in_process/__init__.py rename src/fetchcode/vcs/pip/_vendor/{pep517 => pyproject_hooks/_in_process}/_in_process.py (68%) delete mode 100644 src/fetchcode/vcs/pip/_vendor/requests.pyi delete mode 100644 src/fetchcode/vcs/pip/_vendor/resolvelib.pyi create mode 100644 src/fetchcode/vcs/pip/_vendor/resolvelib/compat/__init__.py create mode 100644 src/fetchcode/vcs/pip/_vendor/resolvelib/compat/collections_abc.py create mode 100644 src/fetchcode/vcs/pip/_vendor/resolvelib/py.typed delete mode 100644 src/fetchcode/vcs/pip/_vendor/retrying.LICENSE delete mode 100644 src/fetchcode/vcs/pip/_vendor/retrying.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/retrying.pyi create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/LICENSE create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/__init__.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/__main__.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/_cell_widths.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/_emoji_codes.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/_emoji_replace.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/_export_format.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/_extension.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/_fileno.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/_inspect.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/_log_render.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/_loop.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/_null_file.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/_palettes.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/_pick.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/_ratio.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/_spinners.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/_stack.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/_timer.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/_win32_console.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/_windows.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/_windows_renderer.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/_wrap.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/abc.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/align.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/ansi.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/bar.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/box.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/cells.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/color.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/color_triplet.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/columns.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/console.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/constrain.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/containers.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/control.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/default_styles.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/diagnose.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/emoji.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/errors.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/file_proxy.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/filesize.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/highlighter.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/json.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/jupyter.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/layout.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/live.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/live_render.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/logging.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/markup.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/measure.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/padding.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/pager.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/palette.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/panel.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/pretty.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/progress.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/progress_bar.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/prompt.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/protocol.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/py.typed create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/region.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/repr.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/rule.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/scope.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/screen.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/segment.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/spinner.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/status.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/style.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/styled.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/syntax.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/table.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/terminal_theme.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/text.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/theme.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/themes.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/traceback.py create mode 100644 src/fetchcode/vcs/pip/_vendor/rich/tree.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/six.LICENSE delete mode 100644 src/fetchcode/vcs/pip/_vendor/six.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/six/__init__.pyi delete mode 100644 src/fetchcode/vcs/pip/_vendor/six/moves/__init__.pyi delete mode 100644 src/fetchcode/vcs/pip/_vendor/six/moves/configparser.pyi delete mode 100644 src/fetchcode/vcs/pip/_vendor/toml.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/toml.pyi delete mode 100644 src/fetchcode/vcs/pip/_vendor/toml/__init__.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/toml/decoder.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/toml/encoder.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/toml/ordered.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/toml/tz.py create mode 100644 src/fetchcode/vcs/pip/_vendor/tomli/LICENSE create mode 100644 src/fetchcode/vcs/pip/_vendor/tomli/__init__.py create mode 100644 src/fetchcode/vcs/pip/_vendor/tomli/_parser.py create mode 100644 src/fetchcode/vcs/pip/_vendor/tomli/_re.py create mode 100644 src/fetchcode/vcs/pip/_vendor/tomli/_types.py create mode 100644 src/fetchcode/vcs/pip/_vendor/tomli/py.typed rename src/fetchcode/vcs/pip/_vendor/{toml => truststore}/LICENSE (82%) create mode 100644 src/fetchcode/vcs/pip/_vendor/truststore/__init__.py create mode 100644 src/fetchcode/vcs/pip/_vendor/truststore/_api.py create mode 100644 src/fetchcode/vcs/pip/_vendor/truststore/_macos.py create mode 100644 src/fetchcode/vcs/pip/_vendor/truststore/_openssl.py create mode 100644 src/fetchcode/vcs/pip/_vendor/truststore/_ssl_constants.py create mode 100644 src/fetchcode/vcs/pip/_vendor/truststore/_windows.py create mode 100644 src/fetchcode/vcs/pip/_vendor/truststore/py.typed create mode 100644 src/fetchcode/vcs/pip/_vendor/typing_extensions.LICENSE create mode 100644 src/fetchcode/vcs/pip/_vendor/typing_extensions.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/urllib3.pyi create mode 100644 src/fetchcode/vcs/pip/_vendor/urllib3/_version.py create mode 100644 src/fetchcode/vcs/pip/_vendor/urllib3/packages/backports/weakref_finalize.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/urllib3/packages/ssl_match_hostname/__init__.py create mode 100644 src/fetchcode/vcs/pip/_vendor/urllib3/util/proxy.py rename src/fetchcode/vcs/pip/_vendor/urllib3/{packages/ssl_match_hostname/_implementation.py => util/ssl_match_hostname.py} (92%) create mode 100644 src/fetchcode/vcs/pip/_vendor/urllib3/util/ssltransport.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/webencodings.pyi delete mode 100644 src/fetchcode/vcs/pip/_vendor/webencodings/__init__.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/webencodings/labels.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/webencodings/mklabels.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/webencodings/tests.py delete mode 100644 src/fetchcode/vcs/pip/_vendor/webencodings/x_user_defined.py create mode 100644 src/fetchcode/vcs/pip/py.typed diff --git a/src/fetchcode/vcs/git.py b/src/fetchcode/vcs/git.py index f602516c..f6a8e862 100644 --- a/src/fetchcode/vcs/git.py +++ b/src/fetchcode/vcs/git.py @@ -30,6 +30,9 @@ def fetch_via_git(url): domain = parsed_url.netloc temp = tempfile.mkdtemp() os.rmdir(temp) + print(url) + print(scheme) + print(Git.schemes) if scheme not in Git.schemes: raise Exception("Not a Git based scheme.") diff --git a/src/fetchcode/vcs/pip.ABOUT b/src/fetchcode/vcs/pip.ABOUT index ba8dfe7d..956440dd 100644 --- a/src/fetchcode/vcs/pip.ABOUT +++ b/src/fetchcode/vcs/pip.ABOUT @@ -4,14 +4,14 @@ copyright: Copyright (c) The Python Software Foundation and the pip authors (see pip-AUTHORS.txt file for a list of pip authors) description: A tool for installing and managing Python packages. notes: this is a copy of pip vendored for stability. -download_url: https://files.pythonhosted.org/packages/08/25/f204a6138dade2f6757b4ae99bc3994aac28a5602c97ddb2a35e0e22fbc4/pip-20.1.1.tar.gz +download_url: https://files.pythonhosted.org/packages/4d/87/fb90046e096a03aeab235e139436b3fe804cdd447ed2093b0d70eba3f7f8/pip-24.2.tar.gz homepage_url: https://pip.pypa.io license_expression: mit AND lgpl-2.1-plus AND python AND mit AND bsd-new AND (bsd-new OR apache-2.0) AND apache-2.0 AND isc name: pip -package_url: pkg:pypi/pip@20.1.1 +package_url: pkg:pypi/pip@24.2 primary_language: Python redistribute: yes track_changes: yes type: pypi -version: 20.1.1 +version: 24.2 diff --git a/src/fetchcode/vcs/pip/__init__.py b/src/fetchcode/vcs/pip/__init__.py index a78bbce8..640e922f 100644 --- a/src/fetchcode/vcs/pip/__init__.py +++ b/src/fetchcode/vcs/pip/__init__.py @@ -1,18 +1,13 @@ -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from typing import List, Optional -if MYPY_CHECK_RUNNING: - from typing import List, Optional +__version__ = "24.2" -__version__ = "20.1.1" - - -def main(args=None): - # type: (Optional[List[str]]) -> int +def main(args: Optional[List[str]] = None) -> int: """This is an internal API only meant for use by pip's own console scripts. For additional details, see https://github.com/pypa/pip/issues/7498. """ - from fetchcode.vcs.pip._internal.utils.entrypoints import _wrapper + from pip._internal.utils.entrypoints import _wrapper return _wrapper(args) diff --git a/src/fetchcode/vcs/pip/__main__.py b/src/fetchcode/vcs/pip/__main__.py index 8bdcf48b..59913261 100644 --- a/src/fetchcode/vcs/pip/__main__.py +++ b/src/fetchcode/vcs/pip/__main__.py @@ -1,5 +1,3 @@ -from __future__ import absolute_import - import os import sys @@ -7,12 +5,12 @@ # of sys.path, if present to avoid using current directory # in pip commands check, freeze, install, list and show, # when invoked as python -m pip -if sys.path[0] in ('', os.getcwd()): +if sys.path[0] in ("", os.getcwd()): sys.path.pop(0) # If we are running from a wheel, add the wheel to sys.path # This allows the usage python pip-*.whl/pip install pip-*.whl -if __package__ == '': +if __package__ == "": # __file__ is pip-*.whl/pip/__main__.py # first dirname call strips of '/__main__.py', second strips off '/pip' # Resulting path is the name of the wheel itself @@ -20,7 +18,7 @@ path = os.path.dirname(os.path.dirname(__file__)) sys.path.insert(0, path) -from fetchcode.vcs.pip._internal.cli.main import main as _main # isort:skip # noqa +if __name__ == "__main__": + from pip._internal.cli.main import main as _main -if __name__ == '__main__': sys.exit(_main()) diff --git a/src/fetchcode/vcs/pip/__pip-runner__.py b/src/fetchcode/vcs/pip/__pip-runner__.py new file mode 100644 index 00000000..c633787f --- /dev/null +++ b/src/fetchcode/vcs/pip/__pip-runner__.py @@ -0,0 +1,50 @@ +"""Execute exactly this copy of pip, within a different environment. + +This file is named as it is, to ensure that this module can't be imported via +an import statement. +""" + +# /!\ This version compatibility check section must be Python 2 compatible. /!\ + +import sys + +# Copied from pyproject.toml +PYTHON_REQUIRES = (3, 8) + + +def version_str(version): # type: ignore + return ".".join(str(v) for v in version) + + +if sys.version_info[:2] < PYTHON_REQUIRES: + raise SystemExit( + "This version of pip does not support python {} (requires >={}).".format( + version_str(sys.version_info[:2]), version_str(PYTHON_REQUIRES) + ) + ) + +# From here on, we can use Python 3 features, but the syntax must remain +# Python 2 compatible. + +import runpy # noqa: E402 +from importlib.machinery import PathFinder # noqa: E402 +from os.path import dirname # noqa: E402 + +PIP_SOURCES_ROOT = dirname(dirname(__file__)) + + +class PipImportRedirectingFinder: + @classmethod + def find_spec(self, fullname, path=None, target=None): # type: ignore + if fullname != "pip": + return None + + spec = PathFinder.find_spec(fullname, [PIP_SOURCES_ROOT], target) + assert spec, (PIP_SOURCES_ROOT, fullname) + return spec + + +sys.meta_path.insert(0, PipImportRedirectingFinder()) + +assert __name__ == "__main__", "Cannot run __pip-runner__.py as a non-main module" +runpy.run_module("pip", run_name="__main__", alter_sys=True) diff --git a/src/fetchcode/vcs/pip/_internal/__init__.py b/src/fetchcode/vcs/pip/_internal/__init__.py index cadc261f..1a5b7f87 100755 --- a/src/fetchcode/vcs/pip/_internal/__init__.py +++ b/src/fetchcode/vcs/pip/_internal/__init__.py @@ -1,17 +1,18 @@ -import fetchcode.vcs.pip._internal.utils.inject_securetransport # noqa -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from typing import List, Optional -if MYPY_CHECK_RUNNING: - from typing import Optional, List +from pip._internal.utils import _log +# init_logging() must be called before any call to logging.getLogger() +# which happens at import of most modules. +_log.init_logging() -def main(args=None): - # type: (Optional[List[str]]) -> int + +def main(args: Optional[List[str]] = None) -> int: """This is preserved for old console scripts that may still be referencing it. For additional details, see https://github.com/pypa/pip/issues/7498. """ - from fetchcode.vcs.pip._internal.utils.entrypoints import _wrapper + from pip._internal.utils.entrypoints import _wrapper return _wrapper(args) diff --git a/src/fetchcode/vcs/pip/_internal/build_env.py b/src/fetchcode/vcs/pip/_internal/build_env.py index e777f9a5..be1e0ca8 100644 --- a/src/fetchcode/vcs/pip/_internal/build_env.py +++ b/src/fetchcode/vcs/pip/_internal/build_env.py @@ -1,70 +1,96 @@ """Build Environment used for isolation during sdist building """ -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False -# mypy: disallow-untyped-defs=False - import logging import os +import pathlib +import site import sys import textwrap from collections import OrderedDict -from distutils.sysconfig import get_python_lib -from sysconfig import get_paths +from types import TracebackType +from typing import TYPE_CHECKING, Iterable, List, Optional, Set, Tuple, Type, Union -from fetchcode.vcs.pip._vendor.pkg_resources import Requirement, VersionConflict, WorkingSet +from pip._vendor.certifi import where +from pip._vendor.packaging.version import Version from pip import __file__ as pip_location -from fetchcode.vcs.pip._internal.cli.spinners import open_spinner -from fetchcode.vcs.pip._internal.utils.subprocess import call_subprocess -from fetchcode.vcs.pip._internal.utils.temp_dir import TempDirectory, tempdir_kinds -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._internal.cli.spinners import open_spinner +from pip._internal.locations import get_platlib, get_purelib, get_scheme +from pip._internal.metadata import get_default_environment, get_environment +from pip._internal.utils.logging import VERBOSE +from pip._internal.utils.packaging import get_requirement +from pip._internal.utils.subprocess import call_subprocess +from pip._internal.utils.temp_dir import TempDirectory, tempdir_kinds -if MYPY_CHECK_RUNNING: - from typing import Tuple, Set, Iterable, Optional, List - from fetchcode.vcs.pip._internal.index.package_finder import PackageFinder +if TYPE_CHECKING: + from pip._internal.index.package_finder import PackageFinder logger = logging.getLogger(__name__) -class _Prefix: +def _dedup(a: str, b: str) -> Union[Tuple[str], Tuple[str, str]]: + return (a, b) if a != b else (a,) + - def __init__(self, path): - # type: (str) -> None +class _Prefix: + def __init__(self, path: str) -> None: self.path = path self.setup = False - self.bin_dir = get_paths( - 'nt' if os.name == 'nt' else 'posix_prefix', - vars={'base': path, 'platbase': path} - )['scripts'] - # Note: prefer distutils' sysconfig to get the - # library paths so PyPy is correctly supported. - purelib = get_python_lib(plat_specific=False, prefix=path) - platlib = get_python_lib(plat_specific=True, prefix=path) - if purelib == platlib: - self.lib_dirs = [purelib] - else: - self.lib_dirs = [purelib, platlib] + scheme = get_scheme("", prefix=path) + self.bin_dir = scheme.scripts + self.lib_dirs = _dedup(scheme.purelib, scheme.platlib) -class BuildEnvironment(object): - """Creates and manages an isolated environment to install build deps +def get_runnable_pip() -> str: + """Get a file to pass to a Python executable, to run the currently-running pip. + + This is used to run a pip subprocess, for installing requirements into the build + environment. """ + source = pathlib.Path(pip_location).resolve().parent + + if not source.is_dir(): + # This would happen if someone is using pip from inside a zip file. In that + # case, we can use that directly. + return str(source) + + return os.fsdecode(source / "__pip-runner__.py") - def __init__(self): - # type: () -> None - temp_dir = TempDirectory( - kind=tempdir_kinds.BUILD_ENV, globally_managed=True - ) - self._prefixes = OrderedDict(( +def _get_system_sitepackages() -> Set[str]: + """Get system site packages + + Usually from site.getsitepackages, + but fallback on `get_purelib()/get_platlib()` if unavailable + (e.g. in a virtualenv created by virtualenv<20) + + Returns normalized set of strings. + """ + if hasattr(site, "getsitepackages"): + system_sites = site.getsitepackages() + else: + # virtualenv < 20 overwrites site.py without getsitepackages + # fallback on get_purelib/get_platlib. + # this is known to miss things, but shouldn't in the cases + # where getsitepackages() has been removed (inside a virtualenv) + system_sites = [get_purelib(), get_platlib()] + return {os.path.normcase(path) for path in system_sites} + + +class BuildEnvironment: + """Creates and manages an isolated environment to install build deps""" + + def __init__(self) -> None: + temp_dir = TempDirectory(kind=tempdir_kinds.BUILD_ENV, globally_managed=True) + + self._prefixes = OrderedDict( (name, _Prefix(os.path.join(temp_dir.path, name))) - for name in ('normal', 'overlay') - )) + for name in ("normal", "overlay") + ) - self._bin_dirs = [] # type: List[str] - self._lib_dirs = [] # type: List[str] + self._bin_dirs: List[str] = [] + self._lib_dirs: List[str] = [] for prefix in reversed(list(self._prefixes.values())): self._bin_dirs.append(prefix.bin_dir) self._lib_dirs.extend(prefix.lib_dirs) @@ -72,18 +98,17 @@ def __init__(self): # Customize site to: # - ensure .pth files are honored # - prevent access to system site packages - system_sites = { - os.path.normcase(site) for site in ( - get_python_lib(plat_specific=False), - get_python_lib(plat_specific=True), - ) - } - self._site_dir = os.path.join(temp_dir.path, 'site') + system_sites = _get_system_sitepackages() + + self._site_dir = os.path.join(temp_dir.path, "site") if not os.path.exists(self._site_dir): os.mkdir(self._site_dir) - with open(os.path.join(self._site_dir, 'sitecustomize.py'), 'w') as fp: - fp.write(textwrap.dedent( - ''' + with open( + os.path.join(self._site_dir, "sitecustomize.py"), "w", encoding="utf-8" + ) as fp: + fp.write( + textwrap.dedent( + """ import os, site, sys # First, drop system-sites related paths. @@ -106,114 +131,185 @@ def __init__(self): for path in {lib_dirs!r}: assert not path in sys.path site.addsitedir(path) - ''' - ).format(system_sites=system_sites, lib_dirs=self._lib_dirs)) + """ + ).format(system_sites=system_sites, lib_dirs=self._lib_dirs) + ) - def __enter__(self): + def __enter__(self) -> None: self._save_env = { name: os.environ.get(name, None) - for name in ('PATH', 'PYTHONNOUSERSITE', 'PYTHONPATH') + for name in ("PATH", "PYTHONNOUSERSITE", "PYTHONPATH") } path = self._bin_dirs[:] - old_path = self._save_env['PATH'] + old_path = self._save_env["PATH"] if old_path: path.extend(old_path.split(os.pathsep)) pythonpath = [self._site_dir] - os.environ.update({ - 'PATH': os.pathsep.join(path), - 'PYTHONNOUSERSITE': '1', - 'PYTHONPATH': os.pathsep.join(pythonpath), - }) + os.environ.update( + { + "PATH": os.pathsep.join(path), + "PYTHONNOUSERSITE": "1", + "PYTHONPATH": os.pathsep.join(pythonpath), + } + ) - def __exit__(self, exc_type, exc_val, exc_tb): + def __exit__( + self, + exc_type: Optional[Type[BaseException]], + exc_val: Optional[BaseException], + exc_tb: Optional[TracebackType], + ) -> None: for varname, old_value in self._save_env.items(): if old_value is None: os.environ.pop(varname, None) else: os.environ[varname] = old_value - def check_requirements(self, reqs): - # type: (Iterable[str]) -> Tuple[Set[Tuple[str, str]], Set[str]] + def check_requirements( + self, reqs: Iterable[str] + ) -> Tuple[Set[Tuple[str, str]], Set[str]]: """Return 2 sets: - - conflicting requirements: set of (installed, wanted) reqs tuples - - missing requirements: set of reqs + - conflicting requirements: set of (installed, wanted) reqs tuples + - missing requirements: set of reqs """ missing = set() conflicting = set() if reqs: - ws = WorkingSet(self._lib_dirs) - for req in reqs: - try: - if ws.find(Requirement.parse(req)) is None: - missing.add(req) - except VersionConflict as e: - conflicting.add((str(e.args[0].as_requirement()), - str(e.args[1]))) + env = ( + get_environment(self._lib_dirs) + if hasattr(self, "_lib_dirs") + else get_default_environment() + ) + for req_str in reqs: + req = get_requirement(req_str) + # We're explicitly evaluating with an empty extra value, since build + # environments are not provided any mechanism to select specific extras. + if req.marker is not None and not req.marker.evaluate({"extra": ""}): + continue + dist = env.get_distribution(req.name) + if not dist: + missing.add(req_str) + continue + if isinstance(dist.version, Version): + installed_req_str = f"{req.name}=={dist.version}" + else: + installed_req_str = f"{req.name}==={dist.version}" + if not req.specifier.contains(dist.version, prereleases=True): + conflicting.add((installed_req_str, req_str)) + # FIXME: Consider direct URL? return conflicting, missing def install_requirements( self, - finder, # type: PackageFinder - requirements, # type: Iterable[str] - prefix_as_string, # type: str - message # type: Optional[str] - ): - # type: (...) -> None + finder: "PackageFinder", + requirements: Iterable[str], + prefix_as_string: str, + *, + kind: str, + ) -> None: prefix = self._prefixes[prefix_as_string] assert not prefix.setup prefix.setup = True if not requirements: return - args = [ - sys.executable, os.path.dirname(pip_location), 'install', - '--ignore-installed', '--no-user', '--prefix', prefix.path, - '--no-warn-script-location', - ] # type: List[str] + self._install_requirements( + get_runnable_pip(), + finder, + requirements, + prefix, + kind=kind, + ) + + @staticmethod + def _install_requirements( + pip_runnable: str, + finder: "PackageFinder", + requirements: Iterable[str], + prefix: _Prefix, + *, + kind: str, + ) -> None: + args: List[str] = [ + sys.executable, + pip_runnable, + "install", + "--ignore-installed", + "--no-user", + "--prefix", + prefix.path, + "--no-warn-script-location", + "--disable-pip-version-check", + ] if logger.getEffectiveLevel() <= logging.DEBUG: - args.append('-v') - for format_control in ('no_binary', 'only_binary'): + args.append("-vv") + elif logger.getEffectiveLevel() <= VERBOSE: + args.append("-v") + for format_control in ("no_binary", "only_binary"): formats = getattr(finder.format_control, format_control) - args.extend(('--' + format_control.replace('_', '-'), - ','.join(sorted(formats or {':none:'})))) + args.extend( + ( + "--" + format_control.replace("_", "-"), + ",".join(sorted(formats or {":none:"})), + ) + ) index_urls = finder.index_urls if index_urls: - args.extend(['-i', index_urls[0]]) + args.extend(["-i", index_urls[0]]) for extra_index in index_urls[1:]: - args.extend(['--extra-index-url', extra_index]) + args.extend(["--extra-index-url", extra_index]) else: - args.append('--no-index') + args.append("--no-index") for link in finder.find_links: - args.extend(['--find-links', link]) + args.extend(["--find-links", link]) for host in finder.trusted_hosts: - args.extend(['--trusted-host', host]) + args.extend(["--trusted-host", host]) if finder.allow_all_prereleases: - args.append('--pre') - args.append('--') + args.append("--pre") + if finder.prefer_binary: + args.append("--prefer-binary") + args.append("--") args.extend(requirements) - with open_spinner(message) as spinner: - call_subprocess(args, spinner=spinner) + extra_environ = {"_PIP_STANDALONE_CERT": where()} + with open_spinner(f"Installing {kind}") as spinner: + call_subprocess( + args, + command_desc=f"pip subprocess to install {kind}", + spinner=spinner, + extra_environ=extra_environ, + ) class NoOpBuildEnvironment(BuildEnvironment): - """A no-op drop-in replacement for BuildEnvironment - """ + """A no-op drop-in replacement for BuildEnvironment""" - def __init__(self): + def __init__(self) -> None: pass - def __enter__(self): + def __enter__(self) -> None: pass - def __exit__(self, exc_type, exc_val, exc_tb): + def __exit__( + self, + exc_type: Optional[Type[BaseException]], + exc_val: Optional[BaseException], + exc_tb: Optional[TracebackType], + ) -> None: pass - def cleanup(self): + def cleanup(self) -> None: pass - def install_requirements(self, finder, requirements, prefix, message): + def install_requirements( + self, + finder: "PackageFinder", + requirements: Iterable[str], + prefix_as_string: str, + *, + kind: str, + ) -> None: raise NotImplementedError() diff --git a/src/fetchcode/vcs/pip/_internal/cache.py b/src/fetchcode/vcs/pip/_internal/cache.py index 37c6198e..6b451267 100644 --- a/src/fetchcode/vcs/pip/_internal/cache.py +++ b/src/fetchcode/vcs/pip/_internal/cache.py @@ -1,98 +1,50 @@ """Cache Management """ -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False - import hashlib import json import logging import os +from pathlib import Path +from typing import Any, Dict, List, Optional -from fetchcode.vcs.pip._vendor.packaging.tags import interpreter_name, interpreter_version -from fetchcode.vcs.pip._vendor.packaging.utils import canonicalize_name - -from fetchcode.vcs.pip._internal.exceptions import InvalidWheelFilename -from fetchcode.vcs.pip._internal.models.link import Link -from fetchcode.vcs.pip._internal.models.wheel import Wheel -from fetchcode.vcs.pip._internal.utils.temp_dir import TempDirectory, tempdir_kinds -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.utils.urls import path_to_url - -if MYPY_CHECK_RUNNING: - from typing import Optional, Set, List, Any, Dict - - from fetchcode.vcs.pip._vendor.packaging.tags import Tag +from pip._vendor.packaging.tags import Tag, interpreter_name, interpreter_version +from pip._vendor.packaging.utils import canonicalize_name - from fetchcode.vcs.pip._internal.models.format_control import FormatControl +from pip._internal.exceptions import InvalidWheelFilename +from pip._internal.models.direct_url import DirectUrl +from pip._internal.models.link import Link +from pip._internal.models.wheel import Wheel +from pip._internal.utils.temp_dir import TempDirectory, tempdir_kinds +from pip._internal.utils.urls import path_to_url logger = logging.getLogger(__name__) +ORIGIN_JSON_NAME = "origin.json" -def _hash_dict(d): - # type: (Dict[str, str]) -> str + +def _hash_dict(d: Dict[str, str]) -> str: """Return a stable sha224 of a dictionary.""" s = json.dumps(d, sort_keys=True, separators=(",", ":"), ensure_ascii=True) return hashlib.sha224(s.encode("ascii")).hexdigest() -class Cache(object): +class Cache: """An abstract class - provides cache directories for data from links - - :param cache_dir: The root of the cache. - :param format_control: An object of FormatControl class to limit - binaries being read from the cache. - :param allowed_formats: which formats of files the cache should store. - ('binary' and 'source' are the only allowed values) + :param cache_dir: The root of the cache. """ - def __init__(self, cache_dir, format_control, allowed_formats): - # type: (str, FormatControl, Set[str]) -> None - super(Cache, self).__init__() + def __init__(self, cache_dir: str) -> None: + super().__init__() assert not cache_dir or os.path.isabs(cache_dir) self.cache_dir = cache_dir or None - self.format_control = format_control - self.allowed_formats = allowed_formats - - _valid_formats = {"source", "binary"} - assert self.allowed_formats.union(_valid_formats) == _valid_formats - - def _get_cache_path_parts_legacy(self, link): - # type: (Link) -> List[str] - """Get parts of part that must be os.path.joined with cache_dir - - Legacy cache key (pip < 20) for compatibility with older caches. - """ - - # We want to generate an url to use as our cache key, we don't want to - # just re-use the URL because it might have other items in the fragment - # and we don't care about those. - key_parts = [link.url_without_fragment] - if link.hash_name is not None and link.hash is not None: - key_parts.append("=".join([link.hash_name, link.hash])) - key_url = "#".join(key_parts) - # Encode our key url with sha224, we'll use this because it has similar - # security properties to sha256, but with a shorter total output (and - # thus less secure). However the differences don't make a lot of - # difference for our use case here. - hashed = hashlib.sha224(key_url.encode()).hexdigest() - - # We want to nest the directories some to prevent having a ton of top - # level directories where we might run out of sub directories on some - # FS. - parts = [hashed[:2], hashed[2:4], hashed[4:6], hashed[6:]] - - return parts - - def _get_cache_path_parts(self, link): - # type: (Link) -> List[str] - """Get parts of part that must be os.path.joined with cache_dir - """ + def _get_cache_path_parts(self, link: Link) -> List[str]: + """Get parts of part that must be os.path.joined with cache_dir""" # We want to generate an url to use as our cache key, we don't want to - # just re-use the URL because it might have other items in the fragment + # just reuse the URL because it might have other items in the fragment # and we don't care about those. key_parts = {"url": link.url_without_fragment} if link.hash_name is not None and link.hash is not None: @@ -121,51 +73,26 @@ def _get_cache_path_parts(self, link): return parts - def _get_candidates(self, link, canonical_package_name): - # type: (Link, Optional[str]) -> List[Any] - can_not_cache = ( - not self.cache_dir or - not canonical_package_name or - not link - ) + def _get_candidates(self, link: Link, canonical_package_name: str) -> List[Any]: + can_not_cache = not self.cache_dir or not canonical_package_name or not link if can_not_cache: return [] - formats = self.format_control.get_allowed_formats( - canonical_package_name - ) - if not self.allowed_formats.intersection(formats): - return [] - - candidates = [] path = self.get_path_for_link(link) if os.path.isdir(path): - for candidate in os.listdir(path): - candidates.append((candidate, path)) - # TODO remove legacy path lookup in pip>=21 - legacy_path = self.get_path_for_link_legacy(link) - if os.path.isdir(legacy_path): - for candidate in os.listdir(legacy_path): - candidates.append((candidate, legacy_path)) - return candidates - - def get_path_for_link_legacy(self, link): - # type: (Link) -> str - raise NotImplementedError() + return [(candidate, path) for candidate in os.listdir(path)] + return [] - def get_path_for_link(self, link): - # type: (Link) -> str - """Return a directory to store cached items in for link. - """ + def get_path_for_link(self, link: Link) -> str: + """Return a directory to store cached items in for link.""" raise NotImplementedError() def get( self, - link, # type: Link - package_name, # type: Optional[str] - supported_tags, # type: List[Tag] - ): - # type: (...) -> Link + link: Link, + package_name: Optional[str], + supported_tags: List[Tag], + ) -> Link: """Returns a link to a cached item if it exists, otherwise returns the passed link. """ @@ -173,22 +100,12 @@ def get( class SimpleWheelCache(Cache): - """A cache of wheels for future installs. - """ - - def __init__(self, cache_dir, format_control): - # type: (str, FormatControl) -> None - super(SimpleWheelCache, self).__init__( - cache_dir, format_control, {"binary"} - ) + """A cache of wheels for future installs.""" - def get_path_for_link_legacy(self, link): - # type: (Link) -> str - parts = self._get_cache_path_parts_legacy(link) - return os.path.join(self.cache_dir, "wheels", *parts) + def __init__(self, cache_dir: str) -> None: + super().__init__(cache_dir) - def get_path_for_link(self, link): - # type: (Link) -> str + def get_path_for_link(self, link: Link) -> str: """Return a directory to store cached wheels for link Because there are M wheels for any one sdist, we provide a directory @@ -204,36 +121,34 @@ def get_path_for_link(self, link): :param link: The link of the sdist for which this will cache wheels. """ parts = self._get_cache_path_parts(link) - + assert self.cache_dir # Store wheels within the root cache_dir return os.path.join(self.cache_dir, "wheels", *parts) def get( self, - link, # type: Link - package_name, # type: Optional[str] - supported_tags, # type: List[Tag] - ): - # type: (...) -> Link + link: Link, + package_name: Optional[str], + supported_tags: List[Tag], + ) -> Link: candidates = [] if not package_name: return link canonical_package_name = canonicalize_name(package_name) - for wheel_name, wheel_dir in self._get_candidates( - link, canonical_package_name - ): + for wheel_name, wheel_dir in self._get_candidates(link, canonical_package_name): try: wheel = Wheel(wheel_name) except InvalidWheelFilename: continue if canonicalize_name(wheel.name) != canonical_package_name: logger.debug( - "Ignoring cached wheel {} for {} as it " - "does not match the expected distribution name {}.".format( - wheel_name, link, package_name - ) + "Ignoring cached wheel %s for %s as it " + "does not match the expected distribution name %s.", + wheel_name, + link, + package_name, ) continue if not wheel.supported(supported_tags): @@ -255,29 +170,39 @@ def get( class EphemWheelCache(SimpleWheelCache): - """A SimpleWheelCache that creates it's own temporary cache directory - """ + """A SimpleWheelCache that creates it's own temporary cache directory""" - def __init__(self, format_control): - # type: (FormatControl) -> None + def __init__(self) -> None: self._temp_dir = TempDirectory( kind=tempdir_kinds.EPHEM_WHEEL_CACHE, globally_managed=True, ) - super(EphemWheelCache, self).__init__( - self._temp_dir.path, format_control - ) + super().__init__(self._temp_dir.path) -class CacheEntry(object): +class CacheEntry: def __init__( self, - link, # type: Link - persistent, # type: bool + link: Link, + persistent: bool, ): self.link = link self.persistent = persistent + self.origin: Optional[DirectUrl] = None + origin_direct_url_path = Path(self.link.file_path).parent / ORIGIN_JSON_NAME + if origin_direct_url_path.exists(): + try: + self.origin = DirectUrl.from_json( + origin_direct_url_path.read_text(encoding="utf-8") + ) + except Exception as e: + logger.warning( + "Ignoring invalid cache entry origin file %s for %s (%s)", + origin_direct_url_path, + link.filename, + e, + ) class WheelCache(Cache): @@ -287,33 +212,23 @@ class WheelCache(Cache): when a certain link is not found in the simple wheel cache first. """ - def __init__(self, cache_dir, format_control): - # type: (str, FormatControl) -> None - super(WheelCache, self).__init__( - cache_dir, format_control, {'binary'} - ) - self._wheel_cache = SimpleWheelCache(cache_dir, format_control) - self._ephem_cache = EphemWheelCache(format_control) - - def get_path_for_link_legacy(self, link): - # type: (Link) -> str - return self._wheel_cache.get_path_for_link_legacy(link) + def __init__(self, cache_dir: str) -> None: + super().__init__(cache_dir) + self._wheel_cache = SimpleWheelCache(cache_dir) + self._ephem_cache = EphemWheelCache() - def get_path_for_link(self, link): - # type: (Link) -> str + def get_path_for_link(self, link: Link) -> str: return self._wheel_cache.get_path_for_link(link) - def get_ephem_path_for_link(self, link): - # type: (Link) -> str + def get_ephem_path_for_link(self, link: Link) -> str: return self._ephem_cache.get_path_for_link(link) def get( self, - link, # type: Link - package_name, # type: Optional[str] - supported_tags, # type: List[Tag] - ): - # type: (...) -> Link + link: Link, + package_name: Optional[str], + supported_tags: List[Tag], + ) -> Link: cache_entry = self.get_cache_entry(link, package_name, supported_tags) if cache_entry is None: return link @@ -321,11 +236,10 @@ def get( def get_cache_entry( self, - link, # type: Link - package_name, # type: Optional[str] - supported_tags, # type: List[Tag] - ): - # type: (...) -> Optional[CacheEntry] + link: Link, + package_name: Optional[str], + supported_tags: List[Tag], + ) -> Optional[CacheEntry]: """Returns a CacheEntry with a link to a cached item if it exists or None. The cache entry indicates if the item was found in the persistent or ephemeral cache. @@ -347,3 +261,30 @@ def get_cache_entry( return CacheEntry(retval, persistent=False) return None + + @staticmethod + def record_download_origin(cache_dir: str, download_info: DirectUrl) -> None: + origin_path = Path(cache_dir) / ORIGIN_JSON_NAME + if origin_path.exists(): + try: + origin = DirectUrl.from_json(origin_path.read_text(encoding="utf-8")) + except Exception as e: + logger.warning( + "Could not read origin file %s in cache entry (%s). " + "Will attempt to overwrite it.", + origin_path, + e, + ) + else: + # TODO: use DirectUrl.equivalent when + # https://github.com/pypa/pip/pull/10564 is merged. + if origin.url != download_info.url: + logger.warning( + "Origin URL %s in cache entry %s does not match download URL " + "%s. This is likely a pip bug or a cache corruption issue. " + "Will overwrite it with the new value.", + origin.url, + cache_dir, + download_info.url, + ) + origin_path.write_text(download_info.to_json(), encoding="utf-8") diff --git a/src/fetchcode/vcs/pip/_internal/cli/autocompletion.py b/src/fetchcode/vcs/pip/_internal/cli/autocompletion.py index bc26912a..f3f70ac8 100644 --- a/src/fetchcode/vcs/pip/_internal/cli/autocompletion.py +++ b/src/fetchcode/vcs/pip/_internal/cli/autocompletion.py @@ -5,36 +5,35 @@ import os import sys from itertools import chain +from typing import Any, Iterable, List, Optional -from fetchcode.vcs.pip._internal.cli.main_parser import create_main_parser -from fetchcode.vcs.pip._internal.commands import commands_dict, create_command -from fetchcode.vcs.pip._internal.utils.misc import get_installed_distributions -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._internal.cli.main_parser import create_main_parser +from pip._internal.commands import commands_dict, create_command +from pip._internal.metadata import get_default_environment -if MYPY_CHECK_RUNNING: - from typing import Any, Iterable, List, Optional - -def autocomplete(): - # type: () -> None - """Entry Point for completion of main and subcommand options. - """ +def autocomplete() -> None: + """Entry Point for completion of main and subcommand options.""" # Don't complete if user hasn't sourced bash_completion file. - if 'PIP_AUTO_COMPLETE' not in os.environ: + if "PIP_AUTO_COMPLETE" not in os.environ: return - cwords = os.environ['COMP_WORDS'].split()[1:] - cword = int(os.environ['COMP_CWORD']) + # Don't complete if autocompletion environment variables + # are not present + if not os.environ.get("COMP_WORDS") or not os.environ.get("COMP_CWORD"): + return + cwords = os.environ["COMP_WORDS"].split()[1:] + cword = int(os.environ["COMP_CWORD"]) try: current = cwords[cword - 1] except IndexError: - current = '' + current = "" parser = create_main_parser() subcommands = list(commands_dict) options = [] # subcommand - subcommand_name = None # type: Optional[str] + subcommand_name: Optional[str] = None for word in cwords: if word in subcommands: subcommand_name = word @@ -42,40 +41,54 @@ def autocomplete(): # subcommand options if subcommand_name is not None: # special case: 'help' subcommand has no options - if subcommand_name == 'help': + if subcommand_name == "help": sys.exit(1) # special case: list locally installed dists for show and uninstall - should_list_installed = ( - subcommand_name in ['show', 'uninstall'] and - not current.startswith('-') - ) + should_list_installed = not current.startswith("-") and subcommand_name in [ + "show", + "uninstall", + ] if should_list_installed: - installed = [] + env = get_default_environment() lc = current.lower() - for dist in get_installed_distributions(local_only=True): - if dist.key.startswith(lc) and dist.key not in cwords[1:]: - installed.append(dist.key) + installed = [ + dist.canonical_name + for dist in env.iter_installed_distributions(local_only=True) + if dist.canonical_name.startswith(lc) + and dist.canonical_name not in cwords[1:] + ] # if there are no dists installed, fall back to option completion if installed: for dist in installed: print(dist) sys.exit(1) + should_list_installables = ( + not current.startswith("-") and subcommand_name == "install" + ) + if should_list_installables: + for path in auto_complete_paths(current, "path"): + print(path) + sys.exit(1) + subcommand = create_command(subcommand_name) for opt in subcommand.parser.option_list_all: if opt.help != optparse.SUPPRESS_HELP: - for opt_str in opt._long_opts + opt._short_opts: - options.append((opt_str, opt.nargs)) + options += [ + (opt_str, opt.nargs) for opt_str in opt._long_opts + opt._short_opts + ] # filter out previously specified options from available options - prev_opts = [x.split('=')[0] for x in cwords[1:cword - 1]] + prev_opts = [x.split("=")[0] for x in cwords[1 : cword - 1]] options = [(x, v) for (x, v) in options if x not in prev_opts] # filter options by current input options = [(k, v) for k, v in options if k.startswith(current)] # get completion type given cwords and available subcommand options completion_type = get_path_completion_type( - cwords, cword, subcommand.parser.option_list_all, + cwords, + cword, + subcommand.parser.option_list_all, ) # get completion files and directories if ``completion_type`` is # ````, ```` or ```` @@ -86,7 +99,7 @@ def autocomplete(): opt_label = option[0] # append '=' to options which require args if option[1] and option[0][:2] == "--": - opt_label += '=' + opt_label += "=" print(opt_label) else: # show main parser options only when necessary @@ -94,24 +107,23 @@ def autocomplete(): opts = [i.option_list for i in parser.option_groups] opts.append(parser.option_list) flattened_opts = chain.from_iterable(opts) - if current.startswith('-'): + if current.startswith("-"): for opt in flattened_opts: if opt.help != optparse.SUPPRESS_HELP: subcommands += opt._long_opts + opt._short_opts else: # get completion type given cwords and all available options - completion_type = get_path_completion_type(cwords, cword, - flattened_opts) + completion_type = get_path_completion_type(cwords, cword, flattened_opts) if completion_type: - subcommands = list(auto_complete_paths(current, - completion_type)) + subcommands = list(auto_complete_paths(current, completion_type)) - print(' '.join([x for x in subcommands if x.startswith(current)])) + print(" ".join([x for x in subcommands if x.startswith(current)])) sys.exit(1) -def get_path_completion_type(cwords, cword, opts): - # type: (List[str], int, Iterable[Any]) -> Optional[str] +def get_path_completion_type( + cwords: List[str], cword: int, opts: Iterable[Any] +) -> Optional[str]: """Get the type of path completion (``file``, ``dir``, ``path`` or None) :param cwords: same as the environmental variable ``COMP_WORDS`` @@ -119,28 +131,27 @@ def get_path_completion_type(cwords, cword, opts): :param opts: The available options to check :return: path completion type (``file``, ``dir``, ``path`` or None) """ - if cword < 2 or not cwords[cword - 2].startswith('-'): + if cword < 2 or not cwords[cword - 2].startswith("-"): return None for opt in opts: if opt.help == optparse.SUPPRESS_HELP: continue - for o in str(opt).split('/'): - if cwords[cword - 2].split('=')[0] == o: + for o in str(opt).split("/"): + if cwords[cword - 2].split("=")[0] == o: if not opt.metavar or any( - x in ('path', 'file', 'dir') - for x in opt.metavar.split('/')): + x in ("path", "file", "dir") for x in opt.metavar.split("/") + ): return opt.metavar return None -def auto_complete_paths(current, completion_type): - # type: (str, str) -> Iterable[str] +def auto_complete_paths(current: str, completion_type: str) -> Iterable[str]: """If ``completion_type`` is ``file`` or ``path``, list all regular files and directories starting with ``current``; otherwise only list directories starting with ``current``. :param current: The word to be completed - :param completion_type: path completion type(`file`, `path` or `dir`)i + :param completion_type: path completion type(``file``, ``path`` or ``dir``) :return: A generator of regular files and/or directories """ directory, filename = os.path.split(current) @@ -150,15 +161,16 @@ def auto_complete_paths(current, completion_type): return filename = os.path.normcase(filename) # list all files that start with ``filename`` - file_list = (x for x in os.listdir(current_path) - if os.path.normcase(x).startswith(filename)) + file_list = ( + x for x in os.listdir(current_path) if os.path.normcase(x).startswith(filename) + ) for f in file_list: opt = os.path.join(current_path, f) comp_file = os.path.normcase(os.path.join(directory, f)) # complete regular files when there is not ```` after option # complete directories when there is ````, ```` or # ````after option - if completion_type != 'dir' and os.path.isfile(opt): + if completion_type != "dir" and os.path.isfile(opt): yield comp_file elif os.path.isdir(opt): - yield os.path.join(comp_file, '') + yield os.path.join(comp_file, "") diff --git a/src/fetchcode/vcs/pip/_internal/cli/base_command.py b/src/fetchcode/vcs/pip/_internal/cli/base_command.py index 576fc4b0..bc1ab659 100644 --- a/src/fetchcode/vcs/pip/_internal/cli/base_command.py +++ b/src/fetchcode/vcs/pip/_internal/cli/base_command.py @@ -1,84 +1,69 @@ """Base Command class, and related routines""" -from __future__ import absolute_import, print_function - import logging import logging.config import optparse import os -import platform import sys import traceback +from optparse import Values +from typing import List, Optional, Tuple -from fetchcode.vcs.pip._internal.cli import cmdoptions -from fetchcode.vcs.pip._internal.cli.command_context import CommandContextMixIn -from fetchcode.vcs.pip._internal.cli.parser import ( - ConfigOptionParser, - UpdatingDefaultsHelpFormatter, -) -from fetchcode.vcs.pip._internal.cli.status_codes import ( +from pip._vendor.rich import reconfigure +from pip._vendor.rich import traceback as rich_traceback + +from pip._internal.cli import cmdoptions +from pip._internal.cli.command_context import CommandContextMixIn +from pip._internal.cli.parser import ConfigOptionParser, UpdatingDefaultsHelpFormatter +from pip._internal.cli.status_codes import ( ERROR, PREVIOUS_BUILD_DIR_ERROR, - SUCCESS, UNKNOWN_ERROR, VIRTUALENV_NOT_FOUND, ) -from fetchcode.vcs.pip._internal.exceptions import ( +from pip._internal.exceptions import ( BadCommand, CommandError, + DiagnosticPipError, InstallationError, + NetworkConnectionError, PreviousBuildDirError, - UninstallationError, ) -from fetchcode.vcs.pip._internal.utils.deprecation import deprecated -from fetchcode.vcs.pip._internal.utils.filesystem import check_path_owner -from fetchcode.vcs.pip._internal.utils.logging import BrokenStdoutLoggingError, setup_logging -from fetchcode.vcs.pip._internal.utils.misc import get_prog, normalize_path -from fetchcode.vcs.pip._internal.utils.temp_dir import ( - global_tempdir_manager, - tempdir_registry, -) -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.utils.virtualenv import running_under_virtualenv - -if MYPY_CHECK_RUNNING: - from typing import List, Optional, Tuple, Any - from optparse import Values +from pip._internal.utils.filesystem import check_path_owner +from pip._internal.utils.logging import BrokenStdoutLoggingError, setup_logging +from pip._internal.utils.misc import get_prog, normalize_path +from pip._internal.utils.temp_dir import TempDirectoryTypeRegistry as TempDirRegistry +from pip._internal.utils.temp_dir import global_tempdir_manager, tempdir_registry +from pip._internal.utils.virtualenv import running_under_virtualenv - from fetchcode.vcs.pip._internal.utils.temp_dir import ( - TempDirectoryTypeRegistry as TempDirRegistry - ) - -__all__ = ['Command'] +__all__ = ["Command"] logger = logging.getLogger(__name__) class Command(CommandContextMixIn): - usage = None # type: str - ignore_require_venv = False # type: bool - - def __init__(self, name, summary, isolated=False): - # type: (str, str, bool) -> None - super(Command, self).__init__() - parser_kw = { - 'usage': self.usage, - 'prog': '{} {}'.format(get_prog(), name), - 'formatter': UpdatingDefaultsHelpFormatter(), - 'add_help_option': False, - 'name': name, - 'description': self.__doc__, - 'isolated': isolated, - } + usage: str = "" + ignore_require_venv: bool = False + + def __init__(self, name: str, summary: str, isolated: bool = False) -> None: + super().__init__() self.name = name self.summary = summary - self.parser = ConfigOptionParser(**parser_kw) + self.parser = ConfigOptionParser( + usage=self.usage, + prog=f"{get_prog()} {name}", + formatter=UpdatingDefaultsHelpFormatter(), + add_help_option=False, + name=name, + description=self.__doc__, + isolated=isolated, + ) - self.tempdir_registry = None # type: Optional[TempDirRegistry] + self.tempdir_registry: Optional[TempDirRegistry] = None # Commands should add options to this option group - optgroup_name = '{} Options'.format(self.name.capitalize()) + optgroup_name = f"{self.name.capitalize()} Options" self.cmd_opts = optparse.OptionGroup(self.parser, optgroup_name) # Add the general options @@ -88,35 +73,92 @@ def __init__(self, name, summary, isolated=False): ) self.parser.add_option_group(gen_opts) - def handle_pip_version_check(self, options): - # type: (Values) -> None + self.add_options() + + def add_options(self) -> None: + pass + + def handle_pip_version_check(self, options: Values) -> None: """ This is a no-op so that commands by default do not do the pip version check. """ # Make sure we do the pip version check if the index_group options # are present. - assert not hasattr(options, 'no_index') + assert not hasattr(options, "no_index") - def run(self, options, args): - # type: (Values, List[Any]) -> Any + def run(self, options: Values, args: List[str]) -> int: raise NotImplementedError - def parse_args(self, args): - # type: (List[str]) -> Tuple[Any, Any] + def _run_wrapper(self, level_number: int, options: Values, args: List[str]) -> int: + def _inner_run() -> int: + try: + return self.run(options, args) + finally: + self.handle_pip_version_check(options) + + if options.debug_mode: + rich_traceback.install(show_locals=True) + return _inner_run() + + try: + status = _inner_run() + assert isinstance(status, int) + return status + except DiagnosticPipError as exc: + logger.error("%s", exc, extra={"rich": True}) + logger.debug("Exception information:", exc_info=True) + + return ERROR + except PreviousBuildDirError as exc: + logger.critical(str(exc)) + logger.debug("Exception information:", exc_info=True) + + return PREVIOUS_BUILD_DIR_ERROR + except ( + InstallationError, + BadCommand, + NetworkConnectionError, + ) as exc: + logger.critical(str(exc)) + logger.debug("Exception information:", exc_info=True) + + return ERROR + except CommandError as exc: + logger.critical("%s", exc) + logger.debug("Exception information:", exc_info=True) + + return ERROR + except BrokenStdoutLoggingError: + # Bypass our logger and write any remaining messages to + # stderr because stdout no longer works. + print("ERROR: Pipe to stdout was broken", file=sys.stderr) + if level_number <= logging.DEBUG: + traceback.print_exc(file=sys.stderr) + + return ERROR + except KeyboardInterrupt: + logger.critical("Operation cancelled by user") + logger.debug("Exception information:", exc_info=True) + + return ERROR + except BaseException: + logger.critical("Exception:", exc_info=True) + + return UNKNOWN_ERROR + + def parse_args(self, args: List[str]) -> Tuple[Values, List[str]]: # factored out for testability return self.parser.parse_args(args) - def main(self, args): - # type: (List[str]) -> int + def main(self, args: List[str]) -> int: try: with self.main_context(): return self._main(args) finally: logging.shutdown() - def _main(self, args): - # type: (List[str]) -> int + def _main(self, args: List[str]) -> int: # We must initialize this before the tempdir manager, otherwise the # configuration would not be accessible by the time we clean up the # tempdir manager. @@ -130,45 +172,47 @@ def _main(self, args): # Set verbosity so that it can be used elsewhere. self.verbosity = options.verbose - options.quiet + reconfigure(no_color=options.no_color) level_number = setup_logging( verbosity=self.verbosity, no_color=options.no_color, user_log_file=options.log, ) - if ( - sys.version_info[:2] == (2, 7) and - not options.no_python_version_warning - ): - message = ( - "pip 21.0 will drop support for Python 2.7 in January 2021. " - "More details about Python 2 support in pip, can be found at " - "https://pip.pypa.io/en/latest/development/release-process/#python-2-support" # noqa + always_enabled_features = set(options.features_enabled) & set( + cmdoptions.ALWAYS_ENABLED_FEATURES + ) + if always_enabled_features: + logger.warning( + "The following features are always enabled: %s. ", + ", ".join(sorted(always_enabled_features)), + ) + + # Make sure that the --python argument isn't specified after the + # subcommand. We can tell, because if --python was specified, + # we should only reach this point if we're running in the created + # subprocess, which has the _PIP_RUNNING_IN_SUBPROCESS environment + # variable set. + if options.python and "_PIP_RUNNING_IN_SUBPROCESS" not in os.environ: + logger.critical( + "The --python option must be placed before the pip subcommand name" ) - if platform.python_implementation() == "CPython": - message = ( - "Python 2.7 reached the end of its life on January " - "1st, 2020. Please upgrade your Python as Python 2.7 " - "is no longer maintained. " - ) + message - deprecated(message, replacement=None, gone_in=None) + sys.exit(ERROR) # TODO: Try to get these passing down from the command? # without resorting to os.environ to hold these. # This also affects isolated builds and it should. if options.no_input: - os.environ['PIP_NO_INPUT'] = '1' + os.environ["PIP_NO_INPUT"] = "1" if options.exists_action: - os.environ['PIP_EXISTS_ACTION'] = ' '.join(options.exists_action) + os.environ["PIP_EXISTS_ACTION"] = " ".join(options.exists_action) if options.require_venv and not self.ignore_require_venv: # If a venv is required check if it can really be found if not running_under_virtualenv(): - logger.critical( - 'Could not find an activated virtualenv (required).' - ) + logger.critical("Could not find an activated virtualenv (required).") sys.exit(VIRTUALENV_NOT_FOUND) if options.cache_dir: @@ -178,51 +222,10 @@ def _main(self, args): "The directory '%s' or its parent directory is not owned " "or is not writable by the current user. The cache " "has been disabled. Check the permissions and owner of " - "that directory. If executing pip with sudo, you may want " - "sudo's -H flag.", + "that directory. If executing pip with sudo, you should " + "use sudo's -H flag.", options.cache_dir, ) options.cache_dir = None - try: - status = self.run(options, args) - # FIXME: all commands should return an exit status - # and when it is done, isinstance is not needed anymore - if isinstance(status, int): - return status - except PreviousBuildDirError as exc: - logger.critical(str(exc)) - logger.debug('Exception information:', exc_info=True) - - return PREVIOUS_BUILD_DIR_ERROR - except (InstallationError, UninstallationError, BadCommand) as exc: - logger.critical(str(exc)) - logger.debug('Exception information:', exc_info=True) - - return ERROR - except CommandError as exc: - logger.critical('%s', exc) - logger.debug('Exception information:', exc_info=True) - - return ERROR - except BrokenStdoutLoggingError: - # Bypass our logger and write any remaining messages to stderr - # because stdout no longer works. - print('ERROR: Pipe to stdout was broken', file=sys.stderr) - if level_number <= logging.DEBUG: - traceback.print_exc(file=sys.stderr) - - return ERROR - except KeyboardInterrupt: - logger.critical('Operation cancelled by user') - logger.debug('Exception information:', exc_info=True) - - return ERROR - except BaseException: - logger.critical('Exception:', exc_info=True) - - return UNKNOWN_ERROR - finally: - self.handle_pip_version_check(options) - - return SUCCESS + return self._run_wrapper(level_number, options, args) diff --git a/src/fetchcode/vcs/pip/_internal/cli/cmdoptions.py b/src/fetchcode/vcs/pip/_internal/cli/cmdoptions.py index 039b6836..0b7cff77 100644 --- a/src/fetchcode/vcs/pip/_internal/cli/cmdoptions.py +++ b/src/fetchcode/vcs/pip/_internal/cli/cmdoptions.py @@ -10,36 +10,30 @@ # The following comment should be removed at some point in the future. # mypy: strict-optional=False -from __future__ import absolute_import - +import importlib.util import logging import os import textwrap -import warnings -from distutils.util import strtobool from functools import partial -from optparse import SUPPRESS_HELP, Option, OptionGroup +from optparse import SUPPRESS_HELP, Option, OptionGroup, OptionParser, Values from textwrap import dedent +from typing import Any, Callable, Dict, Optional, Tuple -from fetchcode.vcs.pip._internal.cli.progress_bars import BAR_TYPES -from fetchcode.vcs.pip._internal.exceptions import CommandError -from fetchcode.vcs.pip._internal.locations import USER_CACHE_DIR, get_src_prefix -from fetchcode.vcs.pip._internal.models.format_control import FormatControl -from fetchcode.vcs.pip._internal.models.index import PyPI -from fetchcode.vcs.pip._internal.models.target_python import TargetPython -from fetchcode.vcs.pip._internal.utils.hashes import STRONG_HASHES -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._vendor.packaging.utils import canonicalize_name -if MYPY_CHECK_RUNNING: - from typing import Any, Callable, Dict, Optional, Tuple - from optparse import OptionParser, Values - from fetchcode.vcs.pip._internal.cli.parser import ConfigOptionParser +from pip._internal.cli.parser import ConfigOptionParser +from pip._internal.exceptions import CommandError +from pip._internal.locations import USER_CACHE_DIR, get_src_prefix +from pip._internal.models.format_control import FormatControl +from pip._internal.models.index import PyPI +from pip._internal.models.target_python import TargetPython +from pip._internal.utils.hashes import STRONG_HASHES +from pip._internal.utils.misc import strtobool logger = logging.getLogger(__name__) -def raise_option_error(parser, option, msg): - # type: (OptionParser, Option, str) -> None +def raise_option_error(parser: OptionParser, option: Option, msg: str) -> None: """ Raise an option parsing error using parser.error(). @@ -48,66 +42,41 @@ def raise_option_error(parser, option, msg): option: an Option instance. msg: the error text. """ - msg = '{} error: {}'.format(option, msg) - msg = textwrap.fill(' '.join(msg.split())) + msg = f"{option} error: {msg}" + msg = textwrap.fill(" ".join(msg.split())) parser.error(msg) -def make_option_group(group, parser): - # type: (Dict[str, Any], ConfigOptionParser) -> OptionGroup +def make_option_group(group: Dict[str, Any], parser: ConfigOptionParser) -> OptionGroup: """ Return an OptionGroup object group -- assumed to be dict with 'name' and 'options' keys parser -- an optparse Parser """ - option_group = OptionGroup(parser, group['name']) - for option in group['options']: + option_group = OptionGroup(parser, group["name"]) + for option in group["options"]: option_group.add_option(option()) return option_group -def check_install_build_global(options, check_options=None): - # type: (Values, Optional[Values]) -> None - """Disable wheels if per-setup.py call options are set. - - :param options: The OptionParser options to update. - :param check_options: The options to check, if not supplied defaults to - options. - """ - if check_options is None: - check_options = options - - def getname(n): - # type: (str) -> Optional[Any] - return getattr(check_options, n, None) - names = ["build_options", "global_options", "install_options"] - if any(map(getname, names)): - control = options.format_control - control.disallow_binaries() - warnings.warn( - 'Disabling all use of wheels due to the use of --build-option ' - '/ --global-option / --install-option.', stacklevel=2, - ) - - -def check_dist_restriction(options, check_target=False): - # type: (Values, bool) -> None +def check_dist_restriction(options: Values, check_target: bool = False) -> None: """Function for determining if custom platform options are allowed. :param options: The OptionParser options. :param check_target: Whether or not to check if --target is being used. """ - dist_restriction_set = any([ - options.python_version, - options.platform, - options.abi, - options.implementation, - ]) - - binary_only = FormatControl(set(), {':all:'}) + dist_restriction_set = any( + [ + options.python_version, + options.platforms, + options.abis, + options.implementation, + ] + ) + + binary_only = FormatControl(set(), {":all:"}) sdist_dependencies_allowed = ( - options.format_control != binary_only and - not options.ignore_dependencies + options.format_control != binary_only and not options.ignore_dependencies ) # Installations or downloads using dist restrictions must not combine @@ -123,21 +92,25 @@ def check_dist_restriction(options, check_target=False): ) if check_target: - if dist_restriction_set and not options.target_dir: + if not options.dry_run and dist_restriction_set and not options.target_dir: raise CommandError( "Can not use any platform or abi specific options unless " - "installing via '--target'" + "installing via '--target' or using '--dry-run'" ) -def _path_option_check(option, opt, value): - # type: (Option, str, str) -> str +def _path_option_check(option: Option, opt: str, value: str) -> str: return os.path.expanduser(value) +def _package_name_option_check(option: Option, opt: str, value: str) -> str: + return canonicalize_name(value) + + class PipOption(Option): - TYPES = Option.TYPES + ("path",) + TYPES = Option.TYPES + ("path", "package_name") TYPE_CHECKER = Option.TYPE_CHECKER.copy() + TYPE_CHECKER["package_name"] = _package_name_option_check TYPE_CHECKER["path"] = _path_option_check @@ -145,15 +118,28 @@ class PipOption(Option): # options # ########### -help_ = partial( +help_: Callable[..., Option] = partial( + Option, + "-h", + "--help", + dest="help", + action="help", + help="Show help.", +) + +debug_mode: Callable[..., Option] = partial( Option, - '-h', '--help', - dest='help', - action='help', - help='Show help.', -) # type: Callable[..., Option] + "--debug", + dest="debug_mode", + action="store_true", + default=False, + help=( + "Let unhandled exceptions propagate outside the main subroutine, " + "instead of logging them to stderr." + ), +) -isolated_mode = partial( +isolated_mode: Callable[..., Option] = partial( Option, "--isolated", dest="isolated_mode", @@ -163,210 +149,248 @@ class PipOption(Option): "Run pip in an isolated mode, ignoring environment variables and user " "configuration." ), -) # type: Callable[..., Option] +) -require_virtualenv = partial( +require_virtualenv: Callable[..., Option] = partial( Option, - # Run only if inside a virtualenv, bail if not. - '--require-virtualenv', '--require-venv', - dest='require_venv', - action='store_true', + "--require-virtualenv", + "--require-venv", + dest="require_venv", + action="store_true", default=False, - help=SUPPRESS_HELP -) # type: Callable[..., Option] + help=( + "Allow pip to only run in a virtual environment; " + "exit with an error otherwise." + ), +) + +override_externally_managed: Callable[..., Option] = partial( + Option, + "--break-system-packages", + dest="override_externally_managed", + action="store_true", + help="Allow pip to modify an EXTERNALLY-MANAGED Python installation", +) -verbose = partial( +python: Callable[..., Option] = partial( Option, - '-v', '--verbose', - dest='verbose', - action='count', + "--python", + dest="python", + help="Run pip with the specified Python interpreter.", +) + +verbose: Callable[..., Option] = partial( + Option, + "-v", + "--verbose", + dest="verbose", + action="count", default=0, - help='Give more output. Option is additive, and can be used up to 3 times.' -) # type: Callable[..., Option] + help="Give more output. Option is additive, and can be used up to 3 times.", +) -no_color = partial( +no_color: Callable[..., Option] = partial( Option, - '--no-color', - dest='no_color', - action='store_true', + "--no-color", + dest="no_color", + action="store_true", default=False, - help="Suppress colored output", -) # type: Callable[..., Option] + help="Suppress colored output.", +) -version = partial( +version: Callable[..., Option] = partial( Option, - '-V', '--version', - dest='version', - action='store_true', - help='Show version and exit.', -) # type: Callable[..., Option] + "-V", + "--version", + dest="version", + action="store_true", + help="Show version and exit.", +) -quiet = partial( +quiet: Callable[..., Option] = partial( Option, - '-q', '--quiet', - dest='quiet', - action='count', + "-q", + "--quiet", + dest="quiet", + action="count", default=0, help=( - 'Give less output. Option is additive, and can be used up to 3' - ' times (corresponding to WARNING, ERROR, and CRITICAL logging' - ' levels).' + "Give less output. Option is additive, and can be used up to 3" + " times (corresponding to WARNING, ERROR, and CRITICAL logging" + " levels)." ), -) # type: Callable[..., Option] +) -progress_bar = partial( +progress_bar: Callable[..., Option] = partial( Option, - '--progress-bar', - dest='progress_bar', - type='choice', - choices=list(BAR_TYPES.keys()), - default='on', - help=( - 'Specify type of progress to be displayed [' + - '|'.join(BAR_TYPES.keys()) + '] (default: %default)' - ), -) # type: Callable[..., Option] - -log = partial( + "--progress-bar", + dest="progress_bar", + type="choice", + choices=["on", "off", "raw"], + default="on", + help="Specify whether the progress bar should be used [on, off, raw] (default: on)", +) + +log: Callable[..., Option] = partial( PipOption, - "--log", "--log-file", "--local-log", + "--log", + "--log-file", + "--local-log", dest="log", metavar="path", type="path", - help="Path to a verbose appending log." -) # type: Callable[..., Option] + help="Path to a verbose appending log.", +) -no_input = partial( +no_input: Callable[..., Option] = partial( Option, # Don't ask for input - '--no-input', - dest='no_input', - action='store_true', + "--no-input", + dest="no_input", + action="store_true", default=False, - help=SUPPRESS_HELP -) # type: Callable[..., Option] + help="Disable prompting for input.", +) -proxy = partial( +keyring_provider: Callable[..., Option] = partial( Option, - '--proxy', - dest='proxy', - type='str', - default='', - help="Specify a proxy in the form [user:passwd@]proxy.server:port." -) # type: Callable[..., Option] + "--keyring-provider", + dest="keyring_provider", + choices=["auto", "disabled", "import", "subprocess"], + default="auto", + help=( + "Enable the credential lookup via the keyring library if user input is allowed." + " Specify which mechanism to use [disabled, import, subprocess]." + " (default: disabled)" + ), +) -retries = partial( +proxy: Callable[..., Option] = partial( + Option, + "--proxy", + dest="proxy", + type="str", + default="", + help="Specify a proxy in the form scheme://[user:passwd@]proxy.server:port.", +) + +retries: Callable[..., Option] = partial( Option, - '--retries', - dest='retries', - type='int', + "--retries", + dest="retries", + type="int", default=5, help="Maximum number of retries each connection should attempt " - "(default %default times).", -) # type: Callable[..., Option] + "(default %default times).", +) -timeout = partial( +timeout: Callable[..., Option] = partial( Option, - '--timeout', '--default-timeout', - metavar='sec', - dest='timeout', - type='float', + "--timeout", + "--default-timeout", + metavar="sec", + dest="timeout", + type="float", default=15, - help='Set the socket timeout (default %default seconds).', -) # type: Callable[..., Option] + help="Set the socket timeout (default %default seconds).", +) -def exists_action(): - # type: () -> Option +def exists_action() -> Option: return Option( # Option when path already exist - '--exists-action', - dest='exists_action', - type='choice', - choices=['s', 'i', 'w', 'b', 'a'], + "--exists-action", + dest="exists_action", + type="choice", + choices=["s", "i", "w", "b", "a"], default=[], - action='append', - metavar='action', + action="append", + metavar="action", help="Default action when a path already exists: " - "(s)witch, (i)gnore, (w)ipe, (b)ackup, (a)bort.", + "(s)witch, (i)gnore, (w)ipe, (b)ackup, (a)bort.", ) -cert = partial( +cert: Callable[..., Option] = partial( PipOption, - '--cert', - dest='cert', - type='path', - metavar='path', - help="Path to alternate CA bundle.", -) # type: Callable[..., Option] - -client_cert = partial( + "--cert", + dest="cert", + type="path", + metavar="path", + help=( + "Path to PEM-encoded CA certificate bundle. " + "If provided, overrides the default. " + "See 'SSL Certificate Verification' in pip documentation " + "for more information." + ), +) + +client_cert: Callable[..., Option] = partial( PipOption, - '--client-cert', - dest='client_cert', - type='path', + "--client-cert", + dest="client_cert", + type="path", default=None, - metavar='path', + metavar="path", help="Path to SSL client certificate, a single file containing the " - "private key and the certificate in PEM format.", -) # type: Callable[..., Option] + "private key and the certificate in PEM format.", +) -index_url = partial( +index_url: Callable[..., Option] = partial( Option, - '-i', '--index-url', '--pypi-url', - dest='index_url', - metavar='URL', + "-i", + "--index-url", + "--pypi-url", + dest="index_url", + metavar="URL", default=PyPI.simple_url, help="Base URL of the Python Package Index (default %default). " - "This should point to a repository compliant with PEP 503 " - "(the simple repository API) or a local directory laid out " - "in the same format.", -) # type: Callable[..., Option] + "This should point to a repository compliant with PEP 503 " + "(the simple repository API) or a local directory laid out " + "in the same format.", +) -def extra_index_url(): - # type: () -> Option +def extra_index_url() -> Option: return Option( - '--extra-index-url', - dest='extra_index_urls', - metavar='URL', - action='append', + "--extra-index-url", + dest="extra_index_urls", + metavar="URL", + action="append", default=[], help="Extra URLs of package indexes to use in addition to " - "--index-url. Should follow the same rules as " - "--index-url.", + "--index-url. Should follow the same rules as " + "--index-url.", ) -no_index = partial( +no_index: Callable[..., Option] = partial( Option, - '--no-index', - dest='no_index', - action='store_true', + "--no-index", + dest="no_index", + action="store_true", default=False, - help='Ignore package index (only looking at --find-links URLs instead).', -) # type: Callable[..., Option] + help="Ignore package index (only looking at --find-links URLs instead).", +) -def find_links(): - # type: () -> Option +def find_links() -> Option: return Option( - '-f', '--find-links', - dest='find_links', - action='append', + "-f", + "--find-links", + dest="find_links", + action="append", default=[], - metavar='url', + metavar="url", help="If a URL or path to an html file, then parse for links to " - "archives such as sdist (.tar.gz) or wheel (.whl) files. " - "If a local path or file:// URL that's a directory, " - "then look for archives in the directory listing. " - "Links to VCS project URLs are not supported.", + "archives such as sdist (.tar.gz) or wheel (.whl) files. " + "If a local path or file:// URL that's a directory, " + "then look for archives in the directory listing. " + "Links to VCS project URLs are not supported.", ) -def trusted_host(): - # type: () -> Option +def trusted_host() -> Option: return Option( "--trusted-host", dest="trusted_hosts", @@ -374,138 +398,154 @@ def trusted_host(): metavar="HOSTNAME", default=[], help="Mark this host or host:port pair as trusted, even though it " - "does not have valid or any HTTPS.", + "does not have valid or any HTTPS.", ) -def constraints(): - # type: () -> Option +def constraints() -> Option: return Option( - '-c', '--constraint', - dest='constraints', - action='append', + "-c", + "--constraint", + dest="constraints", + action="append", default=[], - metavar='file', - help='Constrain versions using the given constraints file. ' - 'This option can be used multiple times.' + metavar="file", + help="Constrain versions using the given constraints file. " + "This option can be used multiple times.", ) -def requirements(): - # type: () -> Option +def requirements() -> Option: return Option( - '-r', '--requirement', - dest='requirements', - action='append', + "-r", + "--requirement", + dest="requirements", + action="append", default=[], - metavar='file', - help='Install from the given requirements file. ' - 'This option can be used multiple times.' + metavar="file", + help="Install from the given requirements file. " + "This option can be used multiple times.", ) -def editable(): - # type: () -> Option +def editable() -> Option: return Option( - '-e', '--editable', - dest='editables', - action='append', + "-e", + "--editable", + dest="editables", + action="append", default=[], - metavar='path/url', - help=('Install a project in editable mode (i.e. setuptools ' - '"develop mode") from a local project path or a VCS url.'), + metavar="path/url", + help=( + "Install a project in editable mode (i.e. setuptools " + '"develop mode") from a local project path or a VCS url.' + ), ) -def _handle_src(option, opt_str, value, parser): - # type: (Option, str, str, OptionParser) -> None +def _handle_src(option: Option, opt_str: str, value: str, parser: OptionParser) -> None: value = os.path.abspath(value) setattr(parser.values, option.dest, value) -src = partial( +src: Callable[..., Option] = partial( PipOption, - '--src', '--source', '--source-dir', '--source-directory', - dest='src_dir', - type='path', - metavar='dir', + "--src", + "--source", + "--source-dir", + "--source-directory", + dest="src_dir", + type="path", + metavar="dir", default=get_src_prefix(), - action='callback', + action="callback", callback=_handle_src, - help='Directory to check out editable projects into. ' + help="Directory to check out editable projects into. " 'The default in a virtualenv is "/src". ' - 'The default for global installs is "/src".' -) # type: Callable[..., Option] + 'The default for global installs is "/src".', +) -def _get_format_control(values, option): - # type: (Values, Option) -> Any +def _get_format_control(values: Values, option: Option) -> Any: """Get a format_control object.""" return getattr(values, option.dest) -def _handle_no_binary(option, opt_str, value, parser): - # type: (Option, str, str, OptionParser) -> None +def _handle_no_binary( + option: Option, opt_str: str, value: str, parser: OptionParser +) -> None: existing = _get_format_control(parser.values, option) FormatControl.handle_mutual_excludes( - value, existing.no_binary, existing.only_binary, + value, + existing.no_binary, + existing.only_binary, ) -def _handle_only_binary(option, opt_str, value, parser): - # type: (Option, str, str, OptionParser) -> None +def _handle_only_binary( + option: Option, opt_str: str, value: str, parser: OptionParser +) -> None: existing = _get_format_control(parser.values, option) FormatControl.handle_mutual_excludes( - value, existing.only_binary, existing.no_binary, + value, + existing.only_binary, + existing.no_binary, ) -def no_binary(): - # type: () -> Option +def no_binary() -> Option: format_control = FormatControl(set(), set()) return Option( - "--no-binary", dest="format_control", action="callback", - callback=_handle_no_binary, type="str", + "--no-binary", + dest="format_control", + action="callback", + callback=_handle_no_binary, + type="str", default=format_control, - help='Do not use binary packages. Can be supplied multiple times, and ' - 'each time adds to the existing value. Accepts either ":all:" to ' - 'disable all binary packages, ":none:" to empty the set (notice ' - 'the colons), or one or more package names with commas between ' - 'them (no colons). Note that some packages are tricky to compile ' - 'and may fail to install when this option is used on them.', + help="Do not use binary packages. Can be supplied multiple times, and " + 'each time adds to the existing value. Accepts either ":all:" to ' + 'disable all binary packages, ":none:" to empty the set (notice ' + "the colons), or one or more package names with commas between " + "them (no colons). Note that some packages are tricky to compile " + "and may fail to install when this option is used on them.", ) -def only_binary(): - # type: () -> Option +def only_binary() -> Option: format_control = FormatControl(set(), set()) return Option( - "--only-binary", dest="format_control", action="callback", - callback=_handle_only_binary, type="str", + "--only-binary", + dest="format_control", + action="callback", + callback=_handle_only_binary, + type="str", default=format_control, - help='Do not use source packages. Can be supplied multiple times, and ' - 'each time adds to the existing value. Accepts either ":all:" to ' - 'disable all source packages, ":none:" to empty the set, or one ' - 'or more package names with commas between them. Packages ' - 'without binary distributions will fail to install when this ' - 'option is used on them.', + help="Do not use source packages. Can be supplied multiple times, and " + 'each time adds to the existing value. Accepts either ":all:" to ' + 'disable all source packages, ":none:" to empty the set, or one ' + "or more package names with commas between them. Packages " + "without binary distributions will fail to install when this " + "option is used on them.", ) -platform = partial( +platforms: Callable[..., Option] = partial( Option, - '--platform', - dest='platform', - metavar='platform', + "--platform", + dest="platforms", + metavar="platform", + action="append", default=None, - help=("Only use wheels compatible with . " - "Defaults to the platform of the running system."), -) # type: Callable[..., Option] + help=( + "Only use wheels compatible with . Defaults to the " + "platform of the running system. Use this option multiple times to " + "specify multiple platforms supported by the target interpreter." + ), +) # This was made a separate function for unit-testing purposes. -def _convert_python_version(value): - # type: (str) -> Tuple[Tuple[int, ...], Optional[str]] +def _convert_python_version(value: str) -> Tuple[Tuple[int, ...], Optional[str]]: """ Convert a version string like "3", "37", or "3.7.3" into a tuple of ints. @@ -516,9 +556,9 @@ def _convert_python_version(value): # The empty string is the same as not providing a value. return (None, None) - parts = value.split('.') + parts = value.split(".") if len(parts) > 3: - return ((), 'at most three version parts are allowed') + return ((), "at most three version parts are allowed") if len(parts) == 1: # Then we are in the case of "3" or "37". @@ -529,119 +569,125 @@ def _convert_python_version(value): try: version_info = tuple(int(part) for part in parts) except ValueError: - return ((), 'each version part must be an integer') + return ((), "each version part must be an integer") return (version_info, None) -def _handle_python_version(option, opt_str, value, parser): - # type: (Option, str, str, OptionParser) -> None +def _handle_python_version( + option: Option, opt_str: str, value: str, parser: OptionParser +) -> None: """ Handle a provided --python-version value. """ version_info, error_msg = _convert_python_version(value) if error_msg is not None: - msg = ( - 'invalid --python-version value: {!r}: {}'.format( - value, error_msg, - ) - ) + msg = f"invalid --python-version value: {value!r}: {error_msg}" raise_option_error(parser, option=option, msg=msg) parser.values.python_version = version_info -python_version = partial( +python_version: Callable[..., Option] = partial( Option, - '--python-version', - dest='python_version', - metavar='python_version', - action='callback', - callback=_handle_python_version, type='str', + "--python-version", + dest="python_version", + metavar="python_version", + action="callback", + callback=_handle_python_version, + type="str", default=None, - help=dedent("""\ + help=dedent( + """\ The Python interpreter version to use for wheel and "Requires-Python" compatibility checks. Defaults to a version derived from the running interpreter. The version can be specified using up to three dot-separated integers (e.g. "3" for 3.0.0, "3.7" for 3.7.0, or "3.7.3"). A major-minor version can also be given as a string without dots (e.g. "37" for 3.7.0). - """), -) # type: Callable[..., Option] + """ + ), +) -implementation = partial( +implementation: Callable[..., Option] = partial( Option, - '--implementation', - dest='implementation', - metavar='implementation', + "--implementation", + dest="implementation", + metavar="implementation", default=None, - help=("Only use wheels compatible with Python " - "implementation , e.g. 'pp', 'jy', 'cp', " - " or 'ip'. If not specified, then the current " - "interpreter implementation is used. Use 'py' to force " - "implementation-agnostic wheels."), -) # type: Callable[..., Option] + help=( + "Only use wheels compatible with Python " + "implementation , e.g. 'pp', 'jy', 'cp', " + " or 'ip'. If not specified, then the current " + "interpreter implementation is used. Use 'py' to force " + "implementation-agnostic wheels." + ), +) -abi = partial( +abis: Callable[..., Option] = partial( Option, - '--abi', - dest='abi', - metavar='abi', + "--abi", + dest="abis", + metavar="abi", + action="append", default=None, - help=("Only use wheels compatible with Python " - "abi , e.g. 'pypy_41'. If not specified, then the " - "current interpreter abi tag is used. Generally " - "you will need to specify --implementation, " - "--platform, and --python-version when using " - "this option."), -) # type: Callable[..., Option] - - -def add_target_python_options(cmd_opts): - # type: (OptionGroup) -> None - cmd_opts.add_option(platform()) + help=( + "Only use wheels compatible with Python abi , e.g. 'pypy_41'. " + "If not specified, then the current interpreter abi tag is used. " + "Use this option multiple times to specify multiple abis supported " + "by the target interpreter. Generally you will need to specify " + "--implementation, --platform, and --python-version when using this " + "option." + ), +) + + +def add_target_python_options(cmd_opts: OptionGroup) -> None: + cmd_opts.add_option(platforms()) cmd_opts.add_option(python_version()) cmd_opts.add_option(implementation()) - cmd_opts.add_option(abi()) + cmd_opts.add_option(abis()) -def make_target_python(options): - # type: (Values) -> TargetPython +def make_target_python(options: Values) -> TargetPython: target_python = TargetPython( - platform=options.platform, + platforms=options.platforms, py_version_info=options.python_version, - abi=options.abi, + abis=options.abis, implementation=options.implementation, ) return target_python -def prefer_binary(): - # type: () -> Option +def prefer_binary() -> Option: return Option( "--prefer-binary", dest="prefer_binary", action="store_true", default=False, - help="Prefer older binary packages over newer source packages." + help=( + "Prefer binary packages over source packages, even if the " + "source packages are newer." + ), ) -cache_dir = partial( +cache_dir: Callable[..., Option] = partial( PipOption, "--cache-dir", dest="cache_dir", default=USER_CACHE_DIR, metavar="dir", - type='path', - help="Store the cache data in ." -) # type: Callable[..., Option] + type="path", + help="Store the cache data in .", +) -def _handle_no_cache_dir(option, opt, value, parser): - # type: (Option, str, str, OptionParser) -> None +def _handle_no_cache_dir( + option: Option, opt: str, value: str, parser: OptionParser +) -> None: """ Process a value provided for the --no-cache-dir option. @@ -668,69 +714,57 @@ def _handle_no_cache_dir(option, opt, value, parser): parser.values.cache_dir = False -no_cache = partial( +no_cache: Callable[..., Option] = partial( Option, "--no-cache-dir", dest="cache_dir", action="callback", callback=_handle_no_cache_dir, help="Disable the cache.", -) # type: Callable[..., Option] +) -no_deps = partial( +no_deps: Callable[..., Option] = partial( Option, - '--no-deps', '--no-dependencies', - dest='ignore_dependencies', - action='store_true', + "--no-deps", + "--no-dependencies", + dest="ignore_dependencies", + action="store_true", default=False, help="Don't install package dependencies.", -) # type: Callable[..., Option] - - -def _handle_build_dir(option, opt, value, parser): - # type: (Option, str, str, OptionParser) -> None - if value: - value = os.path.abspath(value) - setattr(parser.values, option.dest, value) +) +ignore_requires_python: Callable[..., Option] = partial( + Option, + "--ignore-requires-python", + dest="ignore_requires_python", + action="store_true", + help="Ignore the Requires-Python information.", +) -build_dir = partial( - PipOption, - '-b', '--build', '--build-dir', '--build-directory', - dest='build_dir', - type='path', - metavar='dir', - action='callback', - callback=_handle_build_dir, - help='Directory to unpack packages into and build in. Note that ' - 'an initial build still takes place in a temporary directory. ' - 'The location of temporary directories can be controlled by setting ' - 'the TMPDIR environment variable (TEMP on Windows) appropriately. ' - 'When passed, build directories are not cleaned in case of failures.' -) # type: Callable[..., Option] - -ignore_requires_python = partial( - Option, - '--ignore-requires-python', - dest='ignore_requires_python', - action='store_true', - help='Ignore the Requires-Python information.' -) # type: Callable[..., Option] - -no_build_isolation = partial( - Option, - '--no-build-isolation', - dest='build_isolation', - action='store_false', +no_build_isolation: Callable[..., Option] = partial( + Option, + "--no-build-isolation", + dest="build_isolation", + action="store_false", default=True, - help='Disable isolation when building a modern source distribution. ' - 'Build dependencies specified by PEP 518 must be already installed ' - 'if this option is used.' -) # type: Callable[..., Option] + help="Disable isolation when building a modern source distribution. " + "Build dependencies specified by PEP 518 must be already installed " + "if this option is used.", +) + +check_build_deps: Callable[..., Option] = partial( + Option, + "--check-build-dependencies", + dest="check_build_deps", + action="store_true", + default=False, + help="Check the build dependencies when PEP517 is used.", +) -def _handle_no_use_pep517(option, opt, value, parser): - # type: (Option, str, str, OptionParser) -> None +def _handle_no_use_pep517( + option: Option, opt: str, value: str, parser: OptionParser +) -> None: """ Process a value provided for the --no-use-pep517 option. @@ -749,192 +783,270 @@ def _handle_no_use_pep517(option, opt, value, parser): """ raise_option_error(parser, option=option, msg=msg) + # If user doesn't wish to use pep517, we check if setuptools and wheel are installed + # and raise error if it is not. + packages = ("setuptools", "wheel") + if not all(importlib.util.find_spec(package) for package in packages): + msg = ( + f"It is not possible to use --no-use-pep517 " + f"without {' and '.join(packages)} installed." + ) + raise_option_error(parser, option=option, msg=msg) + # Otherwise, --no-use-pep517 was passed via the command-line. parser.values.use_pep517 = False -use_pep517 = partial( +use_pep517: Any = partial( Option, - '--use-pep517', - dest='use_pep517', - action='store_true', + "--use-pep517", + dest="use_pep517", + action="store_true", default=None, - help='Use PEP 517 for building source distributions ' - '(use --no-use-pep517 to force legacy behaviour).' -) # type: Any + help="Use PEP 517 for building source distributions " + "(use --no-use-pep517 to force legacy behaviour).", +) -no_use_pep517 = partial( +no_use_pep517: Any = partial( Option, - '--no-use-pep517', - dest='use_pep517', - action='callback', + "--no-use-pep517", + dest="use_pep517", + action="callback", callback=_handle_no_use_pep517, default=None, - help=SUPPRESS_HELP -) # type: Any - -install_options = partial( - Option, - '--install-option', - dest='install_options', - action='append', - metavar='options', - help="Extra arguments to be supplied to the setup.py install " - "command (use like --install-option=\"--install-scripts=/usr/local/" - "bin\"). Use multiple --install-option options to pass multiple " - "options to setup.py install. If you are using an option with a " - "directory path, be sure to use absolute path.", -) # type: Callable[..., Option] - -global_options = partial( - Option, - '--global-option', - dest='global_options', - action='append', - metavar='options', + help=SUPPRESS_HELP, +) + + +def _handle_config_settings( + option: Option, opt_str: str, value: str, parser: OptionParser +) -> None: + key, sep, val = value.partition("=") + if sep != "=": + parser.error(f"Arguments to {opt_str} must be of the form KEY=VAL") + dest = getattr(parser.values, option.dest) + if dest is None: + dest = {} + setattr(parser.values, option.dest, dest) + if key in dest: + if isinstance(dest[key], list): + dest[key].append(val) + else: + dest[key] = [dest[key], val] + else: + dest[key] = val + + +config_settings: Callable[..., Option] = partial( + Option, + "-C", + "--config-settings", + dest="config_settings", + type=str, + action="callback", + callback=_handle_config_settings, + metavar="settings", + help="Configuration settings to be passed to the PEP 517 build backend. " + "Settings take the form KEY=VALUE. Use multiple --config-settings options " + "to pass multiple keys to the backend.", +) + +build_options: Callable[..., Option] = partial( + Option, + "--build-option", + dest="build_options", + metavar="options", + action="append", + help="Extra arguments to be supplied to 'setup.py bdist_wheel'.", +) + +global_options: Callable[..., Option] = partial( + Option, + "--global-option", + dest="global_options", + action="append", + metavar="options", help="Extra global options to be supplied to the setup.py " - "call before the install command.", -) # type: Callable[..., Option] + "call before the install or bdist_wheel command.", +) -no_clean = partial( +no_clean: Callable[..., Option] = partial( Option, - '--no-clean', - action='store_true', + "--no-clean", + action="store_true", default=False, - help="Don't clean up build directories." -) # type: Callable[..., Option] + help="Don't clean up build directories.", +) -pre = partial( +pre: Callable[..., Option] = partial( Option, - '--pre', - action='store_true', + "--pre", + action="store_true", default=False, help="Include pre-release and development versions. By default, " - "pip only finds stable versions.", -) # type: Callable[..., Option] + "pip only finds stable versions.", +) -disable_pip_version_check = partial( +disable_pip_version_check: Callable[..., Option] = partial( Option, "--disable-pip-version-check", dest="disable_pip_version_check", action="store_true", default=False, help="Don't periodically check PyPI to determine whether a new version " - "of pip is available for download. Implied with --no-index.", -) # type: Callable[..., Option] - + "of pip is available for download. Implied with --no-index.", +) -# Deprecated, Remove later -always_unzip = partial( +root_user_action: Callable[..., Option] = partial( Option, - '-Z', '--always-unzip', - dest='always_unzip', - action='store_true', - help=SUPPRESS_HELP, -) # type: Callable[..., Option] + "--root-user-action", + dest="root_user_action", + default="warn", + choices=["warn", "ignore"], + help="Action if pip is run as a root user [warn, ignore] (default: warn)", +) -def _handle_merge_hash(option, opt_str, value, parser): - # type: (Option, str, str, OptionParser) -> None +def _handle_merge_hash( + option: Option, opt_str: str, value: str, parser: OptionParser +) -> None: """Given a value spelled "algo:digest", append the digest to a list pointed to in a dict by the algo name.""" if not parser.values.hashes: parser.values.hashes = {} try: - algo, digest = value.split(':', 1) + algo, digest = value.split(":", 1) except ValueError: - parser.error('Arguments to {} must be a hash name ' - 'followed by a value, like --hash=sha256:' - 'abcde...'.format(opt_str)) + parser.error( + f"Arguments to {opt_str} must be a hash name " + "followed by a value, like --hash=sha256:" + "abcde..." + ) if algo not in STRONG_HASHES: - parser.error('Allowed hash algorithms for {} are {}.'.format( - opt_str, ', '.join(STRONG_HASHES))) + parser.error( + "Allowed hash algorithms for {} are {}.".format( + opt_str, ", ".join(STRONG_HASHES) + ) + ) parser.values.hashes.setdefault(algo, []).append(digest) -hash = partial( +hash: Callable[..., Option] = partial( Option, - '--hash', + "--hash", # Hash values eventually end up in InstallRequirement.hashes due to # __dict__ copying in process_line(). - dest='hashes', - action='callback', + dest="hashes", + action="callback", callback=_handle_merge_hash, - type='string', + type="string", help="Verify that the package's archive matches this " - 'hash before installing. Example: --hash=sha256:abcdef...', -) # type: Callable[..., Option] + "hash before installing. Example: --hash=sha256:abcdef...", +) -require_hashes = partial( +require_hashes: Callable[..., Option] = partial( Option, - '--require-hashes', - dest='require_hashes', - action='store_true', + "--require-hashes", + dest="require_hashes", + action="store_true", default=False, - help='Require a hash to check each requirement against, for ' - 'repeatable installs. This option is implied when any package in a ' - 'requirements file has a --hash option.', -) # type: Callable[..., Option] + help="Require a hash to check each requirement against, for " + "repeatable installs. This option is implied when any package in a " + "requirements file has a --hash option.", +) -list_path = partial( +list_path: Callable[..., Option] = partial( PipOption, - '--path', - dest='path', - type='path', - action='append', - help='Restrict to the specified installation path for listing ' - 'packages (can be used multiple times).' -) # type: Callable[..., Option] + "--path", + dest="path", + type="path", + action="append", + help="Restrict to the specified installation path for listing " + "packages (can be used multiple times).", +) -def check_list_path_option(options): - # type: (Values) -> None +def check_list_path_option(options: Values) -> None: if options.path and (options.user or options.local): - raise CommandError( - "Cannot combine '--path' with '--user' or '--local'" - ) + raise CommandError("Cannot combine '--path' with '--user' or '--local'") + + +list_exclude: Callable[..., Option] = partial( + PipOption, + "--exclude", + dest="excludes", + action="append", + metavar="package", + type="package_name", + help="Exclude specified package from the output", +) -no_python_version_warning = partial( +no_python_version_warning: Callable[..., Option] = partial( Option, - '--no-python-version-warning', - dest='no_python_version_warning', - action='store_true', + "--no-python-version-warning", + dest="no_python_version_warning", + action="store_true", default=False, - help='Silence deprecation warnings for upcoming unsupported Pythons.', -) # type: Callable[..., Option] + help="Silence deprecation warnings for upcoming unsupported Pythons.", +) + +# Features that are now always on. A warning is printed if they are used. +ALWAYS_ENABLED_FEATURES = [ + "truststore", # always on since 24.2 + "no-binary-enable-wheel-cache", # always on since 23.1 +] -unstable_feature = partial( +use_new_feature: Callable[..., Option] = partial( Option, - '--unstable-feature', - dest='unstable_features', - metavar='feature', - action='append', + "--use-feature", + dest="features_enabled", + metavar="feature", + action="append", default=[], - choices=['resolver'], - help=SUPPRESS_HELP, # TODO: Enable this when the resolver actually works. - # help='Enable unstable feature(s) that may be backward incompatible.', -) # type: Callable[..., Option] + choices=[ + "fast-deps", + ] + + ALWAYS_ENABLED_FEATURES, + help="Enable new functionality, that may be backward incompatible.", +) + +use_deprecated_feature: Callable[..., Option] = partial( + Option, + "--use-deprecated", + dest="deprecated_features_enabled", + metavar="feature", + action="append", + default=[], + choices=[ + "legacy-resolver", + "legacy-certs", + ], + help=("Enable deprecated functionality, that will be removed in the future."), +) ########## # groups # ########## -general_group = { - 'name': 'General Options', - 'options': [ +general_group: Dict[str, Any] = { + "name": "General Options", + "options": [ help_, + debug_mode, isolated_mode, require_virtualenv, + python, verbose, version, quiet, log, no_input, + keyring_provider, proxy, retries, timeout, @@ -947,16 +1059,17 @@ def check_list_path_option(options): disable_pip_version_check, no_color, no_python_version_warning, - unstable_feature, - ] -} # type: Dict[str, Any] - -index_group = { - 'name': 'Package Index Options', - 'options': [ + use_new_feature, + use_deprecated_feature, + ], +} + +index_group: Dict[str, Any] = { + "name": "Package Index Options", + "options": [ index_url, extra_index_url, no_index, find_links, - ] -} # type: Dict[str, Any] + ], +} diff --git a/src/fetchcode/vcs/pip/_internal/cli/command_context.py b/src/fetchcode/vcs/pip/_internal/cli/command_context.py index 2cd6e8db..139995ac 100644 --- a/src/fetchcode/vcs/pip/_internal/cli/command_context.py +++ b/src/fetchcode/vcs/pip/_internal/cli/command_context.py @@ -1,25 +1,17 @@ -from contextlib import contextmanager +from contextlib import ExitStack, contextmanager +from typing import ContextManager, Generator, TypeVar -from fetchcode.vcs.pip._vendor.contextlib2 import ExitStack +_T = TypeVar("_T", covariant=True) -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -if MYPY_CHECK_RUNNING: - from typing import Iterator, ContextManager, TypeVar - - _T = TypeVar('_T', covariant=True) - - -class CommandContextMixIn(object): - def __init__(self): - # type: () -> None - super(CommandContextMixIn, self).__init__() +class CommandContextMixIn: + def __init__(self) -> None: + super().__init__() self._in_main_context = False self._main_context = ExitStack() @contextmanager - def main_context(self): - # type: () -> Iterator[None] + def main_context(self) -> Generator[None, None, None]: assert not self._in_main_context self._in_main_context = True @@ -29,8 +21,7 @@ def main_context(self): finally: self._in_main_context = False - def enter_context(self, context_provider): - # type: (ContextManager[_T]) -> _T + def enter_context(self, context_provider: ContextManager[_T]) -> _T: assert self._in_main_context return self._main_context.enter_context(context_provider) diff --git a/src/fetchcode/vcs/pip/_internal/cli/index_command.py b/src/fetchcode/vcs/pip/_internal/cli/index_command.py new file mode 100644 index 00000000..226f8da1 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/cli/index_command.py @@ -0,0 +1,170 @@ +""" +Contains command classes which may interact with an index / the network. + +Unlike its sister module, req_command, this module still uses lazy imports +so commands which don't always hit the network (e.g. list w/o --outdated or +--uptodate) don't need waste time importing PipSession and friends. +""" + +import logging +import os +import sys +from optparse import Values +from typing import TYPE_CHECKING, List, Optional + +from pip._vendor import certifi + +from pip._internal.cli.base_command import Command +from pip._internal.cli.command_context import CommandContextMixIn + +if TYPE_CHECKING: + from ssl import SSLContext + + from pip._internal.network.session import PipSession + +logger = logging.getLogger(__name__) + + +def _create_truststore_ssl_context() -> Optional["SSLContext"]: + if sys.version_info < (3, 10): + logger.debug("Disabling truststore because Python version isn't 3.10+") + return None + + try: + import ssl + except ImportError: + logger.warning("Disabling truststore since ssl support is missing") + return None + + try: + from pip._vendor import truststore + except ImportError: + logger.warning("Disabling truststore because platform isn't supported") + return None + + ctx = truststore.SSLContext(ssl.PROTOCOL_TLS_CLIENT) + ctx.load_verify_locations(certifi.where()) + return ctx + + +class SessionCommandMixin(CommandContextMixIn): + """ + A class mixin for command classes needing _build_session(). + """ + + def __init__(self) -> None: + super().__init__() + self._session: Optional["PipSession"] = None + + @classmethod + def _get_index_urls(cls, options: Values) -> Optional[List[str]]: + """Return a list of index urls from user-provided options.""" + index_urls = [] + if not getattr(options, "no_index", False): + url = getattr(options, "index_url", None) + if url: + index_urls.append(url) + urls = getattr(options, "extra_index_urls", None) + if urls: + index_urls.extend(urls) + # Return None rather than an empty list + return index_urls or None + + def get_default_session(self, options: Values) -> "PipSession": + """Get a default-managed session.""" + if self._session is None: + self._session = self.enter_context(self._build_session(options)) + # there's no type annotation on requests.Session, so it's + # automatically ContextManager[Any] and self._session becomes Any, + # then https://github.com/python/mypy/issues/7696 kicks in + assert self._session is not None + return self._session + + def _build_session( + self, + options: Values, + retries: Optional[int] = None, + timeout: Optional[int] = None, + ) -> "PipSession": + from pip._internal.network.session import PipSession + + cache_dir = options.cache_dir + assert not cache_dir or os.path.isabs(cache_dir) + + if "legacy-certs" not in options.deprecated_features_enabled: + ssl_context = _create_truststore_ssl_context() + else: + ssl_context = None + + session = PipSession( + cache=os.path.join(cache_dir, "http-v2") if cache_dir else None, + retries=retries if retries is not None else options.retries, + trusted_hosts=options.trusted_hosts, + index_urls=self._get_index_urls(options), + ssl_context=ssl_context, + ) + + # Handle custom ca-bundles from the user + if options.cert: + session.verify = options.cert + + # Handle SSL client certificate + if options.client_cert: + session.cert = options.client_cert + + # Handle timeouts + if options.timeout or timeout: + session.timeout = timeout if timeout is not None else options.timeout + + # Handle configured proxies + if options.proxy: + session.proxies = { + "http": options.proxy, + "https": options.proxy, + } + session.trust_env = False + + # Determine if we can prompt the user for authentication or not + session.auth.prompting = not options.no_input + session.auth.keyring_provider = options.keyring_provider + + return session + + +def _pip_self_version_check(session: "PipSession", options: Values) -> None: + from pip._internal.self_outdated_check import pip_self_version_check as check + + check(session, options) + + +class IndexGroupCommand(Command, SessionCommandMixin): + """ + Abstract base class for commands with the index_group options. + + This also corresponds to the commands that permit the pip version check. + """ + + def handle_pip_version_check(self, options: Values) -> None: + """ + Do the pip version check if not disabled. + + This overrides the default behavior of not doing the check. + """ + # Make sure the index_group options are present. + assert hasattr(options, "no_index") + + if options.disable_pip_version_check or options.no_index: + return + + try: + # Otherwise, check if we're using the latest version of pip available. + session = self._build_session( + options, + retries=0, + timeout=min(5, options.timeout), + ) + with session: + _pip_self_version_check(session, options) + except Exception: + logger.warning("There was an error checking the latest version of pip.") + logger.debug("See below for error", exc_info=True) diff --git a/src/fetchcode/vcs/pip/_internal/cli/main.py b/src/fetchcode/vcs/pip/_internal/cli/main.py index 2adc761c..563ac79c 100644 --- a/src/fetchcode/vcs/pip/_internal/cli/main.py +++ b/src/fetchcode/vcs/pip/_internal/cli/main.py @@ -1,21 +1,18 @@ """Primary application entrypoint. """ -from __future__ import absolute_import import locale import logging import os import sys +import warnings +from typing import List, Optional -from fetchcode.vcs.pip._internal.cli.autocompletion import autocomplete -from fetchcode.vcs.pip._internal.cli.main_parser import parse_command -from fetchcode.vcs.pip._internal.commands import create_command -from fetchcode.vcs.pip._internal.exceptions import PipError -from fetchcode.vcs.pip._internal.utils import deprecation -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import List, Optional +from pip._internal.cli.autocompletion import autocomplete +from pip._internal.cli.main_parser import parse_command +from pip._internal.commands import create_command +from pip._internal.exceptions import PipError +from pip._internal.utils import deprecation logger = logging.getLogger(__name__) @@ -46,11 +43,19 @@ # call to main. As it is not safe to do any processing after calling # main, this should not be an issue in practice. -def main(args=None): - # type: (Optional[List[str]]) -> int + +def main(args: Optional[List[str]] = None) -> int: if args is None: args = sys.argv[1:] + # Suppress the pkg_resources deprecation warning + # Note - we use a module of .*pkg_resources to cover + # the normal case (pip._vendor.pkg_resources) and the + # devendored case (a bare pkg_resources) + warnings.filterwarnings( + action="ignore", category=DeprecationWarning, module=".*pkg_resources" + ) + # Configure our deprecation warnings to be sent through loggers deprecation.install_warning_logger() @@ -59,14 +64,14 @@ def main(args=None): try: cmd_name, cmd_args = parse_command(args) except PipError as exc: - sys.stderr.write("ERROR: {}".format(exc)) + sys.stderr.write(f"ERROR: {exc}") sys.stderr.write(os.linesep) sys.exit(1) # Needed for locale.getpreferredencoding(False) to work # in pip._internal.utils.encoding.auto_decode try: - locale.setlocale(locale.LC_ALL, '') + locale.setlocale(locale.LC_ALL, "") except locale.Error as e: # setlocale can apparently crash if locale are uninitialized logger.debug("Ignoring error %s when setting locale", e) diff --git a/src/fetchcode/vcs/pip/_internal/cli/main_parser.py b/src/fetchcode/vcs/pip/_internal/cli/main_parser.py index b464f43e..5ade356b 100644 --- a/src/fetchcode/vcs/pip/_internal/cli/main_parser.py +++ b/src/fetchcode/vcs/pip/_internal/cli/main_parser.py @@ -2,39 +2,30 @@ """ import os +import subprocess import sys +from typing import List, Optional, Tuple -from fetchcode.vcs.pip._internal.cli import cmdoptions -from fetchcode.vcs.pip._internal.cli.parser import ( - ConfigOptionParser, - UpdatingDefaultsHelpFormatter, -) -from fetchcode.vcs.pip._internal.commands import commands_dict, get_similar_commands -from fetchcode.vcs.pip._internal.exceptions import CommandError -from fetchcode.vcs.pip._internal.utils.misc import get_pip_version, get_prog -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import Tuple, List - +from pip._internal.build_env import get_runnable_pip +from pip._internal.cli import cmdoptions +from pip._internal.cli.parser import ConfigOptionParser, UpdatingDefaultsHelpFormatter +from pip._internal.commands import commands_dict, get_similar_commands +from pip._internal.exceptions import CommandError +from pip._internal.utils.misc import get_pip_version, get_prog __all__ = ["create_main_parser", "parse_command"] -def create_main_parser(): - # type: () -> ConfigOptionParser - """Creates and returns the main parser for pip's CLI - """ +def create_main_parser() -> ConfigOptionParser: + """Creates and returns the main parser for pip's CLI""" - parser_kw = { - 'usage': '\n%prog [options]', - 'add_help_option': False, - 'formatter': UpdatingDefaultsHelpFormatter(), - 'name': 'global', - 'prog': get_prog(), - } - - parser = ConfigOptionParser(**parser_kw) + parser = ConfigOptionParser( + usage="\n%prog [options]", + add_help_option=False, + formatter=UpdatingDefaultsHelpFormatter(), + name="global", + prog=get_prog(), + ) parser.disable_interspersed_args() parser.version = get_pip_version() @@ -47,17 +38,35 @@ def create_main_parser(): parser.main = True # type: ignore # create command listing for description - description = [''] + [ - '{name:27} {command_info.summary}'.format(**locals()) + description = [""] + [ + f"{name:27} {command_info.summary}" for name, command_info in commands_dict.items() ] - parser.description = '\n'.join(description) + parser.description = "\n".join(description) return parser -def parse_command(args): - # type: (List[str]) -> Tuple[str, List[str]] +def identify_python_interpreter(python: str) -> Optional[str]: + # If the named file exists, use it. + # If it's a directory, assume it's a virtual environment and + # look for the environment's Python executable. + if os.path.exists(python): + if os.path.isdir(python): + # bin/python for Unix, Scripts/python.exe for Windows + # Try both in case of odd cases like cygwin. + for exe in ("bin/python", "Scripts/python.exe"): + py = os.path.join(python, exe) + if os.path.exists(py): + return py + else: + return python + + # Could not find the interpreter specified + return None + + +def parse_command(args: List[str]) -> Tuple[str, List[str]]: parser = create_main_parser() # Note: parser calls disable_interspersed_args(), so the result of this @@ -69,14 +78,40 @@ def parse_command(args): # args_else: ['install', '--user', 'INITools'] general_options, args_else = parser.parse_args(args) + # --python + if general_options.python and "_PIP_RUNNING_IN_SUBPROCESS" not in os.environ: + # Re-invoke pip using the specified Python interpreter + interpreter = identify_python_interpreter(general_options.python) + if interpreter is None: + raise CommandError( + f"Could not locate Python interpreter {general_options.python}" + ) + + pip_cmd = [ + interpreter, + get_runnable_pip(), + ] + pip_cmd.extend(args) + + # Set a flag so the child doesn't re-invoke itself, causing + # an infinite loop. + os.environ["_PIP_RUNNING_IN_SUBPROCESS"] = "1" + returncode = 0 + try: + proc = subprocess.run(pip_cmd) + returncode = proc.returncode + except (subprocess.SubprocessError, OSError) as exc: + raise CommandError(f"Failed to run pip under {interpreter}: {exc}") + sys.exit(returncode) + # --version if general_options.version: - sys.stdout.write(parser.version) # type: ignore + sys.stdout.write(parser.version) sys.stdout.write(os.linesep) sys.exit() # pip || pip help -> print_help() - if not args_else or (args_else[0] == 'help' and len(args_else) == 1): + if not args_else or (args_else[0] == "help" and len(args_else) == 1): parser.print_help() sys.exit() @@ -86,11 +121,11 @@ def parse_command(args): if cmd_name not in commands_dict: guess = get_similar_commands(cmd_name) - msg = ['unknown command "{}"'.format(cmd_name)] + msg = [f'unknown command "{cmd_name}"'] if guess: - msg.append('maybe you meant "{}"'.format(guess)) + msg.append(f'maybe you meant "{guess}"') - raise CommandError(' - '.join(msg)) + raise CommandError(" - ".join(msg)) # all the args without the subcommand cmd_args = args[:] diff --git a/src/fetchcode/vcs/pip/_internal/cli/parser.py b/src/fetchcode/vcs/pip/_internal/cli/parser.py index 49d03fcc..b7d7c1f6 100644 --- a/src/fetchcode/vcs/pip/_internal/cli/parser.py +++ b/src/fetchcode/vcs/pip/_internal/cli/parser.py @@ -1,21 +1,16 @@ """Base option parser setup""" -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import - import logging import optparse +import shutil import sys import textwrap -from distutils.util import strtobool - -from fetchcode.vcs.pip._vendor.six import string_types +from contextlib import suppress +from typing import Any, Dict, Generator, List, Optional, Tuple -from fetchcode.vcs.pip._internal.cli.status_codes import UNKNOWN_ERROR -from fetchcode.vcs.pip._internal.configuration import Configuration, ConfigurationError -from fetchcode.vcs.pip._internal.utils.compat import get_terminal_size +from pip._internal.cli.status_codes import UNKNOWN_ERROR +from pip._internal.configuration import Configuration, ConfigurationError +from pip._internal.utils.misc import redact_auth_from_url, strtobool logger = logging.getLogger(__name__) @@ -23,17 +18,19 @@ class PrettyHelpFormatter(optparse.IndentedHelpFormatter): """A prettier/less verbose help formatter for optparse.""" - def __init__(self, *args, **kwargs): + def __init__(self, *args: Any, **kwargs: Any) -> None: # help position must be aligned with __init__.parseopts.description - kwargs['max_help_position'] = 30 - kwargs['indent_increment'] = 1 - kwargs['width'] = get_terminal_size()[0] - 2 - optparse.IndentedHelpFormatter.__init__(self, *args, **kwargs) + kwargs["max_help_position"] = 30 + kwargs["indent_increment"] = 1 + kwargs["width"] = shutil.get_terminal_size()[0] - 2 + super().__init__(*args, **kwargs) - def format_option_strings(self, option): + def format_option_strings(self, option: optparse.Option) -> str: return self._format_option_strings(option) - def _format_option_strings(self, option, mvarfmt=' <{}>', optsep=', '): + def _format_option_strings( + self, option: optparse.Option, mvarfmt: str = " <{}>", optsep: str = ", " + ) -> str: """ Return a comma-separated list of option strings and metavars. @@ -51,52 +48,52 @@ def _format_option_strings(self, option, mvarfmt=' <{}>', optsep=', '): opts.insert(1, optsep) if option.takes_value(): + assert option.dest is not None metavar = option.metavar or option.dest.lower() opts.append(mvarfmt.format(metavar.lower())) - return ''.join(opts) + return "".join(opts) - def format_heading(self, heading): - if heading == 'Options': - return '' - return heading + ':\n' + def format_heading(self, heading: str) -> str: + if heading == "Options": + return "" + return heading + ":\n" - def format_usage(self, usage): + def format_usage(self, usage: str) -> str: """ Ensure there is only one newline between usage and the first heading if there is no description. """ - msg = '\nUsage: {}\n'.format( - self.indent_lines(textwrap.dedent(usage), " ")) + msg = "\nUsage: {}\n".format(self.indent_lines(textwrap.dedent(usage), " ")) return msg - def format_description(self, description): + def format_description(self, description: Optional[str]) -> str: # leave full control over description to us if description: - if hasattr(self.parser, 'main'): - label = 'Commands' + if hasattr(self.parser, "main"): + label = "Commands" else: - label = 'Description' + label = "Description" # some doc strings have initial newlines, some don't - description = description.lstrip('\n') + description = description.lstrip("\n") # some doc strings have final newlines and spaces, some don't description = description.rstrip() # dedent, then reindent description = self.indent_lines(textwrap.dedent(description), " ") - description = '{}:\n{}\n'.format(label, description) + description = f"{label}:\n{description}\n" return description else: - return '' + return "" - def format_epilog(self, epilog): + def format_epilog(self, epilog: Optional[str]) -> str: # leave full control over epilog to us if epilog: return epilog else: - return '' + return "" - def indent_lines(self, text, indent): - new_lines = [indent + line for line in text.split('\n')] + def indent_lines(self, text: str, indent: str) -> str: + new_lines = [indent + line for line in text.split("\n")] return "\n".join(new_lines) @@ -105,17 +102,37 @@ class UpdatingDefaultsHelpFormatter(PrettyHelpFormatter): This is updates the defaults before expanding them, allowing them to show up correctly in the help listing. + + Also redact auth from url type options """ - def expand_default(self, option): + def expand_default(self, option: optparse.Option) -> str: + default_values = None if self.parser is not None: + assert isinstance(self.parser, ConfigOptionParser) self.parser._update_defaults(self.parser.defaults) - return optparse.IndentedHelpFormatter.expand_default(self, option) + assert option.dest is not None + default_values = self.parser.defaults.get(option.dest) + help_text = super().expand_default(option) + if default_values and option.metavar == "URL": + if isinstance(default_values, str): + default_values = [default_values] -class CustomOptionParser(optparse.OptionParser): + # If its not a list, we should abort and just return the help text + if not isinstance(default_values, list): + default_values = [] - def insert_option_group(self, idx, *args, **kwargs): + for val in default_values: + help_text = help_text.replace(val, redact_auth_from_url(val)) + + return help_text + + +class CustomOptionParser(optparse.OptionParser): + def insert_option_group( + self, idx: int, *args: Any, **kwargs: Any + ) -> optparse.OptionGroup: """Insert an OptionGroup at a given position.""" group = self.add_option_group(*args, **kwargs) @@ -125,7 +142,7 @@ def insert_option_group(self, idx, *args, **kwargs): return group @property - def option_list_all(self): + def option_list_all(self) -> List[optparse.Option]: """Get a list of all options, including those in option groups.""" res = self.option_list[:] for i in self.option_groups: @@ -138,34 +155,42 @@ class ConfigOptionParser(CustomOptionParser): """Custom option parser which updates its defaults by checking the configuration files and environmental variables""" - def __init__(self, *args, **kwargs): - self.name = kwargs.pop('name') - - isolated = kwargs.pop("isolated", False) + def __init__( + self, + *args: Any, + name: str, + isolated: bool = False, + **kwargs: Any, + ) -> None: + self.name = name self.config = Configuration(isolated) assert self.name - optparse.OptionParser.__init__(self, *args, **kwargs) + super().__init__(*args, **kwargs) - def check_default(self, option, key, val): + def check_default(self, option: optparse.Option, key: str, val: Any) -> Any: try: return option.check_value(key, val) except optparse.OptionValueError as exc: - print("An error occurred during configuration: {}".format(exc)) + print(f"An error occurred during configuration: {exc}") sys.exit(3) - def _get_ordered_configuration_items(self): + def _get_ordered_configuration_items( + self, + ) -> Generator[Tuple[str, Any], None, None]: # Configuration gives keys in an unordered manner. Order them. override_order = ["global", self.name, ":env:"] # Pool the options into different groups - section_items = {name: [] for name in override_order} + section_items: Dict[str, List[Tuple[str, Any]]] = { + name: [] for name in override_order + } for section_key, val in self.config.items(): # ignore empty values if not val: logger.debug( "Ignoring configuration key '%s' as it's value is empty.", - section_key + section_key, ) continue @@ -178,7 +203,7 @@ def _get_ordered_configuration_items(self): for key, val in section_items[section]: yield key, val - def _update_defaults(self, defaults): + def _update_defaults(self, defaults: Dict[str, Any]) -> Dict[str, Any]: """Updates the given defaults with values from the config files and the environ. Does a little special handling for certain types of options (lists).""" @@ -189,7 +214,7 @@ def _update_defaults(self, defaults): # Then set the options with those values for key, val in self._get_ordered_configuration_items(): # '--' because configuration supports only long names - option = self.get_option('--' + key) + option = self.get_option("--" + key) # Ignore options not present in this parser. E.g. non-globals put # in [global] by users that want them to apply to all applicable @@ -197,19 +222,34 @@ def _update_defaults(self, defaults): if option is None: continue - if option.action in ('store_true', 'store_false', 'count'): + assert option.dest is not None + + if option.action in ("store_true", "store_false"): try: val = strtobool(val) except ValueError: - error_msg = invalid_config_error_message( - option.action, key, val + self.error( + f"{val} is not a valid value for {key} option, " + "please specify a boolean value like yes/no, " + "true/false or 1/0 instead." ) - self.error(error_msg) - - elif option.action == 'append': + elif option.action == "count": + with suppress(ValueError): + val = strtobool(val) + with suppress(ValueError): + val = int(val) + if not isinstance(val, int) or val < 0: + self.error( + f"{val} is not a valid value for {key} option, " + "please instead specify either a non-negative integer " + "or a boolean value like yes/no or false/true " + "which is equivalent to 1/0." + ) + elif option.action == "append": val = val.split() val = [self.check_default(option, key, v) for v in val] - elif option.action == 'callback': + elif option.action == "callback": + assert option.callback is not None late_eval.add(option.dest) opt_str = option.get_opt_string() val = option.convert_value(opt_str, val) @@ -227,7 +267,7 @@ def _update_defaults(self, defaults): self.values = None return defaults - def get_default_values(self): + def get_default_values(self) -> optparse.Values: """Overriding to make updating the defaults after instantiation of the option parser possible, _update_defaults() does the dirty work.""" if not self.process_default_values: @@ -242,25 +282,13 @@ def get_default_values(self): defaults = self._update_defaults(self.defaults.copy()) # ours for option in self._get_all_options(): + assert option.dest is not None default = defaults.get(option.dest) - if isinstance(default, string_types): + if isinstance(default, str): opt_str = option.get_opt_string() defaults[option.dest] = option.check_value(opt_str, default) return optparse.Values(defaults) - def error(self, msg): + def error(self, msg: str) -> None: self.print_usage(sys.stderr) - self.exit(UNKNOWN_ERROR, "{}\n".format(msg)) - - -def invalid_config_error_message(action, key, val): - """Returns a better error message when invalid configuration option - is provided.""" - if action in ('store_true', 'store_false'): - return ("{0} is not a valid value for {1} option, " - "please specify a boolean value like yes/no, " - "true/false or 1/0 instead.").format(val, key) - - return ("{0} is not a valid value for {1} option, " - "please specify a numerical value like 1/0 " - "instead.").format(val, key) + self.exit(UNKNOWN_ERROR, f"{msg}\n") diff --git a/src/fetchcode/vcs/pip/_internal/cli/progress_bars.py b/src/fetchcode/vcs/pip/_internal/cli/progress_bars.py index fed44847..883359c9 100644 --- a/src/fetchcode/vcs/pip/_internal/cli/progress_bars.py +++ b/src/fetchcode/vcs/pip/_internal/cli/progress_bars.py @@ -1,277 +1,94 @@ -from __future__ import division - -import itertools +import functools import sys -from signal import SIGINT, default_int_handler, signal - -from fetchcode.vcs.pip._vendor import six -from fetchcode.vcs.pip._vendor.progress.bar import Bar, FillingCirclesBar, IncrementalBar -from fetchcode.vcs.pip._vendor.progress.spinner import Spinner - -from fetchcode.vcs.pip._internal.utils.compat import WINDOWS -from fetchcode.vcs.pip._internal.utils.logging import get_indentation -from fetchcode.vcs.pip._internal.utils.misc import format_size -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import Any, Dict, List - -try: - from fetchcode.vcs.pip._vendor import colorama -# Lots of different errors can come from this, including SystemError and -# ImportError. -except Exception: - colorama = None - - -def _select_progress_class(preferred, fallback): - # type: (Bar, Bar) -> Bar - encoding = getattr(preferred.file, "encoding", None) - - # If we don't know what encoding this file is in, then we'll just assume - # that it doesn't support unicode and use the ASCII bar. - if not encoding: - return fallback - - # Collect all of the possible characters we want to use with the preferred - # bar. - characters = [ - getattr(preferred, "empty_fill", six.text_type()), - getattr(preferred, "fill", six.text_type()), - ] - characters += list(getattr(preferred, "phases", [])) - - # Try to decode the characters we're using for the bar using the encoding - # of the given file, if this works then we'll assume that we can use the - # fancier bar and if not we'll fall back to the plaintext bar. - try: - six.text_type().join(characters).encode(encoding) - except UnicodeEncodeError: - return fallback - else: - return preferred - - -_BaseBar = _select_progress_class(IncrementalBar, Bar) # type: Any - - -class InterruptibleMixin(object): - """ - Helper to ensure that self.finish() gets called on keyboard interrupt. - - This allows downloads to be interrupted without leaving temporary state - (like hidden cursors) behind. - - This class is similar to the progress library's existing SigIntMixin - helper, but as of version 1.2, that helper has the following problems: - - 1. It calls sys.exit(). - 2. It discards the existing SIGINT handler completely. - 3. It leaves its own handler in place even after an uninterrupted finish, - which will have unexpected delayed effects if the user triggers an - unrelated keyboard interrupt some time after a progress-displaying - download has already completed, for example. - """ - - def __init__(self, *args, **kwargs): - # type: (List[Any], Dict[Any, Any]) -> None - """ - Save the original SIGINT handler for later. - """ - super(InterruptibleMixin, self).__init__( # type: ignore - *args, - **kwargs +from typing import Callable, Generator, Iterable, Iterator, Optional, Tuple + +from pip._vendor.rich.progress import ( + BarColumn, + DownloadColumn, + FileSizeColumn, + Progress, + ProgressColumn, + SpinnerColumn, + TextColumn, + TimeElapsedColumn, + TimeRemainingColumn, + TransferSpeedColumn, +) + +from pip._internal.cli.spinners import RateLimiter +from pip._internal.utils.logging import get_indentation + +DownloadProgressRenderer = Callable[[Iterable[bytes]], Iterator[bytes]] + + +def _rich_progress_bar( + iterable: Iterable[bytes], + *, + bar_type: str, + size: int, +) -> Generator[bytes, None, None]: + assert bar_type == "on", "This should only be used in the default mode." + + if not size: + total = float("inf") + columns: Tuple[ProgressColumn, ...] = ( + TextColumn("[progress.description]{task.description}"), + SpinnerColumn("line", speed=1.5), + FileSizeColumn(), + TransferSpeedColumn(), + TimeElapsedColumn(), ) - - self.original_handler = signal(SIGINT, self.handle_sigint) - - # If signal() returns None, the previous handler was not installed from - # Python, and we cannot restore it. This probably should not happen, - # but if it does, we must restore something sensible instead, at least. - # The least bad option should be Python's default SIGINT handler, which - # just raises KeyboardInterrupt. - if self.original_handler is None: - self.original_handler = default_int_handler - - def finish(self): - # type: () -> None - """ - Restore the original SIGINT handler after finishing. - - This should happen regardless of whether the progress display finishes - normally, or gets interrupted. - """ - super(InterruptibleMixin, self).finish() # type: ignore - signal(SIGINT, self.original_handler) - - def handle_sigint(self, signum, frame): # type: ignore - """ - Call self.finish() before delegating to the original SIGINT handler. - - This handler should only be in place while the progress display is - active. - """ - self.finish() - self.original_handler(signum, frame) - - -class SilentBar(Bar): - - def update(self): - # type: () -> None - pass - - -class BlueEmojiBar(IncrementalBar): - - suffix = "%(percent)d%%" - bar_prefix = " " - bar_suffix = " " - phases = (u"\U0001F539", u"\U0001F537", u"\U0001F535") # type: Any - - -class DownloadProgressMixin(object): - - def __init__(self, *args, **kwargs): - # type: (List[Any], Dict[Any, Any]) -> None - super(DownloadProgressMixin, self).__init__( # type: ignore - *args, - **kwargs + else: + total = size + columns = ( + TextColumn("[progress.description]{task.description}"), + BarColumn(), + DownloadColumn(), + TransferSpeedColumn(), + TextColumn("eta"), + TimeRemainingColumn(), ) - self.message = (" " * ( - get_indentation() + 2 - )) + self.message # type: str - - @property - def downloaded(self): - # type: () -> str - return format_size(self.index) # type: ignore - - @property - def download_speed(self): - # type: () -> str - # Avoid zero division errors... - if self.avg == 0.0: # type: ignore - return "..." - return format_size(1 / self.avg) + "/s" # type: ignore - - @property - def pretty_eta(self): - # type: () -> str - if self.eta: # type: ignore - return "eta {}".format(self.eta_td) # type: ignore - return "" - - def iter(self, it): # type: ignore - for x in it: - yield x - self.next(len(x)) - self.finish() - - -class WindowsMixin(object): - - def __init__(self, *args, **kwargs): - # type: (List[Any], Dict[Any, Any]) -> None - # The Windows terminal does not support the hide/show cursor ANSI codes - # even with colorama. So we'll ensure that hide_cursor is False on - # Windows. - # This call needs to go before the super() call, so that hide_cursor - # is set in time. The base progress bar class writes the "hide cursor" - # code to the terminal in its init, so if we don't set this soon - # enough, we get a "hide" with no corresponding "show"... - if WINDOWS and self.hide_cursor: # type: ignore - self.hide_cursor = False - - super(WindowsMixin, self).__init__(*args, **kwargs) # type: ignore - - # Check if we are running on Windows and we have the colorama module, - # if we do then wrap our file with it. - if WINDOWS and colorama: - self.file = colorama.AnsiToWin32(self.file) # type: ignore - # The progress code expects to be able to call self.file.isatty() - # but the colorama.AnsiToWin32() object doesn't have that, so we'll - # add it. - self.file.isatty = lambda: self.file.wrapped.isatty() - # The progress code expects to be able to call self.file.flush() - # but the colorama.AnsiToWin32() object doesn't have that, so we'll - # add it. - self.file.flush = lambda: self.file.wrapped.flush() - - -class BaseDownloadProgressBar(WindowsMixin, InterruptibleMixin, - DownloadProgressMixin): - - file = sys.stdout - message = "%(percent)d%%" - suffix = "%(downloaded)s %(download_speed)s %(pretty_eta)s" - -# NOTE: The "type: ignore" comments on the following classes are there to -# work around https://github.com/python/typing/issues/241 - - -class DefaultDownloadProgressBar(BaseDownloadProgressBar, - _BaseBar): - pass - -class DownloadSilentBar(BaseDownloadProgressBar, SilentBar): # type: ignore - pass - - -class DownloadBar(BaseDownloadProgressBar, # type: ignore - Bar): - pass - - -class DownloadFillingCirclesBar(BaseDownloadProgressBar, # type: ignore - FillingCirclesBar): - pass - - -class DownloadBlueEmojiProgressBar(BaseDownloadProgressBar, # type: ignore - BlueEmojiBar): - pass - - -class DownloadProgressSpinner(WindowsMixin, InterruptibleMixin, - DownloadProgressMixin, Spinner): - - file = sys.stdout - suffix = "%(downloaded)s %(download_speed)s" - - def next_phase(self): # type: ignore - if not hasattr(self, "_phaser"): - self._phaser = itertools.cycle(self.phases) - return next(self._phaser) - - def update(self): - # type: () -> None - message = self.message % self - phase = self.next_phase() - suffix = self.suffix % self - line = ''.join([ - message, - " " if message else "", - phase, - " " if suffix else "", - suffix, - ]) - - self.writeln(line) - - -BAR_TYPES = { - "off": (DownloadSilentBar, DownloadSilentBar), - "on": (DefaultDownloadProgressBar, DownloadProgressSpinner), - "ascii": (DownloadBar, DownloadProgressSpinner), - "pretty": (DownloadFillingCirclesBar, DownloadProgressSpinner), - "emoji": (DownloadBlueEmojiProgressBar, DownloadProgressSpinner) -} - - -def DownloadProgressProvider(progress_bar, max=None): # type: ignore - if max is None or max == 0: - return BAR_TYPES[progress_bar][1]().iter + progress = Progress(*columns, refresh_per_second=5) + task_id = progress.add_task(" " * (get_indentation() + 2), total=total) + with progress: + for chunk in iterable: + yield chunk + progress.update(task_id, advance=len(chunk)) + + +def _raw_progress_bar( + iterable: Iterable[bytes], + *, + size: Optional[int], +) -> Generator[bytes, None, None]: + def write_progress(current: int, total: int) -> None: + sys.stdout.write("Progress %d of %d\n" % (current, total)) + sys.stdout.flush() + + current = 0 + total = size or 0 + rate_limiter = RateLimiter(0.25) + + write_progress(current, total) + for chunk in iterable: + current += len(chunk) + if rate_limiter.ready() or current == total: + write_progress(current, total) + rate_limiter.reset() + yield chunk + + +def get_download_progress_renderer( + *, bar_type: str, size: Optional[int] = None +) -> DownloadProgressRenderer: + """Get an object that can be used to render the download progress. + + Returns a callable, that takes an iterable to "wrap". + """ + if bar_type == "on": + return functools.partial(_rich_progress_bar, bar_type=bar_type, size=size) + elif bar_type == "raw": + return functools.partial(_raw_progress_bar, size=size) else: - return BAR_TYPES[progress_bar][0](max=max).iter + return iter # no-op, when passed an iterator diff --git a/src/fetchcode/vcs/pip/_internal/cli/req_command.py b/src/fetchcode/vcs/pip/_internal/cli/req_command.py index f442ed8e..92900f94 100644 --- a/src/fetchcode/vcs/pip/_internal/cli/req_command.py +++ b/src/fetchcode/vcs/pip/_internal/cli/req_command.py @@ -1,163 +1,45 @@ -"""Contains the Command base classes that depend on PipSession. +"""Contains the RequirementCommand base class. -The classes in this module are in a separate module so the commands not -needing download / PackageFinder capability don't unnecessarily import the +This class is in a separate module so the commands that do not always +need PackageFinder capability don't unnecessarily import the PackageFinder machinery and all its vendored dependencies, etc. """ import logging -import os from functools import partial - -from fetchcode.vcs.pip._internal.cli import cmdoptions -from fetchcode.vcs.pip._internal.cli.base_command import Command -from fetchcode.vcs.pip._internal.cli.command_context import CommandContextMixIn -from fetchcode.vcs.pip._internal.exceptions import CommandError, PreviousBuildDirError -from fetchcode.vcs.pip._internal.index.package_finder import PackageFinder -from fetchcode.vcs.pip._internal.models.selection_prefs import SelectionPreferences -from fetchcode.vcs.pip._internal.network.download import Downloader -from fetchcode.vcs.pip._internal.network.session import PipSession -from fetchcode.vcs.pip._internal.operations.prepare import RequirementPreparer -from fetchcode.vcs.pip._internal.req.constructors import ( +from optparse import Values +from typing import Any, List, Optional, Tuple + +from pip._internal.cache import WheelCache +from pip._internal.cli import cmdoptions +from pip._internal.cli.index_command import IndexGroupCommand +from pip._internal.cli.index_command import SessionCommandMixin as SessionCommandMixin +from pip._internal.exceptions import CommandError, PreviousBuildDirError +from pip._internal.index.collector import LinkCollector +from pip._internal.index.package_finder import PackageFinder +from pip._internal.models.selection_prefs import SelectionPreferences +from pip._internal.models.target_python import TargetPython +from pip._internal.network.session import PipSession +from pip._internal.operations.build.build_tracker import BuildTracker +from pip._internal.operations.prepare import RequirementPreparer +from pip._internal.req.constructors import ( install_req_from_editable, install_req_from_line, install_req_from_parsed_requirement, install_req_from_req_string, ) -from fetchcode.vcs.pip._internal.req.req_file import parse_requirements -from fetchcode.vcs.pip._internal.req.req_set import RequirementSet -from fetchcode.vcs.pip._internal.self_outdated_check import ( - make_link_collector, - pip_self_version_check, +from pip._internal.req.req_file import parse_requirements +from pip._internal.req.req_install import InstallRequirement +from pip._internal.resolution.base import BaseResolver +from pip._internal.utils.temp_dir import ( + TempDirectory, + TempDirectoryTypeRegistry, + tempdir_kinds, ) -from fetchcode.vcs.pip._internal.utils.temp_dir import tempdir_kinds -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from optparse import Values - from typing import Any, List, Optional, Tuple - - from fetchcode.vcs.pip._internal.cache import WheelCache - from fetchcode.vcs.pip._internal.models.target_python import TargetPython - from fetchcode.vcs.pip._internal.req.req_install import InstallRequirement - from fetchcode.vcs.pip._internal.req.req_tracker import RequirementTracker - from fetchcode.vcs.pip._internal.resolution.base import BaseResolver - from fetchcode.vcs.pip._internal.utils.temp_dir import ( - TempDirectory, - TempDirectoryTypeRegistry, - ) - logger = logging.getLogger(__name__) -class SessionCommandMixin(CommandContextMixIn): - - """ - A class mixin for command classes needing _build_session(). - """ - def __init__(self): - # type: () -> None - super(SessionCommandMixin, self).__init__() - self._session = None # Optional[PipSession] - - @classmethod - def _get_index_urls(cls, options): - # type: (Values) -> Optional[List[str]] - """Return a list of index urls from user-provided options.""" - index_urls = [] - if not getattr(options, "no_index", False): - url = getattr(options, "index_url", None) - if url: - index_urls.append(url) - urls = getattr(options, "extra_index_urls", None) - if urls: - index_urls.extend(urls) - # Return None rather than an empty list - return index_urls or None - - def get_default_session(self, options): - # type: (Values) -> PipSession - """Get a default-managed session.""" - if self._session is None: - self._session = self.enter_context(self._build_session(options)) - # there's no type annotation on requests.Session, so it's - # automatically ContextManager[Any] and self._session becomes Any, - # then https://github.com/python/mypy/issues/7696 kicks in - assert self._session is not None - return self._session - - def _build_session(self, options, retries=None, timeout=None): - # type: (Values, Optional[int], Optional[int]) -> PipSession - assert not options.cache_dir or os.path.isabs(options.cache_dir) - session = PipSession( - cache=( - os.path.join(options.cache_dir, "http") - if options.cache_dir else None - ), - retries=retries if retries is not None else options.retries, - trusted_hosts=options.trusted_hosts, - index_urls=self._get_index_urls(options), - ) - - # Handle custom ca-bundles from the user - if options.cert: - session.verify = options.cert - - # Handle SSL client certificate - if options.client_cert: - session.cert = options.client_cert - - # Handle timeouts - if options.timeout or timeout: - session.timeout = ( - timeout if timeout is not None else options.timeout - ) - - # Handle configured proxies - if options.proxy: - session.proxies = { - "http": options.proxy, - "https": options.proxy, - } - - # Determine if we can prompt the user for authentication or not - session.auth.prompting = not options.no_input - - return session - - -class IndexGroupCommand(Command, SessionCommandMixin): - - """ - Abstract base class for commands with the index_group options. - - This also corresponds to the commands that permit the pip version check. - """ - - def handle_pip_version_check(self, options): - # type: (Values) -> None - """ - Do the pip version check if not disabled. - - This overrides the default behavior of not doing the check. - """ - # Make sure the index_group options are present. - assert hasattr(options, 'no_index') - - if options.disable_pip_version_check or options.no_index: - return - - # Otherwise, check if we're using the latest version of pip available. - session = self._build_session( - options, - retries=0, - timeout=min(5, options.timeout) - ) - with session: - pip_self_version_check(session, options) - - KEEPABLE_TEMPDIR_TYPES = [ tempdir_kinds.BUILD_ENV, tempdir_kinds.EPHEM_WHEEL_CACHE, @@ -165,18 +47,18 @@ def handle_pip_version_check(self, options): ] -def with_cleanup(func): - # type: (Any) -> Any +def with_cleanup(func: Any) -> Any: """Decorator for common logic related to managing temporary directories. """ - def configure_tempdir_registry(registry): - # type: (TempDirectoryTypeRegistry) -> None + + def configure_tempdir_registry(registry: TempDirectoryTypeRegistry) -> None: for t in KEEPABLE_TEMPDIR_TYPES: registry.set_delete(t, False) - def wrapper(self, options, args): - # type: (RequirementCommand, Values, List[Any]) -> Optional[int] + def wrapper( + self: RequirementCommand, options: Values, args: List[Any] + ) -> Optional[int]: assert self.tempdir_registry is not None if options.no_clean: configure_tempdir_registry(self.tempdir_registry) @@ -194,61 +76,89 @@ def wrapper(self, options, args): class RequirementCommand(IndexGroupCommand): - - def __init__(self, *args, **kw): - # type: (Any, Any) -> None - super(RequirementCommand, self).__init__(*args, **kw) + def __init__(self, *args: Any, **kw: Any) -> None: + super().__init__(*args, **kw) self.cmd_opts.add_option(cmdoptions.no_clean()) @staticmethod + def determine_resolver_variant(options: Values) -> str: + """Determines which resolver should be used, based on the given options.""" + if "legacy-resolver" in options.deprecated_features_enabled: + return "legacy" + + return "resolvelib" + + @classmethod def make_requirement_preparer( - temp_build_dir, # type: TempDirectory - options, # type: Values - req_tracker, # type: RequirementTracker - session, # type: PipSession - finder, # type: PackageFinder - use_user_site, # type: bool - download_dir=None, # type: str - wheel_download_dir=None, # type: str - ): - # type: (...) -> RequirementPreparer + cls, + temp_build_dir: TempDirectory, + options: Values, + build_tracker: BuildTracker, + session: PipSession, + finder: PackageFinder, + use_user_site: bool, + download_dir: Optional[str] = None, + verbosity: int = 0, + ) -> RequirementPreparer: """ Create a RequirementPreparer instance for the given parameters. """ - downloader = Downloader(session, progress_bar=options.progress_bar) - temp_build_dir_path = temp_build_dir.path assert temp_build_dir_path is not None + legacy_resolver = False + + resolver_variant = cls.determine_resolver_variant(options) + if resolver_variant == "resolvelib": + lazy_wheel = "fast-deps" in options.features_enabled + if lazy_wheel: + logger.warning( + "pip is using lazily downloaded wheels using HTTP " + "range requests to obtain dependency information. " + "This experimental feature is enabled through " + "--use-feature=fast-deps and it is not ready for " + "production." + ) + else: + legacy_resolver = True + lazy_wheel = False + if "fast-deps" in options.features_enabled: + logger.warning( + "fast-deps has no effect when used with the legacy resolver." + ) return RequirementPreparer( build_dir=temp_build_dir_path, src_dir=options.src_dir, download_dir=download_dir, - wheel_download_dir=wheel_download_dir, build_isolation=options.build_isolation, - req_tracker=req_tracker, - downloader=downloader, + check_build_deps=options.check_build_deps, + build_tracker=build_tracker, + session=session, + progress_bar=options.progress_bar, finder=finder, require_hashes=options.require_hashes, use_user_site=use_user_site, + lazy_wheel=lazy_wheel, + verbosity=verbosity, + legacy_resolver=legacy_resolver, ) - @staticmethod + @classmethod def make_resolver( - preparer, # type: RequirementPreparer - finder, # type: PackageFinder - options, # type: Values - wheel_cache=None, # type: Optional[WheelCache] - use_user_site=False, # type: bool - ignore_installed=True, # type: bool - ignore_requires_python=False, # type: bool - force_reinstall=False, # type: bool - upgrade_strategy="to-satisfy-only", # type: str - use_pep517=None, # type: Optional[bool] - py_version_info=None # type: Optional[Tuple[int, ...]] - ): - # type: (...) -> BaseResolver + cls, + preparer: RequirementPreparer, + finder: PackageFinder, + options: Values, + wheel_cache: Optional[WheelCache] = None, + use_user_site: bool = False, + ignore_installed: bool = True, + ignore_requires_python: bool = False, + force_reinstall: bool = False, + upgrade_strategy: str = "to-satisfy-only", + use_pep517: Optional[bool] = None, + py_version_info: Optional[Tuple[int, ...]] = None, + ) -> BaseResolver: """ Create a Resolver instance for the given parameters. """ @@ -257,11 +167,13 @@ def make_resolver( isolated=options.isolated_mode, use_pep517=use_pep517, ) + resolver_variant = cls.determine_resolver_variant(options) # The long import name and duplicated invocation is needed to convince # Mypy into correctly typechecking. Otherwise it would complain the # "Resolver" class being redefined. - if 'resolver' in options.unstable_features: - import fetchcode.vcs.pip._internal.resolution.resolvelib.resolver + if resolver_variant == "resolvelib": + import pip._internal.resolution.resolvelib.resolver + return pip._internal.resolution.resolvelib.resolver.Resolver( preparer=preparer, finder=finder, @@ -275,7 +187,8 @@ def make_resolver( upgrade_strategy=upgrade_strategy, py_version_info=py_version_info, ) - import fetchcode.vcs.pip._internal.resolution.legacy.resolver + import pip._internal.resolution.legacy.resolver + return pip._internal.resolution.legacy.resolver.Resolver( preparer=preparer, finder=finder, @@ -292,83 +205,92 @@ def make_resolver( def get_requirements( self, - args, # type: List[str] - options, # type: Values - finder, # type: PackageFinder - session, # type: PipSession - check_supported_wheels=True, # type: bool - ): - # type: (...) -> List[InstallRequirement] + args: List[str], + options: Values, + finder: PackageFinder, + session: PipSession, + ) -> List[InstallRequirement]: """ Parse command-line arguments into the corresponding requirements. """ - requirement_set = RequirementSet( - check_supported_wheels=check_supported_wheels - ) + requirements: List[InstallRequirement] = [] for filename in options.constraints: for parsed_req in parse_requirements( - filename, - constraint=True, finder=finder, options=options, - session=session): + filename, + constraint=True, + finder=finder, + options=options, + session=session, + ): req_to_add = install_req_from_parsed_requirement( parsed_req, isolated=options.isolated_mode, + user_supplied=False, ) - req_to_add.is_direct = True - requirement_set.add_requirement(req_to_add) + requirements.append(req_to_add) for req in args: req_to_add = install_req_from_line( - req, None, isolated=options.isolated_mode, + req, + comes_from=None, + isolated=options.isolated_mode, use_pep517=options.use_pep517, + user_supplied=True, + config_settings=getattr(options, "config_settings", None), ) - req_to_add.is_direct = True - requirement_set.add_requirement(req_to_add) + requirements.append(req_to_add) for req in options.editables: req_to_add = install_req_from_editable( req, + user_supplied=True, isolated=options.isolated_mode, use_pep517=options.use_pep517, + config_settings=getattr(options, "config_settings", None), ) - req_to_add.is_direct = True - requirement_set.add_requirement(req_to_add) + requirements.append(req_to_add) # NOTE: options.require_hashes may be set if --require-hashes is True for filename in options.requirements: for parsed_req in parse_requirements( - filename, - finder=finder, options=options, session=session): + filename, finder=finder, options=options, session=session + ): req_to_add = install_req_from_parsed_requirement( parsed_req, isolated=options.isolated_mode, - use_pep517=options.use_pep517 + use_pep517=options.use_pep517, + user_supplied=True, + config_settings=( + parsed_req.options.get("config_settings") + if parsed_req.options + else None + ), ) - req_to_add.is_direct = True - requirement_set.add_requirement(req_to_add) + requirements.append(req_to_add) # If any requirement has hash options, enable hash checking. - requirements = requirement_set.all_requirements if any(req.has_hash_options for req in requirements): options.require_hashes = True if not (args or options.editables or options.requirements): - opts = {'name': self.name} + opts = {"name": self.name} if options.find_links: raise CommandError( - 'You must give at least one requirement to {name} ' + "You must give at least one requirement to {name} " '(maybe you meant "pip {name} {links}"?)'.format( - **dict(opts, links=' '.join(options.find_links)))) + **dict(opts, links=" ".join(options.find_links)) + ) + ) else: raise CommandError( - 'You must give at least one requirement to {name} ' - '(see "pip help {name}")'.format(**opts)) + "You must give at least one requirement to {name} " + '(see "pip help {name}")'.format(**opts) + ) return requirements @staticmethod - def trace_basic_info(finder): - # type: (PackageFinder) -> None + def trace_basic_info(finder: PackageFinder) -> None: """ Trace basic information about the provided objects. """ @@ -380,19 +302,18 @@ def trace_basic_info(finder): def _build_package_finder( self, - options, # type: Values - session, # type: PipSession - target_python=None, # type: Optional[TargetPython] - ignore_requires_python=None, # type: Optional[bool] - ): - # type: (...) -> PackageFinder + options: Values, + session: PipSession, + target_python: Optional[TargetPython] = None, + ignore_requires_python: Optional[bool] = None, + ) -> PackageFinder: """ Create a package finder appropriate to this requirement command. :param ignore_requires_python: Whether to ignore incompatible "Requires-Python" values in links. Defaults to False. """ - link_collector = make_link_collector(session, options=options) + link_collector = LinkCollector.create(session, options=options) selection_prefs = SelectionPreferences( allow_yanked=True, format_control=options.format_control, diff --git a/src/fetchcode/vcs/pip/_internal/cli/spinners.py b/src/fetchcode/vcs/pip/_internal/cli/spinners.py index 57e30a7b..cf2b976f 100644 --- a/src/fetchcode/vcs/pip/_internal/cli/spinners.py +++ b/src/fetchcode/vcs/pip/_internal/cli/spinners.py @@ -1,38 +1,33 @@ -from __future__ import absolute_import, division - import contextlib import itertools import logging import sys import time +from typing import IO, Generator, Optional -from fetchcode.vcs.pip._vendor.progress import HIDE_CURSOR, SHOW_CURSOR - -from fetchcode.vcs.pip._internal.utils.compat import WINDOWS -from fetchcode.vcs.pip._internal.utils.logging import get_indentation -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import Iterator, IO +from pip._internal.utils.compat import WINDOWS +from pip._internal.utils.logging import get_indentation logger = logging.getLogger(__name__) -class SpinnerInterface(object): - def spin(self): - # type: () -> None +class SpinnerInterface: + def spin(self) -> None: raise NotImplementedError() - def finish(self, final_status): - # type: (str) -> None + def finish(self, final_status: str) -> None: raise NotImplementedError() class InteractiveSpinner(SpinnerInterface): - def __init__(self, message, file=None, spin_chars="-\\|/", - # Empirically, 8 updates/second looks nice - min_update_interval_seconds=0.125): - # type: (str, IO[str], str, float) -> None + def __init__( + self, + message: str, + file: Optional[IO[str]] = None, + spin_chars: str = "-\\|/", + # Empirically, 8 updates/second looks nice + min_update_interval_seconds: float = 0.125, + ): self._message = message if file is None: file = sys.stdout @@ -45,8 +40,7 @@ def __init__(self, message, file=None, spin_chars="-\\|/", self._file.write(" " * get_indentation() + self._message + " ... ") self._width = 0 - def _write(self, status): - # type: (str) -> None + def _write(self, status: str) -> None: assert not self._finished # Erase what we wrote before by backspacing to the beginning, writing # spaces to overwrite the old text, and then backspacing again @@ -58,16 +52,14 @@ def _write(self, status): self._file.flush() self._rate_limiter.reset() - def spin(self): - # type: () -> None + def spin(self) -> None: if self._finished: return if not self._rate_limiter.ready(): return self._write(next(self._spin_cycle)) - def finish(self, final_status): - # type: (str) -> None + def finish(self, final_status: str) -> None: if self._finished: return self._write(final_status) @@ -81,63 +73,54 @@ def finish(self, final_status): # act as a keep-alive for systems like Travis-CI that take lack-of-output as # an indication that a task has frozen. class NonInteractiveSpinner(SpinnerInterface): - def __init__(self, message, min_update_interval_seconds=60): - # type: (str, float) -> None + def __init__(self, message: str, min_update_interval_seconds: float = 60.0) -> None: self._message = message self._finished = False self._rate_limiter = RateLimiter(min_update_interval_seconds) self._update("started") - def _update(self, status): - # type: (str) -> None + def _update(self, status: str) -> None: assert not self._finished self._rate_limiter.reset() logger.info("%s: %s", self._message, status) - def spin(self): - # type: () -> None + def spin(self) -> None: if self._finished: return if not self._rate_limiter.ready(): return self._update("still running...") - def finish(self, final_status): - # type: (str) -> None + def finish(self, final_status: str) -> None: if self._finished: return - self._update( - "finished with status '{final_status}'".format(**locals())) + self._update(f"finished with status '{final_status}'") self._finished = True -class RateLimiter(object): - def __init__(self, min_update_interval_seconds): - # type: (float) -> None +class RateLimiter: + def __init__(self, min_update_interval_seconds: float) -> None: self._min_update_interval_seconds = min_update_interval_seconds - self._last_update = 0 # type: float + self._last_update: float = 0 - def ready(self): - # type: () -> bool + def ready(self) -> bool: now = time.time() delta = now - self._last_update return delta >= self._min_update_interval_seconds - def reset(self): - # type: () -> None + def reset(self) -> None: self._last_update = time.time() @contextlib.contextmanager -def open_spinner(message): - # type: (str) -> Iterator[SpinnerInterface] +def open_spinner(message: str) -> Generator[SpinnerInterface, None, None]: # Interactive spinner goes directly to sys.stdout rather than being routed # through the logging system, but it acts like it has level INFO, # i.e. it's only displayed if we're at level INFO or better. # Non-interactive spinner goes through the logging system, so it is always # in sync with logging configuration. if sys.stdout.isatty() and logger.getEffectiveLevel() <= logging.INFO: - spinner = InteractiveSpinner(message) # type: SpinnerInterface + spinner: SpinnerInterface = InteractiveSpinner(message) else: spinner = NonInteractiveSpinner(message) try: @@ -153,9 +136,12 @@ def open_spinner(message): spinner.finish("done") +HIDE_CURSOR = "\x1b[?25l" +SHOW_CURSOR = "\x1b[?25h" + + @contextlib.contextmanager -def hidden_cursor(file): - # type: (IO[str]) -> Iterator[None] +def hidden_cursor(file: IO[str]) -> Generator[None, None, None]: # The Windows terminal does not support the hide/show cursor ANSI codes, # even via colorama. So don't even try. if WINDOWS: diff --git a/src/fetchcode/vcs/pip/_internal/cli/status_codes.py b/src/fetchcode/vcs/pip/_internal/cli/status_codes.py index 275360a3..5e29502c 100644 --- a/src/fetchcode/vcs/pip/_internal/cli/status_codes.py +++ b/src/fetchcode/vcs/pip/_internal/cli/status_codes.py @@ -1,5 +1,3 @@ -from __future__ import absolute_import - SUCCESS = 0 ERROR = 1 UNKNOWN_ERROR = 2 diff --git a/src/fetchcode/vcs/pip/_internal/commands/__init__.py b/src/fetchcode/vcs/pip/_internal/commands/__init__.py index 317cf671..858a4101 100644 --- a/src/fetchcode/vcs/pip/_internal/commands/__init__.py +++ b/src/fetchcode/vcs/pip/_internal/commands/__init__.py @@ -2,101 +2,111 @@ Package containing all pip commands """ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False -# There is currently a bug in python/typeshed mentioned at -# https://github.com/python/typeshed/issues/3906 which causes the -# return type of difflib.get_close_matches to be reported -# as List[Sequence[str]] whereas it should have been List[str] - -from __future__ import absolute_import - import importlib -from collections import OrderedDict, namedtuple - -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import Any - from fetchcode.vcs.pip._internal.cli.base_command import Command - - -CommandInfo = namedtuple('CommandInfo', 'module_path, class_name, summary') - -# The ordering matters for help display. -# Also, even though the module path starts with the same -# "pip._internal.commands" prefix in each case, we include the full path -# because it makes testing easier (specifically when modifying commands_dict -# in test setup / teardown by adding info for a FakeCommand class defined -# in a test-related module). -# Finally, we need to pass an iterable of pairs here rather than a dict -# so that the ordering won't be lost when using Python 2.7. -commands_dict = OrderedDict([ - ('install', CommandInfo( - 'pip._internal.commands.install', 'InstallCommand', - 'Install packages.', - )), - ('download', CommandInfo( - 'pip._internal.commands.download', 'DownloadCommand', - 'Download packages.', - )), - ('uninstall', CommandInfo( - 'pip._internal.commands.uninstall', 'UninstallCommand', - 'Uninstall packages.', - )), - ('freeze', CommandInfo( - 'pip._internal.commands.freeze', 'FreezeCommand', - 'Output installed packages in requirements format.', - )), - ('list', CommandInfo( - 'pip._internal.commands.list', 'ListCommand', - 'List installed packages.', - )), - ('show', CommandInfo( - 'pip._internal.commands.show', 'ShowCommand', - 'Show information about installed packages.', - )), - ('check', CommandInfo( - 'pip._internal.commands.check', 'CheckCommand', - 'Verify installed packages have compatible dependencies.', - )), - ('config', CommandInfo( - 'pip._internal.commands.configuration', 'ConfigurationCommand', - 'Manage local and global configuration.', - )), - ('search', CommandInfo( - 'pip._internal.commands.search', 'SearchCommand', - 'Search PyPI for packages.', - )), - ('cache', CommandInfo( - 'pip._internal.commands.cache', 'CacheCommand', +from collections import namedtuple +from typing import Any, Dict, Optional + +from pip._internal.cli.base_command import Command + +CommandInfo = namedtuple("CommandInfo", "module_path, class_name, summary") + +# This dictionary does a bunch of heavy lifting for help output: +# - Enables avoiding additional (costly) imports for presenting `--help`. +# - The ordering matters for help display. +# +# Even though the module path starts with the same "pip._internal.commands" +# prefix, the full path makes testing easier (specifically when modifying +# `commands_dict` in test setup / teardown). +commands_dict: Dict[str, CommandInfo] = { + "install": CommandInfo( + "pip._internal.commands.install", + "InstallCommand", + "Install packages.", + ), + "download": CommandInfo( + "pip._internal.commands.download", + "DownloadCommand", + "Download packages.", + ), + "uninstall": CommandInfo( + "pip._internal.commands.uninstall", + "UninstallCommand", + "Uninstall packages.", + ), + "freeze": CommandInfo( + "pip._internal.commands.freeze", + "FreezeCommand", + "Output installed packages in requirements format.", + ), + "inspect": CommandInfo( + "pip._internal.commands.inspect", + "InspectCommand", + "Inspect the python environment.", + ), + "list": CommandInfo( + "pip._internal.commands.list", + "ListCommand", + "List installed packages.", + ), + "show": CommandInfo( + "pip._internal.commands.show", + "ShowCommand", + "Show information about installed packages.", + ), + "check": CommandInfo( + "pip._internal.commands.check", + "CheckCommand", + "Verify installed packages have compatible dependencies.", + ), + "config": CommandInfo( + "pip._internal.commands.configuration", + "ConfigurationCommand", + "Manage local and global configuration.", + ), + "search": CommandInfo( + "pip._internal.commands.search", + "SearchCommand", + "Search PyPI for packages.", + ), + "cache": CommandInfo( + "pip._internal.commands.cache", + "CacheCommand", "Inspect and manage pip's wheel cache.", - )), - ('wheel', CommandInfo( - 'pip._internal.commands.wheel', 'WheelCommand', - 'Build wheels from your requirements.', - )), - ('hash', CommandInfo( - 'pip._internal.commands.hash', 'HashCommand', - 'Compute hashes of package archives.', - )), - ('completion', CommandInfo( - 'pip._internal.commands.completion', 'CompletionCommand', - 'A helper command used for command completion.', - )), - ('debug', CommandInfo( - 'pip._internal.commands.debug', 'DebugCommand', - 'Show information useful for debugging.', - )), - ('help', CommandInfo( - 'pip._internal.commands.help', 'HelpCommand', - 'Show help for commands.', - )), -]) # type: OrderedDict[str, CommandInfo] - - -def create_command(name, **kwargs): - # type: (str, **Any) -> Command + ), + "index": CommandInfo( + "pip._internal.commands.index", + "IndexCommand", + "Inspect information available from package indexes.", + ), + "wheel": CommandInfo( + "pip._internal.commands.wheel", + "WheelCommand", + "Build wheels from your requirements.", + ), + "hash": CommandInfo( + "pip._internal.commands.hash", + "HashCommand", + "Compute hashes of package archives.", + ), + "completion": CommandInfo( + "pip._internal.commands.completion", + "CompletionCommand", + "A helper command used for command completion.", + ), + "debug": CommandInfo( + "pip._internal.commands.debug", + "DebugCommand", + "Show information useful for debugging.", + ), + "help": CommandInfo( + "pip._internal.commands.help", + "HelpCommand", + "Show help for commands.", + ), +} + + +def create_command(name: str, **kwargs: Any) -> Command: """ Create an instance of the Command class with the given name. """ @@ -108,7 +118,7 @@ def create_command(name, **kwargs): return command -def get_similar_commands(name): +def get_similar_commands(name: str) -> Optional[str]: """Command name auto-correct.""" from difflib import get_close_matches @@ -119,4 +129,4 @@ def get_similar_commands(name): if close_commands: return close_commands[0] else: - return False + return None diff --git a/src/fetchcode/vcs/pip/_internal/commands/cache.py b/src/fetchcode/vcs/pip/_internal/commands/cache.py index b0a8926e..32833615 100644 --- a/src/fetchcode/vcs/pip/_internal/commands/cache.py +++ b/src/fetchcode/vcs/pip/_internal/commands/cache.py @@ -1,21 +1,15 @@ -from __future__ import absolute_import - -import logging import os import textwrap +from optparse import Values +from typing import Any, List -import fetchcode.vcs.pip._internal.utils.filesystem as filesystem -from fetchcode.vcs.pip._internal.cli.base_command import Command -from fetchcode.vcs.pip._internal.cli.status_codes import ERROR, SUCCESS -from fetchcode.vcs.pip._internal.exceptions import CommandError, PipError -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from optparse import Values - from typing import Any, List +from pip._internal.cli.base_command import Command +from pip._internal.cli.status_codes import ERROR, SUCCESS +from pip._internal.exceptions import CommandError, PipError +from pip._internal.utils import filesystem +from pip._internal.utils.logging import getLogger - -logger = logging.getLogger(__name__) +logger = getLogger(__name__) class CacheCommand(Command): @@ -24,26 +18,37 @@ class CacheCommand(Command): Subcommands: - dir: Show the cache directory. - info: Show information about the cache. - list: List filenames of packages stored in the cache. - remove: Remove one or more package from the cache. - purge: Remove all items from the cache. + - dir: Show the cache directory. + - info: Show information about the cache. + - list: List filenames of packages stored in the cache. + - remove: Remove one or more package from the cache. + - purge: Remove all items from the cache. - can be a glob expression or a package name. + ```` can be a glob expression or a package name. """ ignore_require_venv = True usage = """ %prog dir %prog info - %prog list [] + %prog list [] [--format=[human, abspath]] %prog remove %prog purge """ - def run(self, options, args): - # type: (Values, List[Any]) -> int + def add_options(self) -> None: + self.cmd_opts.add_option( + "--format", + action="store", + dest="list_format", + default="human", + choices=("human", "abspath"), + help="Select the output format among: human (default) or abspath", + ) + + self.parser.insert_option_group(0, self.cmd_opts) + + def run(self, options: Values, args: List[str]) -> int: handlers = { "dir": self.get_cache_dir, "info": self.get_cache_info, @@ -53,14 +58,14 @@ def run(self, options, args): } if not options.cache_dir: - logger.error("pip cache commands can not " - "function since cache is disabled.") + logger.error("pip cache commands can not function since cache is disabled.") return ERROR # Determine action if not args or args[0] not in handlers: - logger.error("Need an action ({}) to perform.".format( - ", ".join(sorted(handlers))) + logger.error( + "Need an action (%s) to perform.", + ", ".join(sorted(handlers)), ) return ERROR @@ -75,90 +80,129 @@ def run(self, options, args): return SUCCESS - def get_cache_dir(self, options, args): - # type: (Values, List[Any]) -> None + def get_cache_dir(self, options: Values, args: List[Any]) -> None: if args: - raise CommandError('Too many arguments') + raise CommandError("Too many arguments") logger.info(options.cache_dir) - def get_cache_info(self, options, args): - # type: (Values, List[Any]) -> None + def get_cache_info(self, options: Values, args: List[Any]) -> None: if args: - raise CommandError('Too many arguments') - - num_packages = len(self._find_wheels(options, '*')) - - cache_location = self._wheels_cache_dir(options) - cache_size = filesystem.format_directory_size(cache_location) - - message = textwrap.dedent(""" - Location: {location} - Size: {size} - Number of wheels: {package_count} - """).format( - location=cache_location, - package_count=num_packages, - size=cache_size, - ).strip() + raise CommandError("Too many arguments") + + num_http_files = len(self._find_http_files(options)) + num_packages = len(self._find_wheels(options, "*")) + + http_cache_location = self._cache_dir(options, "http-v2") + old_http_cache_location = self._cache_dir(options, "http") + wheels_cache_location = self._cache_dir(options, "wheels") + http_cache_size = filesystem.format_size( + filesystem.directory_size(http_cache_location) + + filesystem.directory_size(old_http_cache_location) + ) + wheels_cache_size = filesystem.format_directory_size(wheels_cache_location) + + message = ( + textwrap.dedent( + """ + Package index page cache location (pip v23.3+): {http_cache_location} + Package index page cache location (older pips): {old_http_cache_location} + Package index page cache size: {http_cache_size} + Number of HTTP files: {num_http_files} + Locally built wheels location: {wheels_cache_location} + Locally built wheels size: {wheels_cache_size} + Number of locally built wheels: {package_count} + """ # noqa: E501 + ) + .format( + http_cache_location=http_cache_location, + old_http_cache_location=old_http_cache_location, + http_cache_size=http_cache_size, + num_http_files=num_http_files, + wheels_cache_location=wheels_cache_location, + package_count=num_packages, + wheels_cache_size=wheels_cache_size, + ) + .strip() + ) logger.info(message) - def list_cache_items(self, options, args): - # type: (Values, List[Any]) -> None + def list_cache_items(self, options: Values, args: List[Any]) -> None: if len(args) > 1: - raise CommandError('Too many arguments') + raise CommandError("Too many arguments") if args: pattern = args[0] else: - pattern = '*' + pattern = "*" files = self._find_wheels(options, pattern) + if options.list_format == "human": + self.format_for_human(files) + else: + self.format_for_abspath(files) + def format_for_human(self, files: List[str]) -> None: if not files: - logger.info('Nothing cached.') + logger.info("No locally built wheels cached.") return results = [] for filename in files: wheel = os.path.basename(filename) size = filesystem.format_file_size(filename) - results.append(' - {} ({})'.format(wheel, size)) - logger.info('Cache contents:\n') - logger.info('\n'.join(sorted(results))) + results.append(f" - {wheel} ({size})") + logger.info("Cache contents:\n") + logger.info("\n".join(sorted(results))) + + def format_for_abspath(self, files: List[str]) -> None: + if files: + logger.info("\n".join(sorted(files))) - def remove_cache_items(self, options, args): - # type: (Values, List[Any]) -> None + def remove_cache_items(self, options: Values, args: List[Any]) -> None: if len(args) > 1: - raise CommandError('Too many arguments') + raise CommandError("Too many arguments") if not args: - raise CommandError('Please provide a pattern') + raise CommandError("Please provide a pattern") files = self._find_wheels(options, args[0]) + + no_matching_msg = "No matching packages" + if args[0] == "*": + # Only fetch http files if no specific pattern given + files += self._find_http_files(options) + else: + # Add the pattern to the log message + no_matching_msg += f' for pattern "{args[0]}"' + if not files: - raise CommandError('No matching packages') + logger.warning(no_matching_msg) for filename in files: os.unlink(filename) - logger.debug('Removed %s', filename) - logger.info('Files removed: %s', len(files)) + logger.verbose("Removed %s", filename) + logger.info("Files removed: %s", len(files)) - def purge_cache(self, options, args): - # type: (Values, List[Any]) -> None + def purge_cache(self, options: Values, args: List[Any]) -> None: if args: - raise CommandError('Too many arguments') + raise CommandError("Too many arguments") + + return self.remove_cache_items(options, ["*"]) - return self.remove_cache_items(options, ['*']) + def _cache_dir(self, options: Values, subdir: str) -> str: + return os.path.join(options.cache_dir, subdir) - def _wheels_cache_dir(self, options): - # type: (Values) -> str - return os.path.join(options.cache_dir, 'wheels') + def _find_http_files(self, options: Values) -> List[str]: + old_http_dir = self._cache_dir(options, "http") + new_http_dir = self._cache_dir(options, "http-v2") + return filesystem.find_files(old_http_dir, "*") + filesystem.find_files( + new_http_dir, "*" + ) - def _find_wheels(self, options, pattern): - # type: (Values, str) -> List[str] - wheel_dir = self._wheels_cache_dir(options) + def _find_wheels(self, options: Values, pattern: str) -> List[str]: + wheel_dir = self._cache_dir(options, "wheels") # The wheel filename format, as specified in PEP 427, is: # {distribution}-{version}(-{build})?-{python}-{abi}-{platform}.whl diff --git a/src/fetchcode/vcs/pip/_internal/commands/check.py b/src/fetchcode/vcs/pip/_internal/commands/check.py index 35af06da..f54a16dc 100644 --- a/src/fetchcode/vcs/pip/_internal/commands/check.py +++ b/src/fetchcode/vcs/pip/_internal/commands/check.py @@ -1,39 +1,46 @@ import logging +from optparse import Values +from typing import List -from fetchcode.vcs.pip._internal.cli.base_command import Command -from fetchcode.vcs.pip._internal.cli.status_codes import ERROR, SUCCESS -from fetchcode.vcs.pip._internal.operations.check import ( +from pip._internal.cli.base_command import Command +from pip._internal.cli.status_codes import ERROR, SUCCESS +from pip._internal.metadata import get_default_environment +from pip._internal.operations.check import ( check_package_set, + check_unsupported, create_package_set_from_installed, ) -from fetchcode.vcs.pip._internal.utils.misc import write_output -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._internal.utils.compatibility_tags import get_supported +from pip._internal.utils.misc import write_output logger = logging.getLogger(__name__) -if MYPY_CHECK_RUNNING: - from typing import List, Any - from optparse import Values - class CheckCommand(Command): """Verify installed packages have compatible dependencies.""" + ignore_require_venv = True usage = """ %prog [options]""" - def run(self, options, args): - # type: (Values, List[Any]) -> int - + def run(self, options: Values, args: List[str]) -> int: package_set, parsing_probs = create_package_set_from_installed() missing, conflicting = check_package_set(package_set) + unsupported = list( + check_unsupported( + get_default_environment().iter_installed_distributions(), + get_supported(), + ) + ) for project_name in missing: version = package_set[project_name].version for dependency in missing[project_name]: write_output( "%s %s requires %s, which is not installed.", - project_name, version, dependency[0], + project_name, + version, + dependency[0], ) for project_name in conflicting: @@ -41,10 +48,19 @@ def run(self, options, args): for dep_name, dep_version, req in conflicting[project_name]: write_output( "%s %s has requirement %s, but you have %s %s.", - project_name, version, req, dep_name, dep_version, + project_name, + version, + req, + dep_name, + dep_version, ) - - if missing or conflicting or parsing_probs: + for package in unsupported: + write_output( + "%s %s is not supported on this platform", + package.raw_name, + package.version, + ) + if missing or conflicting or parsing_probs or unsupported: return ERROR else: write_output("No broken requirements found.") diff --git a/src/fetchcode/vcs/pip/_internal/commands/completion.py b/src/fetchcode/vcs/pip/_internal/commands/completion.py index a3c04ab9..9e89e279 100644 --- a/src/fetchcode/vcs/pip/_internal/commands/completion.py +++ b/src/fetchcode/vcs/pip/_internal/commands/completion.py @@ -1,20 +1,18 @@ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import - import sys import textwrap +from optparse import Values +from typing import List -from fetchcode.vcs.pip._internal.cli.base_command import Command -from fetchcode.vcs.pip._internal.utils.misc import get_prog +from pip._internal.cli.base_command import Command +from pip._internal.cli.status_codes import SUCCESS +from pip._internal.utils.misc import get_prog BASE_COMPLETION = """ # pip {shell} completion start{script}# pip {shell} completion end """ COMPLETION_SCRIPTS = { - 'bash': """ + "bash": """ _pip_completion() {{ COMPREPLY=( $( COMP_WORDS="${{COMP_WORDS[*]}}" \\ @@ -23,18 +21,22 @@ }} complete -o default -F _pip_completion {prog} """, - 'zsh': """ - function _pip_completion {{ - local words cword - read -Ac words - read -cn cword - reply=( $( COMP_WORDS="$words[*]" \\ - COMP_CWORD=$(( cword-1 )) \\ - PIP_AUTO_COMPLETE=1 $words[1] 2>/dev/null )) + "zsh": """ + #compdef -P pip[0-9.]# + __pip() {{ + compadd $( COMP_WORDS="$words[*]" \\ + COMP_CWORD=$((CURRENT-1)) \\ + PIP_AUTO_COMPLETE=1 $words[1] 2>/dev/null ) }} - compctl -K _pip_completion {prog} + if [[ $zsh_eval_context[-1] == loadautofunc ]]; then + # autoload from fpath, call function directly + __pip "$@" + else + # eval/source/. command, register function for later + compdef __pip -P 'pip[0-9.]#' + fi """, - 'fish': """ + "fish": """ function __fish_complete_pip set -lx COMP_WORDS (commandline -o) "" set -lx COMP_CWORD ( \\ @@ -45,6 +47,28 @@ end complete -fa "(__fish_complete_pip)" -c {prog} """, + "powershell": """ + if ((Test-Path Function:\\TabExpansion) -and -not ` + (Test-Path Function:\\_pip_completeBackup)) {{ + Rename-Item Function:\\TabExpansion _pip_completeBackup + }} + function TabExpansion($line, $lastWord) {{ + $lastBlock = [regex]::Split($line, '[|;]')[-1].TrimStart() + if ($lastBlock.StartsWith("{prog} ")) {{ + $Env:COMP_WORDS=$lastBlock + $Env:COMP_CWORD=$lastBlock.Split().Length - 1 + $Env:PIP_AUTO_COMPLETE=1 + (& {prog}).Split() + Remove-Item Env:COMP_WORDS + Remove-Item Env:COMP_CWORD + Remove-Item Env:PIP_AUTO_COMPLETE + }} + elseif (Test-Path Function:\\_pip_completeBackup) {{ + # Fall back on existing tab expansion + _pip_completeBackup $line $lastWord + }} + }} + """, } @@ -53,43 +77,54 @@ class CompletionCommand(Command): ignore_require_venv = True - def __init__(self, *args, **kw): - super(CompletionCommand, self).__init__(*args, **kw) - - cmd_opts = self.cmd_opts - - cmd_opts.add_option( - '--bash', '-b', - action='store_const', - const='bash', - dest='shell', - help='Emit completion code for bash') - cmd_opts.add_option( - '--zsh', '-z', - action='store_const', - const='zsh', - dest='shell', - help='Emit completion code for zsh') - cmd_opts.add_option( - '--fish', '-f', - action='store_const', - const='fish', - dest='shell', - help='Emit completion code for fish') + def add_options(self) -> None: + self.cmd_opts.add_option( + "--bash", + "-b", + action="store_const", + const="bash", + dest="shell", + help="Emit completion code for bash", + ) + self.cmd_opts.add_option( + "--zsh", + "-z", + action="store_const", + const="zsh", + dest="shell", + help="Emit completion code for zsh", + ) + self.cmd_opts.add_option( + "--fish", + "-f", + action="store_const", + const="fish", + dest="shell", + help="Emit completion code for fish", + ) + self.cmd_opts.add_option( + "--powershell", + "-p", + action="store_const", + const="powershell", + dest="shell", + help="Emit completion code for powershell", + ) - self.parser.insert_option_group(0, cmd_opts) + self.parser.insert_option_group(0, self.cmd_opts) - def run(self, options, args): + def run(self, options: Values, args: List[str]) -> int: """Prints the completion code of the given shell""" shells = COMPLETION_SCRIPTS.keys() - shell_options = ['--' + shell for shell in sorted(shells)] + shell_options = ["--" + shell for shell in sorted(shells)] if options.shell in shells: script = textwrap.dedent( - COMPLETION_SCRIPTS.get(options.shell, '').format( - prog=get_prog()) + COMPLETION_SCRIPTS.get(options.shell, "").format(prog=get_prog()) ) print(BASE_COMPLETION.format(script=script, shell=options.shell)) + return SUCCESS else: sys.stderr.write( - 'ERROR: You must pass {}\n' .format(' or '.join(shell_options)) + "ERROR: You must pass {}\n".format(" or ".join(shell_options)) ) + return SUCCESS diff --git a/src/fetchcode/vcs/pip/_internal/commands/configuration.py b/src/fetchcode/vcs/pip/_internal/commands/configuration.py index 8371cfe6..1a1dc6b6 100644 --- a/src/fetchcode/vcs/pip/_internal/commands/configuration.py +++ b/src/fetchcode/vcs/pip/_internal/commands/configuration.py @@ -1,37 +1,46 @@ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - import logging import os import subprocess +from optparse import Values +from typing import Any, List, Optional -from fetchcode.vcs.pip._internal.cli.base_command import Command -from fetchcode.vcs.pip._internal.cli.status_codes import ERROR, SUCCESS -from fetchcode.vcs.pip._internal.configuration import ( +from pip._internal.cli.base_command import Command +from pip._internal.cli.status_codes import ERROR, SUCCESS +from pip._internal.configuration import ( Configuration, + Kind, get_configuration_files, kinds, ) -from fetchcode.vcs.pip._internal.exceptions import PipError -from fetchcode.vcs.pip._internal.utils.misc import get_prog, write_output +from pip._internal.exceptions import PipError +from pip._internal.utils.logging import indent_log +from pip._internal.utils.misc import get_prog, write_output logger = logging.getLogger(__name__) class ConfigurationCommand(Command): - """Manage local and global configuration. + """ + Manage local and global configuration. Subcommands: - list: List the active configuration (or from the file specified) - edit: Edit the configuration file in an editor - get: Get the value associated with name - set: Set the name=value - unset: Unset the value associated with name + - list: List the active configuration (or from the file specified) + - edit: Edit the configuration file in an editor + - get: Get the value associated with command.option + - set: Set the command.option=value + - unset: Unset the value associated with command.option + - debug: List the configuration files and values defined under them + + Configuration keys should be dot separated command and option name, + with the special prefix "global" affecting any command. For example, + "pip config set global.index-url https://example.org/" would configure + the index url for all commands, but "pip config set download.timeout 10" + would configure a 10 second timeout only for "pip download" commands. If none of --user, --global and --site are passed, a virtual environment configuration file is used if one is active and the file - exists. Otherwise, all modifications happen on the to the user file by + exists. Otherwise, all modifications happen to the user file by default. """ @@ -40,66 +49,65 @@ class ConfigurationCommand(Command): %prog [] list %prog [] [--editor ] edit - %prog [] get name - %prog [] set name value - %prog [] unset name + %prog [] get command.option + %prog [] set command.option value + %prog [] unset command.option + %prog [] debug """ - def __init__(self, *args, **kwargs): - super(ConfigurationCommand, self).__init__(*args, **kwargs) - - self.configuration = None - + def add_options(self) -> None: self.cmd_opts.add_option( - '--editor', - dest='editor', - action='store', + "--editor", + dest="editor", + action="store", default=None, help=( - 'Editor to use to edit the file. Uses VISUAL or EDITOR ' - 'environment variables if not provided.' - ) + "Editor to use to edit the file. Uses VISUAL or EDITOR " + "environment variables if not provided." + ), ) self.cmd_opts.add_option( - '--global', - dest='global_file', - action='store_true', + "--global", + dest="global_file", + action="store_true", default=False, - help='Use the system-wide configuration file only' + help="Use the system-wide configuration file only", ) self.cmd_opts.add_option( - '--user', - dest='user_file', - action='store_true', + "--user", + dest="user_file", + action="store_true", default=False, - help='Use the user configuration file only' + help="Use the user configuration file only", ) self.cmd_opts.add_option( - '--site', - dest='site_file', - action='store_true', + "--site", + dest="site_file", + action="store_true", default=False, - help='Use the current environment configuration file only' + help="Use the current environment configuration file only", ) self.parser.insert_option_group(0, self.cmd_opts) - def run(self, options, args): + def run(self, options: Values, args: List[str]) -> int: handlers = { "list": self.list_values, "edit": self.open_in_editor, "get": self.get_name, "set": self.set_name_value, - "unset": self.unset_name + "unset": self.unset_name, + "debug": self.list_config_values, } # Determine action if not args or args[0] not in handlers: - logger.error("Need an action ({}) to perform.".format( - ", ".join(sorted(handlers))) + logger.error( + "Need an action (%s) to perform.", + ", ".join(sorted(handlers)), ) return ERROR @@ -130,12 +138,16 @@ def run(self, options, args): return SUCCESS - def _determine_file(self, options, need_value): - file_options = [key for key, value in ( - (kinds.USER, options.user_file), - (kinds.GLOBAL, options.global_file), - (kinds.SITE, options.site_file), - ) if value] + def _determine_file(self, options: Values, need_value: bool) -> Optional[Kind]: + file_options = [ + key + for key, value in ( + (kinds.USER, options.user_file), + (kinds.GLOBAL, options.global_file), + (kinds.SITE, options.site_file), + ) + if value + ] if not file_options: if not need_value: @@ -156,53 +168,89 @@ def _determine_file(self, options, need_value): "(--user, --site, --global) to perform." ) - def list_values(self, options, args): + def list_values(self, options: Values, args: List[str]) -> None: self._get_n_args(args, "list", n=0) for key, value in sorted(self.configuration.items()): write_output("%s=%r", key, value) - def get_name(self, options, args): + def get_name(self, options: Values, args: List[str]) -> None: key = self._get_n_args(args, "get [name]", n=1) value = self.configuration.get_value(key) write_output("%s", value) - def set_name_value(self, options, args): + def set_name_value(self, options: Values, args: List[str]) -> None: key, value = self._get_n_args(args, "set [name] [value]", n=2) self.configuration.set_value(key, value) self._save_configuration() - def unset_name(self, options, args): + def unset_name(self, options: Values, args: List[str]) -> None: key = self._get_n_args(args, "unset [name]", n=1) self.configuration.unset_value(key) self._save_configuration() - def open_in_editor(self, options, args): + def list_config_values(self, options: Values, args: List[str]) -> None: + """List config key-value pairs across different config files""" + self._get_n_args(args, "debug", n=0) + + self.print_env_var_values() + # Iterate over config files and print if they exist, and the + # key-value pairs present in them if they do + for variant, files in sorted(self.configuration.iter_config_files()): + write_output("%s:", variant) + for fname in files: + with indent_log(): + file_exists = os.path.exists(fname) + write_output("%s, exists: %r", fname, file_exists) + if file_exists: + self.print_config_file_values(variant) + + def print_config_file_values(self, variant: Kind) -> None: + """Get key-value pairs from the file of a variant""" + for name, value in self.configuration.get_values_in_config(variant).items(): + with indent_log(): + write_output("%s: %s", name, value) + + def print_env_var_values(self) -> None: + """Get key-values pairs present as environment variables""" + write_output("%s:", "env_var") + with indent_log(): + for key, value in sorted(self.configuration.get_environ_vars()): + env_var = f"PIP_{key.upper()}" + write_output("%s=%r", env_var, value) + + def open_in_editor(self, options: Values, args: List[str]) -> None: editor = self._determine_editor(options) fname = self.configuration.get_file_to_edit() if fname is None: raise PipError("Could not determine appropriate file.") + elif '"' in fname: + # This shouldn't happen, unless we see a username like that. + # If that happens, we'd appreciate a pull request fixing this. + raise PipError( + f'Can not open an editor for a file name containing "\n{fname}' + ) try: - subprocess.check_call([editor, fname]) + subprocess.check_call(f'{editor} "{fname}"', shell=True) + except FileNotFoundError as e: + if not e.filename: + e.filename = editor + raise except subprocess.CalledProcessError as e: - raise PipError( - "Editor Subprocess exited with exit code {}" - .format(e.returncode) - ) + raise PipError(f"Editor Subprocess exited with exit code {e.returncode}") - def _get_n_args(self, args, example, n): - """Helper to make sure the command got the right number of arguments - """ + def _get_n_args(self, args: List[str], example: str, n: int) -> Any: + """Helper to make sure the command got the right number of arguments""" if len(args) != n: msg = ( - 'Got unexpected number of arguments, expected {}. ' - '(example: "{} config {}")' - ).format(n, get_prog(), example) + f"Got unexpected number of arguments, expected {n}. " + f'(example: "{get_prog()} config {example}")' + ) raise PipError(msg) if n == 1: @@ -210,19 +258,18 @@ def _get_n_args(self, args, example, n): else: return args - def _save_configuration(self): + def _save_configuration(self) -> None: # We successfully ran a modifying command. Need to save the # configuration. try: self.configuration.save() except Exception: - logger.error( - "Unable to save configuration. Please report this as a bug.", - exc_info=1 + logger.exception( + "Unable to save configuration. Please report this as a bug." ) raise PipError("Internal Error.") - def _determine_editor(self, options): + def _determine_editor(self, options: Values) -> str: if options.editor is not None: return options.editor elif "VISUAL" in os.environ: diff --git a/src/fetchcode/vcs/pip/_internal/commands/debug.py b/src/fetchcode/vcs/pip/_internal/commands/debug.py index 4ccff438..567ca967 100644 --- a/src/fetchcode/vcs/pip/_internal/commands/debug.py +++ b/src/fetchcode/vcs/pip/_internal/commands/debug.py @@ -1,153 +1,126 @@ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import - import locale import logging import os import sys - -import fetchcode.vcs.pip._vendor -from fetchcode.vcs.pip._vendor import pkg_resources -from fetchcode.vcs.pip._vendor.certifi import where - -from pip import __file__ as pip_location -from fetchcode.vcs.pip._internal.cli import cmdoptions -from fetchcode.vcs.pip._internal.cli.base_command import Command -from fetchcode.vcs.pip._internal.cli.cmdoptions import make_target_python -from fetchcode.vcs.pip._internal.cli.status_codes import SUCCESS -from fetchcode.vcs.pip._internal.utils.logging import indent_log -from fetchcode.vcs.pip._internal.utils.misc import get_pip_version -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from types import ModuleType - from typing import Any, List, Optional, Dict - from optparse import Values +from optparse import Values +from types import ModuleType +from typing import Any, Dict, List, Optional + +import pip._vendor +from pip._vendor.certifi import where +from pip._vendor.packaging.version import parse as parse_version + +from pip._internal.cli import cmdoptions +from pip._internal.cli.base_command import Command +from pip._internal.cli.cmdoptions import make_target_python +from pip._internal.cli.status_codes import SUCCESS +from pip._internal.configuration import Configuration +from pip._internal.metadata import get_environment +from pip._internal.utils.compat import open_text_resource +from pip._internal.utils.logging import indent_log +from pip._internal.utils.misc import get_pip_version logger = logging.getLogger(__name__) -def show_value(name, value): - # type: (str, Optional[str]) -> None - logger.info('{}: {}'.format(name, value)) +def show_value(name: str, value: Any) -> None: + logger.info("%s: %s", name, value) -def show_sys_implementation(): - # type: () -> None - logger.info('sys.implementation:') - if hasattr(sys, 'implementation'): - implementation = sys.implementation # type: ignore - implementation_name = implementation.name - else: - implementation_name = '' - +def show_sys_implementation() -> None: + logger.info("sys.implementation:") + implementation_name = sys.implementation.name with indent_log(): - show_value('name', implementation_name) + show_value("name", implementation_name) -def create_vendor_txt_map(): - # type: () -> Dict[str, str] - vendor_txt_path = os.path.join( - os.path.dirname(pip_location), - '_vendor', - 'vendor.txt' - ) - - with open(vendor_txt_path) as f: +def create_vendor_txt_map() -> Dict[str, str]: + with open_text_resource("pip._vendor", "vendor.txt") as f: # Purge non version specifying lines. # Also, remove any space prefix or suffixes (including comments). - lines = [line.strip().split(' ', 1)[0] - for line in f.readlines() if '==' in line] + lines = [ + line.strip().split(" ", 1)[0] for line in f.readlines() if "==" in line + ] # Transform into "module" -> version dict. - return dict(line.split('==', 1) for line in lines) # type: ignore - + return dict(line.split("==", 1) for line in lines) -def get_module_from_module_name(module_name): - # type: (str) -> ModuleType +def get_module_from_module_name(module_name: str) -> Optional[ModuleType]: # Module name can be uppercase in vendor.txt for some reason... - module_name = module_name.lower() + module_name = module_name.lower().replace("-", "_") # PATCH: setuptools is actually only pkg_resources. - if module_name == 'setuptools': - module_name = 'pkg_resources' + if module_name == "setuptools": + module_name = "pkg_resources" - __import__( - 'pip._vendor.{}'.format(module_name), - globals(), - locals(), - level=0 - ) - return getattr(pip._vendor, module_name) + try: + __import__(f"pip._vendor.{module_name}", globals(), locals(), level=0) + return getattr(pip._vendor, module_name) + except ImportError: + # We allow 'truststore' to fail to import due + # to being unavailable on Python 3.9 and earlier. + if module_name == "truststore" and sys.version_info < (3, 10): + return None + raise -def get_vendor_version_from_module(module_name): - # type: (str) -> str - +def get_vendor_version_from_module(module_name: str) -> Optional[str]: module = get_module_from_module_name(module_name) - version = getattr(module, '__version__', None) + version = getattr(module, "__version__", None) - if not version: - # Try to find version in debundled module info - pkg_set = pkg_resources.WorkingSet( - [os.path.dirname(getattr(module, '__file__'))] - ) - package = pkg_set.find(pkg_resources.Requirement.parse(module_name)) - version = getattr(package, 'version', None) + if module and not version: + # Try to find version in debundled module info. + assert module.__file__ is not None + env = get_environment([os.path.dirname(module.__file__)]) + dist = env.get_distribution(module_name) + if dist: + version = str(dist.version) return version -def show_actual_vendor_versions(vendor_txt_versions): - # type: (Dict[str, str]) -> None - # Logs the actual version and print extra info - # if there is a conflict or if the actual version could not be imported. - +def show_actual_vendor_versions(vendor_txt_versions: Dict[str, str]) -> None: + """Log the actual version and print extra info if there is + a conflict or if the actual version could not be imported. + """ for module_name, expected_version in vendor_txt_versions.items(): - extra_message = '' + extra_message = "" actual_version = get_vendor_version_from_module(module_name) if not actual_version: - extra_message = ' (Unable to locate actual module version, using'\ - ' vendor.txt specified version)' + extra_message = ( + " (Unable to locate actual module version, using" + " vendor.txt specified version)" + ) actual_version = expected_version - elif actual_version != expected_version: - extra_message = ' (CONFLICT: vendor.txt suggests version should'\ - ' be {})'.format(expected_version) - - logger.info( - '{name}=={actual}{extra}'.format( - name=module_name, - actual=actual_version, - extra=extra_message + elif parse_version(actual_version) != parse_version(expected_version): + extra_message = ( + " (CONFLICT: vendor.txt suggests version should" + f" be {expected_version})" ) - ) + logger.info("%s==%s%s", module_name, actual_version, extra_message) -def show_vendor_versions(): - # type: () -> None - logger.info('vendored library versions:') +def show_vendor_versions() -> None: + logger.info("vendored library versions:") vendor_txt_versions = create_vendor_txt_map() with indent_log(): show_actual_vendor_versions(vendor_txt_versions) -def show_tags(options): - # type: (Values) -> None +def show_tags(options: Values) -> None: tag_limit = 10 target_python = make_target_python(options) - tags = target_python.get_tags() + tags = target_python.get_sorted_tags() # Display the target options that were explicitly provided. formatted_target = target_python.format_given() - suffix = '' + suffix = "" if formatted_target: - suffix = ' (target: {})'.format(formatted_target) + suffix = f" (target: {formatted_target})" - msg = 'Compatible tags: {}{}'.format(len(tags), suffix) + msg = f"Compatible tags: {len(tags)}{suffix}" logger.info(msg) if options.verbose < 1 and len(tags) > tag_limit: @@ -161,30 +134,24 @@ def show_tags(options): logger.info(str(tag)) if tags_limited: - msg = ( - '...\n' - '[First {tag_limit} tags shown. Pass --verbose to show all.]' - ).format(tag_limit=tag_limit) + msg = f"...\n[First {tag_limit} tags shown. Pass --verbose to show all.]" logger.info(msg) -def ca_bundle_info(config): - levels = set() - for key, value in config.items(): - levels.add(key.split('.')[0]) - +def ca_bundle_info(config: Configuration) -> str: + levels = {key.split(".", 1)[0] for key, _ in config.items()} if not levels: return "Not specified" - levels_that_override_global = ['install', 'wheel', 'download'] + levels_that_override_global = ["install", "wheel", "download"] global_overriding_level = [ level for level in levels if level in levels_that_override_global ] if not global_overriding_level: - return 'global' + return "global" - if 'global' in levels: - levels.remove('global') + if "global" in levels: + levels.remove("global") return ", ".join(levels) @@ -197,36 +164,33 @@ class DebugCommand(Command): %prog """ ignore_require_venv = True - def __init__(self, *args, **kw): - super(DebugCommand, self).__init__(*args, **kw) - - cmd_opts = self.cmd_opts - cmdoptions.add_target_python_options(cmd_opts) - self.parser.insert_option_group(0, cmd_opts) + def add_options(self) -> None: + cmdoptions.add_target_python_options(self.cmd_opts) + self.parser.insert_option_group(0, self.cmd_opts) self.parser.config.load() - def run(self, options, args): - # type: (Values, List[Any]) -> int + def run(self, options: Values, args: List[str]) -> int: logger.warning( "This command is only meant for debugging. " "Do not use this with automation for parsing and getting these " "details, since the output and options of this command may " "change without notice." ) - show_value('pip version', get_pip_version()) - show_value('sys.version', sys.version) - show_value('sys.executable', sys.executable) - show_value('sys.getdefaultencoding', sys.getdefaultencoding()) - show_value('sys.getfilesystemencoding', sys.getfilesystemencoding()) + show_value("pip version", get_pip_version()) + show_value("sys.version", sys.version) + show_value("sys.executable", sys.executable) + show_value("sys.getdefaultencoding", sys.getdefaultencoding()) + show_value("sys.getfilesystemencoding", sys.getfilesystemencoding()) show_value( - 'locale.getpreferredencoding', locale.getpreferredencoding(), + "locale.getpreferredencoding", + locale.getpreferredencoding(), ) - show_value('sys.platform', sys.platform) + show_value("sys.platform", sys.platform) show_sys_implementation() show_value("'cert' config value", ca_bundle_info(self.parser.config)) - show_value("REQUESTS_CA_BUNDLE", os.environ.get('REQUESTS_CA_BUNDLE')) - show_value("CURL_CA_BUNDLE", os.environ.get('CURL_CA_BUNDLE')) + show_value("REQUESTS_CA_BUNDLE", os.environ.get("REQUESTS_CA_BUNDLE")) + show_value("CURL_CA_BUNDLE", os.environ.get("CURL_CA_BUNDLE")) show_value("pip._vendor.certifi.where()", where()) show_value("pip._vendor.DEBUNDLED", pip._vendor.DEBUNDLED) diff --git a/src/fetchcode/vcs/pip/_internal/commands/download.py b/src/fetchcode/vcs/pip/_internal/commands/download.py index 5c2f19c2..917bbb91 100644 --- a/src/fetchcode/vcs/pip/_internal/commands/download.py +++ b/src/fetchcode/vcs/pip/_internal/commands/download.py @@ -1,17 +1,16 @@ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import - import logging import os - -from fetchcode.vcs.pip._internal.cli import cmdoptions -from fetchcode.vcs.pip._internal.cli.cmdoptions import make_target_python -from fetchcode.vcs.pip._internal.cli.req_command import RequirementCommand, with_cleanup -from fetchcode.vcs.pip._internal.req.req_tracker import get_requirement_tracker -from fetchcode.vcs.pip._internal.utils.misc import ensure_dir, normalize_path, write_output -from fetchcode.vcs.pip._internal.utils.temp_dir import TempDirectory +from optparse import Values +from typing import List + +from pip._internal.cli import cmdoptions +from pip._internal.cli.cmdoptions import make_target_python +from pip._internal.cli.req_command import RequirementCommand, with_cleanup +from pip._internal.cli.status_codes import SUCCESS +from pip._internal.operations.build.build_tracker import get_build_tracker +from pip._internal.req.req_install import check_legacy_setup_py_options +from pip._internal.utils.misc import ensure_dir, normalize_path, write_output +from pip._internal.utils.temp_dir import TempDirectory logger = logging.getLogger(__name__) @@ -36,36 +35,36 @@ class DownloadCommand(RequirementCommand): %prog [options] ... %prog [options] ...""" - def __init__(self, *args, **kw): - super(DownloadCommand, self).__init__(*args, **kw) - - cmd_opts = self.cmd_opts - - cmd_opts.add_option(cmdoptions.constraints()) - cmd_opts.add_option(cmdoptions.requirements()) - cmd_opts.add_option(cmdoptions.build_dir()) - cmd_opts.add_option(cmdoptions.no_deps()) - cmd_opts.add_option(cmdoptions.global_options()) - cmd_opts.add_option(cmdoptions.no_binary()) - cmd_opts.add_option(cmdoptions.only_binary()) - cmd_opts.add_option(cmdoptions.prefer_binary()) - cmd_opts.add_option(cmdoptions.src()) - cmd_opts.add_option(cmdoptions.pre()) - cmd_opts.add_option(cmdoptions.require_hashes()) - cmd_opts.add_option(cmdoptions.progress_bar()) - cmd_opts.add_option(cmdoptions.no_build_isolation()) - cmd_opts.add_option(cmdoptions.use_pep517()) - cmd_opts.add_option(cmdoptions.no_use_pep517()) - - cmd_opts.add_option( - '-d', '--dest', '--destination-dir', '--destination-directory', - dest='download_dir', - metavar='dir', + def add_options(self) -> None: + self.cmd_opts.add_option(cmdoptions.constraints()) + self.cmd_opts.add_option(cmdoptions.requirements()) + self.cmd_opts.add_option(cmdoptions.no_deps()) + self.cmd_opts.add_option(cmdoptions.global_options()) + self.cmd_opts.add_option(cmdoptions.no_binary()) + self.cmd_opts.add_option(cmdoptions.only_binary()) + self.cmd_opts.add_option(cmdoptions.prefer_binary()) + self.cmd_opts.add_option(cmdoptions.src()) + self.cmd_opts.add_option(cmdoptions.pre()) + self.cmd_opts.add_option(cmdoptions.require_hashes()) + self.cmd_opts.add_option(cmdoptions.progress_bar()) + self.cmd_opts.add_option(cmdoptions.no_build_isolation()) + self.cmd_opts.add_option(cmdoptions.use_pep517()) + self.cmd_opts.add_option(cmdoptions.no_use_pep517()) + self.cmd_opts.add_option(cmdoptions.check_build_deps()) + self.cmd_opts.add_option(cmdoptions.ignore_requires_python()) + + self.cmd_opts.add_option( + "-d", + "--dest", + "--destination-dir", + "--destination-directory", + dest="download_dir", + metavar="dir", default=os.curdir, - help=("Download packages into ."), + help="Download packages into .", ) - cmdoptions.add_target_python_options(cmd_opts) + cmdoptions.add_target_python_options(self.cmd_opts) index_opts = cmdoptions.make_option_group( cmdoptions.index_group, @@ -73,10 +72,10 @@ def __init__(self, *args, **kw): ) self.parser.insert_option_group(0, index_opts) - self.parser.insert_option_group(0, cmd_opts) + self.parser.insert_option_group(0, self.cmd_opts) @with_cleanup - def run(self, options, args): + def run(self, options: Values, args: List[str]) -> int: options.ignore_installed = True # editable doesn't really make sense for `pip download`, but the bowels # of the RequirementSet code require that property. @@ -85,7 +84,6 @@ def run(self, options, args): cmdoptions.check_dist_restriction(options) options.download_dir = normalize_path(options.download_dir) - ensure_dir(options.download_dir) session = self.get_default_session(options) @@ -95,48 +93,54 @@ def run(self, options, args): options=options, session=session, target_python=target_python, + ignore_requires_python=options.ignore_requires_python, ) - build_delete = (not (options.no_clean or options.build_dir)) - req_tracker = self.enter_context(get_requirement_tracker()) + build_tracker = self.enter_context(get_build_tracker()) directory = TempDirectory( - options.build_dir, - delete=build_delete, + delete=not options.no_clean, kind="download", globally_managed=True, ) reqs = self.get_requirements(args, options, finder, session) + check_legacy_setup_py_options(options, reqs) preparer = self.make_requirement_preparer( temp_build_dir=directory, options=options, - req_tracker=req_tracker, + build_tracker=build_tracker, session=session, finder=finder, download_dir=options.download_dir, use_user_site=False, + verbosity=self.verbosity, ) resolver = self.make_resolver( preparer=preparer, finder=finder, options=options, + ignore_requires_python=options.ignore_requires_python, + use_pep517=options.use_pep517, py_version_info=options.python_version, ) self.trace_basic_info(finder) - requirement_set = resolver.resolve( - reqs, check_supported_wheels=True - ) + requirement_set = resolver.resolve(reqs, check_supported_wheels=True) + + downloaded: List[str] = [] + for req in requirement_set.requirements.values(): + if req.satisfied_by is None: + assert req.name is not None + preparer.save_linked_requirement(req) + downloaded.append(req.name) + + preparer.prepare_linked_requirements_more(requirement_set.requirements.values()) - downloaded = ' '.join([ - req.name for req in requirement_set.requirements.values() - if req.successfully_downloaded - ]) if downloaded: - write_output('Successfully downloaded %s', downloaded) + write_output("Successfully downloaded %s", " ".join(downloaded)) - return requirement_set + return SUCCESS diff --git a/src/fetchcode/vcs/pip/_internal/commands/freeze.py b/src/fetchcode/vcs/pip/_internal/commands/freeze.py index c7f07cc8..885fdfeb 100644 --- a/src/fetchcode/vcs/pip/_internal/commands/freeze.py +++ b/src/fetchcode/vcs/pip/_internal/commands/freeze.py @@ -1,18 +1,25 @@ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False +import sys +from optparse import Values +from typing import AbstractSet, List -from __future__ import absolute_import +from pip._internal.cli import cmdoptions +from pip._internal.cli.base_command import Command +from pip._internal.cli.status_codes import SUCCESS +from pip._internal.operations.freeze import freeze +from pip._internal.utils.compat import stdlib_pkgs + + +def _should_suppress_build_backends() -> bool: + return sys.version_info < (3, 12) -import sys -from fetchcode.vcs.pip._internal.cache import WheelCache -from fetchcode.vcs.pip._internal.cli import cmdoptions -from fetchcode.vcs.pip._internal.cli.base_command import Command -from fetchcode.vcs.pip._internal.models.format_control import FormatControl -from fetchcode.vcs.pip._internal.operations.freeze import freeze -from fetchcode.vcs.pip._internal.utils.compat import stdlib_pkgs +def _dev_pkgs() -> AbstractSet[str]: + pkgs = {"pip"} -DEV_PKGS = {'pip', 'setuptools', 'distribute', 'wheel'} + if _should_suppress_build_backends(): + pkgs |= {"setuptools", "distribute", "wheel"} + + return pkgs class FreezeCommand(Command): @@ -22,78 +29,81 @@ class FreezeCommand(Command): packages are listed in a case-insensitive sorted order. """ + ignore_require_venv = True usage = """ %prog [options]""" log_streams = ("ext://sys.stderr", "ext://sys.stderr") - def __init__(self, *args, **kw): - super(FreezeCommand, self).__init__(*args, **kw) - - self.cmd_opts.add_option( - '-r', '--requirement', - dest='requirements', - action='append', - default=[], - metavar='file', - help="Use the order in the given requirements file and its " - "comments when generating output. This option can be " - "used multiple times.") + def add_options(self) -> None: self.cmd_opts.add_option( - '-f', '--find-links', - dest='find_links', - action='append', + "-r", + "--requirement", + dest="requirements", + action="append", default=[], - metavar='URL', - help='URL for finding packages, which will be added to the ' - 'output.') + metavar="file", + help=( + "Use the order in the given requirements file and its " + "comments when generating output. This option can be " + "used multiple times." + ), + ) self.cmd_opts.add_option( - '-l', '--local', - dest='local', - action='store_true', + "-l", + "--local", + dest="local", + action="store_true", default=False, - help='If in a virtualenv that has global access, do not output ' - 'globally-installed packages.') + help=( + "If in a virtualenv that has global access, do not output " + "globally-installed packages." + ), + ) self.cmd_opts.add_option( - '--user', - dest='user', - action='store_true', + "--user", + dest="user", + action="store_true", default=False, - help='Only output packages installed in user-site.') + help="Only output packages installed in user-site.", + ) self.cmd_opts.add_option(cmdoptions.list_path()) self.cmd_opts.add_option( - '--all', - dest='freeze_all', - action='store_true', - help='Do not skip these packages in the output:' - ' {}'.format(', '.join(DEV_PKGS))) + "--all", + dest="freeze_all", + action="store_true", + help=( + "Do not skip these packages in the output:" + " {}".format(", ".join(_dev_pkgs())) + ), + ) self.cmd_opts.add_option( - '--exclude-editable', - dest='exclude_editable', - action='store_true', - help='Exclude editable package from output.') + "--exclude-editable", + dest="exclude_editable", + action="store_true", + help="Exclude editable package from output.", + ) + self.cmd_opts.add_option(cmdoptions.list_exclude()) self.parser.insert_option_group(0, self.cmd_opts) - def run(self, options, args): - format_control = FormatControl(set(), set()) - wheel_cache = WheelCache(options.cache_dir, format_control) + def run(self, options: Values, args: List[str]) -> int: skip = set(stdlib_pkgs) if not options.freeze_all: - skip.update(DEV_PKGS) + skip.update(_dev_pkgs()) + + if options.excludes: + skip.update(options.excludes) cmdoptions.check_list_path_option(options) - freeze_kwargs = dict( + for line in freeze( requirement=options.requirements, - find_links=options.find_links, local_only=options.local, user_only=options.user, paths=options.path, isolated=options.isolated_mode, - wheel_cache=wheel_cache, skip=skip, exclude_editable=options.exclude_editable, - ) - - for line in freeze(**freeze_kwargs): - sys.stdout.write(line + '\n') + ): + sys.stdout.write(line + "\n") + return SUCCESS diff --git a/src/fetchcode/vcs/pip/_internal/commands/hash.py b/src/fetchcode/vcs/pip/_internal/commands/hash.py index fd49bf45..042dac81 100644 --- a/src/fetchcode/vcs/pip/_internal/commands/hash.py +++ b/src/fetchcode/vcs/pip/_internal/commands/hash.py @@ -1,16 +1,13 @@ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import - import hashlib import logging import sys +from optparse import Values +from typing import List -from fetchcode.vcs.pip._internal.cli.base_command import Command -from fetchcode.vcs.pip._internal.cli.status_codes import ERROR -from fetchcode.vcs.pip._internal.utils.hashes import FAVORITE_HASH, STRONG_HASHES -from fetchcode.vcs.pip._internal.utils.misc import read_chunks, write_output +from pip._internal.cli.base_command import Command +from pip._internal.cli.status_codes import ERROR, SUCCESS +from pip._internal.utils.hashes import FAVORITE_HASH, STRONG_HASHES +from pip._internal.utils.misc import read_chunks, write_output logger = logging.getLogger(__name__) @@ -23,35 +20,39 @@ class HashCommand(Command): installs. """ - usage = '%prog [options] ...' + usage = "%prog [options] ..." ignore_require_venv = True - def __init__(self, *args, **kw): - super(HashCommand, self).__init__(*args, **kw) + def add_options(self) -> None: self.cmd_opts.add_option( - '-a', '--algorithm', - dest='algorithm', + "-a", + "--algorithm", + dest="algorithm", choices=STRONG_HASHES, - action='store', + action="store", default=FAVORITE_HASH, - help='The hash algorithm to use: one of {}'.format( - ', '.join(STRONG_HASHES))) + help="The hash algorithm to use: one of {}".format( + ", ".join(STRONG_HASHES) + ), + ) self.parser.insert_option_group(0, self.cmd_opts) - def run(self, options, args): + def run(self, options: Values, args: List[str]) -> int: if not args: self.parser.print_usage(sys.stderr) return ERROR algorithm = options.algorithm for path in args: - write_output('%s:\n--hash=%s:%s', - path, algorithm, _hash_of_file(path, algorithm)) + write_output( + "%s:\n--hash=%s:%s", path, algorithm, _hash_of_file(path, algorithm) + ) + return SUCCESS -def _hash_of_file(path, algorithm): +def _hash_of_file(path: str, algorithm: str) -> str: """Return the hash digest of a file.""" - with open(path, 'rb') as archive: + with open(path, "rb") as archive: hash = hashlib.new(algorithm) for chunk in read_chunks(archive): hash.update(chunk) diff --git a/src/fetchcode/vcs/pip/_internal/commands/help.py b/src/fetchcode/vcs/pip/_internal/commands/help.py index fc6288bc..62066318 100644 --- a/src/fetchcode/vcs/pip/_internal/commands/help.py +++ b/src/fetchcode/vcs/pip/_internal/commands/help.py @@ -1,11 +1,9 @@ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False +from optparse import Values +from typing import List -from __future__ import absolute_import - -from fetchcode.vcs.pip._internal.cli.base_command import Command -from fetchcode.vcs.pip._internal.cli.status_codes import SUCCESS -from fetchcode.vcs.pip._internal.exceptions import CommandError +from pip._internal.cli.base_command import Command +from pip._internal.cli.status_codes import SUCCESS +from pip._internal.exceptions import CommandError class HelpCommand(Command): @@ -15,9 +13,11 @@ class HelpCommand(Command): %prog """ ignore_require_venv = True - def run(self, options, args): - from fetchcode.vcs.pip._internal.commands import ( - commands_dict, create_command, get_similar_commands, + def run(self, options: Values, args: List[str]) -> int: + from pip._internal.commands import ( + commands_dict, + create_command, + get_similar_commands, ) try: @@ -29,11 +29,11 @@ def run(self, options, args): if cmd_name not in commands_dict: guess = get_similar_commands(cmd_name) - msg = ['unknown command "{}"'.format(cmd_name)] + msg = [f'unknown command "{cmd_name}"'] if guess: - msg.append('maybe you meant "{}"'.format(guess)) + msg.append(f'maybe you meant "{guess}"') - raise CommandError(' - '.join(msg)) + raise CommandError(" - ".join(msg)) command = create_command(cmd_name) command.parser.print_help() diff --git a/src/fetchcode/vcs/pip/_internal/commands/index.py b/src/fetchcode/vcs/pip/_internal/commands/index.py new file mode 100644 index 00000000..2e2661bb --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/commands/index.py @@ -0,0 +1,139 @@ +import logging +from optparse import Values +from typing import Any, Iterable, List, Optional + +from pip._vendor.packaging.version import Version + +from pip._internal.cli import cmdoptions +from pip._internal.cli.req_command import IndexGroupCommand +from pip._internal.cli.status_codes import ERROR, SUCCESS +from pip._internal.commands.search import print_dist_installation_info +from pip._internal.exceptions import CommandError, DistributionNotFound, PipError +from pip._internal.index.collector import LinkCollector +from pip._internal.index.package_finder import PackageFinder +from pip._internal.models.selection_prefs import SelectionPreferences +from pip._internal.models.target_python import TargetPython +from pip._internal.network.session import PipSession +from pip._internal.utils.misc import write_output + +logger = logging.getLogger(__name__) + + +class IndexCommand(IndexGroupCommand): + """ + Inspect information available from package indexes. + """ + + ignore_require_venv = True + usage = """ + %prog versions + """ + + def add_options(self) -> None: + cmdoptions.add_target_python_options(self.cmd_opts) + + self.cmd_opts.add_option(cmdoptions.ignore_requires_python()) + self.cmd_opts.add_option(cmdoptions.pre()) + self.cmd_opts.add_option(cmdoptions.no_binary()) + self.cmd_opts.add_option(cmdoptions.only_binary()) + + index_opts = cmdoptions.make_option_group( + cmdoptions.index_group, + self.parser, + ) + + self.parser.insert_option_group(0, index_opts) + self.parser.insert_option_group(0, self.cmd_opts) + + def run(self, options: Values, args: List[str]) -> int: + handlers = { + "versions": self.get_available_package_versions, + } + + logger.warning( + "pip index is currently an experimental command. " + "It may be removed/changed in a future release " + "without prior warning." + ) + + # Determine action + if not args or args[0] not in handlers: + logger.error( + "Need an action (%s) to perform.", + ", ".join(sorted(handlers)), + ) + return ERROR + + action = args[0] + + # Error handling happens here, not in the action-handlers. + try: + handlers[action](options, args[1:]) + except PipError as e: + logger.error(e.args[0]) + return ERROR + + return SUCCESS + + def _build_package_finder( + self, + options: Values, + session: PipSession, + target_python: Optional[TargetPython] = None, + ignore_requires_python: Optional[bool] = None, + ) -> PackageFinder: + """ + Create a package finder appropriate to the index command. + """ + link_collector = LinkCollector.create(session, options=options) + + # Pass allow_yanked=False to ignore yanked versions. + selection_prefs = SelectionPreferences( + allow_yanked=False, + allow_all_prereleases=options.pre, + ignore_requires_python=ignore_requires_python, + ) + + return PackageFinder.create( + link_collector=link_collector, + selection_prefs=selection_prefs, + target_python=target_python, + ) + + def get_available_package_versions(self, options: Values, args: List[Any]) -> None: + if len(args) != 1: + raise CommandError("You need to specify exactly one argument") + + target_python = cmdoptions.make_target_python(options) + query = args[0] + + with self._build_session(options) as session: + finder = self._build_package_finder( + options=options, + session=session, + target_python=target_python, + ignore_requires_python=options.ignore_requires_python, + ) + + versions: Iterable[Version] = ( + candidate.version for candidate in finder.find_all_candidates(query) + ) + + if not options.pre: + # Remove prereleases + versions = ( + version for version in versions if not version.is_prerelease + ) + versions = set(versions) + + if not versions: + raise DistributionNotFound( + f"No matching distribution found for {query}" + ) + + formatted_versions = [str(ver) for ver in sorted(versions, reverse=True)] + latest = formatted_versions[0] + + write_output(f"{query} ({latest})") + write_output("Available versions: {}".format(", ".join(formatted_versions))) + print_dist_installation_info(query, latest) diff --git a/src/fetchcode/vcs/pip/_internal/commands/inspect.py b/src/fetchcode/vcs/pip/_internal/commands/inspect.py new file mode 100644 index 00000000..e810c131 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/commands/inspect.py @@ -0,0 +1,92 @@ +import logging +from optparse import Values +from typing import Any, Dict, List + +from pip._vendor.packaging.markers import default_environment +from pip._vendor.rich import print_json + +from pip import __version__ +from pip._internal.cli import cmdoptions +from pip._internal.cli.base_command import Command +from pip._internal.cli.status_codes import SUCCESS +from pip._internal.metadata import BaseDistribution, get_environment +from pip._internal.utils.compat import stdlib_pkgs +from pip._internal.utils.urls import path_to_url + +logger = logging.getLogger(__name__) + + +class InspectCommand(Command): + """ + Inspect the content of a Python environment and produce a report in JSON format. + """ + + ignore_require_venv = True + usage = """ + %prog [options]""" + + def add_options(self) -> None: + self.cmd_opts.add_option( + "--local", + action="store_true", + default=False, + help=( + "If in a virtualenv that has global access, do not list " + "globally-installed packages." + ), + ) + self.cmd_opts.add_option( + "--user", + dest="user", + action="store_true", + default=False, + help="Only output packages installed in user-site.", + ) + self.cmd_opts.add_option(cmdoptions.list_path()) + self.parser.insert_option_group(0, self.cmd_opts) + + def run(self, options: Values, args: List[str]) -> int: + cmdoptions.check_list_path_option(options) + dists = get_environment(options.path).iter_installed_distributions( + local_only=options.local, + user_only=options.user, + skip=set(stdlib_pkgs), + ) + output = { + "version": "1", + "pip_version": __version__, + "installed": [self._dist_to_dict(dist) for dist in dists], + "environment": default_environment(), + # TODO tags? scheme? + } + print_json(data=output) + return SUCCESS + + def _dist_to_dict(self, dist: BaseDistribution) -> Dict[str, Any]: + res: Dict[str, Any] = { + "metadata": dist.metadata_dict, + "metadata_location": dist.info_location, + } + # direct_url. Note that we don't have download_info (as in the installation + # report) since it is not recorded in installed metadata. + direct_url = dist.direct_url + if direct_url is not None: + res["direct_url"] = direct_url.to_dict() + else: + # Emulate direct_url for legacy editable installs. + editable_project_location = dist.editable_project_location + if editable_project_location is not None: + res["direct_url"] = { + "url": path_to_url(editable_project_location), + "dir_info": { + "editable": True, + }, + } + # installer + installer = dist.installer + if dist.installer: + res["installer"] = installer + # requested + if dist.installed_with_dist_info: + res["requested"] = dist.requested + return res diff --git a/src/fetchcode/vcs/pip/_internal/commands/install.py b/src/fetchcode/vcs/pip/_internal/commands/install.py index 9225cdd0..ad45a2f2 100644 --- a/src/fetchcode/vcs/pip/_internal/commands/install.py +++ b/src/fetchcode/vcs/pip/_internal/commands/install.py @@ -1,70 +1,53 @@ -# The following comment should be removed at some point in the future. -# It's included for now because without it InstallCommand.run() has a -# couple errors where we have to know req.name is str rather than -# Optional[str] for the InstallRequirement req. -# mypy: strict-optional=False -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import - import errno -import logging +import json import operator import os import shutil import site -from optparse import SUPPRESS_HELP - -from fetchcode.vcs.pip._vendor import pkg_resources -from fetchcode.vcs.pip._vendor.packaging.utils import canonicalize_name - -from fetchcode.vcs.pip._internal.cache import WheelCache -from fetchcode.vcs.pip._internal.cli import cmdoptions -from fetchcode.vcs.pip._internal.cli.cmdoptions import make_target_python -from fetchcode.vcs.pip._internal.cli.req_command import RequirementCommand, with_cleanup -from fetchcode.vcs.pip._internal.cli.status_codes import ERROR, SUCCESS -from fetchcode.vcs.pip._internal.exceptions import CommandError, InstallationError -from fetchcode.vcs.pip._internal.locations import distutils_scheme -from fetchcode.vcs.pip._internal.operations.check import check_install_conflicts -from fetchcode.vcs.pip._internal.req import install_given_reqs -from fetchcode.vcs.pip._internal.req.req_tracker import get_requirement_tracker -from fetchcode.vcs.pip._internal.utils.deprecation import deprecated -from fetchcode.vcs.pip._internal.utils.distutils_args import parse_distutils_args -from fetchcode.vcs.pip._internal.utils.filesystem import test_writable_dir -from fetchcode.vcs.pip._internal.utils.misc import ( +from optparse import SUPPRESS_HELP, Values +from typing import List, Optional + +from pip._vendor.packaging.utils import canonicalize_name +from pip._vendor.rich import print_json + +from pip._internal.cache import WheelCache +from pip._internal.cli import cmdoptions +from pip._internal.cli.cmdoptions import make_target_python +from pip._internal.cli.req_command import ( + RequirementCommand, + with_cleanup, +) +from pip._internal.cli.status_codes import ERROR, SUCCESS +from pip._internal.exceptions import CommandError, InstallationError +from pip._internal.locations import get_scheme +from pip._internal.metadata import get_environment +from pip._internal.models.installation_report import InstallationReport +from pip._internal.operations.build.build_tracker import get_build_tracker +from pip._internal.operations.check import ConflictDetails, check_install_conflicts +from pip._internal.req import install_given_reqs +from pip._internal.req.req_install import ( + InstallRequirement, + check_legacy_setup_py_options, +) +from pip._internal.utils.compat import WINDOWS +from pip._internal.utils.filesystem import test_writable_dir +from pip._internal.utils.logging import getLogger +from pip._internal.utils.misc import ( + check_externally_managed, ensure_dir, - get_installed_version, + get_pip_version, protect_pip_from_modification_on_windows, + warn_if_run_as_root, write_output, ) -from fetchcode.vcs.pip._internal.utils.temp_dir import TempDirectory -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.utils.virtualenv import virtualenv_no_global -from fetchcode.vcs.pip._internal.wheel_builder import build, should_build_for_install_command - -if MYPY_CHECK_RUNNING: - from optparse import Values - from typing import Any, Iterable, List, Optional - - from fetchcode.vcs.pip._internal.models.format_control import FormatControl - from fetchcode.vcs.pip._internal.req.req_install import InstallRequirement - from fetchcode.vcs.pip._internal.wheel_builder import BinaryAllowedPredicate - - -logger = logging.getLogger(__name__) - - -def get_check_binary_allowed(format_control): - # type: (FormatControl) -> BinaryAllowedPredicate - def check_binary_allowed(req): - # type: (InstallRequirement) -> bool - if req.use_pep517: - return True - canonical_name = canonicalize_name(req.name) - allowed_formats = format_control.get_allowed_formats(canonical_name) - return "binary" in allowed_formats +from pip._internal.utils.temp_dir import TempDirectory +from pip._internal.utils.virtualenv import ( + running_under_virtualenv, + virtualenv_no_global, +) +from pip._internal.wheel_builder import build, should_build_for_install_command - return check_binary_allowed +logger = getLogger(__name__) class InstallCommand(RequirementCommand): @@ -87,110 +70,140 @@ class InstallCommand(RequirementCommand): %prog [options] [-e] ... %prog [options] ...""" - def __init__(self, *args, **kw): - super(InstallCommand, self).__init__(*args, **kw) + def add_options(self) -> None: + self.cmd_opts.add_option(cmdoptions.requirements()) + self.cmd_opts.add_option(cmdoptions.constraints()) + self.cmd_opts.add_option(cmdoptions.no_deps()) + self.cmd_opts.add_option(cmdoptions.pre()) - cmd_opts = self.cmd_opts - - cmd_opts.add_option(cmdoptions.requirements()) - cmd_opts.add_option(cmdoptions.constraints()) - cmd_opts.add_option(cmdoptions.no_deps()) - cmd_opts.add_option(cmdoptions.pre()) - - cmd_opts.add_option(cmdoptions.editable()) - cmd_opts.add_option( - '-t', '--target', - dest='target_dir', - metavar='dir', + self.cmd_opts.add_option(cmdoptions.editable()) + self.cmd_opts.add_option( + "--dry-run", + action="store_true", + dest="dry_run", + default=False, + help=( + "Don't actually install anything, just print what would be. " + "Can be used in combination with --ignore-installed " + "to 'resolve' the requirements." + ), + ) + self.cmd_opts.add_option( + "-t", + "--target", + dest="target_dir", + metavar="dir", default=None, - help='Install packages into . ' - 'By default this will not replace existing files/folders in ' - '. Use --upgrade to replace existing packages in ' - 'with new versions.' + help=( + "Install packages into . " + "By default this will not replace existing files/folders in " + ". Use --upgrade to replace existing packages in " + "with new versions." + ), + ) + cmdoptions.add_target_python_options(self.cmd_opts) + + self.cmd_opts.add_option( + "--user", + dest="use_user_site", + action="store_true", + help=( + "Install to the Python user install directory for your " + "platform. Typically ~/.local/, or %APPDATA%\\Python on " + "Windows. (See the Python documentation for site.USER_BASE " + "for full details.)" + ), ) - cmdoptions.add_target_python_options(cmd_opts) - - cmd_opts.add_option( - '--user', - dest='use_user_site', - action='store_true', - help="Install to the Python user install directory for your " - "platform. Typically ~/.local/, or %APPDATA%\\Python on " - "Windows. (See the Python documentation for site.USER_BASE " - "for full details.)") - cmd_opts.add_option( - '--no-user', - dest='use_user_site', - action='store_false', - help=SUPPRESS_HELP) - cmd_opts.add_option( - '--root', - dest='root_path', - metavar='dir', + self.cmd_opts.add_option( + "--no-user", + dest="use_user_site", + action="store_false", + help=SUPPRESS_HELP, + ) + self.cmd_opts.add_option( + "--root", + dest="root_path", + metavar="dir", default=None, - help="Install everything relative to this alternate root " - "directory.") - cmd_opts.add_option( - '--prefix', - dest='prefix_path', - metavar='dir', + help="Install everything relative to this alternate root directory.", + ) + self.cmd_opts.add_option( + "--prefix", + dest="prefix_path", + metavar="dir", default=None, - help="Installation prefix where lib, bin and other top-level " - "folders are placed") + help=( + "Installation prefix where lib, bin and other top-level " + "folders are placed. Note that the resulting installation may " + "contain scripts and other resources which reference the " + "Python interpreter of pip, and not that of ``--prefix``. " + "See also the ``--python`` option if the intention is to " + "install packages into another (possibly pip-free) " + "environment." + ), + ) - cmd_opts.add_option(cmdoptions.build_dir()) + self.cmd_opts.add_option(cmdoptions.src()) - cmd_opts.add_option(cmdoptions.src()) + self.cmd_opts.add_option( + "-U", + "--upgrade", + dest="upgrade", + action="store_true", + help=( + "Upgrade all specified packages to the newest available " + "version. The handling of dependencies depends on the " + "upgrade-strategy used." + ), + ) - cmd_opts.add_option( - '-U', '--upgrade', - dest='upgrade', - action='store_true', - help='Upgrade all specified packages to the newest available ' - 'version. The handling of dependencies depends on the ' - 'upgrade-strategy used.' + self.cmd_opts.add_option( + "--upgrade-strategy", + dest="upgrade_strategy", + default="only-if-needed", + choices=["only-if-needed", "eager"], + help=( + "Determines how dependency upgrading should be handled " + "[default: %default]. " + '"eager" - dependencies are upgraded regardless of ' + "whether the currently installed version satisfies the " + "requirements of the upgraded package(s). " + '"only-if-needed" - are upgraded only when they do not ' + "satisfy the requirements of the upgraded package(s)." + ), ) - cmd_opts.add_option( - '--upgrade-strategy', - dest='upgrade_strategy', - default='only-if-needed', - choices=['only-if-needed', 'eager'], - help='Determines how dependency upgrading should be handled ' - '[default: %default]. ' - '"eager" - dependencies are upgraded regardless of ' - 'whether the currently installed version satisfies the ' - 'requirements of the upgraded package(s). ' - '"only-if-needed" - are upgraded only when they do not ' - 'satisfy the requirements of the upgraded package(s).' + self.cmd_opts.add_option( + "--force-reinstall", + dest="force_reinstall", + action="store_true", + help="Reinstall all packages even if they are already up-to-date.", ) - cmd_opts.add_option( - '--force-reinstall', - dest='force_reinstall', - action='store_true', - help='Reinstall all packages even if they are already ' - 'up-to-date.') - - cmd_opts.add_option( - '-I', '--ignore-installed', - dest='ignore_installed', - action='store_true', - help='Ignore the installed packages, overwriting them. ' - 'This can break your system if the existing package ' - 'is of a different version or was installed ' - 'with a different package manager!' + self.cmd_opts.add_option( + "-I", + "--ignore-installed", + dest="ignore_installed", + action="store_true", + help=( + "Ignore the installed packages, overwriting them. " + "This can break your system if the existing package " + "is of a different version or was installed " + "with a different package manager!" + ), ) - cmd_opts.add_option(cmdoptions.ignore_requires_python()) - cmd_opts.add_option(cmdoptions.no_build_isolation()) - cmd_opts.add_option(cmdoptions.use_pep517()) - cmd_opts.add_option(cmdoptions.no_use_pep517()) + self.cmd_opts.add_option(cmdoptions.ignore_requires_python()) + self.cmd_opts.add_option(cmdoptions.no_build_isolation()) + self.cmd_opts.add_option(cmdoptions.use_pep517()) + self.cmd_opts.add_option(cmdoptions.no_use_pep517()) + self.cmd_opts.add_option(cmdoptions.check_build_deps()) + self.cmd_opts.add_option(cmdoptions.override_externally_managed()) - cmd_opts.add_option(cmdoptions.install_options()) - cmd_opts.add_option(cmdoptions.global_options()) + self.cmd_opts.add_option(cmdoptions.config_settings()) + self.cmd_opts.add_option(cmdoptions.global_options()) - cmd_opts.add_option( + self.cmd_opts.add_option( "--compile", action="store_true", dest="compile", @@ -198,33 +211,33 @@ def __init__(self, *args, **kw): help="Compile Python source files to bytecode", ) - cmd_opts.add_option( + self.cmd_opts.add_option( "--no-compile", action="store_false", dest="compile", help="Do not compile Python source files to bytecode", ) - cmd_opts.add_option( + self.cmd_opts.add_option( "--no-warn-script-location", action="store_false", dest="warn_script_location", default=True, help="Do not warn when installing scripts outside PATH", ) - cmd_opts.add_option( + self.cmd_opts.add_option( "--no-warn-conflicts", action="store_false", dest="warn_about_conflicts", default=True, help="Do not warn about broken dependencies", ) - - cmd_opts.add_option(cmdoptions.no_binary()) - cmd_opts.add_option(cmdoptions.only_binary()) - cmd_opts.add_option(cmdoptions.prefer_binary()) - cmd_opts.add_option(cmdoptions.require_hashes()) - cmd_opts.add_option(cmdoptions.progress_bar()) + self.cmd_opts.add_option(cmdoptions.no_binary()) + self.cmd_opts.add_option(cmdoptions.only_binary()) + self.cmd_opts.add_option(cmdoptions.prefer_binary()) + self.cmd_opts.add_option(cmdoptions.require_hashes()) + self.cmd_opts.add_option(cmdoptions.progress_bar()) + self.cmd_opts.add_option(cmdoptions.root_user_action()) index_opts = cmdoptions.make_option_group( cmdoptions.index_group, @@ -232,23 +245,53 @@ def __init__(self, *args, **kw): ) self.parser.insert_option_group(0, index_opts) - self.parser.insert_option_group(0, cmd_opts) + self.parser.insert_option_group(0, self.cmd_opts) + + self.cmd_opts.add_option( + "--report", + dest="json_report_file", + metavar="file", + default=None, + help=( + "Generate a JSON file describing what pip did to install " + "the provided requirements. " + "Can be used in combination with --dry-run and --ignore-installed " + "to 'resolve' the requirements. " + "When - is used as file name it writes to stdout. " + "When writing to stdout, please combine with the --quiet option " + "to avoid mixing pip logging output with JSON output." + ), + ) @with_cleanup - def run(self, options, args): - # type: (Values, List[Any]) -> int + def run(self, options: Values, args: List[str]) -> int: if options.use_user_site and options.target_dir is not None: raise CommandError("Can not combine '--user' and '--target'") - cmdoptions.check_install_build_global(options) + # Check whether the environment we're installing into is externally + # managed, as specified in PEP 668. Specifying --root, --target, or + # --prefix disables the check, since there's no reliable way to locate + # the EXTERNALLY-MANAGED file for those cases. An exception is also + # made specifically for "--dry-run --report" for convenience. + installing_into_current_environment = ( + not (options.dry_run and options.json_report_file) + and options.root_path is None + and options.target_dir is None + and options.prefix_path is None + ) + if ( + installing_into_current_environment + and not options.override_externally_managed + ): + check_externally_managed() + upgrade_strategy = "to-satisfy-only" if options.upgrade: upgrade_strategy = options.upgrade_strategy cmdoptions.check_dist_restriction(options, check_target=True) - install_options = options.install_options or [] - + logger.verbose("Using %s", get_pip_version()) options.use_user_site = decide_user_install( options.use_user_site, prefix_path=options.prefix_path, @@ -257,21 +300,25 @@ def run(self, options, args): isolated_mode=options.isolated_mode, ) - target_temp_dir = None # type: Optional[TempDirectory] - target_temp_dir_path = None # type: Optional[str] + target_temp_dir: Optional[TempDirectory] = None + target_temp_dir_path: Optional[str] = None if options.target_dir: options.ignore_installed = True options.target_dir = os.path.abspath(options.target_dir) - if (os.path.exists(options.target_dir) and not - os.path.isdir(options.target_dir)): + if ( + # fmt: off + os.path.exists(options.target_dir) and + not os.path.isdir(options.target_dir) + # fmt: on + ): raise CommandError( - "Target path exists but is not a directory, will not " - "continue." + "Target path exists but is not a directory, will not continue." ) # Create a target directory for using with the target option target_temp_dir = TempDirectory(kind="target") target_temp_dir_path = target_temp_dir.path + self.enter_context(target_temp_dir) global_options = options.global_options or [] @@ -284,35 +331,34 @@ def run(self, options, args): target_python=target_python, ignore_requires_python=options.ignore_requires_python, ) - build_delete = (not (options.no_clean or options.build_dir)) - wheel_cache = WheelCache(options.cache_dir, options.format_control) - - req_tracker = self.enter_context(get_requirement_tracker()) + build_tracker = self.enter_context(get_build_tracker()) directory = TempDirectory( - options.build_dir, - delete=build_delete, + delete=not options.no_clean, kind="install", globally_managed=True, ) try: - reqs = self.get_requirements( - args, options, finder, session, - check_supported_wheels=not options.target_dir, - ) + reqs = self.get_requirements(args, options, finder, session) + check_legacy_setup_py_options(options, reqs) - warn_deprecated_install_options( - reqs, options.install_options - ) + wheel_cache = WheelCache(options.cache_dir) + + # Only when installing is it permitted to use PEP 660. + # In other circumstances (pip wheel, pip download) we generate + # regular (i.e. non editable) metadata and wheels. + for req in reqs: + req.permit_editable_wheels = True preparer = self.make_requirement_preparer( temp_build_dir=directory, options=options, - req_tracker=req_tracker, + build_tracker=build_tracker, session=session, finder=finder, use_user_site=options.use_user_site, + verbosity=self.verbosity, ) resolver = self.make_resolver( preparer=preparer, @@ -325,6 +371,7 @@ def run(self, options, args): force_reinstall=options.force_reinstall, upgrade_strategy=upgrade_strategy, use_pep517=options.use_pep517, + py_version_info=options.python_version, ) self.trace_basic_info(finder) @@ -333,78 +380,89 @@ def run(self, options, args): reqs, check_supported_wheels=not options.target_dir ) + if options.json_report_file: + report = InstallationReport(requirement_set.requirements_to_install) + if options.json_report_file == "-": + print_json(data=report.to_dict()) + else: + with open(options.json_report_file, "w", encoding="utf-8") as f: + json.dump(report.to_dict(), f, indent=2, ensure_ascii=False) + + if options.dry_run: + would_install_items = sorted( + (r.metadata["name"], r.metadata["version"]) + for r in requirement_set.requirements_to_install + ) + if would_install_items: + write_output( + "Would install %s", + " ".join("-".join(item) for item in would_install_items), + ) + return SUCCESS + try: pip_req = requirement_set.get_requirement("pip") except KeyError: - modifying_pip = None + modifying_pip = False else: # If we're not replacing an already installed pip, # we're not modifying it. modifying_pip = pip_req.satisfied_by is None - protect_pip_from_modification_on_windows( - modifying_pip=modifying_pip - ) - - check_binary_allowed = get_check_binary_allowed( - finder.format_control - ) + if modifying_pip: + # Eagerly import this module to avoid crashes. Otherwise, this + # module would be imported *after* pip was replaced, resulting in + # crashes if the new self_outdated_check module was incompatible + # with the rest of pip that's already imported. + import pip._internal.self_outdated_check # noqa: F401 + protect_pip_from_modification_on_windows(modifying_pip=modifying_pip) reqs_to_build = [ - r for r in requirement_set.requirements.values() - if should_build_for_install_command( - r, check_binary_allowed - ) + r + for r in requirement_set.requirements.values() + if should_build_for_install_command(r) ] _, build_failures = build( reqs_to_build, wheel_cache=wheel_cache, + verify=True, build_options=[], - global_options=[], + global_options=global_options, ) - # If we're using PEP 517, we cannot do a direct install - # so we fail here. - # We don't care about failures building legacy - # requirements, as we'll fall through to a direct - # install for those. - pep517_build_failures = [ - r for r in build_failures if r.use_pep517 - ] - if pep517_build_failures: + if build_failures: raise InstallationError( - "Could not build wheels for {} which use" - " PEP 517 and cannot be installed directly".format( - ", ".join(r.name for r in pep517_build_failures))) + "ERROR: Failed to build installable wheels for some " + "pyproject.toml based projects ({})".format( + ", ".join(r.name for r in build_failures) # type: ignore + ) + ) - to_install = resolver.get_installation_order( - requirement_set - ) + to_install = resolver.get_installation_order(requirement_set) - # Consistency Checking of the package set we're installing. + # Check for conflicts in the package set we're installing. + conflicts: Optional[ConflictDetails] = None should_warn_about_conflicts = ( - not options.ignore_dependencies and - options.warn_about_conflicts + not options.ignore_dependencies and options.warn_about_conflicts ) if should_warn_about_conflicts: - self._warn_about_conflicts(to_install) + conflicts = self._determine_conflicts(to_install) # Don't warn about script install locations if - # --target has been specified + # --target or --prefix has been specified warn_script_location = options.warn_script_location - if options.target_dir: + if options.target_dir or options.prefix_path: warn_script_location = False installed = install_given_reqs( to_install, - install_options, global_options, root=options.root_path, home=target_temp_dir_path, prefix=options.prefix_path, - pycompile=options.compile, warn_script_location=warn_script_location, use_user_site=options.use_user_site, + pycompile=options.compile, ) lib_locations = get_lib_location_guesses( @@ -414,145 +472,206 @@ def run(self, options, args): prefix=options.prefix_path, isolated=options.isolated_mode, ) - working_set = pkg_resources.WorkingSet(lib_locations) - - installed.sort(key=operator.attrgetter('name')) - items = [] - for result in installed: - item = result.name - try: - installed_version = get_installed_version( - result.name, working_set=working_set - ) - if installed_version: - item += '-' + installed_version - except Exception: - pass - items.append(item) - installed_desc = ' '.join(items) + env = get_environment(lib_locations) + + # Display a summary of installed packages, with extra care to + # display a package name as it was requested by the user. + installed.sort(key=operator.attrgetter("name")) + summary = [] + installed_versions = {} + for distribution in env.iter_all_distributions(): + installed_versions[distribution.canonical_name] = distribution.version + for package in installed: + display_name = package.name + version = installed_versions.get(canonicalize_name(display_name), None) + if version: + text = f"{display_name}-{version}" + else: + text = display_name + summary.append(text) + + if conflicts is not None: + self._warn_about_conflicts( + conflicts, + resolver_variant=self.determine_resolver_variant(options), + ) + + installed_desc = " ".join(summary) if installed_desc: write_output( - 'Successfully installed %s', installed_desc, + "Successfully installed %s", + installed_desc, ) - except EnvironmentError as error: - show_traceback = (self.verbosity >= 1) + except OSError as error: + show_traceback = self.verbosity >= 1 - message = create_env_error_message( - error, show_traceback, options.use_user_site, + message = create_os_error_message( + error, + show_traceback, + options.use_user_site, ) logger.error(message, exc_info=show_traceback) return ERROR if options.target_dir: + assert target_temp_dir self._handle_target_dir( options.target_dir, target_temp_dir, options.upgrade ) - + if options.root_user_action == "warn": + warn_if_run_as_root() return SUCCESS - def _handle_target_dir(self, target_dir, target_temp_dir, upgrade): + def _handle_target_dir( + self, target_dir: str, target_temp_dir: TempDirectory, upgrade: bool + ) -> None: ensure_dir(target_dir) # Checking both purelib and platlib directories for installed # packages to be moved to target directory lib_dir_list = [] - with target_temp_dir: - # Checking both purelib and platlib directories for installed - # packages to be moved to target directory - scheme = distutils_scheme('', home=target_temp_dir.path) - purelib_dir = scheme['purelib'] - platlib_dir = scheme['platlib'] - data_dir = scheme['data'] - - if os.path.exists(purelib_dir): - lib_dir_list.append(purelib_dir) - if os.path.exists(platlib_dir) and platlib_dir != purelib_dir: - lib_dir_list.append(platlib_dir) - if os.path.exists(data_dir): - lib_dir_list.append(data_dir) - - for lib_dir in lib_dir_list: - for item in os.listdir(lib_dir): - if lib_dir == data_dir: - ddir = os.path.join(data_dir, item) - if any(s.startswith(ddir) for s in lib_dir_list[:-1]): - continue - target_item_dir = os.path.join(target_dir, item) - if os.path.exists(target_item_dir): - if not upgrade: - logger.warning( - 'Target directory %s already exists. Specify ' - '--upgrade to force replacement.', - target_item_dir - ) - continue - if os.path.islink(target_item_dir): - logger.warning( - 'Target directory %s already exists and is ' - 'a link. pip will not automatically replace ' - 'links, please remove if replacement is ' - 'desired.', - target_item_dir - ) - continue - if os.path.isdir(target_item_dir): - shutil.rmtree(target_item_dir) - else: - os.remove(target_item_dir) - - shutil.move( - os.path.join(lib_dir, item), - target_item_dir - ) - - def _warn_about_conflicts(self, to_install): + # Checking both purelib and platlib directories for installed + # packages to be moved to target directory + scheme = get_scheme("", home=target_temp_dir.path) + purelib_dir = scheme.purelib + platlib_dir = scheme.platlib + data_dir = scheme.data + + if os.path.exists(purelib_dir): + lib_dir_list.append(purelib_dir) + if os.path.exists(platlib_dir) and platlib_dir != purelib_dir: + lib_dir_list.append(platlib_dir) + if os.path.exists(data_dir): + lib_dir_list.append(data_dir) + + for lib_dir in lib_dir_list: + for item in os.listdir(lib_dir): + if lib_dir == data_dir: + ddir = os.path.join(data_dir, item) + if any(s.startswith(ddir) for s in lib_dir_list[:-1]): + continue + target_item_dir = os.path.join(target_dir, item) + if os.path.exists(target_item_dir): + if not upgrade: + logger.warning( + "Target directory %s already exists. Specify " + "--upgrade to force replacement.", + target_item_dir, + ) + continue + if os.path.islink(target_item_dir): + logger.warning( + "Target directory %s already exists and is " + "a link. pip will not automatically replace " + "links, please remove if replacement is " + "desired.", + target_item_dir, + ) + continue + if os.path.isdir(target_item_dir): + shutil.rmtree(target_item_dir) + else: + os.remove(target_item_dir) + + shutil.move(os.path.join(lib_dir, item), target_item_dir) + + def _determine_conflicts( + self, to_install: List[InstallRequirement] + ) -> Optional[ConflictDetails]: try: - package_set, _dep_info = check_install_conflicts(to_install) + return check_install_conflicts(to_install) except Exception: - logger.error("Error checking for conflicts.", exc_info=True) + logger.exception( + "Error while checking for conflicts. Please file an issue on " + "pip's issue tracker: https://github.com/pypa/pip/issues/new" + ) + return None + + def _warn_about_conflicts( + self, conflict_details: ConflictDetails, resolver_variant: str + ) -> None: + package_set, (missing, conflicting) = conflict_details + if not missing and not conflicting: return - missing, conflicting = _dep_info - # NOTE: There is some duplication here from pip check + parts: List[str] = [] + if resolver_variant == "legacy": + parts.append( + "pip's legacy dependency resolver does not consider dependency " + "conflicts when selecting packages. This behaviour is the " + "source of the following dependency conflicts." + ) + else: + assert resolver_variant == "resolvelib" + parts.append( + "pip's dependency resolver does not currently take into account " + "all the packages that are installed. This behaviour is the " + "source of the following dependency conflicts." + ) + + # NOTE: There is some duplication here, with commands/check.py for project_name in missing: version = package_set[project_name][0] for dependency in missing[project_name]: - logger.critical( - "%s %s requires %s, which is not installed.", - project_name, version, dependency[1], + message = ( + f"{project_name} {version} requires {dependency[1]}, " + "which is not installed." ) + parts.append(message) for project_name in conflicting: version = package_set[project_name][0] for dep_name, dep_version, req in conflicting[project_name]: - logger.critical( - "%s %s has requirement %s, but you'll have %s %s which is " - "incompatible.", - project_name, version, req, dep_name, dep_version, + message = ( + "{name} {version} requires {requirement}, but {you} have " + "{dep_name} {dep_version} which is incompatible." + ).format( + name=project_name, + version=version, + requirement=req, + dep_name=dep_name, + dep_version=dep_version, + you=("you" if resolver_variant == "resolvelib" else "you'll"), ) + parts.append(message) + + logger.critical("\n".join(parts)) + + +def get_lib_location_guesses( + user: bool = False, + home: Optional[str] = None, + root: Optional[str] = None, + isolated: bool = False, + prefix: Optional[str] = None, +) -> List[str]: + scheme = get_scheme( + "", + user=user, + home=home, + root=root, + isolated=isolated, + prefix=prefix, + ) + return [scheme.purelib, scheme.platlib] -def get_lib_location_guesses(*args, **kwargs): - scheme = distutils_scheme('', *args, **kwargs) - return [scheme['purelib'], scheme['platlib']] - - -def site_packages_writable(**kwargs): +def site_packages_writable(root: Optional[str], isolated: bool) -> bool: return all( - test_writable_dir(d) for d in set(get_lib_location_guesses(**kwargs)) + test_writable_dir(d) + for d in set(get_lib_location_guesses(root=root, isolated=isolated)) ) def decide_user_install( - use_user_site, # type: Optional[bool] - prefix_path=None, # type: Optional[str] - target_dir=None, # type: Optional[str] - root_path=None, # type: Optional[str] - isolated_mode=False, # type: bool -): - # type: (...) -> bool + use_user_site: Optional[bool], + prefix_path: Optional[str] = None, + target_dir: Optional[str] = None, + root_path: Optional[str] = None, + isolated_mode: bool = False, +) -> bool: """Determine whether to do a user install based on the input options. If use_user_site is False, no additional checks are done. @@ -600,70 +719,24 @@ def decide_user_install( logger.debug("Non-user install because site-packages writeable") return False - logger.info("Defaulting to user installation because normal site-packages " - "is not writeable") - return True - - -def warn_deprecated_install_options(requirements, options): - # type: (List[InstallRequirement], Optional[List[str]]) -> None - """If any location-changing --install-option arguments were passed for - requirements or on the command-line, then show a deprecation warning. - """ - def format_options(option_names): - # type: (Iterable[str]) -> List[str] - return ["--{}".format(name.replace("_", "-")) for name in option_names] - - offenders = [] - - for requirement in requirements: - install_options = requirement.install_options - location_options = parse_distutils_args(install_options) - if location_options: - offenders.append( - "{!r} from {}".format( - format_options(location_options.keys()), requirement - ) - ) - - if options: - location_options = parse_distutils_args(options) - if location_options: - offenders.append( - "{!r} from command line".format( - format_options(location_options.keys()) - ) - ) - - if not offenders: - return - - deprecated( - reason=( - "Location-changing options found in --install-option: {}. " - "This configuration may cause unexpected behavior and is " - "unsupported.".format( - "; ".join(offenders) - ) - ), - replacement=( - "using pip-level options like --user, --prefix, --root, and " - "--target" - ), - gone_in="20.2", - issue=7309, + logger.info( + "Defaulting to user installation because normal site-packages " + "is not writeable" ) + return True -def create_env_error_message(error, show_traceback, using_user_site): - """Format an error message for an EnvironmentError +def create_os_error_message( + error: OSError, show_traceback: bool, using_user_site: bool +) -> str: + """Format an error message for an OSError It may occur anytime during the execution of the install command. """ parts = [] # Mention the error if we are not going to show a traceback - parts.append("Could not install packages due to an EnvironmentError") + parts.append("Could not install packages due to an OSError") if not show_traceback: parts.append(": ") parts.append(str(error)) @@ -679,13 +752,32 @@ def create_env_error_message(error, show_traceback, using_user_site): user_option_part = "Consider using the `--user` option" permissions_part = "Check the permissions" - if not using_user_site: - parts.extend([ - user_option_part, " or ", - permissions_part.lower(), - ]) + if not running_under_virtualenv() and not using_user_site: + parts.extend( + [ + user_option_part, + " or ", + permissions_part.lower(), + ] + ) else: parts.append(permissions_part) parts.append(".\n") + # Suggest the user to enable Long Paths if path length is + # more than 260 + if ( + WINDOWS + and error.errno == errno.ENOENT + and error.filename + and len(error.filename) > 260 + ): + parts.append( + "HINT: This error might have occurred since " + "this system does not have Windows Long Path " + "support enabled. You can find information on " + "how to enable this at " + "https://pip.pypa.io/warnings/enable-long-paths\n" + ) + return "".join(parts).strip() + "\n" diff --git a/src/fetchcode/vcs/pip/_internal/commands/list.py b/src/fetchcode/vcs/pip/_internal/commands/list.py index fe1ee21e..82fc46a1 100644 --- a/src/fetchcode/vcs/pip/_internal/commands/list.py +++ b/src/fetchcode/vcs/pip/_internal/commands/list.py @@ -1,26 +1,36 @@ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import - import json import logging +from optparse import Values +from typing import TYPE_CHECKING, Generator, List, Optional, Sequence, Tuple, cast + +from pip._vendor.packaging.utils import canonicalize_name +from pip._vendor.packaging.version import Version + +from pip._internal.cli import cmdoptions +from pip._internal.cli.index_command import IndexGroupCommand +from pip._internal.cli.status_codes import SUCCESS +from pip._internal.exceptions import CommandError +from pip._internal.metadata import BaseDistribution, get_environment +from pip._internal.models.selection_prefs import SelectionPreferences +from pip._internal.utils.compat import stdlib_pkgs +from pip._internal.utils.misc import tabulate, write_output + +if TYPE_CHECKING: + from pip._internal.index.package_finder import PackageFinder + from pip._internal.network.session import PipSession + + class _DistWithLatestInfo(BaseDistribution): + """Give the distribution object a couple of extra fields. + + These will be populated during ``get_outdated()``. This is dirty but + makes the rest of the code much cleaner. + """ + + latest_version: Version + latest_filetype: str + + _ProcessedDists = Sequence[_DistWithLatestInfo] -from fetchcode.vcs.pip._vendor import six - -from fetchcode.vcs.pip._internal.cli import cmdoptions -from fetchcode.vcs.pip._internal.cli.req_command import IndexGroupCommand -from fetchcode.vcs.pip._internal.exceptions import CommandError -from fetchcode.vcs.pip._internal.index.package_finder import PackageFinder -from fetchcode.vcs.pip._internal.models.selection_prefs import SelectionPreferences -from fetchcode.vcs.pip._internal.self_outdated_check import make_link_collector -from fetchcode.vcs.pip._internal.utils.misc import ( - dist_is_editable, - get_installed_distributions, - tabulate, - write_output, -) -from fetchcode.vcs.pip._internal.utils.packaging import get_installer logger = logging.getLogger(__name__) @@ -32,94 +42,113 @@ class ListCommand(IndexGroupCommand): Packages are listed in a case-insensitive sorted order. """ + ignore_require_venv = True usage = """ %prog [options]""" - def __init__(self, *args, **kw): - super(ListCommand, self).__init__(*args, **kw) - - cmd_opts = self.cmd_opts - - cmd_opts.add_option( - '-o', '--outdated', - action='store_true', + def add_options(self) -> None: + self.cmd_opts.add_option( + "-o", + "--outdated", + action="store_true", default=False, - help='List outdated packages') - cmd_opts.add_option( - '-u', '--uptodate', - action='store_true', + help="List outdated packages", + ) + self.cmd_opts.add_option( + "-u", + "--uptodate", + action="store_true", default=False, - help='List uptodate packages') - cmd_opts.add_option( - '-e', '--editable', - action='store_true', + help="List uptodate packages", + ) + self.cmd_opts.add_option( + "-e", + "--editable", + action="store_true", default=False, - help='List editable projects.') - cmd_opts.add_option( - '-l', '--local', - action='store_true', + help="List editable projects.", + ) + self.cmd_opts.add_option( + "-l", + "--local", + action="store_true", default=False, - help=('If in a virtualenv that has global access, do not list ' - 'globally-installed packages.'), + help=( + "If in a virtualenv that has global access, do not list " + "globally-installed packages." + ), ) self.cmd_opts.add_option( - '--user', - dest='user', - action='store_true', + "--user", + dest="user", + action="store_true", default=False, - help='Only output packages installed in user-site.') - cmd_opts.add_option(cmdoptions.list_path()) - cmd_opts.add_option( - '--pre', - action='store_true', + help="Only output packages installed in user-site.", + ) + self.cmd_opts.add_option(cmdoptions.list_path()) + self.cmd_opts.add_option( + "--pre", + action="store_true", default=False, - help=("Include pre-release and development versions. By default, " - "pip only finds stable versions."), + help=( + "Include pre-release and development versions. By default, " + "pip only finds stable versions." + ), ) - cmd_opts.add_option( - '--format', - action='store', - dest='list_format', + self.cmd_opts.add_option( + "--format", + action="store", + dest="list_format", default="columns", - choices=('columns', 'freeze', 'json'), - help="Select the output format among: columns (default), freeze, " - "or json", + choices=("columns", "freeze", "json"), + help=( + "Select the output format among: columns (default), freeze, or json. " + "The 'freeze' format cannot be used with the --outdated option." + ), ) - cmd_opts.add_option( - '--not-required', - action='store_true', - dest='not_required', - help="List packages that are not dependencies of " - "installed packages.", + self.cmd_opts.add_option( + "--not-required", + action="store_true", + dest="not_required", + help="List packages that are not dependencies of installed packages.", ) - cmd_opts.add_option( - '--exclude-editable', - action='store_false', - dest='include_editable', - help='Exclude editable package from output.', + self.cmd_opts.add_option( + "--exclude-editable", + action="store_false", + dest="include_editable", + help="Exclude editable package from output.", ) - cmd_opts.add_option( - '--include-editable', - action='store_true', - dest='include_editable', - help='Include editable package from output.', + self.cmd_opts.add_option( + "--include-editable", + action="store_true", + dest="include_editable", + help="Include editable package from output.", default=True, ) - index_opts = cmdoptions.make_option_group( - cmdoptions.index_group, self.parser - ) + self.cmd_opts.add_option(cmdoptions.list_exclude()) + index_opts = cmdoptions.make_option_group(cmdoptions.index_group, self.parser) self.parser.insert_option_group(0, index_opts) - self.parser.insert_option_group(0, cmd_opts) + self.parser.insert_option_group(0, self.cmd_opts) - def _build_package_finder(self, options, session): + def handle_pip_version_check(self, options: Values) -> None: + if options.outdated or options.uptodate: + super().handle_pip_version_check(options) + + def _build_package_finder( + self, options: Values, session: "PipSession" + ) -> "PackageFinder": """ Create a package finder appropriate to this list command. """ - link_collector = make_link_collector(session, options=options) + # Lazy import the heavy index modules as most list invocations won't need 'em. + from pip._internal.index.collector import LinkCollector + from pip._internal.index.package_finder import PackageFinder + + link_collector = LinkCollector.create(session, options=options) # Pass allow_yanked=False to ignore yanked versions. selection_prefs = SelectionPreferences( @@ -132,20 +161,31 @@ def _build_package_finder(self, options, session): selection_prefs=selection_prefs, ) - def run(self, options, args): + def run(self, options: Values, args: List[str]) -> int: if options.outdated and options.uptodate: + raise CommandError("Options --outdated and --uptodate cannot be combined.") + + if options.outdated and options.list_format == "freeze": raise CommandError( - "Options --outdated and --uptodate cannot be combined.") + "List format 'freeze' cannot be used with the --outdated option." + ) cmdoptions.check_list_path_option(options) - packages = get_installed_distributions( - local_only=options.local, - user_only=options.user, - editables_only=options.editable, - include_editables=options.include_editable, - paths=options.path, - ) + skip = set(stdlib_pkgs) + if options.excludes: + skip.update(canonicalize_name(n) for n in options.excludes) + + packages: "_ProcessedDists" = [ + cast("_DistWithLatestInfo", d) + for d in get_environment(options.path).iter_installed_distributions( + local_only=options.local, + user_only=options.user, + editables_only=options.editable, + include_editables=options.include_editable, + skip=skip, + ) + ] # get_not_required must be called firstly in order to find and # filter out all dependencies correctly. Otherwise a package @@ -160,39 +200,60 @@ def run(self, options, args): packages = self.get_uptodate(packages, options) self.output_package_listing(packages, options) + return SUCCESS - def get_outdated(self, packages, options): + def get_outdated( + self, packages: "_ProcessedDists", options: Values + ) -> "_ProcessedDists": return [ - dist for dist in self.iter_packages_latest_infos(packages, options) - if dist.latest_version > dist.parsed_version + dist + for dist in self.iter_packages_latest_infos(packages, options) + if dist.latest_version > dist.version ] - def get_uptodate(self, packages, options): + def get_uptodate( + self, packages: "_ProcessedDists", options: Values + ) -> "_ProcessedDists": return [ - dist for dist in self.iter_packages_latest_infos(packages, options) - if dist.latest_version == dist.parsed_version + dist + for dist in self.iter_packages_latest_infos(packages, options) + if dist.latest_version == dist.version ] - def get_not_required(self, packages, options): - dep_keys = set() - for dist in packages: - dep_keys.update(requirement.key for requirement in dist.requires()) - return {pkg for pkg in packages if pkg.key not in dep_keys} + def get_not_required( + self, packages: "_ProcessedDists", options: Values + ) -> "_ProcessedDists": + dep_keys = { + canonicalize_name(dep.name) + for dist in packages + for dep in (dist.iter_dependencies() or ()) + } - def iter_packages_latest_infos(self, packages, options): + # Create a set to remove duplicate packages, and cast it to a list + # to keep the return type consistent with get_outdated and + # get_uptodate + return list({pkg for pkg in packages if pkg.canonical_name not in dep_keys}) + + def iter_packages_latest_infos( + self, packages: "_ProcessedDists", options: Values + ) -> Generator["_DistWithLatestInfo", None, None]: with self._build_session(options) as session: finder = self._build_package_finder(options, session) - def latest_info(dist): - typ = 'unknown' - all_candidates = finder.find_all_candidates(dist.key) + def latest_info( + dist: "_DistWithLatestInfo", + ) -> Optional["_DistWithLatestInfo"]: + all_candidates = finder.find_all_candidates(dist.canonical_name) if not options.pre: # Remove prereleases - all_candidates = [candidate for candidate in all_candidates - if not candidate.version.is_prerelease] + all_candidates = [ + candidate + for candidate in all_candidates + if not candidate.version.is_prerelease + ] evaluator = finder.make_candidate_evaluator( - project_name=dist.project_name, + project_name=dist.canonical_name, ) best_candidate = evaluator.sort_best_candidate(all_candidates) if best_candidate is None: @@ -200,10 +261,9 @@ def latest_info(dist): remote_version = best_candidate.version if best_candidate.link.is_wheel: - typ = 'wheel' + typ = "wheel" else: - typ = 'sdist' - # This is dirty but makes the rest of the code much cleaner + typ = "sdist" dist.latest_version = remote_version dist.latest_filetype = typ return dist @@ -212,25 +272,30 @@ def latest_info(dist): if dist is not None: yield dist - def output_package_listing(self, packages, options): + def output_package_listing( + self, packages: "_ProcessedDists", options: Values + ) -> None: packages = sorted( packages, - key=lambda dist: dist.project_name.lower(), + key=lambda dist: dist.canonical_name, ) - if options.list_format == 'columns' and packages: + if options.list_format == "columns" and packages: data, header = format_for_columns(packages, options) self.output_package_listing_columns(data, header) - elif options.list_format == 'freeze': + elif options.list_format == "freeze": for dist in packages: if options.verbose >= 1: - write_output("%s==%s (%s)", dist.project_name, - dist.version, dist.location) + write_output( + "%s==%s (%s)", dist.raw_name, dist.version, dist.location + ) else: - write_output("%s==%s", dist.project_name, dist.version) - elif options.list_format == 'json': + write_output("%s==%s", dist.raw_name, dist.version) + elif options.list_format == "json": write_output(format_for_json(packages, options)) - def output_package_listing_columns(self, data, header): + def output_package_listing_columns( + self, data: List[List[str]], header: List[str] + ) -> None: # insert the header first: we need to know the size of column names if len(data) > 0: data.insert(0, header) @@ -239,61 +304,72 @@ def output_package_listing_columns(self, data, header): # Create and add a separator. if len(data) > 0: - pkg_strings.insert(1, " ".join(map(lambda x: '-' * x, sizes))) + pkg_strings.insert(1, " ".join("-" * x for x in sizes)) for val in pkg_strings: write_output(val) -def format_for_columns(pkgs, options): +def format_for_columns( + pkgs: "_ProcessedDists", options: Values +) -> Tuple[List[List[str]], List[str]]: """ Convert the package data into something usable by output_package_listing_columns. """ + header = ["Package", "Version"] + running_outdated = options.outdated - # Adjust the header for the `pip list --outdated` case. if running_outdated: - header = ["Package", "Version", "Latest", "Type"] - else: - header = ["Package", "Version"] + header.extend(["Latest", "Type"]) - data = [] - if options.verbose >= 1 or any(dist_is_editable(x) for x in pkgs): + has_editables = any(x.editable for x in pkgs) + if has_editables: + header.append("Editable project location") + + if options.verbose >= 1: header.append("Location") if options.verbose >= 1: header.append("Installer") + data = [] for proj in pkgs: # if we're working on the 'outdated' list, separate out the # latest_version and type - row = [proj.project_name, proj.version] + row = [proj.raw_name, proj.raw_version] if running_outdated: - row.append(proj.latest_version) + row.append(str(proj.latest_version)) row.append(proj.latest_filetype) - if options.verbose >= 1 or dist_is_editable(proj): - row.append(proj.location) + if has_editables: + row.append(proj.editable_project_location or "") + + if options.verbose >= 1: + row.append(proj.location or "") if options.verbose >= 1: - row.append(get_installer(proj)) + row.append(proj.installer) data.append(row) return data, header -def format_for_json(packages, options): +def format_for_json(packages: "_ProcessedDists", options: Values) -> str: data = [] for dist in packages: info = { - 'name': dist.project_name, - 'version': six.text_type(dist.version), + "name": dist.raw_name, + "version": str(dist.version), } if options.verbose >= 1: - info['location'] = dist.location - info['installer'] = get_installer(dist) + info["location"] = dist.location or "" + info["installer"] = dist.installer if options.outdated: - info['latest_version'] = six.text_type(dist.latest_version) - info['latest_filetype'] = dist.latest_filetype + info["latest_version"] = str(dist.latest_version) + info["latest_filetype"] = dist.latest_filetype + editable_project_location = dist.editable_project_location + if editable_project_location: + info["editable_project_location"] = editable_project_location data.append(info) return json.dumps(data) diff --git a/src/fetchcode/vcs/pip/_internal/commands/search.py b/src/fetchcode/vcs/pip/_internal/commands/search.py index 5ccbcd13..e0d329d5 100644 --- a/src/fetchcode/vcs/pip/_internal/commands/search.py +++ b/src/fetchcode/vcs/pip/_internal/commands/search.py @@ -1,28 +1,31 @@ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import - import logging +import shutil import sys import textwrap +import xmlrpc.client from collections import OrderedDict +from optparse import Values +from typing import TYPE_CHECKING, Dict, List, Optional, TypedDict + +from pip._vendor.packaging.version import parse as parse_version + +from pip._internal.cli.base_command import Command +from pip._internal.cli.req_command import SessionCommandMixin +from pip._internal.cli.status_codes import NO_MATCHES_FOUND, SUCCESS +from pip._internal.exceptions import CommandError +from pip._internal.metadata import get_default_environment +from pip._internal.models.index import PyPI +from pip._internal.network.xmlrpc import PipXmlrpcTransport +from pip._internal.utils.logging import indent_log +from pip._internal.utils.misc import write_output + +if TYPE_CHECKING: + + class TransformedHit(TypedDict): + name: str + summary: str + versions: List[str] -from fetchcode.vcs.pip._vendor import pkg_resources -from fetchcode.vcs.pip._vendor.packaging.version import parse as parse_version -# NOTE: XMLRPC Client is not annotated in typeshed as on 2017-07-17, which is -# why we ignore the type on this import -from fetchcode.vcs.pip._vendor.six.moves import xmlrpc_client # type: ignore - -from fetchcode.vcs.pip._internal.cli.base_command import Command -from fetchcode.vcs.pip._internal.cli.req_command import SessionCommandMixin -from fetchcode.vcs.pip._internal.cli.status_codes import NO_MATCHES_FOUND, SUCCESS -from fetchcode.vcs.pip._internal.exceptions import CommandError -from fetchcode.vcs.pip._internal.models.index import PyPI -from fetchcode.vcs.pip._internal.network.xmlrpc import PipXmlrpcTransport -from fetchcode.vcs.pip._internal.utils.compat import get_terminal_size -from fetchcode.vcs.pip._internal.utils.logging import indent_log -from fetchcode.vcs.pip._internal.utils.misc import write_output logger = logging.getLogger(__name__) @@ -34,113 +37,136 @@ class SearchCommand(Command, SessionCommandMixin): %prog [options] """ ignore_require_venv = True - def __init__(self, *args, **kw): - super(SearchCommand, self).__init__(*args, **kw) + def add_options(self) -> None: self.cmd_opts.add_option( - '-i', '--index', - dest='index', - metavar='URL', + "-i", + "--index", + dest="index", + metavar="URL", default=PyPI.pypi_url, - help='Base URL of Python Package Index (default %default)') + help="Base URL of Python Package Index (default %default)", + ) self.parser.insert_option_group(0, self.cmd_opts) - def run(self, options, args): + def run(self, options: Values, args: List[str]) -> int: if not args: - raise CommandError('Missing required argument (search query).') + raise CommandError("Missing required argument (search query).") query = args pypi_hits = self.search(query, options) hits = transform_hits(pypi_hits) terminal_width = None if sys.stdout.isatty(): - terminal_width = get_terminal_size()[0] + terminal_width = shutil.get_terminal_size()[0] print_results(hits, terminal_width=terminal_width) if pypi_hits: return SUCCESS return NO_MATCHES_FOUND - def search(self, query, options): + def search(self, query: List[str], options: Values) -> List[Dict[str, str]]: index_url = options.index session = self.get_default_session(options) transport = PipXmlrpcTransport(index_url, session) - pypi = xmlrpc_client.ServerProxy(index_url, transport) - hits = pypi.search({'name': query, 'summary': query}, 'or') + pypi = xmlrpc.client.ServerProxy(index_url, transport) + try: + hits = pypi.search({"name": query, "summary": query}, "or") + except xmlrpc.client.Fault as fault: + message = ( + f"XMLRPC request failed [code: {fault.faultCode}]\n{fault.faultString}" + ) + raise CommandError(message) + assert isinstance(hits, list) return hits -def transform_hits(hits): +def transform_hits(hits: List[Dict[str, str]]) -> List["TransformedHit"]: """ The list from pypi is really a list of versions. We want a list of packages with the list of versions stored inline. This converts the list from pypi into one we can use. """ - packages = OrderedDict() + packages: Dict[str, "TransformedHit"] = OrderedDict() for hit in hits: - name = hit['name'] - summary = hit['summary'] - version = hit['version'] + name = hit["name"] + summary = hit["summary"] + version = hit["version"] if name not in packages.keys(): packages[name] = { - 'name': name, - 'summary': summary, - 'versions': [version], + "name": name, + "summary": summary, + "versions": [version], } else: - packages[name]['versions'].append(version) + packages[name]["versions"].append(version) # if this is the highest version, replace summary and score - if version == highest_version(packages[name]['versions']): - packages[name]['summary'] = summary + if version == highest_version(packages[name]["versions"]): + packages[name]["summary"] = summary return list(packages.values()) -def print_results(hits, name_column_width=None, terminal_width=None): +def print_dist_installation_info(name: str, latest: str) -> None: + env = get_default_environment() + dist = env.get_distribution(name) + if dist is not None: + with indent_log(): + if dist.version == latest: + write_output("INSTALLED: %s (latest)", dist.version) + else: + write_output("INSTALLED: %s", dist.version) + if parse_version(latest).pre: + write_output( + "LATEST: %s (pre-release; install" + " with `pip install --pre`)", + latest, + ) + else: + write_output("LATEST: %s", latest) + + +def print_results( + hits: List["TransformedHit"], + name_column_width: Optional[int] = None, + terminal_width: Optional[int] = None, +) -> None: if not hits: return if name_column_width is None: - name_column_width = max([ - len(hit['name']) + len(highest_version(hit.get('versions', ['-']))) - for hit in hits - ]) + 4 + name_column_width = ( + max( + [ + len(hit["name"]) + len(highest_version(hit.get("versions", ["-"]))) + for hit in hits + ] + ) + + 4 + ) - installed_packages = [p.project_name for p in pkg_resources.working_set] for hit in hits: - name = hit['name'] - summary = hit['summary'] or '' - latest = highest_version(hit.get('versions', ['-'])) + name = hit["name"] + summary = hit["summary"] or "" + latest = highest_version(hit.get("versions", ["-"])) if terminal_width is not None: target_width = terminal_width - name_column_width - 5 if target_width > 10: # wrap and indent summary to fit terminal - summary = textwrap.wrap(summary, target_width) - summary = ('\n' + ' ' * (name_column_width + 3)).join(summary) + summary_lines = textwrap.wrap(summary, target_width) + summary = ("\n" + " " * (name_column_width + 3)).join(summary_lines) - line = '{name_latest:{name_column_width}} - {summary}'.format( - name_latest='{name} ({latest})'.format(**locals()), - **locals()) + name_latest = f"{name} ({latest})" + line = f"{name_latest:{name_column_width}} - {summary}" try: write_output(line) - if name in installed_packages: - dist = pkg_resources.get_distribution(name) - with indent_log(): - if dist.version == latest: - write_output('INSTALLED: %s (latest)', dist.version) - else: - write_output('INSTALLED: %s', dist.version) - if parse_version(latest).pre: - write_output('LATEST: %s (pre-release; install' - ' with "pip install --pre")', latest) - else: - write_output('LATEST: %s', latest) + print_dist_installation_info(name, latest) except UnicodeEncodeError: pass -def highest_version(versions): +def highest_version(versions: List[str]) -> str: return max(versions, key=parse_version) diff --git a/src/fetchcode/vcs/pip/_internal/commands/show.py b/src/fetchcode/vcs/pip/_internal/commands/show.py index 2ccfbe9b..c54d548f 100644 --- a/src/fetchcode/vcs/pip/_internal/commands/show.py +++ b/src/fetchcode/vcs/pip/_internal/commands/show.py @@ -1,18 +1,14 @@ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import - import logging -import os -from email.parser import FeedParser +from optparse import Values +from typing import Generator, Iterable, Iterator, List, NamedTuple, Optional -from fetchcode.vcs.pip._vendor import pkg_resources -from fetchcode.vcs.pip._vendor.packaging.utils import canonicalize_name +from pip._vendor.packaging.requirements import InvalidRequirement +from pip._vendor.packaging.utils import canonicalize_name -from fetchcode.vcs.pip._internal.cli.base_command import Command -from fetchcode.vcs.pip._internal.cli.status_codes import ERROR, SUCCESS -from fetchcode.vcs.pip._internal.utils.misc import write_output +from pip._internal.cli.base_command import Command +from pip._internal.cli.status_codes import ERROR, SUCCESS +from pip._internal.metadata import BaseDistribution, get_default_environment +from pip._internal.utils.misc import write_output logger = logging.getLogger(__name__) @@ -28,119 +24,153 @@ class ShowCommand(Command): %prog [options] ...""" ignore_require_venv = True - def __init__(self, *args, **kw): - super(ShowCommand, self).__init__(*args, **kw) + def add_options(self) -> None: self.cmd_opts.add_option( - '-f', '--files', - dest='files', - action='store_true', + "-f", + "--files", + dest="files", + action="store_true", default=False, - help='Show the full list of installed files for each package.') + help="Show the full list of installed files for each package.", + ) self.parser.insert_option_group(0, self.cmd_opts) - def run(self, options, args): + def run(self, options: Values, args: List[str]) -> int: if not args: - logger.warning('ERROR: Please provide a package name or names.') + logger.warning("ERROR: Please provide a package name or names.") return ERROR query = args results = search_packages_info(query) if not print_results( - results, list_files=options.files, verbose=options.verbose): + results, list_files=options.files, verbose=options.verbose + ): return ERROR return SUCCESS -def search_packages_info(query): +class _PackageInfo(NamedTuple): + name: str + version: str + location: str + editable_project_location: Optional[str] + requires: List[str] + required_by: List[str] + installer: str + metadata_version: str + classifiers: List[str] + summary: str + homepage: str + project_urls: List[str] + author: str + author_email: str + license: str + entry_points: List[str] + files: Optional[List[str]] + + +def search_packages_info(query: List[str]) -> Generator[_PackageInfo, None, None]: """ Gather details from installed distributions. Print distribution name, version, location, and installed files. Installed files requires a pip generated 'installed-files.txt' in the distributions '.egg-info' directory. """ - installed = {} - for p in pkg_resources.working_set: - installed[canonicalize_name(p.project_name)] = p + env = get_default_environment() + installed = {dist.canonical_name: dist for dist in env.iter_all_distributions()} query_names = [canonicalize_name(name) for name in query] missing = sorted( [name for name, pkg in zip(query, query_names) if pkg not in installed] ) if missing: - logger.warning('Package(s) not found: %s', ', '.join(missing)) - - def get_requiring_packages(package_name): - canonical_name = canonicalize_name(package_name) - return [ - pkg.project_name for pkg in pkg_resources.working_set - if canonical_name in - [canonicalize_name(required.name) for required in - pkg.requires()] - ] - - for dist in [installed[pkg] for pkg in query_names if pkg in installed]: - package = { - 'name': dist.project_name, - 'version': dist.version, - 'location': dist.location, - 'requires': [dep.project_name for dep in dist.requires()], - 'required_by': get_requiring_packages(dist.project_name) - } - file_list = None - metadata = None - if isinstance(dist, pkg_resources.DistInfoDistribution): - # RECORDs should be part of .dist-info metadatas - if dist.has_metadata('RECORD'): - lines = dist.get_metadata_lines('RECORD') - paths = [l.split(',')[0] for l in lines] - paths = [os.path.join(dist.location, p) for p in paths] - file_list = [os.path.relpath(p, dist.location) for p in paths] - - if dist.has_metadata('METADATA'): - metadata = dist.get_metadata('METADATA') + logger.warning("Package(s) not found: %s", ", ".join(missing)) + + def _get_requiring_packages(current_dist: BaseDistribution) -> Iterator[str]: + return ( + dist.metadata["Name"] or "UNKNOWN" + for dist in installed.values() + if current_dist.canonical_name + in {canonicalize_name(d.name) for d in dist.iter_dependencies()} + ) + + for query_name in query_names: + try: + dist = installed[query_name] + except KeyError: + continue + + try: + requires = sorted( + # Avoid duplicates in requirements (e.g. due to environment markers). + {req.name for req in dist.iter_dependencies()}, + key=str.lower, + ) + except InvalidRequirement: + requires = sorted(dist.iter_raw_dependencies(), key=str.lower) + + try: + required_by = sorted(_get_requiring_packages(dist), key=str.lower) + except InvalidRequirement: + required_by = ["#N/A"] + + try: + entry_points_text = dist.read_text("entry_points.txt") + entry_points = entry_points_text.splitlines(keepends=False) + except FileNotFoundError: + entry_points = [] + + files_iter = dist.iter_declared_entries() + if files_iter is None: + files: Optional[List[str]] = None else: - # Otherwise use pip's log for .egg-info's - if dist.has_metadata('installed-files.txt'): - paths = dist.get_metadata_lines('installed-files.txt') - paths = [os.path.join(dist.egg_info, p) for p in paths] - file_list = [os.path.relpath(p, dist.location) for p in paths] - - if dist.has_metadata('PKG-INFO'): - metadata = dist.get_metadata('PKG-INFO') - - if dist.has_metadata('entry_points.txt'): - entry_points = dist.get_metadata_lines('entry_points.txt') - package['entry_points'] = entry_points - - if dist.has_metadata('INSTALLER'): - for line in dist.get_metadata_lines('INSTALLER'): - if line.strip(): - package['installer'] = line.strip() + files = sorted(files_iter) + + metadata = dist.metadata + + project_urls = metadata.get_all("Project-URL", []) + homepage = metadata.get("Home-page", "") + if not homepage: + # It's common that there is a "homepage" Project-URL, but Home-page + # remains unset (especially as PEP 621 doesn't surface the field). + # + # This logic was taken from PyPI's codebase. + for url in project_urls: + url_label, url = url.split(",", maxsplit=1) + normalized_label = ( + url_label.casefold().replace("-", "").replace("_", "").strip() + ) + if normalized_label == "homepage": + homepage = url.strip() break - # @todo: Should pkg_resources.Distribution have a - # `get_pkg_info` method? - feed_parser = FeedParser() - feed_parser.feed(metadata) - pkg_info_dict = feed_parser.close() - for key in ('metadata-version', 'summary', - 'home-page', 'author', 'author-email', 'license'): - package[key] = pkg_info_dict.get(key) - - # It looks like FeedParser cannot deal with repeated headers - classifiers = [] - for line in metadata.splitlines(): - if line.startswith('Classifier: '): - classifiers.append(line[len('Classifier: '):]) - package['classifiers'] = classifiers - - if file_list: - package['files'] = sorted(file_list) - yield package - - -def print_results(distributions, list_files=False, verbose=False): + yield _PackageInfo( + name=dist.raw_name, + version=dist.raw_version, + location=dist.location or "", + editable_project_location=dist.editable_project_location, + requires=requires, + required_by=required_by, + installer=dist.installer, + metadata_version=dist.metadata_version or "", + classifiers=metadata.get_all("Classifier", []), + summary=metadata.get("Summary", ""), + homepage=homepage, + project_urls=project_urls, + author=metadata.get("Author", ""), + author_email=metadata.get("Author-email", ""), + license=metadata.get("License", ""), + entry_points=entry_points, + files=files, + ) + + +def print_results( + distributions: Iterable[_PackageInfo], + list_files: bool, + verbose: bool, +) -> bool: """ Print the information from installed distributions found. """ @@ -150,31 +180,38 @@ def print_results(distributions, list_files=False, verbose=False): if i > 0: write_output("---") - write_output("Name: %s", dist.get('name', '')) - write_output("Version: %s", dist.get('version', '')) - write_output("Summary: %s", dist.get('summary', '')) - write_output("Home-page: %s", dist.get('home-page', '')) - write_output("Author: %s", dist.get('author', '')) - write_output("Author-email: %s", dist.get('author-email', '')) - write_output("License: %s", dist.get('license', '')) - write_output("Location: %s", dist.get('location', '')) - write_output("Requires: %s", ', '.join(dist.get('requires', []))) - write_output("Required-by: %s", ', '.join(dist.get('required_by', []))) + write_output("Name: %s", dist.name) + write_output("Version: %s", dist.version) + write_output("Summary: %s", dist.summary) + write_output("Home-page: %s", dist.homepage) + write_output("Author: %s", dist.author) + write_output("Author-email: %s", dist.author_email) + write_output("License: %s", dist.license) + write_output("Location: %s", dist.location) + if dist.editable_project_location is not None: + write_output( + "Editable project location: %s", dist.editable_project_location + ) + write_output("Requires: %s", ", ".join(dist.requires)) + write_output("Required-by: %s", ", ".join(dist.required_by)) if verbose: - write_output("Metadata-Version: %s", - dist.get('metadata-version', '')) - write_output("Installer: %s", dist.get('installer', '')) + write_output("Metadata-Version: %s", dist.metadata_version) + write_output("Installer: %s", dist.installer) write_output("Classifiers:") - for classifier in dist.get('classifiers', []): + for classifier in dist.classifiers: write_output(" %s", classifier) write_output("Entry-points:") - for entry in dist.get('entry_points', []): + for entry in dist.entry_points: write_output(" %s", entry.strip()) + write_output("Project-URLs:") + for project_url in dist.project_urls: + write_output(" %s", project_url) if list_files: write_output("Files:") - for line in dist.get('files', []): - write_output(" %s", line.strip()) - if "files" not in dist: - write_output("Cannot locate installed-files.txt") + if dist.files is None: + write_output("Cannot locate RECORD or installed-files.txt") + else: + for line in dist.files: + write_output(" %s", line.strip()) return results_printed diff --git a/src/fetchcode/vcs/pip/_internal/commands/uninstall.py b/src/fetchcode/vcs/pip/_internal/commands/uninstall.py index adbd9a97..bc0edeac 100644 --- a/src/fetchcode/vcs/pip/_internal/commands/uninstall.py +++ b/src/fetchcode/vcs/pip/_internal/commands/uninstall.py @@ -1,19 +1,26 @@ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False +import logging +from optparse import Values +from typing import List -from __future__ import absolute_import +from pip._vendor.packaging.utils import canonicalize_name -from fetchcode.vcs.pip._vendor.packaging.utils import canonicalize_name - -from fetchcode.vcs.pip._internal.cli.base_command import Command -from fetchcode.vcs.pip._internal.cli.req_command import SessionCommandMixin -from fetchcode.vcs.pip._internal.exceptions import InstallationError -from fetchcode.vcs.pip._internal.req import parse_requirements -from fetchcode.vcs.pip._internal.req.constructors import ( +from pip._internal.cli import cmdoptions +from pip._internal.cli.base_command import Command +from pip._internal.cli.index_command import SessionCommandMixin +from pip._internal.cli.status_codes import SUCCESS +from pip._internal.exceptions import InstallationError +from pip._internal.req import parse_requirements +from pip._internal.req.constructors import ( install_req_from_line, install_req_from_parsed_requirement, ) -from fetchcode.vcs.pip._internal.utils.misc import protect_pip_from_modification_on_windows +from pip._internal.utils.misc import ( + check_externally_managed, + protect_pip_from_modification_on_windows, + warn_if_run_as_root, +) + +logger = logging.getLogger(__name__) class UninstallCommand(Command, SessionCommandMixin): @@ -31,59 +38,77 @@ class UninstallCommand(Command, SessionCommandMixin): %prog [options] ... %prog [options] -r ...""" - def __init__(self, *args, **kw): - super(UninstallCommand, self).__init__(*args, **kw) + def add_options(self) -> None: self.cmd_opts.add_option( - '-r', '--requirement', - dest='requirements', - action='append', + "-r", + "--requirement", + dest="requirements", + action="append", default=[], - metavar='file', - help='Uninstall all the packages listed in the given requirements ' - 'file. This option can be used multiple times.', + metavar="file", + help=( + "Uninstall all the packages listed in the given requirements " + "file. This option can be used multiple times." + ), ) self.cmd_opts.add_option( - '-y', '--yes', - dest='yes', - action='store_true', - help="Don't ask for confirmation of uninstall deletions.") - + "-y", + "--yes", + dest="yes", + action="store_true", + help="Don't ask for confirmation of uninstall deletions.", + ) + self.cmd_opts.add_option(cmdoptions.root_user_action()) + self.cmd_opts.add_option(cmdoptions.override_externally_managed()) self.parser.insert_option_group(0, self.cmd_opts) - def run(self, options, args): + def run(self, options: Values, args: List[str]) -> int: session = self.get_default_session(options) reqs_to_uninstall = {} for name in args: req = install_req_from_line( - name, isolated=options.isolated_mode, + name, + isolated=options.isolated_mode, ) if req.name: reqs_to_uninstall[canonicalize_name(req.name)] = req + else: + logger.warning( + "Invalid requirement: %r ignored -" + " the uninstall command expects named" + " requirements.", + name, + ) for filename in options.requirements: for parsed_req in parse_requirements( - filename, - options=options, - session=session): + filename, options=options, session=session + ): req = install_req_from_parsed_requirement( - parsed_req, - isolated=options.isolated_mode + parsed_req, isolated=options.isolated_mode ) if req.name: reqs_to_uninstall[canonicalize_name(req.name)] = req if not reqs_to_uninstall: raise InstallationError( - 'You must give at least one requirement to {self.name} (see ' - '"pip help {self.name}")'.format(**locals()) + f"You must give at least one requirement to {self.name} (see " + f'"pip help {self.name}")' ) + if not options.override_externally_managed: + check_externally_managed() + protect_pip_from_modification_on_windows( modifying_pip="pip" in reqs_to_uninstall ) for req in reqs_to_uninstall.values(): uninstall_pathset = req.uninstall( - auto_confirm=options.yes, verbose=self.verbosity > 0, + auto_confirm=options.yes, + verbose=self.verbosity > 0, ) if uninstall_pathset: uninstall_pathset.commit() + if options.root_user_action == "warn": + warn_if_run_as_root() + return SUCCESS diff --git a/src/fetchcode/vcs/pip/_internal/commands/wheel.py b/src/fetchcode/vcs/pip/_internal/commands/wheel.py index cfd5d9a6..278719f4 100644 --- a/src/fetchcode/vcs/pip/_internal/commands/wheel.py +++ b/src/fetchcode/vcs/pip/_internal/commands/wheel.py @@ -1,28 +1,22 @@ -# -*- coding: utf-8 -*- - -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import - import logging import os import shutil - -from fetchcode.vcs.pip._internal.cache import WheelCache -from fetchcode.vcs.pip._internal.cli import cmdoptions -from fetchcode.vcs.pip._internal.cli.req_command import RequirementCommand, with_cleanup -from fetchcode.vcs.pip._internal.exceptions import CommandError -from fetchcode.vcs.pip._internal.req.req_tracker import get_requirement_tracker -from fetchcode.vcs.pip._internal.utils.misc import ensure_dir, normalize_path -from fetchcode.vcs.pip._internal.utils.temp_dir import TempDirectory -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.wheel_builder import build, should_build_for_wheel_command - -if MYPY_CHECK_RUNNING: - from optparse import Values - from typing import Any, List - +from optparse import Values +from typing import List + +from pip._internal.cache import WheelCache +from pip._internal.cli import cmdoptions +from pip._internal.cli.req_command import RequirementCommand, with_cleanup +from pip._internal.cli.status_codes import SUCCESS +from pip._internal.exceptions import CommandError +from pip._internal.operations.build.build_tracker import get_build_tracker +from pip._internal.req.req_install import ( + InstallRequirement, + check_legacy_setup_py_options, +) +from pip._internal.utils.misc import ensure_dir, normalize_path +from pip._internal.utils.temp_dir import TempDirectory +from pip._internal.wheel_builder import build, should_build_for_wheel_command logger = logging.getLogger(__name__) @@ -35,10 +29,8 @@ class WheelCommand(RequirementCommand): recompiling your software during every install. For more details, see the wheel docs: https://wheel.readthedocs.io/en/latest/ - Requirements: setuptools>=0.8, and wheel. - - 'pip wheel' uses the bdist_wheel setuptools extension from the wheel - package to build individual wheels. + 'pip wheel' uses the build system interface as described here: + https://pip.pypa.io/en/stable/reference/build-system/ """ @@ -49,58 +41,56 @@ class WheelCommand(RequirementCommand): %prog [options] [-e] ... %prog [options] ...""" - def __init__(self, *args, **kw): - super(WheelCommand, self).__init__(*args, **kw) - - cmd_opts = self.cmd_opts - - cmd_opts.add_option( - '-w', '--wheel-dir', - dest='wheel_dir', - metavar='dir', + def add_options(self) -> None: + self.cmd_opts.add_option( + "-w", + "--wheel-dir", + dest="wheel_dir", + metavar="dir", default=os.curdir, - help=("Build wheels into , where the default is the " - "current working directory."), + help=( + "Build wheels into , where the default is the " + "current working directory." + ), ) - cmd_opts.add_option(cmdoptions.no_binary()) - cmd_opts.add_option(cmdoptions.only_binary()) - cmd_opts.add_option(cmdoptions.prefer_binary()) - cmd_opts.add_option( - '--build-option', - dest='build_options', - metavar='options', - action='append', - help="Extra arguments to be supplied to 'setup.py bdist_wheel'.", + self.cmd_opts.add_option(cmdoptions.no_binary()) + self.cmd_opts.add_option(cmdoptions.only_binary()) + self.cmd_opts.add_option(cmdoptions.prefer_binary()) + self.cmd_opts.add_option(cmdoptions.no_build_isolation()) + self.cmd_opts.add_option(cmdoptions.use_pep517()) + self.cmd_opts.add_option(cmdoptions.no_use_pep517()) + self.cmd_opts.add_option(cmdoptions.check_build_deps()) + self.cmd_opts.add_option(cmdoptions.constraints()) + self.cmd_opts.add_option(cmdoptions.editable()) + self.cmd_opts.add_option(cmdoptions.requirements()) + self.cmd_opts.add_option(cmdoptions.src()) + self.cmd_opts.add_option(cmdoptions.ignore_requires_python()) + self.cmd_opts.add_option(cmdoptions.no_deps()) + self.cmd_opts.add_option(cmdoptions.progress_bar()) + + self.cmd_opts.add_option( + "--no-verify", + dest="no_verify", + action="store_true", + default=False, + help="Don't verify if built wheel is valid.", ) - cmd_opts.add_option(cmdoptions.no_build_isolation()) - cmd_opts.add_option(cmdoptions.use_pep517()) - cmd_opts.add_option(cmdoptions.no_use_pep517()) - cmd_opts.add_option(cmdoptions.constraints()) - cmd_opts.add_option(cmdoptions.editable()) - cmd_opts.add_option(cmdoptions.requirements()) - cmd_opts.add_option(cmdoptions.src()) - cmd_opts.add_option(cmdoptions.ignore_requires_python()) - cmd_opts.add_option(cmdoptions.no_deps()) - cmd_opts.add_option(cmdoptions.build_dir()) - cmd_opts.add_option(cmdoptions.progress_bar()) - - cmd_opts.add_option( - '--global-option', - dest='global_options', - action='append', - metavar='options', - help="Extra global options to be supplied to the setup.py " - "call before the 'bdist_wheel' command.") - - cmd_opts.add_option( - '--pre', - action='store_true', + + self.cmd_opts.add_option(cmdoptions.config_settings()) + self.cmd_opts.add_option(cmdoptions.build_options()) + self.cmd_opts.add_option(cmdoptions.global_options()) + + self.cmd_opts.add_option( + "--pre", + action="store_true", default=False, - help=("Include pre-release and development versions. By default, " - "pip only finds stable versions."), + help=( + "Include pre-release and development versions. By default, " + "pip only finds stable versions." + ), ) - cmd_opts.add_option(cmdoptions.require_hashes()) + self.cmd_opts.add_option(cmdoptions.require_hashes()) index_opts = cmdoptions.make_option_group( cmdoptions.index_group, @@ -108,41 +98,39 @@ def __init__(self, *args, **kw): ) self.parser.insert_option_group(0, index_opts) - self.parser.insert_option_group(0, cmd_opts) + self.parser.insert_option_group(0, self.cmd_opts) @with_cleanup - def run(self, options, args): - # type: (Values, List[Any]) -> None - cmdoptions.check_install_build_global(options) - + def run(self, options: Values, args: List[str]) -> int: session = self.get_default_session(options) finder = self._build_package_finder(options, session) - build_delete = (not (options.no_clean or options.build_dir)) - wheel_cache = WheelCache(options.cache_dir, options.format_control) options.wheel_dir = normalize_path(options.wheel_dir) ensure_dir(options.wheel_dir) - req_tracker = self.enter_context(get_requirement_tracker()) + build_tracker = self.enter_context(get_build_tracker()) directory = TempDirectory( - options.build_dir, - delete=build_delete, + delete=not options.no_clean, kind="wheel", globally_managed=True, ) reqs = self.get_requirements(args, options, finder, session) + check_legacy_setup_py_options(options, reqs) + + wheel_cache = WheelCache(options.cache_dir) preparer = self.make_requirement_preparer( temp_build_dir=directory, options=options, - req_tracker=req_tracker, + build_tracker=build_tracker, session=session, finder=finder, - wheel_download_dir=options.wheel_dir, + download_dir=options.wheel_dir, use_user_site=False, + verbosity=self.verbosity, ) resolver = self.make_resolver( @@ -156,19 +144,22 @@ def run(self, options, args): self.trace_basic_info(finder) - requirement_set = resolver.resolve( - reqs, check_supported_wheels=True - ) + requirement_set = resolver.resolve(reqs, check_supported_wheels=True) - reqs_to_build = [ - r for r in requirement_set.requirements.values() - if should_build_for_wheel_command(r) - ] + reqs_to_build: List[InstallRequirement] = [] + for req in requirement_set.requirements.values(): + if req.is_wheel: + preparer.save_linked_requirement(req) + elif should_build_for_wheel_command(req): + reqs_to_build.append(req) + + preparer.prepare_linked_requirements_more(requirement_set.requirements.values()) # build wheels build_successes, build_failures = build( reqs_to_build, wheel_cache=wheel_cache, + verify=(not options.no_verify), build_options=options.build_options or [], global_options=options.global_options or [], ) @@ -181,10 +172,11 @@ def run(self, options, args): except OSError as e: logger.warning( "Building wheel for %s failed: %s", - req.name, e, + req.name, + e, ) build_failures.append(req) if len(build_failures) != 0: - raise CommandError( - "Failed to build one or more wheels" - ) + raise CommandError("Failed to build one or more wheels") + + return SUCCESS diff --git a/src/fetchcode/vcs/pip/_internal/configuration.py b/src/fetchcode/vcs/pip/_internal/configuration.py index 58e00015..c25273d5 100644 --- a/src/fetchcode/vcs/pip/_internal/configuration.py +++ b/src/fetchcode/vcs/pip/_internal/configuration.py @@ -11,87 +11,72 @@ A single word describing where the configuration key-value pair came from """ -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False - +import configparser import locale -import logging import os import sys +from typing import Any, Dict, Iterable, List, NewType, Optional, Tuple -from fetchcode.vcs.pip._vendor.six.moves import configparser - -from fetchcode.vcs.pip._internal.exceptions import ( +from pip._internal.exceptions import ( ConfigurationError, ConfigurationFileCouldNotBeLoaded, ) -from fetchcode.vcs.pip._internal.utils import appdirs -from fetchcode.vcs.pip._internal.utils.compat import WINDOWS, expanduser -from fetchcode.vcs.pip._internal.utils.misc import ensure_dir, enum -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import ( - Any, Dict, Iterable, List, NewType, Optional, Tuple - ) +from pip._internal.utils import appdirs +from pip._internal.utils.compat import WINDOWS +from pip._internal.utils.logging import getLogger +from pip._internal.utils.misc import ensure_dir, enum - RawConfigParser = configparser.RawConfigParser # Shorthand - Kind = NewType("Kind", str) +RawConfigParser = configparser.RawConfigParser # Shorthand +Kind = NewType("Kind", str) + +CONFIG_BASENAME = "pip.ini" if WINDOWS else "pip.conf" +ENV_NAMES_IGNORED = "version", "help" + +# The kinds of configurations there are. +kinds = enum( + USER="user", # User Specific + GLOBAL="global", # System Wide + SITE="site", # [Virtual] Environment Specific + ENV="env", # from PIP_CONFIG_FILE + ENV_VAR="env-var", # from Environment Variables +) +OVERRIDE_ORDER = kinds.GLOBAL, kinds.USER, kinds.SITE, kinds.ENV, kinds.ENV_VAR +VALID_LOAD_ONLY = kinds.USER, kinds.GLOBAL, kinds.SITE -logger = logging.getLogger(__name__) +logger = getLogger(__name__) # NOTE: Maybe use the optionx attribute to normalize keynames. -def _normalize_name(name): - # type: (str) -> str - """Make a name consistent regardless of source (environment or file) - """ - name = name.lower().replace('_', '-') - if name.startswith('--'): +def _normalize_name(name: str) -> str: + """Make a name consistent regardless of source (environment or file)""" + name = name.lower().replace("_", "-") + if name.startswith("--"): name = name[2:] # only prefer long opts return name -def _disassemble_key(name): - # type: (str) -> List[str] +def _disassemble_key(name: str) -> List[str]: if "." not in name: error_message = ( "Key does not contain dot separated section and key. " - "Perhaps you wanted to use 'global.{}' instead?" - ).format(name) + f"Perhaps you wanted to use 'global.{name}' instead?" + ) raise ConfigurationError(error_message) return name.split(".", 1) -# The kinds of configurations there are. -kinds = enum( - USER="user", # User Specific - GLOBAL="global", # System Wide - SITE="site", # [Virtual] Environment Specific - ENV="env", # from PIP_CONFIG_FILE - ENV_VAR="env-var", # from Environment Variables -) - - -CONFIG_BASENAME = 'pip.ini' if WINDOWS else 'pip.conf' - - -def get_configuration_files(): - # type: () -> Dict[Kind, List[str]] +def get_configuration_files() -> Dict[Kind, List[str]]: global_config_files = [ - os.path.join(path, CONFIG_BASENAME) - for path in appdirs.site_config_dirs('pip') + os.path.join(path, CONFIG_BASENAME) for path in appdirs.site_config_dirs("pip") ] site_config_file = os.path.join(sys.prefix, CONFIG_BASENAME) legacy_config_file = os.path.join( - expanduser('~'), - 'pip' if WINDOWS else '.pip', + os.path.expanduser("~"), + "pip" if WINDOWS else ".pip", CONFIG_BASENAME, ) - new_config_file = os.path.join( - appdirs.user_config_dir("pip"), CONFIG_BASENAME - ) + new_config_file = os.path.join(appdirs.user_config_dir("pip"), CONFIG_BASENAME) return { kinds.GLOBAL: global_config_files, kinds.SITE: [site_config_file], @@ -99,7 +84,7 @@ def get_configuration_files(): } -class Configuration(object): +class Configuration: """Handles management of configuration. Provides an interface to accessing and managing configuration files. @@ -113,78 +98,66 @@ class Configuration(object): and the data stored is also nice. """ - def __init__(self, isolated, load_only=None): - # type: (bool, Kind) -> None - super(Configuration, self).__init__() + def __init__(self, isolated: bool, load_only: Optional[Kind] = None) -> None: + super().__init__() - _valid_load_only = [kinds.USER, kinds.GLOBAL, kinds.SITE, None] - if load_only not in _valid_load_only: + if load_only is not None and load_only not in VALID_LOAD_ONLY: raise ConfigurationError( "Got invalid value for load_only - should be one of {}".format( - ", ".join(map(repr, _valid_load_only[:-1])) + ", ".join(map(repr, VALID_LOAD_ONLY)) ) ) - self.isolated = isolated # type: bool - self.load_only = load_only # type: Optional[Kind] - - # The order here determines the override order. - self._override_order = [ - kinds.GLOBAL, kinds.USER, kinds.SITE, kinds.ENV, kinds.ENV_VAR - ] - - self._ignore_env_names = ["version", "help"] + self.isolated = isolated + self.load_only = load_only # Because we keep track of where we got the data from - self._parsers = { - variant: [] for variant in self._override_order - } # type: Dict[Kind, List[Tuple[str, RawConfigParser]]] - self._config = { - variant: {} for variant in self._override_order - } # type: Dict[Kind, Dict[str, Any]] - self._modified_parsers = [] # type: List[Tuple[str, RawConfigParser]] - - def load(self): - # type: () -> None - """Loads configuration from configuration files and environment - """ + self._parsers: Dict[Kind, List[Tuple[str, RawConfigParser]]] = { + variant: [] for variant in OVERRIDE_ORDER + } + self._config: Dict[Kind, Dict[str, Any]] = { + variant: {} for variant in OVERRIDE_ORDER + } + self._modified_parsers: List[Tuple[str, RawConfigParser]] = [] + + def load(self) -> None: + """Loads configuration from configuration files and environment""" self._load_config_files() if not self.isolated: self._load_environment_vars() - def get_file_to_edit(self): - # type: () -> Optional[str] - """Returns the file with highest priority in configuration - """ - assert self.load_only is not None, \ - "Need to be specified a file to be editing" + def get_file_to_edit(self) -> Optional[str]: + """Returns the file with highest priority in configuration""" + assert self.load_only is not None, "Need to be specified a file to be editing" try: return self._get_parser_to_modify()[0] except IndexError: return None - def items(self): - # type: () -> Iterable[Tuple[str, Any]] + def items(self) -> Iterable[Tuple[str, Any]]: """Returns key-value pairs like dict.items() representing the loaded configuration """ return self._dictionary.items() - def get_value(self, key): - # type: (str) -> Any - """Get a value from the configuration. - """ + def get_value(self, key: str) -> Any: + """Get a value from the configuration.""" + orig_key = key + key = _normalize_name(key) try: return self._dictionary[key] except KeyError: - raise ConfigurationError("No such key - {}".format(key)) - - def set_value(self, key, value): - # type: (str, Any) -> None - """Modify a value in the configuration. - """ + # disassembling triggers a more useful error message than simply + # "No such key" in the case that the key isn't in the form command.option + _disassemble_key(key) + raise ConfigurationError(f"No such key - {orig_key}") + + def set_value(self, key: str, value: Any) -> None: + """Modify a value in the configuration.""" + key = _normalize_name(key) self._ensure_have_load_only() + assert self.load_only fname, parser = self._get_parser_to_modify() if parser is not None: @@ -198,49 +171,37 @@ def set_value(self, key, value): self._config[self.load_only][key] = value self._mark_as_modified(fname, parser) - def unset_value(self, key): - # type: (str) -> None - """Unset a value in the configuration. - """ + def unset_value(self, key: str) -> None: + """Unset a value in the configuration.""" + orig_key = key + key = _normalize_name(key) self._ensure_have_load_only() + assert self.load_only if key not in self._config[self.load_only]: - raise ConfigurationError("No such key - {}".format(key)) + raise ConfigurationError(f"No such key - {orig_key}") fname, parser = self._get_parser_to_modify() if parser is not None: section, name = _disassemble_key(key) - - # Remove the key in the parser - modified_something = False - if parser.has_section(section): - # Returns whether the option was removed or not - modified_something = parser.remove_option(section, name) - - if modified_something: - # name removed from parser, section may now be empty - section_iter = iter(parser.items(section)) - try: - val = next(section_iter) - except StopIteration: - val = None - - if val is None: - parser.remove_section(section) - - self._mark_as_modified(fname, parser) - else: + if not ( + parser.has_section(section) and parser.remove_option(section, name) + ): + # The option was not removed. raise ConfigurationError( "Fatal Internal error [id=1]. Please report as a bug." ) + # The section may be empty after the option was removed. + if not parser.items(section): + parser.remove_section(section) + self._mark_as_modified(fname, parser) + del self._config[self.load_only][key] - def save(self): - # type: () -> None - """Save the current in-memory state. - """ + def save(self) -> None: + """Save the current in-memory state.""" self._ensure_have_load_only() for fname, parser in self._modified_parsers: @@ -249,38 +210,40 @@ def save(self): # Ensure directory exists. ensure_dir(os.path.dirname(fname)) - with open(fname, "w") as f: - parser.write(f) + # Ensure directory's permission(need to be writeable) + try: + with open(fname, "w") as f: + parser.write(f) + except OSError as error: + raise ConfigurationError( + f"An error occurred while writing to the configuration file " + f"{fname}: {error}" + ) # # Private routines # - def _ensure_have_load_only(self): - # type: () -> None + def _ensure_have_load_only(self) -> None: if self.load_only is None: raise ConfigurationError("Needed a specific file to be modifying.") logger.debug("Will be working with %s variant only", self.load_only) @property - def _dictionary(self): - # type: () -> Dict[str, Any] - """A dictionary representing the loaded configuration. - """ + def _dictionary(self) -> Dict[str, Any]: + """A dictionary representing the loaded configuration.""" # NOTE: Dictionaries are not populated if not loaded. So, conditionals # are not needed here. retval = {} - for variant in self._override_order: + for variant in OVERRIDE_ORDER: retval.update(self._config[variant]) return retval - def _load_config_files(self): - # type: () -> None - """Loads configuration from configuration files - """ - config_files = dict(self._iter_config_files()) + def _load_config_files(self) -> None: + """Loads configuration from configuration files""" + config_files = dict(self.iter_config_files()) if config_files[kinds.ENV][0:1] == [os.devnull]: logger.debug( "Skipping loading configuration files due to " @@ -293,9 +256,7 @@ def _load_config_files(self): # If there's specific variant set in `load_only`, load only # that variant, not the others. if self.load_only is not None and variant != self.load_only: - logger.debug( - "Skipping file '%s' (variant: %s)", fname, variant - ) + logger.debug("Skipping file '%s' (variant: %s)", fname, variant) continue parser = self._load_file(variant, fname) @@ -303,9 +264,8 @@ def _load_config_files(self): # Keeping track of the parsers used self._parsers[variant].append((fname, parser)) - def _load_file(self, variant, fname): - # type: (Kind, str) -> RawConfigParser - logger.debug("For variant '%s', will try loading '%s'", variant, fname) + def _load_file(self, variant: Kind, fname: str) -> RawConfigParser: + logger.verbose("For variant '%s', will try loading '%s'", variant, fname) parser = self._construct_parser(fname) for section in parser.sections(): @@ -314,22 +274,20 @@ def _load_file(self, variant, fname): return parser - def _construct_parser(self, fname): - # type: (str) -> RawConfigParser + def _construct_parser(self, fname: str) -> RawConfigParser: parser = configparser.RawConfigParser() # If there is no such file, don't bother reading it but create the # parser anyway, to hold the data. # Doing this is useful when modifying and saving files, where we don't # need to construct a parser. if os.path.exists(fname): + locale_encoding = locale.getpreferredencoding(False) try: - parser.read(fname) + parser.read(fname, encoding=locale_encoding) except UnicodeDecodeError: # See https://github.com/pypa/pip/issues/4963 raise ConfigurationFileCouldNotBeLoaded( - reason="contains invalid {} characters".format( - locale.getpreferredencoding(False) - ), + reason=f"contains invalid {locale_encoding} characters", fname=fname, ) except configparser.Error as error: @@ -337,16 +295,15 @@ def _construct_parser(self, fname): raise ConfigurationFileCouldNotBeLoaded(error=error) return parser - def _load_environment_vars(self): - # type: () -> None - """Loads configuration from environment variables - """ + def _load_environment_vars(self) -> None: + """Loads configuration from environment variables""" self._config[kinds.ENV_VAR].update( - self._normalized_keys(":env:", self._get_environ_vars()) + self._normalized_keys(":env:", self.get_environ_vars()) ) - def _normalized_keys(self, section, items): - # type: (str, Iterable[Tuple[str, Any]]) -> Dict[str, Any] + def _normalized_keys( + self, section: str, items: Iterable[Tuple[str, Any]] + ) -> Dict[str, Any]: """Normalizes items to construct a dictionary with normalized keys. This routine is where the names become keys and are made the same @@ -358,52 +315,54 @@ def _normalized_keys(self, section, items): normalized[key] = val return normalized - def _get_environ_vars(self): - # type: () -> Iterable[Tuple[str, str]] + def get_environ_vars(self) -> Iterable[Tuple[str, str]]: """Returns a generator with all environmental vars with prefix PIP_""" for key, val in os.environ.items(): - should_be_yielded = ( - key.startswith("PIP_") and - key[4:].lower() not in self._ignore_env_names - ) - if should_be_yielded: - yield key[4:].lower(), val + if key.startswith("PIP_"): + name = key[4:].lower() + if name not in ENV_NAMES_IGNORED: + yield name, val # XXX: This is patched in the tests. - def _iter_config_files(self): - # type: () -> Iterable[Tuple[Kind, List[str]]] + def iter_config_files(self) -> Iterable[Tuple[Kind, List[str]]]: """Yields variant and configuration files associated with it. - This should be treated like items of a dictionary. + This should be treated like items of a dictionary. The order + here doesn't affect what gets overridden. That is controlled + by OVERRIDE_ORDER. However this does control the order they are + displayed to the user. It's probably most ergononmic to display + things in the same order as OVERRIDE_ORDER """ # SMELL: Move the conditions out of this function - # environment variables have the lowest priority - config_file = os.environ.get('PIP_CONFIG_FILE', None) - if config_file is not None: - yield kinds.ENV, [config_file] - else: - yield kinds.ENV, [] - + env_config_file = os.environ.get("PIP_CONFIG_FILE", None) config_files = get_configuration_files() - # at the base we have any global configuration yield kinds.GLOBAL, config_files[kinds.GLOBAL] - # per-user configuration next + # per-user config is not loaded when env_config_file exists should_load_user_config = not self.isolated and not ( - config_file and os.path.exists(config_file) + env_config_file and os.path.exists(env_config_file) ) if should_load_user_config: # The legacy config file is overridden by the new config file yield kinds.USER, config_files[kinds.USER] - # finally virtualenv configuration first trumping others + # virtualenv config yield kinds.SITE, config_files[kinds.SITE] - def _get_parser_to_modify(self): - # type: () -> Tuple[str, RawConfigParser] + if env_config_file is not None: + yield kinds.ENV, [env_config_file] + else: + yield kinds.ENV, [] + + def get_values_in_config(self, variant: Kind) -> Dict[str, Any]: + """Get values present in a config file""" + return self._config[variant] + + def _get_parser_to_modify(self) -> Tuple[str, RawConfigParser]: # Determine which parser to modify + assert self.load_only parsers = self._parsers[self.load_only] if not parsers: # This should not happen if everything works correctly. @@ -415,12 +374,10 @@ def _get_parser_to_modify(self): return parsers[-1] # XXX: This is patched in the tests. - def _mark_as_modified(self, fname, parser): - # type: (str, RawConfigParser) -> None + def _mark_as_modified(self, fname: str, parser: RawConfigParser) -> None: file_parser_tuple = (fname, parser) if file_parser_tuple not in self._modified_parsers: self._modified_parsers.append(file_parser_tuple) - def __repr__(self): - # type: () -> str - return "{}({!r})".format(self.__class__.__name__, self._dictionary) + def __repr__(self) -> str: + return f"{self.__class__.__name__}({self._dictionary!r})" diff --git a/src/fetchcode/vcs/pip/_internal/distributions/__init__.py b/src/fetchcode/vcs/pip/_internal/distributions/__init__.py index 3a02295c..9a89a838 100644 --- a/src/fetchcode/vcs/pip/_internal/distributions/__init__.py +++ b/src/fetchcode/vcs/pip/_internal/distributions/__init__.py @@ -1,16 +1,13 @@ -from fetchcode.vcs.pip._internal.distributions.sdist import SourceDistribution -from fetchcode.vcs.pip._internal.distributions.wheel import WheelDistribution -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._internal.distributions.base import AbstractDistribution +from pip._internal.distributions.sdist import SourceDistribution +from pip._internal.distributions.wheel import WheelDistribution +from pip._internal.req.req_install import InstallRequirement -if MYPY_CHECK_RUNNING: - from fetchcode.vcs.pip._internal.distributions.base import AbstractDistribution - from fetchcode.vcs.pip._internal.req.req_install import InstallRequirement - -def make_distribution_for_install_requirement(install_req): - # type: (InstallRequirement) -> AbstractDistribution - """Returns a Distribution for the given InstallRequirement - """ +def make_distribution_for_install_requirement( + install_req: InstallRequirement, +) -> AbstractDistribution: + """Returns a Distribution for the given InstallRequirement""" # Editable requirements will always be source distributions. They use the # legacy logic until we create a modern standard for them. if install_req.editable: diff --git a/src/fetchcode/vcs/pip/_internal/distributions/base.py b/src/fetchcode/vcs/pip/_internal/distributions/base.py index 356e24d4..6e4d0c91 100644 --- a/src/fetchcode/vcs/pip/_internal/distributions/base.py +++ b/src/fetchcode/vcs/pip/_internal/distributions/base.py @@ -1,19 +1,14 @@ import abc +from typing import TYPE_CHECKING, Optional -from fetchcode.vcs.pip._vendor.six import add_metaclass +from pip._internal.metadata.base import BaseDistribution +from pip._internal.req import InstallRequirement -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +if TYPE_CHECKING: + from pip._internal.index.package_finder import PackageFinder -if MYPY_CHECK_RUNNING: - from typing import Optional - from fetchcode.vcs.pip._vendor.pkg_resources import Distribution - from fetchcode.vcs.pip._internal.req import InstallRequirement - from fetchcode.vcs.pip._internal.index.package_finder import PackageFinder - - -@add_metaclass(abc.ABCMeta) -class AbstractDistribution(object): +class AbstractDistribution(metaclass=abc.ABCMeta): """A base class for handling installable artifacts. The requirements for anything installable are as follows: @@ -27,19 +22,32 @@ class AbstractDistribution(object): - we must be able to create a Distribution object exposing the above metadata. + + - if we need to do work in the build tracker, we must be able to generate a unique + string to identify the requirement in the build tracker. """ - def __init__(self, req): - # type: (InstallRequirement) -> None - super(AbstractDistribution, self).__init__() + def __init__(self, req: InstallRequirement) -> None: + super().__init__() self.req = req + @abc.abstractproperty + def build_tracker_id(self) -> Optional[str]: + """A string that uniquely identifies this requirement to the build tracker. + + If None, then this dist has no work to do in the build tracker, and + ``.prepare_distribution_metadata()`` will not be called.""" + raise NotImplementedError() + @abc.abstractmethod - def get_pkg_resources_distribution(self): - # type: () -> Optional[Distribution] + def get_metadata_distribution(self) -> BaseDistribution: raise NotImplementedError() @abc.abstractmethod - def prepare_distribution_metadata(self, finder, build_isolation): - # type: (PackageFinder, bool) -> None + def prepare_distribution_metadata( + self, + finder: "PackageFinder", + build_isolation: bool, + check_build_deps: bool, + ) -> None: raise NotImplementedError() diff --git a/src/fetchcode/vcs/pip/_internal/distributions/installed.py b/src/fetchcode/vcs/pip/_internal/distributions/installed.py index 72e01380..ab8d53be 100644 --- a/src/fetchcode/vcs/pip/_internal/distributions/installed.py +++ b/src/fetchcode/vcs/pip/_internal/distributions/installed.py @@ -1,11 +1,8 @@ -from fetchcode.vcs.pip._internal.distributions.base import AbstractDistribution -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from typing import Optional -if MYPY_CHECK_RUNNING: - from typing import Optional - - from fetchcode.vcs.pip._vendor.pkg_resources import Distribution - from fetchcode.vcs.pip._internal.index.package_finder import PackageFinder +from pip._internal.distributions.base import AbstractDistribution +from pip._internal.index.package_finder import PackageFinder +from pip._internal.metadata import BaseDistribution class InstalledDistribution(AbstractDistribution): @@ -15,10 +12,18 @@ class InstalledDistribution(AbstractDistribution): been computed. """ - def get_pkg_resources_distribution(self): - # type: () -> Optional[Distribution] + @property + def build_tracker_id(self) -> Optional[str]: + return None + + def get_metadata_distribution(self) -> BaseDistribution: + assert self.req.satisfied_by is not None, "not actually installed" return self.req.satisfied_by - def prepare_distribution_metadata(self, finder, build_isolation): - # type: (PackageFinder, bool) -> None + def prepare_distribution_metadata( + self, + finder: PackageFinder, + build_isolation: bool, + check_build_deps: bool, + ) -> None: pass diff --git a/src/fetchcode/vcs/pip/_internal/distributions/sdist.py b/src/fetchcode/vcs/pip/_internal/distributions/sdist.py index af05dfb5..28ea5cea 100644 --- a/src/fetchcode/vcs/pip/_internal/distributions/sdist.py +++ b/src/fetchcode/vcs/pip/_internal/distributions/sdist.py @@ -1,17 +1,14 @@ import logging +from typing import TYPE_CHECKING, Iterable, Optional, Set, Tuple -from fetchcode.vcs.pip._internal.build_env import BuildEnvironment -from fetchcode.vcs.pip._internal.distributions.base import AbstractDistribution -from fetchcode.vcs.pip._internal.exceptions import InstallationError -from fetchcode.vcs.pip._internal.utils.subprocess import runner_with_spinner_message -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import Set, Tuple - - from fetchcode.vcs.pip._vendor.pkg_resources import Distribution - from fetchcode.vcs.pip._internal.index.package_finder import PackageFinder +from pip._internal.build_env import BuildEnvironment +from pip._internal.distributions.base import AbstractDistribution +from pip._internal.exceptions import InstallationError +from pip._internal.metadata import BaseDistribution +from pip._internal.utils.subprocess import runner_with_spinner_message +if TYPE_CHECKING: + from pip._internal.index.package_finder import PackageFinder logger = logging.getLogger(__name__) @@ -23,40 +20,55 @@ class SourceDistribution(AbstractDistribution): generated, either using PEP 517 or using the legacy `setup.py egg_info`. """ - def get_pkg_resources_distribution(self): - # type: () -> Distribution + @property + def build_tracker_id(self) -> Optional[str]: + """Identify this requirement uniquely by its link.""" + assert self.req.link + return self.req.link.url_without_fragment + + def get_metadata_distribution(self) -> BaseDistribution: return self.req.get_dist() - def prepare_distribution_metadata(self, finder, build_isolation): - # type: (PackageFinder, bool) -> None + def prepare_distribution_metadata( + self, + finder: "PackageFinder", + build_isolation: bool, + check_build_deps: bool, + ) -> None: # Load pyproject.toml, to determine whether PEP 517 is to be used self.req.load_pyproject_toml() # Set up the build isolation, if this requirement should be isolated should_isolate = self.req.use_pep517 and build_isolation if should_isolate: - self._setup_isolation(finder) - - self.req.prepare_metadata() - - def _setup_isolation(self, finder): - # type: (PackageFinder) -> None - def _raise_conflicts(conflicting_with, conflicting_reqs): - # type: (str, Set[Tuple[str, str]]) -> None - format_string = ( - "Some build dependencies for {requirement} " - "conflict with {conflicting_with}: {description}." - ) - error_message = format_string.format( - requirement=self.req, - conflicting_with=conflicting_with, - description=', '.join( - '{} is incompatible with {}'.format(installed, wanted) - for installed, wanted in sorted(conflicting) - ) + # Setup an isolated environment and install the build backend static + # requirements in it. + self._prepare_build_backend(finder) + # Check that if the requirement is editable, it either supports PEP 660 or + # has a setup.py or a setup.cfg. This cannot be done earlier because we need + # to setup the build backend to verify it supports build_editable, nor can + # it be done later, because we want to avoid installing build requirements + # needlessly. Doing it here also works around setuptools generating + # UNKNOWN.egg-info when running get_requires_for_build_wheel on a directory + # without setup.py nor setup.cfg. + self.req.isolated_editable_sanity_check() + # Install the dynamic build requirements. + self._install_build_reqs(finder) + # Check if the current environment provides build dependencies + should_check_deps = self.req.use_pep517 and check_build_deps + if should_check_deps: + pyproject_requires = self.req.pyproject_requires + assert pyproject_requires is not None + conflicting, missing = self.req.build_env.check_requirements( + pyproject_requires ) - raise InstallationError(error_message) + if conflicting: + self._raise_conflicts("the backend dependencies", conflicting) + if missing: + self._raise_missing_reqs(missing) + self.req.prepare_metadata() + def _prepare_build_backend(self, finder: "PackageFinder") -> None: # Isolate in a BuildEnvironment and install the build-time # requirements. pyproject_requires = self.req.pyproject_requires @@ -64,15 +76,13 @@ def _raise_conflicts(conflicting_with, conflicting_reqs): self.req.build_env = BuildEnvironment() self.req.build_env.install_requirements( - finder, pyproject_requires, 'overlay', - "Installing build dependencies" + finder, pyproject_requires, "overlay", kind="build dependencies" ) conflicting, missing = self.req.build_env.check_requirements( self.req.requirements_to_check ) if conflicting: - _raise_conflicts("PEP 517/518 supported requirements", - conflicting) + self._raise_conflicts("PEP 517/518 supported requirements", conflicting) if missing: logger.warning( "Missing build requirements in pyproject.toml for %s.", @@ -81,24 +91,68 @@ def _raise_conflicts(conflicting_with, conflicting_reqs): logger.warning( "The project does not specify a build backend, and " "pip cannot fall back to setuptools without %s.", - " and ".join(map(repr, sorted(missing))) + " and ".join(map(repr, sorted(missing))), ) - # Install any extra build dependencies that the backend requests. - # This must be done in a second pass, as the pyproject.toml - # dependencies must be installed before we can call the backend. + + def _get_build_requires_wheel(self) -> Iterable[str]: + with self.req.build_env: + runner = runner_with_spinner_message("Getting requirements to build wheel") + backend = self.req.pep517_backend + assert backend is not None + with backend.subprocess_runner(runner): + return backend.get_requires_for_build_wheel() + + def _get_build_requires_editable(self) -> Iterable[str]: with self.req.build_env: runner = runner_with_spinner_message( - "Getting requirements to build wheel" + "Getting requirements to build editable" ) backend = self.req.pep517_backend assert backend is not None with backend.subprocess_runner(runner): - reqs = backend.get_requires_for_build_wheel() + return backend.get_requires_for_build_editable() - conflicting, missing = self.req.build_env.check_requirements(reqs) + def _install_build_reqs(self, finder: "PackageFinder") -> None: + # Install any extra build dependencies that the backend requests. + # This must be done in a second pass, as the pyproject.toml + # dependencies must be installed before we can call the backend. + if ( + self.req.editable + and self.req.permit_editable_wheels + and self.req.supports_pyproject_editable + ): + build_reqs = self._get_build_requires_editable() + else: + build_reqs = self._get_build_requires_wheel() + conflicting, missing = self.req.build_env.check_requirements(build_reqs) if conflicting: - _raise_conflicts("the backend dependencies", conflicting) + self._raise_conflicts("the backend dependencies", conflicting) self.req.build_env.install_requirements( - finder, missing, 'normal', - "Installing backend dependencies" + finder, missing, "normal", kind="backend dependencies" + ) + + def _raise_conflicts( + self, conflicting_with: str, conflicting_reqs: Set[Tuple[str, str]] + ) -> None: + format_string = ( + "Some build dependencies for {requirement} " + "conflict with {conflicting_with}: {description}." + ) + error_message = format_string.format( + requirement=self.req, + conflicting_with=conflicting_with, + description=", ".join( + f"{installed} is incompatible with {wanted}" + for installed, wanted in sorted(conflicting_reqs) + ), + ) + raise InstallationError(error_message) + + def _raise_missing_reqs(self, missing: Set[str]) -> None: + format_string = ( + "Some build dependencies for {requirement} are missing: {missing}." + ) + error_message = format_string.format( + requirement=self.req, missing=", ".join(map(repr, sorted(missing))) ) + raise InstallationError(error_message) diff --git a/src/fetchcode/vcs/pip/_internal/distributions/wheel.py b/src/fetchcode/vcs/pip/_internal/distributions/wheel.py index a8b58ec1..bfadd39d 100644 --- a/src/fetchcode/vcs/pip/_internal/distributions/wheel.py +++ b/src/fetchcode/vcs/pip/_internal/distributions/wheel.py @@ -1,12 +1,16 @@ -from zipfile import ZipFile +from typing import TYPE_CHECKING, Optional -from fetchcode.vcs.pip._internal.distributions.base import AbstractDistribution -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.utils.wheel import pkg_resources_distribution_for_wheel +from pip._vendor.packaging.utils import canonicalize_name -if MYPY_CHECK_RUNNING: - from fetchcode.vcs.pip._vendor.pkg_resources import Distribution - from fetchcode.vcs.pip._internal.index.package_finder import PackageFinder +from pip._internal.distributions.base import AbstractDistribution +from pip._internal.metadata import ( + BaseDistribution, + FilesystemWheel, + get_wheel_distribution, +) + +if TYPE_CHECKING: + from pip._internal.index.package_finder import PackageFinder class WheelDistribution(AbstractDistribution): @@ -15,22 +19,24 @@ class WheelDistribution(AbstractDistribution): This does not need any preparation as wheels can be directly unpacked. """ - def get_pkg_resources_distribution(self): - # type: () -> Distribution + @property + def build_tracker_id(self) -> Optional[str]: + return None + + def get_metadata_distribution(self) -> BaseDistribution: """Loads the metadata from the wheel file into memory and returns a Distribution that uses it, not relying on the wheel file or requirement. """ - # Set as part of preparation during download. - assert self.req.local_file_path - # Wheels are never unnamed. - assert self.req.name - - with ZipFile(self.req.local_file_path, allowZip64=True) as z: - return pkg_resources_distribution_for_wheel( - z, self.req.name, self.req.local_file_path - ) - - def prepare_distribution_metadata(self, finder, build_isolation): - # type: (PackageFinder, bool) -> None + assert self.req.local_file_path, "Set as part of preparation during download" + assert self.req.name, "Wheels are never unnamed" + wheel = FilesystemWheel(self.req.local_file_path) + return get_wheel_distribution(wheel, canonicalize_name(self.req.name)) + + def prepare_distribution_metadata( + self, + finder: "PackageFinder", + build_isolation: bool, + check_build_deps: bool, + ) -> None: pass diff --git a/src/fetchcode/vcs/pip/_internal/exceptions.py b/src/fetchcode/vcs/pip/_internal/exceptions.py index d1b8f6c3..2587740f 100644 --- a/src/fetchcode/vcs/pip/_internal/exceptions.py +++ b/src/fetchcode/vcs/pip/_internal/exceptions.py @@ -1,26 +1,181 @@ -"""Exceptions used throughout package""" +"""Exceptions used throughout package. + +This module MUST NOT try to import from anything within `pip._internal` to +operate. This is expected to be importable from any/all files within the +subpackage and, thus, should not depend on them. +""" + +import configparser +import contextlib +import locale +import logging +import pathlib +import re +import sys +from itertools import chain, groupby, repeat +from typing import TYPE_CHECKING, Dict, Iterator, List, Literal, Optional, Union -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False +from pip._vendor.rich.console import Console, ConsoleOptions, RenderResult +from pip._vendor.rich.markup import escape +from pip._vendor.rich.text import Text -from __future__ import absolute_import +if TYPE_CHECKING: + from hashlib import _Hash -from itertools import chain, groupby, repeat + from pip._vendor.requests.models import Request, Response + + from pip._internal.metadata import BaseDistribution + from pip._internal.req.req_install import InstallRequirement + +logger = logging.getLogger(__name__) -from fetchcode.vcs.pip._vendor.six import iteritems -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +# +# Scaffolding +# +def _is_kebab_case(s: str) -> bool: + return re.match(r"^[a-z]+(-[a-z]+)*$", s) is not None -if MYPY_CHECK_RUNNING: - from typing import Optional - from fetchcode.vcs.pip._vendor.pkg_resources import Distribution - from fetchcode.vcs.pip._internal.req.req_install import InstallRequirement + +def _prefix_with_indent( + s: Union[Text, str], + console: Console, + *, + prefix: str, + indent: str, +) -> Text: + if isinstance(s, Text): + text = s + else: + text = console.render_str(s) + + return console.render_str(prefix, overflow="ignore") + console.render_str( + f"\n{indent}", overflow="ignore" + ).join(text.split(allow_blank=True)) class PipError(Exception): - """Base pip exception""" + """The base pip error.""" + + +class DiagnosticPipError(PipError): + """An error, that presents diagnostic information to the user. + + This contains a bunch of logic, to enable pretty presentation of our error + messages. Each error gets a unique reference. Each error can also include + additional context, a hint and/or a note -- which are presented with the + main error message in a consistent style. + + This is adapted from the error output styling in `sphinx-theme-builder`. + """ + + reference: str + + def __init__( + self, + *, + kind: 'Literal["error", "warning"]' = "error", + reference: Optional[str] = None, + message: Union[str, Text], + context: Optional[Union[str, Text]], + hint_stmt: Optional[Union[str, Text]], + note_stmt: Optional[Union[str, Text]] = None, + link: Optional[str] = None, + ) -> None: + # Ensure a proper reference is provided. + if reference is None: + assert hasattr(self, "reference"), "error reference not provided!" + reference = self.reference + assert _is_kebab_case(reference), "error reference must be kebab-case!" + + self.kind = kind + self.reference = reference + + self.message = message + self.context = context + self.note_stmt = note_stmt + self.hint_stmt = hint_stmt + self.link = link + + super().__init__(f"<{self.__class__.__name__}: {self.reference}>") + + def __repr__(self) -> str: + return ( + f"<{self.__class__.__name__}(" + f"reference={self.reference!r}, " + f"message={self.message!r}, " + f"context={self.context!r}, " + f"note_stmt={self.note_stmt!r}, " + f"hint_stmt={self.hint_stmt!r}" + ")>" + ) + + def __rich_console__( + self, + console: Console, + options: ConsoleOptions, + ) -> RenderResult: + colour = "red" if self.kind == "error" else "yellow" + + yield f"[{colour} bold]{self.kind}[/]: [bold]{self.reference}[/]" + yield "" + + if not options.ascii_only: + # Present the main message, with relevant context indented. + if self.context is not None: + yield _prefix_with_indent( + self.message, + console, + prefix=f"[{colour}]×[/] ", + indent=f"[{colour}]│[/] ", + ) + yield _prefix_with_indent( + self.context, + console, + prefix=f"[{colour}]╰─>[/] ", + indent=f"[{colour}] [/] ", + ) + else: + yield _prefix_with_indent( + self.message, + console, + prefix="[red]×[/] ", + indent=" ", + ) + else: + yield self.message + if self.context is not None: + yield "" + yield self.context + + if self.note_stmt is not None or self.hint_stmt is not None: + yield "" + + if self.note_stmt is not None: + yield _prefix_with_indent( + self.note_stmt, + console, + prefix="[magenta bold]note[/]: ", + indent=" ", + ) + if self.hint_stmt is not None: + yield _prefix_with_indent( + self.hint_stmt, + console, + prefix="[cyan bold]hint[/]: ", + indent=" ", + ) + + if self.link is not None: + yield "" + yield f"Link: {self.link}" + + +# +# Actual Errors +# class ConfigurationError(PipError): """General exception in configuration""" @@ -29,21 +184,54 @@ class InstallationError(PipError): """General exception during installation""" -class UninstallationError(PipError): - """General exception during uninstallation""" +class MissingPyProjectBuildRequires(DiagnosticPipError): + """Raised when pyproject.toml has `build-system`, but no `build-system.requires`.""" + + reference = "missing-pyproject-build-system-requires" + + def __init__(self, *, package: str) -> None: + super().__init__( + message=f"Can not process {escape(package)}", + context=Text( + "This package has an invalid pyproject.toml file.\n" + "The [build-system] table is missing the mandatory `requires` key." + ), + note_stmt="This is an issue with the package mentioned above, not pip.", + hint_stmt=Text("See PEP 518 for the detailed specification."), + ) + + +class InvalidPyProjectBuildRequires(DiagnosticPipError): + """Raised when pyproject.toml an invalid `build-system.requires`.""" + + reference = "invalid-pyproject-build-system-requires" + + def __init__(self, *, package: str, reason: str) -> None: + super().__init__( + message=f"Can not process {escape(package)}", + context=Text( + "This package has an invalid `build-system.requires` key in " + f"pyproject.toml.\n{reason}" + ), + note_stmt="This is an issue with the package mentioned above, not pip.", + hint_stmt=Text("See PEP 518 for the detailed specification."), + ) class NoneMetadataError(PipError): - """ - Raised when accessing "METADATA" or "PKG-INFO" metadata for a - pip._vendor.pkg_resources.Distribution object and - `dist.has_metadata('METADATA')` returns True but - `dist.get_metadata('METADATA')` returns None (and similarly for - "PKG-INFO"). + """Raised when accessing a Distribution's "METADATA" or "PKG-INFO". + + This signifies an inconsistency, when the Distribution claims to have + the metadata file (if not, raise ``FileNotFoundError`` instead), but is + not actually able to produce its content. This may be due to permission + errors. """ - def __init__(self, dist, metadata_name): - # type: (Distribution, str) -> None + def __init__( + self, + dist: "BaseDistribution", + metadata_name: str, + ) -> None: """ :param dist: A Distribution object. :param metadata_name: The name of the metadata being accessed @@ -52,15 +240,23 @@ def __init__(self, dist, metadata_name): self.dist = dist self.metadata_name = metadata_name - def __str__(self): - # type: () -> str + def __str__(self) -> str: # Use `dist` in the error message because its stringification # includes more information, like the version and location. - return ( - 'None {} metadata found for distribution: {}'.format( - self.metadata_name, self.dist, - ) - ) + return f"None {self.metadata_name} metadata found for distribution: {self.dist}" + + +class UserInstallationInvalid(InstallationError): + """A --user install is requested on an environment without user site.""" + + def __str__(self) -> str: + return "User base directory is not specified" + + +class InvalidSchemeCombination(InstallationError): + def __str__(self) -> str: + before = ", ".join(str(a) for a in self.args[:-1]) + return f"Cannot set {before} and {self.args[-1]} together" class DistributionNotFound(InstallationError): @@ -88,6 +284,34 @@ class PreviousBuildDirError(PipError): """Raised when there's a previous conflicting build directory""" +class NetworkConnectionError(PipError): + """HTTP connection error""" + + def __init__( + self, + error_msg: str, + response: Optional["Response"] = None, + request: Optional["Request"] = None, + ) -> None: + """ + Initialize NetworkConnectionError with `request` and `response` + objects. + """ + self.response = response + self.request = request + self.error_msg = error_msg + if ( + self.response is not None + and not self.request + and hasattr(response, "request") + ): + self.request = self.response.request + super().__init__(error_msg, response, request) + + def __str__(self) -> str: + return str(self.error_msg) + + class InvalidWheelFilename(InstallationError): """Invalid wheel filename.""" @@ -96,30 +320,133 @@ class UnsupportedWheel(InstallationError): """Unsupported wheel.""" +class InvalidWheel(InstallationError): + """Invalid (e.g. corrupt) wheel.""" + + def __init__(self, location: str, name: str): + self.location = location + self.name = name + + def __str__(self) -> str: + return f"Wheel '{self.name}' located at {self.location} is invalid." + + +class MetadataInconsistent(InstallationError): + """Built metadata contains inconsistent information. + + This is raised when the metadata contains values (e.g. name and version) + that do not match the information previously obtained from sdist filename, + user-supplied ``#egg=`` value, or an install requirement name. + """ + + def __init__( + self, ireq: "InstallRequirement", field: str, f_val: str, m_val: str + ) -> None: + self.ireq = ireq + self.field = field + self.f_val = f_val + self.m_val = m_val + + def __str__(self) -> str: + return ( + f"Requested {self.ireq} has inconsistent {self.field}: " + f"expected {self.f_val!r}, but metadata has {self.m_val!r}" + ) + + +class MetadataInvalid(InstallationError): + """Metadata is invalid.""" + + def __init__(self, ireq: "InstallRequirement", error: str) -> None: + self.ireq = ireq + self.error = error + + def __str__(self) -> str: + return f"Requested {self.ireq} has invalid metadata: {self.error}" + + +class InstallationSubprocessError(DiagnosticPipError, InstallationError): + """A subprocess call failed.""" + + reference = "subprocess-exited-with-error" + + def __init__( + self, + *, + command_description: str, + exit_code: int, + output_lines: Optional[List[str]], + ) -> None: + if output_lines is None: + output_prompt = Text("See above for output.") + else: + output_prompt = ( + Text.from_markup(f"[red][{len(output_lines)} lines of output][/]\n") + + Text("".join(output_lines)) + + Text.from_markup(R"[red]\[end of output][/]") + ) + + super().__init__( + message=( + f"[green]{escape(command_description)}[/] did not run successfully.\n" + f"exit code: {exit_code}" + ), + context=output_prompt, + hint_stmt=None, + note_stmt=( + "This error originates from a subprocess, and is likely not a " + "problem with pip." + ), + ) + + self.command_description = command_description + self.exit_code = exit_code + + def __str__(self) -> str: + return f"{self.command_description} exited with {self.exit_code}" + + +class MetadataGenerationFailed(InstallationSubprocessError, InstallationError): + reference = "metadata-generation-failed" + + def __init__( + self, + *, + package_details: str, + ) -> None: + super(InstallationSubprocessError, self).__init__( + message="Encountered error while generating package metadata.", + context=escape(package_details), + hint_stmt="See above for details.", + note_stmt="This is an issue with the package mentioned above, not pip.", + ) + + def __str__(self) -> str: + return "metadata generation failed" + + class HashErrors(InstallationError): """Multiple HashError instances rolled into one for reporting""" - def __init__(self): - self.errors = [] + def __init__(self) -> None: + self.errors: List["HashError"] = [] - def append(self, error): + def append(self, error: "HashError") -> None: self.errors.append(error) - def __str__(self): + def __str__(self) -> str: lines = [] self.errors.sort(key=lambda e: e.order) for cls, errors_of_cls in groupby(self.errors, lambda e: e.__class__): lines.append(cls.head) lines.extend(e.body() for e in errors_of_cls) if lines: - return '\n'.join(lines) + return "\n".join(lines) + return "" - def __nonzero__(self): + def __bool__(self) -> bool: return bool(self.errors) - def __bool__(self): - return self.__nonzero__() - class HashError(InstallationError): """ @@ -137,10 +464,12 @@ class HashError(InstallationError): typically available earlier. """ - req = None # type: Optional[InstallRequirement] - head = '' - def body(self): + req: Optional["InstallRequirement"] = None + head = "" + order: int = -1 + + def body(self) -> str: """Return a summary of me for display under the heading. This default implementation simply prints a description of the @@ -150,19 +479,19 @@ def body(self): its link already populated by the resolver's _populate_link(). """ - return ' {}'.format(self._requirement_name()) + return f" {self._requirement_name()}" - def __str__(self): - return '{}\n{}'.format(self.head, self.body()) + def __str__(self) -> str: + return f"{self.head}\n{self.body()}" - def _requirement_name(self): + def _requirement_name(self) -> str: """Return a description of the requirement that triggered me. This default implementation returns long description of the req, with line numbers """ - return str(self.req) if self.req else 'unknown package' + return str(self.req) if self.req else "unknown package" class VcsHashUnsupported(HashError): @@ -170,8 +499,10 @@ class VcsHashUnsupported(HashError): we don't have a method for hashing those.""" order = 0 - head = ("Can't verify hashes for these requirements because we don't " - "have a way to hash version control repositories:") + head = ( + "Can't verify hashes for these requirements because we don't " + "have a way to hash version control repositories:" + ) class DirectoryUrlHashUnsupported(HashError): @@ -179,45 +510,52 @@ class DirectoryUrlHashUnsupported(HashError): we don't have a method for hashing those.""" order = 1 - head = ("Can't verify hashes for these file:// requirements because they " - "point to directories:") + head = ( + "Can't verify hashes for these file:// requirements because they " + "point to directories:" + ) class HashMissing(HashError): """A hash was needed for a requirement but is absent.""" order = 2 - head = ('Hashes are required in --require-hashes mode, but they are ' - 'missing from some requirements. Here is a list of those ' - 'requirements along with the hashes their downloaded archives ' - 'actually had. Add lines like these to your requirements files to ' - 'prevent tampering. (If you did not enable --require-hashes ' - 'manually, note that it turns on automatically when any package ' - 'has a hash.)') - - def __init__(self, gotten_hash): + head = ( + "Hashes are required in --require-hashes mode, but they are " + "missing from some requirements. Here is a list of those " + "requirements along with the hashes their downloaded archives " + "actually had. Add lines like these to your requirements files to " + "prevent tampering. (If you did not enable --require-hashes " + "manually, note that it turns on automatically when any package " + "has a hash.)" + ) + + def __init__(self, gotten_hash: str) -> None: """ :param gotten_hash: The hash of the (possibly malicious) archive we just downloaded """ self.gotten_hash = gotten_hash - def body(self): + def body(self) -> str: # Dodge circular import. - from fetchcode.vcs.pip._internal.utils.hashes import FAVORITE_HASH + from pip._internal.utils.hashes import FAVORITE_HASH package = None if self.req: # In the case of URL-based requirements, display the original URL # seen in the requirements file rather than the package name, # so the output can be directly copied into the requirements file. - package = (self.req.original_link if self.req.original_link - # In case someone feeds something downright stupid - # to InstallRequirement's constructor. - else getattr(self.req, 'req', None)) - return ' {} --hash={}:{}'.format(package or 'unknown package', - FAVORITE_HASH, - self.gotten_hash) + package = ( + self.req.original_link + if self.req.is_direct + # In case someone feeds something downright stupid + # to InstallRequirement's constructor. + else getattr(self.req, "req", None) + ) + return " {} --hash={}:{}".format( + package or "unknown package", FAVORITE_HASH, self.gotten_hash + ) class HashUnpinned(HashError): @@ -225,8 +563,10 @@ class HashUnpinned(HashError): version.""" order = 3 - head = ('In --require-hashes mode, all requirements must have their ' - 'versions pinned with ==. These do not:') + head = ( + "In --require-hashes mode, all requirements must have their " + "versions pinned with ==. These do not:" + ) class HashMismatch(HashError): @@ -238,13 +578,16 @@ class HashMismatch(HashError): improve its error message. """ - order = 4 - head = ('THESE PACKAGES DO NOT MATCH THE HASHES FROM THE REQUIREMENTS ' - 'FILE. If you have updated the package versions, please update ' - 'the hashes. Otherwise, examine the package contents carefully; ' - 'someone may have tampered with them.') - def __init__(self, allowed, gots): + order = 4 + head = ( + "THESE PACKAGES DO NOT MATCH THE HASHES FROM THE REQUIREMENTS " + "FILE. If you have updated the package versions, please update " + "the hashes. Otherwise, examine the package contents carefully; " + "someone may have tampered with them." + ) + + def __init__(self, allowed: Dict[str, List[str]], gots: Dict[str, "_Hash"]) -> None: """ :param allowed: A dict of algorithm names pointing to lists of allowed hex digests @@ -254,11 +597,10 @@ def __init__(self, allowed, gots): self.allowed = allowed self.gots = gots - def body(self): - return ' {}:\n{}'.format(self._requirement_name(), - self._hash_comparison()) + def body(self) -> str: + return f" {self._requirement_name()}:\n{self._hash_comparison()}" - def _hash_comparison(self): + def _hash_comparison(self) -> str: """ Return a comparison of actual and expected hash values. @@ -269,19 +611,20 @@ def _hash_comparison(self): Got bcdefbcdefbcdefbcdefbcdefbcdefbcdefbcdefbcdef """ - def hash_then_or(hash_name): + + def hash_then_or(hash_name: str) -> "chain[str]": # For now, all the decent hashes have 6-char names, so we can get # away with hard-coding space literals. - return chain([hash_name], repeat(' or')) + return chain([hash_name], repeat(" or")) - lines = [] - for hash_name, expecteds in iteritems(self.allowed): + lines: List[str] = [] + for hash_name, expecteds in self.allowed.items(): prefix = hash_then_or(hash_name) - lines.extend((' Expected {} {}'.format(next(prefix), e)) - for e in expecteds) - lines.append(' Got {}\n'.format( - self.gots[hash_name].hexdigest())) - return '\n'.join(lines) + lines.extend((f" Expected {next(prefix)} {e}") for e in expecteds) + lines.append( + f" Got {self.gots[hash_name].hexdigest()}\n" + ) + return "\n".join(lines) class UnsupportedPythonVersion(InstallationError): @@ -290,19 +633,145 @@ class UnsupportedPythonVersion(InstallationError): class ConfigurationFileCouldNotBeLoaded(ConfigurationError): - """When there are errors while loading a configuration file - """ - - def __init__(self, reason="could not be loaded", fname=None, error=None): - super(ConfigurationFileCouldNotBeLoaded, self).__init__(error) + """When there are errors while loading a configuration file""" + + def __init__( + self, + reason: str = "could not be loaded", + fname: Optional[str] = None, + error: Optional[configparser.Error] = None, + ) -> None: + super().__init__(error) self.reason = reason self.fname = fname self.error = error - def __str__(self): + def __str__(self) -> str: if self.fname is not None: - message_part = " in {}.".format(self.fname) + message_part = f" in {self.fname}." else: assert self.error is not None - message_part = ".\n{}\n".format(self.error.message) - return "Configuration file {}{}".format(self.reason, message_part) + message_part = f".\n{self.error}\n" + return f"Configuration file {self.reason}{message_part}" + + +_DEFAULT_EXTERNALLY_MANAGED_ERROR = f"""\ +The Python environment under {sys.prefix} is managed externally, and may not be +manipulated by the user. Please use specific tooling from the distributor of +the Python installation to interact with this environment instead. +""" + + +class ExternallyManagedEnvironment(DiagnosticPipError): + """The current environment is externally managed. + + This is raised when the current environment is externally managed, as + defined by `PEP 668`_. The ``EXTERNALLY-MANAGED`` configuration is checked + and displayed when the error is bubbled up to the user. + + :param error: The error message read from ``EXTERNALLY-MANAGED``. + """ + + reference = "externally-managed-environment" + + def __init__(self, error: Optional[str]) -> None: + if error is None: + context = Text(_DEFAULT_EXTERNALLY_MANAGED_ERROR) + else: + context = Text(error) + super().__init__( + message="This environment is externally managed", + context=context, + note_stmt=( + "If you believe this is a mistake, please contact your " + "Python installation or OS distribution provider. " + "You can override this, at the risk of breaking your Python " + "installation or OS, by passing --break-system-packages." + ), + hint_stmt=Text("See PEP 668 for the detailed specification."), + ) + + @staticmethod + def _iter_externally_managed_error_keys() -> Iterator[str]: + # LC_MESSAGES is in POSIX, but not the C standard. The most common + # platform that does not implement this category is Windows, where + # using other categories for console message localization is equally + # unreliable, so we fall back to the locale-less vendor message. This + # can always be re-evaluated when a vendor proposes a new alternative. + try: + category = locale.LC_MESSAGES + except AttributeError: + lang: Optional[str] = None + else: + lang, _ = locale.getlocale(category) + if lang is not None: + yield f"Error-{lang}" + for sep in ("-", "_"): + before, found, _ = lang.partition(sep) + if not found: + continue + yield f"Error-{before}" + yield "Error" + + @classmethod + def from_config( + cls, + config: Union[pathlib.Path, str], + ) -> "ExternallyManagedEnvironment": + parser = configparser.ConfigParser(interpolation=None) + try: + parser.read(config, encoding="utf-8") + section = parser["externally-managed"] + for key in cls._iter_externally_managed_error_keys(): + with contextlib.suppress(KeyError): + return cls(section[key]) + except KeyError: + pass + except (OSError, UnicodeDecodeError, configparser.ParsingError): + from pip._internal.utils._log import VERBOSE + + exc_info = logger.isEnabledFor(VERBOSE) + logger.warning("Failed to read %s", config, exc_info=exc_info) + return cls(None) + + +class UninstallMissingRecord(DiagnosticPipError): + reference = "uninstall-no-record-file" + + def __init__(self, *, distribution: "BaseDistribution") -> None: + installer = distribution.installer + if not installer or installer == "pip": + dep = f"{distribution.raw_name}=={distribution.version}" + hint = Text.assemble( + "You might be able to recover from this via: ", + (f"pip install --force-reinstall --no-deps {dep}", "green"), + ) + else: + hint = Text( + f"The package was installed by {installer}. " + "You should check if it can uninstall the package." + ) + + super().__init__( + message=Text(f"Cannot uninstall {distribution}"), + context=( + "The package's contents are unknown: " + f"no RECORD file was found for {distribution.raw_name}." + ), + hint_stmt=hint, + ) + + +class LegacyDistutilsInstall(DiagnosticPipError): + reference = "uninstall-distutils-installed-package" + + def __init__(self, *, distribution: "BaseDistribution") -> None: + super().__init__( + message=Text(f"Cannot uninstall {distribution}"), + context=( + "It is a distutils installed project and thus we cannot accurately " + "determine which files belong to it which would lead to only a partial " + "uninstall." + ), + hint_stmt=None, + ) diff --git a/src/fetchcode/vcs/pip/_internal/index/collector.py b/src/fetchcode/vcs/pip/_internal/index/collector.py index 23365621..5f8fdee3 100644 --- a/src/fetchcode/vcs/pip/_internal/index/collector.py +++ b/src/fetchcode/vcs/pip/_internal/index/collector.py @@ -1,155 +1,143 @@ """ -The main purpose of this module is to expose LinkCollector.collect_links(). +The main purpose of this module is to expose LinkCollector.collect_sources(). """ -import cgi +import collections +import email.message import functools import itertools +import json import logging -import mimetypes import os -import re -from collections import OrderedDict - -from fetchcode.vcs.pip._vendor import html5lib, requests -from fetchcode.vcs.pip._vendor.distlib.compat import unescape -from fetchcode.vcs.pip._vendor.requests.exceptions import HTTPError, RetryError, SSLError -from fetchcode.vcs.pip._vendor.six.moves.urllib import parse as urllib_parse -from fetchcode.vcs.pip._vendor.six.moves.urllib import request as urllib_request - -from fetchcode.vcs.pip._internal.models.link import Link -from fetchcode.vcs.pip._internal.utils.filetypes import ARCHIVE_EXTENSIONS -from fetchcode.vcs.pip._internal.utils.misc import pairwise, redact_auth_from_url -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.utils.urls import path_to_url, url_to_path -from fetchcode.vcs.pip._internal.vcs import is_url, vcs - -if MYPY_CHECK_RUNNING: - from typing import ( - Callable, Iterable, List, MutableMapping, Optional, - Protocol, Sequence, Tuple, TypeVar, Union, - ) - import xml.etree.ElementTree - - from fetchcode.vcs.pip._vendor.requests import Response - - from fetchcode.vcs.pip._internal.models.search_scope import SearchScope - from fetchcode.vcs.pip._internal.network.session import PipSession - - HTMLElement = xml.etree.ElementTree.Element - ResponseHeaders = MutableMapping[str, str] - - # Used in the @lru_cache polyfill. - F = TypeVar('F') - - class LruCache(Protocol): - def __call__(self, maxsize=None): - # type: (Optional[int]) -> Callable[[F], F] - raise NotImplementedError - +import urllib.parse +import urllib.request +from dataclasses import dataclass +from html.parser import HTMLParser +from optparse import Values +from typing import ( + Callable, + Dict, + Iterable, + List, + MutableMapping, + NamedTuple, + Optional, + Protocol, + Sequence, + Tuple, + Union, +) + +from pip._vendor import requests +from pip._vendor.requests import Response +from pip._vendor.requests.exceptions import RetryError, SSLError + +from pip._internal.exceptions import NetworkConnectionError +from pip._internal.models.link import Link +from pip._internal.models.search_scope import SearchScope +from pip._internal.network.session import PipSession +from pip._internal.network.utils import raise_for_status +from pip._internal.utils.filetypes import is_archive_file +from pip._internal.utils.misc import redact_auth_from_url +from pip._internal.vcs import vcs + +from .sources import CandidatesFromPage, LinkSource, build_source logger = logging.getLogger(__name__) - -# Fallback to noop_lru_cache in Python 2 -# TODO: this can be removed when python 2 support is dropped! -def noop_lru_cache(maxsize=None): - # type: (Optional[int]) -> Callable[[F], F] - def _wrapper(f): - # type: (F) -> F - return f - return _wrapper +ResponseHeaders = MutableMapping[str, str] -_lru_cache = getattr(functools, "lru_cache", noop_lru_cache) # type: LruCache - - -def _match_vcs_scheme(url): - # type: (str) -> Optional[str] +def _match_vcs_scheme(url: str) -> Optional[str]: """Look for VCS schemes in the URL. Returns the matched VCS scheme, or None if there's no match. """ for scheme in vcs.schemes: - if url.lower().startswith(scheme) and url[len(scheme)] in '+:': + if url.lower().startswith(scheme) and url[len(scheme)] in "+:": return scheme return None -def _is_url_like_archive(url): - # type: (str) -> bool - """Return whether the URL looks like an archive. - """ - filename = Link(url).filename - for bad_ext in ARCHIVE_EXTENSIONS: - if filename.endswith(bad_ext): - return True - return False - - -class _NotHTML(Exception): - def __init__(self, content_type, request_desc): - # type: (str, str) -> None - super(_NotHTML, self).__init__(content_type, request_desc) +class _NotAPIContent(Exception): + def __init__(self, content_type: str, request_desc: str) -> None: + super().__init__(content_type, request_desc) self.content_type = content_type self.request_desc = request_desc -def _ensure_html_header(response): - # type: (Response) -> None - """Check the Content-Type header to ensure the response contains HTML. +def _ensure_api_header(response: Response) -> None: + """ + Check the Content-Type header to ensure the response contains a Simple + API Response. - Raises `_NotHTML` if the content type is not text/html. + Raises `_NotAPIContent` if the content type is not a valid content-type. """ - content_type = response.headers.get("Content-Type", "") - if not content_type.lower().startswith("text/html"): - raise _NotHTML(content_type, response.request.method) + content_type = response.headers.get("Content-Type", "Unknown") + + content_type_l = content_type.lower() + if content_type_l.startswith( + ( + "text/html", + "application/vnd.pypi.simple.v1+html", + "application/vnd.pypi.simple.v1+json", + ) + ): + return + + raise _NotAPIContent(content_type, response.request.method) class _NotHTTP(Exception): pass -def _ensure_html_response(url, session): - # type: (str, PipSession) -> None - """Send a HEAD request to the URL, and ensure the response contains HTML. +def _ensure_api_response(url: str, session: PipSession) -> None: + """ + Send a HEAD request to the URL, and ensure the response contains a simple + API Response. Raises `_NotHTTP` if the URL is not available for a HEAD request, or - `_NotHTML` if the content type is not text/html. + `_NotAPIContent` if the content type is not a valid content type. """ - scheme, netloc, path, query, fragment = urllib_parse.urlsplit(url) - if scheme not in {'http', 'https'}: + scheme, netloc, path, query, fragment = urllib.parse.urlsplit(url) + if scheme not in {"http", "https"}: raise _NotHTTP() resp = session.head(url, allow_redirects=True) - resp.raise_for_status() + raise_for_status(resp) - _ensure_html_header(resp) + _ensure_api_header(resp) -def _get_html_response(url, session): - # type: (str, PipSession) -> Response - """Access an HTML page with GET, and return the response. +def _get_simple_response(url: str, session: PipSession) -> Response: + """Access an Simple API response with GET, and return the response. This consists of three parts: 1. If the URL looks suspiciously like an archive, send a HEAD first to - check the Content-Type is HTML, to avoid downloading a large file. - Raise `_NotHTTP` if the content type cannot be determined, or - `_NotHTML` if it is not HTML. + check the Content-Type is HTML or Simple API, to avoid downloading a + large file. Raise `_NotHTTP` if the content type cannot be determined, or + `_NotAPIContent` if it is not HTML or a Simple API. 2. Actually perform the request. Raise HTTP exceptions on network failures. - 3. Check the Content-Type header to make sure we got HTML, and raise - `_NotHTML` otherwise. + 3. Check the Content-Type header to make sure we got a Simple API response, + and raise `_NotAPIContent` otherwise. """ - if _is_url_like_archive(url): - _ensure_html_response(url, session=session) + if is_archive_file(Link(url).filename): + _ensure_api_response(url, session=session) - logger.debug('Getting page %s', redact_auth_from_url(url)) + logger.debug("Getting page %s", redact_auth_from_url(url)) resp = session.get( url, headers={ - "Accept": "text/html", + "Accept": ", ".join( + [ + "application/vnd.pypi.simple.v1+json", + "application/vnd.pypi.simple.v1+html; q=0.1", + "text/html; q=0.01", + ] + ), # We don't want to blindly returned cached data for # /simple/, because authors generally expecting that # twine upload && pip install will function, but if @@ -166,181 +154,65 @@ def _get_html_response(url, session): "Cache-Control": "max-age=0", }, ) - resp.raise_for_status() + raise_for_status(resp) # The check for archives above only works if the url ends with # something that looks like an archive. However that is not a # requirement of an url. Unless we issue a HEAD request on every - # url we cannot know ahead of time for sure if something is HTML - # or not. However we can check after we've downloaded it. - _ensure_html_header(resp) + # url we cannot know ahead of time for sure if something is a + # Simple API response or not. However we can check after we've + # downloaded it. + _ensure_api_header(resp) + + logger.debug( + "Fetched page %s as %s", + redact_auth_from_url(url), + resp.headers.get("Content-Type", "Unknown"), + ) return resp -def _get_encoding_from_headers(headers): - # type: (ResponseHeaders) -> Optional[str] - """Determine if we have any encoding information in our headers. - """ +def _get_encoding_from_headers(headers: ResponseHeaders) -> Optional[str]: + """Determine if we have any encoding information in our headers.""" if headers and "Content-Type" in headers: - content_type, params = cgi.parse_header(headers["Content-Type"]) - if "charset" in params: - return params['charset'] + m = email.message.Message() + m["content-type"] = headers["Content-Type"] + charset = m.get_param("charset") + if charset: + return str(charset) return None -def _determine_base_url(document, page_url): - # type: (HTMLElement, str) -> str - """Determine the HTML document's base URL. - - This looks for a ```` tag in the HTML document. If present, its href - attribute denotes the base URL of anchor tags in the document. If there is - no such tag (or if it does not have a valid href attribute), the HTML - file's URL is used as the base URL. - - :param document: An HTML document representation. The current - implementation expects the result of ``html5lib.parse()``. - :param page_url: The URL of the HTML document. - """ - for base in document.findall(".//base"): - href = base.get("href") - if href is not None: - return href - return page_url - - -def _clean_url_path_part(part): - # type: (str) -> str - """ - Clean a "part" of a URL path (i.e. after splitting on "@" characters). - """ - # We unquote prior to quoting to make sure nothing is double quoted. - return urllib_parse.quote(urllib_parse.unquote(part)) - - -def _clean_file_url_path(part): - # type: (str) -> str - """ - Clean the first part of a URL path that corresponds to a local - filesystem path (i.e. the first part after splitting on "@" characters). - """ - # We unquote prior to quoting to make sure nothing is double quoted. - # Also, on Windows the path part might contain a drive letter which - # should not be quoted. On Linux where drive letters do not - # exist, the colon should be quoted. We rely on urllib.request - # to do the right thing here. - return urllib_request.pathname2url(urllib_request.url2pathname(part)) - - -# percent-encoded: / -_reserved_chars_re = re.compile('(@|%2F)', re.IGNORECASE) - - -def _clean_url_path(path, is_local_path): - # type: (str, bool) -> str - """ - Clean the path portion of a URL. - """ - if is_local_path: - clean_func = _clean_file_url_path - else: - clean_func = _clean_url_path_part - - # Split on the reserved characters prior to cleaning so that - # revision strings in VCS URLs are properly preserved. - parts = _reserved_chars_re.split(path) - - cleaned_parts = [] - for to_clean, reserved in pairwise(itertools.chain(parts, [''])): - cleaned_parts.append(clean_func(to_clean)) - # Normalize %xx escapes (e.g. %2f -> %2F) - cleaned_parts.append(reserved.upper()) - - return ''.join(cleaned_parts) - - -def _clean_link(url): - # type: (str) -> str - """ - Make sure a link is fully quoted. - For example, if ' ' occurs in the URL, it will be replaced with "%20", - and without double-quoting other characters. - """ - # Split the URL into parts according to the general structure - # `scheme://netloc/path;parameters?query#fragment`. - result = urllib_parse.urlparse(url) - # If the netloc is empty, then the URL refers to a local filesystem path. - is_local_path = not result.netloc - path = _clean_url_path(result.path, is_local_path=is_local_path) - return urllib_parse.urlunparse(result._replace(path=path)) - - -def _create_link_from_element( - anchor, # type: HTMLElement - page_url, # type: str - base_url, # type: str -): - # type: (...) -> Optional[Link] - """ - Convert an anchor element in a simple repository page to a Link. - """ - href = anchor.get("href") - if not href: - return None - - url = _clean_link(urllib_parse.urljoin(base_url, href)) - pyrequire = anchor.get('data-requires-python') - pyrequire = unescape(pyrequire) if pyrequire else None - - yanked_reason = anchor.get('data-yanked') - if yanked_reason: - # This is a unicode string in Python 2 (and 3). - yanked_reason = unescape(yanked_reason) - - link = Link( - url, - comes_from=page_url, - requires_python=pyrequire, - yanked_reason=yanked_reason, - ) - - return link - - -class CacheablePageContent(object): - def __init__(self, page): - # type: (HTMLPage) -> None +class CacheablePageContent: + def __init__(self, page: "IndexContent") -> None: assert page.cache_link_parsing self.page = page - def __eq__(self, other): - # type: (object) -> bool - return (isinstance(other, type(self)) and - self.page.url == other.page.url) + def __eq__(self, other: object) -> bool: + return isinstance(other, type(self)) and self.page.url == other.page.url - def __hash__(self): - # type: () -> int + def __hash__(self) -> int: return hash(self.page.url) -def with_cached_html_pages( - fn, # type: Callable[[HTMLPage], Iterable[Link]] -): - # type: (...) -> Callable[[HTMLPage], List[Link]] +class ParseLinks(Protocol): + def __call__(self, page: "IndexContent") -> Iterable[Link]: ... + + +def with_cached_index_content(fn: ParseLinks) -> ParseLinks: """ - Given a function that parses an Iterable[Link] from an HTMLPage, cache the - function's result (keyed by CacheablePageContent), unless the HTMLPage + Given a function that parses an Iterable[Link] from an IndexContent, cache the + function's result (keyed by CacheablePageContent), unless the IndexContent `page` has `page.cache_link_parsing == False`. """ - @_lru_cache(maxsize=None) - def wrapper(cacheable_page): - # type: (CacheablePageContent) -> List[Link] + @functools.lru_cache(maxsize=None) + def wrapper(cacheable_page: CacheablePageContent) -> List[Link]: return list(fn(cacheable_page.page)) @functools.wraps(fn) - def wrapper_wrapper(page): - # type: (HTMLPage) -> List[Link] + def wrapper_wrapper(page: "IndexContent") -> List[Link]: if page.cache_link_parsing: return wrapper(CacheablePageContent(page)) return list(fn(page)) @@ -348,314 +220,275 @@ def wrapper_wrapper(page): return wrapper_wrapper -@with_cached_html_pages -def parse_links(page): - # type: (HTMLPage) -> Iterable[Link] +@with_cached_index_content +def parse_links(page: "IndexContent") -> Iterable[Link]: """ - Parse an HTML document, and yield its anchor elements as Link objects. + Parse a Simple API's Index Content, and yield its anchor elements as Link objects. """ - document = html5lib.parse( - page.content, - transport_encoding=page.encoding, - namespaceHTMLElements=False, - ) + + content_type_l = page.content_type.lower() + if content_type_l.startswith("application/vnd.pypi.simple.v1+json"): + data = json.loads(page.content) + for file in data.get("files", []): + link = Link.from_json(file, page.url) + if link is None: + continue + yield link + return + + parser = HTMLLinkParser(page.url) + encoding = page.encoding or "utf-8" + parser.feed(page.content.decode(encoding)) url = page.url - base_url = _determine_base_url(document, url) - for anchor in document.findall(".//a"): - link = _create_link_from_element( - anchor, - page_url=url, - base_url=base_url, - ) + base_url = parser.base_url or url + for anchor in parser.anchors: + link = Link.from_element(anchor, page_url=url, base_url=base_url) if link is None: continue yield link -class HTMLPage(object): - """Represents one page, along with its URL""" +@dataclass(frozen=True) +class IndexContent: + """Represents one response (or page), along with its URL. - def __init__( - self, - content, # type: bytes - encoding, # type: Optional[str] - url, # type: str - cache_link_parsing=True, # type: bool - ): - # type: (...) -> None - """ - :param encoding: the encoding to decode the given content. - :param url: the URL from which the HTML was downloaded. - :param cache_link_parsing: whether links parsed from this page's url - should be cached. PyPI index urls should - have this set to False, for example. - """ - self.content = content - self.encoding = encoding - self.url = url - self.cache_link_parsing = cache_link_parsing + :param encoding: the encoding to decode the given content. + :param url: the URL from which the HTML was downloaded. + :param cache_link_parsing: whether links parsed from this page's url + should be cached. PyPI index urls should + have this set to False, for example. + """ - def __str__(self): - # type: () -> str + content: bytes + content_type: str + encoding: Optional[str] + url: str + cache_link_parsing: bool = True + + def __str__(self) -> str: return redact_auth_from_url(self.url) -def _handle_get_page_fail( - link, # type: Link - reason, # type: Union[str, Exception] - meth=None # type: Optional[Callable[..., None]] -): - # type: (...) -> None +class HTMLLinkParser(HTMLParser): + """ + HTMLParser that keeps the first base HREF and a list of all anchor + elements' attributes. + """ + + def __init__(self, url: str) -> None: + super().__init__(convert_charrefs=True) + + self.url: str = url + self.base_url: Optional[str] = None + self.anchors: List[Dict[str, Optional[str]]] = [] + + def handle_starttag(self, tag: str, attrs: List[Tuple[str, Optional[str]]]) -> None: + if tag == "base" and self.base_url is None: + href = self.get_href(attrs) + if href is not None: + self.base_url = href + elif tag == "a": + self.anchors.append(dict(attrs)) + + def get_href(self, attrs: List[Tuple[str, Optional[str]]]) -> Optional[str]: + for name, value in attrs: + if name == "href": + return value + return None + + +def _handle_get_simple_fail( + link: Link, + reason: Union[str, Exception], + meth: Optional[Callable[..., None]] = None, +) -> None: if meth is None: meth = logger.debug meth("Could not fetch URL %s: %s - skipping", link, reason) -def _make_html_page(response, cache_link_parsing=True): - # type: (Response, bool) -> HTMLPage +def _make_index_content( + response: Response, cache_link_parsing: bool = True +) -> IndexContent: encoding = _get_encoding_from_headers(response.headers) - return HTMLPage( + return IndexContent( response.content, + response.headers["Content-Type"], encoding=encoding, url=response.url, - cache_link_parsing=cache_link_parsing) - + cache_link_parsing=cache_link_parsing, + ) -def _get_html_page(link, session=None): - # type: (Link, Optional[PipSession]) -> Optional[HTMLPage] - if session is None: - raise TypeError( - "_get_html_page() missing 1 required keyword argument: 'session'" - ) - url = link.url.split('#', 1)[0] +def _get_index_content(link: Link, *, session: PipSession) -> Optional["IndexContent"]: + url = link.url.split("#", 1)[0] # Check for VCS schemes that do not support lookup as web pages. vcs_scheme = _match_vcs_scheme(url) if vcs_scheme: - logger.debug('Cannot look at %s URL %s', vcs_scheme, link) + logger.warning( + "Cannot look at %s URL %s because it does not support lookup as web pages.", + vcs_scheme, + link, + ) return None # Tack index.html onto file:// URLs that point to directories - scheme, _, path, _, _, _ = urllib_parse.urlparse(url) - if (scheme == 'file' and os.path.isdir(urllib_request.url2pathname(path))): + scheme, _, path, _, _, _ = urllib.parse.urlparse(url) + if scheme == "file" and os.path.isdir(urllib.request.url2pathname(path)): # add trailing slash if not present so urljoin doesn't trim # final segment - if not url.endswith('/'): - url += '/' - url = urllib_parse.urljoin(url, 'index.html') - logger.debug(' file: URL is directory, getting %s', url) + if not url.endswith("/"): + url += "/" + # TODO: In the future, it would be nice if pip supported PEP 691 + # style responses in the file:// URLs, however there's no + # standard file extension for application/vnd.pypi.simple.v1+json + # so we'll need to come up with something on our own. + url = urllib.parse.urljoin(url, "index.html") + logger.debug(" file: URL is directory, getting %s", url) try: - resp = _get_html_response(url, session=session) + resp = _get_simple_response(url, session=session) except _NotHTTP: - logger.debug( - 'Skipping page %s because it looks like an archive, and cannot ' - 'be checked by HEAD.', link, + logger.warning( + "Skipping page %s because it looks like an archive, and cannot " + "be checked by a HTTP HEAD request.", + link, ) - except _NotHTML as exc: - logger.debug( - 'Skipping page %s because the %s request got Content-Type: %s', - link, exc.request_desc, exc.content_type, + except _NotAPIContent as exc: + logger.warning( + "Skipping page %s because the %s request got Content-Type: %s. " + "The only supported Content-Types are application/vnd.pypi.simple.v1+json, " + "application/vnd.pypi.simple.v1+html, and text/html", + link, + exc.request_desc, + exc.content_type, ) - except HTTPError as exc: - _handle_get_page_fail(link, exc) + except NetworkConnectionError as exc: + _handle_get_simple_fail(link, exc) except RetryError as exc: - _handle_get_page_fail(link, exc) + _handle_get_simple_fail(link, exc) except SSLError as exc: reason = "There was a problem confirming the ssl certificate: " reason += str(exc) - _handle_get_page_fail(link, reason, meth=logger.info) + _handle_get_simple_fail(link, reason, meth=logger.info) except requests.ConnectionError as exc: - _handle_get_page_fail(link, "connection error: {}".format(exc)) + _handle_get_simple_fail(link, f"connection error: {exc}") except requests.Timeout: - _handle_get_page_fail(link, "timed out") + _handle_get_simple_fail(link, "timed out") else: - return _make_html_page(resp, - cache_link_parsing=link.cache_link_parsing) + return _make_index_content(resp, cache_link_parsing=link.cache_link_parsing) return None -def _remove_duplicate_links(links): - # type: (Iterable[Link]) -> List[Link] - """ - Return a list of links, with duplicates removed and ordering preserved. - """ - # We preserve the ordering when removing duplicates because we can. - return list(OrderedDict.fromkeys(links)) - - -def group_locations(locations, expand_dir=False): - # type: (Sequence[str], bool) -> Tuple[List[str], List[str]] - """ - Divide a list of locations into two groups: "files" (archives) and "urls." - - :return: A pair of lists (files, urls). - """ - files = [] - urls = [] - - # puts the url for the given file path into the appropriate list - def sort_path(path): - # type: (str) -> None - url = path_to_url(path) - if mimetypes.guess_type(url, strict=False)[0] == 'text/html': - urls.append(url) - else: - files.append(url) - - for url in locations: - - is_local_path = os.path.exists(url) - is_file_url = url.startswith('file:') - - if is_local_path or is_file_url: - if is_local_path: - path = url - else: - path = url_to_path(url) - if os.path.isdir(path): - if expand_dir: - path = os.path.realpath(path) - for item in os.listdir(path): - sort_path(os.path.join(path, item)) - elif is_file_url: - urls.append(url) - else: - logger.warning( - "Path '{0}' is ignored: " - "it is a directory.".format(path), - ) - elif os.path.isfile(path): - sort_path(path) - else: - logger.warning( - "Url '%s' is ignored: it is neither a file " - "nor a directory.", url, - ) - elif is_url(url): - # Only add url with clear scheme - urls.append(url) - else: - logger.warning( - "Url '%s' is ignored. It is either a non-existing " - "path or lacks a specific scheme.", url, - ) - - return files, urls - - -class CollectedLinks(object): +class CollectedSources(NamedTuple): + find_links: Sequence[Optional[LinkSource]] + index_urls: Sequence[Optional[LinkSource]] - """ - Encapsulates the return value of a call to LinkCollector.collect_links(). - - The return value includes both URLs to project pages containing package - links, as well as individual package Link objects collected from other - sources. - - This info is stored separately as: - - (1) links from the configured file locations, - (2) links from the configured find_links, and - (3) urls to HTML project pages, as described by the PEP 503 simple - repository API. - """ - - def __init__( - self, - files, # type: List[Link] - find_links, # type: List[Link] - project_urls, # type: List[Link] - ): - # type: (...) -> None - """ - :param files: Links from file locations. - :param find_links: Links from find_links. - :param project_urls: URLs to HTML project pages, as described by - the PEP 503 simple repository API. - """ - self.files = files - self.find_links = find_links - self.project_urls = project_urls - - -class LinkCollector(object): +class LinkCollector: """ Responsible for collecting Link objects from all configured locations, making network requests as needed. - The class's main method is its collect_links() method. + The class's main method is its collect_sources() method. """ def __init__( self, - session, # type: PipSession - search_scope, # type: SearchScope - ): - # type: (...) -> None + session: PipSession, + search_scope: SearchScope, + ) -> None: self.search_scope = search_scope self.session = session + @classmethod + def create( + cls, + session: PipSession, + options: Values, + suppress_no_index: bool = False, + ) -> "LinkCollector": + """ + :param session: The Session to use to make requests. + :param suppress_no_index: Whether to ignore the --no-index option + when constructing the SearchScope object. + """ + index_urls = [options.index_url] + options.extra_index_urls + if options.no_index and not suppress_no_index: + logger.debug( + "Ignoring indexes: %s", + ",".join(redact_auth_from_url(url) for url in index_urls), + ) + index_urls = [] + + # Make sure find_links is a list before passing to create(). + find_links = options.find_links or [] + + search_scope = SearchScope.create( + find_links=find_links, + index_urls=index_urls, + no_index=options.no_index, + ) + link_collector = LinkCollector( + session=session, + search_scope=search_scope, + ) + return link_collector + @property - def find_links(self): - # type: () -> List[str] + def find_links(self) -> List[str]: return self.search_scope.find_links - def fetch_page(self, location): - # type: (Link) -> Optional[HTMLPage] + def fetch_response(self, location: Link) -> Optional[IndexContent]: """ Fetch an HTML page containing package links. """ - return _get_html_page(location, session=self.session) - - def collect_links(self, project_name): - # type: (str) -> CollectedLinks - """Find all available links for the given project name. + return _get_index_content(location, session=self.session) - :return: All the Link objects (unfiltered), as a CollectedLinks object. - """ - search_scope = self.search_scope - index_locations = search_scope.get_index_urls_locations(project_name) - index_file_loc, index_url_loc = group_locations(index_locations) - fl_file_loc, fl_url_loc = group_locations( - self.find_links, expand_dir=True, - ) - - file_links = [ - Link(url) for url in itertools.chain(index_file_loc, fl_file_loc) - ] - - # We trust every directly linked archive in find_links - find_link_links = [Link(url, '-f') for url in self.find_links] - - # We trust every url that the user has given us whether it was given - # via --index-url or --find-links. - # We want to filter out anything that does not have a secure origin. - url_locations = [ - link for link in itertools.chain( - # Mark PyPI indices as "cache_link_parsing == False" -- this - # will avoid caching the result of parsing the page for links. - (Link(url, cache_link_parsing=False) for url in index_url_loc), - (Link(url) for url in fl_url_loc), + def collect_sources( + self, + project_name: str, + candidates_from_page: CandidatesFromPage, + ) -> CollectedSources: + # The OrderedDict calls deduplicate sources by URL. + index_url_sources = collections.OrderedDict( + build_source( + loc, + candidates_from_page=candidates_from_page, + page_validator=self.session.is_secure_origin, + expand_dir=False, + cache_link_parsing=False, + project_name=project_name, ) - if self.session.is_secure_origin(link) - ] - - url_locations = _remove_duplicate_links(url_locations) - lines = [ - '{} location(s) to search for versions of {}:'.format( - len(url_locations), project_name, - ), - ] - for link in url_locations: - lines.append('* {}'.format(link)) - logger.debug('\n'.join(lines)) - - return CollectedLinks( - files=file_links, - find_links=find_link_links, - project_urls=url_locations, + for loc in self.search_scope.get_index_urls_locations(project_name) + ).values() + find_links_sources = collections.OrderedDict( + build_source( + loc, + candidates_from_page=candidates_from_page, + page_validator=self.session.is_secure_origin, + expand_dir=True, + cache_link_parsing=True, + project_name=project_name, + ) + for loc in self.find_links + ).values() + + if logger.isEnabledFor(logging.DEBUG): + lines = [ + f"* {s.link}" + for s in itertools.chain(find_links_sources, index_url_sources) + if s is not None and s.link is not None + ] + lines = [ + f"{len(lines)} location(s) to search " + f"for versions of {project_name}:" + ] + lines + logger.debug("\n".join(lines)) + + return CollectedSources( + find_links=list(find_links_sources), + index_urls=list(index_url_sources), ) diff --git a/src/fetchcode/vcs/pip/_internal/index/package_finder.py b/src/fetchcode/vcs/pip/_internal/index/package_finder.py index 58e09b4f..0d65ce35 100644 --- a/src/fetchcode/vcs/pip/_internal/index/package_finder.py +++ b/src/fetchcode/vcs/pip/_internal/index/package_finder.py @@ -1,69 +1,59 @@ """Routines related to PyPI, indexes""" -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False - -from __future__ import absolute_import - +import enum +import functools +import itertools import logging import re +from dataclasses import dataclass +from typing import TYPE_CHECKING, FrozenSet, Iterable, List, Optional, Set, Tuple, Union -from fetchcode.vcs.pip._vendor.packaging import specifiers -from fetchcode.vcs.pip._vendor.packaging.utils import canonicalize_name -from fetchcode.vcs.pip._vendor.packaging.version import parse as parse_version +from pip._vendor.packaging import specifiers +from pip._vendor.packaging.tags import Tag +from pip._vendor.packaging.utils import canonicalize_name +from pip._vendor.packaging.version import InvalidVersion, _BaseVersion +from pip._vendor.packaging.version import parse as parse_version -from fetchcode.vcs.pip._internal.exceptions import ( +from pip._internal.exceptions import ( BestVersionAlreadyInstalled, DistributionNotFound, InvalidWheelFilename, UnsupportedWheel, ) -from fetchcode.vcs.pip._internal.index.collector import parse_links -from fetchcode.vcs.pip._internal.models.candidate import InstallationCandidate -from fetchcode.vcs.pip._internal.models.format_control import FormatControl -from fetchcode.vcs.pip._internal.models.link import Link -from fetchcode.vcs.pip._internal.models.selection_prefs import SelectionPreferences -from fetchcode.vcs.pip._internal.models.target_python import TargetPython -from fetchcode.vcs.pip._internal.models.wheel import Wheel -from fetchcode.vcs.pip._internal.utils.filetypes import WHEEL_EXTENSION -from fetchcode.vcs.pip._internal.utils.logging import indent_log -from fetchcode.vcs.pip._internal.utils.misc import build_netloc -from fetchcode.vcs.pip._internal.utils.packaging import check_requires_python -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.utils.unpacking import SUPPORTED_EXTENSIONS -from fetchcode.vcs.pip._internal.utils.urls import url_to_path - -if MYPY_CHECK_RUNNING: - from typing import ( - FrozenSet, Iterable, List, Optional, Set, Text, Tuple, Union, - ) - - from fetchcode.vcs.pip._vendor.packaging.tags import Tag - from fetchcode.vcs.pip._vendor.packaging.version import _BaseVersion +from pip._internal.index.collector import LinkCollector, parse_links +from pip._internal.models.candidate import InstallationCandidate +from pip._internal.models.format_control import FormatControl +from pip._internal.models.link import Link +from pip._internal.models.search_scope import SearchScope +from pip._internal.models.selection_prefs import SelectionPreferences +from pip._internal.models.target_python import TargetPython +from pip._internal.models.wheel import Wheel +from pip._internal.req import InstallRequirement +from pip._internal.utils._log import getLogger +from pip._internal.utils.filetypes import WHEEL_EXTENSION +from pip._internal.utils.hashes import Hashes +from pip._internal.utils.logging import indent_log +from pip._internal.utils.misc import build_netloc +from pip._internal.utils.packaging import check_requires_python +from pip._internal.utils.unpacking import SUPPORTED_EXTENSIONS - from fetchcode.vcs.pip._internal.index.collector import LinkCollector - from fetchcode.vcs.pip._internal.models.search_scope import SearchScope - from fetchcode.vcs.pip._internal.req import InstallRequirement - from fetchcode.vcs.pip._internal.utils.hashes import Hashes - - BuildTag = Union[Tuple[()], Tuple[int, str]] - CandidateSortingKey = ( - Tuple[int, int, int, _BaseVersion, BuildTag, Optional[int]] - ) +if TYPE_CHECKING: + from pip._vendor.typing_extensions import TypeGuard +__all__ = ["FormatControl", "BestCandidateResult", "PackageFinder"] -__all__ = ['FormatControl', 'BestCandidateResult', 'PackageFinder'] +logger = getLogger(__name__) -logger = logging.getLogger(__name__) +BuildTag = Union[Tuple[()], Tuple[int, str]] +CandidateSortingKey = Tuple[int, int, int, _BaseVersion, Optional[int], BuildTag] def _check_link_requires_python( - link, # type: Link - version_info, # type: Tuple[int, int, int] - ignore_requires_python=False, # type: bool -): - # type: (...) -> bool + link: Link, + version_info: Tuple[int, int, int], + ignore_requires_python: bool = False, +) -> bool: """ Return whether the given Python version is compatible with a link's "Requires-Python" value. @@ -75,39 +65,53 @@ def _check_link_requires_python( """ try: is_compatible = check_requires_python( - link.requires_python, version_info=version_info, + link.requires_python, + version_info=version_info, ) except specifiers.InvalidSpecifier: logger.debug( "Ignoring invalid Requires-Python (%r) for link: %s", - link.requires_python, link, + link.requires_python, + link, ) else: if not is_compatible: - version = '.'.join(map(str, version_info)) + version = ".".join(map(str, version_info)) if not ignore_requires_python: - logger.debug( - 'Link requires a different Python (%s not in: %r): %s', - version, link.requires_python, link, + logger.verbose( + "Link requires a different Python (%s not in: %r): %s", + version, + link.requires_python, + link, ) return False logger.debug( - 'Ignoring failed Requires-Python check (%s not in: %r) ' - 'for link: %s', - version, link.requires_python, link, + "Ignoring failed Requires-Python check (%s not in: %r) for link: %s", + version, + link.requires_python, + link, ) return True -class LinkEvaluator(object): +class LinkType(enum.Enum): + candidate = enum.auto() + different_project = enum.auto() + yanked = enum.auto() + format_unsupported = enum.auto() + format_invalid = enum.auto() + platform_mismatch = enum.auto() + requires_python_mismatch = enum.auto() + +class LinkEvaluator: """ Responsible for evaluating links for a particular project. """ - _py_version_re = re.compile(r'-py([123]\.?[0-9]?)$') + _py_version_re = re.compile(r"-py([123]\.?[0-9]?)$") # Don't include an allow_yanked default value to make sure each call # site considers whether yanked releases are allowed. This also causes @@ -115,14 +119,13 @@ class LinkEvaluator(object): # people when reading the code. def __init__( self, - project_name, # type: str - canonical_name, # type: str - formats, # type: FrozenSet[str] - target_python, # type: TargetPython - allow_yanked, # type: bool - ignore_requires_python=None, # type: Optional[bool] - ): - # type: (...) -> None + project_name: str, + canonical_name: str, + formats: FrozenSet[str], + target_python: TargetPython, + allow_yanked: bool, + ignore_requires_python: Optional[bool] = None, + ) -> None: """ :param project_name: The user supplied package name. :param canonical_name: The canonical package name. @@ -151,23 +154,20 @@ def __init__( self.project_name = project_name - def evaluate_link(self, link): - # type: (Link) -> Tuple[bool, Optional[Text]] + def evaluate_link(self, link: Link) -> Tuple[LinkType, str]: """ Determine whether a link is a candidate for installation. - :return: A tuple (is_candidate, result), where `result` is (1) a - version string if `is_candidate` is True, and (2) if - `is_candidate` is False, an optional string to log the reason - the link fails to qualify. + :return: A tuple (result, detail), where *result* is an enum + representing whether the evaluation found a candidate, or the reason + why one is not found. If a candidate is found, *detail* will be the + candidate's version string; if one is not found, it contains the + reason the link fails to qualify. """ version = None if link.is_yanked and not self._allow_yanked: - reason = link.yanked_reason or '' - # Mark this as a unicode string to prevent "UnicodeEncodeError: - # 'ascii' codec can't encode character" in Python 2 when - # the reason contains non-ascii characters. - return (False, u'yanked for reason: {}'.format(reason)) + reason = link.yanked_reason or "" + return (LinkType.yanked, f"yanked for reason: {reason}") if link.egg_fragment: egg_info = link.egg_fragment @@ -175,79 +175,85 @@ def evaluate_link(self, link): else: egg_info, ext = link.splitext() if not ext: - return (False, 'not a file') + return (LinkType.format_unsupported, "not a file") if ext not in SUPPORTED_EXTENSIONS: - return (False, 'unsupported archive format: {}'.format(ext)) + return ( + LinkType.format_unsupported, + f"unsupported archive format: {ext}", + ) if "binary" not in self._formats and ext == WHEEL_EXTENSION: - reason = 'No binaries permitted for {}'.format( - self.project_name) - return (False, reason) - if "macosx10" in link.path and ext == '.zip': - return (False, 'macosx10 one') + reason = f"No binaries permitted for {self.project_name}" + return (LinkType.format_unsupported, reason) + if "macosx10" in link.path and ext == ".zip": + return (LinkType.format_unsupported, "macosx10 one") if ext == WHEEL_EXTENSION: try: wheel = Wheel(link.filename) except InvalidWheelFilename: - return (False, 'invalid wheel filename') + return ( + LinkType.format_invalid, + "invalid wheel filename", + ) if canonicalize_name(wheel.name) != self._canonical_name: - reason = 'wrong project name (not {})'.format( - self.project_name) - return (False, reason) + reason = f"wrong project name (not {self.project_name})" + return (LinkType.different_project, reason) - supported_tags = self._target_python.get_tags() + supported_tags = self._target_python.get_unsorted_tags() if not wheel.supported(supported_tags): # Include the wheel's tags in the reason string to # simplify troubleshooting compatibility issues. - file_tags = wheel.get_formatted_file_tags() + file_tags = ", ".join(wheel.get_formatted_file_tags()) reason = ( - "none of the wheel's tags match: {}".format( - ', '.join(file_tags) - ) + f"none of the wheel's tags ({file_tags}) are compatible " + f"(run pip debug --verbose to show compatible tags)" ) - return (False, reason) + return (LinkType.platform_mismatch, reason) version = wheel.version # This should be up by the self.ok_binary check, but see issue 2700. if "source" not in self._formats and ext != WHEEL_EXTENSION: - reason = 'No sources permitted for {}'.format(self.project_name) - return (False, reason) + reason = f"No sources permitted for {self.project_name}" + return (LinkType.format_unsupported, reason) if not version: version = _extract_version_from_fragment( - egg_info, self._canonical_name, + egg_info, + self._canonical_name, ) if not version: - reason = 'Missing project version for {}'.format(self.project_name) - return (False, reason) + reason = f"Missing project version for {self.project_name}" + return (LinkType.format_invalid, reason) match = self._py_version_re.search(version) if match: - version = version[:match.start()] + version = version[: match.start()] py_version = match.group(1) if py_version != self._target_python.py_version: - return (False, 'Python version is incorrect') + return ( + LinkType.platform_mismatch, + "Python version is incorrect", + ) supports_python = _check_link_requires_python( - link, version_info=self._target_python.py_version_info, + link, + version_info=self._target_python.py_version_info, ignore_requires_python=self._ignore_requires_python, ) if not supports_python: - # Return None for the reason text to suppress calling - # _log_skipped_link(). - return (False, None) + reason = f"{version} Requires-Python {link.requires_python}" + return (LinkType.requires_python_mismatch, reason) - logger.debug('Found link %s, version: %s', link, version) + logger.debug("Found link %s, version: %s", link, version) - return (True, version) + return (LinkType.candidate, version) def filter_unallowed_hashes( - candidates, # type: List[InstallationCandidate] - hashes, # type: Hashes - project_name, # type: str -): - # type: (...) -> List[InstallationCandidate] + candidates: List[InstallationCandidate], + hashes: Optional[Hashes], + project_name: str, +) -> List[InstallationCandidate]: """ Filter out candidates whose hashes aren't allowed, and return a new list of candidates. @@ -265,8 +271,8 @@ def filter_unallowed_hashes( """ if not hashes: logger.debug( - 'Given no hashes to check %s links for project %r: ' - 'discarding no candidates', + "Given no hashes to check %s links for project %r: " + "discarding no candidates", len(candidates), project_name, ) @@ -296,48 +302,39 @@ def filter_unallowed_hashes( filtered = list(candidates) if len(filtered) == len(candidates): - discard_message = 'discarding no candidates' + discard_message = "discarding no candidates" else: - discard_message = 'discarding {} non-matches:\n {}'.format( + discard_message = "discarding {} non-matches:\n {}".format( len(non_matches), - '\n '.join(str(candidate.link) for candidate in non_matches) + "\n ".join(str(candidate.link) for candidate in non_matches), ) logger.debug( - 'Checked %s links for project %r against %s hashes ' - '(%s matches, %s no digest): %s', + "Checked %s links for project %r against %s hashes " + "(%s matches, %s no digest): %s", len(candidates), project_name, hashes.digest_count, match_count, len(matches_or_no_digest) - match_count, - discard_message + discard_message, ) return filtered -class CandidatePreferences(object): - +@dataclass +class CandidatePreferences: """ Encapsulates some of the preferences for filtering and sorting InstallationCandidate objects. """ - def __init__( - self, - prefer_binary=False, # type: bool - allow_all_prereleases=False, # type: bool - ): - # type: (...) -> None - """ - :param allow_all_prereleases: Whether to allow all pre-releases. - """ - self.allow_all_prereleases = allow_all_prereleases - self.prefer_binary = prefer_binary + prefer_binary: bool = False + allow_all_prereleases: bool = False -class BestCandidateResult(object): +class BestCandidateResult: """A collection of candidates, returned by `PackageFinder.find_best_candidate`. This class is only intended to be instantiated by CandidateEvaluator's @@ -346,11 +343,10 @@ class BestCandidateResult(object): def __init__( self, - candidates, # type: List[InstallationCandidate] - applicable_candidates, # type: List[InstallationCandidate] - best_candidate, # type: Optional[InstallationCandidate] - ): - # type: (...) -> None + candidates: List[InstallationCandidate], + applicable_candidates: List[InstallationCandidate], + best_candidate: Optional[InstallationCandidate], + ) -> None: """ :param candidates: A sequence of all available candidates found. :param applicable_candidates: The applicable candidates. @@ -369,21 +365,16 @@ def __init__( self.best_candidate = best_candidate - def iter_all(self): - # type: () -> Iterable[InstallationCandidate] - """Iterate through all candidates. - """ + def iter_all(self) -> Iterable[InstallationCandidate]: + """Iterate through all candidates.""" return iter(self._candidates) - def iter_applicable(self): - # type: () -> Iterable[InstallationCandidate] - """Iterate through the applicable candidates. - """ + def iter_applicable(self) -> Iterable[InstallationCandidate]: + """Iterate through the applicable candidates.""" return iter(self._applicable_candidates) -class CandidateEvaluator(object): - +class CandidateEvaluator: """ Responsible for filtering and sorting candidates for installation based on what tags are valid. @@ -392,14 +383,13 @@ class CandidateEvaluator(object): @classmethod def create( cls, - project_name, # type: str - target_python=None, # type: Optional[TargetPython] - prefer_binary=False, # type: bool - allow_all_prereleases=False, # type: bool - specifier=None, # type: Optional[specifiers.BaseSpecifier] - hashes=None, # type: Optional[Hashes] - ): - # type: (...) -> CandidateEvaluator + project_name: str, + target_python: Optional[TargetPython] = None, + prefer_binary: bool = False, + allow_all_prereleases: bool = False, + specifier: Optional[specifiers.BaseSpecifier] = None, + hashes: Optional[Hashes] = None, + ) -> "CandidateEvaluator": """Create a CandidateEvaluator object. :param target_python: The target Python interpreter to use when @@ -415,7 +405,7 @@ def create( if specifier is None: specifier = specifiers.SpecifierSet() - supported_tags = target_python.get_tags() + supported_tags = target_python.get_sorted_tags() return cls( project_name=project_name, @@ -428,14 +418,13 @@ def create( def __init__( self, - project_name, # type: str - supported_tags, # type: List[Tag] - specifier, # type: specifiers.BaseSpecifier - prefer_binary=False, # type: bool - allow_all_prereleases=False, # type: bool - hashes=None, # type: Optional[Hashes] - ): - # type: (...) -> None + project_name: str, + supported_tags: List[Tag], + specifier: specifiers.BaseSpecifier, + prefer_binary: bool = False, + allow_all_prereleases: bool = False, + hashes: Optional[Hashes] = None, + ) -> None: """ :param supported_tags: The PEP 425 tags supported by the target Python in order of preference (most preferred first). @@ -446,37 +435,40 @@ def __init__( self._project_name = project_name self._specifier = specifier self._supported_tags = supported_tags + # Since the index of the tag in the _supported_tags list is used + # as a priority, precompute a map from tag to index/priority to be + # used in wheel.find_most_preferred_tag. + self._wheel_tag_preferences = { + tag: idx for idx, tag in enumerate(supported_tags) + } def get_applicable_candidates( self, - candidates, # type: List[InstallationCandidate] - ): - # type: (...) -> List[InstallationCandidate] + candidates: List[InstallationCandidate], + ) -> List[InstallationCandidate]: """ Return the applicable candidates from a list of candidates. """ # Using None infers from the specifier instead. allow_prereleases = self._allow_all_prereleases or None specifier = self._specifier - versions = { - str(v) for v in specifier.filter( - # We turn the version object into a str here because otherwise - # when we're debundled but setuptools isn't, Python will see - # packaging.version.Version and - # pkg_resources._vendor.packaging.version.Version as different - # types. This way we'll use a str as a common data interchange - # format. If we stop using the pkg_resources provided specifier - # and start using our own, we can drop the cast to str(). - (str(c.version) for c in candidates), + + # We turn the version object into a str here because otherwise + # when we're debundled but setuptools isn't, Python will see + # packaging.version.Version and + # pkg_resources._vendor.packaging.version.Version as different + # types. This way we'll use a str as a common data interchange + # format. If we stop using the pkg_resources provided specifier + # and start using our own, we can drop the cast to str(). + candidates_and_versions = [(c, str(c.version)) for c in candidates] + versions = set( + specifier.filter( + (v for _, v in candidates_and_versions), prereleases=allow_prereleases, ) - } - - # Again, converting version to str to deal with debundling. - applicable_candidates = [ - c for c in candidates if str(c.version) in versions - ] + ) + applicable_candidates = [c for c, v in candidates_and_versions if v in versions] filtered_applicable_candidates = filter_unallowed_hashes( candidates=applicable_candidates, hashes=self._hashes, @@ -485,8 +477,7 @@ def get_applicable_candidates( return sorted(filtered_applicable_candidates, key=self._sort_key) - def _sort_key(self, candidate): - # type: (InstallationCandidate) -> CandidateSortingKey + def _sort_key(self, candidate: InstallationCandidate) -> CandidateSortingKey: """ Function to pass as the `key` argument to a call to sorted() to sort InstallationCandidates by preference. @@ -518,22 +509,28 @@ def _sort_key(self, candidate): """ valid_tags = self._supported_tags support_num = len(valid_tags) - build_tag = () # type: BuildTag + build_tag: BuildTag = () binary_preference = 0 link = candidate.link if link.is_wheel: # can raise InvalidWheelFilename wheel = Wheel(link.filename) - if not wheel.supported(valid_tags): + try: + pri = -( + wheel.find_most_preferred_tag( + valid_tags, self._wheel_tag_preferences + ) + ) + except ValueError: raise UnsupportedWheel( - "{} is not a supported wheel for this platform. It " - "can't be sorted.".format(wheel.filename) + f"{wheel.filename} is not a supported wheel for this platform. It " + "can't be sorted." ) if self._prefer_binary: binary_preference = 1 - pri = -(wheel.support_index_min(valid_tags)) if wheel.build_tag is not None: - match = re.match(r'^(\d+)(.*)$', wheel.build_tag) + match = re.match(r"^(\d+)(.*)$", wheel.build_tag) + assert match is not None, "guaranteed by filename validation" build_tag_groups = match.groups() build_tag = (int(build_tag_groups[0]), build_tag_groups[1]) else: # sdist @@ -541,45 +538,31 @@ def _sort_key(self, candidate): has_allowed_hash = int(link.is_hash_allowed(self._hashes)) yank_value = -1 * int(link.is_yanked) # -1 for yanked. return ( - has_allowed_hash, yank_value, binary_preference, candidate.version, - build_tag, pri, + has_allowed_hash, + yank_value, + binary_preference, + candidate.version, + pri, + build_tag, ) def sort_best_candidate( self, - candidates, # type: List[InstallationCandidate] - ): - # type: (...) -> Optional[InstallationCandidate] + candidates: List[InstallationCandidate], + ) -> Optional[InstallationCandidate]: """ Return the best candidate per the instance's sort order, or None if no candidate is acceptable. """ if not candidates: return None - best_candidate = max(candidates, key=self._sort_key) - - # Log a warning per PEP 592 if necessary before returning. - link = best_candidate.link - if link.is_yanked: - reason = link.yanked_reason or '' - msg = ( - # Mark this as a unicode string to prevent - # "UnicodeEncodeError: 'ascii' codec can't encode character" - # in Python 2 when the reason contains non-ascii characters. - u'The candidate selected for download or install is a ' - 'yanked version: {candidate}\n' - 'Reason for being yanked: {reason}' - ).format(candidate=best_candidate, reason=reason) - logger.warning(msg) - return best_candidate def compute_best_candidate( self, - candidates, # type: List[InstallationCandidate] - ): - # type: (...) -> BestCandidateResult + candidates: List[InstallationCandidate], + ) -> BestCandidateResult: """ Compute and return a `BestCandidateResult` instance. """ @@ -594,7 +577,7 @@ def compute_best_candidate( ) -class PackageFinder(object): +class PackageFinder: """This finds packages. This is meant to match easy_install's technique for looking for @@ -603,14 +586,13 @@ class PackageFinder(object): def __init__( self, - link_collector, # type: LinkCollector - target_python, # type: TargetPython - allow_yanked, # type: bool - format_control=None, # type: Optional[FormatControl] - candidate_prefs=None, # type: CandidatePreferences - ignore_requires_python=None, # type: Optional[bool] - ): - # type: (...) -> None + link_collector: LinkCollector, + target_python: TargetPython, + allow_yanked: bool, + format_control: Optional[FormatControl] = None, + candidate_prefs: Optional[CandidatePreferences] = None, + ignore_requires_python: Optional[bool] = None, + ) -> None: """ This constructor is primarily meant to be used by the create() class method and from tests. @@ -635,7 +617,7 @@ def __init__( self.format_control = format_control # These are boring links that have already been logged somehow. - self._logged_links = set() # type: Set[Link] + self._logged_links: Set[Tuple[Link, LinkType, str]] = set() # Don't include an allow_yanked default value to make sure each call # site considers whether yanked releases are allowed. This also causes @@ -644,11 +626,10 @@ def __init__( @classmethod def create( cls, - link_collector, # type: LinkCollector - selection_prefs, # type: SelectionPreferences - target_python=None, # type: Optional[TargetPython] - ): - # type: (...) -> PackageFinder + link_collector: LinkCollector, + selection_prefs: SelectionPreferences, + target_python: Optional[TargetPython] = None, + ) -> "PackageFinder": """Create a PackageFinder. :param selection_prefs: The candidate selection preferences, as a @@ -675,42 +656,53 @@ def create( ) @property - def search_scope(self): - # type: () -> SearchScope + def target_python(self) -> TargetPython: + return self._target_python + + @property + def search_scope(self) -> SearchScope: return self._link_collector.search_scope @search_scope.setter - def search_scope(self, search_scope): - # type: (SearchScope) -> None + def search_scope(self, search_scope: SearchScope) -> None: self._link_collector.search_scope = search_scope @property - def find_links(self): - # type: () -> List[str] + def find_links(self) -> List[str]: return self._link_collector.find_links @property - def index_urls(self): - # type: () -> List[str] + def index_urls(self) -> List[str]: return self.search_scope.index_urls @property - def trusted_hosts(self): - # type: () -> Iterable[str] + def trusted_hosts(self) -> Iterable[str]: for host_port in self._link_collector.session.pip_trusted_origins: yield build_netloc(*host_port) @property - def allow_all_prereleases(self): - # type: () -> bool + def allow_all_prereleases(self) -> bool: return self._candidate_prefs.allow_all_prereleases - def set_allow_all_prereleases(self): - # type: () -> None + def set_allow_all_prereleases(self) -> None: self._candidate_prefs.allow_all_prereleases = True - def make_link_evaluator(self, project_name): - # type: (str) -> LinkEvaluator + @property + def prefer_binary(self) -> bool: + return self._candidate_prefs.prefer_binary + + def set_prefer_binary(self) -> None: + self._candidate_prefs.prefer_binary = True + + def requires_python_skipped_reasons(self) -> List[str]: + reasons = { + detail + for _, result, detail in self._logged_links + if result == LinkType.requires_python_mismatch + } + return sorted(reasons) + + def make_link_evaluator(self, project_name: str) -> LinkEvaluator: canonical_name = canonicalize_name(project_name) formats = self.format_control.get_allowed_formats(canonical_name) @@ -723,14 +715,13 @@ def make_link_evaluator(self, project_name): ignore_requires_python=self._ignore_requires_python, ) - def _sort_links(self, links): - # type: (Iterable[Link]) -> List[Link] + def _sort_links(self, links: Iterable[Link]) -> List[Link]: """ Returns elements of links in order, non-egg links first, egg links second, while eliminating duplicates """ eggs, no_eggs = [], [] - seen = set() # type: Set[Link] + seen: Set[Link] = set() for link in links: if link not in seen: seen.add(link) @@ -740,39 +731,38 @@ def _sort_links(self, links): no_eggs.append(link) return no_eggs + eggs - def _log_skipped_link(self, link, reason): - # type: (Link, Text) -> None - if link not in self._logged_links: - # Mark this as a unicode string to prevent "UnicodeEncodeError: - # 'ascii' codec can't encode character" in Python 2 when - # the reason contains non-ascii characters. - # Also, put the link at the end so the reason is more visible - # and because the link string is usually very long. - logger.debug(u'Skipping link: %s: %s', reason, link) - self._logged_links.add(link) - - def get_install_candidate(self, link_evaluator, link): - # type: (LinkEvaluator, Link) -> Optional[InstallationCandidate] + def _log_skipped_link(self, link: Link, result: LinkType, detail: str) -> None: + entry = (link, result, detail) + if entry not in self._logged_links: + # Put the link at the end so the reason is more visible and because + # the link string is usually very long. + logger.debug("Skipping link: %s: %s", detail, link) + self._logged_links.add(entry) + + def get_install_candidate( + self, link_evaluator: LinkEvaluator, link: Link + ) -> Optional[InstallationCandidate]: """ If the link is a candidate for install, convert it to an InstallationCandidate and return it. Otherwise, return None. """ - is_candidate, result = link_evaluator.evaluate_link(link) - if not is_candidate: - if result: - self._log_skipped_link(link, reason=result) + result, detail = link_evaluator.evaluate_link(link) + if result != LinkType.candidate: + self._log_skipped_link(link, result, detail) return None - return InstallationCandidate( - name=link_evaluator.project_name, - link=link, - # Convert the Text result to str since InstallationCandidate - # accepts str. - version=str(result), - ) + try: + return InstallationCandidate( + name=link_evaluator.project_name, + link=link, + version=detail, + ) + except InvalidVersion: + return None - def evaluate_links(self, link_evaluator, links): - # type: (LinkEvaluator, Iterable[Link]) -> List[InstallationCandidate] + def evaluate_links( + self, link_evaluator: LinkEvaluator, links: Iterable[Link] + ) -> List[InstallationCandidate]: """ Convert links that are candidates to InstallationCandidate objects. """ @@ -784,16 +774,18 @@ def evaluate_links(self, link_evaluator, links): return candidates - def process_project_url(self, project_url, link_evaluator): - # type: (Link, LinkEvaluator) -> List[InstallationCandidate] + def process_project_url( + self, project_url: Link, link_evaluator: LinkEvaluator + ) -> List[InstallationCandidate]: logger.debug( - 'Fetching project page and analyzing links: %s', project_url, + "Fetching project page and analyzing links: %s", + project_url, ) - html_page = self._link_collector.fetch_page(project_url) - if html_page is None: + index_response = self._link_collector.fetch_response(project_url) + if index_response is None: return [] - page_links = list(parse_links(html_page)) + page_links = list(parse_links(index_response)) with indent_log(): package_links = self.evaluate_links( @@ -803,8 +795,8 @@ def process_project_url(self, project_url, link_evaluator): return package_links - def find_all_candidates(self, project_name): - # type: (str) -> List[InstallationCandidate] + @functools.lru_cache(maxsize=None) + def find_all_candidates(self, project_name: str) -> List[InstallationCandidate]: """Find all available InstallationCandidate for project_name This checks index_urls and find_links. @@ -813,48 +805,56 @@ def find_all_candidates(self, project_name): See LinkEvaluator.evaluate_link() for details on which files are accepted. """ - collected_links = self._link_collector.collect_links(project_name) - link_evaluator = self.make_link_evaluator(project_name) - find_links_versions = self.evaluate_links( - link_evaluator, - links=collected_links.find_links, + collected_sources = self._link_collector.collect_sources( + project_name=project_name, + candidates_from_page=functools.partial( + self.process_project_url, + link_evaluator=link_evaluator, + ), ) - page_versions = [] - for project_url in collected_links.project_urls: - package_links = self.process_project_url( - project_url, link_evaluator=link_evaluator, - ) - page_versions.extend(package_links) + page_candidates_it = itertools.chain.from_iterable( + source.page_candidates() + for sources in collected_sources + for source in sources + if source is not None + ) + page_candidates = list(page_candidates_it) - file_versions = self.evaluate_links( + file_links_it = itertools.chain.from_iterable( + source.file_links() + for sources in collected_sources + for source in sources + if source is not None + ) + file_candidates = self.evaluate_links( link_evaluator, - links=collected_links.files, + sorted(file_links_it, reverse=True), ) - if file_versions: - file_versions.sort(reverse=True) - logger.debug( - 'Local files found: %s', - ', '.join([ - url_to_path(candidate.link.url) - for candidate in file_versions - ]) - ) + + if logger.isEnabledFor(logging.DEBUG) and file_candidates: + paths = [] + for candidate in file_candidates: + assert candidate.link.url # we need to have a URL + try: + paths.append(candidate.link.file_path) + except Exception: + paths.append(candidate.link.url) # it's not a local file + + logger.debug("Local files found: %s", ", ".join(paths)) # This is an intentional priority ordering - return file_versions + find_links_versions + page_versions + return file_candidates + page_candidates def make_candidate_evaluator( self, - project_name, # type: str - specifier=None, # type: Optional[specifiers.BaseSpecifier] - hashes=None, # type: Optional[Hashes] - ): - # type: (...) -> CandidateEvaluator - """Create a CandidateEvaluator object to use. - """ + project_name: str, + specifier: Optional[specifiers.BaseSpecifier] = None, + hashes: Optional[Hashes] = None, + ) -> CandidateEvaluator: + """Create a CandidateEvaluator object to use.""" candidate_prefs = self._candidate_prefs return CandidateEvaluator.create( project_name=project_name, @@ -865,13 +865,13 @@ def make_candidate_evaluator( hashes=hashes, ) + @functools.lru_cache(maxsize=None) def find_best_candidate( self, - project_name, # type: str - specifier=None, # type: Optional[specifiers.BaseSpecifier] - hashes=None, # type: Optional[Hashes] - ): - # type: (...) -> BestCandidateResult + project_name: str, + specifier: Optional[specifiers.BaseSpecifier] = None, + hashes: Optional[Hashes] = None, + ) -> BestCandidateResult: """Find matches for the given project and specifier. :param specifier: An optional object implementing `filter` @@ -888,90 +888,95 @@ def find_best_candidate( ) return candidate_evaluator.compute_best_candidate(candidates) - def find_requirement(self, req, upgrade): - # type: (InstallRequirement, bool) -> Optional[Link] + def find_requirement( + self, req: InstallRequirement, upgrade: bool + ) -> Optional[InstallationCandidate]: """Try to find a Link matching req Expects req, an InstallRequirement and upgrade, a boolean - Returns a Link if found, + Returns a InstallationCandidate if found, Raises DistributionNotFound or BestVersionAlreadyInstalled otherwise """ hashes = req.hashes(trust_internet=False) best_candidate_result = self.find_best_candidate( - req.name, specifier=req.specifier, hashes=hashes, + req.name, + specifier=req.specifier, + hashes=hashes, ) best_candidate = best_candidate_result.best_candidate - installed_version = None # type: Optional[_BaseVersion] + installed_version: Optional[_BaseVersion] = None if req.satisfied_by is not None: - installed_version = parse_version(req.satisfied_by.version) + installed_version = req.satisfied_by.version - def _format_versions(cand_iter): - # type: (Iterable[InstallationCandidate]) -> str + def _format_versions(cand_iter: Iterable[InstallationCandidate]) -> str: # This repeated parse_version and str() conversion is needed to # handle different vendoring sources from pip and pkg_resources. # If we stop using the pkg_resources provided specifier and start # using our own, we can drop the cast to str(). - return ", ".join(sorted( - {str(c.version) for c in cand_iter}, - key=parse_version, - )) or "none" + return ( + ", ".join( + sorted( + {str(c.version) for c in cand_iter}, + key=parse_version, + ) + ) + or "none" + ) if installed_version is None and best_candidate is None: logger.critical( - 'Could not find a version that satisfies the requirement %s ' - '(from versions: %s)', + "Could not find a version that satisfies the requirement %s " + "(from versions: %s)", req, _format_versions(best_candidate_result.iter_all()), ) - raise DistributionNotFound( - 'No matching distribution found for {}'.format( - req) - ) + raise DistributionNotFound(f"No matching distribution found for {req}") - best_installed = False - if installed_version and ( - best_candidate is None or - best_candidate.version <= installed_version): - best_installed = True + def _should_install_candidate( + candidate: Optional[InstallationCandidate], + ) -> "TypeGuard[InstallationCandidate]": + if installed_version is None: + return True + if best_candidate is None: + return False + return best_candidate.version > installed_version if not upgrade and installed_version is not None: - if best_installed: + if _should_install_candidate(best_candidate): logger.debug( - 'Existing installed version (%s) is most up-to-date and ' - 'satisfies requirement', + "Existing installed version (%s) satisfies requirement " + "(most up-to-date version is %s)", installed_version, + best_candidate.version, ) else: logger.debug( - 'Existing installed version (%s) satisfies requirement ' - '(most up-to-date version is %s)', + "Existing installed version (%s) is most up-to-date and " + "satisfies requirement", installed_version, - best_candidate.version, ) return None - if best_installed: - # We have an existing version, and its the best version + if _should_install_candidate(best_candidate): logger.debug( - 'Installed version (%s) is most up-to-date (past versions: ' - '%s)', - installed_version, + "Using version %s (newest of versions: %s)", + best_candidate.version, _format_versions(best_candidate_result.iter_applicable()), ) - raise BestVersionAlreadyInstalled + return best_candidate + # We have an existing version, and its the best version logger.debug( - 'Using version %s (newest of versions: %s)', - best_candidate.version, + "Installed version (%s) is most up-to-date (past versions: %s)", + installed_version, _format_versions(best_candidate_result.iter_applicable()), ) - return best_candidate.link + raise BestVersionAlreadyInstalled -def _find_name_version_sep(fragment, canonical_name): - # type: (str, str) -> int +def _find_name_version_sep(fragment: str, canonical_name: str) -> int: """Find the separator's index based on the package's canonical name. :param fragment: A + filename "fragment" (stem) or @@ -994,11 +999,10 @@ def _find_name_version_sep(fragment, canonical_name): continue if canonicalize_name(fragment[:i]) == canonical_name: return i - raise ValueError("{} does not match {}".format(fragment, canonical_name)) + raise ValueError(f"{fragment} does not match {canonical_name}") -def _extract_version_from_fragment(fragment, canonical_name): - # type: (str, str) -> Optional[str] +def _extract_version_from_fragment(fragment: str, canonical_name: str) -> Optional[str]: """Parse the version string from a + filename "fragment" (stem) or egg fragment. diff --git a/src/fetchcode/vcs/pip/_internal/index/sources.py b/src/fetchcode/vcs/pip/_internal/index/sources.py new file mode 100644 index 00000000..f4626d71 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/index/sources.py @@ -0,0 +1,285 @@ +import logging +import mimetypes +import os +from collections import defaultdict +from typing import Callable, Dict, Iterable, List, Optional, Tuple + +from pip._vendor.packaging.utils import ( + InvalidSdistFilename, + InvalidVersion, + InvalidWheelFilename, + canonicalize_name, + parse_sdist_filename, + parse_wheel_filename, +) + +from pip._internal.models.candidate import InstallationCandidate +from pip._internal.models.link import Link +from pip._internal.utils.urls import path_to_url, url_to_path +from pip._internal.vcs import is_url + +logger = logging.getLogger(__name__) + +FoundCandidates = Iterable[InstallationCandidate] +FoundLinks = Iterable[Link] +CandidatesFromPage = Callable[[Link], Iterable[InstallationCandidate]] +PageValidator = Callable[[Link], bool] + + +class LinkSource: + @property + def link(self) -> Optional[Link]: + """Returns the underlying link, if there's one.""" + raise NotImplementedError() + + def page_candidates(self) -> FoundCandidates: + """Candidates found by parsing an archive listing HTML file.""" + raise NotImplementedError() + + def file_links(self) -> FoundLinks: + """Links found by specifying archives directly.""" + raise NotImplementedError() + + +def _is_html_file(file_url: str) -> bool: + return mimetypes.guess_type(file_url, strict=False)[0] == "text/html" + + +class _FlatDirectoryToUrls: + """Scans directory and caches results""" + + def __init__(self, path: str) -> None: + self._path = path + self._page_candidates: List[str] = [] + self._project_name_to_urls: Dict[str, List[str]] = defaultdict(list) + self._scanned_directory = False + + def _scan_directory(self) -> None: + """Scans directory once and populates both page_candidates + and project_name_to_urls at the same time + """ + for entry in os.scandir(self._path): + url = path_to_url(entry.path) + if _is_html_file(url): + self._page_candidates.append(url) + continue + + # File must have a valid wheel or sdist name, + # otherwise not worth considering as a package + try: + project_filename = parse_wheel_filename(entry.name)[0] + except (InvalidWheelFilename, InvalidVersion): + try: + project_filename = parse_sdist_filename(entry.name)[0] + except (InvalidSdistFilename, InvalidVersion): + continue + + self._project_name_to_urls[project_filename].append(url) + self._scanned_directory = True + + @property + def page_candidates(self) -> List[str]: + if not self._scanned_directory: + self._scan_directory() + + return self._page_candidates + + @property + def project_name_to_urls(self) -> Dict[str, List[str]]: + if not self._scanned_directory: + self._scan_directory() + + return self._project_name_to_urls + + +class _FlatDirectorySource(LinkSource): + """Link source specified by ``--find-links=``. + + This looks the content of the directory, and returns: + + * ``page_candidates``: Links listed on each HTML file in the directory. + * ``file_candidates``: Archives in the directory. + """ + + _paths_to_urls: Dict[str, _FlatDirectoryToUrls] = {} + + def __init__( + self, + candidates_from_page: CandidatesFromPage, + path: str, + project_name: str, + ) -> None: + self._candidates_from_page = candidates_from_page + self._project_name = canonicalize_name(project_name) + + # Get existing instance of _FlatDirectoryToUrls if it exists + if path in self._paths_to_urls: + self._path_to_urls = self._paths_to_urls[path] + else: + self._path_to_urls = _FlatDirectoryToUrls(path=path) + self._paths_to_urls[path] = self._path_to_urls + + @property + def link(self) -> Optional[Link]: + return None + + def page_candidates(self) -> FoundCandidates: + for url in self._path_to_urls.page_candidates: + yield from self._candidates_from_page(Link(url)) + + def file_links(self) -> FoundLinks: + for url in self._path_to_urls.project_name_to_urls[self._project_name]: + yield Link(url) + + +class _LocalFileSource(LinkSource): + """``--find-links=`` or ``--[extra-]index-url=``. + + If a URL is supplied, it must be a ``file:`` URL. If a path is supplied to + the option, it is converted to a URL first. This returns: + + * ``page_candidates``: Links listed on an HTML file. + * ``file_candidates``: The non-HTML file. + """ + + def __init__( + self, + candidates_from_page: CandidatesFromPage, + link: Link, + ) -> None: + self._candidates_from_page = candidates_from_page + self._link = link + + @property + def link(self) -> Optional[Link]: + return self._link + + def page_candidates(self) -> FoundCandidates: + if not _is_html_file(self._link.url): + return + yield from self._candidates_from_page(self._link) + + def file_links(self) -> FoundLinks: + if _is_html_file(self._link.url): + return + yield self._link + + +class _RemoteFileSource(LinkSource): + """``--find-links=`` or ``--[extra-]index-url=``. + + This returns: + + * ``page_candidates``: Links listed on an HTML file. + * ``file_candidates``: The non-HTML file. + """ + + def __init__( + self, + candidates_from_page: CandidatesFromPage, + page_validator: PageValidator, + link: Link, + ) -> None: + self._candidates_from_page = candidates_from_page + self._page_validator = page_validator + self._link = link + + @property + def link(self) -> Optional[Link]: + return self._link + + def page_candidates(self) -> FoundCandidates: + if not self._page_validator(self._link): + return + yield from self._candidates_from_page(self._link) + + def file_links(self) -> FoundLinks: + yield self._link + + +class _IndexDirectorySource(LinkSource): + """``--[extra-]index-url=``. + + This is treated like a remote URL; ``candidates_from_page`` contains logic + for this by appending ``index.html`` to the link. + """ + + def __init__( + self, + candidates_from_page: CandidatesFromPage, + link: Link, + ) -> None: + self._candidates_from_page = candidates_from_page + self._link = link + + @property + def link(self) -> Optional[Link]: + return self._link + + def page_candidates(self) -> FoundCandidates: + yield from self._candidates_from_page(self._link) + + def file_links(self) -> FoundLinks: + return () + + +def build_source( + location: str, + *, + candidates_from_page: CandidatesFromPage, + page_validator: PageValidator, + expand_dir: bool, + cache_link_parsing: bool, + project_name: str, +) -> Tuple[Optional[str], Optional[LinkSource]]: + path: Optional[str] = None + url: Optional[str] = None + if os.path.exists(location): # Is a local path. + url = path_to_url(location) + path = location + elif location.startswith("file:"): # A file: URL. + url = location + path = url_to_path(location) + elif is_url(location): + url = location + + if url is None: + msg = ( + "Location '%s' is ignored: " + "it is either a non-existing path or lacks a specific scheme." + ) + logger.warning(msg, location) + return (None, None) + + if path is None: + source: LinkSource = _RemoteFileSource( + candidates_from_page=candidates_from_page, + page_validator=page_validator, + link=Link(url, cache_link_parsing=cache_link_parsing), + ) + return (url, source) + + if os.path.isdir(path): + if expand_dir: + source = _FlatDirectorySource( + candidates_from_page=candidates_from_page, + path=path, + project_name=project_name, + ) + else: + source = _IndexDirectorySource( + candidates_from_page=candidates_from_page, + link=Link(url, cache_link_parsing=cache_link_parsing), + ) + return (url, source) + elif os.path.isfile(path): + source = _LocalFileSource( + candidates_from_page=candidates_from_page, + link=Link(url, cache_link_parsing=cache_link_parsing), + ) + return (url, source) + logger.warning( + "Location '%s' is ignored: it is neither a file nor a directory.", + location, + ) + return (url, None) diff --git a/src/fetchcode/vcs/pip/_internal/locations.py b/src/fetchcode/vcs/pip/_internal/locations.py deleted file mode 100644 index 1ae96343..00000000 --- a/src/fetchcode/vcs/pip/_internal/locations.py +++ /dev/null @@ -1,194 +0,0 @@ -"""Locations where we look for configs, install stuff, etc""" - -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False - -from __future__ import absolute_import - -import os -import os.path -import platform -import site -import sys -import sysconfig -from distutils import sysconfig as distutils_sysconfig -from distutils.command.install import SCHEME_KEYS # type: ignore -from distutils.command.install import install as distutils_install_command - -from fetchcode.vcs.pip._internal.models.scheme import Scheme -from fetchcode.vcs.pip._internal.utils import appdirs -from fetchcode.vcs.pip._internal.utils.compat import WINDOWS -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING, cast -from fetchcode.vcs.pip._internal.utils.virtualenv import running_under_virtualenv - -if MYPY_CHECK_RUNNING: - from typing import Dict, List, Optional, Union - - from distutils.cmd import Command as DistutilsCommand - - -# Application Directories -USER_CACHE_DIR = appdirs.user_cache_dir("pip") - - -def get_major_minor_version(): - # type: () -> str - """ - Return the major-minor version of the current Python as a string, e.g. - "3.7" or "3.10". - """ - return '{}.{}'.format(*sys.version_info) - - -def get_src_prefix(): - # type: () -> str - if running_under_virtualenv(): - src_prefix = os.path.join(sys.prefix, 'src') - else: - # FIXME: keep src in cwd for now (it is not a temporary folder) - try: - src_prefix = os.path.join(os.getcwd(), 'src') - except OSError: - # In case the current working directory has been renamed or deleted - sys.exit( - "The folder you are executing pip from can no longer be found." - ) - - # under macOS + virtualenv sys.prefix is not properly resolved - # it is something like /path/to/python/bin/.. - return os.path.abspath(src_prefix) - - -# FIXME doesn't account for venv linked to global site-packages - -site_packages = sysconfig.get_path("purelib") # type: Optional[str] - -# This is because of a bug in PyPy's sysconfig module, see -# https://bitbucket.org/pypy/pypy/issues/2506/sysconfig-returns-incorrect-paths -# for more information. -if platform.python_implementation().lower() == "pypy": - site_packages = distutils_sysconfig.get_python_lib() -try: - # Use getusersitepackages if this is present, as it ensures that the - # value is initialised properly. - user_site = site.getusersitepackages() -except AttributeError: - user_site = site.USER_SITE - -if WINDOWS: - bin_py = os.path.join(sys.prefix, 'Scripts') - bin_user = os.path.join(user_site, 'Scripts') - # buildout uses 'bin' on Windows too? - if not os.path.exists(bin_py): - bin_py = os.path.join(sys.prefix, 'bin') - bin_user = os.path.join(user_site, 'bin') -else: - bin_py = os.path.join(sys.prefix, 'bin') - bin_user = os.path.join(user_site, 'bin') - - # Forcing to use /usr/local/bin for standard macOS framework installs - # Also log to ~/Library/Logs/ for use with the Console.app log viewer - if sys.platform[:6] == 'darwin' and sys.prefix[:16] == '/System/Library/': - bin_py = '/usr/local/bin' - - -def distutils_scheme( - dist_name, user=False, home=None, root=None, isolated=False, prefix=None -): - # type:(str, bool, str, str, bool, str) -> Dict[str, str] - """ - Return a distutils install scheme - """ - from distutils.dist import Distribution - - dist_args = {'name': dist_name} # type: Dict[str, Union[str, List[str]]] - if isolated: - dist_args["script_args"] = ["--no-user-cfg"] - - d = Distribution(dist_args) - d.parse_config_files() - obj = None # type: Optional[DistutilsCommand] - obj = d.get_command_obj('install', create=True) - assert obj is not None - i = cast(distutils_install_command, obj) - # NOTE: setting user or home has the side-effect of creating the home dir - # or user base for installations during finalize_options() - # ideally, we'd prefer a scheme class that has no side-effects. - assert not (user and prefix), "user={} prefix={}".format(user, prefix) - assert not (home and prefix), "home={} prefix={}".format(home, prefix) - i.user = user or i.user - if user or home: - i.prefix = "" - i.prefix = prefix or i.prefix - i.home = home or i.home - i.root = root or i.root - i.finalize_options() - - scheme = {} - for key in SCHEME_KEYS: - scheme[key] = getattr(i, 'install_' + key) - - # install_lib specified in setup.cfg should install *everything* - # into there (i.e. it takes precedence over both purelib and - # platlib). Note, i.install_lib is *always* set after - # finalize_options(); we only want to override here if the user - # has explicitly requested it hence going back to the config - if 'install_lib' in d.get_option_dict('install'): - scheme.update(dict(purelib=i.install_lib, platlib=i.install_lib)) - - if running_under_virtualenv(): - scheme['headers'] = os.path.join( - sys.prefix, - 'include', - 'site', - 'python{}'.format(get_major_minor_version()), - dist_name, - ) - - if root is not None: - path_no_drive = os.path.splitdrive( - os.path.abspath(scheme["headers"]))[1] - scheme["headers"] = os.path.join( - root, - path_no_drive[1:], - ) - - return scheme - - -def get_scheme( - dist_name, # type: str - user=False, # type: bool - home=None, # type: Optional[str] - root=None, # type: Optional[str] - isolated=False, # type: bool - prefix=None, # type: Optional[str] -): - # type: (...) -> Scheme - """ - Get the "scheme" corresponding to the input parameters. The distutils - documentation provides the context for the available schemes: - https://docs.python.org/3/install/index.html#alternate-installation - - :param dist_name: the name of the package to retrieve the scheme for, used - in the headers scheme path - :param user: indicates to use the "user" scheme - :param home: indicates to use the "home" scheme and provides the base - directory for the same - :param root: root under which other directories are re-based - :param isolated: equivalent to --no-user-cfg, i.e. do not consider - ~/.pydistutils.cfg (posix) or ~/pydistutils.cfg (non-posix) for - scheme paths - :param prefix: indicates to use the "prefix" scheme and provides the - base directory for the same - """ - scheme = distutils_scheme( - dist_name, user, home, root, isolated, prefix - ) - return Scheme( - platlib=scheme["platlib"], - purelib=scheme["purelib"], - headers=scheme["headers"], - scripts=scheme["scripts"], - data=scheme["data"], - ) diff --git a/src/fetchcode/vcs/pip/_internal/locations/__init__.py b/src/fetchcode/vcs/pip/_internal/locations/__init__.py new file mode 100644 index 00000000..32382be7 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/locations/__init__.py @@ -0,0 +1,456 @@ +import functools +import logging +import os +import pathlib +import sys +import sysconfig +from typing import Any, Dict, Generator, Optional, Tuple + +from pip._internal.models.scheme import SCHEME_KEYS, Scheme +from pip._internal.utils.compat import WINDOWS +from pip._internal.utils.deprecation import deprecated +from pip._internal.utils.virtualenv import running_under_virtualenv + +from . import _sysconfig +from .base import ( + USER_CACHE_DIR, + get_major_minor_version, + get_src_prefix, + is_osx_framework, + site_packages, + user_site, +) + +__all__ = [ + "USER_CACHE_DIR", + "get_bin_prefix", + "get_bin_user", + "get_major_minor_version", + "get_platlib", + "get_purelib", + "get_scheme", + "get_src_prefix", + "site_packages", + "user_site", +] + + +logger = logging.getLogger(__name__) + + +_PLATLIBDIR: str = getattr(sys, "platlibdir", "lib") + +_USE_SYSCONFIG_DEFAULT = sys.version_info >= (3, 10) + + +def _should_use_sysconfig() -> bool: + """This function determines the value of _USE_SYSCONFIG. + + By default, pip uses sysconfig on Python 3.10+. + But Python distributors can override this decision by setting: + sysconfig._PIP_USE_SYSCONFIG = True / False + Rationale in https://github.com/pypa/pip/issues/10647 + + This is a function for testability, but should be constant during any one + run. + """ + return bool(getattr(sysconfig, "_PIP_USE_SYSCONFIG", _USE_SYSCONFIG_DEFAULT)) + + +_USE_SYSCONFIG = _should_use_sysconfig() + +if not _USE_SYSCONFIG: + # Import distutils lazily to avoid deprecation warnings, + # but import it soon enough that it is in memory and available during + # a pip reinstall. + from . import _distutils + +# Be noisy about incompatibilities if this platforms "should" be using +# sysconfig, but is explicitly opting out and using distutils instead. +if _USE_SYSCONFIG_DEFAULT and not _USE_SYSCONFIG: + _MISMATCH_LEVEL = logging.WARNING +else: + _MISMATCH_LEVEL = logging.DEBUG + + +def _looks_like_bpo_44860() -> bool: + """The resolution to bpo-44860 will change this incorrect platlib. + + See . + """ + from distutils.command.install import INSTALL_SCHEMES + + try: + unix_user_platlib = INSTALL_SCHEMES["unix_user"]["platlib"] + except KeyError: + return False + return unix_user_platlib == "$usersite" + + +def _looks_like_red_hat_patched_platlib_purelib(scheme: Dict[str, str]) -> bool: + platlib = scheme["platlib"] + if "/$platlibdir/" in platlib: + platlib = platlib.replace("/$platlibdir/", f"/{_PLATLIBDIR}/") + if "/lib64/" not in platlib: + return False + unpatched = platlib.replace("/lib64/", "/lib/") + return unpatched.replace("$platbase/", "$base/") == scheme["purelib"] + + +@functools.lru_cache(maxsize=None) +def _looks_like_red_hat_lib() -> bool: + """Red Hat patches platlib in unix_prefix and unix_home, but not purelib. + + This is the only way I can see to tell a Red Hat-patched Python. + """ + from distutils.command.install import INSTALL_SCHEMES + + return all( + k in INSTALL_SCHEMES + and _looks_like_red_hat_patched_platlib_purelib(INSTALL_SCHEMES[k]) + for k in ("unix_prefix", "unix_home") + ) + + +@functools.lru_cache(maxsize=None) +def _looks_like_debian_scheme() -> bool: + """Debian adds two additional schemes.""" + from distutils.command.install import INSTALL_SCHEMES + + return "deb_system" in INSTALL_SCHEMES and "unix_local" in INSTALL_SCHEMES + + +@functools.lru_cache(maxsize=None) +def _looks_like_red_hat_scheme() -> bool: + """Red Hat patches ``sys.prefix`` and ``sys.exec_prefix``. + + Red Hat's ``00251-change-user-install-location.patch`` changes the install + command's ``prefix`` and ``exec_prefix`` to append ``"/local"``. This is + (fortunately?) done quite unconditionally, so we create a default command + object without any configuration to detect this. + """ + from distutils.command.install import install + from distutils.dist import Distribution + + cmd: Any = install(Distribution()) + cmd.finalize_options() + return ( + cmd.exec_prefix == f"{os.path.normpath(sys.exec_prefix)}/local" + and cmd.prefix == f"{os.path.normpath(sys.prefix)}/local" + ) + + +@functools.lru_cache(maxsize=None) +def _looks_like_slackware_scheme() -> bool: + """Slackware patches sysconfig but fails to patch distutils and site. + + Slackware changes sysconfig's user scheme to use ``"lib64"`` for the lib + path, but does not do the same to the site module. + """ + if user_site is None: # User-site not available. + return False + try: + paths = sysconfig.get_paths(scheme="posix_user", expand=False) + except KeyError: # User-site not available. + return False + return "/lib64/" in paths["purelib"] and "/lib64/" not in user_site + + +@functools.lru_cache(maxsize=None) +def _looks_like_msys2_mingw_scheme() -> bool: + """MSYS2 patches distutils and sysconfig to use a UNIX-like scheme. + + However, MSYS2 incorrectly patches sysconfig ``nt`` scheme. The fix is + likely going to be included in their 3.10 release, so we ignore the warning. + See msys2/MINGW-packages#9319. + + MSYS2 MINGW's patch uses lowercase ``"lib"`` instead of the usual uppercase, + and is missing the final ``"site-packages"``. + """ + paths = sysconfig.get_paths("nt", expand=False) + return all( + "Lib" not in p and "lib" in p and not p.endswith("site-packages") + for p in (paths[key] for key in ("platlib", "purelib")) + ) + + +def _fix_abiflags(parts: Tuple[str]) -> Generator[str, None, None]: + ldversion = sysconfig.get_config_var("LDVERSION") + abiflags = getattr(sys, "abiflags", None) + + # LDVERSION does not end with sys.abiflags. Just return the path unchanged. + if not ldversion or not abiflags or not ldversion.endswith(abiflags): + yield from parts + return + + # Strip sys.abiflags from LDVERSION-based path components. + for part in parts: + if part.endswith(ldversion): + part = part[: (0 - len(abiflags))] + yield part + + +@functools.lru_cache(maxsize=None) +def _warn_mismatched(old: pathlib.Path, new: pathlib.Path, *, key: str) -> None: + issue_url = "https://github.com/pypa/pip/issues/10151" + message = ( + "Value for %s does not match. Please report this to <%s>" + "\ndistutils: %s" + "\nsysconfig: %s" + ) + logger.log(_MISMATCH_LEVEL, message, key, issue_url, old, new) + + +def _warn_if_mismatch(old: pathlib.Path, new: pathlib.Path, *, key: str) -> bool: + if old == new: + return False + _warn_mismatched(old, new, key=key) + return True + + +@functools.lru_cache(maxsize=None) +def _log_context( + *, + user: bool = False, + home: Optional[str] = None, + root: Optional[str] = None, + prefix: Optional[str] = None, +) -> None: + parts = [ + "Additional context:", + "user = %r", + "home = %r", + "root = %r", + "prefix = %r", + ] + + logger.log(_MISMATCH_LEVEL, "\n".join(parts), user, home, root, prefix) + + +def get_scheme( + dist_name: str, + user: bool = False, + home: Optional[str] = None, + root: Optional[str] = None, + isolated: bool = False, + prefix: Optional[str] = None, +) -> Scheme: + new = _sysconfig.get_scheme( + dist_name, + user=user, + home=home, + root=root, + isolated=isolated, + prefix=prefix, + ) + if _USE_SYSCONFIG: + return new + + old = _distutils.get_scheme( + dist_name, + user=user, + home=home, + root=root, + isolated=isolated, + prefix=prefix, + ) + + warning_contexts = [] + for k in SCHEME_KEYS: + old_v = pathlib.Path(getattr(old, k)) + new_v = pathlib.Path(getattr(new, k)) + + if old_v == new_v: + continue + + # distutils incorrectly put PyPy packages under ``site-packages/python`` + # in the ``posix_home`` scheme, but PyPy devs said they expect the + # directory name to be ``pypy`` instead. So we treat this as a bug fix + # and not warn about it. See bpo-43307 and python/cpython#24628. + skip_pypy_special_case = ( + sys.implementation.name == "pypy" + and home is not None + and k in ("platlib", "purelib") + and old_v.parent == new_v.parent + and old_v.name.startswith("python") + and new_v.name.startswith("pypy") + ) + if skip_pypy_special_case: + continue + + # sysconfig's ``osx_framework_user`` does not include ``pythonX.Y`` in + # the ``include`` value, but distutils's ``headers`` does. We'll let + # CPython decide whether this is a bug or feature. See bpo-43948. + skip_osx_framework_user_special_case = ( + user + and is_osx_framework() + and k == "headers" + and old_v.parent.parent == new_v.parent + and old_v.parent.name.startswith("python") + ) + if skip_osx_framework_user_special_case: + continue + + # On Red Hat and derived Linux distributions, distutils is patched to + # use "lib64" instead of "lib" for platlib. + if k == "platlib" and _looks_like_red_hat_lib(): + continue + + # On Python 3.9+, sysconfig's posix_user scheme sets platlib against + # sys.platlibdir, but distutils's unix_user incorrectly coninutes + # using the same $usersite for both platlib and purelib. This creates a + # mismatch when sys.platlibdir is not "lib". + skip_bpo_44860 = ( + user + and k == "platlib" + and not WINDOWS + and sys.version_info >= (3, 9) + and _PLATLIBDIR != "lib" + and _looks_like_bpo_44860() + ) + if skip_bpo_44860: + continue + + # Slackware incorrectly patches posix_user to use lib64 instead of lib, + # but not usersite to match the location. + skip_slackware_user_scheme = ( + user + and k in ("platlib", "purelib") + and not WINDOWS + and _looks_like_slackware_scheme() + ) + if skip_slackware_user_scheme: + continue + + # Both Debian and Red Hat patch Python to place the system site under + # /usr/local instead of /usr. Debian also places lib in dist-packages + # instead of site-packages, but the /usr/local check should cover it. + skip_linux_system_special_case = ( + not (user or home or prefix or running_under_virtualenv()) + and old_v.parts[1:3] == ("usr", "local") + and len(new_v.parts) > 1 + and new_v.parts[1] == "usr" + and (len(new_v.parts) < 3 or new_v.parts[2] != "local") + and (_looks_like_red_hat_scheme() or _looks_like_debian_scheme()) + ) + if skip_linux_system_special_case: + continue + + # MSYS2 MINGW's sysconfig patch does not include the "site-packages" + # part of the path. This is incorrect and will be fixed in MSYS. + skip_msys2_mingw_bug = ( + WINDOWS and k in ("platlib", "purelib") and _looks_like_msys2_mingw_scheme() + ) + if skip_msys2_mingw_bug: + continue + + # CPython's POSIX install script invokes pip (via ensurepip) against the + # interpreter located in the source tree, not the install site. This + # triggers special logic in sysconfig that's not present in distutils. + # https://github.com/python/cpython/blob/8c21941ddaf/Lib/sysconfig.py#L178-L194 + skip_cpython_build = ( + sysconfig.is_python_build(check_home=True) + and not WINDOWS + and k in ("headers", "include", "platinclude") + ) + if skip_cpython_build: + continue + + warning_contexts.append((old_v, new_v, f"scheme.{k}")) + + if not warning_contexts: + return old + + # Check if this path mismatch is caused by distutils config files. Those + # files will no longer work once we switch to sysconfig, so this raises a + # deprecation message for them. + default_old = _distutils.distutils_scheme( + dist_name, + user, + home, + root, + isolated, + prefix, + ignore_config_files=True, + ) + if any(default_old[k] != getattr(old, k) for k in SCHEME_KEYS): + deprecated( + reason=( + "Configuring installation scheme with distutils config files " + "is deprecated and will no longer work in the near future. If you " + "are using a Homebrew or Linuxbrew Python, please see discussion " + "at https://github.com/Homebrew/homebrew-core/issues/76621" + ), + replacement=None, + gone_in=None, + ) + return old + + # Post warnings about this mismatch so user can report them back. + for old_v, new_v, key in warning_contexts: + _warn_mismatched(old_v, new_v, key=key) + _log_context(user=user, home=home, root=root, prefix=prefix) + + return old + + +def get_bin_prefix() -> str: + new = _sysconfig.get_bin_prefix() + if _USE_SYSCONFIG: + return new + + old = _distutils.get_bin_prefix() + if _warn_if_mismatch(pathlib.Path(old), pathlib.Path(new), key="bin_prefix"): + _log_context() + return old + + +def get_bin_user() -> str: + return _sysconfig.get_scheme("", user=True).scripts + + +def _looks_like_deb_system_dist_packages(value: str) -> bool: + """Check if the value is Debian's APT-controlled dist-packages. + + Debian's ``distutils.sysconfig.get_python_lib()`` implementation returns the + default package path controlled by APT, but does not patch ``sysconfig`` to + do the same. This is similar to the bug worked around in ``get_scheme()``, + but here the default is ``deb_system`` instead of ``unix_local``. Ultimately + we can't do anything about this Debian bug, and this detection allows us to + skip the warning when needed. + """ + if not _looks_like_debian_scheme(): + return False + if value == "/usr/lib/python3/dist-packages": + return True + return False + + +def get_purelib() -> str: + """Return the default pure-Python lib location.""" + new = _sysconfig.get_purelib() + if _USE_SYSCONFIG: + return new + + old = _distutils.get_purelib() + if _looks_like_deb_system_dist_packages(old): + return old + if _warn_if_mismatch(pathlib.Path(old), pathlib.Path(new), key="purelib"): + _log_context() + return old + + +def get_platlib() -> str: + """Return the default platform-shared lib location.""" + new = _sysconfig.get_platlib() + if _USE_SYSCONFIG: + return new + + from . import _distutils + + old = _distutils.get_platlib() + if _looks_like_deb_system_dist_packages(old): + return old + if _warn_if_mismatch(pathlib.Path(old), pathlib.Path(new), key="platlib"): + _log_context() + return old diff --git a/src/fetchcode/vcs/pip/_internal/locations/_distutils.py b/src/fetchcode/vcs/pip/_internal/locations/_distutils.py new file mode 100644 index 00000000..0e18c6e1 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/locations/_distutils.py @@ -0,0 +1,172 @@ +"""Locations where we look for configs, install stuff, etc""" + +# The following comment should be removed at some point in the future. +# mypy: strict-optional=False + +# If pip's going to use distutils, it should not be using the copy that setuptools +# might have injected into the environment. This is done by removing the injected +# shim, if it's injected. +# +# See https://github.com/pypa/pip/issues/8761 for the original discussion and +# rationale for why this is done within pip. +try: + __import__("_distutils_hack").remove_shim() +except (ImportError, AttributeError): + pass + +import logging +import os +import sys +from distutils.cmd import Command as DistutilsCommand +from distutils.command.install import SCHEME_KEYS +from distutils.command.install import install as distutils_install_command +from distutils.sysconfig import get_python_lib +from typing import Dict, List, Optional, Union, cast + +from pip._internal.models.scheme import Scheme +from pip._internal.utils.compat import WINDOWS +from pip._internal.utils.virtualenv import running_under_virtualenv + +from .base import get_major_minor_version + +logger = logging.getLogger(__name__) + + +def distutils_scheme( + dist_name: str, + user: bool = False, + home: Optional[str] = None, + root: Optional[str] = None, + isolated: bool = False, + prefix: Optional[str] = None, + *, + ignore_config_files: bool = False, +) -> Dict[str, str]: + """ + Return a distutils install scheme + """ + from distutils.dist import Distribution + + dist_args: Dict[str, Union[str, List[str]]] = {"name": dist_name} + if isolated: + dist_args["script_args"] = ["--no-user-cfg"] + + d = Distribution(dist_args) + if not ignore_config_files: + try: + d.parse_config_files() + except UnicodeDecodeError: + paths = d.find_config_files() + logger.warning( + "Ignore distutils configs in %s due to encoding errors.", + ", ".join(os.path.basename(p) for p in paths), + ) + obj: Optional[DistutilsCommand] = None + obj = d.get_command_obj("install", create=True) + assert obj is not None + i = cast(distutils_install_command, obj) + # NOTE: setting user or home has the side-effect of creating the home dir + # or user base for installations during finalize_options() + # ideally, we'd prefer a scheme class that has no side-effects. + assert not (user and prefix), f"user={user} prefix={prefix}" + assert not (home and prefix), f"home={home} prefix={prefix}" + i.user = user or i.user + if user or home: + i.prefix = "" + i.prefix = prefix or i.prefix + i.home = home or i.home + i.root = root or i.root + i.finalize_options() + + scheme = {} + for key in SCHEME_KEYS: + scheme[key] = getattr(i, "install_" + key) + + # install_lib specified in setup.cfg should install *everything* + # into there (i.e. it takes precedence over both purelib and + # platlib). Note, i.install_lib is *always* set after + # finalize_options(); we only want to override here if the user + # has explicitly requested it hence going back to the config + if "install_lib" in d.get_option_dict("install"): + scheme.update({"purelib": i.install_lib, "platlib": i.install_lib}) + + if running_under_virtualenv(): + if home: + prefix = home + elif user: + prefix = i.install_userbase + else: + prefix = i.prefix + scheme["headers"] = os.path.join( + prefix, + "include", + "site", + f"python{get_major_minor_version()}", + dist_name, + ) + + if root is not None: + path_no_drive = os.path.splitdrive(os.path.abspath(scheme["headers"]))[1] + scheme["headers"] = os.path.join(root, path_no_drive[1:]) + + return scheme + + +def get_scheme( + dist_name: str, + user: bool = False, + home: Optional[str] = None, + root: Optional[str] = None, + isolated: bool = False, + prefix: Optional[str] = None, +) -> Scheme: + """ + Get the "scheme" corresponding to the input parameters. The distutils + documentation provides the context for the available schemes: + https://docs.python.org/3/install/index.html#alternate-installation + + :param dist_name: the name of the package to retrieve the scheme for, used + in the headers scheme path + :param user: indicates to use the "user" scheme + :param home: indicates to use the "home" scheme and provides the base + directory for the same + :param root: root under which other directories are re-based + :param isolated: equivalent to --no-user-cfg, i.e. do not consider + ~/.pydistutils.cfg (posix) or ~/pydistutils.cfg (non-posix) for + scheme paths + :param prefix: indicates to use the "prefix" scheme and provides the + base directory for the same + """ + scheme = distutils_scheme(dist_name, user, home, root, isolated, prefix) + return Scheme( + platlib=scheme["platlib"], + purelib=scheme["purelib"], + headers=scheme["headers"], + scripts=scheme["scripts"], + data=scheme["data"], + ) + + +def get_bin_prefix() -> str: + # XXX: In old virtualenv versions, sys.prefix can contain '..' components, + # so we need to call normpath to eliminate them. + prefix = os.path.normpath(sys.prefix) + if WINDOWS: + bin_py = os.path.join(prefix, "Scripts") + # buildout uses 'bin' on Windows too? + if not os.path.exists(bin_py): + bin_py = os.path.join(prefix, "bin") + return bin_py + # Forcing to use /usr/local/bin for standard macOS framework installs + # Also log to ~/Library/Logs/ for use with the Console.app log viewer + if sys.platform[:6] == "darwin" and prefix[:16] == "/System/Library/": + return "/usr/local/bin" + return os.path.join(prefix, "bin") + + +def get_purelib() -> str: + return get_python_lib(plat_specific=False) + + +def get_platlib() -> str: + return get_python_lib(plat_specific=True) diff --git a/src/fetchcode/vcs/pip/_internal/locations/_sysconfig.py b/src/fetchcode/vcs/pip/_internal/locations/_sysconfig.py new file mode 100644 index 00000000..ca860ea5 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/locations/_sysconfig.py @@ -0,0 +1,214 @@ +import logging +import os +import sys +import sysconfig +import typing + +from pip._internal.exceptions import InvalidSchemeCombination, UserInstallationInvalid +from pip._internal.models.scheme import SCHEME_KEYS, Scheme +from pip._internal.utils.virtualenv import running_under_virtualenv + +from .base import change_root, get_major_minor_version, is_osx_framework + +logger = logging.getLogger(__name__) + + +# Notes on _infer_* functions. +# Unfortunately ``get_default_scheme()`` didn't exist before 3.10, so there's no +# way to ask things like "what is the '_prefix' scheme on this platform". These +# functions try to answer that with some heuristics while accounting for ad-hoc +# platforms not covered by CPython's default sysconfig implementation. If the +# ad-hoc implementation does not fully implement sysconfig, we'll fall back to +# a POSIX scheme. + +_AVAILABLE_SCHEMES = set(sysconfig.get_scheme_names()) + +_PREFERRED_SCHEME_API = getattr(sysconfig, "get_preferred_scheme", None) + + +def _should_use_osx_framework_prefix() -> bool: + """Check for Apple's ``osx_framework_library`` scheme. + + Python distributed by Apple's Command Line Tools has this special scheme + that's used when: + + * This is a framework build. + * We are installing into the system prefix. + + This does not account for ``pip install --prefix`` (also means we're not + installing to the system prefix), which should use ``posix_prefix``, but + logic here means ``_infer_prefix()`` outputs ``osx_framework_library``. But + since ``prefix`` is not available for ``sysconfig.get_default_scheme()``, + which is the stdlib replacement for ``_infer_prefix()``, presumably Apple + wouldn't be able to magically switch between ``osx_framework_library`` and + ``posix_prefix``. ``_infer_prefix()`` returning ``osx_framework_library`` + means its behavior is consistent whether we use the stdlib implementation + or our own, and we deal with this special case in ``get_scheme()`` instead. + """ + return ( + "osx_framework_library" in _AVAILABLE_SCHEMES + and not running_under_virtualenv() + and is_osx_framework() + ) + + +def _infer_prefix() -> str: + """Try to find a prefix scheme for the current platform. + + This tries: + + * A special ``osx_framework_library`` for Python distributed by Apple's + Command Line Tools, when not running in a virtual environment. + * Implementation + OS, used by PyPy on Windows (``pypy_nt``). + * Implementation without OS, used by PyPy on POSIX (``pypy``). + * OS + "prefix", used by CPython on POSIX (``posix_prefix``). + * Just the OS name, used by CPython on Windows (``nt``). + + If none of the above works, fall back to ``posix_prefix``. + """ + if _PREFERRED_SCHEME_API: + return _PREFERRED_SCHEME_API("prefix") + if _should_use_osx_framework_prefix(): + return "osx_framework_library" + implementation_suffixed = f"{sys.implementation.name}_{os.name}" + if implementation_suffixed in _AVAILABLE_SCHEMES: + return implementation_suffixed + if sys.implementation.name in _AVAILABLE_SCHEMES: + return sys.implementation.name + suffixed = f"{os.name}_prefix" + if suffixed in _AVAILABLE_SCHEMES: + return suffixed + if os.name in _AVAILABLE_SCHEMES: # On Windows, prefx is just called "nt". + return os.name + return "posix_prefix" + + +def _infer_user() -> str: + """Try to find a user scheme for the current platform.""" + if _PREFERRED_SCHEME_API: + return _PREFERRED_SCHEME_API("user") + if is_osx_framework() and not running_under_virtualenv(): + suffixed = "osx_framework_user" + else: + suffixed = f"{os.name}_user" + if suffixed in _AVAILABLE_SCHEMES: + return suffixed + if "posix_user" not in _AVAILABLE_SCHEMES: # User scheme unavailable. + raise UserInstallationInvalid() + return "posix_user" + + +def _infer_home() -> str: + """Try to find a home for the current platform.""" + if _PREFERRED_SCHEME_API: + return _PREFERRED_SCHEME_API("home") + suffixed = f"{os.name}_home" + if suffixed in _AVAILABLE_SCHEMES: + return suffixed + return "posix_home" + + +# Update these keys if the user sets a custom home. +_HOME_KEYS = [ + "installed_base", + "base", + "installed_platbase", + "platbase", + "prefix", + "exec_prefix", +] +if sysconfig.get_config_var("userbase") is not None: + _HOME_KEYS.append("userbase") + + +def get_scheme( + dist_name: str, + user: bool = False, + home: typing.Optional[str] = None, + root: typing.Optional[str] = None, + isolated: bool = False, + prefix: typing.Optional[str] = None, +) -> Scheme: + """ + Get the "scheme" corresponding to the input parameters. + + :param dist_name: the name of the package to retrieve the scheme for, used + in the headers scheme path + :param user: indicates to use the "user" scheme + :param home: indicates to use the "home" scheme + :param root: root under which other directories are re-based + :param isolated: ignored, but kept for distutils compatibility (where + this controls whether the user-site pydistutils.cfg is honored) + :param prefix: indicates to use the "prefix" scheme and provides the + base directory for the same + """ + if user and prefix: + raise InvalidSchemeCombination("--user", "--prefix") + if home and prefix: + raise InvalidSchemeCombination("--home", "--prefix") + + if home is not None: + scheme_name = _infer_home() + elif user: + scheme_name = _infer_user() + else: + scheme_name = _infer_prefix() + + # Special case: When installing into a custom prefix, use posix_prefix + # instead of osx_framework_library. See _should_use_osx_framework_prefix() + # docstring for details. + if prefix is not None and scheme_name == "osx_framework_library": + scheme_name = "posix_prefix" + + if home is not None: + variables = {k: home for k in _HOME_KEYS} + elif prefix is not None: + variables = {k: prefix for k in _HOME_KEYS} + else: + variables = {} + + paths = sysconfig.get_paths(scheme=scheme_name, vars=variables) + + # Logic here is very arbitrary, we're doing it for compatibility, don't ask. + # 1. Pip historically uses a special header path in virtual environments. + # 2. If the distribution name is not known, distutils uses 'UNKNOWN'. We + # only do the same when not running in a virtual environment because + # pip's historical header path logic (see point 1) did not do this. + if running_under_virtualenv(): + if user: + base = variables.get("userbase", sys.prefix) + else: + base = variables.get("base", sys.prefix) + python_xy = f"python{get_major_minor_version()}" + paths["include"] = os.path.join(base, "include", "site", python_xy) + elif not dist_name: + dist_name = "UNKNOWN" + + scheme = Scheme( + platlib=paths["platlib"], + purelib=paths["purelib"], + headers=os.path.join(paths["include"], dist_name), + scripts=paths["scripts"], + data=paths["data"], + ) + if root is not None: + converted_keys = {} + for key in SCHEME_KEYS: + converted_keys[key] = change_root(root, getattr(scheme, key)) + scheme = Scheme(**converted_keys) + return scheme + + +def get_bin_prefix() -> str: + # Forcing to use /usr/local/bin for standard macOS framework installs. + if sys.platform[:6] == "darwin" and sys.prefix[:16] == "/System/Library/": + return "/usr/local/bin" + return sysconfig.get_paths()["scripts"] + + +def get_purelib() -> str: + return sysconfig.get_paths()["purelib"] + + +def get_platlib() -> str: + return sysconfig.get_paths()["platlib"] diff --git a/src/fetchcode/vcs/pip/_internal/locations/base.py b/src/fetchcode/vcs/pip/_internal/locations/base.py new file mode 100644 index 00000000..3f9f896e --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/locations/base.py @@ -0,0 +1,81 @@ +import functools +import os +import site +import sys +import sysconfig +import typing + +from pip._internal.exceptions import InstallationError +from pip._internal.utils import appdirs +from pip._internal.utils.virtualenv import running_under_virtualenv + +# Application Directories +USER_CACHE_DIR = appdirs.user_cache_dir("pip") + +# FIXME doesn't account for venv linked to global site-packages +site_packages: str = sysconfig.get_path("purelib") + + +def get_major_minor_version() -> str: + """ + Return the major-minor version of the current Python as a string, e.g. + "3.7" or "3.10". + """ + return "{}.{}".format(*sys.version_info) + + +def change_root(new_root: str, pathname: str) -> str: + """Return 'pathname' with 'new_root' prepended. + + If 'pathname' is relative, this is equivalent to os.path.join(new_root, pathname). + Otherwise, it requires making 'pathname' relative and then joining the + two, which is tricky on DOS/Windows and Mac OS. + + This is borrowed from Python's standard library's distutils module. + """ + if os.name == "posix": + if not os.path.isabs(pathname): + return os.path.join(new_root, pathname) + else: + return os.path.join(new_root, pathname[1:]) + + elif os.name == "nt": + (drive, path) = os.path.splitdrive(pathname) + if path[0] == "\\": + path = path[1:] + return os.path.join(new_root, path) + + else: + raise InstallationError( + f"Unknown platform: {os.name}\n" + "Can not change root path prefix on unknown platform." + ) + + +def get_src_prefix() -> str: + if running_under_virtualenv(): + src_prefix = os.path.join(sys.prefix, "src") + else: + # FIXME: keep src in cwd for now (it is not a temporary folder) + try: + src_prefix = os.path.join(os.getcwd(), "src") + except OSError: + # In case the current working directory has been renamed or deleted + sys.exit("The folder you are executing pip from can no longer be found.") + + # under macOS + virtualenv sys.prefix is not properly resolved + # it is something like /path/to/python/bin/.. + return os.path.abspath(src_prefix) + + +try: + # Use getusersitepackages if this is present, as it ensures that the + # value is initialised properly. + user_site: typing.Optional[str] = site.getusersitepackages() +except AttributeError: + user_site = site.USER_SITE + + +@functools.lru_cache(maxsize=None) +def is_osx_framework() -> bool: + return bool(sysconfig.get_config_var("PYTHONFRAMEWORK")) diff --git a/src/fetchcode/vcs/pip/_internal/main.py b/src/fetchcode/vcs/pip/_internal/main.py index 93c10650..33c6d24c 100644 --- a/src/fetchcode/vcs/pip/_internal/main.py +++ b/src/fetchcode/vcs/pip/_internal/main.py @@ -1,16 +1,12 @@ -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from typing import List, Optional -if MYPY_CHECK_RUNNING: - from typing import Optional, List - -def main(args=None): - # type: (Optional[List[str]]) -> int +def main(args: Optional[List[str]] = None) -> int: """This is preserved for old console scripts that may still be referencing it. For additional details, see https://github.com/pypa/pip/issues/7498. """ - from fetchcode.vcs.pip._internal.utils.entrypoints import _wrapper + from pip._internal.utils.entrypoints import _wrapper return _wrapper(args) diff --git a/src/fetchcode/vcs/pip/_internal/metadata/__init__.py b/src/fetchcode/vcs/pip/_internal/metadata/__init__.py new file mode 100644 index 00000000..aa232b6c --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/metadata/__init__.py @@ -0,0 +1,128 @@ +import contextlib +import functools +import os +import sys +from typing import TYPE_CHECKING, List, Optional, Type, cast + +from pip._internal.utils.misc import strtobool + +from .base import BaseDistribution, BaseEnvironment, FilesystemWheel, MemoryWheel, Wheel + +if TYPE_CHECKING: + from typing import Literal, Protocol +else: + Protocol = object + +__all__ = [ + "BaseDistribution", + "BaseEnvironment", + "FilesystemWheel", + "MemoryWheel", + "Wheel", + "get_default_environment", + "get_environment", + "get_wheel_distribution", + "select_backend", +] + + +def _should_use_importlib_metadata() -> bool: + """Whether to use the ``importlib.metadata`` or ``pkg_resources`` backend. + + By default, pip uses ``importlib.metadata`` on Python 3.11+, and + ``pkg_resourcess`` otherwise. This can be overridden by a couple of ways: + + * If environment variable ``_PIP_USE_IMPORTLIB_METADATA`` is set, it + dictates whether ``importlib.metadata`` is used, regardless of Python + version. + * On Python 3.11+, Python distributors can patch ``importlib.metadata`` + to add a global constant ``_PIP_USE_IMPORTLIB_METADATA = False``. This + makes pip use ``pkg_resources`` (unless the user set the aforementioned + environment variable to *True*). + """ + with contextlib.suppress(KeyError, ValueError): + return bool(strtobool(os.environ["_PIP_USE_IMPORTLIB_METADATA"])) + if sys.version_info < (3, 11): + return False + import importlib.metadata + + return bool(getattr(importlib.metadata, "_PIP_USE_IMPORTLIB_METADATA", True)) + + +class Backend(Protocol): + NAME: 'Literal["importlib", "pkg_resources"]' + Distribution: Type[BaseDistribution] + Environment: Type[BaseEnvironment] + + +@functools.lru_cache(maxsize=None) +def select_backend() -> Backend: + if _should_use_importlib_metadata(): + from . import importlib + + return cast(Backend, importlib) + from . import pkg_resources + + return cast(Backend, pkg_resources) + + +def get_default_environment() -> BaseEnvironment: + """Get the default representation for the current environment. + + This returns an Environment instance from the chosen backend. The default + Environment instance should be built from ``sys.path`` and may use caching + to share instance state accorss calls. + """ + return select_backend().Environment.default() + + +def get_environment(paths: Optional[List[str]]) -> BaseEnvironment: + """Get a representation of the environment specified by ``paths``. + + This returns an Environment instance from the chosen backend based on the + given import paths. The backend must build a fresh instance representing + the state of installed distributions when this function is called. + """ + return select_backend().Environment.from_paths(paths) + + +def get_directory_distribution(directory: str) -> BaseDistribution: + """Get the distribution metadata representation in the specified directory. + + This returns a Distribution instance from the chosen backend based on + the given on-disk ``.dist-info`` directory. + """ + return select_backend().Distribution.from_directory(directory) + + +def get_wheel_distribution(wheel: Wheel, canonical_name: str) -> BaseDistribution: + """Get the representation of the specified wheel's distribution metadata. + + This returns a Distribution instance from the chosen backend based on + the given wheel's ``.dist-info`` directory. + + :param canonical_name: Normalized project name of the given wheel. + """ + return select_backend().Distribution.from_wheel(wheel, canonical_name) + + +def get_metadata_distribution( + metadata_contents: bytes, + filename: str, + canonical_name: str, +) -> BaseDistribution: + """Get the dist representation of the specified METADATA file contents. + + This returns a Distribution instance from the chosen backend sourced from the data + in `metadata_contents`. + + :param metadata_contents: Contents of a METADATA file within a dist, or one served + via PEP 658. + :param filename: Filename for the dist this metadata represents. + :param canonical_name: Normalized project name of the given dist. + """ + return select_backend().Distribution.from_metadata_file_contents( + metadata_contents, + filename, + canonical_name, + ) diff --git a/src/fetchcode/vcs/pip/_internal/metadata/_json.py b/src/fetchcode/vcs/pip/_internal/metadata/_json.py new file mode 100644 index 00000000..9097dd58 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/metadata/_json.py @@ -0,0 +1,84 @@ +# Extracted from https://github.com/pfmoore/pkg_metadata + +from email.header import Header, decode_header, make_header +from email.message import Message +from typing import Any, Dict, List, Union, cast + +METADATA_FIELDS = [ + # Name, Multiple-Use + ("Metadata-Version", False), + ("Name", False), + ("Version", False), + ("Dynamic", True), + ("Platform", True), + ("Supported-Platform", True), + ("Summary", False), + ("Description", False), + ("Description-Content-Type", False), + ("Keywords", False), + ("Home-page", False), + ("Download-URL", False), + ("Author", False), + ("Author-email", False), + ("Maintainer", False), + ("Maintainer-email", False), + ("License", False), + ("Classifier", True), + ("Requires-Dist", True), + ("Requires-Python", False), + ("Requires-External", True), + ("Project-URL", True), + ("Provides-Extra", True), + ("Provides-Dist", True), + ("Obsoletes-Dist", True), +] + + +def json_name(field: str) -> str: + return field.lower().replace("-", "_") + + +def msg_to_json(msg: Message) -> Dict[str, Any]: + """Convert a Message object into a JSON-compatible dictionary.""" + + def sanitise_header(h: Union[Header, str]) -> str: + if isinstance(h, Header): + chunks = [] + for bytes, encoding in decode_header(h): + if encoding == "unknown-8bit": + try: + # See if UTF-8 works + bytes.decode("utf-8") + encoding = "utf-8" + except UnicodeDecodeError: + # If not, latin1 at least won't fail + encoding = "latin1" + chunks.append((bytes, encoding)) + return str(make_header(chunks)) + return str(h) + + result = {} + for field, multi in METADATA_FIELDS: + if field not in msg: + continue + key = json_name(field) + if multi: + value: Union[str, List[str]] = [ + sanitise_header(v) for v in msg.get_all(field) # type: ignore + ] + else: + value = sanitise_header(msg.get(field)) # type: ignore + if key == "keywords": + # Accept both comma-separated and space-separated + # forms, for better compatibility with old data. + if "," in value: + value = [v.strip() for v in value.split(",")] + else: + value = value.split() + result[key] = value + + payload = cast(str, msg.get_payload()) + if payload: + result["description"] = payload + + return result diff --git a/src/fetchcode/vcs/pip/_internal/metadata/base.py b/src/fetchcode/vcs/pip/_internal/metadata/base.py new file mode 100644 index 00000000..9eabcdb2 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/metadata/base.py @@ -0,0 +1,688 @@ +import csv +import email.message +import functools +import json +import logging +import pathlib +import re +import zipfile +from typing import ( + IO, + Any, + Collection, + Container, + Dict, + Iterable, + Iterator, + List, + NamedTuple, + Optional, + Protocol, + Tuple, + Union, +) + +from pip._vendor.packaging.requirements import Requirement +from pip._vendor.packaging.specifiers import InvalidSpecifier, SpecifierSet +from pip._vendor.packaging.utils import NormalizedName, canonicalize_name +from pip._vendor.packaging.version import Version + +from pip._internal.exceptions import NoneMetadataError +from pip._internal.locations import site_packages, user_site +from pip._internal.models.direct_url import ( + DIRECT_URL_METADATA_NAME, + DirectUrl, + DirectUrlValidationError, +) +from pip._internal.utils.compat import stdlib_pkgs # TODO: Move definition here. +from pip._internal.utils.egg_link import egg_link_path_from_sys_path +from pip._internal.utils.misc import is_local, normalize_path +from pip._internal.utils.urls import url_to_path + +from ._json import msg_to_json + +InfoPath = Union[str, pathlib.PurePath] + +logger = logging.getLogger(__name__) + + +class BaseEntryPoint(Protocol): + @property + def name(self) -> str: + raise NotImplementedError() + + @property + def value(self) -> str: + raise NotImplementedError() + + @property + def group(self) -> str: + raise NotImplementedError() + + +def _convert_installed_files_path( + entry: Tuple[str, ...], + info: Tuple[str, ...], +) -> str: + """Convert a legacy installed-files.txt path into modern RECORD path. + + The legacy format stores paths relative to the info directory, while the + modern format stores paths relative to the package root, e.g. the + site-packages directory. + + :param entry: Path parts of the installed-files.txt entry. + :param info: Path parts of the egg-info directory relative to package root. + :returns: The converted entry. + + For best compatibility with symlinks, this does not use ``abspath()`` or + ``Path.resolve()``, but tries to work with path parts: + + 1. While ``entry`` starts with ``..``, remove the equal amounts of parts + from ``info``; if ``info`` is empty, start appending ``..`` instead. + 2. Join the two directly. + """ + while entry and entry[0] == "..": + if not info or info[-1] == "..": + info += ("..",) + else: + info = info[:-1] + entry = entry[1:] + return str(pathlib.Path(*info, *entry)) + + +class RequiresEntry(NamedTuple): + requirement: str + extra: str + marker: str + + +class BaseDistribution(Protocol): + @classmethod + def from_directory(cls, directory: str) -> "BaseDistribution": + """Load the distribution from a metadata directory. + + :param directory: Path to a metadata directory, e.g. ``.dist-info``. + """ + raise NotImplementedError() + + @classmethod + def from_metadata_file_contents( + cls, + metadata_contents: bytes, + filename: str, + project_name: str, + ) -> "BaseDistribution": + """Load the distribution from the contents of a METADATA file. + + This is used to implement PEP 658 by generating a "shallow" dist object that can + be used for resolution without downloading or building the actual dist yet. + + :param metadata_contents: The contents of a METADATA file. + :param filename: File name for the dist with this metadata. + :param project_name: Name of the project this dist represents. + """ + raise NotImplementedError() + + @classmethod + def from_wheel(cls, wheel: "Wheel", name: str) -> "BaseDistribution": + """Load the distribution from a given wheel. + + :param wheel: A concrete wheel definition. + :param name: File name of the wheel. + + :raises InvalidWheel: Whenever loading of the wheel causes a + :py:exc:`zipfile.BadZipFile` exception to be thrown. + :raises UnsupportedWheel: If the wheel is a valid zip, but malformed + internally. + """ + raise NotImplementedError() + + def __repr__(self) -> str: + return f"{self.raw_name} {self.raw_version} ({self.location})" + + def __str__(self) -> str: + return f"{self.raw_name} {self.raw_version}" + + @property + def location(self) -> Optional[str]: + """Where the distribution is loaded from. + + A string value is not necessarily a filesystem path, since distributions + can be loaded from other sources, e.g. arbitrary zip archives. ``None`` + means the distribution is created in-memory. + + Do not canonicalize this value with e.g. ``pathlib.Path.resolve()``. If + this is a symbolic link, we want to preserve the relative path between + it and files in the distribution. + """ + raise NotImplementedError() + + @property + def editable_project_location(self) -> Optional[str]: + """The project location for editable distributions. + + This is the directory where pyproject.toml or setup.py is located. + None if the distribution is not installed in editable mode. + """ + # TODO: this property is relatively costly to compute, memoize it ? + direct_url = self.direct_url + if direct_url: + if direct_url.is_local_editable(): + return url_to_path(direct_url.url) + else: + # Search for an .egg-link file by walking sys.path, as it was + # done before by dist_is_editable(). + egg_link_path = egg_link_path_from_sys_path(self.raw_name) + if egg_link_path: + # TODO: get project location from second line of egg_link file + # (https://github.com/pypa/pip/issues/10243) + return self.location + return None + + @property + def installed_location(self) -> Optional[str]: + """The distribution's "installed" location. + + This should generally be a ``site-packages`` directory. This is + usually ``dist.location``, except for legacy develop-installed packages, + where ``dist.location`` is the source code location, and this is where + the ``.egg-link`` file is. + + The returned location is normalized (in particular, with symlinks removed). + """ + raise NotImplementedError() + + @property + def info_location(self) -> Optional[str]: + """Location of the .[egg|dist]-info directory or file. + + Similarly to ``location``, a string value is not necessarily a + filesystem path. ``None`` means the distribution is created in-memory. + + For a modern .dist-info installation on disk, this should be something + like ``{location}/{raw_name}-{version}.dist-info``. + + Do not canonicalize this value with e.g. ``pathlib.Path.resolve()``. If + this is a symbolic link, we want to preserve the relative path between + it and other files in the distribution. + """ + raise NotImplementedError() + + @property + def installed_by_distutils(self) -> bool: + """Whether this distribution is installed with legacy distutils format. + + A distribution installed with "raw" distutils not patched by setuptools + uses one single file at ``info_location`` to store metadata. We need to + treat this specially on uninstallation. + """ + info_location = self.info_location + if not info_location: + return False + return pathlib.Path(info_location).is_file() + + @property + def installed_as_egg(self) -> bool: + """Whether this distribution is installed as an egg. + + This usually indicates the distribution was installed by (older versions + of) easy_install. + """ + location = self.location + if not location: + return False + return location.endswith(".egg") + + @property + def installed_with_setuptools_egg_info(self) -> bool: + """Whether this distribution is installed with the ``.egg-info`` format. + + This usually indicates the distribution was installed with setuptools + with an old pip version or with ``single-version-externally-managed``. + + Note that this ensure the metadata store is a directory. distutils can + also installs an ``.egg-info``, but as a file, not a directory. This + property is *False* for that case. Also see ``installed_by_distutils``. + """ + info_location = self.info_location + if not info_location: + return False + if not info_location.endswith(".egg-info"): + return False + return pathlib.Path(info_location).is_dir() + + @property + def installed_with_dist_info(self) -> bool: + """Whether this distribution is installed with the "modern format". + + This indicates a "modern" installation, e.g. storing metadata in the + ``.dist-info`` directory. This applies to installations made by + setuptools (but through pip, not directly), or anything using the + standardized build backend interface (PEP 517). + """ + info_location = self.info_location + if not info_location: + return False + if not info_location.endswith(".dist-info"): + return False + return pathlib.Path(info_location).is_dir() + + @property + def canonical_name(self) -> NormalizedName: + raise NotImplementedError() + + @property + def version(self) -> Version: + raise NotImplementedError() + + @property + def raw_version(self) -> str: + raise NotImplementedError() + + @property + def setuptools_filename(self) -> str: + """Convert a project name to its setuptools-compatible filename. + + This is a copy of ``pkg_resources.to_filename()`` for compatibility. + """ + return self.raw_name.replace("-", "_") + + @property + def direct_url(self) -> Optional[DirectUrl]: + """Obtain a DirectUrl from this distribution. + + Returns None if the distribution has no `direct_url.json` metadata, + or if `direct_url.json` is invalid. + """ + try: + content = self.read_text(DIRECT_URL_METADATA_NAME) + except FileNotFoundError: + return None + try: + return DirectUrl.from_json(content) + except ( + UnicodeDecodeError, + json.JSONDecodeError, + DirectUrlValidationError, + ) as e: + logger.warning( + "Error parsing %s for %s: %s", + DIRECT_URL_METADATA_NAME, + self.canonical_name, + e, + ) + return None + + @property + def installer(self) -> str: + try: + installer_text = self.read_text("INSTALLER") + except (OSError, ValueError, NoneMetadataError): + return "" # Fail silently if the installer file cannot be read. + for line in installer_text.splitlines(): + cleaned_line = line.strip() + if cleaned_line: + return cleaned_line + return "" + + @property + def requested(self) -> bool: + return self.is_file("REQUESTED") + + @property + def editable(self) -> bool: + return bool(self.editable_project_location) + + @property + def local(self) -> bool: + """If distribution is installed in the current virtual environment. + + Always True if we're not in a virtualenv. + """ + if self.installed_location is None: + return False + return is_local(self.installed_location) + + @property + def in_usersite(self) -> bool: + if self.installed_location is None or user_site is None: + return False + return self.installed_location.startswith(normalize_path(user_site)) + + @property + def in_site_packages(self) -> bool: + if self.installed_location is None or site_packages is None: + return False + return self.installed_location.startswith(normalize_path(site_packages)) + + def is_file(self, path: InfoPath) -> bool: + """Check whether an entry in the info directory is a file.""" + raise NotImplementedError() + + def iter_distutils_script_names(self) -> Iterator[str]: + """Find distutils 'scripts' entries metadata. + + If 'scripts' is supplied in ``setup.py``, distutils records those in the + installed distribution's ``scripts`` directory, a file for each script. + """ + raise NotImplementedError() + + def read_text(self, path: InfoPath) -> str: + """Read a file in the info directory. + + :raise FileNotFoundError: If ``path`` does not exist in the directory. + :raise NoneMetadataError: If ``path`` exists in the info directory, but + cannot be read. + """ + raise NotImplementedError() + + def iter_entry_points(self) -> Iterable[BaseEntryPoint]: + raise NotImplementedError() + + def _metadata_impl(self) -> email.message.Message: + raise NotImplementedError() + + @functools.cached_property + def metadata(self) -> email.message.Message: + """Metadata of distribution parsed from e.g. METADATA or PKG-INFO. + + This should return an empty message if the metadata file is unavailable. + + :raises NoneMetadataError: If the metadata file is available, but does + not contain valid metadata. + """ + metadata = self._metadata_impl() + self._add_egg_info_requires(metadata) + return metadata + + @property + def metadata_dict(self) -> Dict[str, Any]: + """PEP 566 compliant JSON-serializable representation of METADATA or PKG-INFO. + + This should return an empty dict if the metadata file is unavailable. + + :raises NoneMetadataError: If the metadata file is available, but does + not contain valid metadata. + """ + return msg_to_json(self.metadata) + + @property + def metadata_version(self) -> Optional[str]: + """Value of "Metadata-Version:" in distribution metadata, if available.""" + return self.metadata.get("Metadata-Version") + + @property + def raw_name(self) -> str: + """Value of "Name:" in distribution metadata.""" + # The metadata should NEVER be missing the Name: key, but if it somehow + # does, fall back to the known canonical name. + return self.metadata.get("Name", self.canonical_name) + + @property + def requires_python(self) -> SpecifierSet: + """Value of "Requires-Python:" in distribution metadata. + + If the key does not exist or contains an invalid value, an empty + SpecifierSet should be returned. + """ + value = self.metadata.get("Requires-Python") + if value is None: + return SpecifierSet() + try: + # Convert to str to satisfy the type checker; this can be a Header object. + spec = SpecifierSet(str(value)) + except InvalidSpecifier as e: + message = "Package %r has an invalid Requires-Python: %s" + logger.warning(message, self.raw_name, e) + return SpecifierSet() + return spec + + def iter_dependencies(self, extras: Collection[str] = ()) -> Iterable[Requirement]: + """Dependencies of this distribution. + + For modern .dist-info distributions, this is the collection of + "Requires-Dist:" entries in distribution metadata. + """ + raise NotImplementedError() + + def iter_raw_dependencies(self) -> Iterable[str]: + """Raw Requires-Dist metadata.""" + return self.metadata.get_all("Requires-Dist", []) + + def iter_provided_extras(self) -> Iterable[NormalizedName]: + """Extras provided by this distribution. + + For modern .dist-info distributions, this is the collection of + "Provides-Extra:" entries in distribution metadata. + + The return value of this function is expected to be normalised names, + per PEP 685, with the returned value being handled appropriately by + `iter_dependencies`. + """ + raise NotImplementedError() + + def _iter_declared_entries_from_record(self) -> Optional[Iterator[str]]: + try: + text = self.read_text("RECORD") + except FileNotFoundError: + return None + # This extra Path-str cast normalizes entries. + return (str(pathlib.Path(row[0])) for row in csv.reader(text.splitlines())) + + def _iter_declared_entries_from_legacy(self) -> Optional[Iterator[str]]: + try: + text = self.read_text("installed-files.txt") + except FileNotFoundError: + return None + paths = (p for p in text.splitlines(keepends=False) if p) + root = self.location + info = self.info_location + if root is None or info is None: + return paths + try: + info_rel = pathlib.Path(info).relative_to(root) + except ValueError: # info is not relative to root. + return paths + if not info_rel.parts: # info *is* root. + return paths + return ( + _convert_installed_files_path(pathlib.Path(p).parts, info_rel.parts) + for p in paths + ) + + def iter_declared_entries(self) -> Optional[Iterator[str]]: + """Iterate through file entries declared in this distribution. + + For modern .dist-info distributions, this is the files listed in the + ``RECORD`` metadata file. For legacy setuptools distributions, this + comes from ``installed-files.txt``, with entries normalized to be + compatible with the format used by ``RECORD``. + + :return: An iterator for listed entries, or None if the distribution + contains neither ``RECORD`` nor ``installed-files.txt``. + """ + return ( + self._iter_declared_entries_from_record() + or self._iter_declared_entries_from_legacy() + ) + + def _iter_requires_txt_entries(self) -> Iterator[RequiresEntry]: + """Parse a ``requires.txt`` in an egg-info directory. + + This is an INI-ish format where an egg-info stores dependencies. A + section name describes extra other environment markers, while each entry + is an arbitrary string (not a key-value pair) representing a dependency + as a requirement string (no markers). + + There is a construct in ``importlib.metadata`` called ``Sectioned`` that + does mostly the same, but the format is currently considered private. + """ + try: + content = self.read_text("requires.txt") + except FileNotFoundError: + return + extra = marker = "" # Section-less entries don't have markers. + for line in content.splitlines(): + line = line.strip() + if not line or line.startswith("#"): # Comment; ignored. + continue + if line.startswith("[") and line.endswith("]"): # A section header. + extra, _, marker = line.strip("[]").partition(":") + continue + yield RequiresEntry(requirement=line, extra=extra, marker=marker) + + def _iter_egg_info_extras(self) -> Iterable[str]: + """Get extras from the egg-info directory.""" + known_extras = {""} + for entry in self._iter_requires_txt_entries(): + extra = canonicalize_name(entry.extra) + if extra in known_extras: + continue + known_extras.add(extra) + yield extra + + def _iter_egg_info_dependencies(self) -> Iterable[str]: + """Get distribution dependencies from the egg-info directory. + + To ease parsing, this converts a legacy dependency entry into a PEP 508 + requirement string. Like ``_iter_requires_txt_entries()``, there is code + in ``importlib.metadata`` that does mostly the same, but not do exactly + what we need. + + Namely, ``importlib.metadata`` does not normalize the extra name before + putting it into the requirement string, which causes marker comparison + to fail because the dist-info format do normalize. This is consistent in + all currently available PEP 517 backends, although not standardized. + """ + for entry in self._iter_requires_txt_entries(): + extra = canonicalize_name(entry.extra) + if extra and entry.marker: + marker = f'({entry.marker}) and extra == "{extra}"' + elif extra: + marker = f'extra == "{extra}"' + elif entry.marker: + marker = entry.marker + else: + marker = "" + if marker: + yield f"{entry.requirement} ; {marker}" + else: + yield entry.requirement + + def _add_egg_info_requires(self, metadata: email.message.Message) -> None: + """Add egg-info requires.txt information to the metadata.""" + if not metadata.get_all("Requires-Dist"): + for dep in self._iter_egg_info_dependencies(): + metadata["Requires-Dist"] = dep + if not metadata.get_all("Provides-Extra"): + for extra in self._iter_egg_info_extras(): + metadata["Provides-Extra"] = extra + + +class BaseEnvironment: + """An environment containing distributions to introspect.""" + + @classmethod + def default(cls) -> "BaseEnvironment": + raise NotImplementedError() + + @classmethod + def from_paths(cls, paths: Optional[List[str]]) -> "BaseEnvironment": + raise NotImplementedError() + + def get_distribution(self, name: str) -> Optional["BaseDistribution"]: + """Given a requirement name, return the installed distributions. + + The name may not be normalized. The implementation must canonicalize + it for lookup. + """ + raise NotImplementedError() + + def _iter_distributions(self) -> Iterator["BaseDistribution"]: + """Iterate through installed distributions. + + This function should be implemented by subclass, but never called + directly. Use the public ``iter_distribution()`` instead, which + implements additional logic to make sure the distributions are valid. + """ + raise NotImplementedError() + + def iter_all_distributions(self) -> Iterator[BaseDistribution]: + """Iterate through all installed distributions without any filtering.""" + for dist in self._iter_distributions(): + # Make sure the distribution actually comes from a valid Python + # packaging distribution. Pip's AdjacentTempDirectory leaves folders + # e.g. ``~atplotlib.dist-info`` if cleanup was interrupted. The + # valid project name pattern is taken from PEP 508. + project_name_valid = re.match( + r"^([A-Z0-9]|[A-Z0-9][A-Z0-9._-]*[A-Z0-9])$", + dist.canonical_name, + flags=re.IGNORECASE, + ) + if not project_name_valid: + logger.warning( + "Ignoring invalid distribution %s (%s)", + dist.canonical_name, + dist.location, + ) + continue + yield dist + + def iter_installed_distributions( + self, + local_only: bool = True, + skip: Container[str] = stdlib_pkgs, + include_editables: bool = True, + editables_only: bool = False, + user_only: bool = False, + ) -> Iterator[BaseDistribution]: + """Return a list of installed distributions. + + This is based on ``iter_all_distributions()`` with additional filtering + options. Note that ``iter_installed_distributions()`` without arguments + is *not* equal to ``iter_all_distributions()``, since some of the + configurations exclude packages by default. + + :param local_only: If True (default), only return installations + local to the current virtualenv, if in a virtualenv. + :param skip: An iterable of canonicalized project names to ignore; + defaults to ``stdlib_pkgs``. + :param include_editables: If False, don't report editables. + :param editables_only: If True, only report editables. + :param user_only: If True, only report installations in the user + site directory. + """ + it = self.iter_all_distributions() + if local_only: + it = (d for d in it if d.local) + if not include_editables: + it = (d for d in it if not d.editable) + if editables_only: + it = (d for d in it if d.editable) + if user_only: + it = (d for d in it if d.in_usersite) + return (d for d in it if d.canonical_name not in skip) + + +class Wheel(Protocol): + location: str + + def as_zipfile(self) -> zipfile.ZipFile: + raise NotImplementedError() + + +class FilesystemWheel(Wheel): + def __init__(self, location: str) -> None: + self.location = location + + def as_zipfile(self) -> zipfile.ZipFile: + return zipfile.ZipFile(self.location, allowZip64=True) + + +class MemoryWheel(Wheel): + def __init__(self, location: str, stream: IO[bytes]) -> None: + self.location = location + self.stream = stream + + def as_zipfile(self) -> zipfile.ZipFile: + return zipfile.ZipFile(self.stream, allowZip64=True) diff --git a/src/fetchcode/vcs/pip/_internal/metadata/importlib/__init__.py b/src/fetchcode/vcs/pip/_internal/metadata/importlib/__init__.py new file mode 100644 index 00000000..a779138d --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/metadata/importlib/__init__.py @@ -0,0 +1,6 @@ +from ._dists import Distribution +from ._envs import Environment + +__all__ = ["NAME", "Distribution", "Environment"] + +NAME = "importlib" diff --git a/src/fetchcode/vcs/pip/_internal/metadata/importlib/_compat.py b/src/fetchcode/vcs/pip/_internal/metadata/importlib/_compat.py new file mode 100644 index 00000000..ec1e815c --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/metadata/importlib/_compat.py @@ -0,0 +1,85 @@ +import importlib.metadata +import os +from typing import Any, Optional, Protocol, Tuple, cast + +from pip._vendor.packaging.utils import NormalizedName, canonicalize_name + + +class BadMetadata(ValueError): + def __init__(self, dist: importlib.metadata.Distribution, *, reason: str) -> None: + self.dist = dist + self.reason = reason + + def __str__(self) -> str: + return f"Bad metadata in {self.dist} ({self.reason})" + + +class BasePath(Protocol): + """A protocol that various path objects conform. + + This exists because importlib.metadata uses both ``pathlib.Path`` and + ``zipfile.Path``, and we need a common base for type hints (Union does not + work well since ``zipfile.Path`` is too new for our linter setup). + + This does not mean to be exhaustive, but only contains things that present + in both classes *that we need*. + """ + + @property + def name(self) -> str: + raise NotImplementedError() + + @property + def parent(self) -> "BasePath": + raise NotImplementedError() + + +def get_info_location(d: importlib.metadata.Distribution) -> Optional[BasePath]: + """Find the path to the distribution's metadata directory. + + HACK: This relies on importlib.metadata's private ``_path`` attribute. Not + all distributions exist on disk, so importlib.metadata is correct to not + expose the attribute as public. But pip's code base is old and not as clean, + so we do this to avoid having to rewrite too many things. Hopefully we can + eliminate this some day. + """ + return getattr(d, "_path", None) + + +def parse_name_and_version_from_info_directory( + dist: importlib.metadata.Distribution, +) -> Tuple[Optional[str], Optional[str]]: + """Get a name and version from the metadata directory name. + + This is much faster than reading distribution metadata. + """ + info_location = get_info_location(dist) + if info_location is None: + return None, None + + stem, suffix = os.path.splitext(info_location.name) + if suffix == ".dist-info": + name, sep, version = stem.partition("-") + if sep: + return name, version + + if suffix == ".egg-info": + name = stem.split("-", 1)[0] + return name, None + + return None, None + + +def get_dist_canonical_name(dist: importlib.metadata.Distribution) -> NormalizedName: + """Get the distribution's normalized name. + + The ``name`` attribute is only available in Python 3.10 or later. We are + targeting exactly that, but Mypy does not know this. + """ + if name := parse_name_and_version_from_info_directory(dist)[0]: + return canonicalize_name(name) + + name = cast(Any, dist).name + if not isinstance(name, str): + raise BadMetadata(dist, reason="invalid metadata entry 'name'") + return canonicalize_name(name) diff --git a/src/fetchcode/vcs/pip/_internal/metadata/importlib/_dists.py b/src/fetchcode/vcs/pip/_internal/metadata/importlib/_dists.py new file mode 100644 index 00000000..36cd3262 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/metadata/importlib/_dists.py @@ -0,0 +1,221 @@ +import email.message +import importlib.metadata +import pathlib +import zipfile +from typing import ( + Collection, + Dict, + Iterable, + Iterator, + Mapping, + Optional, + Sequence, + cast, +) + +from pip._vendor.packaging.requirements import Requirement +from pip._vendor.packaging.utils import NormalizedName, canonicalize_name +from pip._vendor.packaging.version import Version +from pip._vendor.packaging.version import parse as parse_version + +from pip._internal.exceptions import InvalidWheel, UnsupportedWheel +from pip._internal.metadata.base import ( + BaseDistribution, + BaseEntryPoint, + InfoPath, + Wheel, +) +from pip._internal.utils.misc import normalize_path +from pip._internal.utils.packaging import get_requirement +from pip._internal.utils.temp_dir import TempDirectory +from pip._internal.utils.wheel import parse_wheel, read_wheel_metadata_file + +from ._compat import ( + BasePath, + get_dist_canonical_name, + parse_name_and_version_from_info_directory, +) + + +class WheelDistribution(importlib.metadata.Distribution): + """An ``importlib.metadata.Distribution`` read from a wheel. + + Although ``importlib.metadata.PathDistribution`` accepts ``zipfile.Path``, + its implementation is too "lazy" for pip's needs (we can't keep the ZipFile + handle open for the entire lifetime of the distribution object). + + This implementation eagerly reads the entire metadata directory into the + memory instead, and operates from that. + """ + + def __init__( + self, + files: Mapping[pathlib.PurePosixPath, bytes], + info_location: pathlib.PurePosixPath, + ) -> None: + self._files = files + self.info_location = info_location + + @classmethod + def from_zipfile( + cls, + zf: zipfile.ZipFile, + name: str, + location: str, + ) -> "WheelDistribution": + info_dir, _ = parse_wheel(zf, name) + paths = ( + (name, pathlib.PurePosixPath(name.split("/", 1)[-1])) + for name in zf.namelist() + if name.startswith(f"{info_dir}/") + ) + files = { + relpath: read_wheel_metadata_file(zf, fullpath) + for fullpath, relpath in paths + } + info_location = pathlib.PurePosixPath(location, info_dir) + return cls(files, info_location) + + def iterdir(self, path: InfoPath) -> Iterator[pathlib.PurePosixPath]: + # Only allow iterating through the metadata directory. + if pathlib.PurePosixPath(str(path)) in self._files: + return iter(self._files) + raise FileNotFoundError(path) + + def read_text(self, filename: str) -> Optional[str]: + try: + data = self._files[pathlib.PurePosixPath(filename)] + except KeyError: + return None + try: + text = data.decode("utf-8") + except UnicodeDecodeError as e: + wheel = self.info_location.parent + error = f"Error decoding metadata for {wheel}: {e} in {filename} file" + raise UnsupportedWheel(error) + return text + + +class Distribution(BaseDistribution): + def __init__( + self, + dist: importlib.metadata.Distribution, + info_location: Optional[BasePath], + installed_location: Optional[BasePath], + ) -> None: + self._dist = dist + self._info_location = info_location + self._installed_location = installed_location + + @classmethod + def from_directory(cls, directory: str) -> BaseDistribution: + info_location = pathlib.Path(directory) + dist = importlib.metadata.Distribution.at(info_location) + return cls(dist, info_location, info_location.parent) + + @classmethod + def from_metadata_file_contents( + cls, + metadata_contents: bytes, + filename: str, + project_name: str, + ) -> BaseDistribution: + # Generate temp dir to contain the metadata file, and write the file contents. + temp_dir = pathlib.Path( + TempDirectory(kind="metadata", globally_managed=True).path + ) + metadata_path = temp_dir / "METADATA" + metadata_path.write_bytes(metadata_contents) + # Construct dist pointing to the newly created directory. + dist = importlib.metadata.Distribution.at(metadata_path.parent) + return cls(dist, metadata_path.parent, None) + + @classmethod + def from_wheel(cls, wheel: Wheel, name: str) -> BaseDistribution: + try: + with wheel.as_zipfile() as zf: + dist = WheelDistribution.from_zipfile(zf, name, wheel.location) + except zipfile.BadZipFile as e: + raise InvalidWheel(wheel.location, name) from e + return cls(dist, dist.info_location, pathlib.PurePosixPath(wheel.location)) + + @property + def location(self) -> Optional[str]: + if self._info_location is None: + return None + return str(self._info_location.parent) + + @property + def info_location(self) -> Optional[str]: + if self._info_location is None: + return None + return str(self._info_location) + + @property + def installed_location(self) -> Optional[str]: + if self._installed_location is None: + return None + return normalize_path(str(self._installed_location)) + + @property + def canonical_name(self) -> NormalizedName: + return get_dist_canonical_name(self._dist) + + @property + def version(self) -> Version: + if version := parse_name_and_version_from_info_directory(self._dist)[1]: + return parse_version(version) + return parse_version(self._dist.version) + + @property + def raw_version(self) -> str: + return self._dist.version + + def is_file(self, path: InfoPath) -> bool: + return self._dist.read_text(str(path)) is not None + + def iter_distutils_script_names(self) -> Iterator[str]: + # A distutils installation is always "flat" (not in e.g. egg form), so + # if this distribution's info location is NOT a pathlib.Path (but e.g. + # zipfile.Path), it can never contain any distutils scripts. + if not isinstance(self._info_location, pathlib.Path): + return + for child in self._info_location.joinpath("scripts").iterdir(): + yield child.name + + def read_text(self, path: InfoPath) -> str: + content = self._dist.read_text(str(path)) + if content is None: + raise FileNotFoundError(path) + return content + + def iter_entry_points(self) -> Iterable[BaseEntryPoint]: + # importlib.metadata's EntryPoint structure sasitfies BaseEntryPoint. + return self._dist.entry_points + + def _metadata_impl(self) -> email.message.Message: + # From Python 3.10+, importlib.metadata declares PackageMetadata as the + # return type. This protocol is unfortunately a disaster now and misses + # a ton of fields that we need, including get() and get_payload(). We + # rely on the implementation that the object is actually a Message now, + # until upstream can improve the protocol. (python/cpython#94952) + return cast(email.message.Message, self._dist.metadata) + + def iter_provided_extras(self) -> Iterable[NormalizedName]: + return [ + canonicalize_name(extra) + for extra in self.metadata.get_all("Provides-Extra", []) + ] + + def iter_dependencies(self, extras: Collection[str] = ()) -> Iterable[Requirement]: + contexts: Sequence[Dict[str, str]] = [{"extra": e} for e in extras] + for req_string in self.metadata.get_all("Requires-Dist", []): + # strip() because email.message.Message.get_all() may return a leading \n + # in case a long header was wrapped. + req = get_requirement(req_string.strip()) + if not req.marker: + yield req + elif not extras and req.marker.evaluate({"extra": ""}): + yield req + elif any(req.marker.evaluate(context) for context in contexts): + yield req diff --git a/src/fetchcode/vcs/pip/_internal/metadata/importlib/_envs.py b/src/fetchcode/vcs/pip/_internal/metadata/importlib/_envs.py new file mode 100644 index 00000000..70cb7a60 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/metadata/importlib/_envs.py @@ -0,0 +1,189 @@ +import functools +import importlib.metadata +import logging +import os +import pathlib +import sys +import zipfile +import zipimport +from typing import Iterator, List, Optional, Sequence, Set, Tuple + +from pip._vendor.packaging.utils import NormalizedName, canonicalize_name + +from pip._internal.metadata.base import BaseDistribution, BaseEnvironment +from pip._internal.models.wheel import Wheel +from pip._internal.utils.deprecation import deprecated +from pip._internal.utils.filetypes import WHEEL_EXTENSION + +from ._compat import BadMetadata, BasePath, get_dist_canonical_name, get_info_location +from ._dists import Distribution + +logger = logging.getLogger(__name__) + + +def _looks_like_wheel(location: str) -> bool: + if not location.endswith(WHEEL_EXTENSION): + return False + if not os.path.isfile(location): + return False + if not Wheel.wheel_file_re.match(os.path.basename(location)): + return False + return zipfile.is_zipfile(location) + + +class _DistributionFinder: + """Finder to locate distributions. + + The main purpose of this class is to memoize found distributions' names, so + only one distribution is returned for each package name. At lot of pip code + assumes this (because it is setuptools's behavior), and not doing the same + can potentially cause a distribution in lower precedence path to override a + higher precedence one if the caller is not careful. + + Eventually we probably want to make it possible to see lower precedence + installations as well. It's useful feature, after all. + """ + + FoundResult = Tuple[importlib.metadata.Distribution, Optional[BasePath]] + + def __init__(self) -> None: + self._found_names: Set[NormalizedName] = set() + + def _find_impl(self, location: str) -> Iterator[FoundResult]: + """Find distributions in a location.""" + # Skip looking inside a wheel. Since a package inside a wheel is not + # always valid (due to .data directories etc.), its .dist-info entry + # should not be considered an installed distribution. + if _looks_like_wheel(location): + return + # To know exactly where we find a distribution, we have to feed in the + # paths one by one, instead of dumping the list to importlib.metadata. + for dist in importlib.metadata.distributions(path=[location]): + info_location = get_info_location(dist) + try: + name = get_dist_canonical_name(dist) + except BadMetadata as e: + logger.warning("Skipping %s due to %s", info_location, e.reason) + continue + if name in self._found_names: + continue + self._found_names.add(name) + yield dist, info_location + + def find(self, location: str) -> Iterator[BaseDistribution]: + """Find distributions in a location. + + The path can be either a directory, or a ZIP archive. + """ + for dist, info_location in self._find_impl(location): + if info_location is None: + installed_location: Optional[BasePath] = None + else: + installed_location = info_location.parent + yield Distribution(dist, info_location, installed_location) + + def find_linked(self, location: str) -> Iterator[BaseDistribution]: + """Read location in egg-link files and return distributions in there. + + The path should be a directory; otherwise this returns nothing. This + follows how setuptools does this for compatibility. The first non-empty + line in the egg-link is read as a path (resolved against the egg-link's + containing directory if relative). Distributions found at that linked + location are returned. + """ + path = pathlib.Path(location) + if not path.is_dir(): + return + for child in path.iterdir(): + if child.suffix != ".egg-link": + continue + with child.open() as f: + lines = (line.strip() for line in f) + target_rel = next((line for line in lines if line), "") + if not target_rel: + continue + target_location = str(path.joinpath(target_rel)) + for dist, info_location in self._find_impl(target_location): + yield Distribution(dist, info_location, path) + + def _find_eggs_in_dir(self, location: str) -> Iterator[BaseDistribution]: + from pip._vendor.pkg_resources import find_distributions + + from pip._internal.metadata import pkg_resources as legacy + + with os.scandir(location) as it: + for entry in it: + if not entry.name.endswith(".egg"): + continue + for dist in find_distributions(entry.path): + yield legacy.Distribution(dist) + + def _find_eggs_in_zip(self, location: str) -> Iterator[BaseDistribution]: + from pip._vendor.pkg_resources import find_eggs_in_zip + + from pip._internal.metadata import pkg_resources as legacy + + try: + importer = zipimport.zipimporter(location) + except zipimport.ZipImportError: + return + for dist in find_eggs_in_zip(importer, location): + yield legacy.Distribution(dist) + + def find_eggs(self, location: str) -> Iterator[BaseDistribution]: + """Find eggs in a location. + + This actually uses the old *pkg_resources* backend. We likely want to + deprecate this so we can eventually remove the *pkg_resources* + dependency entirely. Before that, this should first emit a deprecation + warning for some versions when using the fallback since importing + *pkg_resources* is slow for those who don't need it. + """ + if os.path.isdir(location): + yield from self._find_eggs_in_dir(location) + if zipfile.is_zipfile(location): + yield from self._find_eggs_in_zip(location) + + +@functools.lru_cache(maxsize=None) # Warn a distribution exactly once. +def _emit_egg_deprecation(location: Optional[str]) -> None: + deprecated( + reason=f"Loading egg at {location} is deprecated.", + replacement="to use pip for package installation", + gone_in="24.3", + issue=12330, + ) + + +class Environment(BaseEnvironment): + def __init__(self, paths: Sequence[str]) -> None: + self._paths = paths + + @classmethod + def default(cls) -> BaseEnvironment: + return cls(sys.path) + + @classmethod + def from_paths(cls, paths: Optional[List[str]]) -> BaseEnvironment: + if paths is None: + return cls(sys.path) + return cls(paths) + + def _iter_distributions(self) -> Iterator[BaseDistribution]: + finder = _DistributionFinder() + for location in self._paths: + yield from finder.find(location) + for dist in finder.find_eggs(location): + _emit_egg_deprecation(dist.location) + yield dist + # This must go last because that's how pkg_resources tie-breaks. + yield from finder.find_linked(location) + + def get_distribution(self, name: str) -> Optional[BaseDistribution]: + canonical_name = canonicalize_name(name) + matches = ( + distribution + for distribution in self.iter_all_distributions() + if distribution.canonical_name == canonical_name + ) + return next(matches, None) diff --git a/src/fetchcode/vcs/pip/_internal/metadata/pkg_resources.py b/src/fetchcode/vcs/pip/_internal/metadata/pkg_resources.py new file mode 100644 index 00000000..4ea84f93 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/metadata/pkg_resources.py @@ -0,0 +1,301 @@ +import email.message +import email.parser +import logging +import os +import zipfile +from typing import ( + Collection, + Iterable, + Iterator, + List, + Mapping, + NamedTuple, + Optional, +) + +from pip._vendor import pkg_resources +from pip._vendor.packaging.requirements import Requirement +from pip._vendor.packaging.utils import NormalizedName, canonicalize_name +from pip._vendor.packaging.version import Version +from pip._vendor.packaging.version import parse as parse_version + +from pip._internal.exceptions import InvalidWheel, NoneMetadataError, UnsupportedWheel +from pip._internal.utils.egg_link import egg_link_path_from_location +from pip._internal.utils.misc import display_path, normalize_path +from pip._internal.utils.wheel import parse_wheel, read_wheel_metadata_file + +from .base import ( + BaseDistribution, + BaseEntryPoint, + BaseEnvironment, + InfoPath, + Wheel, +) + +__all__ = ["NAME", "Distribution", "Environment"] + +logger = logging.getLogger(__name__) + +NAME = "pkg_resources" + + +class EntryPoint(NamedTuple): + name: str + value: str + group: str + + +class InMemoryMetadata: + """IMetadataProvider that reads metadata files from a dictionary. + + This also maps metadata decoding exceptions to our internal exception type. + """ + + def __init__(self, metadata: Mapping[str, bytes], wheel_name: str) -> None: + self._metadata = metadata + self._wheel_name = wheel_name + + def has_metadata(self, name: str) -> bool: + return name in self._metadata + + def get_metadata(self, name: str) -> str: + try: + return self._metadata[name].decode() + except UnicodeDecodeError as e: + # Augment the default error with the origin of the file. + raise UnsupportedWheel( + f"Error decoding metadata for {self._wheel_name}: {e} in {name} file" + ) + + def get_metadata_lines(self, name: str) -> Iterable[str]: + return pkg_resources.yield_lines(self.get_metadata(name)) + + def metadata_isdir(self, name: str) -> bool: + return False + + def metadata_listdir(self, name: str) -> List[str]: + return [] + + def run_script(self, script_name: str, namespace: str) -> None: + pass + + +class Distribution(BaseDistribution): + def __init__(self, dist: pkg_resources.Distribution) -> None: + self._dist = dist + # This is populated lazily, to avoid loading metadata for all possible + # distributions eagerly. + self.__extra_mapping: Optional[Mapping[NormalizedName, str]] = None + + @property + def _extra_mapping(self) -> Mapping[NormalizedName, str]: + if self.__extra_mapping is None: + self.__extra_mapping = { + canonicalize_name(extra): extra for extra in self._dist.extras + } + + return self.__extra_mapping + + @classmethod + def from_directory(cls, directory: str) -> BaseDistribution: + dist_dir = directory.rstrip(os.sep) + + # Build a PathMetadata object, from path to metadata. :wink: + base_dir, dist_dir_name = os.path.split(dist_dir) + metadata = pkg_resources.PathMetadata(base_dir, dist_dir) + + # Determine the correct Distribution object type. + if dist_dir.endswith(".egg-info"): + dist_cls = pkg_resources.Distribution + dist_name = os.path.splitext(dist_dir_name)[0] + else: + assert dist_dir.endswith(".dist-info") + dist_cls = pkg_resources.DistInfoDistribution + dist_name = os.path.splitext(dist_dir_name)[0].split("-")[0] + + dist = dist_cls(base_dir, project_name=dist_name, metadata=metadata) + return cls(dist) + + @classmethod + def from_metadata_file_contents( + cls, + metadata_contents: bytes, + filename: str, + project_name: str, + ) -> BaseDistribution: + metadata_dict = { + "METADATA": metadata_contents, + } + dist = pkg_resources.DistInfoDistribution( + location=filename, + metadata=InMemoryMetadata(metadata_dict, filename), + project_name=project_name, + ) + return cls(dist) + + @classmethod + def from_wheel(cls, wheel: Wheel, name: str) -> BaseDistribution: + try: + with wheel.as_zipfile() as zf: + info_dir, _ = parse_wheel(zf, name) + metadata_dict = { + path.split("/", 1)[-1]: read_wheel_metadata_file(zf, path) + for path in zf.namelist() + if path.startswith(f"{info_dir}/") + } + except zipfile.BadZipFile as e: + raise InvalidWheel(wheel.location, name) from e + except UnsupportedWheel as e: + raise UnsupportedWheel(f"{name} has an invalid wheel, {e}") + dist = pkg_resources.DistInfoDistribution( + location=wheel.location, + metadata=InMemoryMetadata(metadata_dict, wheel.location), + project_name=name, + ) + return cls(dist) + + @property + def location(self) -> Optional[str]: + return self._dist.location + + @property + def installed_location(self) -> Optional[str]: + egg_link = egg_link_path_from_location(self.raw_name) + if egg_link: + location = egg_link + elif self.location: + location = self.location + else: + return None + return normalize_path(location) + + @property + def info_location(self) -> Optional[str]: + return self._dist.egg_info + + @property + def installed_by_distutils(self) -> bool: + # A distutils-installed distribution is provided by FileMetadata. This + # provider has a "path" attribute not present anywhere else. Not the + # best introspection logic, but pip has been doing this for a long time. + try: + return bool(self._dist._provider.path) + except AttributeError: + return False + + @property + def canonical_name(self) -> NormalizedName: + return canonicalize_name(self._dist.project_name) + + @property + def version(self) -> Version: + return parse_version(self._dist.version) + + @property + def raw_version(self) -> str: + return self._dist.version + + def is_file(self, path: InfoPath) -> bool: + return self._dist.has_metadata(str(path)) + + def iter_distutils_script_names(self) -> Iterator[str]: + yield from self._dist.metadata_listdir("scripts") + + def read_text(self, path: InfoPath) -> str: + name = str(path) + if not self._dist.has_metadata(name): + raise FileNotFoundError(name) + content = self._dist.get_metadata(name) + if content is None: + raise NoneMetadataError(self, name) + return content + + def iter_entry_points(self) -> Iterable[BaseEntryPoint]: + for group, entries in self._dist.get_entry_map().items(): + for name, entry_point in entries.items(): + name, _, value = str(entry_point).partition("=") + yield EntryPoint(name=name.strip(), value=value.strip(), group=group) + + def _metadata_impl(self) -> email.message.Message: + """ + :raises NoneMetadataError: if the distribution reports `has_metadata()` + True but `get_metadata()` returns None. + """ + if isinstance(self._dist, pkg_resources.DistInfoDistribution): + metadata_name = "METADATA" + else: + metadata_name = "PKG-INFO" + try: + metadata = self.read_text(metadata_name) + except FileNotFoundError: + if self.location: + displaying_path = display_path(self.location) + else: + displaying_path = repr(self.location) + logger.warning("No metadata found in %s", displaying_path) + metadata = "" + feed_parser = email.parser.FeedParser() + feed_parser.feed(metadata) + return feed_parser.close() + + def iter_dependencies(self, extras: Collection[str] = ()) -> Iterable[Requirement]: + if extras: + relevant_extras = set(self._extra_mapping) & set( + map(canonicalize_name, extras) + ) + extras = [self._extra_mapping[extra] for extra in relevant_extras] + return self._dist.requires(extras) + + def iter_provided_extras(self) -> Iterable[NormalizedName]: + return self._extra_mapping.keys() + + +class Environment(BaseEnvironment): + def __init__(self, ws: pkg_resources.WorkingSet) -> None: + self._ws = ws + + @classmethod + def default(cls) -> BaseEnvironment: + return cls(pkg_resources.working_set) + + @classmethod + def from_paths(cls, paths: Optional[List[str]]) -> BaseEnvironment: + return cls(pkg_resources.WorkingSet(paths)) + + def _iter_distributions(self) -> Iterator[BaseDistribution]: + for dist in self._ws: + yield Distribution(dist) + + def _search_distribution(self, name: str) -> Optional[BaseDistribution]: + """Find a distribution matching the ``name`` in the environment. + + This searches from *all* distributions available in the environment, to + match the behavior of ``pkg_resources.get_distribution()``. + """ + canonical_name = canonicalize_name(name) + for dist in self.iter_all_distributions(): + if dist.canonical_name == canonical_name: + return dist + return None + + def get_distribution(self, name: str) -> Optional[BaseDistribution]: + # Search the distribution by looking through the working set. + dist = self._search_distribution(name) + if dist: + return dist + + # If distribution could not be found, call working_set.require to + # update the working set, and try to find the distribution again. + # This might happen for e.g. when you install a package twice, once + # using setup.py develop and again using setup.py install. Now when + # running pip uninstall twice, the package gets removed from the + # working set in the first uninstall, so we have to populate the + # working set again so that pip knows about it and the packages gets + # picked up and is successfully uninstalled the second time too. + try: + # We didn't pass in any version specifiers, so this can never + # raise pkg_resources.VersionConflict. + self._ws.require(name) + except pkg_resources.DistributionNotFound: + return None + return self._search_distribution(name) diff --git a/src/fetchcode/vcs/pip/_internal/models/candidate.py b/src/fetchcode/vcs/pip/_internal/models/candidate.py index f643d2d3..f27f2831 100644 --- a/src/fetchcode/vcs/pip/_internal/models/candidate.py +++ b/src/fetchcode/vcs/pip/_internal/models/candidate.py @@ -1,36 +1,25 @@ -from fetchcode.vcs.pip._vendor.packaging.version import parse as parse_version - -from fetchcode.vcs.pip._internal.utils.models import KeyBasedCompareMixin -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from fetchcode.vcs.pip._vendor.packaging.version import _BaseVersion - from fetchcode.vcs.pip._internal.models.link import Link - - -class InstallationCandidate(KeyBasedCompareMixin): - """Represents a potential "candidate" for installation. - """ - - def __init__(self, name, version, link): - # type: (str, str, Link) -> None - self.name = name - self.version = parse_version(version) # type: _BaseVersion - self.link = link - - super(InstallationCandidate, self).__init__( - key=(self.name, self.version, self.link), - defining_class=InstallationCandidate - ) - - def __repr__(self): - # type: () -> str - return "".format( - self.name, self.version, self.link, - ) - - def __str__(self): - # type: () -> str - return '{!r} candidate (version {} at {})'.format( - self.name, self.version, self.link, - ) +from dataclasses import dataclass + +from pip._vendor.packaging.version import Version +from pip._vendor.packaging.version import parse as parse_version + +from pip._internal.models.link import Link + + +@dataclass(frozen=True) +class InstallationCandidate: + """Represents a potential "candidate" for installation.""" + + __slots__ = ["name", "version", "link"] + + name: str + version: Version + link: Link + + def __init__(self, name: str, version: str, link: Link) -> None: + object.__setattr__(self, "name", name) + object.__setattr__(self, "version", parse_version(version)) + object.__setattr__(self, "link", link) + + def __str__(self) -> str: + return f"{self.name!r} candidate (version {self.version} at {self.link})" diff --git a/src/fetchcode/vcs/pip/_internal/models/direct_url.py b/src/fetchcode/vcs/pip/_internal/models/direct_url.py index 3030f9d1..fc5ec8d4 100644 --- a/src/fetchcode/vcs/pip/_internal/models/direct_url.py +++ b/src/fetchcode/vcs/pip/_internal/models/direct_url.py @@ -1,22 +1,10 @@ """ PEP 610 """ + import json import re - -from fetchcode.vcs.pip._vendor import six -from fetchcode.vcs.pip._vendor.six.moves.urllib import parse as urllib_parse - -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import ( - Any, Dict, Iterable, Optional, Type, TypeVar, Union - ) - - T = TypeVar("T") - - -DIRECT_URL_METADATA_NAME = "direct_url.json" -ENV_VAR_RE = re.compile(r"^\$\{[A-Za-z0-9-_]+\}(:\$\{[A-Za-z0-9-_]+\})?$") +import urllib.parse +from dataclasses import dataclass +from typing import Any, ClassVar, Dict, Iterable, Optional, Type, TypeVar, Union __all__ = [ "DirectUrl", @@ -26,38 +14,40 @@ "VcsInfo", ] +T = TypeVar("T") + +DIRECT_URL_METADATA_NAME = "direct_url.json" +ENV_VAR_RE = re.compile(r"^\$\{[A-Za-z0-9-_]+\}(:\$\{[A-Za-z0-9-_]+\})?$") + class DirectUrlValidationError(Exception): pass -def _get(d, expected_type, key, default=None): - # type: (Dict[str, Any], Type[T], str, Optional[T]) -> Optional[T] +def _get( + d: Dict[str, Any], expected_type: Type[T], key: str, default: Optional[T] = None +) -> Optional[T]: """Get value from dictionary and verify expected type.""" if key not in d: return default value = d[key] - if six.PY2 and expected_type is str: - expected_type = six.string_types # type: ignore if not isinstance(value, expected_type): raise DirectUrlValidationError( - "{!r} has unexpected type for {} (expected {})".format( - value, key, expected_type - ) + f"{value!r} has unexpected type for {key} (expected {expected_type})" ) return value -def _get_required(d, expected_type, key, default=None): - # type: (Dict[str, Any], Type[T], str, Optional[T]) -> T +def _get_required( + d: Dict[str, Any], expected_type: Type[T], key: str, default: Optional[T] = None +) -> T: value = _get(d, expected_type, key, default) if value is None: - raise DirectUrlValidationError("{} must have a value".format(key)) + raise DirectUrlValidationError(f"{key} must have a value") return value -def _exactly_one_of(infos): - # type: (Iterable[Optional[InfoType]]) -> InfoType +def _exactly_one_of(infos: Iterable[Optional["InfoType"]]) -> "InfoType": infos = [info for info in infos if info is not None] if not infos: raise DirectUrlValidationError( @@ -71,122 +61,114 @@ def _exactly_one_of(infos): return infos[0] -def _filter_none(**kwargs): - # type: (Any) -> Dict[str, Any] +def _filter_none(**kwargs: Any) -> Dict[str, Any]: """Make dict excluding None values.""" return {k: v for k, v in kwargs.items() if v is not None} -class VcsInfo(object): - name = "vcs_info" +@dataclass +class VcsInfo: + name: ClassVar = "vcs_info" - def __init__( - self, - vcs, # type: str - commit_id, # type: str - requested_revision=None, # type: Optional[str] - resolved_revision=None, # type: Optional[str] - resolved_revision_type=None, # type: Optional[str] - ): - self.vcs = vcs - self.requested_revision = requested_revision - self.commit_id = commit_id - self.resolved_revision = resolved_revision - self.resolved_revision_type = resolved_revision_type + vcs: str + commit_id: str + requested_revision: Optional[str] = None @classmethod - def _from_dict(cls, d): - # type: (Optional[Dict[str, Any]]) -> Optional[VcsInfo] + def _from_dict(cls, d: Optional[Dict[str, Any]]) -> Optional["VcsInfo"]: if d is None: return None return cls( vcs=_get_required(d, str, "vcs"), commit_id=_get_required(d, str, "commit_id"), requested_revision=_get(d, str, "requested_revision"), - resolved_revision=_get(d, str, "resolved_revision"), - resolved_revision_type=_get(d, str, "resolved_revision_type"), ) - def _to_dict(self): - # type: () -> Dict[str, Any] + def _to_dict(self) -> Dict[str, Any]: return _filter_none( vcs=self.vcs, requested_revision=self.requested_revision, commit_id=self.commit_id, - resolved_revision=self.resolved_revision, - resolved_revision_type=self.resolved_revision_type, ) -class ArchiveInfo(object): +class ArchiveInfo: name = "archive_info" def __init__( self, - hash=None, # type: Optional[str] - ): + hash: Optional[str] = None, + hashes: Optional[Dict[str, str]] = None, + ) -> None: + # set hashes before hash, since the hash setter will further populate hashes + self.hashes = hashes self.hash = hash + @property + def hash(self) -> Optional[str]: + return self._hash + + @hash.setter + def hash(self, value: Optional[str]) -> None: + if value is not None: + # Auto-populate the hashes key to upgrade to the new format automatically. + # We don't back-populate the legacy hash key from hashes. + try: + hash_name, hash_value = value.split("=", 1) + except ValueError: + raise DirectUrlValidationError( + f"invalid archive_info.hash format: {value!r}" + ) + if self.hashes is None: + self.hashes = {hash_name: hash_value} + elif hash_name not in self.hashes: + self.hashes = self.hashes.copy() + self.hashes[hash_name] = hash_value + self._hash = value + @classmethod - def _from_dict(cls, d): - # type: (Optional[Dict[str, Any]]) -> Optional[ArchiveInfo] + def _from_dict(cls, d: Optional[Dict[str, Any]]) -> Optional["ArchiveInfo"]: if d is None: return None - return cls(hash=_get(d, str, "hash")) + return cls(hash=_get(d, str, "hash"), hashes=_get(d, dict, "hashes")) - def _to_dict(self): - # type: () -> Dict[str, Any] - return _filter_none(hash=self.hash) + def _to_dict(self) -> Dict[str, Any]: + return _filter_none(hash=self.hash, hashes=self.hashes) -class DirInfo(object): - name = "dir_info" +@dataclass +class DirInfo: + name: ClassVar = "dir_info" - def __init__( - self, - editable=False, # type: bool - ): - self.editable = editable + editable: bool = False @classmethod - def _from_dict(cls, d): - # type: (Optional[Dict[str, Any]]) -> Optional[DirInfo] + def _from_dict(cls, d: Optional[Dict[str, Any]]) -> Optional["DirInfo"]: if d is None: return None - return cls( - editable=_get_required(d, bool, "editable", default=False) - ) + return cls(editable=_get_required(d, bool, "editable", default=False)) - def _to_dict(self): - # type: () -> Dict[str, Any] + def _to_dict(self) -> Dict[str, Any]: return _filter_none(editable=self.editable or None) -if MYPY_CHECK_RUNNING: - InfoType = Union[ArchiveInfo, DirInfo, VcsInfo] +InfoType = Union[ArchiveInfo, DirInfo, VcsInfo] -class DirectUrl(object): +@dataclass +class DirectUrl: + url: str + info: InfoType + subdirectory: Optional[str] = None - def __init__( - self, - url, # type: str - info, # type: InfoType - subdirectory=None, # type: Optional[str] - ): - self.url = url - self.info = info - self.subdirectory = subdirectory - - def _remove_auth_from_netloc(self, netloc): - # type: (str) -> str + def _remove_auth_from_netloc(self, netloc: str) -> str: if "@" not in netloc: return netloc user_pass, netloc_no_user_pass = netloc.split("@", 1) if ( - isinstance(self.info, VcsInfo) and - self.info.vcs == "git" and - user_pass == "git" + isinstance(self.info, VcsInfo) + and self.info.vcs == "git" + and user_pass == "git" ): return netloc if ENV_VAR_RE.match(user_pass): @@ -194,26 +176,23 @@ def _remove_auth_from_netloc(self, netloc): return netloc_no_user_pass @property - def redacted_url(self): - # type: () -> str + def redacted_url(self) -> str: """url with user:password part removed unless it is formed with environment variables as specified in PEP 610, or it is ``git`` in the case of a git URL. """ - purl = urllib_parse.urlsplit(self.url) + purl = urllib.parse.urlsplit(self.url) netloc = self._remove_auth_from_netloc(purl.netloc) - surl = urllib_parse.urlunsplit( + surl = urllib.parse.urlunsplit( (purl.scheme, netloc, purl.path, purl.query, purl.fragment) ) return surl - def validate(self): - # type: () -> None + def validate(self) -> None: self.from_dict(self.to_dict()) @classmethod - def from_dict(cls, d): - # type: (Dict[str, Any]) -> DirectUrl + def from_dict(cls, d: Dict[str, Any]) -> "DirectUrl": return DirectUrl( url=_get_required(d, str, "url"), subdirectory=_get(d, str, "subdirectory"), @@ -226,8 +205,7 @@ def from_dict(cls, d): ), ) - def to_dict(self): - # type: () -> Dict[str, Any] + def to_dict(self) -> Dict[str, Any]: res = _filter_none( url=self.redacted_url, subdirectory=self.subdirectory, @@ -236,10 +214,11 @@ def to_dict(self): return res @classmethod - def from_json(cls, s): - # type: (str) -> DirectUrl + def from_json(cls, s: str) -> "DirectUrl": return cls.from_dict(json.loads(s)) - def to_json(self): - # type: () -> str + def to_json(self) -> str: return json.dumps(self.to_dict(), sort_keys=True) + + def is_local_editable(self) -> bool: + return isinstance(self.info, DirInfo) and self.info.editable diff --git a/src/fetchcode/vcs/pip/_internal/models/format_control.py b/src/fetchcode/vcs/pip/_internal/models/format_control.py index 1db1130b..ccd11272 100644 --- a/src/fetchcode/vcs/pip/_internal/models/format_control.py +++ b/src/fetchcode/vcs/pip/_internal/models/format_control.py @@ -1,21 +1,20 @@ -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False +from typing import FrozenSet, Optional, Set -from fetchcode.vcs.pip._vendor.packaging.utils import canonicalize_name +from pip._vendor.packaging.utils import canonicalize_name -from fetchcode.vcs.pip._internal.exceptions import CommandError -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._internal.exceptions import CommandError -if MYPY_CHECK_RUNNING: - from typing import Optional, Set, FrozenSet +class FormatControl: + """Helper for managing formats from which a package can be installed.""" -class FormatControl(object): - """Helper for managing formats from which a package can be installed. - """ + __slots__ = ["no_binary", "only_binary"] - def __init__(self, no_binary=None, only_binary=None): - # type: (Optional[Set[str]], Optional[Set[str]]) -> None + def __init__( + self, + no_binary: Optional[Set[str]] = None, + only_binary: Optional[Set[str]] = None, + ) -> None: if no_binary is None: no_binary = set() if only_binary is None: @@ -24,61 +23,56 @@ def __init__(self, no_binary=None, only_binary=None): self.no_binary = no_binary self.only_binary = only_binary - def __eq__(self, other): - # type: (object) -> bool - return self.__dict__ == other.__dict__ + def __eq__(self, other: object) -> bool: + if not isinstance(other, self.__class__): + return NotImplemented - def __ne__(self, other): - # type: (object) -> bool - return not self.__eq__(other) + if self.__slots__ != other.__slots__: + return False - def __repr__(self): - # type: () -> str - return "{}({}, {})".format( - self.__class__.__name__, - self.no_binary, - self.only_binary - ) + return all(getattr(self, k) == getattr(other, k) for k in self.__slots__) + + def __repr__(self) -> str: + return f"{self.__class__.__name__}({self.no_binary}, {self.only_binary})" @staticmethod - def handle_mutual_excludes(value, target, other): - # type: (str, Optional[Set[str]], Optional[Set[str]]) -> None - if value.startswith('-'): + def handle_mutual_excludes(value: str, target: Set[str], other: Set[str]) -> None: + if value.startswith("-"): raise CommandError( "--no-binary / --only-binary option requires 1 argument." ) - new = value.split(',') - while ':all:' in new: + new = value.split(",") + while ":all:" in new: other.clear() target.clear() - target.add(':all:') - del new[:new.index(':all:') + 1] + target.add(":all:") + del new[: new.index(":all:") + 1] # Without a none, we want to discard everything as :all: covers it - if ':none:' not in new: + if ":none:" not in new: return for name in new: - if name == ':none:': + if name == ":none:": target.clear() continue name = canonicalize_name(name) other.discard(name) target.add(name) - def get_allowed_formats(self, canonical_name): - # type: (str) -> FrozenSet[str] + def get_allowed_formats(self, canonical_name: str) -> FrozenSet[str]: result = {"binary", "source"} if canonical_name in self.only_binary: - result.discard('source') + result.discard("source") elif canonical_name in self.no_binary: - result.discard('binary') - elif ':all:' in self.only_binary: - result.discard('source') - elif ':all:' in self.no_binary: - result.discard('binary') + result.discard("binary") + elif ":all:" in self.only_binary: + result.discard("source") + elif ":all:" in self.no_binary: + result.discard("binary") return frozenset(result) - def disallow_binaries(self): - # type: () -> None + def disallow_binaries(self) -> None: self.handle_mutual_excludes( - ':all:', self.no_binary, self.only_binary, + ":all:", + self.no_binary, + self.only_binary, ) diff --git a/src/fetchcode/vcs/pip/_internal/models/index.py b/src/fetchcode/vcs/pip/_internal/models/index.py index 845ac017..b94c3251 100644 --- a/src/fetchcode/vcs/pip/_internal/models/index.py +++ b/src/fetchcode/vcs/pip/_internal/models/index.py @@ -1,31 +1,28 @@ -from fetchcode.vcs.pip._vendor.six.moves.urllib import parse as urllib_parse +import urllib.parse -class PackageIndex(object): - """Represents a Package Index and provides easier access to endpoints - """ +class PackageIndex: + """Represents a Package Index and provides easier access to endpoints""" - def __init__(self, url, file_storage_domain): - # type: (str, str) -> None - super(PackageIndex, self).__init__() + __slots__ = ["url", "netloc", "simple_url", "pypi_url", "file_storage_domain"] + + def __init__(self, url: str, file_storage_domain: str) -> None: + super().__init__() self.url = url - self.netloc = urllib_parse.urlsplit(url).netloc - self.simple_url = self._url_for_path('simple') - self.pypi_url = self._url_for_path('pypi') + self.netloc = urllib.parse.urlsplit(url).netloc + self.simple_url = self._url_for_path("simple") + self.pypi_url = self._url_for_path("pypi") # This is part of a temporary hack used to block installs of PyPI # packages which depend on external urls only necessary until PyPI can # block such packages themselves self.file_storage_domain = file_storage_domain - def _url_for_path(self, path): - # type: (str) -> str - return urllib_parse.urljoin(self.url, path) + def _url_for_path(self, path: str) -> str: + return urllib.parse.urljoin(self.url, path) -PyPI = PackageIndex( - 'https://pypi.org/', file_storage_domain='files.pythonhosted.org' -) +PyPI = PackageIndex("https://pypi.org/", file_storage_domain="files.pythonhosted.org") TestPyPI = PackageIndex( - 'https://test.pypi.org/', file_storage_domain='test-files.pythonhosted.org' + "https://test.pypi.org/", file_storage_domain="test-files.pythonhosted.org" ) diff --git a/src/fetchcode/vcs/pip/_internal/models/installation_report.py b/src/fetchcode/vcs/pip/_internal/models/installation_report.py new file mode 100644 index 00000000..b9c6330d --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/models/installation_report.py @@ -0,0 +1,56 @@ +from typing import Any, Dict, Sequence + +from pip._vendor.packaging.markers import default_environment + +from pip import __version__ +from pip._internal.req.req_install import InstallRequirement + + +class InstallationReport: + def __init__(self, install_requirements: Sequence[InstallRequirement]): + self._install_requirements = install_requirements + + @classmethod + def _install_req_to_dict(cls, ireq: InstallRequirement) -> Dict[str, Any]: + assert ireq.download_info, f"No download_info for {ireq}" + res = { + # PEP 610 json for the download URL. download_info.archive_info.hashes may + # be absent when the requirement was installed from the wheel cache + # and the cache entry was populated by an older pip version that did not + # record origin.json. + "download_info": ireq.download_info.to_dict(), + # is_direct is true if the requirement was a direct URL reference (which + # includes editable requirements), and false if the requirement was + # downloaded from a PEP 503 index or --find-links. + "is_direct": ireq.is_direct, + # is_yanked is true if the requirement was yanked from the index, but + # was still selected by pip to conform to PEP 592. + "is_yanked": ireq.link.is_yanked if ireq.link else False, + # requested is true if the requirement was specified by the user (aka + # top level requirement), and false if it was installed as a dependency of a + # requirement. https://peps.python.org/pep-0376/#requested + "requested": ireq.user_supplied, + # PEP 566 json encoding for metadata + # https://www.python.org/dev/peps/pep-0566/#json-compatible-metadata + "metadata": ireq.get_dist().metadata_dict, + } + if ireq.user_supplied and ireq.extras: + # For top level requirements, the list of requested extras, if any. + res["requested_extras"] = sorted(ireq.extras) + return res + + def to_dict(self) -> Dict[str, Any]: + return { + "version": "1", + "pip_version": __version__, + "install": [ + self._install_req_to_dict(ireq) for ireq in self._install_requirements + ], + # https://peps.python.org/pep-0508/#environment-markers + # TODO: currently, the resolver uses the default environment to evaluate + # environment markers, so that is what we report here. In the future, it + # should also take into account options such as --python-version or + # --platform, perhaps under the form of an environment_override field? + # https://github.com/pypa/pip/issues/11198 + "environment": default_environment(), + } diff --git a/src/fetchcode/vcs/pip/_internal/models/link.py b/src/fetchcode/vcs/pip/_internal/models/link.py index 5ded7ae6..2f41f2f6 100644 --- a/src/fetchcode/vcs/pip/_internal/models/link.py +++ b/src/fetchcode/vcs/pip/_internal/models/link.py @@ -1,41 +1,212 @@ +import functools +import itertools +import logging import os import posixpath import re +import urllib.parse +from dataclasses import dataclass +from typing import ( + TYPE_CHECKING, + Any, + Dict, + List, + Mapping, + NamedTuple, + Optional, + Tuple, + Union, +) -from fetchcode.vcs.pip._vendor.six.moves.urllib import parse as urllib_parse - -from fetchcode.vcs.pip._internal.utils.filetypes import WHEEL_EXTENSION -from fetchcode.vcs.pip._internal.utils.misc import ( +from pip._internal.utils.deprecation import deprecated +from pip._internal.utils.filetypes import WHEEL_EXTENSION +from pip._internal.utils.hashes import Hashes +from pip._internal.utils.misc import ( + pairwise, redact_auth_from_url, split_auth_from_netloc, splitext, ) -from fetchcode.vcs.pip._internal.utils.models import KeyBasedCompareMixin -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.utils.urls import path_to_url, url_to_path +from pip._internal.utils.urls import path_to_url, url_to_path + +if TYPE_CHECKING: + from pip._internal.index.collector import IndexContent + +logger = logging.getLogger(__name__) + + +# Order matters, earlier hashes have a precedence over later hashes for what +# we will pick to use. +_SUPPORTED_HASHES = ("sha512", "sha384", "sha256", "sha224", "sha1", "md5") + + +@dataclass(frozen=True) +class LinkHash: + """Links to content may have embedded hash values. This class parses those. + + `name` must be any member of `_SUPPORTED_HASHES`. + + This class can be converted to and from `ArchiveInfo`. While ArchiveInfo intends to + be JSON-serializable to conform to PEP 610, this class contains the logic for + parsing a hash name and value for correctness, and then checking whether that hash + conforms to a schema with `.is_hash_allowed()`.""" + + name: str + value: str + + _hash_url_fragment_re = re.compile( + # NB: we do not validate that the second group (.*) is a valid hex + # digest. Instead, we simply keep that string in this class, and then check it + # against Hashes when hash-checking is needed. This is easier to debug than + # proactively discarding an invalid hex digest, as we handle incorrect hashes + # and malformed hashes in the same place. + r"[#&]({choices})=([^&]*)".format( + choices="|".join(re.escape(hash_name) for hash_name in _SUPPORTED_HASHES) + ), + ) + + def __post_init__(self) -> None: + assert self.name in _SUPPORTED_HASHES + + @classmethod + @functools.lru_cache(maxsize=None) + def find_hash_url_fragment(cls, url: str) -> Optional["LinkHash"]: + """Search a string for a checksum algorithm name and encoded output value.""" + match = cls._hash_url_fragment_re.search(url) + if match is None: + return None + name, value = match.groups() + return cls(name=name, value=value) + + def as_dict(self) -> Dict[str, str]: + return {self.name: self.value} + + def as_hashes(self) -> Hashes: + """Return a Hashes instance which checks only for the current hash.""" + return Hashes({self.name: [self.value]}) + + def is_hash_allowed(self, hashes: Optional[Hashes]) -> bool: + """ + Return True if the current hash is allowed by `hashes`. + """ + if hashes is None: + return False + return hashes.is_hash_allowed(self.name, hex_digest=self.value) + + +@dataclass(frozen=True) +class MetadataFile: + """Information about a core metadata file associated with a distribution.""" + + hashes: Optional[Dict[str, str]] + + def __post_init__(self) -> None: + if self.hashes is not None: + assert all(name in _SUPPORTED_HASHES for name in self.hashes) + + +def supported_hashes(hashes: Optional[Dict[str, str]]) -> Optional[Dict[str, str]]: + # Remove any unsupported hash types from the mapping. If this leaves no + # supported hashes, return None + if hashes is None: + return None + hashes = {n: v for n, v in hashes.items() if n in _SUPPORTED_HASHES} + if not hashes: + return None + return hashes + + +def _clean_url_path_part(part: str) -> str: + """ + Clean a "part" of a URL path (i.e. after splitting on "@" characters). + """ + # We unquote prior to quoting to make sure nothing is double quoted. + return urllib.parse.quote(urllib.parse.unquote(part)) + + +def _clean_file_url_path(part: str) -> str: + """ + Clean the first part of a URL path that corresponds to a local + filesystem path (i.e. the first part after splitting on "@" characters). + """ + # We unquote prior to quoting to make sure nothing is double quoted. + # Also, on Windows the path part might contain a drive letter which + # should not be quoted. On Linux where drive letters do not + # exist, the colon should be quoted. We rely on urllib.request + # to do the right thing here. + return urllib.request.pathname2url(urllib.request.url2pathname(part)) + -if MYPY_CHECK_RUNNING: - from typing import Optional, Text, Tuple, Union - from fetchcode.vcs.pip._internal.index.collector import HTMLPage - from fetchcode.vcs.pip._internal.utils.hashes import Hashes +# percent-encoded: / +_reserved_chars_re = re.compile("(@|%2F)", re.IGNORECASE) -class Link(KeyBasedCompareMixin): - """Represents a parsed link from a Package Index's simple URL +def _clean_url_path(path: str, is_local_path: bool) -> str: """ + Clean the path portion of a URL. + """ + if is_local_path: + clean_func = _clean_file_url_path + else: + clean_func = _clean_url_path_part + + # Split on the reserved characters prior to cleaning so that + # revision strings in VCS URLs are properly preserved. + parts = _reserved_chars_re.split(path) + + cleaned_parts = [] + for to_clean, reserved in pairwise(itertools.chain(parts, [""])): + cleaned_parts.append(clean_func(to_clean)) + # Normalize %xx escapes (e.g. %2f -> %2F) + cleaned_parts.append(reserved.upper()) + + return "".join(cleaned_parts) + + +def _ensure_quoted_url(url: str) -> str: + """ + Make sure a link is fully quoted. + For example, if ' ' occurs in the URL, it will be replaced with "%20", + and without double-quoting other characters. + """ + # Split the URL into parts according to the general structure + # `scheme://netloc/path;parameters?query#fragment`. + result = urllib.parse.urlparse(url) + # If the netloc is empty, then the URL refers to a local filesystem path. + is_local_path = not result.netloc + path = _clean_url_path(result.path, is_local_path=is_local_path) + return urllib.parse.urlunparse(result._replace(path=path)) + + +@functools.total_ordering +class Link: + """Represents a parsed link from a Package Index's simple URL""" + + __slots__ = [ + "_parsed_url", + "_url", + "_hashes", + "comes_from", + "requires_python", + "yanked_reason", + "metadata_file_data", + "cache_link_parsing", + "egg_fragment", + ] def __init__( self, - url, # type: str - comes_from=None, # type: Optional[Union[str, HTMLPage]] - requires_python=None, # type: Optional[str] - yanked_reason=None, # type: Optional[Text] - cache_link_parsing=True, # type: bool - ): - # type: (...) -> None + url: str, + comes_from: Optional[Union[str, "IndexContent"]] = None, + requires_python: Optional[str] = None, + yanked_reason: Optional[str] = None, + metadata_file_data: Optional[MetadataFile] = None, + cache_link_parsing: bool = True, + hashes: Optional[Mapping[str, str]] = None, + ) -> None: """ :param url: url of the resource pointed to (href of the link) - :param comes_from: instance of HTMLPage where the link was found, + :param comes_from: instance of IndexContent where the link was found, or string. :param requires_python: String containing the `Requires-Python` metadata field, specified in PEP 345. This may be specified by @@ -47,55 +218,181 @@ def __init__( a simple repository HTML link. If the file has been yanked but no reason was provided, this should be the empty string. See PEP 592 for more information and the specification. + :param metadata_file_data: the metadata attached to the file, or None if + no such metadata is provided. This argument, if not None, indicates + that a separate metadata file exists, and also optionally supplies + hashes for that file. :param cache_link_parsing: A flag that is used elsewhere to determine - whether resources retrieved from this link - should be cached. PyPI index urls should - generally have this set to False, for - example. + whether resources retrieved from this link should be cached. PyPI + URLs should generally have this set to False, for example. + :param hashes: A mapping of hash names to digests to allow us to + determine the validity of a download. """ + # The comes_from, requires_python, and metadata_file_data arguments are + # only used by classmethods of this class, and are not used in client + # code directly. + # url can be a UNC windows share - if url.startswith('\\\\'): + if url.startswith("\\\\"): url = path_to_url(url) - self._parsed_url = urllib_parse.urlsplit(url) + self._parsed_url = urllib.parse.urlsplit(url) # Store the url as a private attribute to prevent accidentally # trying to set a new value. self._url = url + link_hash = LinkHash.find_hash_url_fragment(url) + hashes_from_link = {} if link_hash is None else link_hash.as_dict() + if hashes is None: + self._hashes = hashes_from_link + else: + self._hashes = {**hashes, **hashes_from_link} + self.comes_from = comes_from self.requires_python = requires_python if requires_python else None self.yanked_reason = yanked_reason - - super(Link, self).__init__(key=url, defining_class=Link) + self.metadata_file_data = metadata_file_data self.cache_link_parsing = cache_link_parsing + self.egg_fragment = self._egg_fragment() + + @classmethod + def from_json( + cls, + file_data: Dict[str, Any], + page_url: str, + ) -> Optional["Link"]: + """ + Convert an pypi json document from a simple repository page into a Link. + """ + file_url = file_data.get("url") + if file_url is None: + return None + + url = _ensure_quoted_url(urllib.parse.urljoin(page_url, file_url)) + pyrequire = file_data.get("requires-python") + yanked_reason = file_data.get("yanked") + hashes = file_data.get("hashes", {}) + + # PEP 714: Indexes must use the name core-metadata, but + # clients should support the old name as a fallback for compatibility. + metadata_info = file_data.get("core-metadata") + if metadata_info is None: + metadata_info = file_data.get("dist-info-metadata") + + # The metadata info value may be a boolean, or a dict of hashes. + if isinstance(metadata_info, dict): + # The file exists, and hashes have been supplied + metadata_file_data = MetadataFile(supported_hashes(metadata_info)) + elif metadata_info: + # The file exists, but there are no hashes + metadata_file_data = MetadataFile(None) + else: + # False or not present: the file does not exist + metadata_file_data = None + + # The Link.yanked_reason expects an empty string instead of a boolean. + if yanked_reason and not isinstance(yanked_reason, str): + yanked_reason = "" + # The Link.yanked_reason expects None instead of False. + elif not yanked_reason: + yanked_reason = None + + return cls( + url, + comes_from=page_url, + requires_python=pyrequire, + yanked_reason=yanked_reason, + hashes=hashes, + metadata_file_data=metadata_file_data, + ) + + @classmethod + def from_element( + cls, + anchor_attribs: Dict[str, Optional[str]], + page_url: str, + base_url: str, + ) -> Optional["Link"]: + """ + Convert an anchor element's attributes in a simple repository page to a Link. + """ + href = anchor_attribs.get("href") + if not href: + return None - def __str__(self): - # type: () -> str + url = _ensure_quoted_url(urllib.parse.urljoin(base_url, href)) + pyrequire = anchor_attribs.get("data-requires-python") + yanked_reason = anchor_attribs.get("data-yanked") + + # PEP 714: Indexes must use the name data-core-metadata, but + # clients should support the old name as a fallback for compatibility. + metadata_info = anchor_attribs.get("data-core-metadata") + if metadata_info is None: + metadata_info = anchor_attribs.get("data-dist-info-metadata") + # The metadata info value may be the string "true", or a string of + # the form "hashname=hashval" + if metadata_info == "true": + # The file exists, but there are no hashes + metadata_file_data = MetadataFile(None) + elif metadata_info is None: + # The file does not exist + metadata_file_data = None + else: + # The file exists, and hashes have been supplied + hashname, sep, hashval = metadata_info.partition("=") + if sep == "=": + metadata_file_data = MetadataFile(supported_hashes({hashname: hashval})) + else: + # Error - data is wrong. Treat as no hashes supplied. + logger.debug( + "Index returned invalid data-dist-info-metadata value: %s", + metadata_info, + ) + metadata_file_data = MetadataFile(None) + + return cls( + url, + comes_from=page_url, + requires_python=pyrequire, + yanked_reason=yanked_reason, + metadata_file_data=metadata_file_data, + ) + + def __str__(self) -> str: if self.requires_python: - rp = ' (requires-python:{})'.format(self.requires_python) + rp = f" (requires-python:{self.requires_python})" else: - rp = '' + rp = "" if self.comes_from: - return '{} (from {}){}'.format( - redact_auth_from_url(self._url), self.comes_from, rp) + return f"{redact_auth_from_url(self._url)} (from {self.comes_from}){rp}" else: return redact_auth_from_url(str(self._url)) - def __repr__(self): - # type: () -> str - return ''.format(self) + def __repr__(self) -> str: + return f"" + + def __hash__(self) -> int: + return hash(self.url) + + def __eq__(self, other: Any) -> bool: + if not isinstance(other, Link): + return NotImplemented + return self.url == other.url + + def __lt__(self, other: Any) -> bool: + if not isinstance(other, Link): + return NotImplemented + return self.url < other.url @property - def url(self): - # type: () -> str + def url(self) -> str: return self._url @property - def filename(self): - # type: () -> str - path = self.path.rstrip('/') + def filename(self) -> str: + path = self.path.rstrip("/") name = posixpath.basename(path) if not name: # Make sure we don't leak auth information if the netloc @@ -103,134 +400,191 @@ def filename(self): netloc, user_pass = split_auth_from_netloc(self.netloc) return netloc - name = urllib_parse.unquote(name) - assert name, ( - 'URL {self._url!r} produced no filename'.format(**locals())) + name = urllib.parse.unquote(name) + assert name, f"URL {self._url!r} produced no filename" return name @property - def file_path(self): - # type: () -> str + def file_path(self) -> str: return url_to_path(self.url) @property - def scheme(self): - # type: () -> str + def scheme(self) -> str: return self._parsed_url.scheme @property - def netloc(self): - # type: () -> str + def netloc(self) -> str: """ This can contain auth information. """ return self._parsed_url.netloc @property - def path(self): - # type: () -> str - return urllib_parse.unquote(self._parsed_url.path) + def path(self) -> str: + return urllib.parse.unquote(self._parsed_url.path) - def splitext(self): - # type: () -> Tuple[str, str] - return splitext(posixpath.basename(self.path.rstrip('/'))) + def splitext(self) -> Tuple[str, str]: + return splitext(posixpath.basename(self.path.rstrip("/"))) @property - def ext(self): - # type: () -> str + def ext(self) -> str: return self.splitext()[1] @property - def url_without_fragment(self): - # type: () -> str + def url_without_fragment(self) -> str: scheme, netloc, path, query, fragment = self._parsed_url - return urllib_parse.urlunsplit((scheme, netloc, path, query, None)) + return urllib.parse.urlunsplit((scheme, netloc, path, query, "")) - _egg_fragment_re = re.compile(r'[#&]egg=([^&]*)') + _egg_fragment_re = re.compile(r"[#&]egg=([^&]*)") - @property - def egg_fragment(self): - # type: () -> Optional[str] + # Per PEP 508. + _project_name_re = re.compile( + r"^([A-Z0-9]|[A-Z0-9][A-Z0-9._-]*[A-Z0-9])$", re.IGNORECASE + ) + + def _egg_fragment(self) -> Optional[str]: match = self._egg_fragment_re.search(self._url) if not match: return None - return match.group(1) - _subdirectory_fragment_re = re.compile(r'[#&]subdirectory=([^&]*)') + # An egg fragment looks like a PEP 508 project name, along with + # an optional extras specifier. Anything else is invalid. + project_name = match.group(1) + if not self._project_name_re.match(project_name): + deprecated( + reason=f"{self} contains an egg fragment with a non-PEP 508 name", + replacement="to use the req @ url syntax, and remove the egg fragment", + gone_in="25.0", + issue=11617, + ) + + return project_name + + _subdirectory_fragment_re = re.compile(r"[#&]subdirectory=([^&]*)") @property - def subdirectory_fragment(self): - # type: () -> Optional[str] + def subdirectory_fragment(self) -> Optional[str]: match = self._subdirectory_fragment_re.search(self._url) if not match: return None return match.group(1) - _hash_re = re.compile( - r'(sha1|sha224|sha384|sha256|sha512|md5)=([a-f0-9]+)' - ) + def metadata_link(self) -> Optional["Link"]: + """Return a link to the associated core metadata file (if any).""" + if self.metadata_file_data is None: + return None + metadata_url = f"{self.url_without_fragment}.metadata" + if self.metadata_file_data.hashes is None: + return Link(metadata_url) + return Link(metadata_url, hashes=self.metadata_file_data.hashes) + + def as_hashes(self) -> Hashes: + return Hashes({k: [v] for k, v in self._hashes.items()}) @property - def hash(self): - # type: () -> Optional[str] - match = self._hash_re.search(self._url) - if match: - return match.group(2) - return None + def hash(self) -> Optional[str]: + return next(iter(self._hashes.values()), None) @property - def hash_name(self): - # type: () -> Optional[str] - match = self._hash_re.search(self._url) - if match: - return match.group(1) - return None + def hash_name(self) -> Optional[str]: + return next(iter(self._hashes), None) @property - def show_url(self): - # type: () -> str - return posixpath.basename(self._url.split('#', 1)[0].split('?', 1)[0]) + def show_url(self) -> str: + return posixpath.basename(self._url.split("#", 1)[0].split("?", 1)[0]) @property - def is_file(self): - # type: () -> bool - return self.scheme == 'file' + def is_file(self) -> bool: + return self.scheme == "file" - def is_existing_dir(self): - # type: () -> bool + def is_existing_dir(self) -> bool: return self.is_file and os.path.isdir(self.file_path) @property - def is_wheel(self): - # type: () -> bool + def is_wheel(self) -> bool: return self.ext == WHEEL_EXTENSION @property - def is_vcs(self): - # type: () -> bool - from fetchcode.vcs.pip._internal.vcs import vcs + def is_vcs(self) -> bool: + from pip._internal.vcs import vcs return self.scheme in vcs.all_schemes @property - def is_yanked(self): - # type: () -> bool + def is_yanked(self) -> bool: return self.yanked_reason is not None @property - def has_hash(self): - # type: () -> bool - return self.hash_name is not None + def has_hash(self) -> bool: + return bool(self._hashes) - def is_hash_allowed(self, hashes): - # type: (Optional[Hashes]) -> bool + def is_hash_allowed(self, hashes: Optional[Hashes]) -> bool: """ - Return True if the link has a hash and it is allowed. + Return True if the link has a hash and it is allowed by `hashes`. """ - if hashes is None or not self.has_hash: + if hashes is None: return False - # Assert non-None so mypy knows self.hash_name and self.hash are str. - assert self.hash_name is not None - assert self.hash is not None + return any(hashes.is_hash_allowed(k, v) for k, v in self._hashes.items()) + + +class _CleanResult(NamedTuple): + """Convert link for equivalency check. + + This is used in the resolver to check whether two URL-specified requirements + likely point to the same distribution and can be considered equivalent. This + equivalency logic avoids comparing URLs literally, which can be too strict + (e.g. "a=1&b=2" vs "b=2&a=1") and produce conflicts unexpecting to users. + + Currently this does three things: + + 1. Drop the basic auth part. This is technically wrong since a server can + serve different content based on auth, but if it does that, it is even + impossible to guarantee two URLs without auth are equivalent, since + the user can input different auth information when prompted. So the + practical solution is to assume the auth doesn't affect the response. + 2. Parse the query to avoid the ordering issue. Note that ordering under the + same key in the query are NOT cleaned; i.e. "a=1&a=2" and "a=2&a=1" are + still considered different. + 3. Explicitly drop most of the fragment part, except ``subdirectory=`` and + hash values, since it should have no impact the downloaded content. Note + that this drops the "egg=" part historically used to denote the requested + project (and extras), which is wrong in the strictest sense, but too many + people are supplying it inconsistently to cause superfluous resolution + conflicts, so we choose to also ignore them. + """ + + parsed: urllib.parse.SplitResult + query: Dict[str, List[str]] + subdirectory: str + hashes: Dict[str, str] + + +def _clean_link(link: Link) -> _CleanResult: + parsed = link._parsed_url + netloc = parsed.netloc.rsplit("@", 1)[-1] + # According to RFC 8089, an empty host in file: means localhost. + if parsed.scheme == "file" and not netloc: + netloc = "localhost" + fragment = urllib.parse.parse_qs(parsed.fragment) + if "egg" in fragment: + logger.debug("Ignoring egg= fragment in %s", link) + try: + # If there are multiple subdirectory values, use the first one. + # This matches the behavior of Link.subdirectory_fragment. + subdirectory = fragment["subdirectory"][0] + except (IndexError, KeyError): + subdirectory = "" + # If there are multiple hash values under the same algorithm, use the + # first one. This matches the behavior of Link.hash_value. + hashes = {k: fragment[k][0] for k in _SUPPORTED_HASHES if k in fragment} + return _CleanResult( + parsed=parsed._replace(netloc=netloc, query="", fragment=""), + query=urllib.parse.parse_qs(parsed.query), + subdirectory=subdirectory, + hashes=hashes, + ) + - return hashes.is_hash_allowed(self.hash_name, hex_digest=self.hash) +@functools.lru_cache(maxsize=None) +def links_equivalent(link1: Link, link2: Link) -> bool: + return _clean_link(link1) == _clean_link(link2) diff --git a/src/fetchcode/vcs/pip/_internal/models/scheme.py b/src/fetchcode/vcs/pip/_internal/models/scheme.py index af07b407..06a9a550 100644 --- a/src/fetchcode/vcs/pip/_internal/models/scheme.py +++ b/src/fetchcode/vcs/pip/_internal/models/scheme.py @@ -5,21 +5,21 @@ https://docs.python.org/3/install/index.html#alternate-installation. """ +from dataclasses import dataclass -class Scheme(object): +SCHEME_KEYS = ["platlib", "purelib", "headers", "scripts", "data"] + + +@dataclass(frozen=True) +class Scheme: """A Scheme holds paths which are used as the base directories for artifacts associated with a Python package. """ - def __init__( - self, - platlib, # type: str - purelib, # type: str - headers, # type: str - scripts, # type: str - data, # type: str - ): - self.platlib = platlib - self.purelib = purelib - self.headers = headers - self.scripts = scripts - self.data = data + + __slots__ = SCHEME_KEYS + + platlib: str + purelib: str + headers: str + scripts: str + data: str diff --git a/src/fetchcode/vcs/pip/_internal/models/search_scope.py b/src/fetchcode/vcs/pip/_internal/models/search_scope.py index df32a080..ee7bc862 100644 --- a/src/fetchcode/vcs/pip/_internal/models/search_scope.py +++ b/src/fetchcode/vcs/pip/_internal/models/search_scope.py @@ -2,35 +2,38 @@ import logging import os import posixpath +import urllib.parse +from dataclasses import dataclass +from typing import List -from fetchcode.vcs.pip._vendor.packaging.utils import canonicalize_name -from fetchcode.vcs.pip._vendor.six.moves.urllib import parse as urllib_parse - -from fetchcode.vcs.pip._internal.models.index import PyPI -from fetchcode.vcs.pip._internal.utils.compat import has_tls -from fetchcode.vcs.pip._internal.utils.misc import normalize_path, redact_auth_from_url -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import List +from pip._vendor.packaging.utils import canonicalize_name +from pip._internal.models.index import PyPI +from pip._internal.utils.compat import has_tls +from pip._internal.utils.misc import normalize_path, redact_auth_from_url logger = logging.getLogger(__name__) -class SearchScope(object): - +@dataclass(frozen=True) +class SearchScope: """ Encapsulates the locations that pip is configured to search. """ + __slots__ = ["find_links", "index_urls", "no_index"] + + find_links: List[str] + index_urls: List[str] + no_index: bool + @classmethod def create( cls, - find_links, # type: List[str] - index_urls, # type: List[str] - ): - # type: (...) -> SearchScope + find_links: List[str], + index_urls: List[str], + no_index: bool, + ) -> "SearchScope": """ Create a SearchScope object after normalizing the `find_links`. """ @@ -39,9 +42,9 @@ def create( # it and if it exists, use the normalized version. # This is deliberately conservative - it might be fine just to # blindly normalize anything starting with a ~... - built_find_links = [] # type: List[str] + built_find_links: List[str] = [] for link in find_links: - if link.startswith('~'): + if link.startswith("~"): new_link = normalize_path(link) if os.path.exists(new_link): link = new_link @@ -51,40 +54,30 @@ def create( # relies on TLS. if not has_tls(): for link in itertools.chain(index_urls, built_find_links): - parsed = urllib_parse.urlparse(link) - if parsed.scheme == 'https': + parsed = urllib.parse.urlparse(link) + if parsed.scheme == "https": logger.warning( - 'pip is configured with locations that require ' - 'TLS/SSL, however the ssl module in Python is not ' - 'available.' + "pip is configured with locations that require " + "TLS/SSL, however the ssl module in Python is not " + "available." ) break return cls( find_links=built_find_links, index_urls=index_urls, + no_index=no_index, ) - def __init__( - self, - find_links, # type: List[str] - index_urls, # type: List[str] - ): - # type: (...) -> None - self.find_links = find_links - self.index_urls = index_urls - - def get_formatted_locations(self): - # type: () -> str + def get_formatted_locations(self) -> str: lines = [] redacted_index_urls = [] if self.index_urls and self.index_urls != [PyPI.simple_url]: for url in self.index_urls: - redacted_index_url = redact_auth_from_url(url) # Parse the URL - purl = urllib_parse.urlsplit(redacted_index_url) + purl = urllib.parse.urlsplit(redacted_index_url) # URL is generally invalid if scheme and netloc is missing # there are issues with Python and URL parsing, so this test @@ -93,41 +86,42 @@ def get_formatted_locations(self): # exceptions for malformed URLs if not purl.scheme and not purl.netloc: logger.warning( - 'The index url "{}" seems invalid, ' - 'please provide a scheme.'.format(redacted_index_url)) + 'The index url "%s" seems invalid, please provide a scheme.', + redacted_index_url, + ) redacted_index_urls.append(redacted_index_url) - lines.append('Looking in indexes: {}'.format( - ', '.join(redacted_index_urls))) + lines.append( + "Looking in indexes: {}".format(", ".join(redacted_index_urls)) + ) if self.find_links: lines.append( - 'Looking in links: {}'.format(', '.join( - redact_auth_from_url(url) for url in self.find_links)) + "Looking in links: {}".format( + ", ".join(redact_auth_from_url(url) for url in self.find_links) + ) ) - return '\n'.join(lines) + return "\n".join(lines) - def get_index_urls_locations(self, project_name): - # type: (str) -> List[str] + def get_index_urls_locations(self, project_name: str) -> List[str]: """Returns the locations found via self.index_urls Checks the url_name on the main (first in the list) index and use this url_name to produce all locations """ - def mkurl_pypi_url(url): - # type: (str) -> str + def mkurl_pypi_url(url: str) -> str: loc = posixpath.join( - url, - urllib_parse.quote(canonicalize_name(project_name))) + url, urllib.parse.quote(canonicalize_name(project_name)) + ) # For maximum compatibility with easy_install, ensure the path # ends in a trailing slash. Although this isn't in the spec # (and PyPI can handle it without the slash) some other index # implementations might break if they relied on easy_install's # behavior. - if not loc.endswith('/'): - loc = loc + '/' + if not loc.endswith("/"): + loc = loc + "/" return loc return [mkurl_pypi_url(url) for url in self.index_urls] diff --git a/src/fetchcode/vcs/pip/_internal/models/selection_prefs.py b/src/fetchcode/vcs/pip/_internal/models/selection_prefs.py index 87ff0f18..e9b50aa5 100644 --- a/src/fetchcode/vcs/pip/_internal/models/selection_prefs.py +++ b/src/fetchcode/vcs/pip/_internal/models/selection_prefs.py @@ -1,30 +1,36 @@ -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from typing import Optional -if MYPY_CHECK_RUNNING: - from typing import Optional - from fetchcode.vcs.pip._internal.models.format_control import FormatControl +from pip._internal.models.format_control import FormatControl -class SelectionPreferences(object): - +# TODO: This needs Python 3.10's improved slots support for dataclasses +# to be converted into a dataclass. +class SelectionPreferences: """ Encapsulates the candidate selection preferences for downloading and installing files. """ + __slots__ = [ + "allow_yanked", + "allow_all_prereleases", + "format_control", + "prefer_binary", + "ignore_requires_python", + ] + # Don't include an allow_yanked default value to make sure each call # site considers whether yanked releases are allowed. This also causes # that decision to be made explicit in the calling code, which helps # people when reading the code. def __init__( self, - allow_yanked, # type: bool - allow_all_prereleases=False, # type: bool - format_control=None, # type: Optional[FormatControl] - prefer_binary=False, # type: bool - ignore_requires_python=None, # type: Optional[bool] - ): - # type: (...) -> None + allow_yanked: bool, + allow_all_prereleases: bool = False, + format_control: Optional[FormatControl] = None, + prefer_binary: bool = False, + ignore_requires_python: Optional[bool] = None, + ) -> None: """Create a SelectionPreferences object. :param allow_yanked: Whether files marked as yanked (in the sense diff --git a/src/fetchcode/vcs/pip/_internal/models/target_python.py b/src/fetchcode/vcs/pip/_internal/models/target_python.py index 30a8e67d..88925a9f 100644 --- a/src/fetchcode/vcs/pip/_internal/models/target_python.py +++ b/src/fetchcode/vcs/pip/_internal/models/target_python.py @@ -1,44 +1,47 @@ import sys +from typing import List, Optional, Set, Tuple -from fetchcode.vcs.pip._internal.utils.compatibility_tags import ( - get_supported, - version_info_to_nodot, -) -from fetchcode.vcs.pip._internal.utils.misc import normalize_version_info -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._vendor.packaging.tags import Tag -if MYPY_CHECK_RUNNING: - from typing import List, Optional, Tuple +from pip._internal.utils.compatibility_tags import get_supported, version_info_to_nodot +from pip._internal.utils.misc import normalize_version_info - from fetchcode.vcs.pip._vendor.packaging.tags import Tag - - -class TargetPython(object): +class TargetPython: """ Encapsulates the properties of a Python interpreter one is targeting for a package install, download, etc. """ + __slots__ = [ + "_given_py_version_info", + "abis", + "implementation", + "platforms", + "py_version", + "py_version_info", + "_valid_tags", + "_valid_tags_set", + ] + def __init__( self, - platform=None, # type: Optional[str] - py_version_info=None, # type: Optional[Tuple[int, ...]] - abi=None, # type: Optional[str] - implementation=None, # type: Optional[str] - ): - # type: (...) -> None + platforms: Optional[List[str]] = None, + py_version_info: Optional[Tuple[int, ...]] = None, + abis: Optional[List[str]] = None, + implementation: Optional[str] = None, + ) -> None: """ - :param platform: A string or None. If None, searches for packages - that are supported by the current system. Otherwise, will find - packages that can be built on the platform passed in. These + :param platforms: A list of strings or None. If None, searches for + packages that are supported by the current system. Otherwise, will + find packages that can be built on the platforms passed in. These packages will only be downloaded for distribution: they will not be built locally. :param py_version_info: An optional tuple of ints representing the Python version information to use (e.g. `sys.version_info[:3]`). This can have length 1, 2, or 3 when provided. - :param abi: A string or None. This is passed to compatibility_tags.py's - get_supported() function as is. + :param abis: A list of strings or None. This is passed to + compatibility_tags.py's get_supported() function as is. :param implementation: A string or None. This is passed to compatibility_tags.py's get_supported() function as is. """ @@ -50,41 +53,39 @@ def __init__( else: py_version_info = normalize_version_info(py_version_info) - py_version = '.'.join(map(str, py_version_info[:2])) + py_version = ".".join(map(str, py_version_info[:2])) - self.abi = abi + self.abis = abis self.implementation = implementation - self.platform = platform + self.platforms = platforms self.py_version = py_version self.py_version_info = py_version_info - # This is used to cache the return value of get_tags(). - self._valid_tags = None # type: Optional[List[Tag]] + # This is used to cache the return value of get_(un)sorted_tags. + self._valid_tags: Optional[List[Tag]] = None + self._valid_tags_set: Optional[Set[Tag]] = None - def format_given(self): - # type: () -> str + def format_given(self) -> str: """ Format the given, non-None attributes for display. """ display_version = None if self._given_py_version_info is not None: - display_version = '.'.join( + display_version = ".".join( str(part) for part in self._given_py_version_info ) key_values = [ - ('platform', self.platform), - ('version_info', display_version), - ('abi', self.abi), - ('implementation', self.implementation), + ("platforms", self.platforms), + ("version_info", display_version), + ("abis", self.abis), + ("implementation", self.implementation), ] - return ' '.join( - '{}={!r}'.format(key, value) for key, value in key_values - if value is not None + return " ".join( + f"{key}={value!r}" for key, value in key_values if value is not None ) - def get_tags(self): - # type: () -> List[Tag] + def get_sorted_tags(self) -> List[Tag]: """ Return the supported PEP 425 tags to check wheel candidates against. @@ -101,10 +102,20 @@ def get_tags(self): tags = get_supported( version=version, - platform=self.platform, - abi=self.abi, + platforms=self.platforms, + abis=self.abis, impl=self.implementation, ) self._valid_tags = tags return self._valid_tags + + def get_unsorted_tags(self) -> Set[Tag]: + """Exactly the same as get_sorted_tags, but returns a set. + + This is important for performance. + """ + if self._valid_tags_set is None: + self._valid_tags_set = set(self.get_sorted_tags()) + + return self._valid_tags_set diff --git a/src/fetchcode/vcs/pip/_internal/models/wheel.py b/src/fetchcode/vcs/pip/_internal/models/wheel.py index 2bfbfb4c..36d4d2e7 100644 --- a/src/fetchcode/vcs/pip/_internal/models/wheel.py +++ b/src/fetchcode/vcs/pip/_internal/models/wheel.py @@ -1,60 +1,52 @@ """Represents a wheel file and provides access to the various parts of the name that have meaning. """ -import re -from fetchcode.vcs.pip._vendor.packaging.tags import Tag +import re +from typing import Dict, Iterable, List -from fetchcode.vcs.pip._internal.exceptions import InvalidWheelFilename -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._vendor.packaging.tags import Tag -if MYPY_CHECK_RUNNING: - from typing import List +from pip._internal.exceptions import InvalidWheelFilename -class Wheel(object): +class Wheel: """A wheel file""" wheel_file_re = re.compile( - r"""^(?P(?P.+?)-(?P.*?)) - ((-(?P\d[^-]*?))?-(?P.+?)-(?P.+?)-(?P.+?) + r"""^(?P(?P[^\s-]+?)-(?P[^\s-]*?)) + ((-(?P\d[^-]*?))?-(?P[^\s-]+?)-(?P[^\s-]+?)-(?P[^\s-]+?) \.whl|\.dist-info)$""", - re.VERBOSE + re.VERBOSE, ) - def __init__(self, filename): - # type: (str) -> None + def __init__(self, filename: str) -> None: """ :raises InvalidWheelFilename: when the filename is invalid for a wheel """ wheel_info = self.wheel_file_re.match(filename) if not wheel_info: - raise InvalidWheelFilename( - "{} is not a valid wheel filename.".format(filename) - ) + raise InvalidWheelFilename(f"{filename} is not a valid wheel filename.") self.filename = filename - self.name = wheel_info.group('name').replace('_', '-') + self.name = wheel_info.group("name").replace("_", "-") # we'll assume "_" means "-" due to wheel naming scheme # (https://github.com/pypa/pip/issues/1150) - self.version = wheel_info.group('ver').replace('_', '-') - self.build_tag = wheel_info.group('build') - self.pyversions = wheel_info.group('pyver').split('.') - self.abis = wheel_info.group('abi').split('.') - self.plats = wheel_info.group('plat').split('.') + self.version = wheel_info.group("ver").replace("_", "-") + self.build_tag = wheel_info.group("build") + self.pyversions = wheel_info.group("pyver").split(".") + self.abis = wheel_info.group("abi").split(".") + self.plats = wheel_info.group("plat").split(".") # All the tag combinations from this file self.file_tags = { - Tag(x, y, z) for x in self.pyversions - for y in self.abis for z in self.plats + Tag(x, y, z) for x in self.pyversions for y in self.abis for z in self.plats } - def get_formatted_file_tags(self): - # type: () -> List[str] + def get_formatted_file_tags(self) -> List[str]: """Return the wheel's tags as a sorted list of strings.""" return sorted(str(tag) for tag in self.file_tags) - def support_index_min(self, tags): - # type: (List[Tag]) -> int + def support_index_min(self, tags: List[Tag]) -> int: """Return the lowest index that one of the wheel's file_tag combinations achieves in the given list of supported tags. @@ -67,10 +59,33 @@ def support_index_min(self, tags): :raises ValueError: If none of the wheel's file tags match one of the supported tags. """ - return min(tags.index(tag) for tag in self.file_tags if tag in tags) + try: + return next(i for i, t in enumerate(tags) if t in self.file_tags) + except StopIteration: + raise ValueError() + + def find_most_preferred_tag( + self, tags: List[Tag], tag_to_priority: Dict[Tag, int] + ) -> int: + """Return the priority of the most preferred tag that one of the wheel's file + tag combinations achieves in the given list of supported tags using the given + tag_to_priority mapping, where lower priorities are more-preferred. + + This is used in place of support_index_min in some cases in order to avoid + an expensive linear scan of a large list of tags. + + :param tags: the PEP 425 tags to check the wheel against. + :param tag_to_priority: a mapping from tag to priority of that tag, where + lower is more preferred. + + :raises ValueError: If none of the wheel's file tags match one of + the supported tags. + """ + return min( + tag_to_priority[tag] for tag in self.file_tags if tag in tag_to_priority + ) - def supported(self, tags): - # type: (List[Tag]) -> bool + def supported(self, tags: Iterable[Tag]) -> bool: """Return whether the wheel is compatible with one of the given tags. :param tags: the PEP 425 tags to check the wheel against. diff --git a/src/fetchcode/vcs/pip/_internal/network/auth.py b/src/fetchcode/vcs/pip/_internal/network/auth.py index 4b0a88e2..1a2606ed 100644 --- a/src/fetchcode/vcs/pip/_internal/network/auth.py +++ b/src/fetchcode/vcs/pip/_internal/network/auth.py @@ -4,89 +4,287 @@ providing credentials in the context of network requests. """ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - import logging - -from fetchcode.vcs.pip._vendor.requests.auth import AuthBase, HTTPBasicAuth -from fetchcode.vcs.pip._vendor.requests.utils import get_netrc_auth -from fetchcode.vcs.pip._vendor.six.moves.urllib import parse as urllib_parse - -from fetchcode.vcs.pip._internal.utils.misc import ( +import os +import shutil +import subprocess +import sysconfig +import typing +import urllib.parse +from abc import ABC, abstractmethod +from functools import lru_cache +from os.path import commonprefix +from pathlib import Path +from typing import Any, Dict, List, NamedTuple, Optional, Tuple + +from pip._vendor.requests.auth import AuthBase, HTTPBasicAuth +from pip._vendor.requests.models import Request, Response +from pip._vendor.requests.utils import get_netrc_auth + +from pip._internal.utils.logging import getLogger +from pip._internal.utils.misc import ( ask, ask_input, ask_password, remove_auth_from_url, split_auth_netloc_from_url, ) -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._internal.vcs.versioncontrol import AuthInfo + +logger = getLogger(__name__) + +KEYRING_DISABLED = False + + +class Credentials(NamedTuple): + url: str + username: str + password: str + -if MYPY_CHECK_RUNNING: - from optparse import Values - from typing import Dict, Optional, Tuple +class KeyRingBaseProvider(ABC): + """Keyring base provider interface""" - from fetchcode.vcs.pip._internal.vcs.versioncontrol import AuthInfo + has_keyring: bool - Credentials = Tuple[str, str, str] + @abstractmethod + def get_auth_info( + self, url: str, username: Optional[str] + ) -> Optional[AuthInfo]: ... -logger = logging.getLogger(__name__) + @abstractmethod + def save_auth_info(self, url: str, username: str, password: str) -> None: ... -try: - import keyring # noqa -except ImportError: - keyring = None -except Exception as exc: - logger.warning( - "Keyring is skipped due to an exception: %s", str(exc), - ) - keyring = None +class KeyRingNullProvider(KeyRingBaseProvider): + """Keyring null provider""" -def get_keyring_auth(url, username): - """Return the tuple auth for a given url from keyring.""" - if not url or not keyring: + has_keyring = False + + def get_auth_info(self, url: str, username: Optional[str]) -> Optional[AuthInfo]: return None - try: - try: - get_credential = keyring.get_credential - except AttributeError: - pass - else: + def save_auth_info(self, url: str, username: str, password: str) -> None: + return None + + +class KeyRingPythonProvider(KeyRingBaseProvider): + """Keyring interface which uses locally imported `keyring`""" + + has_keyring = True + + def __init__(self) -> None: + import keyring + + self.keyring = keyring + + def get_auth_info(self, url: str, username: Optional[str]) -> Optional[AuthInfo]: + # Support keyring's get_credential interface which supports getting + # credentials without a username. This is only available for + # keyring>=15.2.0. + if hasattr(self.keyring, "get_credential"): logger.debug("Getting credentials from keyring for %s", url) - cred = get_credential(url, username) + cred = self.keyring.get_credential(url, username) if cred is not None: return cred.username, cred.password return None - if username: + if username is not None: logger.debug("Getting password from keyring for %s", url) - password = keyring.get_password(url, username) + password = self.keyring.get_password(url, username) if password: return username, password + return None + + def save_auth_info(self, url: str, username: str, password: str) -> None: + self.keyring.set_password(url, username, password) + + +class KeyRingCliProvider(KeyRingBaseProvider): + """Provider which uses `keyring` cli + + Instead of calling the keyring package installed alongside pip + we call keyring on the command line which will enable pip to + use which ever installation of keyring is available first in + PATH. + """ + + has_keyring = True + + def __init__(self, cmd: str) -> None: + self.keyring = cmd + + def get_auth_info(self, url: str, username: Optional[str]) -> Optional[AuthInfo]: + # This is the default implementation of keyring.get_credential + # https://github.com/jaraco/keyring/blob/97689324abcf01bd1793d49063e7ca01e03d7d07/keyring/backend.py#L134-L139 + if username is not None: + password = self._get_password(url, username) + if password is not None: + return username, password + return None + + def save_auth_info(self, url: str, username: str, password: str) -> None: + return self._set_password(url, username, password) + + def _get_password(self, service_name: str, username: str) -> Optional[str]: + """Mirror the implementation of keyring.get_password using cli""" + if self.keyring is None: + return None - except Exception as exc: - logger.warning( - "Keyring is skipped due to an exception: %s", str(exc), + cmd = [self.keyring, "get", service_name, username] + env = os.environ.copy() + env["PYTHONIOENCODING"] = "utf-8" + res = subprocess.run( + cmd, + stdin=subprocess.DEVNULL, + stdout=subprocess.PIPE, + env=env, ) + if res.returncode: + return None + return res.stdout.decode("utf-8").strip(os.linesep) + + def _set_password(self, service_name: str, username: str, password: str) -> None: + """Mirror the implementation of keyring.set_password using cli""" + if self.keyring is None: + return None + env = os.environ.copy() + env["PYTHONIOENCODING"] = "utf-8" + subprocess.run( + [self.keyring, "set", service_name, username], + input=f"{password}{os.linesep}".encode(), + env=env, + check=True, + ) + return None -class MultiDomainBasicAuth(AuthBase): +@lru_cache(maxsize=None) +def get_keyring_provider(provider: str) -> KeyRingBaseProvider: + logger.verbose("Keyring provider requested: %s", provider) - def __init__(self, prompting=True, index_urls=None): - # type: (bool, Optional[Values]) -> None + # keyring has previously failed and been disabled + if KEYRING_DISABLED: + provider = "disabled" + if provider in ["import", "auto"]: + try: + impl = KeyRingPythonProvider() + logger.verbose("Keyring provider set: import") + return impl + except ImportError: + pass + except Exception as exc: + # In the event of an unexpected exception + # we should warn the user + msg = "Installed copy of keyring fails with exception %s" + if provider == "auto": + msg = msg + ", trying to find a keyring executable as a fallback" + logger.warning(msg, exc, exc_info=logger.isEnabledFor(logging.DEBUG)) + if provider in ["subprocess", "auto"]: + cli = shutil.which("keyring") + if cli and cli.startswith(sysconfig.get_path("scripts")): + # all code within this function is stolen from shutil.which implementation + @typing.no_type_check + def PATH_as_shutil_which_determines_it() -> str: + path = os.environ.get("PATH", None) + if path is None: + try: + path = os.confstr("CS_PATH") + except (AttributeError, ValueError): + # os.confstr() or CS_PATH is not available + path = os.defpath + # bpo-35755: Don't use os.defpath if the PATH environment variable is + # set to an empty string + + return path + + scripts = Path(sysconfig.get_path("scripts")) + + paths = [] + for path in PATH_as_shutil_which_determines_it().split(os.pathsep): + p = Path(path) + try: + if not p.samefile(scripts): + paths.append(path) + except FileNotFoundError: + pass + + path = os.pathsep.join(paths) + + cli = shutil.which("keyring", path=path) + + if cli: + logger.verbose("Keyring provider set: subprocess with executable %s", cli) + return KeyRingCliProvider(cli) + + logger.verbose("Keyring provider set: disabled") + return KeyRingNullProvider() + + +class MultiDomainBasicAuth(AuthBase): + def __init__( + self, + prompting: bool = True, + index_urls: Optional[List[str]] = None, + keyring_provider: str = "auto", + ) -> None: self.prompting = prompting self.index_urls = index_urls - self.passwords = {} # type: Dict[str, AuthInfo] + self.keyring_provider = keyring_provider # type: ignore[assignment] + self.passwords: Dict[str, AuthInfo] = {} # When the user is prompted to enter credentials and keyring is # available, we will offer to save them. If the user accepts, # this value is set to the credentials they entered. After the # request authenticates, the caller should call # ``save_credentials`` to save these. - self._credentials_to_save = None # type: Optional[Credentials] + self._credentials_to_save: Optional[Credentials] = None + + @property + def keyring_provider(self) -> KeyRingBaseProvider: + return get_keyring_provider(self._keyring_provider) + + @keyring_provider.setter + def keyring_provider(self, provider: str) -> None: + # The free function get_keyring_provider has been decorated with + # functools.cache. If an exception occurs in get_keyring_auth that + # cache will be cleared and keyring disabled, take that into account + # if you want to remove this indirection. + self._keyring_provider = provider + + @property + def use_keyring(self) -> bool: + # We won't use keyring when --no-input is passed unless + # a specific provider is requested because it might require + # user interaction + return self.prompting or self._keyring_provider not in ["auto", "disabled"] + + def _get_keyring_auth( + self, + url: Optional[str], + username: Optional[str], + ) -> Optional[AuthInfo]: + """Return the tuple auth for a given url from keyring.""" + # Do nothing if no url was provided + if not url: + return None + + try: + return self.keyring_provider.get_auth_info(url, username) + except Exception as exc: + # Log the full exception (with stacktrace) at debug, so it'll only + # show up when running in verbose mode. + logger.debug("Keyring is skipped due to an exception", exc_info=True) + # Always log a shortened version of the exception. + logger.warning( + "Keyring is skipped due to an exception: %s", + str(exc), + ) + global KEYRING_DISABLED + KEYRING_DISABLED = True + get_keyring_provider.cache_clear() + return None - def _get_index_url(self, url): + def _get_index_url(self, url: str) -> Optional[str]: """Return the original index URL matching the requested URL. Cached or dynamically generated credentials may work against @@ -102,13 +300,45 @@ def _get_index_url(self, url): if not url or not self.index_urls: return None - for u in self.index_urls: - prefix = remove_auth_from_url(u).rstrip("/") + "/" - if url.startswith(prefix): - return u + url = remove_auth_from_url(url).rstrip("/") + "/" + parsed_url = urllib.parse.urlsplit(url) + + candidates = [] + + for index in self.index_urls: + index = index.rstrip("/") + "/" + parsed_index = urllib.parse.urlsplit(remove_auth_from_url(index)) + if parsed_url == parsed_index: + return index + + if parsed_url.netloc != parsed_index.netloc: + continue + + candidate = urllib.parse.urlsplit(index) + candidates.append(candidate) + + if not candidates: + return None - def _get_new_credentials(self, original_url, allow_netrc=True, - allow_keyring=True): + candidates.sort( + reverse=True, + key=lambda candidate: commonprefix( + [ + parsed_url.path, + candidate.path, + ] + ).rfind("/"), + ) + + return urllib.parse.urlunsplit(candidates[0]) + + def _get_new_credentials( + self, + original_url: str, + *, + allow_netrc: bool = True, + allow_keyring: bool = False, + ) -> AuthInfo: """Find and return credentials for the specified URL.""" # Split the credentials and netloc from the url. url, netloc, url_user_password = split_auth_netloc_from_url( @@ -147,17 +377,21 @@ def _get_new_credentials(self, original_url, allow_netrc=True, # If we don't have a password and keyring is available, use it. if allow_keyring: # The index url is more specific than the netloc, so try it first + # fmt: off kr_auth = ( - get_keyring_auth(index_url, username) or - get_keyring_auth(netloc, username) + self._get_keyring_auth(index_url, username) or + self._get_keyring_auth(netloc, username) ) + # fmt: on if kr_auth: logger.debug("Found credentials in keyring for %s", netloc) return kr_auth return username, password - def _get_url_and_credentials(self, original_url): + def _get_url_and_credentials( + self, original_url: str + ) -> Tuple[str, Optional[str], Optional[str]]: """Return the credentials to use for the provided URL. If allowed, netrc and keyring may be used to obtain the @@ -169,13 +403,19 @@ def _get_url_and_credentials(self, original_url): """ url, netloc, _ = split_auth_netloc_from_url(original_url) - # Use any stored credentials that we have for this netloc - username, password = self.passwords.get(netloc, (None, None)) + # Try to get credentials from original url + username, password = self._get_new_credentials(original_url) - if username is None and password is None: - # No stored credentials. Acquire new credentials without prompting - # the user. (e.g. from netrc, keyring, or the URL itself) - username, password = self._get_new_credentials(original_url) + # If credentials not found, use any stored credentials for this netloc. + # Do this if either the username or the password is missing. + # This accounts for the situation in which the user has specified + # the username in the index url, but the password comes from keyring. + if (username is None or password is None) and netloc in self.passwords: + un, pw = self.passwords[netloc] + # It is possible that the cached credentials are for a different username, + # in which case the cache should be ignored. + if username is None or username == un: + username, password = un, pw if username is not None or password is not None: # Convert the username and password if they're None, so that @@ -190,14 +430,14 @@ def _get_url_and_credentials(self, original_url): assert ( # Credentials were found - (username is not None and password is not None) or + (username is not None and password is not None) # Credentials were not found - (username is None and password is None) - ), "Could not load credentials from url: {}".format(original_url) + or (username is None and password is None) + ), f"Could not load credentials from url: {original_url}" return url, username, password - def __call__(self, req): + def __call__(self, req: Request) -> Request: # Get credentials for this request url, username, password = self._get_url_and_credentials(req.url) @@ -214,36 +454,55 @@ def __call__(self, req): return req # Factored out to allow for easy patching in tests - def _prompt_for_password(self, netloc): - username = ask_input("User for {}: ".format(netloc)) + def _prompt_for_password( + self, netloc: str + ) -> Tuple[Optional[str], Optional[str], bool]: + username = ask_input(f"User for {netloc}: ") if self.prompting else None if not username: - return None, None - auth = get_keyring_auth(netloc, username) - if auth: - return auth[0], auth[1], False + return None, None, False + if self.use_keyring: + auth = self._get_keyring_auth(netloc, username) + if auth and auth[0] is not None and auth[1] is not None: + return auth[0], auth[1], False password = ask_password("Password: ") return username, password, True # Factored out to allow for easy patching in tests - def _should_save_password_to_keyring(self): - if not keyring: + def _should_save_password_to_keyring(self) -> bool: + if ( + not self.prompting + or not self.use_keyring + or not self.keyring_provider.has_keyring + ): return False return ask("Save credentials to keyring [y/N]: ", ["y", "n"]) == "y" - def handle_401(self, resp, **kwargs): + def handle_401(self, resp: Response, **kwargs: Any) -> Response: # We only care about 401 responses, anything else we want to just # pass through the actual response if resp.status_code != 401: return resp + username, password = None, None + + # Query the keyring for credentials: + if self.use_keyring: + username, password = self._get_new_credentials( + resp.url, + allow_netrc=False, + allow_keyring=True, + ) + # We are not able to prompt the user so simply return the response - if not self.prompting: + if not self.prompting and not username and not password: return resp - parsed = urllib_parse.urlparse(resp.url) + parsed = urllib.parse.urlparse(resp.url) # Prompt the user for a new username and password - username, password, save = self._prompt_for_password(parsed.netloc) + save = False + if not username and not password: + username, password, save = self._prompt_for_password(parsed.netloc) # Store the new username and password to use for future requests self._credentials_to_save = None @@ -252,11 +511,17 @@ def handle_401(self, resp, **kwargs): # Prompt to save the password to keyring if save and self._should_save_password_to_keyring(): - self._credentials_to_save = (parsed.netloc, username, password) + self._credentials_to_save = Credentials( + url=parsed.netloc, + username=username, + password=password, + ) # Consume content and release the original connection to allow our new # request to reuse the same one. - resp.content + # The result of the assignment isn't used, it's just needed to consume + # the content. + _ = resp.content resp.raw.release_conn() # Add our new username and password to the request @@ -275,24 +540,27 @@ def handle_401(self, resp, **kwargs): return new_resp - def warn_on_401(self, resp, **kwargs): + def warn_on_401(self, resp: Response, **kwargs: Any) -> None: """Response callback to warn about incorrect credentials.""" if resp.status_code == 401: logger.warning( - '401 Error, Credentials not correct for %s', resp.request.url, + "401 Error, Credentials not correct for %s", + resp.request.url, ) - def save_credentials(self, resp, **kwargs): + def save_credentials(self, resp: Response, **kwargs: Any) -> None: """Response callback to save credentials on success.""" - assert keyring is not None, "should never reach here without keyring" - if not keyring: - return + assert ( + self.keyring_provider.has_keyring + ), "should never reach here without keyring" creds = self._credentials_to_save self._credentials_to_save = None if creds and resp.status_code < 400: try: - logger.info('Saving credentials to keyring') - keyring.set_password(*creds) + logger.info("Saving credentials to keyring") + self.keyring_provider.save_auth_info( + creds.url, creds.username, creds.password + ) except Exception: - logger.exception('Failed to save credentials') + logger.exception("Failed to save credentials") diff --git a/src/fetchcode/vcs/pip/_internal/network/cache.py b/src/fetchcode/vcs/pip/_internal/network/cache.py index a5b1f4ef..4d0fb545 100644 --- a/src/fetchcode/vcs/pip/_internal/network/cache.py +++ b/src/fetchcode/vcs/pip/_internal/network/cache.py @@ -1,81 +1,106 @@ """HTTP cache implementation. """ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - import os from contextlib import contextmanager +from datetime import datetime +from typing import BinaryIO, Generator, Optional, Union -from fetchcode.vcs.pip._vendor.cachecontrol.cache import BaseCache -from fetchcode.vcs.pip._vendor.cachecontrol.caches import FileCache -from fetchcode.vcs.pip._vendor.requests.models import Response - -from fetchcode.vcs.pip._internal.utils.filesystem import adjacent_tmp_file, replace -from fetchcode.vcs.pip._internal.utils.misc import ensure_dir -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._vendor.cachecontrol.cache import SeparateBodyBaseCache +from pip._vendor.cachecontrol.caches import SeparateBodyFileCache +from pip._vendor.requests.models import Response -if MYPY_CHECK_RUNNING: - from typing import Optional +from pip._internal.utils.filesystem import adjacent_tmp_file, replace +from pip._internal.utils.misc import ensure_dir -def is_from_cache(response): - # type: (Response) -> bool +def is_from_cache(response: Response) -> bool: return getattr(response, "from_cache", False) @contextmanager -def suppressed_cache_errors(): +def suppressed_cache_errors() -> Generator[None, None, None]: """If we can't access the cache then we can just skip caching and process requests as if caching wasn't enabled. """ try: yield - except (OSError, IOError): + except OSError: pass -class SafeFileCache(BaseCache): +class SafeFileCache(SeparateBodyBaseCache): """ A file based cache which is safe to use even when the target directory may not be accessible or writable. + + There is a race condition when two processes try to write and/or read the + same entry at the same time, since each entry consists of two separate + files (https://github.com/psf/cachecontrol/issues/324). We therefore have + additional logic that makes sure that both files to be present before + returning an entry; this fixes the read side of the race condition. + + For the write side, we assume that the server will only ever return the + same data for the same URL, which ought to be the case for files pip is + downloading. PyPI does not have a mechanism to swap out a wheel for + another wheel, for example. If this assumption is not true, the + CacheControl issue will need to be fixed. """ - def __init__(self, directory): - # type: (str) -> None + def __init__(self, directory: str) -> None: assert directory is not None, "Cache directory must not be None." - super(SafeFileCache, self).__init__() + super().__init__() self.directory = directory - def _get_cache_path(self, name): - # type: (str) -> str + def _get_cache_path(self, name: str) -> str: # From cachecontrol.caches.file_cache.FileCache._fn, brought into our # class for backwards-compatibility and to avoid using a non-public # method. - hashed = FileCache.encode(name) + hashed = SeparateBodyFileCache.encode(name) parts = list(hashed[:5]) + [hashed] return os.path.join(self.directory, *parts) - def get(self, key): - # type: (str) -> Optional[bytes] - path = self._get_cache_path(key) + def get(self, key: str) -> Optional[bytes]: + # The cache entry is only valid if both metadata and body exist. + metadata_path = self._get_cache_path(key) + body_path = metadata_path + ".body" + if not (os.path.exists(metadata_path) and os.path.exists(body_path)): + return None with suppressed_cache_errors(): - with open(path, 'rb') as f: + with open(metadata_path, "rb") as f: return f.read() - def set(self, key, value): - # type: (str, bytes) -> None - path = self._get_cache_path(key) + def _write(self, path: str, data: bytes) -> None: with suppressed_cache_errors(): ensure_dir(os.path.dirname(path)) with adjacent_tmp_file(path) as f: - f.write(value) + f.write(data) replace(f.name, path) - def delete(self, key): - # type: (str) -> None + def set( + self, key: str, value: bytes, expires: Union[int, datetime, None] = None + ) -> None: + path = self._get_cache_path(key) + self._write(path, value) + + def delete(self, key: str) -> None: path = self._get_cache_path(key) with suppressed_cache_errors(): os.remove(path) + with suppressed_cache_errors(): + os.remove(path + ".body") + + def get_body(self, key: str) -> Optional[BinaryIO]: + # The cache entry is only valid if both metadata and body exist. + metadata_path = self._get_cache_path(key) + body_path = metadata_path + ".body" + if not (os.path.exists(metadata_path) and os.path.exists(body_path)): + return None + with suppressed_cache_errors(): + return open(body_path, "rb") + + def set_body(self, key: str, body: bytes) -> None: + path = self._get_cache_path(key) + ".body" + self._write(path, body) diff --git a/src/fetchcode/vcs/pip/_internal/network/download.py b/src/fetchcode/vcs/pip/_internal/network/download.py index 937078f5..5c3bce3d 100644 --- a/src/fetchcode/vcs/pip/_internal/network/download.py +++ b/src/fetchcode/vcs/pip/_internal/network/download.py @@ -1,49 +1,38 @@ """Download files with progress indicators. """ -import cgi + +import email.message import logging import mimetypes import os +from typing import Iterable, Optional, Tuple -from fetchcode.vcs.pip._vendor import requests -from fetchcode.vcs.pip._vendor.requests.models import CONTENT_CHUNK_SIZE - -from fetchcode.vcs.pip._internal.cli.progress_bars import DownloadProgressProvider -from fetchcode.vcs.pip._internal.models.index import PyPI -from fetchcode.vcs.pip._internal.network.cache import is_from_cache -from fetchcode.vcs.pip._internal.network.utils import response_chunks -from fetchcode.vcs.pip._internal.utils.misc import ( - format_size, - redact_auth_from_url, - splitext, -) -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import Iterable, Optional - - from fetchcode.vcs.pip._vendor.requests.models import Response +from pip._vendor.requests.models import Response - from fetchcode.vcs.pip._internal.models.link import Link - from fetchcode.vcs.pip._internal.network.session import PipSession +from pip._internal.cli.progress_bars import get_download_progress_renderer +from pip._internal.exceptions import NetworkConnectionError +from pip._internal.models.index import PyPI +from pip._internal.models.link import Link +from pip._internal.network.cache import is_from_cache +from pip._internal.network.session import PipSession +from pip._internal.network.utils import HEADERS, raise_for_status, response_chunks +from pip._internal.utils.misc import format_size, redact_auth_from_url, splitext logger = logging.getLogger(__name__) -def _get_http_response_size(resp): - # type: (Response) -> Optional[int] +def _get_http_response_size(resp: Response) -> Optional[int]: try: - return int(resp.headers['content-length']) + return int(resp.headers["content-length"]) except (ValueError, KeyError, TypeError): return None def _prepare_download( - resp, # type: Response - link, # type: Link - progress_bar # type: str -): - # type: (...) -> Iterable[bytes] + resp: Response, + link: Link, + progress_bar: str, +) -> Iterable[bytes]: total_length = _get_http_response_size(resp) if link.netloc == PyPI.file_storage_domain: @@ -54,7 +43,7 @@ def _prepare_download( logged_url = redact_auth_from_url(url) if total_length: - logged_url = '{} ({})'.format(logged_url, format_size(total_length)) + logged_url = f"{logged_url} ({format_size(total_length)})" if is_from_cache(resp): logger.info("Using cached %s", logged_url) @@ -67,59 +56,54 @@ def _prepare_download( show_progress = False elif not total_length: show_progress = True - elif total_length > (40 * 1000): + elif total_length > (512 * 1024): show_progress = True else: show_progress = False - chunks = response_chunks(resp, CONTENT_CHUNK_SIZE) + chunks = response_chunks(resp) if not show_progress: return chunks - return DownloadProgressProvider( - progress_bar, max=total_length - )(chunks) + renderer = get_download_progress_renderer(bar_type=progress_bar, size=total_length) + return renderer(chunks) -def sanitize_content_filename(filename): - # type: (str) -> str +def sanitize_content_filename(filename: str) -> str: """ Sanitize the "filename" value from a Content-Disposition header. """ return os.path.basename(filename) -def parse_content_disposition(content_disposition, default_filename): - # type: (str, str) -> str +def parse_content_disposition(content_disposition: str, default_filename: str) -> str: """ Parse the "filename" value from a Content-Disposition header, and return the default filename if the result is empty. """ - _type, params = cgi.parse_header(content_disposition) - filename = params.get('filename') + m = email.message.Message() + m["content-type"] = content_disposition + filename = m.get_param("filename") if filename: # We need to sanitize the filename to prevent directory traversal # in case the filename contains ".." path parts. - filename = sanitize_content_filename(filename) + filename = sanitize_content_filename(str(filename)) return filename or default_filename -def _get_http_response_filename(resp, link): - # type: (Response, Link) -> str +def _get_http_response_filename(resp: Response, link: Link) -> str: """Get an ideal filename from the given HTTP response, falling back to the link filename if not provided. """ filename = link.filename # fallback # Have a look at the Content-Disposition header for a better guess - content_disposition = resp.headers.get('content-disposition') + content_disposition = resp.headers.get("content-disposition") if content_disposition: filename = parse_content_disposition(content_disposition, filename) - ext = splitext(filename)[1] # type: Optional[str] + ext: Optional[str] = splitext(filename)[1] if not ext: - ext = mimetypes.guess_extension( - resp.headers.get('content-type', '') - ) + ext = mimetypes.guess_extension(resp.headers.get("content-type", "")) if ext: filename += ext if not ext and link.url != resp.url: @@ -129,72 +113,75 @@ def _get_http_response_filename(resp, link): return filename -def _http_get_download(session, link): - # type: (PipSession, Link) -> Response - target_url = link.url.split('#', 1)[0] - resp = session.get( - target_url, - # We use Accept-Encoding: identity here because requests - # defaults to accepting compressed responses. This breaks in - # a variety of ways depending on how the server is configured. - # - Some servers will notice that the file isn't a compressible - # file and will leave the file alone and with an empty - # Content-Encoding - # - Some servers will notice that the file is already - # compressed and will leave the file alone and will add a - # Content-Encoding: gzip header - # - Some servers won't notice anything at all and will take - # a file that's already been compressed and compress it again - # and set the Content-Encoding: gzip header - # By setting this to request only the identity encoding We're - # hoping to eliminate the third case. Hopefully there does not - # exist a server which when given a file will notice it is - # already compressed and that you're not asking for a - # compressed file and will then decompress it before sending - # because if that's the case I don't think it'll ever be - # possible to make this work. - headers={"Accept-Encoding": "identity"}, - stream=True, - ) - resp.raise_for_status() +def _http_get_download(session: PipSession, link: Link) -> Response: + target_url = link.url.split("#", 1)[0] + resp = session.get(target_url, headers=HEADERS, stream=True) + raise_for_status(resp) return resp -class Download(object): - def __init__( - self, - response, # type: Response - filename, # type: str - chunks, # type: Iterable[bytes] - ): - # type: (...) -> None - self.response = response - self.filename = filename - self.chunks = chunks - - -class Downloader(object): +class Downloader: def __init__( self, - session, # type: PipSession - progress_bar, # type: str - ): - # type: (...) -> None + session: PipSession, + progress_bar: str, + ) -> None: self._session = session self._progress_bar = progress_bar - def __call__(self, link): - # type: (Link) -> Download + def __call__(self, link: Link, location: str) -> Tuple[str, str]: + """Download the file given by link into location.""" try: resp = _http_get_download(self._session, link) - except requests.HTTPError as e: + except NetworkConnectionError as e: + assert e.response is not None logger.critical( "HTTP error %s while getting %s", e.response.status_code, link ) raise - return Download( - resp, - _get_http_response_filename(resp, link), - _prepare_download(resp, link, self._progress_bar), - ) + filename = _get_http_response_filename(resp, link) + filepath = os.path.join(location, filename) + + chunks = _prepare_download(resp, link, self._progress_bar) + with open(filepath, "wb") as content_file: + for chunk in chunks: + content_file.write(chunk) + content_type = resp.headers.get("Content-Type", "") + return filepath, content_type + + +class BatchDownloader: + def __init__( + self, + session: PipSession, + progress_bar: str, + ) -> None: + self._session = session + self._progress_bar = progress_bar + + def __call__( + self, links: Iterable[Link], location: str + ) -> Iterable[Tuple[Link, Tuple[str, str]]]: + """Download the files given by links into location.""" + for link in links: + try: + resp = _http_get_download(self._session, link) + except NetworkConnectionError as e: + assert e.response is not None + logger.critical( + "HTTP error %s while getting %s", + e.response.status_code, + link, + ) + raise + + filename = _get_http_response_filename(resp, link) + filepath = os.path.join(location, filename) + + chunks = _prepare_download(resp, link, self._progress_bar) + with open(filepath, "wb") as content_file: + for chunk in chunks: + content_file.write(chunk) + content_type = resp.headers.get("Content-Type", "") + yield link, (filepath, content_type) diff --git a/src/fetchcode/vcs/pip/_internal/network/lazy_wheel.py b/src/fetchcode/vcs/pip/_internal/network/lazy_wheel.py new file mode 100644 index 00000000..82ec50d5 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/network/lazy_wheel.py @@ -0,0 +1,210 @@ +"""Lazy ZIP over HTTP""" + +__all__ = ["HTTPRangeRequestUnsupported", "dist_from_wheel_url"] + +from bisect import bisect_left, bisect_right +from contextlib import contextmanager +from tempfile import NamedTemporaryFile +from typing import Any, Dict, Generator, List, Optional, Tuple +from zipfile import BadZipFile, ZipFile + +from pip._vendor.packaging.utils import canonicalize_name +from pip._vendor.requests.models import CONTENT_CHUNK_SIZE, Response + +from pip._internal.metadata import BaseDistribution, MemoryWheel, get_wheel_distribution +from pip._internal.network.session import PipSession +from pip._internal.network.utils import HEADERS, raise_for_status, response_chunks + + +class HTTPRangeRequestUnsupported(Exception): + pass + + +def dist_from_wheel_url(name: str, url: str, session: PipSession) -> BaseDistribution: + """Return a distribution object from the given wheel URL. + + This uses HTTP range requests to only fetch the portion of the wheel + containing metadata, just enough for the object to be constructed. + If such requests are not supported, HTTPRangeRequestUnsupported + is raised. + """ + with LazyZipOverHTTP(url, session) as zf: + # For read-only ZIP files, ZipFile only needs methods read, + # seek, seekable and tell, not the whole IO protocol. + wheel = MemoryWheel(zf.name, zf) # type: ignore + # After context manager exit, wheel.name + # is an invalid file by intention. + return get_wheel_distribution(wheel, canonicalize_name(name)) + + +class LazyZipOverHTTP: + """File-like object mapped to a ZIP file over HTTP. + + This uses HTTP range requests to lazily fetch the file's content, + which is supposed to be fed to ZipFile. If such requests are not + supported by the server, raise HTTPRangeRequestUnsupported + during initialization. + """ + + def __init__( + self, url: str, session: PipSession, chunk_size: int = CONTENT_CHUNK_SIZE + ) -> None: + head = session.head(url, headers=HEADERS) + raise_for_status(head) + assert head.status_code == 200 + self._session, self._url, self._chunk_size = session, url, chunk_size + self._length = int(head.headers["Content-Length"]) + self._file = NamedTemporaryFile() + self.truncate(self._length) + self._left: List[int] = [] + self._right: List[int] = [] + if "bytes" not in head.headers.get("Accept-Ranges", "none"): + raise HTTPRangeRequestUnsupported("range request is not supported") + self._check_zip() + + @property + def mode(self) -> str: + """Opening mode, which is always rb.""" + return "rb" + + @property + def name(self) -> str: + """Path to the underlying file.""" + return self._file.name + + def seekable(self) -> bool: + """Return whether random access is supported, which is True.""" + return True + + def close(self) -> None: + """Close the file.""" + self._file.close() + + @property + def closed(self) -> bool: + """Whether the file is closed.""" + return self._file.closed + + def read(self, size: int = -1) -> bytes: + """Read up to size bytes from the object and return them. + + As a convenience, if size is unspecified or -1, + all bytes until EOF are returned. Fewer than + size bytes may be returned if EOF is reached. + """ + download_size = max(size, self._chunk_size) + start, length = self.tell(), self._length + stop = length if size < 0 else min(start + download_size, length) + start = max(0, stop - download_size) + self._download(start, stop - 1) + return self._file.read(size) + + def readable(self) -> bool: + """Return whether the file is readable, which is True.""" + return True + + def seek(self, offset: int, whence: int = 0) -> int: + """Change stream position and return the new absolute position. + + Seek to offset relative position indicated by whence: + * 0: Start of stream (the default). pos should be >= 0; + * 1: Current position - pos may be negative; + * 2: End of stream - pos usually negative. + """ + return self._file.seek(offset, whence) + + def tell(self) -> int: + """Return the current position.""" + return self._file.tell() + + def truncate(self, size: Optional[int] = None) -> int: + """Resize the stream to the given size in bytes. + + If size is unspecified resize to the current position. + The current stream position isn't changed. + + Return the new file size. + """ + return self._file.truncate(size) + + def writable(self) -> bool: + """Return False.""" + return False + + def __enter__(self) -> "LazyZipOverHTTP": + self._file.__enter__() + return self + + def __exit__(self, *exc: Any) -> None: + self._file.__exit__(*exc) + + @contextmanager + def _stay(self) -> Generator[None, None, None]: + """Return a context manager keeping the position. + + At the end of the block, seek back to original position. + """ + pos = self.tell() + try: + yield + finally: + self.seek(pos) + + def _check_zip(self) -> None: + """Check and download until the file is a valid ZIP.""" + end = self._length - 1 + for start in reversed(range(0, end, self._chunk_size)): + self._download(start, end) + with self._stay(): + try: + # For read-only ZIP files, ZipFile only needs + # methods read, seek, seekable and tell. + ZipFile(self) # type: ignore + except BadZipFile: + pass + else: + break + + def _stream_response( + self, start: int, end: int, base_headers: Dict[str, str] = HEADERS + ) -> Response: + """Return HTTP response to a range request from start to end.""" + headers = base_headers.copy() + headers["Range"] = f"bytes={start}-{end}" + # TODO: Get range requests to be correctly cached + headers["Cache-Control"] = "no-cache" + return self._session.get(self._url, headers=headers, stream=True) + + def _merge( + self, start: int, end: int, left: int, right: int + ) -> Generator[Tuple[int, int], None, None]: + """Return a generator of intervals to be fetched. + + Args: + start (int): Start of needed interval + end (int): End of needed interval + left (int): Index of first overlapping downloaded data + right (int): Index after last overlapping downloaded data + """ + lslice, rslice = self._left[left:right], self._right[left:right] + i = start = min([start] + lslice[:1]) + end = max([end] + rslice[-1:]) + for j, k in zip(lslice, rslice): + if j > i: + yield i, j - 1 + i = k + 1 + if i <= end: + yield i, end + self._left[left:right], self._right[left:right] = [start], [end] + + def _download(self, start: int, end: int) -> None: + """Download bytes from start to end inclusively.""" + with self._stay(): + left = bisect_left(self._right, start) + right = bisect_right(self._left, end) + for start, end in self._merge(start, end, left, right): + response = self._stream_response(start, end) + response.raise_for_status() + self.seek(start) + for chunk in response_chunks(response, self._chunk_size): + self._file.write(chunk) diff --git a/src/fetchcode/vcs/pip/_internal/network/session.py b/src/fetchcode/vcs/pip/_internal/network/session.py index e1702092..1765b4f6 100644 --- a/src/fetchcode/vcs/pip/_internal/network/session.py +++ b/src/fetchcode/vcs/pip/_internal/network/session.py @@ -2,58 +2,70 @@ network request configuration and behavior. """ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - import email.utils +import functools +import io +import ipaddress import json import logging import mimetypes import os import platform +import shutil +import subprocess import sys +import urllib.parse import warnings +from typing import ( + TYPE_CHECKING, + Any, + Dict, + Generator, + List, + Mapping, + Optional, + Sequence, + Tuple, + Union, +) -from fetchcode.vcs.pip._vendor import requests, six, urllib3 -from fetchcode.vcs.pip._vendor.cachecontrol import CacheControlAdapter -from fetchcode.vcs.pip._vendor.requests.adapters import BaseAdapter, HTTPAdapter -from fetchcode.vcs.pip._vendor.requests.models import Response -from fetchcode.vcs.pip._vendor.requests.structures import CaseInsensitiveDict -from fetchcode.vcs.pip._vendor.six.moves.urllib import parse as urllib_parse -from fetchcode.vcs.pip._vendor.urllib3.exceptions import InsecureRequestWarning +from pip._vendor import requests, urllib3 +from pip._vendor.cachecontrol import CacheControlAdapter as _BaseCacheControlAdapter +from pip._vendor.requests.adapters import DEFAULT_POOLBLOCK, BaseAdapter +from pip._vendor.requests.adapters import HTTPAdapter as _BaseHTTPAdapter +from pip._vendor.requests.models import PreparedRequest, Response +from pip._vendor.requests.structures import CaseInsensitiveDict +from pip._vendor.urllib3.connectionpool import ConnectionPool +from pip._vendor.urllib3.exceptions import InsecureRequestWarning from pip import __version__ -from fetchcode.vcs.pip._internal.network.auth import MultiDomainBasicAuth -from fetchcode.vcs.pip._internal.network.cache import SafeFileCache -# Import ssl from compat so the initial import occurs in only one place. -from fetchcode.vcs.pip._internal.utils.compat import has_tls, ipaddress -from fetchcode.vcs.pip._internal.utils.glibc import libc_ver -from fetchcode.vcs.pip._internal.utils.misc import ( - build_url_from_netloc, - get_installed_version, - parse_netloc, -) -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.utils.urls import url_to_path +from pip._internal.metadata import get_default_environment +from pip._internal.models.link import Link +from pip._internal.network.auth import MultiDomainBasicAuth +from pip._internal.network.cache import SafeFileCache -if MYPY_CHECK_RUNNING: - from typing import ( - Iterator, List, Optional, Tuple, Union, - ) +# Import ssl from compat so the initial import occurs in only one place. +from pip._internal.utils.compat import has_tls +from pip._internal.utils.glibc import libc_ver +from pip._internal.utils.misc import build_url_from_netloc, parse_netloc +from pip._internal.utils.urls import url_to_path - from fetchcode.vcs.pip._internal.models.link import Link +if TYPE_CHECKING: + from ssl import SSLContext - SecureOrigin = Tuple[str, str, Optional[Union[int, str]]] + from pip._vendor.urllib3.poolmanager import PoolManager logger = logging.getLogger(__name__) +SecureOrigin = Tuple[str, str, Optional[Union[int, str]]] + # Ignore warning raised when using --trusted-host. warnings.filterwarnings("ignore", category=InsecureRequestWarning) -SECURE_ORIGINS = [ +SECURE_ORIGINS: List[SecureOrigin] = [ # protocol, hostname, port # Taken from Chrome's list of secure origins (See: http://bit.ly/1qrySKC) ("https", "*", "*"), @@ -63,7 +75,7 @@ ("file", "*", None), # ssh is always secure. ("ssh", "*", "*"), -] # type: List[SecureOrigin] +] # These are environment variables present when running under various @@ -75,18 +87,17 @@ # For more background, see: https://github.com/pypa/pip/issues/5499 CI_ENVIRONMENT_VARIABLES = ( # Azure Pipelines - 'BUILD_BUILDID', + "BUILD_BUILDID", # Jenkins - 'BUILD_ID', + "BUILD_ID", # AppVeyor, CircleCI, Codeship, Gitlab CI, Shippable, Travis CI - 'CI', + "CI", # Explicit environment variable. - 'PIP_IS_CI', + "PIP_IS_CI", ) -def looks_like_ci(): - # type: () -> bool +def looks_like_ci() -> bool: """ Return whether it looks like pip is running under CI. """ @@ -96,11 +107,12 @@ def looks_like_ci(): return any(name in os.environ for name in CI_ENVIRONMENT_VARIABLES) -def user_agent(): +@functools.lru_cache(maxsize=1) +def user_agent() -> str: """ Return a string representing the user agent. """ - data = { + data: Dict[str, Any] = { "installer": {"name": "pip", "version": __version__}, "python": platform.python_version(), "implementation": { @@ -108,33 +120,38 @@ def user_agent(): }, } - if data["implementation"]["name"] == 'CPython': + if data["implementation"]["name"] == "CPython": data["implementation"]["version"] = platform.python_version() - elif data["implementation"]["name"] == 'PyPy': - if sys.pypy_version_info.releaselevel == 'final': - pypy_version_info = sys.pypy_version_info[:3] - else: - pypy_version_info = sys.pypy_version_info + elif data["implementation"]["name"] == "PyPy": + pypy_version_info = sys.pypy_version_info # type: ignore + if pypy_version_info.releaselevel == "final": + pypy_version_info = pypy_version_info[:3] data["implementation"]["version"] = ".".join( [str(x) for x in pypy_version_info] ) - elif data["implementation"]["name"] == 'Jython': + elif data["implementation"]["name"] == "Jython": # Complete Guess data["implementation"]["version"] = platform.python_version() - elif data["implementation"]["name"] == 'IronPython': + elif data["implementation"]["name"] == "IronPython": # Complete Guess data["implementation"]["version"] = platform.python_version() if sys.platform.startswith("linux"): - from fetchcode.vcs.pip._vendor import distro - distro_infos = dict(filter( - lambda x: x[1], - zip(["name", "version", "id"], distro.linux_distribution()), - )) - libc = dict(filter( - lambda x: x[1], - zip(["lib", "version"], libc_ver()), - )) + from pip._vendor import distro + + linux_distribution = distro.name(), distro.version(), distro.codename() + distro_infos: Dict[str, Any] = dict( + filter( + lambda x: x[1], + zip(["name", "version", "id"], linux_distribution), + ) + ) + libc = dict( + filter( + lambda x: x[1], + zip(["lib", "version"], libc_ver()), + ) + ) if libc: distro_infos["libc"] = libc if distro_infos: @@ -154,11 +171,27 @@ def user_agent(): if has_tls(): import _ssl as ssl + data["openssl_version"] = ssl.OPENSSL_VERSION - setuptools_version = get_installed_version("setuptools") - if setuptools_version is not None: - data["setuptools_version"] = setuptools_version + setuptools_dist = get_default_environment().get_distribution("setuptools") + if setuptools_dist is not None: + data["setuptools_version"] = str(setuptools_dist.version) + + if shutil.which("rustc") is not None: + # If for any reason `rustc --version` fails, silently ignore it + try: + rustc_output = subprocess.check_output( + ["rustc", "--version"], stderr=subprocess.STDOUT, timeout=0.5 + ) + except Exception: + pass + else: + if rustc_output.startswith(b"rustc "): + # The format of `rustc --version` is: + # `b'rustc 1.52.1 (9bc8c42bb 2021-05-09)\n'` + # We extract just the middle (1.52.1) part + data["rustc_version"] = rustc_output.split(b" ")[1].decode() # Use None rather than False so as not to give the impression that # pip knows it is not being run under CI. Rather, it is a null or @@ -177,9 +210,15 @@ def user_agent(): class LocalFSAdapter(BaseAdapter): - - def send(self, request, stream=None, timeout=None, verify=None, cert=None, - proxies=None): + def send( + self, + request: PreparedRequest, + stream: bool = False, + timeout: Optional[Union[float, Tuple[float, float]]] = None, + verify: Union[bool, str] = True, + cert: Optional[Union[str, Tuple[str, str]]] = None, + proxies: Optional[Mapping[str, str]] = None, + ) -> Response: pathname = url_to_path(request.url) resp = Response() @@ -189,61 +228,117 @@ def send(self, request, stream=None, timeout=None, verify=None, cert=None, try: stats = os.stat(pathname) except OSError as exc: + # format the exception raised as a io.BytesIO object, + # to return a better error message: resp.status_code = 404 - resp.raw = exc + resp.reason = type(exc).__name__ + resp.raw = io.BytesIO(f"{resp.reason}: {exc}".encode()) else: modified = email.utils.formatdate(stats.st_mtime, usegmt=True) content_type = mimetypes.guess_type(pathname)[0] or "text/plain" - resp.headers = CaseInsensitiveDict({ - "Content-Type": content_type, - "Content-Length": stats.st_size, - "Last-Modified": modified, - }) + resp.headers = CaseInsensitiveDict( + { + "Content-Type": content_type, + "Content-Length": stats.st_size, + "Last-Modified": modified, + } + ) resp.raw = open(pathname, "rb") resp.close = resp.raw.close return resp - def close(self): + def close(self) -> None: pass -class InsecureHTTPAdapter(HTTPAdapter): +class _SSLContextAdapterMixin: + """Mixin to add the ``ssl_context`` constructor argument to HTTP adapters. - def cert_verify(self, conn, url, verify, cert): - super(InsecureHTTPAdapter, self).cert_verify( - conn=conn, url=url, verify=False, cert=cert + The additional argument is forwarded directly to the pool manager. This allows us + to dynamically decide what SSL store to use at runtime, which is used to implement + the optional ``truststore`` backend. + """ + + def __init__( + self, + *, + ssl_context: Optional["SSLContext"] = None, + **kwargs: Any, + ) -> None: + self._ssl_context = ssl_context + super().__init__(**kwargs) + + def init_poolmanager( + self, + connections: int, + maxsize: int, + block: bool = DEFAULT_POOLBLOCK, + **pool_kwargs: Any, + ) -> "PoolManager": + if self._ssl_context is not None: + pool_kwargs.setdefault("ssl_context", self._ssl_context) + return super().init_poolmanager( # type: ignore[misc] + connections=connections, + maxsize=maxsize, + block=block, + **pool_kwargs, ) -class InsecureCacheControlAdapter(CacheControlAdapter): +class HTTPAdapter(_SSLContextAdapterMixin, _BaseHTTPAdapter): + pass - def cert_verify(self, conn, url, verify, cert): - super(InsecureCacheControlAdapter, self).cert_verify( - conn=conn, url=url, verify=False, cert=cert - ) +class CacheControlAdapter(_SSLContextAdapterMixin, _BaseCacheControlAdapter): + pass -class PipSession(requests.Session): - timeout = None # type: Optional[int] +class InsecureHTTPAdapter(HTTPAdapter): + def cert_verify( + self, + conn: ConnectionPool, + url: str, + verify: Union[bool, str], + cert: Optional[Union[str, Tuple[str, str]]], + ) -> None: + super().cert_verify(conn=conn, url=url, verify=False, cert=cert) + + +class InsecureCacheControlAdapter(CacheControlAdapter): + def cert_verify( + self, + conn: ConnectionPool, + url: str, + verify: Union[bool, str], + cert: Optional[Union[str, Tuple[str, str]]], + ) -> None: + super().cert_verify(conn=conn, url=url, verify=False, cert=cert) + - def __init__(self, *args, **kwargs): +class PipSession(requests.Session): + timeout: Optional[int] = None + + def __init__( + self, + *args: Any, + retries: int = 0, + cache: Optional[str] = None, + trusted_hosts: Sequence[str] = (), + index_urls: Optional[List[str]] = None, + ssl_context: Optional["SSLContext"] = None, + **kwargs: Any, + ) -> None: """ :param trusted_hosts: Domains not to emit warnings for when not using HTTPS. """ - retries = kwargs.pop("retries", 0) - cache = kwargs.pop("cache", None) - trusted_hosts = kwargs.pop("trusted_hosts", []) # type: List[str] - index_urls = kwargs.pop("index_urls", None) - - super(PipSession, self).__init__(*args, **kwargs) + super().__init__(*args, **kwargs) # Namespace the attribute with "pip_" just in case to prevent # possible conflicts with the base class. - self.pip_trusted_origins = [] # type: List[Tuple[str, Optional[int]]] + self.pip_trusted_origins: List[Tuple[str, Optional[int]]] = [] # Attach our User Agent to the request self.headers["User-Agent"] = user_agent() @@ -257,19 +352,18 @@ def __init__(self, *args, **kwargs): # Set the total number of retries that a particular request can # have. total=retries, - # A 503 error from PyPI typically means that the Fastly -> Origin # connection got interrupted in some way. A 503 error in general # is typically considered a transient error so we'll go ahead and # retry it. # A 500 may indicate transient error in Amazon S3 + # A 502 may be a transient error from a CDN like CloudFlare or CloudFront # A 520 or 527 - may indicate transient error in CloudFlare - status_forcelist=[500, 503, 520, 527], - + status_forcelist=[500, 502, 503, 520, 527], # Add a small amount of back off between failed requests in # order to prevent hammering the service. backoff_factor=0.25, - ) + ) # type: ignore # Our Insecure HTTPAdapter disables HTTPS validation. It does not # support caching so we'll use it for all http:// URLs. @@ -287,13 +381,14 @@ def __init__(self, *args, **kwargs): secure_adapter = CacheControlAdapter( cache=SafeFileCache(cache), max_retries=retries, + ssl_context=ssl_context, ) self._trusted_host_adapter = InsecureCacheControlAdapter( cache=SafeFileCache(cache), max_retries=retries, ) else: - secure_adapter = HTTPAdapter(max_retries=retries) + secure_adapter = HTTPAdapter(max_retries=retries, ssl_context=ssl_context) self._trusted_host_adapter = insecure_adapter self.mount("https://", secure_adapter) @@ -305,8 +400,16 @@ def __init__(self, *args, **kwargs): for host in trusted_hosts: self.add_trusted_host(host, suppress_logging=True) - def add_trusted_host(self, host, source=None, suppress_logging=False): - # type: (str, Optional[str], bool) -> None + def update_index_urls(self, new_index_urls: List[str]) -> None: + """ + :param new_index_urls: New index urls to update the authentication + handler with. + """ + self.auth.index_urls = new_index_urls + + def add_trusted_host( + self, host: str, source: Optional[str] = None, suppress_logging: bool = False + ) -> None: """ :param host: It is okay to provide a host that has previously been added. @@ -314,46 +417,48 @@ def add_trusted_host(self, host, source=None, suppress_logging=False): string came from. """ if not suppress_logging: - msg = 'adding trusted host: {!r}'.format(host) + msg = f"adding trusted host: {host!r}" if source is not None: - msg += ' (from {})'.format(source) + msg += f" (from {source})" logger.info(msg) - host_port = parse_netloc(host) - if host_port not in self.pip_trusted_origins: - self.pip_trusted_origins.append(host_port) + parsed_host, parsed_port = parse_netloc(host) + if parsed_host is None: + raise ValueError(f"Trusted host URL must include a host part: {host!r}") + if (parsed_host, parsed_port) not in self.pip_trusted_origins: + self.pip_trusted_origins.append((parsed_host, parsed_port)) self.mount( - build_url_from_netloc(host) + '/', - self._trusted_host_adapter + build_url_from_netloc(host, scheme="http") + "/", self._trusted_host_adapter ) - if not host_port[1]: - # Mount wildcard ports for the same host. + self.mount(build_url_from_netloc(host) + "/", self._trusted_host_adapter) + if not parsed_port: self.mount( - build_url_from_netloc(host) + ':', - self._trusted_host_adapter + build_url_from_netloc(host, scheme="http") + ":", + self._trusted_host_adapter, ) + # Mount wildcard ports for the same host. + self.mount(build_url_from_netloc(host) + ":", self._trusted_host_adapter) - def iter_secure_origins(self): - # type: () -> Iterator[SecureOrigin] - for secure_origin in SECURE_ORIGINS: - yield secure_origin + def iter_secure_origins(self) -> Generator[SecureOrigin, None, None]: + yield from SECURE_ORIGINS for host, port in self.pip_trusted_origins: - yield ('*', host, '*' if port is None else port) + yield ("*", host, "*" if port is None else port) - def is_secure_origin(self, location): - # type: (Link) -> bool + def is_secure_origin(self, location: Link) -> bool: # Determine if this url used a secure transport mechanism - parsed = urllib_parse.urlparse(str(location)) + parsed = urllib.parse.urlparse(str(location)) origin_protocol, origin_host, origin_port = ( - parsed.scheme, parsed.hostname, parsed.port, + parsed.scheme, + parsed.hostname, + parsed.port, ) # The protocol to use to see if the protocol matches. # Don't count the repository type as part of the protocol: in # cases such as "git+ssh", only use "ssh". (I.e., Only verify against # the last scheme.) - origin_protocol = origin_protocol.rsplit('+', 1)[-1] + origin_protocol = origin_protocol.rsplit("+", 1)[-1] # Determine if our origin is a secure origin by looking through our # hardcoded list of secure origins, as well as any additional ones @@ -364,21 +469,15 @@ def is_secure_origin(self, location): continue try: - addr = ipaddress.ip_address( - None - if origin_host is None - else six.ensure_text(origin_host) - ) - network = ipaddress.ip_network( - six.ensure_text(secure_host) - ) + addr = ipaddress.ip_address(origin_host or "") + network = ipaddress.ip_network(secure_host) except ValueError: # We don't have both a valid address or a valid network, so # we'll check this origin against hostnames. if ( - origin_host and - origin_host.lower() != secure_host.lower() and - secure_host != "*" + origin_host + and origin_host.lower() != secure_host.lower() + and secure_host != "*" ): continue else: @@ -389,9 +488,9 @@ def is_secure_origin(self, location): # Check to see if the port matches. if ( - origin_port != secure_port and - secure_port != "*" and - secure_port is not None + origin_port != secure_port + and secure_port != "*" + and secure_port is not None ): continue @@ -413,9 +512,11 @@ def is_secure_origin(self, location): return False - def request(self, method, url, *args, **kwargs): + def request(self, method: str, url: str, *args: Any, **kwargs: Any) -> Response: # Allow setting a default timeout on a session kwargs.setdefault("timeout", self.timeout) + # Allow setting a default proxies on a session + kwargs.setdefault("proxies", self.proxies) # Dispatch the actual request - return super(PipSession, self).request(method, url, *args, **kwargs) + return super().request(method, url, *args, **kwargs) diff --git a/src/fetchcode/vcs/pip/_internal/network/utils.py b/src/fetchcode/vcs/pip/_internal/network/utils.py index 05f8b1a0..bba4c265 100644 --- a/src/fetchcode/vcs/pip/_internal/network/utils.py +++ b/src/fetchcode/vcs/pip/_internal/network/utils.py @@ -1,15 +1,65 @@ -from fetchcode.vcs.pip._vendor.requests.models import CONTENT_CHUNK_SIZE, Response +from typing import Dict, Generator -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._vendor.requests.models import Response -if MYPY_CHECK_RUNNING: - from typing import Iterator +from pip._internal.exceptions import NetworkConnectionError +# The following comments and HTTP headers were originally added by +# Donald Stufft in git commit 22c562429a61bb77172039e480873fb239dd8c03. +# +# We use Accept-Encoding: identity here because requests defaults to +# accepting compressed responses. This breaks in a variety of ways +# depending on how the server is configured. +# - Some servers will notice that the file isn't a compressible file +# and will leave the file alone and with an empty Content-Encoding +# - Some servers will notice that the file is already compressed and +# will leave the file alone, adding a Content-Encoding: gzip header +# - Some servers won't notice anything at all and will take a file +# that's already been compressed and compress it again, and set +# the Content-Encoding: gzip header +# By setting this to request only the identity encoding we're hoping +# to eliminate the third case. Hopefully there does not exist a server +# which when given a file will notice it is already compressed and that +# you're not asking for a compressed file and will then decompress it +# before sending because if that's the case I don't think it'll ever be +# possible to make this work. +HEADERS: Dict[str, str] = {"Accept-Encoding": "identity"} -def response_chunks(response, chunk_size=CONTENT_CHUNK_SIZE): - # type: (Response, int) -> Iterator[bytes] - """Given a requests Response, provide the data chunks. - """ +DOWNLOAD_CHUNK_SIZE = 256 * 1024 + + +def raise_for_status(resp: Response) -> None: + http_error_msg = "" + if isinstance(resp.reason, bytes): + # We attempt to decode utf-8 first because some servers + # choose to localize their reason strings. If the string + # isn't utf-8, we fall back to iso-8859-1 for all other + # encodings. + try: + reason = resp.reason.decode("utf-8") + except UnicodeDecodeError: + reason = resp.reason.decode("iso-8859-1") + else: + reason = resp.reason + + if 400 <= resp.status_code < 500: + http_error_msg = ( + f"{resp.status_code} Client Error: {reason} for url: {resp.url}" + ) + + elif 500 <= resp.status_code < 600: + http_error_msg = ( + f"{resp.status_code} Server Error: {reason} for url: {resp.url}" + ) + + if http_error_msg: + raise NetworkConnectionError(http_error_msg, response=resp) + + +def response_chunks( + response: Response, chunk_size: int = DOWNLOAD_CHUNK_SIZE +) -> Generator[bytes, None, None]: + """Given a requests Response, provide the data chunks.""" try: # Special case for urllib3. for chunk in response.raw.stream( diff --git a/src/fetchcode/vcs/pip/_internal/network/xmlrpc.py b/src/fetchcode/vcs/pip/_internal/network/xmlrpc.py index 1b43a594..22ec8d2f 100644 --- a/src/fetchcode/vcs/pip/_internal/network/xmlrpc.py +++ b/src/fetchcode/vcs/pip/_internal/network/xmlrpc.py @@ -1,44 +1,62 @@ """xmlrpclib.Transport implementation """ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - import logging +import urllib.parse +import xmlrpc.client +from typing import TYPE_CHECKING, Tuple + +from pip._internal.exceptions import NetworkConnectionError +from pip._internal.network.session import PipSession +from pip._internal.network.utils import raise_for_status -from fetchcode.vcs.pip._vendor import requests -# NOTE: XMLRPC Client is not annotated in typeshed as on 2017-07-17, which is -# why we ignore the type on this import -from fetchcode.vcs.pip._vendor.six.moves import xmlrpc_client # type: ignore -from fetchcode.vcs.pip._vendor.six.moves.urllib import parse as urllib_parse +if TYPE_CHECKING: + from xmlrpc.client import _HostType, _Marshallable + + from _typeshed import SizedBuffer logger = logging.getLogger(__name__) -class PipXmlrpcTransport(xmlrpc_client.Transport): +class PipXmlrpcTransport(xmlrpc.client.Transport): """Provide a `xmlrpclib.Transport` implementation via a `PipSession` object. """ - def __init__(self, index_url, session, use_datetime=False): - xmlrpc_client.Transport.__init__(self, use_datetime) - index_parts = urllib_parse.urlparse(index_url) + def __init__( + self, index_url: str, session: PipSession, use_datetime: bool = False + ) -> None: + super().__init__(use_datetime) + index_parts = urllib.parse.urlparse(index_url) self._scheme = index_parts.scheme self._session = session - def request(self, host, handler, request_body, verbose=False): + def request( + self, + host: "_HostType", + handler: str, + request_body: "SizedBuffer", + verbose: bool = False, + ) -> Tuple["_Marshallable", ...]: + assert isinstance(host, str) parts = (self._scheme, host, handler, None, None, None) - url = urllib_parse.urlunparse(parts) + url = urllib.parse.urlunparse(parts) try: - headers = {'Content-Type': 'text/xml'} - response = self._session.post(url, data=request_body, - headers=headers, stream=True) - response.raise_for_status() + headers = {"Content-Type": "text/xml"} + response = self._session.post( + url, + data=request_body, + headers=headers, + stream=True, + ) + raise_for_status(response) self.verbose = verbose return self.parse_response(response.raw) - except requests.HTTPError as exc: + except NetworkConnectionError as exc: + assert exc.response logger.critical( "HTTP error %s while getting %s", - exc.response.status_code, url, + exc.response.status_code, + url, ) raise diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib/filters/__init__.py b/src/fetchcode/vcs/pip/_internal/operations/build/__init__.py similarity index 100% rename from src/fetchcode/vcs/pip/_vendor/html5lib/filters/__init__.py rename to src/fetchcode/vcs/pip/_internal/operations/build/__init__.py diff --git a/src/fetchcode/vcs/pip/_internal/operations/build/build_tracker.py b/src/fetchcode/vcs/pip/_internal/operations/build/build_tracker.py new file mode 100644 index 00000000..0ed8dd23 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/operations/build/build_tracker.py @@ -0,0 +1,138 @@ +import contextlib +import hashlib +import logging +import os +from types import TracebackType +from typing import Dict, Generator, Optional, Type, Union + +from pip._internal.req.req_install import InstallRequirement +from pip._internal.utils.temp_dir import TempDirectory + +logger = logging.getLogger(__name__) + + +@contextlib.contextmanager +def update_env_context_manager(**changes: str) -> Generator[None, None, None]: + target = os.environ + + # Save values from the target and change them. + non_existent_marker = object() + saved_values: Dict[str, Union[object, str]] = {} + for name, new_value in changes.items(): + try: + saved_values[name] = target[name] + except KeyError: + saved_values[name] = non_existent_marker + target[name] = new_value + + try: + yield + finally: + # Restore original values in the target. + for name, original_value in saved_values.items(): + if original_value is non_existent_marker: + del target[name] + else: + assert isinstance(original_value, str) # for mypy + target[name] = original_value + + +@contextlib.contextmanager +def get_build_tracker() -> Generator["BuildTracker", None, None]: + root = os.environ.get("PIP_BUILD_TRACKER") + with contextlib.ExitStack() as ctx: + if root is None: + root = ctx.enter_context(TempDirectory(kind="build-tracker")).path + ctx.enter_context(update_env_context_manager(PIP_BUILD_TRACKER=root)) + logger.debug("Initialized build tracking at %s", root) + + with BuildTracker(root) as tracker: + yield tracker + + +class TrackerId(str): + """Uniquely identifying string provided to the build tracker.""" + + +class BuildTracker: + """Ensure that an sdist cannot request itself as a setup requirement. + + When an sdist is prepared, it identifies its setup requirements in the + context of ``BuildTracker.track()``. If a requirement shows up recursively, this + raises an exception. + + This stops fork bombs embedded in malicious packages.""" + + def __init__(self, root: str) -> None: + self._root = root + self._entries: Dict[TrackerId, InstallRequirement] = {} + logger.debug("Created build tracker: %s", self._root) + + def __enter__(self) -> "BuildTracker": + logger.debug("Entered build tracker: %s", self._root) + return self + + def __exit__( + self, + exc_type: Optional[Type[BaseException]], + exc_val: Optional[BaseException], + exc_tb: Optional[TracebackType], + ) -> None: + self.cleanup() + + def _entry_path(self, key: TrackerId) -> str: + hashed = hashlib.sha224(key.encode()).hexdigest() + return os.path.join(self._root, hashed) + + def add(self, req: InstallRequirement, key: TrackerId) -> None: + """Add an InstallRequirement to build tracking.""" + + # Get the file to write information about this requirement. + entry_path = self._entry_path(key) + + # Try reading from the file. If it exists and can be read from, a build + # is already in progress, so a LookupError is raised. + try: + with open(entry_path) as fp: + contents = fp.read() + except FileNotFoundError: + pass + else: + message = f"{req.link} is already being built: {contents}" + raise LookupError(message) + + # If we're here, req should really not be building already. + assert key not in self._entries + + # Start tracking this requirement. + with open(entry_path, "w", encoding="utf-8") as fp: + fp.write(str(req)) + self._entries[key] = req + + logger.debug("Added %s to build tracker %r", req, self._root) + + def remove(self, req: InstallRequirement, key: TrackerId) -> None: + """Remove an InstallRequirement from build tracking.""" + + # Delete the created file and the corresponding entry. + os.unlink(self._entry_path(key)) + del self._entries[key] + + logger.debug("Removed %s from build tracker %r", req, self._root) + + def cleanup(self) -> None: + for key, req in list(self._entries.items()): + self.remove(req, key) + + logger.debug("Removed build tracker: %r", self._root) + + @contextlib.contextmanager + def track(self, req: InstallRequirement, key: str) -> Generator[None, None, None]: + """Ensure that `key` cannot install itself as a setup requirement. + + :raises LookupError: If `key` was already provided in a parent invocation of + the context introduced by this method.""" + tracker_id = TrackerId(key) + self.add(req, tracker_id) + yield + self.remove(req, tracker_id) diff --git a/src/fetchcode/vcs/pip/_internal/operations/build/metadata.py b/src/fetchcode/vcs/pip/_internal/operations/build/metadata.py new file mode 100644 index 00000000..c66ac354 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/operations/build/metadata.py @@ -0,0 +1,39 @@ +"""Metadata generation logic for source distributions. +""" + +import os + +from pip._vendor.pyproject_hooks import BuildBackendHookCaller + +from pip._internal.build_env import BuildEnvironment +from pip._internal.exceptions import ( + InstallationSubprocessError, + MetadataGenerationFailed, +) +from pip._internal.utils.subprocess import runner_with_spinner_message +from pip._internal.utils.temp_dir import TempDirectory + + +def generate_metadata( + build_env: BuildEnvironment, backend: BuildBackendHookCaller, details: str +) -> str: + """Generate metadata using mechanisms described in PEP 517. + + Returns the generated metadata directory. + """ + metadata_tmpdir = TempDirectory(kind="modern-metadata", globally_managed=True) + + metadata_dir = metadata_tmpdir.path + + with build_env: + # Note that BuildBackendHookCaller implements a fallback for + # prepare_metadata_for_build_wheel, so we don't have to + # consider the possibility that this hook doesn't exist. + runner = runner_with_spinner_message("Preparing metadata (pyproject.toml)") + with backend.subprocess_runner(runner): + try: + distinfo_dir = backend.prepare_metadata_for_build_wheel(metadata_dir) + except InstallationSubprocessError as error: + raise MetadataGenerationFailed(package_details=details) from error + + return os.path.join(metadata_dir, distinfo_dir) diff --git a/src/fetchcode/vcs/pip/_internal/operations/build/metadata_editable.py b/src/fetchcode/vcs/pip/_internal/operations/build/metadata_editable.py new file mode 100644 index 00000000..27c69f0d --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/operations/build/metadata_editable.py @@ -0,0 +1,41 @@ +"""Metadata generation logic for source distributions. +""" + +import os + +from pip._vendor.pyproject_hooks import BuildBackendHookCaller + +from pip._internal.build_env import BuildEnvironment +from pip._internal.exceptions import ( + InstallationSubprocessError, + MetadataGenerationFailed, +) +from pip._internal.utils.subprocess import runner_with_spinner_message +from pip._internal.utils.temp_dir import TempDirectory + + +def generate_editable_metadata( + build_env: BuildEnvironment, backend: BuildBackendHookCaller, details: str +) -> str: + """Generate metadata using mechanisms described in PEP 660. + + Returns the generated metadata directory. + """ + metadata_tmpdir = TempDirectory(kind="modern-metadata", globally_managed=True) + + metadata_dir = metadata_tmpdir.path + + with build_env: + # Note that BuildBackendHookCaller implements a fallback for + # prepare_metadata_for_build_wheel/editable, so we don't have to + # consider the possibility that this hook doesn't exist. + runner = runner_with_spinner_message( + "Preparing editable metadata (pyproject.toml)" + ) + with backend.subprocess_runner(runner): + try: + distinfo_dir = backend.prepare_metadata_for_build_editable(metadata_dir) + except InstallationSubprocessError as error: + raise MetadataGenerationFailed(package_details=details) from error + + return os.path.join(metadata_dir, distinfo_dir) diff --git a/src/fetchcode/vcs/pip/_internal/operations/build/metadata_legacy.py b/src/fetchcode/vcs/pip/_internal/operations/build/metadata_legacy.py new file mode 100644 index 00000000..c01dd1c6 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/operations/build/metadata_legacy.py @@ -0,0 +1,74 @@ +"""Metadata generation logic for legacy source distributions. +""" + +import logging +import os + +from pip._internal.build_env import BuildEnvironment +from pip._internal.cli.spinners import open_spinner +from pip._internal.exceptions import ( + InstallationError, + InstallationSubprocessError, + MetadataGenerationFailed, +) +from pip._internal.utils.setuptools_build import make_setuptools_egg_info_args +from pip._internal.utils.subprocess import call_subprocess +from pip._internal.utils.temp_dir import TempDirectory + +logger = logging.getLogger(__name__) + + +def _find_egg_info(directory: str) -> str: + """Find an .egg-info subdirectory in `directory`.""" + filenames = [f for f in os.listdir(directory) if f.endswith(".egg-info")] + + if not filenames: + raise InstallationError(f"No .egg-info directory found in {directory}") + + if len(filenames) > 1: + raise InstallationError( + f"More than one .egg-info directory found in {directory}" + ) + + return os.path.join(directory, filenames[0]) + + +def generate_metadata( + build_env: BuildEnvironment, + setup_py_path: str, + source_dir: str, + isolated: bool, + details: str, +) -> str: + """Generate metadata using setup.py-based defacto mechanisms. + + Returns the generated metadata directory. + """ + logger.debug( + "Running setup.py (path:%s) egg_info for package %s", + setup_py_path, + details, + ) + + egg_info_dir = TempDirectory(kind="pip-egg-info", globally_managed=True).path + + args = make_setuptools_egg_info_args( + setup_py_path, + egg_info_dir=egg_info_dir, + no_user_config=isolated, + ) + + with build_env: + with open_spinner("Preparing metadata (setup.py)") as spinner: + try: + call_subprocess( + args, + cwd=source_dir, + command_desc="python setup.py egg_info", + spinner=spinner, + ) + except InstallationSubprocessError as error: + raise MetadataGenerationFailed(package_details=details) from error + + # Return the .egg-info directory. + return _find_egg_info(egg_info_dir) diff --git a/src/fetchcode/vcs/pip/_internal/operations/build/wheel.py b/src/fetchcode/vcs/pip/_internal/operations/build/wheel.py new file mode 100644 index 00000000..064811ad --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/operations/build/wheel.py @@ -0,0 +1,37 @@ +import logging +import os +from typing import Optional + +from pip._vendor.pyproject_hooks import BuildBackendHookCaller + +from pip._internal.utils.subprocess import runner_with_spinner_message + +logger = logging.getLogger(__name__) + + +def build_wheel_pep517( + name: str, + backend: BuildBackendHookCaller, + metadata_directory: str, + tempd: str, +) -> Optional[str]: + """Build one InstallRequirement using the PEP 517 build process. + + Returns path to wheel if successfully built. Otherwise, returns None. + """ + assert metadata_directory is not None + try: + logger.debug("Destination directory: %s", tempd) + + runner = runner_with_spinner_message( + f"Building wheel for {name} (pyproject.toml)" + ) + with backend.subprocess_runner(runner): + wheel_name = backend.build_wheel( + tempd, + metadata_directory=metadata_directory, + ) + except Exception: + logger.error("Failed building wheel for %s", name) + return None + return os.path.join(tempd, wheel_name) diff --git a/src/fetchcode/vcs/pip/_internal/operations/build/wheel_editable.py b/src/fetchcode/vcs/pip/_internal/operations/build/wheel_editable.py new file mode 100644 index 00000000..719d69dd --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/operations/build/wheel_editable.py @@ -0,0 +1,46 @@ +import logging +import os +from typing import Optional + +from pip._vendor.pyproject_hooks import BuildBackendHookCaller, HookMissing + +from pip._internal.utils.subprocess import runner_with_spinner_message + +logger = logging.getLogger(__name__) + + +def build_wheel_editable( + name: str, + backend: BuildBackendHookCaller, + metadata_directory: str, + tempd: str, +) -> Optional[str]: + """Build one InstallRequirement using the PEP 660 build process. + + Returns path to wheel if successfully built. Otherwise, returns None. + """ + assert metadata_directory is not None + try: + logger.debug("Destination directory: %s", tempd) + + runner = runner_with_spinner_message( + f"Building editable for {name} (pyproject.toml)" + ) + with backend.subprocess_runner(runner): + try: + wheel_name = backend.build_editable( + tempd, + metadata_directory=metadata_directory, + ) + except HookMissing as e: + logger.error( + "Cannot build editable %s because the build " + "backend does not have the %s hook", + name, + e, + ) + return None + except Exception: + logger.error("Failed building editable for %s", name) + return None + return os.path.join(tempd, wheel_name) diff --git a/src/fetchcode/vcs/pip/_internal/operations/build/wheel_legacy.py b/src/fetchcode/vcs/pip/_internal/operations/build/wheel_legacy.py new file mode 100644 index 00000000..3ee2a705 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/operations/build/wheel_legacy.py @@ -0,0 +1,102 @@ +import logging +import os.path +from typing import List, Optional + +from pip._internal.cli.spinners import open_spinner +from pip._internal.utils.setuptools_build import make_setuptools_bdist_wheel_args +from pip._internal.utils.subprocess import call_subprocess, format_command_args + +logger = logging.getLogger(__name__) + + +def format_command_result( + command_args: List[str], + command_output: str, +) -> str: + """Format command information for logging.""" + command_desc = format_command_args(command_args) + text = f"Command arguments: {command_desc}\n" + + if not command_output: + text += "Command output: None" + elif logger.getEffectiveLevel() > logging.DEBUG: + text += "Command output: [use --verbose to show]" + else: + if not command_output.endswith("\n"): + command_output += "\n" + text += f"Command output:\n{command_output}" + + return text + + +def get_legacy_build_wheel_path( + names: List[str], + temp_dir: str, + name: str, + command_args: List[str], + command_output: str, +) -> Optional[str]: + """Return the path to the wheel in the temporary build directory.""" + # Sort for determinism. + names = sorted(names) + if not names: + msg = f"Legacy build of wheel for {name!r} created no files.\n" + msg += format_command_result(command_args, command_output) + logger.warning(msg) + return None + + if len(names) > 1: + msg = ( + f"Legacy build of wheel for {name!r} created more than one file.\n" + f"Filenames (choosing first): {names}\n" + ) + msg += format_command_result(command_args, command_output) + logger.warning(msg) + + return os.path.join(temp_dir, names[0]) + + +def build_wheel_legacy( + name: str, + setup_py_path: str, + source_dir: str, + global_options: List[str], + build_options: List[str], + tempd: str, +) -> Optional[str]: + """Build one unpacked package using the "legacy" build process. + + Returns path to wheel if successfully built. Otherwise, returns None. + """ + wheel_args = make_setuptools_bdist_wheel_args( + setup_py_path, + global_options=global_options, + build_options=build_options, + destination_dir=tempd, + ) + + spin_message = f"Building wheel for {name} (setup.py)" + with open_spinner(spin_message) as spinner: + logger.debug("Destination directory: %s", tempd) + + try: + output = call_subprocess( + wheel_args, + command_desc="python setup.py bdist_wheel", + cwd=source_dir, + spinner=spinner, + ) + except Exception: + spinner.finish("error") + logger.error("Failed building wheel for %s", name) + return None + + names = os.listdir(tempd) + wheel_path = get_legacy_build_wheel_path( + names=names, + temp_dir=tempd, + name=name, + command_args=wheel_args, + command_output=output, + ) + return wheel_path diff --git a/src/fetchcode/vcs/pip/_internal/operations/check.py b/src/fetchcode/vcs/pip/_internal/operations/check.py index b9425d79..4b6fbc4c 100644 --- a/src/fetchcode/vcs/pip/_internal/operations/check.py +++ b/src/fetchcode/vcs/pip/_internal/operations/check.py @@ -1,99 +1,103 @@ """Validation of dependencies of packages """ -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False -# mypy: disallow-untyped-defs=False - import logging -from collections import namedtuple +from contextlib import suppress +from email.parser import Parser +from functools import reduce +from typing import ( + Callable, + Dict, + FrozenSet, + Generator, + Iterable, + List, + NamedTuple, + Optional, + Set, + Tuple, +) -from fetchcode.vcs.pip._vendor.packaging.utils import canonicalize_name -from fetchcode.vcs.pip._vendor.pkg_resources import RequirementParseError +from pip._vendor.packaging.requirements import Requirement +from pip._vendor.packaging.tags import Tag, parse_tag +from pip._vendor.packaging.utils import NormalizedName, canonicalize_name +from pip._vendor.packaging.version import Version -from fetchcode.vcs.pip._internal.distributions import ( - make_distribution_for_install_requirement, -) -from fetchcode.vcs.pip._internal.utils.misc import get_installed_distributions -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._internal.distributions import make_distribution_for_install_requirement +from pip._internal.metadata import get_default_environment +from pip._internal.metadata.base import BaseDistribution +from pip._internal.req.req_install import InstallRequirement logger = logging.getLogger(__name__) -if MYPY_CHECK_RUNNING: - from fetchcode.vcs.pip._internal.req.req_install import InstallRequirement - from typing import ( - Any, Callable, Dict, Optional, Set, Tuple, List - ) - # Shorthands - PackageSet = Dict[str, 'PackageDetails'] - Missing = Tuple[str, Any] - Conflicting = Tuple[str, str, Any] +class PackageDetails(NamedTuple): + version: Version + dependencies: List[Requirement] - MissingDict = Dict[str, List[Missing]] - ConflictingDict = Dict[str, List[Conflicting]] - CheckResult = Tuple[MissingDict, ConflictingDict] -PackageDetails = namedtuple('PackageDetails', ['version', 'requires']) +# Shorthands +PackageSet = Dict[NormalizedName, PackageDetails] +Missing = Tuple[NormalizedName, Requirement] +Conflicting = Tuple[NormalizedName, Version, Requirement] +MissingDict = Dict[NormalizedName, List[Missing]] +ConflictingDict = Dict[NormalizedName, List[Conflicting]] +CheckResult = Tuple[MissingDict, ConflictingDict] +ConflictDetails = Tuple[PackageSet, CheckResult] -def create_package_set_from_installed(**kwargs): - # type: (**Any) -> Tuple[PackageSet, bool] - """Converts a list of distributions into a PackageSet. - """ - # Default to using all packages installed on the system - if kwargs == {}: - kwargs = {"local_only": False, "skip": ()} +def create_package_set_from_installed() -> Tuple[PackageSet, bool]: + """Converts a list of distributions into a PackageSet.""" package_set = {} problems = False - for dist in get_installed_distributions(**kwargs): - name = canonicalize_name(dist.project_name) + env = get_default_environment() + for dist in env.iter_installed_distributions(local_only=False, skip=()): + name = dist.canonical_name try: - package_set[name] = PackageDetails(dist.version, dist.requires()) - except RequirementParseError as e: - # Don't crash on broken metadata - logger.warning("Error parsing requirements for %s: %s", name, e) + dependencies = list(dist.iter_dependencies()) + package_set[name] = PackageDetails(dist.version, dependencies) + except (OSError, ValueError) as e: + # Don't crash on unreadable or broken metadata. + logger.warning("Error parsing dependencies of %s: %s", name, e) problems = True return package_set, problems -def check_package_set(package_set, should_ignore=None): - # type: (PackageSet, Optional[Callable[[str], bool]]) -> CheckResult +def check_package_set( + package_set: PackageSet, should_ignore: Optional[Callable[[str], bool]] = None +) -> CheckResult: """Check if a package set is consistent If should_ignore is passed, it should be a callable that takes a package name and returns a boolean. """ - if should_ignore is None: - def should_ignore(name): - return False missing = {} conflicting = {} - for package_name in package_set: + for package_name, package_detail in package_set.items(): # Info about dependencies of package_name - missing_deps = set() # type: Set[Missing] - conflicting_deps = set() # type: Set[Conflicting] + missing_deps: Set[Missing] = set() + conflicting_deps: Set[Conflicting] = set() - if should_ignore(package_name): + if should_ignore and should_ignore(package_name): continue - for req in package_set[package_name].requires: - name = canonicalize_name(req.project_name) # type: str + for req in package_detail.dependencies: + name = canonicalize_name(req.name) # Check if it's missing if name not in package_set: missed = True if req.marker is not None: - missed = req.marker.evaluate() + missed = req.marker.evaluate({"extra": ""}) if missed: missing_deps.add((name, req)) continue # Check if there's a conflict - version = package_set[name].version # type: str + version = package_set[name].version if not req.specifier.contains(version, prereleases=True): conflicting_deps.add((name, version, req)) @@ -105,8 +109,7 @@ def should_ignore(name): return missing, conflicting -def check_install_conflicts(to_install): - # type: (List[InstallRequirement]) -> Tuple[PackageSet, CheckResult] +def check_install_conflicts(to_install: List[InstallRequirement]) -> ConflictDetails: """For checking if the dependency graph would be consistent after \ installing given requirements """ @@ -122,40 +125,55 @@ def check_install_conflicts(to_install): package_set, check_package_set( package_set, should_ignore=lambda name: name not in whitelist - ) + ), ) -def _simulate_installation_of(to_install, package_set): - # type: (List[InstallRequirement], PackageSet) -> Set[str] - """Computes the version of packages after installing to_install. - """ - +def check_unsupported( + packages: Iterable[BaseDistribution], + supported_tags: Iterable[Tag], +) -> Generator[BaseDistribution, None, None]: + for p in packages: + with suppress(FileNotFoundError): + wheel_file = p.read_text("WHEEL") + wheel_tags: FrozenSet[Tag] = reduce( + frozenset.union, + map(parse_tag, Parser().parsestr(wheel_file).get_all("Tag", [])), + frozenset(), + ) + if wheel_tags.isdisjoint(supported_tags): + yield p + + +def _simulate_installation_of( + to_install: List[InstallRequirement], package_set: PackageSet +) -> Set[NormalizedName]: + """Computes the version of packages after installing to_install.""" # Keep track of packages that were installed installed = set() # Modify it as installing requirement_set would (assuming no errors) for inst_req in to_install: abstract_dist = make_distribution_for_install_requirement(inst_req) - dist = abstract_dist.get_pkg_resources_distribution() - - name = canonicalize_name(dist.key) - package_set[name] = PackageDetails(dist.version, dist.requires()) + dist = abstract_dist.get_metadata_distribution() + name = dist.canonical_name + package_set[name] = PackageDetails(dist.version, list(dist.iter_dependencies())) installed.add(name) return installed -def _create_whitelist(would_be_installed, package_set): - # type: (Set[str], PackageSet) -> Set[str] +def _create_whitelist( + would_be_installed: Set[NormalizedName], package_set: PackageSet +) -> Set[NormalizedName]: packages_affected = set(would_be_installed) for package_name in package_set: if package_name in packages_affected: continue - for req in package_set[package_name].requires: + for req in package_set[package_name].dependencies: if canonicalize_name(req.name) in packages_affected: packages_affected.add(package_name) break diff --git a/src/fetchcode/vcs/pip/_internal/operations/freeze.py b/src/fetchcode/vcs/pip/_internal/operations/freeze.py index da3247c2..bb1039fb 100644 --- a/src/fetchcode/vcs/pip/_internal/operations/freeze.py +++ b/src/fetchcode/vcs/pip/_internal/operations/freeze.py @@ -1,81 +1,46 @@ -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import - import collections import logging import os +from typing import Container, Dict, Generator, Iterable, List, NamedTuple, Optional, Set -from fetchcode.vcs.pip._vendor import six -from fetchcode.vcs.pip._vendor.packaging.utils import canonicalize_name -from fetchcode.vcs.pip._vendor.pkg_resources import RequirementParseError +from pip._vendor.packaging.utils import canonicalize_name +from pip._vendor.packaging.version import InvalidVersion -from fetchcode.vcs.pip._internal.exceptions import BadCommand, InstallationError -from fetchcode.vcs.pip._internal.req.constructors import ( +from pip._internal.exceptions import BadCommand, InstallationError +from pip._internal.metadata import BaseDistribution, get_environment +from pip._internal.req.constructors import ( install_req_from_editable, install_req_from_line, ) -from fetchcode.vcs.pip._internal.req.req_file import COMMENT_RE -from fetchcode.vcs.pip._internal.utils.direct_url_helpers import ( - direct_url_as_pep440_direct_reference, - dist_get_direct_url, -) -from fetchcode.vcs.pip._internal.utils.misc import ( - dist_is_editable, - get_installed_distributions, -) -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._internal.req.req_file import COMMENT_RE +from pip._internal.utils.direct_url_helpers import direct_url_as_pep440_direct_reference -if MYPY_CHECK_RUNNING: - from typing import ( - Iterator, Optional, List, Container, Set, Dict, Tuple, Iterable, Union - ) - from fetchcode.vcs.pip._internal.cache import WheelCache - from fetchcode.vcs.pip._vendor.pkg_resources import ( - Distribution, Requirement - ) +logger = logging.getLogger(__name__) - RequirementInfo = Tuple[Optional[Union[str, Requirement]], bool, List[str]] - -logger = logging.getLogger(__name__) +class _EditableInfo(NamedTuple): + requirement: str + comments: List[str] def freeze( - requirement=None, # type: Optional[List[str]] - find_links=None, # type: Optional[List[str]] - local_only=None, # type: Optional[bool] - user_only=None, # type: Optional[bool] - paths=None, # type: Optional[List[str]] - isolated=False, # type: bool - wheel_cache=None, # type: Optional[WheelCache] - exclude_editable=False, # type: bool - skip=() # type: Container[str] -): - # type: (...) -> Iterator[str] - find_links = find_links or [] + requirement: Optional[List[str]] = None, + local_only: bool = False, + user_only: bool = False, + paths: Optional[List[str]] = None, + isolated: bool = False, + exclude_editable: bool = False, + skip: Container[str] = (), +) -> Generator[str, None, None]: + installations: Dict[str, FrozenRequirement] = {} - for link in find_links: - yield '-f {}'.format(link) - installations = {} # type: Dict[str, FrozenRequirement] - for dist in get_installed_distributions(local_only=local_only, - skip=(), - user_only=user_only, - paths=paths): - try: - req = FrozenRequirement.from_dist(dist) - except RequirementParseError as exc: - # We include dist rather than dist.project_name because the - # dist string includes more information, like the version and - # location. We also include the exception message to aid - # troubleshooting. - logger.warning( - 'Could not generate requirement for distribution %r: %s', - dist, exc - ) - continue + dists = get_environment(paths).iter_installed_distributions( + local_only=local_only, + skip=(), + user_only=user_only, + ) + for dist in dists: + req = FrozenRequirement.from_dist(dist) if exclude_editable and req.editable: continue installations[req.canonical_name] = req @@ -85,42 +50,50 @@ def freeze( # should only be emitted once, even if the same option is in multiple # requirements files, so we need to keep track of what has been emitted # so that we don't emit it again if it's seen again - emitted_options = set() # type: Set[str] + emitted_options: Set[str] = set() # keep track of which files a requirement is in so that we can # give an accurate warning if a requirement appears multiple times. - req_files = collections.defaultdict(list) # type: Dict[str, List[str]] + req_files: Dict[str, List[str]] = collections.defaultdict(list) for req_file_path in requirement: with open(req_file_path) as req_file: for line in req_file: - if (not line.strip() or - line.strip().startswith('#') or - line.startswith(( - '-r', '--requirement', - '-Z', '--always-unzip', - '-f', '--find-links', - '-i', '--index-url', - '--pre', - '--trusted-host', - '--process-dependency-links', - '--extra-index-url'))): + if ( + not line.strip() + or line.strip().startswith("#") + or line.startswith( + ( + "-r", + "--requirement", + "-f", + "--find-links", + "-i", + "--index-url", + "--pre", + "--trusted-host", + "--process-dependency-links", + "--extra-index-url", + "--use-feature", + ) + ) + ): line = line.rstrip() if line not in emitted_options: emitted_options.add(line) yield line continue - if line.startswith('-e') or line.startswith('--editable'): - if line.startswith('-e'): + if line.startswith("-e") or line.startswith("--editable"): + if line.startswith("-e"): line = line[2:].strip() else: - line = line[len('--editable'):].strip().lstrip('=') + line = line[len("--editable") :].strip().lstrip("=") line_req = install_req_from_editable( line, isolated=isolated, ) else: line_req = install_req_from_line( - COMMENT_RE.sub('', line).strip(), + COMMENT_RE.sub("", line).strip(), isolated=isolated, ) @@ -128,15 +101,15 @@ def freeze( logger.info( "Skipping line in requirement file [%s] because " "it's not clear what it would install: %s", - req_file_path, line.strip(), + req_file_path, + line.strip(), ) logger.info( " (add #egg=PackageName to the URL to avoid" " this warning)" ) else: - line_req_canonical_name = canonicalize_name( - line_req.name) + line_req_canonical_name = canonicalize_name(line_req.name) if line_req_canonical_name not in installations: # either it's not installed, or it is installed # but has been processed already @@ -145,99 +118,116 @@ def freeze( "Requirement file [%s] contains %s, but " "package %r is not installed", req_file_path, - COMMENT_RE.sub('', line).strip(), - line_req.name + COMMENT_RE.sub("", line).strip(), + line_req.name, ) else: req_files[line_req.name].append(req_file_path) else: - yield str(installations[ - line_req_canonical_name]).rstrip() + yield str(installations[line_req_canonical_name]).rstrip() del installations[line_req_canonical_name] req_files[line_req.name].append(req_file_path) # Warn about requirements that were included multiple times (in a # single requirements file or in different requirements files). - for name, files in six.iteritems(req_files): + for name, files in req_files.items(): if len(files) > 1: - logger.warning("Requirement %s included multiple times [%s]", - name, ', '.join(sorted(set(files)))) + logger.warning( + "Requirement %s included multiple times [%s]", + name, + ", ".join(sorted(set(files))), + ) - yield( - '## The following requirements were added by ' - 'pip freeze:' - ) - for installation in sorted( - installations.values(), key=lambda x: x.name.lower()): + yield ("## The following requirements were added by pip freeze:") + for installation in sorted(installations.values(), key=lambda x: x.name.lower()): if installation.canonical_name not in skip: yield str(installation).rstrip() -def get_requirement_info(dist): - # type: (Distribution) -> RequirementInfo +def _format_as_name_version(dist: BaseDistribution) -> str: + try: + dist_version = dist.version + except InvalidVersion: + # legacy version + return f"{dist.raw_name}==={dist.raw_version}" + else: + return f"{dist.raw_name}=={dist_version}" + + +def _get_editable_info(dist: BaseDistribution) -> _EditableInfo: """ - Compute and return values (req, editable, comments) for use in + Compute and return values (req, comments) for use in FrozenRequirement.from_dist(). """ - if not dist_is_editable(dist): - return (None, False, []) + editable_project_location = dist.editable_project_location + assert editable_project_location + location = os.path.normcase(os.path.abspath(editable_project_location)) - location = os.path.normcase(os.path.abspath(dist.location)) + from pip._internal.vcs import RemoteNotFoundError, RemoteNotValidError, vcs - from fetchcode.vcs.pip._internal.vcs import vcs, RemoteNotFoundError vcs_backend = vcs.get_backend_for_dir(location) if vcs_backend is None: - req = dist.as_requirement() + display = _format_as_name_version(dist) logger.debug( - 'No VCS found for editable requirement "%s" in: %r', req, + 'No VCS found for editable requirement "%s" in: %r', + display, location, ) - comments = [ - '# Editable install with no version control ({})'.format(req) - ] - return (location, True, comments) + return _EditableInfo( + requirement=location, + comments=[f"# Editable install with no version control ({display})"], + ) + + vcs_name = type(vcs_backend).__name__ try: - req = vcs_backend.get_src_requirement(location, dist.project_name) + req = vcs_backend.get_src_requirement(location, dist.raw_name) except RemoteNotFoundError: - req = dist.as_requirement() - comments = [ - '# Editable {} install with no remote ({})'.format( - type(vcs_backend).__name__, req, - ) - ] - return (location, True, comments) - + display = _format_as_name_version(dist) + return _EditableInfo( + requirement=location, + comments=[f"# Editable {vcs_name} install with no remote ({display})"], + ) + except RemoteNotValidError as ex: + display = _format_as_name_version(dist) + return _EditableInfo( + requirement=location, + comments=[ + f"# Editable {vcs_name} install ({display}) with either a deleted " + f"local remote or invalid URI:", + f"# '{ex.url}'", + ], + ) except BadCommand: logger.warning( - 'cannot determine version of editable source in %s ' - '(%s command not found in path)', + "cannot determine version of editable source in %s " + "(%s command not found in path)", location, vcs_backend.name, ) - return (None, True, []) - + return _EditableInfo(requirement=location, comments=[]) except InstallationError as exc: - logger.warning( - "Error when trying to get requirement for VCS system %s, " - "falling back to uneditable format", exc - ) + logger.warning("Error when trying to get requirement for VCS system %s", exc) else: - if req is not None: - return (req, True, []) + return _EditableInfo(requirement=req, comments=[]) - logger.warning( - 'Could not determine repository location of %s', location - ) - comments = ['## !! Could not determine repository location'] + logger.warning("Could not determine repository location of %s", location) - return (None, False, comments) + return _EditableInfo( + requirement=location, + comments=["## !! Could not determine repository location"], + ) -class FrozenRequirement(object): - def __init__(self, name, req, editable, comments=()): - # type: (str, Union[str, Requirement], bool, Iterable[str]) -> None +class FrozenRequirement: + def __init__( + self, + name: str, + req: str, + editable: bool, + comments: Iterable[str] = (), + ) -> None: self.name = name self.canonical_name = canonicalize_name(name) self.req = req @@ -245,28 +235,24 @@ def __init__(self, name, req, editable, comments=()): self.comments = comments @classmethod - def from_dist(cls, dist): - # type: (Distribution) -> FrozenRequirement - # TODO `get_requirement_info` is taking care of editable requirements. - # TODO This should be refactored when we will add detection of - # editable that provide .dist-info metadata. - req, editable, comments = get_requirement_info(dist) - if req is None and not editable: - # if PEP 610 metadata is present, attempt to use it - direct_url = dist_get_direct_url(dist) + def from_dist(cls, dist: BaseDistribution) -> "FrozenRequirement": + editable = dist.editable + if editable: + req, comments = _get_editable_info(dist) + else: + comments = [] + direct_url = dist.direct_url if direct_url: - req = direct_url_as_pep440_direct_reference( - direct_url, dist.project_name - ) - comments = [] - if req is None: - # name==version requirement - req = dist.as_requirement() + # if PEP 610 metadata is present, use it + req = direct_url_as_pep440_direct_reference(direct_url, dist.raw_name) + else: + # name==version requirement + req = _format_as_name_version(dist) - return cls(dist.project_name, req, editable, comments=comments) + return cls(dist.raw_name, req, editable, comments=comments) - def __str__(self): + def __str__(self) -> str: req = self.req if self.editable: - req = '-e {}'.format(req) - return '\n'.join(list(self.comments) + [str(req)]) + '\n' + req = f"-e {req}" + return "\n".join(list(self.comments) + [str(req)]) + "\n" diff --git a/src/fetchcode/vcs/pip/_internal/operations/install/editable_legacy.py b/src/fetchcode/vcs/pip/_internal/operations/install/editable_legacy.py index c24c7e1c..9aaa699a 100644 --- a/src/fetchcode/vcs/pip/_internal/operations/install/editable_legacy.py +++ b/src/fetchcode/vcs/pip/_internal/operations/install/editable_legacy.py @@ -1,43 +1,37 @@ """Legacy editable installation process, i.e. `setup.py develop`. """ -import logging - -from fetchcode.vcs.pip._internal.utils.logging import indent_log -from fetchcode.vcs.pip._internal.utils.setuptools_build import make_setuptools_develop_args -from fetchcode.vcs.pip._internal.utils.subprocess import call_subprocess -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -if MYPY_CHECK_RUNNING: - from typing import List, Optional, Sequence - - from fetchcode.vcs.pip._internal.build_env import BuildEnvironment +import logging +from typing import Optional, Sequence +from pip._internal.build_env import BuildEnvironment +from pip._internal.utils.logging import indent_log +from pip._internal.utils.setuptools_build import make_setuptools_develop_args +from pip._internal.utils.subprocess import call_subprocess logger = logging.getLogger(__name__) def install_editable( - install_options, # type: List[str] - global_options, # type: Sequence[str] - prefix, # type: Optional[str] - home, # type: Optional[str] - use_user_site, # type: bool - name, # type: str - setup_py_path, # type: str - isolated, # type: bool - build_env, # type: BuildEnvironment - unpacked_source_directory, # type: str -): - # type: (...) -> None + *, + global_options: Sequence[str], + prefix: Optional[str], + home: Optional[str], + use_user_site: bool, + name: str, + setup_py_path: str, + isolated: bool, + build_env: BuildEnvironment, + unpacked_source_directory: str, +) -> None: """Install a package in editable mode. Most arguments are pass-through to setuptools. """ - logger.info('Running setup.py develop for %s', name) + logger.info("Running setup.py develop for %s", name) args = make_setuptools_develop_args( setup_py_path, global_options=global_options, - install_options=install_options, no_user_config=isolated, prefix=prefix, home=home, @@ -48,5 +42,6 @@ def install_editable( with build_env: call_subprocess( args, + command_desc="python setup.py develop", cwd=unpacked_source_directory, ) diff --git a/src/fetchcode/vcs/pip/_internal/operations/install/legacy.py b/src/fetchcode/vcs/pip/_internal/operations/install/legacy.py deleted file mode 100644 index a899ec72..00000000 --- a/src/fetchcode/vcs/pip/_internal/operations/install/legacy.py +++ /dev/null @@ -1,142 +0,0 @@ -"""Legacy installation process, i.e. `setup.py install`. -""" - -import logging -import os -import sys -from distutils.util import change_root - -from fetchcode.vcs.pip._internal.utils.deprecation import deprecated -from fetchcode.vcs.pip._internal.utils.logging import indent_log -from fetchcode.vcs.pip._internal.utils.misc import ensure_dir -from fetchcode.vcs.pip._internal.utils.setuptools_build import make_setuptools_install_args -from fetchcode.vcs.pip._internal.utils.subprocess import runner_with_spinner_message -from fetchcode.vcs.pip._internal.utils.temp_dir import TempDirectory -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import List, Optional, Sequence - - from fetchcode.vcs.pip._internal.build_env import BuildEnvironment - from fetchcode.vcs.pip._internal.models.scheme import Scheme - - -logger = logging.getLogger(__name__) - - -class LegacyInstallFailure(Exception): - def __init__(self): - # type: () -> None - self.parent = sys.exc_info() - - -def install( - install_options, # type: List[str] - global_options, # type: Sequence[str] - root, # type: Optional[str] - home, # type: Optional[str] - prefix, # type: Optional[str] - use_user_site, # type: bool - pycompile, # type: bool - scheme, # type: Scheme - setup_py_path, # type: str - isolated, # type: bool - req_name, # type: str - build_env, # type: BuildEnvironment - unpacked_source_directory, # type: str - req_description, # type: str -): - # type: (...) -> bool - - header_dir = scheme.headers - - with TempDirectory(kind="record") as temp_dir: - try: - record_filename = os.path.join(temp_dir.path, 'install-record.txt') - install_args = make_setuptools_install_args( - setup_py_path, - global_options=global_options, - install_options=install_options, - record_filename=record_filename, - root=root, - prefix=prefix, - header_dir=header_dir, - home=home, - use_user_site=use_user_site, - no_user_config=isolated, - pycompile=pycompile, - ) - - runner = runner_with_spinner_message( - "Running setup.py install for {}".format(req_name) - ) - with indent_log(), build_env: - runner( - cmd=install_args, - cwd=unpacked_source_directory, - ) - - if not os.path.exists(record_filename): - logger.debug('Record file %s not found', record_filename) - # Signal to the caller that we didn't install the new package - return False - - except Exception: - # Signal to the caller that we didn't install the new package - raise LegacyInstallFailure - - # At this point, we have successfully installed the requirement. - - # We intentionally do not use any encoding to read the file because - # setuptools writes the file using distutils.file_util.write_file, - # which does not specify an encoding. - with open(record_filename) as f: - record_lines = f.read().splitlines() - - def prepend_root(path): - # type: (str) -> str - if root is None or not os.path.isabs(path): - return path - else: - return change_root(root, path) - - for line in record_lines: - directory = os.path.dirname(line) - if directory.endswith('.egg-info'): - egg_info_dir = prepend_root(directory) - break - else: - deprecated( - reason=( - "{} did not indicate that it installed an " - ".egg-info directory. Only setup.py projects " - "generating .egg-info directories are supported." - ).format(req_description), - replacement=( - "for maintainers: updating the setup.py of {0}. " - "For users: contact the maintainers of {0} to let " - "them know to update their setup.py.".format( - req_name - ) - ), - gone_in="20.2", - issue=6998, - ) - # FIXME: put the record somewhere - return True - - new_lines = [] - for line in record_lines: - filename = line.strip() - if os.path.isdir(filename): - filename += os.path.sep - new_lines.append( - os.path.relpath(prepend_root(filename), egg_info_dir) - ) - new_lines.sort() - ensure_dir(egg_info_dir) - inst_files_path = os.path.join(egg_info_dir, 'installed-files.txt') - with open(inst_files_path, 'w') as f: - f.write('\n'.join(new_lines) + '\n') - - return True diff --git a/src/fetchcode/vcs/pip/_internal/operations/install/wheel.py b/src/fetchcode/vcs/pip/_internal/operations/install/wheel.py index b9ac38b9..aef42aa9 100644 --- a/src/fetchcode/vcs/pip/_internal/operations/install/wheel.py +++ b/src/fetchcode/vcs/pip/_internal/operations/install/wheel.py @@ -1,147 +1,132 @@ """Support for installing and building the "wheel" binary package format. """ -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False - -from __future__ import absolute_import - import collections import compileall import contextlib import csv +import importlib import logging import os.path import re import shutil -import stat import sys import warnings from base64 import urlsafe_b64encode -from itertools import starmap -from zipfile import ZipFile - -from fetchcode.vcs.pip._vendor import pkg_resources -from fetchcode.vcs.pip._vendor.distlib.scripts import ScriptMaker -from fetchcode.vcs.pip._vendor.distlib.util import get_export_entry -from fetchcode.vcs.pip._vendor.six import StringIO - -from fetchcode.vcs.pip._internal.exceptions import InstallationError -from fetchcode.vcs.pip._internal.locations import get_major_minor_version -from fetchcode.vcs.pip._internal.models.direct_url import DIRECT_URL_METADATA_NAME, DirectUrl -from fetchcode.vcs.pip._internal.utils.filesystem import adjacent_tmp_file, replace -from fetchcode.vcs.pip._internal.utils.misc import captured_stdout, ensure_dir, hash_file -from fetchcode.vcs.pip._internal.utils.temp_dir import TempDirectory -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.utils.unpacking import current_umask, unpack_file -from fetchcode.vcs.pip._internal.utils.wheel import parse_wheel - -if MYPY_CHECK_RUNNING: - from email.message import Message - from typing import ( - Dict, List, Optional, Sequence, Tuple, Any, - Iterable, Iterator, Callable, Set, - ) - - from fetchcode.vcs.pip._internal.models.scheme import Scheme - from fetchcode.vcs.pip._internal.utils.filesystem import NamedTemporaryFileResult - - InstalledCSVRow = Tuple[str, ...] +from email.message import Message +from itertools import chain, filterfalse, starmap +from typing import ( + IO, + TYPE_CHECKING, + Any, + BinaryIO, + Callable, + Dict, + Generator, + Iterable, + Iterator, + List, + NewType, + Optional, + Protocol, + Sequence, + Set, + Tuple, + Union, + cast, +) +from zipfile import ZipFile, ZipInfo + +from pip._vendor.distlib.scripts import ScriptMaker +from pip._vendor.distlib.util import get_export_entry +from pip._vendor.packaging.utils import canonicalize_name + +from pip._internal.exceptions import InstallationError +from pip._internal.locations import get_major_minor_version +from pip._internal.metadata import ( + BaseDistribution, + FilesystemWheel, + get_wheel_distribution, +) +from pip._internal.models.direct_url import DIRECT_URL_METADATA_NAME, DirectUrl +from pip._internal.models.scheme import SCHEME_KEYS, Scheme +from pip._internal.utils.filesystem import adjacent_tmp_file, replace +from pip._internal.utils.misc import StreamWrapper, ensure_dir, hash_file, partition +from pip._internal.utils.unpacking import ( + current_umask, + is_within_directory, + set_extracted_file_to_default_mode_plus_executable, + zip_item_is_executable, +) +from pip._internal.utils.wheel import parse_wheel + +if TYPE_CHECKING: + + class File(Protocol): + src_record_path: "RecordPath" + dest_path: str + changed: bool + + def save(self) -> None: + pass logger = logging.getLogger(__name__) - -def normpath(src, p): - # type: (str, str) -> str - return os.path.relpath(src, p).replace(os.path.sep, '/') +RecordPath = NewType("RecordPath", str) +InstalledCSVRow = Tuple[RecordPath, str, Union[int, str]] -def rehash(path, blocksize=1 << 20): - # type: (str, int) -> Tuple[str, str] +def rehash(path: str, blocksize: int = 1 << 20) -> Tuple[str, str]: """Return (encoded_digest, length) for path using hashlib.sha256()""" h, length = hash_file(path, blocksize) - digest = 'sha256=' + urlsafe_b64encode( - h.digest() - ).decode('latin1').rstrip('=') - # unicode/str python2 issues - return (digest, str(length)) # type: ignore + digest = "sha256=" + urlsafe_b64encode(h.digest()).decode("latin1").rstrip("=") + return (digest, str(length)) -def csv_io_kwargs(mode): - # type: (str) -> Dict[str, Any] +def csv_io_kwargs(mode: str) -> Dict[str, Any]: """Return keyword arguments to properly open a CSV file in the given mode. """ - if sys.version_info.major < 3: - return {'mode': '{}b'.format(mode)} - else: - return {'mode': mode, 'newline': ''} + return {"mode": mode, "newline": "", "encoding": "utf-8"} -def fix_script(path): - # type: (str) -> Optional[bool] +def fix_script(path: str) -> bool: """Replace #!python with #!/path/to/python Return True if file was changed. """ # XXX RECORD hashes will need to be updated - if os.path.isfile(path): - with open(path, 'rb') as script: - firstline = script.readline() - if not firstline.startswith(b'#!python'): - return False - exename = sys.executable.encode(sys.getfilesystemencoding()) - firstline = b'#!' + exename + os.linesep.encode("ascii") - rest = script.read() - with open(path, 'wb') as script: - script.write(firstline) - script.write(rest) - return True - return None - - -def wheel_root_is_purelib(metadata): - # type: (Message) -> bool + assert os.path.isfile(path) + + with open(path, "rb") as script: + firstline = script.readline() + if not firstline.startswith(b"#!python"): + return False + exename = sys.executable.encode(sys.getfilesystemencoding()) + firstline = b"#!" + exename + os.linesep.encode("ascii") + rest = script.read() + with open(path, "wb") as script: + script.write(firstline) + script.write(rest) + return True + + +def wheel_root_is_purelib(metadata: Message) -> bool: return metadata.get("Root-Is-Purelib", "").lower() == "true" -def get_entrypoints(filename): - # type: (str) -> Tuple[Dict[str, str], Dict[str, str]] - if not os.path.exists(filename): - return {}, {} - - # This is done because you can pass a string to entry_points wrappers which - # means that they may or may not be valid INI files. The attempt here is to - # strip leading and trailing whitespace in order to make them valid INI - # files. - with open(filename) as fp: - data = StringIO() - for line in fp: - data.write(line.strip()) - data.write("\n") - data.seek(0) - - # get the entry points and then the script names - entry_points = pkg_resources.EntryPoint.parse_map(data) - console = entry_points.get('console_scripts', {}) - gui = entry_points.get('gui_scripts', {}) - - def _split_ep(s): - # type: (pkg_resources.EntryPoint) -> Tuple[str, str] - """get the string representation of EntryPoint, - remove space and split on '=' - """ - split_parts = str(s).replace(" ", "").split("=") - return split_parts[0], split_parts[1] - - # convert the EntryPoint objects into strings with module:function - console = dict(_split_ep(v) for v in console.values()) - gui = dict(_split_ep(v) for v in gui.values()) - return console, gui - - -def message_about_scripts_not_on_PATH(scripts): - # type: (Sequence[str]) -> Optional[str] +def get_entrypoints(dist: BaseDistribution) -> Tuple[Dict[str, str], Dict[str, str]]: + console_scripts = {} + gui_scripts = {} + for entry_point in dist.iter_entry_points(): + if entry_point.group == "console_scripts": + console_scripts[entry_point.name] = entry_point.value + elif entry_point.group == "gui_scripts": + gui_scripts[entry_point.name] = entry_point.value + return console_scripts, gui_scripts + + +def message_about_scripts_not_on_PATH(scripts: Sequence[str]) -> Optional[str]: """Determine if any scripts are not on PATH and format a warning. Returns a warning message if one or more scripts are not on PATH, otherwise None. @@ -150,7 +135,7 @@ def message_about_scripts_not_on_PATH(scripts): return None # Group scripts by the path they were installed in - grouped_by_dir = collections.defaultdict(set) # type: Dict[str, Set[str]] + grouped_by_dir: Dict[str, Set[str]] = collections.defaultdict(set) for destfile in scripts: parent_dir = os.path.dirname(destfile) script_name = os.path.basename(destfile) @@ -158,33 +143,35 @@ def message_about_scripts_not_on_PATH(scripts): # We don't want to warn for directories that are on PATH. not_warn_dirs = [ - os.path.normcase(i).rstrip(os.sep) for i in - os.environ.get("PATH", "").split(os.pathsep) + os.path.normcase(os.path.normpath(i)).rstrip(os.sep) + for i in os.environ.get("PATH", "").split(os.pathsep) ] # If an executable sits with sys.executable, we don't warn for it. # This covers the case of venv invocations without activating the venv. - not_warn_dirs.append(os.path.normcase(os.path.dirname(sys.executable))) - warn_for = { - parent_dir: scripts for parent_dir, scripts in grouped_by_dir.items() - if os.path.normcase(parent_dir) not in not_warn_dirs - } # type: Dict[str, Set[str]] + not_warn_dirs.append( + os.path.normcase(os.path.normpath(os.path.dirname(sys.executable))) + ) + warn_for: Dict[str, Set[str]] = { + parent_dir: scripts + for parent_dir, scripts in grouped_by_dir.items() + if os.path.normcase(os.path.normpath(parent_dir)) not in not_warn_dirs + } if not warn_for: return None # Format a message msg_lines = [] for parent_dir, dir_scripts in warn_for.items(): - sorted_scripts = sorted(dir_scripts) # type: List[str] + sorted_scripts: List[str] = sorted(dir_scripts) if len(sorted_scripts) == 1: - start_text = "script {} is".format(sorted_scripts[0]) + start_text = f"script {sorted_scripts[0]} is" else: start_text = "scripts {} are".format( ", ".join(sorted_scripts[:-1]) + " and " + sorted_scripts[-1] ) msg_lines.append( - "The {} installed in '{}' which is not on PATH." - .format(start_text, parent_dir) + f"The {start_text} installed in '{parent_dir}' which is not on PATH." ) last_line_fmt = ( @@ -211,9 +198,13 @@ def message_about_scripts_not_on_PATH(scripts): return "\n".join(msg_lines) -def sorted_outrows(outrows): - # type: (Iterable[InstalledCSVRow]) -> List[InstalledCSVRow] - """Return the given rows of a RECORD file in sorted order. +def _normalized_outrows( + outrows: Iterable[InstalledCSVRow], +) -> List[Tuple[str, str, str]]: + """Normalize the given rows of a RECORD file. + + Items in each row are converted into str. Rows are then sorted to make + the value more predictable for tests. Each row is a 3-tuple (path, hash, size) and corresponds to a record of a RECORD file (see PEP 376 and PEP 427 for details). For the rows @@ -228,78 +219,219 @@ def sorted_outrows(outrows): # coerce each element to a string to avoid a TypeError in this case. # For additional background, see-- # https://github.com/pypa/pip/issues/5868 - return sorted(outrows, key=lambda row: tuple(str(x) for x in row)) + return sorted( + (record_path, hash_, str(size)) for record_path, hash_, size in outrows + ) + + +def _record_to_fs_path(record_path: RecordPath, lib_dir: str) -> str: + return os.path.join(lib_dir, record_path) + + +def _fs_to_record_path(path: str, lib_dir: str) -> RecordPath: + # On Windows, do not handle relative paths if they belong to different + # logical disks + if os.path.splitdrive(path)[0].lower() == os.path.splitdrive(lib_dir)[0].lower(): + path = os.path.relpath(path, lib_dir) + + path = path.replace(os.path.sep, "/") + return cast("RecordPath", path) def get_csv_rows_for_installed( - old_csv_rows, # type: Iterable[List[str]] - installed, # type: Dict[str, str] - changed, # type: Set[str] - generated, # type: List[str] - lib_dir, # type: str -): - # type: (...) -> List[InstalledCSVRow] + old_csv_rows: List[List[str]], + installed: Dict[RecordPath, RecordPath], + changed: Set[RecordPath], + generated: List[str], + lib_dir: str, +) -> List[InstalledCSVRow]: """ :param installed: A map from archive RECORD path to installation RECORD path. """ - installed_rows = [] # type: List[InstalledCSVRow] + installed_rows: List[InstalledCSVRow] = [] for row in old_csv_rows: if len(row) > 3: - logger.warning( - 'RECORD line has more than three elements: {}'.format(row) - ) - # Make a copy because we are mutating the row. - row = list(row) - old_path = row[0] - new_path = installed.pop(old_path, old_path) - row[0] = new_path - if new_path in changed: - digest, length = rehash(new_path) - row[1] = digest - row[2] = length - installed_rows.append(tuple(row)) + logger.warning("RECORD line has more than three elements: %s", row) + old_record_path = cast("RecordPath", row[0]) + new_record_path = installed.pop(old_record_path, old_record_path) + if new_record_path in changed: + digest, length = rehash(_record_to_fs_path(new_record_path, lib_dir)) + else: + digest = row[1] if len(row) > 1 else "" + length = row[2] if len(row) > 2 else "" + installed_rows.append((new_record_path, digest, length)) for f in generated: + path = _fs_to_record_path(f, lib_dir) digest, length = rehash(f) - installed_rows.append((normpath(f, lib_dir), digest, str(length))) - for f in installed: - installed_rows.append((installed[f], '', '')) - return installed_rows + installed_rows.append((path, digest, length)) + return installed_rows + [ + (installed_record_path, "", "") for installed_record_path in installed.values() + ] + + +def get_console_script_specs(console: Dict[str, str]) -> List[str]: + """ + Given the mapping from entrypoint name to callable, return the relevant + console script specs. + """ + # Don't mutate caller's version + console = console.copy() + + scripts_to_generate = [] + + # Special case pip and setuptools to generate versioned wrappers + # + # The issue is that some projects (specifically, pip and setuptools) use + # code in setup.py to create "versioned" entry points - pip2.7 on Python + # 2.7, pip3.3 on Python 3.3, etc. But these entry points are baked into + # the wheel metadata at build time, and so if the wheel is installed with + # a *different* version of Python the entry points will be wrong. The + # correct fix for this is to enhance the metadata to be able to describe + # such versioned entry points. + # Currently, projects using versioned entry points will either have + # incorrect versioned entry points, or they will not be able to distribute + # "universal" wheels (i.e., they will need a wheel per Python version). + # + # Because setuptools and pip are bundled with _ensurepip and virtualenv, + # we need to use universal wheels. As a workaround, we + # override the versioned entry points in the wheel and generate the + # correct ones. + # + # To add the level of hack in this section of code, in order to support + # ensurepip this code will look for an ``ENSUREPIP_OPTIONS`` environment + # variable which will control which version scripts get installed. + # + # ENSUREPIP_OPTIONS=altinstall + # - Only pipX.Y and easy_install-X.Y will be generated and installed + # ENSUREPIP_OPTIONS=install + # - pipX.Y, pipX, easy_install-X.Y will be generated and installed. Note + # that this option is technically if ENSUREPIP_OPTIONS is set and is + # not altinstall + # DEFAULT + # - The default behavior is to install pip, pipX, pipX.Y, easy_install + # and easy_install-X.Y. + pip_script = console.pop("pip", None) + if pip_script: + if "ENSUREPIP_OPTIONS" not in os.environ: + scripts_to_generate.append("pip = " + pip_script) + if os.environ.get("ENSUREPIP_OPTIONS", "") != "altinstall": + scripts_to_generate.append(f"pip{sys.version_info[0]} = {pip_script}") -class MissingCallableSuffix(Exception): - pass + scripts_to_generate.append(f"pip{get_major_minor_version()} = {pip_script}") + # Delete any other versioned pip entry points + pip_ep = [k for k in console if re.match(r"pip(\d+(\.\d+)?)?$", k)] + for k in pip_ep: + del console[k] + easy_install_script = console.pop("easy_install", None) + if easy_install_script: + if "ENSUREPIP_OPTIONS" not in os.environ: + scripts_to_generate.append("easy_install = " + easy_install_script) + + scripts_to_generate.append( + f"easy_install-{get_major_minor_version()} = {easy_install_script}" + ) + # Delete any other versioned easy_install entry points + easy_install_ep = [ + k for k in console if re.match(r"easy_install(-\d+\.\d+)?$", k) + ] + for k in easy_install_ep: + del console[k] + + # Generate the console entry points specified in the wheel + scripts_to_generate.extend(starmap("{} = {}".format, console.items())) + + return scripts_to_generate + + +class ZipBackedFile: + def __init__( + self, src_record_path: RecordPath, dest_path: str, zip_file: ZipFile + ) -> None: + self.src_record_path = src_record_path + self.dest_path = dest_path + self._zip_file = zip_file + self.changed = False + + def _getinfo(self) -> ZipInfo: + return self._zip_file.getinfo(self.src_record_path) + + def save(self) -> None: + # When we open the output file below, any existing file is truncated + # before we start writing the new contents. This is fine in most + # cases, but can cause a segfault if pip has loaded a shared + # object (e.g. from pyopenssl through its vendored urllib3) + # Since the shared object is mmap'd an attempt to call a + # symbol in it will then cause a segfault. Unlinking the file + # allows writing of new contents while allowing the process to + # continue to use the old copy. + if os.path.exists(self.dest_path): + os.unlink(self.dest_path) + + zipinfo = self._getinfo() + + # optimization: the file is created by open(), + # skip the decompression when there is 0 bytes to decompress. + with open(self.dest_path, "wb") as dest: + if zipinfo.file_size > 0: + with self._zip_file.open(zipinfo) as f: + blocksize = min(zipinfo.file_size, 1024 * 1024) + shutil.copyfileobj(f, dest, blocksize) + + if zip_item_is_executable(zipinfo): + set_extracted_file_to_default_mode_plus_executable(self.dest_path) + + +class ScriptFile: + def __init__(self, file: "File") -> None: + self._file = file + self.src_record_path = self._file.src_record_path + self.dest_path = self._file.dest_path + self.changed = False + + def save(self) -> None: + self._file.save() + self.changed = fix_script(self.dest_path) + + +class MissingCallableSuffix(InstallationError): + def __init__(self, entry_point: str) -> None: + super().__init__( + f"Invalid script entry point: {entry_point} - A callable " + "suffix is required. Cf https://packaging.python.org/" + "specifications/entry-points/#use-for-scripts for more " + "information." + ) -def _raise_for_invalid_entrypoint(specification): - # type: (str) -> None +def _raise_for_invalid_entrypoint(specification: str) -> None: entry = get_export_entry(specification) if entry is not None and entry.suffix is None: raise MissingCallableSuffix(str(entry)) class PipScriptMaker(ScriptMaker): - def make(self, specification, options=None): - # type: (str, Dict[str, Any]) -> List[str] + def make( + self, specification: str, options: Optional[Dict[str, Any]] = None + ) -> List[str]: _raise_for_invalid_entrypoint(specification) - return super(PipScriptMaker, self).make(specification, options) - - -def install_unpacked_wheel( - name, # type: str - wheeldir, # type: str - wheel_zip, # type: ZipFile - scheme, # type: Scheme - req_description, # type: str - pycompile=True, # type: bool - warn_script_location=True, # type: bool - direct_url=None, # type: Optional[DirectUrl] -): - # type: (...) -> None + return super().make(specification, options) + + +def _install_wheel( # noqa: C901, PLR0915 function is too long + name: str, + wheel_zip: ZipFile, + wheel_path: str, + scheme: Scheme, + pycompile: bool = True, + warn_script_location: bool = True, + direct_url: Optional[DirectUrl] = None, + requested: bool = False, +) -> None: """Install a wheel. :param name: Name of the project to install - :param wheeldir: Base directory of the unpacked wheel :param wheel_zip: open ZipFile for wheel being installed :param scheme: Distutils scheme dictating the install directories :param req_description: String used in place of the requirement, for @@ -312,12 +444,6 @@ def install_unpacked_wheel( Wheel-Version * when the .dist-info dir does not match the wheel """ - # TODO: Investigate and break this up. - # TODO: Look into moving this into a dedicated class for representing an - # installation. - - source = wheeldir.rstrip(os.path.sep) + os.path.sep - info_dir, metadata = parse_wheel(wheel_zip, name) if wheel_root_is_purelib(metadata): @@ -325,132 +451,179 @@ def install_unpacked_wheel( else: lib_dir = scheme.platlib - subdirs = os.listdir(source) - data_dirs = [s for s in subdirs if s.endswith('.data')] - # Record details of the files moved # installed = files copied from the wheel to the destination # changed = files changed while installing (scripts #! line typically) # generated = files newly generated during the install (script wrappers) - installed = {} # type: Dict[str, str] - changed = set() - generated = [] # type: List[str] + installed: Dict[RecordPath, RecordPath] = {} + changed: Set[RecordPath] = set() + generated: List[str] = [] - # Compile all of the pyc files that we're going to be installing - if pycompile: - with captured_stdout() as stdout: - with warnings.catch_warnings(): - warnings.filterwarnings('ignore') - compileall.compile_dir(source, force=True, quiet=True) - logger.debug(stdout.getvalue()) - - def record_installed(srcfile, destfile, modified=False): - # type: (str, str, bool) -> None + def record_installed( + srcfile: RecordPath, destfile: str, modified: bool = False + ) -> None: """Map archive RECORD paths to installation RECORD paths.""" - oldpath = normpath(srcfile, wheeldir) - newpath = normpath(destfile, lib_dir) - installed[oldpath] = newpath + newpath = _fs_to_record_path(destfile, lib_dir) + installed[srcfile] = newpath if modified: - changed.add(destfile) - - def clobber( - source, # type: str - dest, # type: str - is_base, # type: bool - fixer=None, # type: Optional[Callable[[str], Any]] - filter=None # type: Optional[Callable[[str], bool]] - ): - # type: (...) -> None - ensure_dir(dest) # common for the 'include' path - - for dir, subdirs, files in os.walk(source): - basedir = dir[len(source):].lstrip(os.path.sep) - destdir = os.path.join(dest, basedir) - if is_base and basedir == '': - subdirs[:] = [s for s in subdirs if not s.endswith('.data')] - for f in files: - # Skip unwanted files - if filter and filter(f): - continue - srcfile = os.path.join(dir, f) - destfile = os.path.join(dest, basedir, f) - # directory creation is lazy and after the file filtering above - # to ensure we don't install empty dirs; empty dirs can't be - # uninstalled. - ensure_dir(destdir) - - # copyfile (called below) truncates the destination if it - # exists and then writes the new contents. This is fine in most - # cases, but can cause a segfault if pip has loaded a shared - # object (e.g. from pyopenssl through its vendored urllib3) - # Since the shared object is mmap'd an attempt to call a - # symbol in it will then cause a segfault. Unlinking the file - # allows writing of new contents while allowing the process to - # continue to use the old copy. - if os.path.exists(destfile): - os.unlink(destfile) - - # We use copyfile (not move, copy, or copy2) to be extra sure - # that we are not moving directories over (copyfile fails for - # directories) as well as to ensure that we are not copying - # over any metadata because we want more control over what - # metadata we actually copy over. - shutil.copyfile(srcfile, destfile) - - # Copy over the metadata for the file, currently this only - # includes the atime and mtime. - st = os.stat(srcfile) - if hasattr(os, "utime"): - os.utime(destfile, (st.st_atime, st.st_mtime)) - - # If our file is executable, then make our destination file - # executable. - if os.access(srcfile, os.X_OK): - st = os.stat(srcfile) - permissions = ( - st.st_mode | stat.S_IXUSR | stat.S_IXGRP | stat.S_IXOTH - ) - os.chmod(destfile, permissions) - - changed = False - if fixer: - changed = fixer(destfile) - record_installed(srcfile, destfile, changed) - - clobber(source, lib_dir, True) + changed.add(newpath) - dest_info_dir = os.path.join(lib_dir, info_dir) + def is_dir_path(path: RecordPath) -> bool: + return path.endswith("/") + + def assert_no_path_traversal(dest_dir_path: str, target_path: str) -> None: + if not is_within_directory(dest_dir_path, target_path): + message = ( + "The wheel {!r} has a file {!r} trying to install" + " outside the target directory {!r}" + ) + raise InstallationError( + message.format(wheel_path, target_path, dest_dir_path) + ) + + def root_scheme_file_maker( + zip_file: ZipFile, dest: str + ) -> Callable[[RecordPath], "File"]: + def make_root_scheme_file(record_path: RecordPath) -> "File": + normed_path = os.path.normpath(record_path) + dest_path = os.path.join(dest, normed_path) + assert_no_path_traversal(dest, dest_path) + return ZipBackedFile(record_path, dest_path, zip_file) + + return make_root_scheme_file + + def data_scheme_file_maker( + zip_file: ZipFile, scheme: Scheme + ) -> Callable[[RecordPath], "File"]: + scheme_paths = {key: getattr(scheme, key) for key in SCHEME_KEYS} + + def make_data_scheme_file(record_path: RecordPath) -> "File": + normed_path = os.path.normpath(record_path) + try: + _, scheme_key, dest_subpath = normed_path.split(os.path.sep, 2) + except ValueError: + message = ( + f"Unexpected file in {wheel_path}: {record_path!r}. .data directory" + " contents should be named like: '/'." + ) + raise InstallationError(message) + + try: + scheme_path = scheme_paths[scheme_key] + except KeyError: + valid_scheme_keys = ", ".join(sorted(scheme_paths)) + message = ( + f"Unknown scheme key used in {wheel_path}: {scheme_key} " + f"(for file {record_path!r}). .data directory contents " + f"should be in subdirectories named with a valid scheme " + f"key ({valid_scheme_keys})" + ) + raise InstallationError(message) + + dest_path = os.path.join(scheme_path, dest_subpath) + assert_no_path_traversal(scheme_path, dest_path) + return ZipBackedFile(record_path, dest_path, zip_file) + + return make_data_scheme_file + + def is_data_scheme_path(path: RecordPath) -> bool: + return path.split("/", 1)[0].endswith(".data") + + paths = cast(List[RecordPath], wheel_zip.namelist()) + file_paths = filterfalse(is_dir_path, paths) + root_scheme_paths, data_scheme_paths = partition(is_data_scheme_path, file_paths) + + make_root_scheme_file = root_scheme_file_maker(wheel_zip, lib_dir) + files: Iterator[File] = map(make_root_scheme_file, root_scheme_paths) + + def is_script_scheme_path(path: RecordPath) -> bool: + parts = path.split("/", 2) + return len(parts) > 2 and parts[0].endswith(".data") and parts[1] == "scripts" + + other_scheme_paths, script_scheme_paths = partition( + is_script_scheme_path, data_scheme_paths + ) + + make_data_scheme_file = data_scheme_file_maker(wheel_zip, scheme) + other_scheme_files = map(make_data_scheme_file, other_scheme_paths) + files = chain(files, other_scheme_files) # Get the defined entry points - ep_file = os.path.join(dest_info_dir, 'entry_points.txt') - console, gui = get_entrypoints(ep_file) + distribution = get_wheel_distribution( + FilesystemWheel(wheel_path), + canonicalize_name(name), + ) + console, gui = get_entrypoints(distribution) - def is_entrypoint_wrapper(name): - # type: (str) -> bool + def is_entrypoint_wrapper(file: "File") -> bool: # EP, EP.exe and EP-script.py are scripts generated for # entry point EP by setuptools - if name.lower().endswith('.exe'): + path = file.dest_path + name = os.path.basename(path) + if name.lower().endswith(".exe"): matchname = name[:-4] - elif name.lower().endswith('-script.py'): + elif name.lower().endswith("-script.py"): matchname = name[:-10] elif name.lower().endswith(".pya"): matchname = name[:-4] else: matchname = name # Ignore setuptools-generated scripts - return (matchname in console or matchname in gui) - - for datadir in data_dirs: - fixer = None - filter = None - for subdir in os.listdir(os.path.join(wheeldir, datadir)): - fixer = None - if subdir == 'scripts': - fixer = fix_script - filter = is_entrypoint_wrapper - source = os.path.join(wheeldir, datadir, subdir) - dest = getattr(scheme, subdir) - clobber(source, dest, False, fixer=fixer, filter=filter) + return matchname in console or matchname in gui + + script_scheme_files: Iterator[File] = map( + make_data_scheme_file, script_scheme_paths + ) + script_scheme_files = filterfalse(is_entrypoint_wrapper, script_scheme_files) + script_scheme_files = map(ScriptFile, script_scheme_files) + files = chain(files, script_scheme_files) + + existing_parents = set() + for file in files: + # directory creation is lazy and after file filtering + # to ensure we don't install empty dirs; empty dirs can't be + # uninstalled. + parent_dir = os.path.dirname(file.dest_path) + if parent_dir not in existing_parents: + ensure_dir(parent_dir) + existing_parents.add(parent_dir) + file.save() + record_installed(file.src_record_path, file.dest_path, file.changed) + + def pyc_source_file_paths() -> Generator[str, None, None]: + # We de-duplicate installation paths, since there can be overlap (e.g. + # file in .data maps to same location as file in wheel root). + # Sorting installation paths makes it easier to reproduce and debug + # issues related to permissions on existing files. + for installed_path in sorted(set(installed.values())): + full_installed_path = os.path.join(lib_dir, installed_path) + if not os.path.isfile(full_installed_path): + continue + if not full_installed_path.endswith(".py"): + continue + yield full_installed_path + + def pyc_output_path(path: str) -> str: + """Return the path the pyc file would have been written to.""" + return importlib.util.cache_from_source(path) + + # Compile all of the pyc files for the installed files + if pycompile: + with contextlib.redirect_stdout( + StreamWrapper.from_stream(sys.stdout) + ) as stdout: + with warnings.catch_warnings(): + warnings.filterwarnings("ignore") + for path in pyc_source_file_paths(): + success = compileall.compile_file(path, force=True, quiet=True) + if success: + pyc_path = pyc_output_path(path) + assert os.path.exists(pyc_path) + pyc_record_path = cast( + "RecordPath", pyc_path.replace(os.path.sep, "/") + ) + record_installed(pyc_record_path, pyc_path) + logger.debug(stdout.getvalue()) maker = PipScriptMaker(None, scheme.scripts) @@ -461,106 +634,22 @@ def is_entrypoint_wrapper(name): # Ensure we don't generate any variants for scripts because this is almost # never what somebody wants. # See https://bitbucket.org/pypa/distlib/issue/35/ - maker.variants = {''} + maker.variants = {""} # This is required because otherwise distlib creates scripts that are not # executable. # See https://bitbucket.org/pypa/distlib/issue/32/ maker.set_mode = True - scripts_to_generate = [] - - # Special case pip and setuptools to generate versioned wrappers - # - # The issue is that some projects (specifically, pip and setuptools) use - # code in setup.py to create "versioned" entry points - pip2.7 on Python - # 2.7, pip3.3 on Python 3.3, etc. But these entry points are baked into - # the wheel metadata at build time, and so if the wheel is installed with - # a *different* version of Python the entry points will be wrong. The - # correct fix for this is to enhance the metadata to be able to describe - # such versioned entry points, but that won't happen till Metadata 2.0 is - # available. - # In the meantime, projects using versioned entry points will either have - # incorrect versioned entry points, or they will not be able to distribute - # "universal" wheels (i.e., they will need a wheel per Python version). - # - # Because setuptools and pip are bundled with _ensurepip and virtualenv, - # we need to use universal wheels. So, as a stopgap until Metadata 2.0, we - # override the versioned entry points in the wheel and generate the - # correct ones. This code is purely a short-term measure until Metadata 2.0 - # is available. - # - # To add the level of hack in this section of code, in order to support - # ensurepip this code will look for an ``ENSUREPIP_OPTIONS`` environment - # variable which will control which version scripts get installed. - # - # ENSUREPIP_OPTIONS=altinstall - # - Only pipX.Y and easy_install-X.Y will be generated and installed - # ENSUREPIP_OPTIONS=install - # - pipX.Y, pipX, easy_install-X.Y will be generated and installed. Note - # that this option is technically if ENSUREPIP_OPTIONS is set and is - # not altinstall - # DEFAULT - # - The default behavior is to install pip, pipX, pipX.Y, easy_install - # and easy_install-X.Y. - pip_script = console.pop('pip', None) - if pip_script: - if "ENSUREPIP_OPTIONS" not in os.environ: - scripts_to_generate.append('pip = ' + pip_script) - - if os.environ.get("ENSUREPIP_OPTIONS", "") != "altinstall": - scripts_to_generate.append( - 'pip{} = {}'.format(sys.version_info[0], pip_script) - ) - - scripts_to_generate.append( - 'pip{} = {}'.format(get_major_minor_version(), pip_script) - ) - # Delete any other versioned pip entry points - pip_ep = [k for k in console if re.match(r'pip(\d(\.\d)?)?$', k)] - for k in pip_ep: - del console[k] - easy_install_script = console.pop('easy_install', None) - if easy_install_script: - if "ENSUREPIP_OPTIONS" not in os.environ: - scripts_to_generate.append( - 'easy_install = ' + easy_install_script - ) - - scripts_to_generate.append( - 'easy_install-{} = {}'.format( - get_major_minor_version(), easy_install_script - ) - ) - # Delete any other versioned easy_install entry points - easy_install_ep = [ - k for k in console if re.match(r'easy_install(-\d\.\d)?$', k) - ] - for k in easy_install_ep: - del console[k] - # Generate the console and GUI entry points specified in the wheel - scripts_to_generate.extend(starmap('{} = {}'.format, console.items())) - - gui_scripts_to_generate = list(starmap('{} = {}'.format, gui.items())) + scripts_to_generate = get_console_script_specs(console) - generated_console_scripts = [] # type: List[str] + gui_scripts_to_generate = list(starmap("{} = {}".format, gui.items())) - try: - generated_console_scripts = maker.make_multiple(scripts_to_generate) - generated.extend(generated_console_scripts) + generated_console_scripts = maker.make_multiple(scripts_to_generate) + generated.extend(generated_console_scripts) - generated.extend( - maker.make_multiple(gui_scripts_to_generate, {'gui': True}) - ) - except MissingCallableSuffix as e: - entry = e.args[0] - raise InstallationError( - "Invalid script entry point: {} for req: {} - A callable " - "suffix is required. Cf https://packaging.python.org/" - "specifications/entry-points/#use-for-scripts for more " - "information.".format(entry, req_description) - ) + generated.extend(maker.make_multiple(gui_scripts_to_generate, {"gui": True})) if warn_script_location: msg = message_about_scripts_not_on_PATH(generated_console_scripts) @@ -570,17 +659,18 @@ def is_entrypoint_wrapper(name): generated_file_mode = 0o666 & ~current_umask() @contextlib.contextmanager - def _generate_file(path, **kwargs): - # type: (str, **Any) -> Iterator[NamedTemporaryFileResult] + def _generate_file(path: str, **kwargs: Any) -> Generator[BinaryIO, None, None]: with adjacent_tmp_file(path, **kwargs) as f: yield f os.chmod(f.name, generated_file_mode) replace(f.name, path) + dest_info_dir = os.path.join(lib_dir, info_dir) + # Record pip as the installer - installer_path = os.path.join(dest_info_dir, 'INSTALLER') + installer_path = os.path.join(dest_info_dir, "INSTALLER") with _generate_file(installer_path) as installer_file: - installer_file.write(b'pip\n') + installer_file.write(b"pip\n") generated.append(installer_path) # Record the PEP 610 direct URL reference @@ -590,42 +680,62 @@ def _generate_file(path, **kwargs): direct_url_file.write(direct_url.to_json().encode("utf-8")) generated.append(direct_url_path) + # Record the REQUESTED file + if requested: + requested_path = os.path.join(dest_info_dir, "REQUESTED") + with open(requested_path, "wb"): + pass + generated.append(requested_path) + + record_text = distribution.read_text("RECORD") + record_rows = list(csv.reader(record_text.splitlines())) + + rows = get_csv_rows_for_installed( + record_rows, + installed=installed, + changed=changed, + generated=generated, + lib_dir=lib_dir, + ) + # Record details of all files installed - record_path = os.path.join(dest_info_dir, 'RECORD') - with open(record_path, **csv_io_kwargs('r')) as record_file: - rows = get_csv_rows_for_installed( - csv.reader(record_file), - installed=installed, - changed=changed, - generated=generated, - lib_dir=lib_dir) - with _generate_file(record_path, **csv_io_kwargs('w')) as record_file: - writer = csv.writer(record_file) - writer.writerows(sorted_outrows(rows)) # sort to simplify testing + record_path = os.path.join(dest_info_dir, "RECORD") + + with _generate_file(record_path, **csv_io_kwargs("w")) as record_file: + # Explicitly cast to typing.IO[str] as a workaround for the mypy error: + # "writer" has incompatible type "BinaryIO"; expected "_Writer" + writer = csv.writer(cast("IO[str]", record_file)) + writer.writerows(_normalized_outrows(rows)) + + +@contextlib.contextmanager +def req_error_context(req_description: str) -> Generator[None, None, None]: + try: + yield + except InstallationError as e: + message = f"For req: {req_description}. {e.args[0]}" + raise InstallationError(message) from e def install_wheel( - name, # type: str - wheel_path, # type: str - scheme, # type: Scheme - req_description, # type: str - pycompile=True, # type: bool - warn_script_location=True, # type: bool - _temp_dir_for_testing=None, # type: Optional[str] - direct_url=None, # type: Optional[DirectUrl] -): - # type: (...) -> None - with TempDirectory( - path=_temp_dir_for_testing, kind="unpacked-wheel" - ) as unpacked_dir, ZipFile(wheel_path, allowZip64=True) as z: - unpack_file(wheel_path, unpacked_dir.path) - install_unpacked_wheel( - name=name, - wheeldir=unpacked_dir.path, - wheel_zip=z, - scheme=scheme, - req_description=req_description, - pycompile=pycompile, - warn_script_location=warn_script_location, - direct_url=direct_url, - ) + name: str, + wheel_path: str, + scheme: Scheme, + req_description: str, + pycompile: bool = True, + warn_script_location: bool = True, + direct_url: Optional[DirectUrl] = None, + requested: bool = False, +) -> None: + with ZipFile(wheel_path, allowZip64=True) as z: + with req_error_context(req_description): + _install_wheel( + name=name, + wheel_zip=z, + wheel_path=wheel_path, + scheme=scheme, + pycompile=pycompile, + warn_script_location=warn_script_location, + direct_url=direct_url, + requested=requested, + ) diff --git a/src/fetchcode/vcs/pip/_internal/operations/prepare.py b/src/fetchcode/vcs/pip/_internal/operations/prepare.py index fd48606e..e6aa3447 100644 --- a/src/fetchcode/vcs/pip/_internal/operations/prepare.py +++ b/src/fetchcode/vcs/pip/_internal/operations/prepare.py @@ -4,202 +4,125 @@ # The following comment should be removed at some point in the future. # mypy: strict-optional=False -import logging import mimetypes import os import shutil +from dataclasses import dataclass +from pathlib import Path +from typing import Dict, Iterable, List, Optional -from fetchcode.vcs.pip._vendor import requests -from fetchcode.vcs.pip._vendor.six import PY2 +from pip._vendor.packaging.utils import canonicalize_name -from fetchcode.vcs.pip._internal.distributions import ( - make_distribution_for_install_requirement, -) -from fetchcode.vcs.pip._internal.distributions.installed import InstalledDistribution -from fetchcode.vcs.pip._internal.exceptions import ( +from pip._internal.distributions import make_distribution_for_install_requirement +from pip._internal.distributions.installed import InstalledDistribution +from pip._internal.exceptions import ( DirectoryUrlHashUnsupported, HashMismatch, HashUnpinned, InstallationError, - PreviousBuildDirError, + MetadataInconsistent, + NetworkConnectionError, VcsHashUnsupported, ) -from fetchcode.vcs.pip._internal.utils.filesystem import copy2_fixed -from fetchcode.vcs.pip._internal.utils.hashes import MissingHashes -from fetchcode.vcs.pip._internal.utils.logging import indent_log -from fetchcode.vcs.pip._internal.utils.misc import ( +from pip._internal.index.package_finder import PackageFinder +from pip._internal.metadata import BaseDistribution, get_metadata_distribution +from pip._internal.models.direct_url import ArchiveInfo +from pip._internal.models.link import Link +from pip._internal.models.wheel import Wheel +from pip._internal.network.download import BatchDownloader, Downloader +from pip._internal.network.lazy_wheel import ( + HTTPRangeRequestUnsupported, + dist_from_wheel_url, +) +from pip._internal.network.session import PipSession +from pip._internal.operations.build.build_tracker import BuildTracker +from pip._internal.req.req_install import InstallRequirement +from pip._internal.utils._log import getLogger +from pip._internal.utils.direct_url_helpers import ( + direct_url_for_editable, + direct_url_from_link, +) +from pip._internal.utils.hashes import Hashes, MissingHashes +from pip._internal.utils.logging import indent_log +from pip._internal.utils.misc import ( display_path, + hash_file, hide_url, - path_to_display, - rmtree, + redact_auth_from_requirement, ) -from fetchcode.vcs.pip._internal.utils.temp_dir import TempDirectory -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.utils.unpacking import unpack_file -from fetchcode.vcs.pip._internal.vcs import vcs - -if MYPY_CHECK_RUNNING: - from typing import ( - Callable, List, Optional, Tuple, - ) - - from mypy_extensions import TypedDict - - from fetchcode.vcs.pip._internal.distributions import AbstractDistribution - from fetchcode.vcs.pip._internal.index.package_finder import PackageFinder - from fetchcode.vcs.pip._internal.models.link import Link - from fetchcode.vcs.pip._internal.network.download import Downloader - from fetchcode.vcs.pip._internal.req.req_install import InstallRequirement - from fetchcode.vcs.pip._internal.req.req_tracker import RequirementTracker - from fetchcode.vcs.pip._internal.utils.hashes import Hashes - - if PY2: - CopytreeKwargs = TypedDict( - 'CopytreeKwargs', - { - 'ignore': Callable[[str, List[str]], List[str]], - 'symlinks': bool, - }, - total=False, - ) - else: - CopytreeKwargs = TypedDict( - 'CopytreeKwargs', - { - 'copy_function': Callable[[str, str], None], - 'ignore': Callable[[str, List[str]], List[str]], - 'ignore_dangling_symlinks': bool, - 'symlinks': bool, - }, - total=False, - ) +from pip._internal.utils.temp_dir import TempDirectory +from pip._internal.utils.unpacking import unpack_file +from pip._internal.vcs import vcs -logger = logging.getLogger(__name__) +logger = getLogger(__name__) def _get_prepared_distribution( - req, # type: InstallRequirement - req_tracker, # type: RequirementTracker - finder, # type: PackageFinder - build_isolation # type: bool -): - # type: (...) -> AbstractDistribution - """Prepare a distribution for installation. - """ + req: InstallRequirement, + build_tracker: BuildTracker, + finder: PackageFinder, + build_isolation: bool, + check_build_deps: bool, +) -> BaseDistribution: + """Prepare a distribution for installation.""" abstract_dist = make_distribution_for_install_requirement(req) - with req_tracker.track(req): - abstract_dist.prepare_distribution_metadata(finder, build_isolation) - return abstract_dist + tracker_id = abstract_dist.build_tracker_id + if tracker_id is not None: + with build_tracker.track(req, tracker_id): + abstract_dist.prepare_distribution_metadata( + finder, build_isolation, check_build_deps + ) + return abstract_dist.get_metadata_distribution() -def unpack_vcs_link(link, location): - # type: (Link, str) -> None +def unpack_vcs_link(link: Link, location: str, verbosity: int) -> None: vcs_backend = vcs.get_backend_for_scheme(link.scheme) assert vcs_backend is not None - vcs_backend.unpack(location, url=hide_url(link.url)) + vcs_backend.unpack(location, url=hide_url(link.url), verbosity=verbosity) + +@dataclass +class File: + path: str + content_type: Optional[str] = None -class File(object): - def __init__(self, path, content_type): - # type: (str, str) -> None - self.path = path - self.content_type = content_type + def __post_init__(self) -> None: + if self.content_type is None: + self.content_type = mimetypes.guess_type(self.path)[0] def get_http_url( - link, # type: Link - downloader, # type: Downloader - download_dir=None, # type: Optional[str] - hashes=None, # type: Optional[Hashes] -): - # type: (...) -> File + link: Link, + download: Downloader, + download_dir: Optional[str] = None, + hashes: Optional[Hashes] = None, +) -> File: temp_dir = TempDirectory(kind="unpack", globally_managed=True) # If a download dir is specified, is the file already downloaded there? already_downloaded_path = None if download_dir: - already_downloaded_path = _check_download_dir( - link, download_dir, hashes - ) + already_downloaded_path = _check_download_dir(link, download_dir, hashes) if already_downloaded_path: from_path = already_downloaded_path - content_type = mimetypes.guess_type(from_path)[0] + content_type = None else: # let's download to a tmp dir - from_path, content_type = _download_http_url( - link, downloader, temp_dir.path, hashes - ) + from_path, content_type = download(link, temp_dir.path) + if hashes: + hashes.check_against_path(from_path) return File(from_path, content_type) -def _copy2_ignoring_special_files(src, dest): - # type: (str, str) -> None - """Copying special files is not supported, but as a convenience to users - we skip errors copying them. This supports tools that may create e.g. - socket files in the project source directory. - """ - try: - copy2_fixed(src, dest) - except shutil.SpecialFileError as e: - # SpecialFileError may be raised due to either the source or - # destination. If the destination was the cause then we would actually - # care, but since the destination directory is deleted prior to - # copy we ignore all of them assuming it is caused by the source. - logger.warning( - "Ignoring special file error '%s' encountered copying %s to %s.", - str(e), - path_to_display(src), - path_to_display(dest), - ) - - -def _copy_source_tree(source, target): - # type: (str, str) -> None - target_abspath = os.path.abspath(target) - target_basename = os.path.basename(target_abspath) - target_dirname = os.path.dirname(target_abspath) - - def ignore(d, names): - # type: (str, List[str]) -> List[str] - skipped = [] # type: List[str] - if d == source: - # Pulling in those directories can potentially be very slow, - # exclude the following directories if they appear in the top - # level dir (and only it). - # See discussion at https://github.com/pypa/pip/pull/6770 - skipped += ['.tox', '.nox'] - if os.path.abspath(d) == target_dirname: - # Prevent an infinite recursion if the target is in source. - # This can happen when TMPDIR is set to ${PWD}/... - # and we copy PWD to TMPDIR. - skipped += [target_basename] - return skipped - - kwargs = dict(ignore=ignore, symlinks=True) # type: CopytreeKwargs - - if not PY2: - # Python 2 does not support copy_function, so we only ignore - # errors on special file copy in Python 3. - kwargs['copy_function'] = _copy2_ignoring_special_files - - shutil.copytree(source, target, **kwargs) - - def get_file_url( - link, # type: Link - download_dir=None, # type: Optional[str] - hashes=None # type: Optional[Hashes] -): - # type: (...) -> File - """Get file and optionally check its hash. - """ + link: Link, download_dir: Optional[str] = None, hashes: Optional[Hashes] = None +) -> File: + """Get file and optionally check its hash.""" # If a download dir is specified, is the file already there and valid? already_downloaded_path = None if download_dir: - already_downloaded_path = _check_download_dir( - link, download_dir, hashes - ) + already_downloaded_path = _check_download_dir(link, download_dir, hashes) if already_downloaded_path: from_path = already_downloaded_path @@ -213,20 +136,17 @@ def get_file_url( # one; no internet-sourced hash will be in `hashes`. if hashes: hashes.check_against_path(from_path) - - content_type = mimetypes.guess_type(from_path)[0] - - return File(from_path, content_type) + return File(from_path, None) def unpack_url( - link, # type: Link - location, # type: str - downloader, # type: Downloader - download_dir=None, # type: Optional[str] - hashes=None, # type: Optional[Hashes] -): - # type: (...) -> Optional[File] + link: Link, + location: str, + download: Downloader, + verbosity: int, + download_dir: Optional[str] = None, + hashes: Optional[Hashes] = None, +) -> Optional[File]: """Unpack link into location, downloading if required. :param hashes: A Hashes object, one of whose embedded hashes must match, @@ -236,15 +156,10 @@ def unpack_url( """ # non-editable vcs urls if link.is_vcs: - unpack_vcs_link(link, location) + unpack_vcs_link(link, location, verbosity=verbosity) return None - # If it's a url to a local directory - if link.is_existing_dir(): - if os.path.isdir(location): - rmtree(location) - _copy_source_tree(link.file_path, location) - return None + assert not link.is_existing_dir() # file urls if link.is_file: @@ -254,43 +169,27 @@ def unpack_url( else: file = get_http_url( link, - downloader, + download, download_dir, hashes=hashes, ) # unpack the archive to the build dir location. even when only downloading - # archives, they have to be unpacked to parse dependencies - unpack_file(file.path, location, file.content_type) + # archives, they have to be unpacked to parse dependencies, except wheels + if not link.is_wheel: + unpack_file(file.path, location, file.content_type) return file -def _download_http_url( - link, # type: Link - downloader, # type: Downloader - temp_dir, # type: str - hashes, # type: Optional[Hashes] -): - # type: (...) -> Tuple[str, str] - """Download link url into temp_dir using provided session""" - download = downloader(link) - - file_path = os.path.join(temp_dir, download.filename) - with open(file_path, 'wb') as content_file: - for chunk in download.chunks: - content_file.write(chunk) - - if hashes: - hashes.check_against_path(file_path) - - return file_path, download.response.headers.get('content-type', '') - - -def _check_download_dir(link, download_dir, hashes): - # type: (Link, str, Optional[Hashes]) -> Optional[str] - """ Check download_dir for previously downloaded file with correct hash - If a correct file is found return its path else None +def _check_download_dir( + link: Link, + download_dir: str, + hashes: Optional[Hashes], + warn_on_hash_mismatch: bool = True, +) -> Optional[str]: + """Check download_dir for previously downloaded file with correct hash + If a correct file is found return its path else None """ download_path = os.path.join(download_dir, link.filename) @@ -298,271 +197,536 @@ def _check_download_dir(link, download_dir, hashes): return None # If already downloaded, does its hash match? - logger.info('File was already downloaded %s', download_path) + logger.info("File was already downloaded %s", download_path) if hashes: try: hashes.check_against_path(download_path) except HashMismatch: - logger.warning( - 'Previously-downloaded file %s has bad hash. ' - 'Re-downloading.', - download_path - ) + if warn_on_hash_mismatch: + logger.warning( + "Previously-downloaded file %s has bad hash. Re-downloading.", + download_path, + ) os.unlink(download_path) return None return download_path -class RequirementPreparer(object): - """Prepares a Requirement - """ +class RequirementPreparer: + """Prepares a Requirement""" def __init__( self, - build_dir, # type: str - download_dir, # type: Optional[str] - src_dir, # type: str - wheel_download_dir, # type: Optional[str] - build_isolation, # type: bool - req_tracker, # type: RequirementTracker - downloader, # type: Downloader - finder, # type: PackageFinder - require_hashes, # type: bool - use_user_site, # type: bool - ): - # type: (...) -> None - super(RequirementPreparer, self).__init__() + build_dir: str, + download_dir: Optional[str], + src_dir: str, + build_isolation: bool, + check_build_deps: bool, + build_tracker: BuildTracker, + session: PipSession, + progress_bar: str, + finder: PackageFinder, + require_hashes: bool, + use_user_site: bool, + lazy_wheel: bool, + verbosity: int, + legacy_resolver: bool, + ) -> None: + super().__init__() self.src_dir = src_dir self.build_dir = build_dir - self.req_tracker = req_tracker - self.downloader = downloader + self.build_tracker = build_tracker + self._session = session + self._download = Downloader(session, progress_bar) + self._batch_download = BatchDownloader(session, progress_bar) self.finder = finder # Where still-packed archives should be written to. If None, they are # not saved, and are deleted immediately after unpacking. self.download_dir = download_dir - # Where still-packed .whl files should be written to. If None, they are - # written to the download_dir parameter. Separate to download_dir to - # permit only keeping wheel archives for pip wheel. - self.wheel_download_dir = wheel_download_dir - - # NOTE - # download_dir and wheel_download_dir overlap semantically and may - # be combined if we're willing to have non-wheel archives present in - # the wheelhouse output by 'pip wheel'. - # Is build isolation allowed? self.build_isolation = build_isolation + # Should check build dependencies? + self.check_build_deps = check_build_deps + # Should hash-checking be required? self.require_hashes = require_hashes # Should install in user site-packages? self.use_user_site = use_user_site - @property - def _download_should_save(self): - # type: () -> bool - if not self.download_dir: - return False + # Should wheels be downloaded lazily? + self.use_lazy_wheel = lazy_wheel - if os.path.exists(self.download_dir): - return True + # How verbose should underlying tooling be? + self.verbosity = verbosity - logger.critical('Could not find download directory') - raise InstallationError( - "Could not find or access download directory '{}'" - .format(self.download_dir)) + # Are we using the legacy resolver? + self.legacy_resolver = legacy_resolver - def prepare_linked_requirement( - self, - req, # type: InstallRequirement - ): - # type: (...) -> AbstractDistribution - """Prepare a requirement that would be obtained from req.link - """ - assert req.link - link = req.link + # Memoized downloaded files, as mapping of url: path. + self._downloaded: Dict[str, str] = {} + + # Previous "header" printed for a link-based InstallRequirement + self._previous_requirement_header = ("", "") - # TODO: Breakup into smaller functions - if link.scheme == 'file': - path = link.file_path - logger.info('Processing %s', display_path(path)) + def _log_preparing_link(self, req: InstallRequirement) -> None: + """Provide context for the requirement being prepared.""" + if req.link.is_file and not req.is_wheel_from_cache: + message = "Processing %s" + information = str(display_path(req.link.file_path)) else: - logger.info('Collecting %s', req.req or req) - - download_dir = self.download_dir - if link.is_wheel and self.wheel_download_dir: - # when doing 'pip wheel` we download wheels to a - # dedicated dir. - download_dir = self.wheel_download_dir - - if link.is_wheel: - if download_dir: - # When downloading, we only unpack wheels to get - # metadata. - autodelete_unpacked = True + message = "Collecting %s" + information = redact_auth_from_requirement(req.req) if req.req else str(req) + + # If we used req.req, inject requirement source if available (this + # would already be included if we used req directly) + if req.req and req.comes_from: + if isinstance(req.comes_from, str): + comes_from: Optional[str] = req.comes_from else: - # When installing a wheel, we use the unpacked - # wheel. - autodelete_unpacked = False - else: - # We always delete unpacked sdists after pip runs. - autodelete_unpacked = True + comes_from = req.comes_from.from_path() + if comes_from: + information += f" (from {comes_from})" + + if (message, information) != self._previous_requirement_header: + self._previous_requirement_header = (message, information) + logger.info(message, information) + + if req.is_wheel_from_cache: + with indent_log(): + logger.info("Using cached %s", req.link.filename) + + def _ensure_link_req_src_dir( + self, req: InstallRequirement, parallel_builds: bool + ) -> None: + """Ensure source_dir of a linked InstallRequirement.""" + # Since source_dir is only set for editable requirements. + if req.link.is_wheel: + # We don't need to unpack wheels, so no need for a source + # directory. + return + assert req.source_dir is None + if req.link.is_existing_dir(): + # build local directories in-tree + req.source_dir = req.link.file_path + return + + # We always delete unpacked sdists after pip runs. + req.ensure_has_source_dir( + self.build_dir, + autodelete=True, + parallel_builds=parallel_builds, + ) + req.ensure_pristine_source_checkout() + + def _get_linked_req_hashes(self, req: InstallRequirement) -> Hashes: + # By the time this is called, the requirement's link should have + # been checked so we can tell what kind of requirements req is + # and raise some more informative errors than otherwise. + # (For example, we can raise VcsHashUnsupported for a VCS URL + # rather than HashMissing.) + if not self.require_hashes: + return req.hashes(trust_internet=True) + + # We could check these first 2 conditions inside unpack_url + # and save repetition of conditions, but then we would + # report less-useful error messages for unhashable + # requirements, complaining that there's no hash provided. + if req.link.is_vcs: + raise VcsHashUnsupported() + if req.link.is_existing_dir(): + raise DirectoryUrlHashUnsupported() + + # Unpinned packages are asking for trouble when a new version + # is uploaded. This isn't a security check, but it saves users + # a surprising hash mismatch in the future. + # file:/// URLs aren't pinnable, so don't complain about them + # not being pinned. + if not req.is_direct and not req.is_pinned: + raise HashUnpinned() + + # If known-good hashes are missing for this requirement, + # shim it with a facade object that will provoke hash + # computation and then raise a HashMissing exception + # showing the user what the hash should be. + return req.hashes(trust_internet=False) or MissingHashes() + + def _fetch_metadata_only( + self, + req: InstallRequirement, + ) -> Optional[BaseDistribution]: + if self.legacy_resolver: + logger.debug( + "Metadata-only fetching is not used in the legacy resolver", + ) + return None + if self.require_hashes: + logger.debug( + "Metadata-only fetching is not used as hash checking is required", + ) + return None + # Try PEP 658 metadata first, then fall back to lazy wheel if unavailable. + return self._fetch_metadata_using_link_data_attr( + req + ) or self._fetch_metadata_using_lazy_wheel(req.link) + + def _fetch_metadata_using_link_data_attr( + self, + req: InstallRequirement, + ) -> Optional[BaseDistribution]: + """Fetch metadata from the data-dist-info-metadata attribute, if possible.""" + # (1) Get the link to the metadata file, if provided by the backend. + metadata_link = req.link.metadata_link() + if metadata_link is None: + return None + assert req.req is not None + logger.verbose( + "Obtaining dependency information for %s from %s", + req.req, + metadata_link, + ) + # (2) Download the contents of the METADATA file, separate from the dist itself. + metadata_file = get_http_url( + metadata_link, + self._download, + hashes=metadata_link.as_hashes(), + ) + with open(metadata_file.path, "rb") as f: + metadata_contents = f.read() + # (3) Generate a dist just from those file contents. + metadata_dist = get_metadata_distribution( + metadata_contents, + req.link.filename, + req.req.name, + ) + # (4) Ensure the Name: field from the METADATA file matches the name from the + # install requirement. + # + # NB: raw_name will fall back to the name from the install requirement if + # the Name: field is not present, but it's noted in the raw_name docstring + # that that should NEVER happen anyway. + if canonicalize_name(metadata_dist.raw_name) != canonicalize_name(req.req.name): + raise MetadataInconsistent( + req, "Name", req.req.name, metadata_dist.raw_name + ) + return metadata_dist + + def _fetch_metadata_using_lazy_wheel( + self, + link: Link, + ) -> Optional[BaseDistribution]: + """Fetch metadata using lazy wheel, if possible.""" + # --use-feature=fast-deps must be provided. + if not self.use_lazy_wheel: + return None + if link.is_file or not link.is_wheel: + logger.debug( + "Lazy wheel is not used as %r does not point to a remote wheel", + link, + ) + return None + + wheel = Wheel(link.filename) + name = canonicalize_name(wheel.name) + logger.info( + "Obtaining dependency information from %s %s", + name, + wheel.version, + ) + url = link.url.split("#", 1)[0] + try: + return dist_from_wheel_url(name, url, self._session) + except HTTPRangeRequestUnsupported: + logger.debug("%s does not support range requests", url) + return None + + def _complete_partial_requirements( + self, + partially_downloaded_reqs: Iterable[InstallRequirement], + parallel_builds: bool = False, + ) -> None: + """Download any requirements which were only fetched by metadata.""" + # Download to a temporary directory. These will be copied over as + # needed for downstream 'download', 'wheel', and 'install' commands. + temp_dir = TempDirectory(kind="unpack", globally_managed=True).path + + # Map each link to the requirement that owns it. This allows us to set + # `req.local_file_path` on the appropriate requirement after passing + # all the links at once into BatchDownloader. + links_to_fully_download: Dict[Link, InstallRequirement] = {} + for req in partially_downloaded_reqs: + assert req.link + links_to_fully_download[req.link] = req + + batch_download = self._batch_download( + links_to_fully_download.keys(), + temp_dir, + ) + for link, (filepath, _) in batch_download: + logger.debug("Downloading link %s to %s", link, filepath) + req = links_to_fully_download[link] + # Record the downloaded file path so wheel reqs can extract a Distribution + # in .get_dist(). + req.local_file_path = filepath + # Record that the file is downloaded so we don't do it again in + # _prepare_linked_requirement(). + self._downloaded[req.link.url] = filepath + + # If this is an sdist, we need to unpack it after downloading, but the + # .source_dir won't be set up until we are in _prepare_linked_requirement(). + # Add the downloaded archive to the install requirement to unpack after + # preparing the source dir. + if not req.is_wheel: + req.needs_unpacked_archive(Path(filepath)) + + # This step is necessary to ensure all lazy wheels are processed + # successfully by the 'download', 'wheel', and 'install' commands. + for req in partially_downloaded_reqs: + self._prepare_linked_requirement(req, parallel_builds) + def prepare_linked_requirement( + self, req: InstallRequirement, parallel_builds: bool = False + ) -> BaseDistribution: + """Prepare a requirement to be obtained from req.link.""" + assert req.link + self._log_preparing_link(req) with indent_log(): - # Since source_dir is only set for editable requirements. - assert req.source_dir is None - req.ensure_has_source_dir(self.build_dir, autodelete_unpacked) - # If a checkout exists, it's unwise to keep going. version - # inconsistencies are logged later, but do not fail the - # installation. - # FIXME: this won't upgrade when there's an existing - # package unpacked in `req.source_dir` - if os.path.exists(os.path.join(req.source_dir, 'setup.py')): - raise PreviousBuildDirError( - "pip can't proceed with requirements '{}' due to a" - " pre-existing build directory ({}). This is " - "likely due to a previous installation that failed" - ". pip is being responsible and not assuming it " - "can delete this. Please delete it and try again." - .format(req, req.source_dir) + # Check if the relevant file is already available + # in the download directory + file_path = None + if self.download_dir is not None and req.link.is_wheel: + hashes = self._get_linked_req_hashes(req) + file_path = _check_download_dir( + req.link, + self.download_dir, + hashes, + # When a locally built wheel has been found in cache, we don't warn + # about re-downloading when the already downloaded wheel hash does + # not match. This is because the hash must be checked against the + # original link, not the cached link. It that case the already + # downloaded file will be removed and re-fetched from cache (which + # implies a hash check against the cache entry's origin.json). + warn_on_hash_mismatch=not req.is_wheel_from_cache, ) - # Now that we have the real link, we can tell what kind of - # requirements we have and raise some more informative errors - # than otherwise. (For example, we can raise VcsHashUnsupported - # for a VCS URL rather than HashMissing.) - if self.require_hashes: - # We could check these first 2 conditions inside - # unpack_url and save repetition of conditions, but then - # we would report less-useful error messages for - # unhashable requirements, complaining that there's no - # hash provided. - if link.is_vcs: - raise VcsHashUnsupported() - elif link.is_existing_dir(): - raise DirectoryUrlHashUnsupported() - if not req.original_link and not req.is_pinned: - # Unpinned packages are asking for trouble when a new - # version is uploaded. This isn't a security check, but - # it saves users a surprising hash mismatch in the - # future. - # - # file:/// URLs aren't pinnable, so don't complain - # about them not being pinned. - raise HashUnpinned() - - hashes = req.hashes(trust_internet=not self.require_hashes) - if self.require_hashes and not hashes: - # Known-good hashes are missing for this requirement, so - # shim it with a facade object that will provoke hash - # computation and then raise a HashMissing exception - # showing the user what the hash should be. - hashes = MissingHashes() + if file_path is not None: + # The file is already available, so mark it as downloaded + self._downloaded[req.link.url] = file_path + else: + # The file is not available, attempt to fetch only metadata + metadata_dist = self._fetch_metadata_only(req) + if metadata_dist is not None: + req.needs_more_preparation = True + return metadata_dist + + # None of the optimizations worked, fully prepare the requirement + return self._prepare_linked_requirement(req, parallel_builds) + + def prepare_linked_requirements_more( + self, reqs: Iterable[InstallRequirement], parallel_builds: bool = False + ) -> None: + """Prepare linked requirements more, if needed.""" + reqs = [req for req in reqs if req.needs_more_preparation] + for req in reqs: + # Determine if any of these requirements were already downloaded. + if self.download_dir is not None and req.link.is_wheel: + hashes = self._get_linked_req_hashes(req) + file_path = _check_download_dir(req.link, self.download_dir, hashes) + if file_path is not None: + self._downloaded[req.link.url] = file_path + req.needs_more_preparation = False + + # Prepare requirements we found were already downloaded for some + # reason. The other downloads will be completed separately. + partially_downloaded_reqs: List[InstallRequirement] = [] + for req in reqs: + if req.needs_more_preparation: + partially_downloaded_reqs.append(req) + else: + self._prepare_linked_requirement(req, parallel_builds) + + # TODO: separate this part out from RequirementPreparer when the v1 + # resolver can be removed! + self._complete_partial_requirements( + partially_downloaded_reqs, + parallel_builds=parallel_builds, + ) + + def _prepare_linked_requirement( + self, req: InstallRequirement, parallel_builds: bool + ) -> BaseDistribution: + assert req.link + link = req.link + hashes = self._get_linked_req_hashes(req) + + if hashes and req.is_wheel_from_cache: + assert req.download_info is not None + assert link.is_wheel + assert link.is_file + # We need to verify hashes, and we have found the requirement in the cache + # of locally built wheels. + if ( + isinstance(req.download_info.info, ArchiveInfo) + and req.download_info.info.hashes + and hashes.has_one_of(req.download_info.info.hashes) + ): + # At this point we know the requirement was built from a hashable source + # artifact, and we verified that the cache entry's hash of the original + # artifact matches one of the hashes we expect. We don't verify hashes + # against the cached wheel, because the wheel is not the original. + hashes = None + else: + logger.warning( + "The hashes of the source archive found in cache entry " + "don't match, ignoring cached built wheel " + "and re-downloading source." + ) + req.link = req.cached_wheel_source_link + link = req.link + + self._ensure_link_req_src_dir(req, parallel_builds) + + if link.is_existing_dir(): + local_file = None + elif link.url not in self._downloaded: try: local_file = unpack_url( - link, req.source_dir, self.downloader, download_dir, - hashes=hashes, - ) - except requests.HTTPError as exc: - logger.critical( - 'Could not install requirement %s because of error %s', - req, - exc, + link, + req.source_dir, + self._download, + self.verbosity, + self.download_dir, + hashes, ) + except NetworkConnectionError as exc: raise InstallationError( - 'Could not install requirement {} because of HTTP ' - 'error {} for URL {}'.format(req, exc, link) + f"Could not install requirement {req} because of HTTP " + f"error {exc} for URL {link}" ) + else: + file_path = self._downloaded[link.url] + if hashes: + hashes.check_against_path(file_path) + local_file = File(file_path, content_type=None) + + # If download_info is set, we got it from the wheel cache. + if req.download_info is None: + # Editables don't go through this function (see + # prepare_editable_requirement). + assert not req.editable + req.download_info = direct_url_from_link(link, req.source_dir) + # Make sure we have a hash in download_info. If we got it as part of the + # URL, it will have been verified and we can rely on it. Otherwise we + # compute it from the downloaded file. + # FIXME: https://github.com/pypa/pip/issues/11943 + if ( + isinstance(req.download_info.info, ArchiveInfo) + and not req.download_info.info.hashes + and local_file + ): + hash = hash_file(local_file.path)[0].hexdigest() + # We populate info.hash for backward compatibility. + # This will automatically populate info.hashes. + req.download_info.info.hash = f"sha256={hash}" + + # For use in later processing, + # preserve the file path on the requirement. + if local_file: + req.local_file_path = local_file.path + + dist = _get_prepared_distribution( + req, + self.build_tracker, + self.finder, + self.build_isolation, + self.check_build_deps, + ) + return dist - # For use in later processing, preserve the file path on the - # requirement. - if local_file: - req.local_file_path = local_file.path - - abstract_dist = _get_prepared_distribution( - req, self.req_tracker, self.finder, self.build_isolation, + def save_linked_requirement(self, req: InstallRequirement) -> None: + assert self.download_dir is not None + assert req.link is not None + link = req.link + if link.is_vcs or (link.is_existing_dir() and req.editable): + # Make a .zip of the source_dir we already created. + req.archive(self.download_dir) + return + + if link.is_existing_dir(): + logger.debug( + "Not copying link to destination directory " + "since it is a directory: %s", + link, ) + return + if req.local_file_path is None: + # No distribution was downloaded for this requirement. + return - if download_dir: - if link.is_existing_dir(): - logger.info('Link is a directory, ignoring download_dir') - elif local_file: - download_location = os.path.join( - download_dir, link.filename - ) - if not os.path.exists(download_location): - shutil.copy(local_file.path, download_location) - logger.info( - 'Saved %s', display_path(download_location) - ) - - if self._download_should_save: - # Make a .zip of the source_dir we already created. - if link.is_vcs: - req.archive(self.download_dir) - return abstract_dist + download_location = os.path.join(self.download_dir, link.filename) + if not os.path.exists(download_location): + shutil.copy(req.local_file_path, download_location) + download_path = display_path(download_location) + logger.info("Saved %s", download_path) def prepare_editable_requirement( self, - req, # type: InstallRequirement - ): - # type: (...) -> AbstractDistribution - """Prepare an editable requirement - """ + req: InstallRequirement, + ) -> BaseDistribution: + """Prepare an editable requirement.""" assert req.editable, "cannot prepare a non-editable req as editable" - logger.info('Obtaining %s', req) + logger.info("Obtaining %s", req) with indent_log(): if self.require_hashes: raise InstallationError( - 'The editable requirement {} cannot be installed when ' - 'requiring hashes, because there is no single file to ' - 'hash.'.format(req) + f"The editable requirement {req} cannot be installed when " + "requiring hashes, because there is no single file to " + "hash." ) req.ensure_has_source_dir(self.src_dir) - req.update_editable(not self._download_should_save) - - abstract_dist = _get_prepared_distribution( - req, self.req_tracker, self.finder, self.build_isolation, + req.update_editable() + assert req.source_dir + req.download_info = direct_url_for_editable(req.unpacked_source_directory) + + dist = _get_prepared_distribution( + req, + self.build_tracker, + self.finder, + self.build_isolation, + self.check_build_deps, ) - if self._download_should_save: - req.archive(self.download_dir) req.check_if_exists(self.use_user_site) - return abstract_dist + return dist def prepare_installed_requirement( self, - req, # type: InstallRequirement - skip_reason # type: str - ): - # type: (...) -> AbstractDistribution - """Prepare an already-installed requirement - """ + req: InstallRequirement, + skip_reason: str, + ) -> BaseDistribution: + """Prepare an already-installed requirement.""" assert req.satisfied_by, "req should have been satisfied but isn't" assert skip_reason is not None, ( "did not get skip reason skipped but req.satisfied_by " - "is set to {}".format(req.satisfied_by) + f"is set to {req.satisfied_by}" ) logger.info( - 'Requirement %s: %s (%s)', - skip_reason, req, req.satisfied_by.version + "Requirement %s: %s (%s)", skip_reason, req, req.satisfied_by.version ) with indent_log(): if self.require_hashes: logger.debug( - 'Since it is already installed, we are trusting this ' - 'package without checking its hash. To ensure a ' - 'completely repeatable environment, install into an ' - 'empty virtualenv.' + "Since it is already installed, we are trusting this " + "package without checking its hash. To ensure a " + "completely repeatable environment, install into an " + "empty virtualenv." ) - abstract_dist = InstalledDistribution(req) - - return abstract_dist + return InstalledDistribution(req).get_metadata_distribution() diff --git a/src/fetchcode/vcs/pip/_internal/pyproject.py b/src/fetchcode/vcs/pip/_internal/pyproject.py index bf76be2d..2a9cad48 100644 --- a/src/fetchcode/vcs/pip/_internal/pyproject.py +++ b/src/fetchcode/vcs/pip/_internal/pyproject.py @@ -1,51 +1,40 @@ -from __future__ import absolute_import - -import io +import importlib.util import os import sys from collections import namedtuple +from typing import Any, List, Optional -from fetchcode.vcs.pip._vendor import six, toml -from fetchcode.vcs.pip._vendor.packaging.requirements import InvalidRequirement, Requirement - -from fetchcode.vcs.pip._internal.exceptions import InstallationError -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import Any, Optional, List +if sys.version_info >= (3, 11): + import tomllib +else: + from pip._vendor import tomli as tomllib +from pip._vendor.packaging.requirements import InvalidRequirement -def _is_list_of_str(obj): - # type: (Any) -> bool - return ( - isinstance(obj, list) and - all(isinstance(item, six.string_types) for item in obj) - ) +from pip._internal.exceptions import ( + InstallationError, + InvalidPyProjectBuildRequires, + MissingPyProjectBuildRequires, +) +from pip._internal.utils.packaging import get_requirement -def make_pyproject_path(unpacked_source_directory): - # type: (str) -> str - path = os.path.join(unpacked_source_directory, 'pyproject.toml') +def _is_list_of_str(obj: Any) -> bool: + return isinstance(obj, list) and all(isinstance(item, str) for item in obj) - # Python2 __file__ should not be unicode - if six.PY2 and isinstance(path, six.text_type): - path = path.encode(sys.getfilesystemencoding()) - return path +def make_pyproject_path(unpacked_source_directory: str) -> str: + return os.path.join(unpacked_source_directory, "pyproject.toml") -BuildSystemDetails = namedtuple('BuildSystemDetails', [ - 'requires', 'backend', 'check', 'backend_path' -]) +BuildSystemDetails = namedtuple( + "BuildSystemDetails", ["requires", "backend", "check", "backend_path"] +) def load_pyproject_toml( - use_pep517, # type: Optional[bool] - pyproject_toml, # type: str - setup_py, # type: str - req_name # type: str -): - # type: (...) -> Optional[BuildSystemDetails] + use_pep517: Optional[bool], pyproject_toml: str, setup_py: str, req_name: str +) -> Optional[BuildSystemDetails]: """Load the pyproject.toml file. Parameters: @@ -70,9 +59,15 @@ def load_pyproject_toml( has_pyproject = os.path.isfile(pyproject_toml) has_setup = os.path.isfile(setup_py) + if not has_pyproject and not has_setup: + raise InstallationError( + f"{req_name} does not appear to be a Python project: " + f"neither 'setup.py' nor 'pyproject.toml' found." + ) + if has_pyproject: - with io.open(pyproject_toml, encoding="utf-8") as f: - pp_toml = toml.load(f) + with open(pyproject_toml, encoding="utf-8") as f: + pp_toml = tomllib.loads(f.read()) build_system = pp_toml.get("build-system") else: build_system = None @@ -95,17 +90,26 @@ def load_pyproject_toml( raise InstallationError( "Disabling PEP 517 processing is invalid: " "project specifies a build backend of {} " - "in pyproject.toml".format( - build_system["build-backend"] - ) + "in pyproject.toml".format(build_system["build-backend"]) ) use_pep517 = True # If we haven't worked out whether to use PEP 517 yet, # and the user hasn't explicitly stated a preference, - # we do so if the project has a pyproject.toml file. + # we do so if the project has a pyproject.toml file + # or if we cannot import setuptools or wheels. + + # We fallback to PEP 517 when without setuptools or without the wheel package, + # so setuptools can be installed as a default build backend. + # For more info see: + # https://discuss.python.org/t/pip-without-setuptools-could-the-experience-be-improved/11810/9 + # https://github.com/pypa/pip/issues/8559 elif use_pep517 is None: - use_pep517 = has_pyproject + use_pep517 = ( + has_pyproject + or not importlib.util.find_spec("setuptools") + or not importlib.util.find_spec("wheel") + ) # At this point, we know whether we're going to use PEP 517. assert use_pep517 is not None @@ -125,7 +129,7 @@ def load_pyproject_toml( # a version of setuptools that supports that backend. build_system = { - "requires": ["setuptools>=40.8.0", "wheel"], + "requires": ["setuptools>=40.8.0"], "build-backend": "setuptools.build_meta:__legacy__", } @@ -137,52 +141,37 @@ def load_pyproject_toml( # Ensure that the build-system section in pyproject.toml conforms # to PEP 518. - error_template = ( - "{package} has a pyproject.toml file that does not comply " - "with PEP 518: {reason}" - ) # Specifying the build-system table but not the requires key is invalid if "requires" not in build_system: - raise InstallationError( - error_template.format(package=req_name, reason=( - "it has a 'build-system' table but not " - "'build-system.requires' which is mandatory in the table" - )) - ) + raise MissingPyProjectBuildRequires(package=req_name) # Error out if requires is not a list of strings requires = build_system["requires"] if not _is_list_of_str(requires): - raise InstallationError(error_template.format( + raise InvalidPyProjectBuildRequires( package=req_name, - reason="'build-system.requires' is not a list of strings.", - )) + reason="It is not a list of strings.", + ) # Each requirement must be valid as per PEP 508 for requirement in requires: try: - Requirement(requirement) - except InvalidRequirement: - raise InstallationError( - error_template.format( - package=req_name, - reason=( - "'build-system.requires' contains an invalid " - "requirement: {!r}".format(requirement) - ), - ) - ) + get_requirement(requirement) + except InvalidRequirement as error: + raise InvalidPyProjectBuildRequires( + package=req_name, + reason=f"It contains an invalid requirement: {requirement!r}", + ) from error backend = build_system.get("build-backend") backend_path = build_system.get("backend-path", []) - check = [] # type: List[str] + check: List[str] = [] if backend is None: # If the user didn't specify a backend, we assume they want to use # the setuptools backend. But we can't be sure they have included - # a version of setuptools which supplies the backend, or wheel - # (which is needed by the backend) in their requirements. So we - # make a note to check that those requirements are present once + # a version of setuptools which supplies the backend. So we + # make a note to check that this requirement is present once # we have set up the environment. # This is quite a lot of work to check for a very specific case. But # the problem is, that case is potentially quite common - projects that @@ -191,6 +180,6 @@ def load_pyproject_toml( # tools themselves. The original PEP 518 code had a similar check (but # implemented in a different way). backend = "setuptools.build_meta:__legacy__" - check = ["setuptools>=40.8.0", "wheel"] + check = ["setuptools>=40.8.0"] return BuildSystemDetails(requires, backend, check, backend_path) diff --git a/src/fetchcode/vcs/pip/_internal/req/__init__.py b/src/fetchcode/vcs/pip/_internal/req/__init__.py index ceb8f927..422d851d 100644 --- a/src/fetchcode/vcs/pip/_internal/req/__init__.py +++ b/src/fetchcode/vcs/pip/_internal/req/__init__.py @@ -1,92 +1,90 @@ -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False - -from __future__ import absolute_import - +import collections import logging +from dataclasses import dataclass +from typing import Generator, List, Optional, Sequence, Tuple -from fetchcode.vcs.pip._internal.utils.logging import indent_log -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._internal.utils.logging import indent_log from .req_file import parse_requirements from .req_install import InstallRequirement from .req_set import RequirementSet -if MYPY_CHECK_RUNNING: - from typing import Any, List, Sequence - __all__ = [ - "RequirementSet", "InstallRequirement", - "parse_requirements", "install_given_reqs", + "RequirementSet", + "InstallRequirement", + "parse_requirements", + "install_given_reqs", ] logger = logging.getLogger(__name__) -class InstallationResult(object): - def __init__(self, name): - # type: (str) -> None - self.name = name +@dataclass(frozen=True) +class InstallationResult: + name: str + - def __repr__(self): - # type: () -> str - return "InstallationResult(name={!r})".format(self.name) +def _validate_requirements( + requirements: List[InstallRequirement], +) -> Generator[Tuple[str, InstallRequirement], None, None]: + for req in requirements: + assert req.name, f"invalid to-be-installed requirement: {req}" + yield req.name, req def install_given_reqs( - to_install, # type: List[InstallRequirement] - install_options, # type: List[str] - global_options=(), # type: Sequence[str] - *args, # type: Any - **kwargs # type: Any -): - # type: (...) -> List[InstallationResult] + requirements: List[InstallRequirement], + global_options: Sequence[str], + root: Optional[str], + home: Optional[str], + prefix: Optional[str], + warn_script_location: bool, + use_user_site: bool, + pycompile: bool, +) -> List[InstallationResult]: """ Install everything in the given list. (to be called after having downloaded and unpacked the packages) """ + to_install = collections.OrderedDict(_validate_requirements(requirements)) if to_install: logger.info( - 'Installing collected packages: %s', - ', '.join([req.name for req in to_install]), + "Installing collected packages: %s", + ", ".join(to_install.keys()), ) installed = [] with indent_log(): - for requirement in to_install: + for req_name, requirement in to_install.items(): if requirement.should_reinstall: - logger.info('Attempting uninstall: %s', requirement.name) + logger.info("Attempting uninstall: %s", req_name) with indent_log(): - uninstalled_pathset = requirement.uninstall( - auto_confirm=True - ) + uninstalled_pathset = requirement.uninstall(auto_confirm=True) + else: + uninstalled_pathset = None + try: requirement.install( - install_options, global_options, - *args, - **kwargs + root=root, + home=home, + prefix=prefix, + warn_script_location=warn_script_location, + use_user_site=use_user_site, + pycompile=pycompile, ) except Exception: - should_rollback = ( - requirement.should_reinstall and - not requirement.install_succeeded - ) # if install did not succeed, rollback previous uninstall - if should_rollback: + if uninstalled_pathset and not requirement.install_succeeded: uninstalled_pathset.rollback() raise else: - should_commit = ( - requirement.should_reinstall and - requirement.install_succeeded - ) - if should_commit: + if uninstalled_pathset and requirement.install_succeeded: uninstalled_pathset.commit() - installed.append(InstallationResult(requirement.name)) + installed.append(InstallationResult(req_name)) return installed diff --git a/src/fetchcode/vcs/pip/_internal/req/constructors.py b/src/fetchcode/vcs/pip/_internal/req/constructors.py index 644e3c0a..d73236e0 100644 --- a/src/fetchcode/vcs/pip/_internal/req/constructors.py +++ b/src/fetchcode/vcs/pip/_internal/req/constructors.py @@ -8,58 +8,41 @@ InstallRequirement. """ -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False - +import copy import logging import os import re - -from fetchcode.vcs.pip._vendor.packaging.markers import Marker -from fetchcode.vcs.pip._vendor.packaging.requirements import InvalidRequirement, Requirement -from fetchcode.vcs.pip._vendor.packaging.specifiers import Specifier -from fetchcode.vcs.pip._vendor.pkg_resources import RequirementParseError, parse_requirements - -from fetchcode.vcs.pip._internal.exceptions import InstallationError -from fetchcode.vcs.pip._internal.models.index import PyPI, TestPyPI -from fetchcode.vcs.pip._internal.models.link import Link -from fetchcode.vcs.pip._internal.models.wheel import Wheel -from fetchcode.vcs.pip._internal.pyproject import make_pyproject_path -from fetchcode.vcs.pip._internal.req.req_install import InstallRequirement -from fetchcode.vcs.pip._internal.utils.filetypes import ARCHIVE_EXTENSIONS -from fetchcode.vcs.pip._internal.utils.misc import is_installable_dir, splitext -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.utils.urls import path_to_url -from fetchcode.vcs.pip._internal.vcs import is_url, vcs - -if MYPY_CHECK_RUNNING: - from typing import ( - Any, Dict, Optional, Set, Tuple, Union, - ) - from fetchcode.vcs.pip._internal.req.req_file import ParsedRequirement - +from dataclasses import dataclass +from typing import Collection, Dict, List, Optional, Set, Tuple, Union + +from pip._vendor.packaging.markers import Marker +from pip._vendor.packaging.requirements import InvalidRequirement, Requirement +from pip._vendor.packaging.specifiers import Specifier + +from pip._internal.exceptions import InstallationError +from pip._internal.models.index import PyPI, TestPyPI +from pip._internal.models.link import Link +from pip._internal.models.wheel import Wheel +from pip._internal.req.req_file import ParsedRequirement +from pip._internal.req.req_install import InstallRequirement +from pip._internal.utils.filetypes import is_archive_file +from pip._internal.utils.misc import is_installable_dir +from pip._internal.utils.packaging import get_requirement +from pip._internal.utils.urls import path_to_url +from pip._internal.vcs import is_url, vcs __all__ = [ - "install_req_from_editable", "install_req_from_line", - "parse_editable" + "install_req_from_editable", + "install_req_from_line", + "parse_editable", ] logger = logging.getLogger(__name__) operators = Specifier._operators.keys() -def is_archive_file(name): - # type: (str) -> bool - """Return True if `name` is a considered as an archive file.""" - ext = splitext(name)[1].lower() - if ext in ARCHIVE_EXTENSIONS: - return True - return False - - -def _strip_extras(path): - # type: (str) -> Tuple[str, Optional[str]] - m = re.match(r'^(.+)(\[[^\]]+\])$', path) +def _strip_extras(path: str) -> Tuple[str, Optional[str]]: + m = re.match(r"^(.+)(\[[^\]]+\])$", path) extras = None if m: path_no_extras = m.group(1) @@ -70,15 +53,38 @@ def _strip_extras(path): return path_no_extras, extras -def convert_extras(extras): - # type: (Optional[str]) -> Set[str] +def convert_extras(extras: Optional[str]) -> Set[str]: if not extras: return set() - return Requirement("placeholder" + extras.lower()).extras + return get_requirement("placeholder" + extras.lower()).extras -def parse_editable(editable_req): - # type: (str) -> Tuple[Optional[str], str, Optional[Set[str]]] +def _set_requirement_extras(req: Requirement, new_extras: Set[str]) -> Requirement: + """ + Returns a new requirement based on the given one, with the supplied extras. If the + given requirement already has extras those are replaced (or dropped if no new extras + are given). + """ + match: Optional[re.Match[str]] = re.fullmatch( + # see https://peps.python.org/pep-0508/#complete-grammar + r"([\w\t .-]+)(\[[^\]]*\])?(.*)", + str(req), + flags=re.ASCII, + ) + # ireq.req is a valid requirement so the regex should always match + assert ( + match is not None + ), f"regex match on requirement {req} failed, this should never happen" + pre: Optional[str] = match.group(1) + post: Optional[str] = match.group(3) + assert ( + pre is not None and post is not None + ), f"regex group selection for requirement {req} failed, this should never happen" + extras: str = "[%s]" % ",".join(sorted(new_extras)) if new_extras else "" + return get_requirement(f"{pre}{extras}{post}") + + +def parse_editable(editable_req: str) -> Tuple[Optional[str], str, Set[str]]: """Parses an editable requirement into: - a requirement name - an URL @@ -95,116 +101,113 @@ def parse_editable(editable_req): url_no_extras, extras = _strip_extras(url) if os.path.isdir(url_no_extras): - if not os.path.exists(os.path.join(url_no_extras, 'setup.py')): - msg = ( - 'File "setup.py" not found. Directory cannot be installed ' - 'in editable mode: {}'.format(os.path.abspath(url_no_extras)) - ) - pyproject_path = make_pyproject_path(url_no_extras) - if os.path.isfile(pyproject_path): - msg += ( - '\n(A "pyproject.toml" file was found, but editable ' - 'mode currently requires a setup.py based build.)' - ) - raise InstallationError(msg) - # Treating it as code that has already been checked out url_no_extras = path_to_url(url_no_extras) - if url_no_extras.lower().startswith('file:'): + if url_no_extras.lower().startswith("file:"): package_name = Link(url_no_extras).egg_fragment if extras: return ( package_name, url_no_extras, - Requirement("placeholder" + extras.lower()).extras, + get_requirement("placeholder" + extras.lower()).extras, ) else: - return package_name, url_no_extras, None + return package_name, url_no_extras, set() for version_control in vcs: - if url.lower().startswith('{}:'.format(version_control)): - url = '{}+{}'.format(version_control, url) + if url.lower().startswith(f"{version_control}:"): + url = f"{version_control}+{url}" break - if '+' not in url: + link = Link(url) + + if not link.is_vcs: + backends = ", ".join(vcs.all_schemes) raise InstallationError( - '{} is not a valid editable requirement. ' - 'It should either be a path to a local project or a VCS URL ' - '(beginning with svn+, git+, hg+, or bzr+).'.format(editable_req) + f"{editable_req} is not a valid editable requirement. " + f"It should either be a path to a local project or a VCS URL " + f"(beginning with {backends})." ) - vc_type = url.split('+', 1)[0].lower() - - if not vcs.get_backend(vc_type): - backends = ", ".join([bends.name + '+URL' for bends in vcs.backends]) - error_message = "For --editable={}, " \ - "only {} are currently supported".format( - editable_req, backends) - raise InstallationError(error_message) - - package_name = Link(url).egg_fragment + package_name = link.egg_fragment if not package_name: raise InstallationError( - "Could not detect requirement name for '{}', please specify one " - "with #egg=your_package_name".format(editable_req) + f"Could not detect requirement name for '{editable_req}', " + "please specify one with #egg=your_package_name" + ) + return package_name, url, set() + + +def check_first_requirement_in_file(filename: str) -> None: + """Check if file is parsable as a requirements file. + + This is heavily based on ``pkg_resources.parse_requirements``, but + simplified to just check the first meaningful line. + + :raises InvalidRequirement: If the first meaningful line cannot be parsed + as an requirement. + """ + with open(filename, encoding="utf-8", errors="ignore") as f: + # Create a steppable iterator, so we can handle \-continuations. + lines = ( + line + for line in (line.strip() for line in f) + if line and not line.startswith("#") # Skip blank lines/comments. ) - return package_name, url, None + for line in lines: + # Drop comments -- a hash without a space may be in a URL. + if " #" in line: + line = line[: line.find(" #")] + # If there is a line continuation, drop it, and append the next line. + if line.endswith("\\"): + line = line[:-2].strip() + next(lines, "") + get_requirement(line) + return -def deduce_helpful_msg(req): - # type: (str) -> str + +def deduce_helpful_msg(req: str) -> str: """Returns helpful msg in case requirements file does not exist, or cannot be parsed. :params req: Requirements file path """ - msg = "" - if os.path.exists(req): - msg = " It does exist." - # Try to parse and check if it is a requirements file. - try: - with open(req, 'r') as fp: - # parse first line only - next(parse_requirements(fp.read())) - msg += ( - "The argument you provided " - "({}) appears to be a" - " requirements file. If that is the" - " case, use the '-r' flag to install" - " the packages specified within it." - ).format(req) - except RequirementParseError: - logger.debug("Cannot parse '{}' as requirements \ - file".format(req), exc_info=True) + if not os.path.exists(req): + return f" File '{req}' does not exist." + msg = " The path does exist. " + # Try to parse and check if it is a requirements file. + try: + check_first_requirement_in_file(req) + except InvalidRequirement: + logger.debug("Cannot parse '%s' as requirements file", req) else: - msg += " File '{}' does not exist.".format(req) + msg += ( + f"The argument you provided " + f"({req}) appears to be a" + f" requirements file. If that is the" + f" case, use the '-r' flag to install" + f" the packages specified within it." + ) return msg -class RequirementParts(object): - def __init__( - self, - requirement, # type: Optional[Requirement] - link, # type: Optional[Link] - markers, # type: Optional[Marker] - extras, # type: Set[str] - ): - self.requirement = requirement - self.link = link - self.markers = markers - self.extras = extras +@dataclass(frozen=True) +class RequirementParts: + requirement: Optional[Requirement] + link: Optional[Link] + markers: Optional[Marker] + extras: Set[str] -def parse_req_from_editable(editable_req): - # type: (str) -> RequirementParts +def parse_req_from_editable(editable_req: str) -> RequirementParts: name, url, extras_override = parse_editable(editable_req) if name is not None: try: - req = Requirement(name) - except InvalidRequirement: - raise InstallationError("Invalid requirement: '{}'".format(name)) + req: Optional[Requirement] = get_requirement(name) + except InvalidRequirement as exc: + raise InstallationError(f"Invalid requirement: {name!r}: {exc}") else: req = None @@ -217,34 +220,38 @@ def parse_req_from_editable(editable_req): def install_req_from_editable( - editable_req, # type: str - comes_from=None, # type: Optional[Union[InstallRequirement, str]] - use_pep517=None, # type: Optional[bool] - isolated=False, # type: bool - options=None, # type: Optional[Dict[str, Any]] - constraint=False # type: bool -): - # type: (...) -> InstallRequirement - + editable_req: str, + comes_from: Optional[Union[InstallRequirement, str]] = None, + *, + use_pep517: Optional[bool] = None, + isolated: bool = False, + global_options: Optional[List[str]] = None, + hash_options: Optional[Dict[str, List[str]]] = None, + constraint: bool = False, + user_supplied: bool = False, + permit_editable_wheels: bool = False, + config_settings: Optional[Dict[str, Union[str, List[str]]]] = None, +) -> InstallRequirement: parts = parse_req_from_editable(editable_req) return InstallRequirement( parts.requirement, comes_from=comes_from, + user_supplied=user_supplied, editable=True, + permit_editable_wheels=permit_editable_wheels, link=parts.link, constraint=constraint, use_pep517=use_pep517, isolated=isolated, - install_options=options.get("install_options", []) if options else [], - global_options=options.get("global_options", []) if options else [], - hash_options=options.get("hashes", {}) if options else {}, + global_options=global_options, + hash_options=hash_options, + config_settings=config_settings, extras=parts.extras, ) -def _looks_like_path(name): - # type: (str) -> bool +def _looks_like_path(name: str) -> bool: """Checks whether the string "looks like" a path on the filesystem. This does not check whether the target actually exists, only judge from the @@ -263,11 +270,10 @@ def _looks_like_path(name): return False -def _get_url_from_path(path, name): - # type: (str, str) -> str +def _get_url_from_path(path: str, name: str) -> Optional[str]: """ - First, it checks whether a provided path is an installable directory - (e.g. it has a setup.py). If it is, returns the path. + First, it checks whether a provided path is an installable directory. If it + is, returns the path. If false, check if the path is an archive file (such as a .whl). The function checks if the path is a file. If false, if the path has @@ -276,33 +282,33 @@ def _get_url_from_path(path, name): if _looks_like_path(name) and os.path.isdir(path): if is_installable_dir(path): return path_to_url(path) + # TODO: The is_installable_dir test here might not be necessary + # now that it is done in load_pyproject_toml too. raise InstallationError( - "Directory {name!r} is not installable. Neither 'setup.py' " - "nor 'pyproject.toml' found.".format(**locals()) + f"Directory {name!r} is not installable. Neither 'setup.py' " + "nor 'pyproject.toml' found." ) if not is_archive_file(path): return None if os.path.isfile(path): return path_to_url(path) - urlreq_parts = name.split('@', 1) + urlreq_parts = name.split("@", 1) if len(urlreq_parts) >= 2 and not _looks_like_path(urlreq_parts[0]): # If the path contains '@' and the part before it does not look # like a path, try to treat it as a PEP 440 URL req instead. return None logger.warning( - 'Requirement %r looks like a filename, but the ' - 'file does not exist', - name + "Requirement %r looks like a filename, but the file does not exist", + name, ) return path_to_url(path) -def parse_req_from_line(name, line_source): - # type: (str, Optional[str]) -> RequirementParts +def parse_req_from_line(name: str, line_source: Optional[str]) -> RequirementParts: if is_url(name): - marker_sep = '; ' + marker_sep = "; " else: - marker_sep = ';' + marker_sep = ";" if marker_sep in name: name, markers_as_string = name.split(marker_sep, 1) markers_as_string = markers_as_string.strip() @@ -329,13 +335,12 @@ def parse_req_from_line(name, line_source): # it's a local file, dir, or url if link: # Handle relative file URLs - if link.scheme == 'file' and re.search(r'\.\./', link.url): - link = Link( - path_to_url(os.path.normpath(os.path.abspath(link.path)))) + if link.scheme == "file" and re.search(r"\.\./", link.url): + link = Link(path_to_url(os.path.normpath(os.path.abspath(link.path)))) # wheel file if link.is_wheel: wheel = Wheel(link.filename) # can raise InvalidWheelFilename - req_as_string = "{wheel.name}=={wheel.version}".format(**locals()) + req_as_string = f"{wheel.name}=={wheel.version}" else: # set the req to the egg fragment. when it's not there, this # will become an 'unnamed' requirement @@ -347,30 +352,31 @@ def parse_req_from_line(name, line_source): extras = convert_extras(extras_as_string) - def with_source(text): - # type: (str) -> str + def with_source(text: str) -> str: if not line_source: return text - return '{} (from {})'.format(text, line_source) + return f"{text} (from {line_source})" - if req_as_string is not None: + def _parse_req_string(req_as_string: str) -> Requirement: try: - req = Requirement(req_as_string) - except InvalidRequirement: + return get_requirement(req_as_string) + except InvalidRequirement as exc: if os.path.sep in req_as_string: add_msg = "It looks like a path." add_msg += deduce_helpful_msg(req_as_string) - elif ('=' in req_as_string and - not any(op in req_as_string for op in operators)): + elif "=" in req_as_string and not any( + op in req_as_string for op in operators + ): add_msg = "= is not a valid operator. Did you mean == ?" else: - add_msg = '' - msg = with_source( - 'Invalid requirement: {!r}'.format(req_as_string) - ) + add_msg = "" + msg = with_source(f"Invalid requirement: {req_as_string!r}: {exc}") if add_msg: - msg += '\nHint: {}'.format(add_msg) + msg += f"\nHint: {add_msg}" raise InstallationError(msg) + + if req_as_string is not None: + req: Optional[Requirement] = _parse_req_string(req_as_string) else: req = None @@ -378,15 +384,18 @@ def with_source(text): def install_req_from_line( - name, # type: str - comes_from=None, # type: Optional[Union[str, InstallRequirement]] - use_pep517=None, # type: Optional[bool] - isolated=False, # type: bool - options=None, # type: Optional[Dict[str, Any]] - constraint=False, # type: bool - line_source=None, # type: Optional[str] -): - # type: (...) -> InstallRequirement + name: str, + comes_from: Optional[Union[str, InstallRequirement]] = None, + *, + use_pep517: Optional[bool] = None, + isolated: bool = False, + global_options: Optional[List[str]] = None, + hash_options: Optional[Dict[str, List[str]]] = None, + constraint: bool = False, + line_source: Optional[str] = None, + user_supplied: bool = False, + config_settings: Optional[Dict[str, Union[str, List[str]]]] = None, +) -> InstallRequirement: """Creates an InstallRequirement from a name, which might be a requirement, directory containing 'setup.py', filename, or URL. @@ -396,52 +405,66 @@ def install_req_from_line( parts = parse_req_from_line(name, line_source) return InstallRequirement( - parts.requirement, comes_from, link=parts.link, markers=parts.markers, - use_pep517=use_pep517, isolated=isolated, - install_options=options.get("install_options", []) if options else [], - global_options=options.get("global_options", []) if options else [], - hash_options=options.get("hashes", {}) if options else {}, + parts.requirement, + comes_from, + link=parts.link, + markers=parts.markers, + use_pep517=use_pep517, + isolated=isolated, + global_options=global_options, + hash_options=hash_options, + config_settings=config_settings, constraint=constraint, extras=parts.extras, + user_supplied=user_supplied, ) def install_req_from_req_string( - req_string, # type: str - comes_from=None, # type: Optional[InstallRequirement] - isolated=False, # type: bool - use_pep517=None # type: Optional[bool] -): - # type: (...) -> InstallRequirement + req_string: str, + comes_from: Optional[InstallRequirement] = None, + isolated: bool = False, + use_pep517: Optional[bool] = None, + user_supplied: bool = False, +) -> InstallRequirement: try: - req = Requirement(req_string) - except InvalidRequirement: - raise InstallationError("Invalid requirement: '{}'".format(req_string)) + req = get_requirement(req_string) + except InvalidRequirement as exc: + raise InstallationError(f"Invalid requirement: {req_string!r}: {exc}") domains_not_allowed = [ PyPI.file_storage_domain, TestPyPI.file_storage_domain, ] - if (req.url and comes_from and comes_from.link and - comes_from.link.netloc in domains_not_allowed): + if ( + req.url + and comes_from + and comes_from.link + and comes_from.link.netloc in domains_not_allowed + ): # Explicitly disallow pypi packages that depend on external urls raise InstallationError( "Packages installed from PyPI cannot depend on packages " "which are not also hosted on PyPI.\n" - "{} depends on {} ".format(comes_from.name, req) + f"{comes_from.name} depends on {req} " ) return InstallRequirement( - req, comes_from, isolated=isolated, use_pep517=use_pep517 + req, + comes_from, + isolated=isolated, + use_pep517=use_pep517, + user_supplied=user_supplied, ) def install_req_from_parsed_requirement( - parsed_req, # type: ParsedRequirement - isolated=False, # type: bool - use_pep517=None # type: Optional[bool] -): - # type: (...) -> InstallRequirement + parsed_req: ParsedRequirement, + isolated: bool = False, + use_pep517: Optional[bool] = None, + user_supplied: bool = False, + config_settings: Optional[Dict[str, Union[str, List[str]]]] = None, +) -> InstallRequirement: if parsed_req.is_editable: req = install_req_from_editable( parsed_req.requirement, @@ -449,6 +472,8 @@ def install_req_from_parsed_requirement( use_pep517=use_pep517, constraint=parsed_req.constraint, isolated=isolated, + user_supplied=user_supplied, + config_settings=config_settings, ) else: @@ -457,8 +482,79 @@ def install_req_from_parsed_requirement( comes_from=parsed_req.comes_from, use_pep517=use_pep517, isolated=isolated, - options=parsed_req.options, + global_options=( + parsed_req.options.get("global_options", []) + if parsed_req.options + else [] + ), + hash_options=( + parsed_req.options.get("hashes", {}) if parsed_req.options else {} + ), constraint=parsed_req.constraint, line_source=parsed_req.line_source, + user_supplied=user_supplied, + config_settings=config_settings, ) return req + + +def install_req_from_link_and_ireq( + link: Link, ireq: InstallRequirement +) -> InstallRequirement: + return InstallRequirement( + req=ireq.req, + comes_from=ireq.comes_from, + editable=ireq.editable, + link=link, + markers=ireq.markers, + use_pep517=ireq.use_pep517, + isolated=ireq.isolated, + global_options=ireq.global_options, + hash_options=ireq.hash_options, + config_settings=ireq.config_settings, + user_supplied=ireq.user_supplied, + ) + + +def install_req_drop_extras(ireq: InstallRequirement) -> InstallRequirement: + """ + Creates a new InstallationRequirement using the given template but without + any extras. Sets the original requirement as the new one's parent + (comes_from). + """ + return InstallRequirement( + req=( + _set_requirement_extras(ireq.req, set()) if ireq.req is not None else None + ), + comes_from=ireq, + editable=ireq.editable, + link=ireq.link, + markers=ireq.markers, + use_pep517=ireq.use_pep517, + isolated=ireq.isolated, + global_options=ireq.global_options, + hash_options=ireq.hash_options, + constraint=ireq.constraint, + extras=[], + config_settings=ireq.config_settings, + user_supplied=ireq.user_supplied, + permit_editable_wheels=ireq.permit_editable_wheels, + ) + + +def install_req_extend_extras( + ireq: InstallRequirement, + extras: Collection[str], +) -> InstallRequirement: + """ + Returns a copy of an installation requirement with some additional extras. + Makes a shallow copy of the ireq object. + """ + result = copy.copy(ireq) + result.extras = {*ireq.extras, *extras} + result.req = ( + _set_requirement_extras(ireq.req, result.extras) + if ireq.req is not None + else None + ) + return result diff --git a/src/fetchcode/vcs/pip/_internal/req/req_file.py b/src/fetchcode/vcs/pip/_internal/req/req_file.py index 2fafcff3..53ad8674 100644 --- a/src/fetchcode/vcs/pip/_internal/req/req_file.py +++ b/src/fetchcode/vcs/pip/_internal/req/req_file.py @@ -2,55 +2,51 @@ Requirements file parsing """ -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False - -from __future__ import absolute_import - +import logging import optparse import os import re import shlex -import sys - -from fetchcode.vcs.pip._vendor.six.moves.urllib import parse as urllib_parse - -from fetchcode.vcs.pip._internal.cli import cmdoptions -from fetchcode.vcs.pip._internal.exceptions import ( - InstallationError, - RequirementsFileParseError, +import urllib.parse +from optparse import Values +from typing import ( + TYPE_CHECKING, + Any, + Callable, + Dict, + Generator, + Iterable, + List, + NoReturn, + Optional, + Tuple, ) -from fetchcode.vcs.pip._internal.models.search_scope import SearchScope -from fetchcode.vcs.pip._internal.utils.encoding import auto_decode -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.utils.urls import get_url_scheme - -if MYPY_CHECK_RUNNING: - from optparse import Values - from typing import ( - Any, Callable, Dict, Iterator, List, NoReturn, Optional, Text, Tuple, - ) - from fetchcode.vcs.pip._internal.index.package_finder import PackageFinder - from fetchcode.vcs.pip._internal.network.session import PipSession +from pip._internal.cli import cmdoptions +from pip._internal.exceptions import InstallationError, RequirementsFileParseError +from pip._internal.models.search_scope import SearchScope +from pip._internal.utils.encoding import auto_decode - ReqFileLines = Iterator[Tuple[int, Text]] +if TYPE_CHECKING: + from pip._internal.index.package_finder import PackageFinder + from pip._internal.network.session import PipSession - LineParser = Callable[[Text], Tuple[str, Values]] +__all__ = ["parse_requirements"] +ReqFileLines = Iterable[Tuple[int, str]] -__all__ = ['parse_requirements'] +LineParser = Callable[[str], Tuple[str, Values]] -SCHEME_RE = re.compile(r'^(http|https|file):', re.I) -COMMENT_RE = re.compile(r'(^|\s+)#.*$') +SCHEME_RE = re.compile(r"^(http|https|file):", re.I) +COMMENT_RE = re.compile(r"(^|\s+)#.*$") # Matches environment variable-style values in '${MY_VARIABLE_1}' with the # variable name consisting of only uppercase letters, digits or the '_' # (underscore). This follows the POSIX standard defined in IEEE Std 1003.1, # 2013 Edition. -ENV_VAR_RE = re.compile(r'(?P\$\{(?P[A-Z0-9_]+)\})') +ENV_VAR_RE = re.compile(r"(?P\$\{(?P[A-Z0-9_]+)\})") -SUPPORTED_OPTIONS = [ +SUPPORTED_OPTIONS: List[Callable[..., optparse.Option]] = [ cmdoptions.index_url, cmdoptions.extra_index_url, cmdoptions.no_index, @@ -60,34 +56,44 @@ cmdoptions.find_links, cmdoptions.no_binary, cmdoptions.only_binary, + cmdoptions.prefer_binary, cmdoptions.require_hashes, cmdoptions.pre, cmdoptions.trusted_host, - cmdoptions.always_unzip, # Deprecated -] # type: List[Callable[..., optparse.Option]] + cmdoptions.use_new_feature, +] # options to be passed to requirements -SUPPORTED_OPTIONS_REQ = [ - cmdoptions.install_options, +SUPPORTED_OPTIONS_REQ: List[Callable[..., optparse.Option]] = [ cmdoptions.global_options, cmdoptions.hash, -] # type: List[Callable[..., optparse.Option]] + cmdoptions.config_settings, +] + +SUPPORTED_OPTIONS_EDITABLE_REQ: List[Callable[..., optparse.Option]] = [ + cmdoptions.config_settings, +] + # the 'dest' string values SUPPORTED_OPTIONS_REQ_DEST = [str(o().dest) for o in SUPPORTED_OPTIONS_REQ] +SUPPORTED_OPTIONS_EDITABLE_REQ_DEST = [ + str(o().dest) for o in SUPPORTED_OPTIONS_EDITABLE_REQ +] + +logger = logging.getLogger(__name__) -class ParsedRequirement(object): +class ParsedRequirement: def __init__( self, - requirement, # type:str - is_editable, # type: bool - comes_from, # type: str - constraint, # type: bool - options=None, # type: Optional[Dict[str, Any]] - line_source=None, # type: Optional[str] - ): - # type: (...) -> None + requirement: str, + is_editable: bool, + comes_from: str, + constraint: bool, + options: Optional[Dict[str, Any]] = None, + line_source: Optional[str] = None, + ) -> None: self.requirement = requirement self.is_editable = is_editable self.comes_from = comes_from @@ -96,20 +102,17 @@ def __init__( self.line_source = line_source -class ParsedLine(object): +class ParsedLine: def __init__( self, - filename, # type: str - lineno, # type: int - comes_from, # type: str - args, # type: str - opts, # type: Values - constraint, # type: bool - ): - # type: (...) -> None + filename: str, + lineno: int, + args: str, + opts: Values, + constraint: bool, + ) -> None: self.filename = filename self.lineno = lineno - self.comes_from = comes_from self.opts = opts self.constraint = constraint @@ -127,45 +130,38 @@ def __init__( def parse_requirements( - filename, # type: str - session, # type: PipSession - finder=None, # type: Optional[PackageFinder] - comes_from=None, # type: Optional[str] - options=None, # type: Optional[optparse.Values] - constraint=False, # type: bool -): - # type: (...) -> Iterator[ParsedRequirement] - """Parse a requirements file and yield InstallRequirement instances. + filename: str, + session: "PipSession", + finder: Optional["PackageFinder"] = None, + options: Optional[optparse.Values] = None, + constraint: bool = False, +) -> Generator[ParsedRequirement, None, None]: + """Parse a requirements file and yield ParsedRequirement instances. :param filename: Path or url of requirements file. :param session: PipSession instance. :param finder: Instance of pip.index.PackageFinder. - :param comes_from: Origin description of requirements. :param options: cli options. :param constraint: If true, parsing a constraint file rather than requirements file. """ line_parser = get_line_parser(finder) - parser = RequirementsFileParser(session, line_parser, comes_from) + parser = RequirementsFileParser(session, line_parser) for parsed_line in parser.parse(filename, constraint): parsed_req = handle_line( - parsed_line, - options=options, - finder=finder, - session=session + parsed_line, options=options, finder=finder, session=session ) if parsed_req is not None: yield parsed_req -def preprocess(content): - # type: (Text) -> ReqFileLines +def preprocess(content: str) -> ReqFileLines: """Split, filter, and join lines, and return a line iterator :param content: the content of the requirements file """ - lines_enum = enumerate(content.splitlines(), start=1) # type: ReqFileLines + lines_enum: ReqFileLines = enumerate(content.splitlines(), start=1) lines_enum = join_lines(lines_enum) lines_enum = ignore_comments(lines_enum) lines_enum = expand_env_variables(lines_enum) @@ -173,72 +169,74 @@ def preprocess(content): def handle_requirement_line( - line, # type: ParsedLine - options=None, # type: Optional[optparse.Values] -): - # type: (...) -> ParsedRequirement - + line: ParsedLine, + options: Optional[optparse.Values] = None, +) -> ParsedRequirement: # preserve for the nested code path - line_comes_from = '{} {} (line {})'.format( - '-c' if line.constraint else '-r', line.filename, line.lineno, + line_comes_from = "{} {} (line {})".format( + "-c" if line.constraint else "-r", + line.filename, + line.lineno, ) assert line.is_requirement + # get the options that apply to requirements if line.is_editable: - # For editable requirements, we don't support per-requirement - # options, so just return the parsed requirement. - return ParsedRequirement( - requirement=line.requirement, - is_editable=line.is_editable, - comes_from=line_comes_from, - constraint=line.constraint, - ) + supported_dest = SUPPORTED_OPTIONS_EDITABLE_REQ_DEST else: - if options: - # Disable wheels if the user has specified build options - cmdoptions.check_install_build_global(options, line.opts) - - # get the options that apply to requirements - req_options = {} - for dest in SUPPORTED_OPTIONS_REQ_DEST: - if dest in line.opts.__dict__ and line.opts.__dict__[dest]: - req_options[dest] = line.opts.__dict__[dest] - - line_source = 'line {} of {}'.format(line.lineno, line.filename) - return ParsedRequirement( - requirement=line.requirement, - is_editable=line.is_editable, - comes_from=line_comes_from, - constraint=line.constraint, - options=req_options, - line_source=line_source, - ) + supported_dest = SUPPORTED_OPTIONS_REQ_DEST + req_options = {} + for dest in supported_dest: + if dest in line.opts.__dict__ and line.opts.__dict__[dest]: + req_options[dest] = line.opts.__dict__[dest] + + line_source = f"line {line.lineno} of {line.filename}" + return ParsedRequirement( + requirement=line.requirement, + is_editable=line.is_editable, + comes_from=line_comes_from, + constraint=line.constraint, + options=req_options, + line_source=line_source, + ) def handle_option_line( - opts, # type: Values - filename, # type: str - lineno, # type: int - finder=None, # type: Optional[PackageFinder] - options=None, # type: Optional[optparse.Values] - session=None, # type: Optional[PipSession] -): - # type: (...) -> None - - # percolate hash-checking option upward - if opts.require_hashes: - options.require_hashes = opts.require_hashes + opts: Values, + filename: str, + lineno: int, + finder: Optional["PackageFinder"] = None, + options: Optional[optparse.Values] = None, + session: Optional["PipSession"] = None, +) -> None: + if opts.hashes: + logger.warning( + "%s line %s has --hash but no requirement, and will be ignored.", + filename, + lineno, + ) + + if options: + # percolate options upward + if opts.require_hashes: + options.require_hashes = opts.require_hashes + if opts.features_enabled: + options.features_enabled.extend( + f for f in opts.features_enabled if f not in options.features_enabled + ) # set finder options - elif finder: + if finder: find_links = finder.find_links index_urls = finder.index_urls - if opts.index_url: - index_urls = [opts.index_url] + no_index = finder.search_scope.no_index if opts.no_index is True: + no_index = True index_urls = [] - if opts.extra_index_urls: + if opts.index_url and not no_index: + index_urls = [opts.index_url] + if opts.extra_index_urls and not no_index: index_urls.extend(opts.extra_index_urls) if opts.find_links: # FIXME: it would be nice to keep track of the source @@ -251,28 +249,35 @@ def handle_option_line( value = relative_to_reqs_file find_links.append(value) + if session: + # We need to update the auth urls in session + session.update_index_urls(index_urls) + search_scope = SearchScope( find_links=find_links, index_urls=index_urls, + no_index=no_index, ) finder.search_scope = search_scope if opts.pre: finder.set_allow_all_prereleases() + if opts.prefer_binary: + finder.set_prefer_binary() + if session: for host in opts.trusted_hosts or []: - source = 'line {} of {}'.format(lineno, filename) + source = f"line {lineno} of {filename}" session.add_trusted_host(host, source=source) def handle_line( - line, # type: ParsedLine - options=None, # type: Optional[optparse.Values] - finder=None, # type: Optional[PackageFinder] - session=None, # type: Optional[PipSession] -): - # type: (...) -> Optional[ParsedRequirement] + line: ParsedLine, + options: Optional[optparse.Values] = None, + finder: Optional["PackageFinder"] = None, + session: Optional["PipSession"] = None, +) -> Optional[ParsedRequirement]: """Handle a single parsed requirements line; This can result in creating/yielding requirements, or updating the finder. @@ -311,31 +316,27 @@ def handle_line( return None -class RequirementsFileParser(object): +class RequirementsFileParser: def __init__( self, - session, # type: PipSession - line_parser, # type: LineParser - comes_from, # type: str - ): - # type: (...) -> None + session: "PipSession", + line_parser: LineParser, + ) -> None: self._session = session self._line_parser = line_parser - self._comes_from = comes_from - def parse(self, filename, constraint): - # type: (str, bool) -> Iterator[ParsedLine] - """Parse a given file, yielding parsed lines. - """ - for line in self._parse_and_recurse(filename, constraint): - yield line + def parse( + self, filename: str, constraint: bool + ) -> Generator[ParsedLine, None, None]: + """Parse a given file, yielding parsed lines.""" + yield from self._parse_and_recurse(filename, constraint) - def _parse_and_recurse(self, filename, constraint): - # type: (str, bool) -> Iterator[ParsedLine] + def _parse_and_recurse( + self, filename: str, constraint: bool + ) -> Generator[ParsedLine, None, None]: for line in self._parse_file(filename, constraint): - if ( - not line.is_requirement and - (line.opts.requirements or line.opts.constraints) + if not line.is_requirement and ( + line.opts.requirements or line.opts.constraints ): # parse a nested requirements file if line.opts.requirements: @@ -348,26 +349,23 @@ def _parse_and_recurse(self, filename, constraint): # original file is over http if SCHEME_RE.search(filename): # do a url join so relative paths work - req_path = urllib_parse.urljoin(filename, req_path) + req_path = urllib.parse.urljoin(filename, req_path) # original file and nested file are paths elif not SCHEME_RE.search(req_path): # do a join so relative paths work req_path = os.path.join( - os.path.dirname(filename), req_path, + os.path.dirname(filename), + req_path, ) - for inner_line in self._parse_and_recurse( - req_path, nested_constraint, - ): - yield inner_line + yield from self._parse_and_recurse(req_path, nested_constraint) else: yield line - def _parse_file(self, filename, constraint): - # type: (str, bool) -> Iterator[ParsedLine] - _, content = get_file_content( - filename, self._session, comes_from=self._comes_from - ) + def _parse_file( + self, filename: str, constraint: bool + ) -> Generator[ParsedLine, None, None]: + _, content = get_file_content(filename, self._session) lines_enum = preprocess(content) @@ -376,23 +374,20 @@ def _parse_file(self, filename, constraint): args_str, opts = self._line_parser(line) except OptionParsingError as e: # add offending line - msg = 'Invalid requirement: {}\n{}'.format(line, e.msg) + msg = f"Invalid requirement: {line}\n{e.msg}" raise RequirementsFileParseError(msg) yield ParsedLine( filename, line_number, - self._comes_from, args_str, opts, constraint, ) -def get_line_parser(finder): - # type: (Optional[PackageFinder]) -> LineParser - def parse_line(line): - # type: (Text) -> Tuple[str, Values] +def get_line_parser(finder: Optional["PackageFinder"]) -> LineParser: + def parse_line(line: str) -> Tuple[str, Values]: # Build new parser for each line since it accumulates appendable # options. parser = build_parser() @@ -402,46 +397,42 @@ def parse_line(line): defaults.format_control = finder.format_control args_str, options_str = break_args_options(line) - # Prior to 2.7.3, shlex cannot deal with unicode entries - if sys.version_info < (2, 7, 3): - # https://github.com/python/mypy/issues/1174 - options_str = options_str.encode('utf8') # type: ignore - # https://github.com/python/mypy/issues/1174 - opts, _ = parser.parse_args( - shlex.split(options_str), defaults) # type: ignore + try: + options = shlex.split(options_str) + except ValueError as e: + raise OptionParsingError(f"Could not split options: {options_str}") from e + + opts, _ = parser.parse_args(options, defaults) return args_str, opts return parse_line -def break_args_options(line): - # type: (Text) -> Tuple[str, Text] +def break_args_options(line: str) -> Tuple[str, str]: """Break up the line into an args and options string. We only want to shlex (and then optparse) the options, not the args. args can contain markers which are corrupted by shlex. """ - tokens = line.split(' ') + tokens = line.split(" ") args = [] options = tokens[:] for token in tokens: - if token.startswith('-') or token.startswith('--'): + if token.startswith("-") or token.startswith("--"): break else: args.append(token) options.pop(0) - return ' '.join(args), ' '.join(options) # type: ignore + return " ".join(args), " ".join(options) class OptionParsingError(Exception): - def __init__(self, msg): - # type: (str) -> None + def __init__(self, msg: str) -> None: self.msg = msg -def build_parser(): - # type: () -> optparse.OptionParser +def build_parser() -> optparse.OptionParser: """ Return a parser for parsing requirement lines """ @@ -454,9 +445,9 @@ def build_parser(): # By default optparse sys.exits on parsing errors. We want to wrap # that in our own exception. - def parser_exit(self, msg): - # type: (Any, str) -> NoReturn + def parser_exit(self: Any, msg: str) -> "NoReturn": raise OptionParsingError(msg) + # NOTE: mypy disallows assigning to a method # https://github.com/python/mypy/issues/2427 parser.exit = parser_exit # type: ignore @@ -464,50 +455,49 @@ def parser_exit(self, msg): return parser -def join_lines(lines_enum): - # type: (ReqFileLines) -> ReqFileLines +def join_lines(lines_enum: ReqFileLines) -> ReqFileLines: """Joins a line ending in '\' with the previous line (except when following comments). The joined line takes on the index of the first line. """ primary_line_number = None - new_line = [] # type: List[Text] + new_line: List[str] = [] for line_number, line in lines_enum: - if not line.endswith('\\') or COMMENT_RE.match(line): + if not line.endswith("\\") or COMMENT_RE.match(line): if COMMENT_RE.match(line): # this ensures comments are always matched later - line = ' ' + line + line = " " + line if new_line: new_line.append(line) - yield primary_line_number, ''.join(new_line) + assert primary_line_number is not None + yield primary_line_number, "".join(new_line) new_line = [] else: yield line_number, line else: if not new_line: primary_line_number = line_number - new_line.append(line.strip('\\')) + new_line.append(line.strip("\\")) # last line contains \ if new_line: - yield primary_line_number, ''.join(new_line) + assert primary_line_number is not None + yield primary_line_number, "".join(new_line) # TODO: handle space after '\'. -def ignore_comments(lines_enum): - # type: (ReqFileLines) -> ReqFileLines +def ignore_comments(lines_enum: ReqFileLines) -> ReqFileLines: """ Strips comments and filter empty lines. """ for line_number, line in lines_enum: - line = COMMENT_RE.sub('', line) + line = COMMENT_RE.sub("", line) line = line.strip() if line: yield line_number, line -def expand_env_variables(lines_enum): - # type: (ReqFileLines) -> ReqFileLines +def expand_env_variables(lines_enum: ReqFileLines) -> ReqFileLines: """Replace all environment variables that can be retrieved via `os.getenv`. The only allowed format for environment variables defined in the @@ -534,49 +524,28 @@ def expand_env_variables(lines_enum): yield line_number, line -def get_file_content(url, session, comes_from=None): - # type: (str, PipSession, Optional[str]) -> Tuple[str, Text] +def get_file_content(url: str, session: "PipSession") -> Tuple[str, str]: """Gets the content of a file; it may be a filename, file: URL, or http: URL. Returns (location, content). Content is unicode. Respects # -*- coding: declarations on the retrieved files. :param url: File path or url. :param session: PipSession instance. - :param comes_from: Origin description of requirements. """ - scheme = get_url_scheme(url) + scheme = urllib.parse.urlsplit(url).scheme + # Pip has special support for file:// URLs (LocalFSAdapter). + if scheme in ["http", "https", "file"]: + # Delay importing heavy network modules until absolutely necessary. + from pip._internal.network.utils import raise_for_status - if scheme in ['http', 'https']: - # FIXME: catch some errors resp = session.get(url) - resp.raise_for_status() + raise_for_status(resp) return resp.url, resp.text - elif scheme == 'file': - if comes_from and comes_from.startswith('http'): - raise InstallationError( - 'Requirements file {} references URL {}, ' - 'which is local'.format(comes_from, url) - ) - - path = url.split(':', 1)[1] - path = path.replace('\\', '/') - match = _url_slash_drive_re.match(path) - if match: - path = match.group(1) + ':' + path.split('|', 1)[1] - path = urllib_parse.unquote(path) - if path.startswith('/'): - path = '/' + path.lstrip('/') - url = path - + # Assume this is a bare path. try: - with open(url, 'rb') as f: + with open(url, "rb") as f: content = auto_decode(f.read()) - except IOError as exc: - raise InstallationError( - 'Could not open requirements file: {}'.format(exc) - ) + except OSError as exc: + raise InstallationError(f"Could not open requirements file: {exc}") return url, content - - -_url_slash_drive_re = re.compile(r'/*([a-z])\|', re.I) diff --git a/src/fetchcode/vcs/pip/_internal/req/req_install.py b/src/fetchcode/vcs/pip/_internal/req/req_install.py index 2b4d76de..834bc513 100644 --- a/src/fetchcode/vcs/pip/_internal/req/req_install.py +++ b/src/fetchcode/vcs/pip/_internal/req/req_install.py @@ -1,96 +1,68 @@ -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False - -from __future__ import absolute_import - +import functools import logging import os import shutil import sys +import uuid import zipfile - -from fetchcode.vcs.pip._vendor import pkg_resources, six -from fetchcode.vcs.pip._vendor.packaging.requirements import Requirement -from fetchcode.vcs.pip._vendor.packaging.utils import canonicalize_name -from fetchcode.vcs.pip._vendor.packaging.version import Version -from fetchcode.vcs.pip._vendor.packaging.version import parse as parse_version -from fetchcode.vcs.pip._vendor.pep517.wrappers import Pep517HookCaller - -from fetchcode.vcs.pip._internal.build_env import NoOpBuildEnvironment -from fetchcode.vcs.pip._internal.exceptions import InstallationError -from fetchcode.vcs.pip._internal.locations import get_scheme -from fetchcode.vcs.pip._internal.models.link import Link -from fetchcode.vcs.pip._internal.operations.build.metadata import generate_metadata -from fetchcode.vcs.pip._internal.operations.build.metadata_legacy import \ - generate_metadata as generate_metadata_legacy -from fetchcode.vcs.pip._internal.operations.install.editable_legacy import \ - install_editable as install_editable_legacy -from fetchcode.vcs.pip._internal.operations.install.legacy import LegacyInstallFailure -from fetchcode.vcs.pip._internal.operations.install.legacy import install as install_legacy -from fetchcode.vcs.pip._internal.operations.install.wheel import install_wheel -from fetchcode.vcs.pip._internal.pyproject import load_pyproject_toml, make_pyproject_path -from fetchcode.vcs.pip._internal.req.req_uninstall import UninstallPathSet -from fetchcode.vcs.pip._internal.utils.deprecation import deprecated -from fetchcode.vcs.pip._internal.utils.direct_url_helpers import direct_url_from_link -from fetchcode.vcs.pip._internal.utils.hashes import Hashes -from fetchcode.vcs.pip._internal.utils.logging import indent_log -from fetchcode.vcs.pip._internal.utils.misc import ( +from optparse import Values +from pathlib import Path +from typing import Any, Collection, Dict, Iterable, List, Optional, Sequence, Union + +from pip._vendor.packaging.markers import Marker +from pip._vendor.packaging.requirements import Requirement +from pip._vendor.packaging.specifiers import SpecifierSet +from pip._vendor.packaging.utils import canonicalize_name +from pip._vendor.packaging.version import Version +from pip._vendor.packaging.version import parse as parse_version +from pip._vendor.pyproject_hooks import BuildBackendHookCaller + +from pip._internal.build_env import BuildEnvironment, NoOpBuildEnvironment +from pip._internal.exceptions import InstallationError, PreviousBuildDirError +from pip._internal.locations import get_scheme +from pip._internal.metadata import ( + BaseDistribution, + get_default_environment, + get_directory_distribution, + get_wheel_distribution, +) +from pip._internal.metadata.base import FilesystemWheel +from pip._internal.models.direct_url import DirectUrl +from pip._internal.models.link import Link +from pip._internal.operations.build.metadata import generate_metadata +from pip._internal.operations.build.metadata_editable import generate_editable_metadata +from pip._internal.operations.build.metadata_legacy import ( + generate_metadata as generate_metadata_legacy, +) +from pip._internal.operations.install.editable_legacy import ( + install_editable as install_editable_legacy, +) +from pip._internal.operations.install.wheel import install_wheel +from pip._internal.pyproject import load_pyproject_toml, make_pyproject_path +from pip._internal.req.req_uninstall import UninstallPathSet +from pip._internal.utils.deprecation import deprecated +from pip._internal.utils.hashes import Hashes +from pip._internal.utils.misc import ( + ConfiguredBuildBackendHookCaller, ask_path_exists, backup_dir, display_path, - dist_in_site_packages, - dist_in_usersite, - get_installed_version, hide_url, + is_installable_dir, + redact_auth_from_requirement, redact_auth_from_url, ) -from fetchcode.vcs.pip._internal.utils.packaging import get_metadata -from fetchcode.vcs.pip._internal.utils.temp_dir import TempDirectory, tempdir_kinds -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.utils.virtualenv import running_under_virtualenv -from fetchcode.vcs.pip._internal.vcs import vcs - -if MYPY_CHECK_RUNNING: - from typing import ( - Any, Dict, Iterable, List, Optional, Sequence, Union, - ) - from fetchcode.vcs.pip._internal.build_env import BuildEnvironment - from fetchcode.vcs.pip._vendor.pkg_resources import Distribution - from fetchcode.vcs.pip._vendor.packaging.specifiers import SpecifierSet - from fetchcode.vcs.pip._vendor.packaging.markers import Marker - +from pip._internal.utils.packaging import get_requirement +from pip._internal.utils.subprocess import runner_with_spinner_message +from pip._internal.utils.temp_dir import TempDirectory, tempdir_kinds +from pip._internal.utils.unpacking import unpack_file +from pip._internal.utils.virtualenv import running_under_virtualenv +from pip._internal.vcs import vcs logger = logging.getLogger(__name__) -def _get_dist(metadata_directory): - # type: (str) -> Distribution - """Return a pkg_resources.Distribution for the provided - metadata directory. - """ - dist_dir = metadata_directory.rstrip(os.sep) - - # Build a PathMetadata object, from path to metadata. :wink: - base_dir, dist_dir_name = os.path.split(dist_dir) - metadata = pkg_resources.PathMetadata(base_dir, dist_dir) - - # Determine the correct Distribution object type. - if dist_dir.endswith(".egg-info"): - dist_cls = pkg_resources.Distribution - dist_name = os.path.splitext(dist_dir_name)[0] - else: - assert dist_dir.endswith(".dist-info") - dist_cls = pkg_resources.DistInfoDistribution - dist_name = os.path.splitext(dist_dir_name)[0].split("-")[0] - - return dist_cls( - base_dir, - project_name=dist_name, - metadata=metadata, - ) - - -class InstallRequirement(object): +class InstallRequirement: """ Represents something that may be installed later on, may have information about where to fetch the relevant requirement and also contains logic for @@ -99,106 +71,108 @@ class InstallRequirement(object): def __init__( self, - req, # type: Optional[Requirement] - comes_from, # type: Optional[Union[str, InstallRequirement]] - editable=False, # type: bool - link=None, # type: Optional[Link] - markers=None, # type: Optional[Marker] - use_pep517=None, # type: Optional[bool] - isolated=False, # type: bool - install_options=None, # type: Optional[List[str]] - global_options=None, # type: Optional[List[str]] - hash_options=None, # type: Optional[Dict[str, List[str]]] - constraint=False, # type: bool - extras=() # type: Iterable[str] - ): - # type: (...) -> None + req: Optional[Requirement], + comes_from: Optional[Union[str, "InstallRequirement"]], + editable: bool = False, + link: Optional[Link] = None, + markers: Optional[Marker] = None, + use_pep517: Optional[bool] = None, + isolated: bool = False, + *, + global_options: Optional[List[str]] = None, + hash_options: Optional[Dict[str, List[str]]] = None, + config_settings: Optional[Dict[str, Union[str, List[str]]]] = None, + constraint: bool = False, + extras: Collection[str] = (), + user_supplied: bool = False, + permit_editable_wheels: bool = False, + ) -> None: assert req is None or isinstance(req, Requirement), req self.req = req self.comes_from = comes_from self.constraint = constraint self.editable = editable + self.permit_editable_wheels = permit_editable_wheels # source_dir is the local directory where the linked requirement is # located, or unpacked. In case unpacking is needed, creating and # populating source_dir is done by the RequirementPreparer. Note this # is not necessarily the directory where pyproject.toml or setup.py is # located - that one is obtained via unpacked_source_directory. - self.source_dir = None # type: Optional[str] + self.source_dir: Optional[str] = None if self.editable: assert link if link.is_file: - self.source_dir = os.path.normpath( - os.path.abspath(link.file_path) - ) + self.source_dir = os.path.normpath(os.path.abspath(link.file_path)) + # original_link is the direct URL that was provided by the user for the + # requirement, either directly or via a constraints file. if link is None and req and req.url: # PEP 508 URL requirement link = Link(req.url) self.link = self.original_link = link - self.original_link_is_in_wheel_cache = False + + # When this InstallRequirement is a wheel obtained from the cache of locally + # built wheels, this is the source link corresponding to the cache entry, which + # was used to download and build the cached wheel. + self.cached_wheel_source_link: Optional[Link] = None + + # Information about the location of the artifact that was downloaded . This + # property is guaranteed to be set in resolver results. + self.download_info: Optional[DirectUrl] = None # Path to any downloaded or already-existing package. - self.local_file_path = None # type: Optional[str] + self.local_file_path: Optional[str] = None if self.link and self.link.is_file: self.local_file_path = self.link.file_path if extras: self.extras = extras elif req: - self.extras = { - pkg_resources.safe_extra(extra) for extra in req.extras - } + self.extras = req.extras else: self.extras = set() if markers is None and req: markers = req.marker self.markers = markers - # This holds the pkg_resources.Distribution object if this requirement - # is already available: - self.satisfied_by = None # type: Optional[Distribution] + # This holds the Distribution object if this requirement is already installed. + self.satisfied_by: Optional[BaseDistribution] = None # Whether the installation process should try to uninstall an existing # distribution before installing this requirement. self.should_reinstall = False # Temporary build location - self._temp_build_dir = None # type: Optional[TempDirectory] + self._temp_build_dir: Optional[TempDirectory] = None # Set to True after successful installation - self.install_succeeded = None # type: Optional[bool] + self.install_succeeded: Optional[bool] = None # Supplied options - self.install_options = install_options if install_options else [] self.global_options = global_options if global_options else [] self.hash_options = hash_options if hash_options else {} + self.config_settings = config_settings # Set to True after successful preparation of this requirement self.prepared = False - self.is_direct = False - - # Set by the legacy resolver when the requirement has been downloaded - # TODO: This introduces a strong coupling between the resolver and the - # requirement (the coupling was previously between the resolver - # and the requirement set). This should be refactored to allow - # the requirement to decide for itself when it has been - # successfully downloaded - but that is more tricky to get right, - # se we are making the change in stages. - self.successfully_downloaded = False + # User supplied requirement are explicitly requested for installation + # by the user via CLI arguments or requirements files, as opposed to, + # e.g. dependencies, extras or constraints. + self.user_supplied = user_supplied self.isolated = isolated - self.build_env = NoOpBuildEnvironment() # type: BuildEnvironment + self.build_env: BuildEnvironment = NoOpBuildEnvironment() # For PEP 517, the directory where we request the project metadata # gets stored. We need this to pass to build_wheel, so the backend # can ensure that the wheel matches the metadata (see the PEP for # details). - self.metadata_directory = None # type: Optional[str] + self.metadata_directory: Optional[str] = None # The static build requirements (from pyproject.toml) - self.pyproject_requires = None # type: Optional[List[str]] + self.pyproject_requires: Optional[List[str]] = None # Build requirements that we will check are available - self.requirements_to_check = [] # type: List[str] + self.requirements_to_check: List[str] = [] # The PEP 517 backend we should use to build the project - self.pep517_backend = None # type: Optional[Pep517HookCaller] + self.pep517_backend: Optional[BuildBackendHookCaller] = None # Are we using PEP 517 for this requirement? # After pyproject.toml has been loaded, the only valid values are True @@ -207,92 +181,116 @@ def __init__( # but after loading this flag should be treated as read only. self.use_pep517 = use_pep517 - def __str__(self): - # type: () -> str + # If config settings are provided, enforce PEP 517. + if self.config_settings: + if self.use_pep517 is False: + logger.warning( + "--no-use-pep517 ignored for %s " + "because --config-settings are specified.", + self, + ) + self.use_pep517 = True + + # This requirement needs more preparation before it can be built + self.needs_more_preparation = False + + # This requirement needs to be unpacked before it can be installed. + self._archive_source: Optional[Path] = None + + def __str__(self) -> str: if self.req: - s = str(self.req) + s = redact_auth_from_requirement(self.req) if self.link: - s += ' from {}'.format(redact_auth_from_url(self.link.url)) + s += f" from {redact_auth_from_url(self.link.url)}" elif self.link: s = redact_auth_from_url(self.link.url) else: - s = '' + s = "" if self.satisfied_by is not None: - s += ' in {}'.format(display_path(self.satisfied_by.location)) + if self.satisfied_by.location is not None: + location = display_path(self.satisfied_by.location) + else: + location = "" + s += f" in {location}" if self.comes_from: - if isinstance(self.comes_from, six.string_types): - comes_from = self.comes_from # type: Optional[str] + if isinstance(self.comes_from, str): + comes_from: Optional[str] = self.comes_from else: comes_from = self.comes_from.from_path() if comes_from: - s += ' (from {})'.format(comes_from) + s += f" (from {comes_from})" return s - def __repr__(self): - # type: () -> str - return '<{} object: {} editable={!r}>'.format( - self.__class__.__name__, str(self), self.editable) + def __repr__(self) -> str: + return ( + f"<{self.__class__.__name__} object: " + f"{str(self)} editable={self.editable!r}>" + ) - def format_debug(self): - # type: () -> str - """An un-tested helper for getting state, for debugging. - """ + def format_debug(self) -> str: + """An un-tested helper for getting state, for debugging.""" attributes = vars(self) names = sorted(attributes) - state = ( - "{}={!r}".format(attr, attributes[attr]) for attr in sorted(names) - ) - return '<{name} object: {{{state}}}>'.format( + state = (f"{attr}={attributes[attr]!r}" for attr in sorted(names)) + return "<{name} object: {{{state}}}>".format( name=self.__class__.__name__, state=", ".join(state), ) # Things that are valid for all kinds of requirements? @property - def name(self): - # type: () -> Optional[str] + def name(self) -> Optional[str]: if self.req is None: return None - return six.ensure_str(pkg_resources.safe_name(self.req.name)) + return self.req.name + + @functools.cached_property + def supports_pyproject_editable(self) -> bool: + if not self.use_pep517: + return False + assert self.pep517_backend + with self.build_env: + runner = runner_with_spinner_message( + "Checking if build backend supports build_editable" + ) + with self.pep517_backend.subprocess_runner(runner): + return "build_editable" in self.pep517_backend._supported_features() @property - def specifier(self): - # type: () -> SpecifierSet + def specifier(self) -> SpecifierSet: + assert self.req is not None return self.req.specifier @property - def is_pinned(self): - # type: () -> bool + def is_direct(self) -> bool: + """Whether this requirement was specified as a direct URL.""" + return self.original_link is not None + + @property + def is_pinned(self) -> bool: """Return whether I am pinned to an exact version. For example, some-package==1.2 is pinned; some-package>1.2 is not. """ - specifiers = self.specifier - return (len(specifiers) == 1 and - next(iter(specifiers)).operator in {'==', '==='}) - - @property - def installed_version(self): - # type: () -> Optional[str] - return get_installed_version(self.name) + assert self.req is not None + specifiers = self.req.specifier + return len(specifiers) == 1 and next(iter(specifiers)).operator in {"==", "==="} - def match_markers(self, extras_requested=None): - # type: (Optional[Iterable[str]]) -> bool + def match_markers(self, extras_requested: Optional[Iterable[str]] = None) -> bool: if not extras_requested: # Provide an extra to safely evaluate the markers # without matching any extra - extras_requested = ('',) + extras_requested = ("",) if self.markers is not None: return any( - self.markers.evaluate({'extra': extra}) - for extra in extras_requested) + self.markers.evaluate({"extra": extra}) for extra in extras_requested + ) else: return True @property - def has_hash_options(self): - # type: () -> bool + def has_hash_options(self) -> bool: """Return whether any known-good hashes are specified as options. These activate --require-hashes mode; hashes specified as part of a @@ -301,8 +299,7 @@ def has_hash_options(self): """ return bool(self.hash_options) - def hashes(self, trust_internet=True): - # type: (bool) -> Hashes + def hashes(self, trust_internet: bool = True) -> Hashes: """Return a hash-comparer that considers my option- and URL-based hashes to be known-good. @@ -318,29 +315,35 @@ def hashes(self, trust_internet=True): """ good_hashes = self.hash_options.copy() - link = self.link if trust_internet else self.original_link + if trust_internet: + link = self.link + elif self.is_direct and self.user_supplied: + link = self.original_link + else: + link = None if link and link.hash: + assert link.hash_name is not None good_hashes.setdefault(link.hash_name, []).append(link.hash) return Hashes(good_hashes) - def from_path(self): - # type: () -> Optional[str] - """Format a nice indicator to show where this "comes from" - """ + def from_path(self) -> Optional[str]: + """Format a nice indicator to show where this "comes from" """ if self.req is None: return None s = str(self.req) if self.comes_from: - if isinstance(self.comes_from, six.string_types): + comes_from: Optional[str] + if isinstance(self.comes_from, str): comes_from = self.comes_from else: comes_from = self.comes_from.from_path() if comes_from: - s += '->' + comes_from + s += "->" + comes_from return s - def ensure_build_location(self, build_dir, autodelete): - # type: (str, bool) -> str + def ensure_build_location( + self, build_dir: str, autodelete: bool, parallel_builds: bool + ) -> str: assert build_dir is not None if self._temp_build_dir is not None: assert self._temp_build_dir.path @@ -354,16 +357,23 @@ def ensure_build_location(self, build_dir, autodelete): ) return self._temp_build_dir.path - if self.editable: - name = self.name.lower() - else: - name = self.name + + # This is the only remaining place where we manually determine the path + # for the temporary directory. It is only needed for editables where + # it is the value of the --src option. + + # When parallel builds are enabled, add a UUID to the build directory + # name so multiple builds do not interfere with each other. + dir_name: str = canonicalize_name(self.req.name) + if parallel_builds: + dir_name = f"{dir_name}_{uuid.uuid4().hex}" + # FIXME: Is there a better place to create the build_dir? (hg and bzr # need this) if not os.path.exists(build_dir): - logger.debug('Creating directory %s', build_dir) + logger.debug("Creating directory %s", build_dir) os.makedirs(build_dir) - actual_build_dir = os.path.join(build_dir, name) + actual_build_dir = os.path.join(build_dir, dir_name) # `None` indicates that we respect the globally-configured deletion # settings, which is what we actually want when auto-deleting. delete_arg = None if autodelete else False @@ -374,10 +384,8 @@ def ensure_build_location(self, build_dir, autodelete): globally_managed=True, ).path - def _set_requirement(self): - # type: () -> None - """Set requirement after generating metadata. - """ + def _set_requirement(self) -> None: + """Set requirement after generating metadata.""" assert self.req is None assert self.metadata is not None assert self.source_dir is not None @@ -388,16 +396,18 @@ def _set_requirement(self): else: op = "===" - self.req = Requirement( - "".join([ - self.metadata["Name"], - op, - self.metadata["Version"], - ]) + self.req = get_requirement( + "".join( + [ + self.metadata["Name"], + op, + self.metadata["Version"], + ] + ) ) - def warn_on_mismatching_name(self): - # type: () -> None + def warn_on_mismatching_name(self) -> None: + assert self.req is not None metadata_name = canonicalize_name(self.metadata["Name"]) if canonicalize_name(self.req.name) == metadata_name: # Everything is fine. @@ -405,90 +415,93 @@ def warn_on_mismatching_name(self): # If we're here, there's a mismatch. Log a warning about it. logger.warning( - 'Generating metadata for package %s ' - 'produced metadata for project name %s. Fix your ' - '#egg=%s fragments.', - self.name, metadata_name, self.name + "Generating metadata for package %s " + "produced metadata for project name %s. Fix your " + "#egg=%s fragments.", + self.name, + metadata_name, + self.name, ) - self.req = Requirement(metadata_name) + self.req = get_requirement(metadata_name) - def check_if_exists(self, use_user_site): - # type: (bool) -> None + def check_if_exists(self, use_user_site: bool) -> None: """Find an installed distribution that satisfies or conflicts with this requirement, and set self.satisfied_by or self.should_reinstall appropriately. """ if self.req is None: return - # get_distribution() will resolve the entire list of requirements - # anyway, and we've already determined that we need the requirement - # in question, so strip the marker so that we don't try to - # evaluate it. - no_marker = Requirement(str(self.req)) - no_marker.marker = None - try: - self.satisfied_by = pkg_resources.get_distribution(str(no_marker)) - except pkg_resources.DistributionNotFound: + existing_dist = get_default_environment().get_distribution(self.req.name) + if not existing_dist: return - except pkg_resources.VersionConflict: - existing_dist = pkg_resources.get_distribution( - self.req.name - ) + + version_compatible = self.req.specifier.contains( + existing_dist.version, + prereleases=True, + ) + if not version_compatible: + self.satisfied_by = None if use_user_site: - if dist_in_usersite(existing_dist): + if existing_dist.in_usersite: self.should_reinstall = True - elif (running_under_virtualenv() and - dist_in_site_packages(existing_dist)): + elif running_under_virtualenv() and existing_dist.in_site_packages: raise InstallationError( - "Will not install to the user site because it will " - "lack sys.path precedence to {} in {}".format( - existing_dist.project_name, existing_dist.location) + f"Will not install to the user site because it will " + f"lack sys.path precedence to {existing_dist.raw_name} " + f"in {existing_dist.location}" ) else: self.should_reinstall = True else: - if self.editable and self.satisfied_by: + if self.editable: self.should_reinstall = True # when installing editables, nothing pre-existing should ever # satisfy self.satisfied_by = None + else: + self.satisfied_by = existing_dist # Things valid for wheels @property - def is_wheel(self): - # type: () -> bool + def is_wheel(self) -> bool: if not self.link: return False return self.link.is_wheel + @property + def is_wheel_from_cache(self) -> bool: + # When True, it means that this InstallRequirement is a local wheel file in the + # cache of locally built wheels. + return self.cached_wheel_source_link is not None + # Things valid for sdists @property - def unpacked_source_directory(self): - # type: () -> str + def unpacked_source_directory(self) -> str: + assert self.source_dir, f"No source dir for {self}" return os.path.join( - self.source_dir, - self.link and self.link.subdirectory_fragment or '') + self.source_dir, self.link and self.link.subdirectory_fragment or "" + ) @property - def setup_py_path(self): - # type: () -> str - assert self.source_dir, "No source dir for {}".format(self) - setup_py = os.path.join(self.unpacked_source_directory, 'setup.py') - - # Python2 __file__ should not be unicode - if six.PY2 and isinstance(setup_py, six.text_type): - setup_py = setup_py.encode(sys.getfilesystemencoding()) + def setup_py_path(self) -> str: + assert self.source_dir, f"No source dir for {self}" + setup_py = os.path.join(self.unpacked_source_directory, "setup.py") return setup_py @property - def pyproject_toml_path(self): - # type: () -> str - assert self.source_dir, "No source dir for {}".format(self) + def setup_cfg_path(self) -> str: + assert self.source_dir, f"No source dir for {self}" + setup_cfg = os.path.join(self.unpacked_source_directory, "setup.cfg") + + return setup_cfg + + @property + def pyproject_toml_path(self) -> str: + assert self.source_dir, f"No source dir for {self}" return make_pyproject_path(self.unpacked_source_directory) - def load_pyproject_toml(self): - # type: () -> None + def load_pyproject_toml(self) -> None: """Load the pyproject.toml file. After calling this routine, all of the attributes related to PEP 517 @@ -497,13 +510,11 @@ def load_pyproject_toml(self): follow the PEP 517 or legacy (setup.py) code path. """ pyproject_toml_data = load_pyproject_toml( - self.use_pep517, - self.pyproject_toml_path, - self.setup_py_path, - str(self) + self.use_pep517, self.pyproject_toml_path, self.setup_py_path, str(self) ) if pyproject_toml_data is None: + assert not self.config_settings self.use_pep517 = False return @@ -511,43 +522,69 @@ def load_pyproject_toml(self): requires, backend, check, backend_path = pyproject_toml_data self.requirements_to_check = check self.pyproject_requires = requires - self.pep517_backend = Pep517HookCaller( - self.unpacked_source_directory, backend, backend_path=backend_path, + self.pep517_backend = ConfiguredBuildBackendHookCaller( + self, + self.unpacked_source_directory, + backend, + backend_path=backend_path, ) - def _generate_metadata(self): - # type: () -> str - """Invokes metadata generator functions, with the required arguments. - """ - if not self.use_pep517: - assert self.unpacked_source_directory + def isolated_editable_sanity_check(self) -> None: + """Check that an editable requirement if valid for use with PEP 517/518. - return generate_metadata_legacy( - build_env=self.build_env, - setup_py_path=self.setup_py_path, - source_dir=self.unpacked_source_directory, - isolated=self.isolated, - details=self.name or "from {}".format(self.link) + This verifies that an editable that has a pyproject.toml either supports PEP 660 + or as a setup.py or a setup.cfg + """ + if ( + self.editable + and self.use_pep517 + and not self.supports_pyproject_editable + and not os.path.isfile(self.setup_py_path) + and not os.path.isfile(self.setup_cfg_path) + ): + raise InstallationError( + f"Project {self} has a 'pyproject.toml' and its build " + f"backend is missing the 'build_editable' hook. Since it does not " + f"have a 'setup.py' nor a 'setup.cfg', " + f"it cannot be installed in editable mode. " + f"Consider using a build backend that supports PEP 660." ) - assert self.pep517_backend is not None - - return generate_metadata( - build_env=self.build_env, - backend=self.pep517_backend, - ) - - def prepare_metadata(self): - # type: () -> None + def prepare_metadata(self) -> None: """Ensure that project metadata is available. - Under PEP 517, call the backend hook to prepare the metadata. + Under PEP 517 and PEP 660, call the backend hook to prepare the metadata. Under legacy processing, call setup.py egg-info. """ - assert self.source_dir - - with indent_log(): - self.metadata_directory = self._generate_metadata() + assert self.source_dir, f"No source dir for {self}" + details = self.name or f"from {self.link}" + + if self.use_pep517: + assert self.pep517_backend is not None + if ( + self.editable + and self.permit_editable_wheels + and self.supports_pyproject_editable + ): + self.metadata_directory = generate_editable_metadata( + build_env=self.build_env, + backend=self.pep517_backend, + details=details, + ) + else: + self.metadata_directory = generate_metadata( + build_env=self.build_env, + backend=self.pep517_backend, + details=details, + ) + else: + self.metadata_directory = generate_metadata_legacy( + build_env=self.build_env, + setup_py_path=self.setup_py_path, + source_dir=self.unpacked_source_directory, + isolated=self.isolated, + details=details, + ) # Act on the newly generated metadata, based on the name and version. if not self.name: @@ -558,38 +595,50 @@ def prepare_metadata(self): self.assert_source_matches_version() @property - def metadata(self): - # type: () -> Any - if not hasattr(self, '_metadata'): - self._metadata = get_metadata(self.get_dist()) + def metadata(self) -> Any: + if not hasattr(self, "_metadata"): + self._metadata = self.get_dist().metadata return self._metadata - def get_dist(self): - # type: () -> Distribution - return _get_dist(self.metadata_directory) + def get_dist(self) -> BaseDistribution: + if self.metadata_directory: + return get_directory_distribution(self.metadata_directory) + elif self.local_file_path and self.is_wheel: + assert self.req is not None + return get_wheel_distribution( + FilesystemWheel(self.local_file_path), + canonicalize_name(self.req.name), + ) + raise AssertionError( + f"InstallRequirement {self} has no metadata directory and no wheel: " + f"can't make a distribution." + ) - def assert_source_matches_version(self): - # type: () -> None - assert self.source_dir - version = self.metadata['version'] - if self.req.specifier and version not in self.req.specifier: + def assert_source_matches_version(self) -> None: + assert self.source_dir, f"No source dir for {self}" + version = self.metadata["version"] + if self.req and self.req.specifier and version not in self.req.specifier: logger.warning( - 'Requested %s, but installing version %s', + "Requested %s, but installing version %s", self, version, ) else: logger.debug( - 'Source in %s has version %s, which satisfies requirement %s', + "Source in %s has version %s, which satisfies requirement %s", display_path(self.source_dir), version, self, ) # For both source distributions and editables - def ensure_has_source_dir(self, parent_dir, autodelete=False): - # type: (str, bool) -> None + def ensure_has_source_dir( + self, + parent_dir: str, + autodelete: bool = False, + parallel_builds: bool = False, + ) -> None: """Ensure that a source_dir is set. This will create a temporary build dir if the name of the requirement @@ -601,56 +650,56 @@ def ensure_has_source_dir(self, parent_dir, autodelete=False): """ if self.source_dir is None: self.source_dir = self.ensure_build_location( - parent_dir, autodelete + parent_dir, + autodelete=autodelete, + parallel_builds=parallel_builds, + ) + + def needs_unpacked_archive(self, archive_source: Path) -> None: + assert self._archive_source is None + self._archive_source = archive_source + + def ensure_pristine_source_checkout(self) -> None: + """Ensure the source directory has not yet been built in.""" + assert self.source_dir is not None + if self._archive_source is not None: + unpack_file(str(self._archive_source), self.source_dir) + elif is_installable_dir(self.source_dir): + # If a checkout exists, it's unwise to keep going. + # version inconsistencies are logged later, but do not fail + # the installation. + raise PreviousBuildDirError( + f"pip can't proceed with requirements '{self}' due to a " + f"pre-existing build directory ({self.source_dir}). This is likely " + "due to a previous installation that failed . pip is " + "being responsible and not assuming it can delete this. " + "Please delete it and try again." ) # For editable installations - def update_editable(self, obtain=True): - # type: (bool) -> None + def update_editable(self) -> None: if not self.link: logger.debug( - "Cannot update repository at %s; repository location is " - "unknown", + "Cannot update repository at %s; repository location is unknown", self.source_dir, ) return assert self.editable assert self.source_dir - if self.link.scheme == 'file': + if self.link.scheme == "file": # Static paths don't get updated return - assert '+' in self.link.url, \ - "bad url: {self.link.url!r}".format(**locals()) - vc_type, url = self.link.url.split('+', 1) - vcs_backend = vcs.get_backend(vc_type) - if vcs_backend: - if not self.link.is_vcs: - reason = ( - "This form of VCS requirement is being deprecated: {}." - ).format( - self.link.url - ) - replacement = None - if self.link.url.startswith("git+git@"): - replacement = ( - "git+https://git@example.com/..., " - "git+ssh://git@example.com/..., " - "or the insecure git+git://git@example.com/..." - ) - deprecated(reason, replacement, gone_in="21.0", issue=7554) - hidden_url = hide_url(self.link.url) - if obtain: - vcs_backend.obtain(self.source_dir, url=hidden_url) - else: - vcs_backend.export(self.source_dir, url=hidden_url) - else: - assert 0, ( - 'Unexpected version control type (in {}): {}'.format( - self.link, vc_type)) + vcs_backend = vcs.get_backend_for_scheme(self.link.scheme) + # Editable requirements are validated in Requirement constructors. + # So here, if it's neither a path nor a valid VCS URL, it's a bug. + assert vcs_backend, f"Unsupported VCS URL {self.link.url}" + hidden_url = hide_url(self.link.url) + vcs_backend.obtain(self.source_dir, url=hidden_url, verbosity=0) # Top-level Actions - def uninstall(self, auto_confirm=False, verbose=False): - # type: (bool, bool) -> Optional[UninstallPathSet] + def uninstall( + self, auto_confirm: bool = False, verbose: bool = False + ) -> Optional[UninstallPathSet]: """ Uninstall the distribution currently satisfying this requirement. @@ -664,110 +713,110 @@ def uninstall(self, auto_confirm=False, verbose=False): """ assert self.req - try: - dist = pkg_resources.get_distribution(self.req.name) - except pkg_resources.DistributionNotFound: + dist = get_default_environment().get_distribution(self.req.name) + if not dist: logger.warning("Skipping %s as it is not installed.", self.name) return None - else: - logger.info('Found existing installation: %s', dist) + logger.info("Found existing installation: %s", dist) uninstalled_pathset = UninstallPathSet.from_dist(dist) uninstalled_pathset.remove(auto_confirm, verbose) return uninstalled_pathset - def _get_archive_name(self, path, parentdir, rootdir): - # type: (str, str, str) -> str - - def _clean_zip_name(name, prefix): - # type: (str, str) -> str - assert name.startswith(prefix + os.path.sep), ( - "name {name!r} doesn't start with prefix {prefix!r}" - .format(**locals()) - ) - name = name[len(prefix) + 1:] - name = name.replace(os.path.sep, '/') + def _get_archive_name(self, path: str, parentdir: str, rootdir: str) -> str: + def _clean_zip_name(name: str, prefix: str) -> str: + assert name.startswith( + prefix + os.path.sep + ), f"name {name!r} doesn't start with prefix {prefix!r}" + name = name[len(prefix) + 1 :] + name = name.replace(os.path.sep, "/") return name + assert self.req is not None path = os.path.join(parentdir, path) name = _clean_zip_name(path, rootdir) - return self.name + '/' + name + return self.req.name + "/" + name - def archive(self, build_dir): - # type: (str) -> None + def archive(self, build_dir: Optional[str]) -> None: """Saves archive to provided build_dir. Used for saving downloaded VCS requirements as part of `pip download`. """ assert self.source_dir + if build_dir is None: + return create_archive = True - archive_name = '{}-{}.zip'.format(self.name, self.metadata["version"]) + archive_name = "{}-{}.zip".format(self.name, self.metadata["version"]) archive_path = os.path.join(build_dir, archive_name) if os.path.exists(archive_path): response = ask_path_exists( - 'The file {} exists. (i)gnore, (w)ipe, ' - '(b)ackup, (a)bort '.format( - display_path(archive_path)), - ('i', 'w', 'b', 'a')) - if response == 'i': + f"The file {display_path(archive_path)} exists. (i)gnore, (w)ipe, " + "(b)ackup, (a)bort ", + ("i", "w", "b", "a"), + ) + if response == "i": create_archive = False - elif response == 'w': - logger.warning('Deleting %s', display_path(archive_path)) + elif response == "w": + logger.warning("Deleting %s", display_path(archive_path)) os.remove(archive_path) - elif response == 'b': + elif response == "b": dest_file = backup_dir(archive_path) logger.warning( - 'Backing up %s to %s', + "Backing up %s to %s", display_path(archive_path), display_path(dest_file), ) shutil.move(archive_path, dest_file) - elif response == 'a': + elif response == "a": sys.exit(-1) if not create_archive: return zip_output = zipfile.ZipFile( - archive_path, 'w', zipfile.ZIP_DEFLATED, allowZip64=True, + archive_path, + "w", + zipfile.ZIP_DEFLATED, + allowZip64=True, ) with zip_output: - dir = os.path.normcase( - os.path.abspath(self.unpacked_source_directory) - ) + dir = os.path.normcase(os.path.abspath(self.unpacked_source_directory)) for dirpath, dirnames, filenames in os.walk(dir): for dirname in dirnames: dir_arcname = self._get_archive_name( - dirname, parentdir=dirpath, rootdir=dir, + dirname, + parentdir=dirpath, + rootdir=dir, ) - zipdir = zipfile.ZipInfo(dir_arcname + '/') + zipdir = zipfile.ZipInfo(dir_arcname + "/") zipdir.external_attr = 0x1ED << 16 # 0o755 - zip_output.writestr(zipdir, '') + zip_output.writestr(zipdir, "") for filename in filenames: file_arcname = self._get_archive_name( - filename, parentdir=dirpath, rootdir=dir, + filename, + parentdir=dirpath, + rootdir=dir, ) filename = os.path.join(dirpath, filename) zip_output.write(filename, file_arcname) - logger.info('Saved %s', display_path(archive_path)) + logger.info("Saved %s", display_path(archive_path)) def install( self, - install_options, # type: List[str] - global_options=None, # type: Optional[Sequence[str]] - root=None, # type: Optional[str] - home=None, # type: Optional[str] - prefix=None, # type: Optional[str] - warn_script_location=True, # type: bool - use_user_site=False, # type: bool - pycompile=True # type: bool - ): - # type: (...) -> None + global_options: Optional[Sequence[str]] = None, + root: Optional[str] = None, + home: Optional[str] = None, + prefix: Optional[str] = None, + warn_script_location: bool = True, + use_user_site: bool = False, + pycompile: bool = True, + ) -> None: + assert self.req is not None scheme = get_scheme( - self.name, + self.req.name, user=use_user_site, home=home, root=root, @@ -775,15 +824,35 @@ def install( prefix=prefix, ) - global_options = global_options if global_options is not None else [] - if self.editable: + if self.editable and not self.is_wheel: + deprecated( + reason=( + f"Legacy editable install of {self} (setup.py develop) " + "is deprecated." + ), + replacement=( + "to add a pyproject.toml or enable --use-pep517, " + "and use setuptools >= 64. " + "If the resulting installation is not behaving as expected, " + "try using --config-settings editable_mode=compat. " + "Please consult the setuptools documentation for more information" + ), + gone_in="25.0", + issue=11457, + ) + if self.config_settings: + logger.warning( + "--config-settings ignored for legacy editable install of %s. " + "Consider upgrading to a version of setuptools " + "that supports PEP 660 (>= 64).", + self, + ) install_editable_legacy( - install_options, - global_options, + global_options=global_options if global_options is not None else [], prefix=prefix, home=home, use_user_site=use_user_site, - name=self.name, + name=self.req.name, setup_py_path=self.setup_py_path, isolated=self.isolated, build_env=self.build_env, @@ -792,59 +861,74 @@ def install( self.install_succeeded = True return - if self.is_wheel: - assert self.local_file_path - direct_url = None - if self.original_link: - direct_url = direct_url_from_link( - self.original_link, - self.source_dir, - self.original_link_is_in_wheel_cache, - ) - install_wheel( - self.name, - self.local_file_path, - scheme=scheme, - req_description=str(self.req), - pycompile=pycompile, - warn_script_location=warn_script_location, - direct_url=direct_url, - ) - self.install_succeeded = True - return + assert self.is_wheel + assert self.local_file_path + + install_wheel( + self.req.name, + self.local_file_path, + scheme=scheme, + req_description=str(self.req), + pycompile=pycompile, + warn_script_location=warn_script_location, + direct_url=self.download_info if self.is_direct else None, + requested=self.user_supplied, + ) + self.install_succeeded = True + + +def check_invalid_constraint_type(req: InstallRequirement) -> str: + # Check for unsupported forms + problem = "" + if not req.name: + problem = "Unnamed requirements are not allowed as constraints" + elif req.editable: + problem = "Editable requirements are not allowed as constraints" + elif req.extras: + problem = "Constraints cannot have extras" + + if problem: + deprecated( + reason=( + "Constraints are only allowed to take the form of a package " + "name and a version specifier. Other forms were originally " + "permitted as an accident of the implementation, but were " + "undocumented. The new implementation of the resolver no " + "longer supports these forms." + ), + replacement="replacing the constraint with a requirement", + # No plan yet for when the new resolver becomes default + gone_in=None, + issue=8210, + ) - # TODO: Why don't we do this for editable installs? - - # Extend the list of global and install options passed on to - # the setup.py call with the ones from the requirements file. - # Options specified in requirements file override those - # specified on the command line, since the last option given - # to setup.py is the one that is used. - global_options = list(global_options) + self.global_options - install_options = list(install_options) + self.install_options - - try: - success = install_legacy( - install_options=install_options, - global_options=global_options, - root=root, - home=home, - prefix=prefix, - use_user_site=use_user_site, - pycompile=pycompile, - scheme=scheme, - setup_py_path=self.setup_py_path, - isolated=self.isolated, - req_name=self.name, - build_env=self.build_env, - unpacked_source_directory=self.unpacked_source_directory, - req_description=str(self.req), - ) - except LegacyInstallFailure as exc: - self.install_succeeded = False - six.reraise(*exc.parent) - except Exception: - self.install_succeeded = True - raise + return problem - self.install_succeeded = success + +def _has_option(options: Values, reqs: List[InstallRequirement], option: str) -> bool: + if getattr(options, option, None): + return True + for req in reqs: + if getattr(req, option, None): + return True + return False + + +def check_legacy_setup_py_options( + options: Values, + reqs: List[InstallRequirement], +) -> None: + has_build_options = _has_option(options, reqs, "build_options") + has_global_options = _has_option(options, reqs, "global_options") + if has_build_options or has_global_options: + deprecated( + reason="--build-option and --global-option are deprecated.", + issue=11859, + replacement="to use --config-settings", + gone_in="25.0", + ) + logger.warning( + "Implying --no-binary=:all: due to the presence of " + "--build-option / --global-option. " + ) + options.format_control.disallow_binaries() diff --git a/src/fetchcode/vcs/pip/_internal/req/req_set.py b/src/fetchcode/vcs/pip/_internal/req/req_set.py index 1e84be3e..ec7a6e07 100644 --- a/src/fetchcode/vcs/pip/_internal/req/req_set.py +++ b/src/fetchcode/vcs/pip/_internal/req/req_set.py @@ -1,202 +1,82 @@ -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False - -from __future__ import absolute_import - import logging from collections import OrderedDict +from typing import Dict, List -from fetchcode.vcs.pip._vendor.packaging.utils import canonicalize_name - -from fetchcode.vcs.pip._internal.exceptions import InstallationError -from fetchcode.vcs.pip._internal.models.wheel import Wheel -from fetchcode.vcs.pip._internal.utils import compatibility_tags -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import Dict, Iterable, List, Optional, Tuple - from fetchcode.vcs.pip._internal.req.req_install import InstallRequirement +from pip._vendor.packaging.utils import canonicalize_name +from pip._internal.req.req_install import InstallRequirement logger = logging.getLogger(__name__) -class RequirementSet(object): +class RequirementSet: + def __init__(self, check_supported_wheels: bool = True) -> None: + """Create a RequirementSet.""" - def __init__(self, check_supported_wheels=True): - # type: (bool) -> None - """Create a RequirementSet. - """ - - self.requirements = OrderedDict() # type: Dict[str, InstallRequirement] # noqa: E501 + self.requirements: Dict[str, InstallRequirement] = OrderedDict() self.check_supported_wheels = check_supported_wheels - self.unnamed_requirements = [] # type: List[InstallRequirement] + self.unnamed_requirements: List[InstallRequirement] = [] - def __str__(self): - # type: () -> str + def __str__(self) -> str: requirements = sorted( (req for req in self.requirements.values() if not req.comes_from), - key=lambda req: canonicalize_name(req.name), + key=lambda req: canonicalize_name(req.name or ""), ) - return ' '.join(str(req.req) for req in requirements) + return " ".join(str(req.req) for req in requirements) - def __repr__(self): - # type: () -> str + def __repr__(self) -> str: requirements = sorted( self.requirements.values(), - key=lambda req: canonicalize_name(req.name), + key=lambda req: canonicalize_name(req.name or ""), ) - format_string = '<{classname} object; {count} requirement(s): {reqs}>' + format_string = "<{classname} object; {count} requirement(s): {reqs}>" return format_string.format( classname=self.__class__.__name__, count=len(requirements), - reqs=', '.join(str(req.req) for req in requirements), + reqs=", ".join(str(req.req) for req in requirements), ) - def add_unnamed_requirement(self, install_req): - # type: (InstallRequirement) -> None + def add_unnamed_requirement(self, install_req: InstallRequirement) -> None: assert not install_req.name self.unnamed_requirements.append(install_req) - def add_named_requirement(self, install_req): - # type: (InstallRequirement) -> None + def add_named_requirement(self, install_req: InstallRequirement) -> None: assert install_req.name project_name = canonicalize_name(install_req.name) self.requirements[project_name] = install_req - def add_requirement( - self, - install_req, # type: InstallRequirement - parent_req_name=None, # type: Optional[str] - extras_requested=None # type: Optional[Iterable[str]] - ): - # type: (...) -> Tuple[List[InstallRequirement], Optional[InstallRequirement]] # noqa: E501 - """Add install_req as a requirement to install. - - :param parent_req_name: The name of the requirement that needed this - added. The name is used because when multiple unnamed requirements - resolve to the same name, we could otherwise end up with dependency - links that point outside the Requirements set. parent_req must - already be added. Note that None implies that this is a user - supplied requirement, vs an inferred one. - :param extras_requested: an iterable of extras used to evaluate the - environment markers. - :return: Additional requirements to scan. That is either [] if - the requirement is not applicable, or [install_req] if the - requirement is applicable and has just been added. - """ - # If the markers do not match, ignore this requirement. - if not install_req.match_markers(extras_requested): - logger.info( - "Ignoring %s: markers '%s' don't match your environment", - install_req.name, install_req.markers, - ) - return [], None - - # If the wheel is not supported, raise an error. - # Should check this after filtering out based on environment markers to - # allow specifying different wheels based on the environment/OS, in a - # single requirements file. - if install_req.link and install_req.link.is_wheel: - wheel = Wheel(install_req.link.filename) - tags = compatibility_tags.get_supported() - if (self.check_supported_wheels and not wheel.supported(tags)): - raise InstallationError( - "{} is not a supported wheel on this platform.".format( - wheel.filename) - ) - - # This next bit is really a sanity check. - assert install_req.is_direct == (parent_req_name is None), ( - "a direct req shouldn't have a parent and also, " - "a non direct req should have a parent" - ) - - # Unnamed requirements are scanned again and the requirement won't be - # added as a dependency until after scanning. - if not install_req.name: - self.add_unnamed_requirement(install_req) - return [install_req], None - - try: - existing_req = self.get_requirement(install_req.name) - except KeyError: - existing_req = None - - has_conflicting_requirement = ( - parent_req_name is None and - existing_req and - not existing_req.constraint and - existing_req.extras == install_req.extras and - existing_req.req.specifier != install_req.req.specifier - ) - if has_conflicting_requirement: - raise InstallationError( - "Double requirement given: {} (already in {}, name={!r})" - .format(install_req, existing_req, install_req.name) - ) - - # When no existing requirement exists, add the requirement as a - # dependency and it will be scanned again after. - if not existing_req: - self.add_named_requirement(install_req) - # We'd want to rescan this requirement later - return [install_req], install_req - - # Assume there's no need to scan, and that we've already - # encountered this for scanning. - if install_req.constraint or not existing_req.constraint: - return [], existing_req - - does_not_satisfy_constraint = ( - install_req.link and - not ( - existing_req.link and - install_req.link.path == existing_req.link.path - ) - ) - if does_not_satisfy_constraint: - raise InstallationError( - "Could not satisfy constraints for '{}': " - "installation from path or url cannot be " - "constrained to a version".format(install_req.name) - ) - # If we're now installing a constraint, mark the existing - # object for real installation. - existing_req.constraint = False - existing_req.extras = tuple(sorted( - set(existing_req.extras) | set(install_req.extras) - )) - logger.debug( - "Setting %s extras to: %s", - existing_req, existing_req.extras, - ) - # Return the existing requirement for addition to the parent and - # scanning again. - return [existing_req], existing_req - - def has_requirement(self, name): - # type: (str) -> bool + def has_requirement(self, name: str) -> bool: project_name = canonicalize_name(name) return ( - project_name in self.requirements and - not self.requirements[project_name].constraint + project_name in self.requirements + and not self.requirements[project_name].constraint ) - def get_requirement(self, name): - # type: (str) -> InstallRequirement + def get_requirement(self, name: str) -> InstallRequirement: project_name = canonicalize_name(name) if project_name in self.requirements: return self.requirements[project_name] - raise KeyError("No project with the name {name!r}".format(**locals())) + raise KeyError(f"No project with the name {name!r}") @property - def all_requirements(self): - # type: () -> List[InstallRequirement] + def all_requirements(self) -> List[InstallRequirement]: return self.unnamed_requirements + list(self.requirements.values()) + + @property + def requirements_to_install(self) -> List[InstallRequirement]: + """Return the list of requirements that need to be installed. + + TODO remove this property together with the legacy resolver, since the new + resolver only returns requirements that need to be installed. + """ + return [ + install_req + for install_req in self.all_requirements + if not install_req.constraint and not install_req.satisfied_by + ] diff --git a/src/fetchcode/vcs/pip/_internal/req/req_tracker.py b/src/fetchcode/vcs/pip/_internal/req/req_tracker.py deleted file mode 100644 index 83cb32d1..00000000 --- a/src/fetchcode/vcs/pip/_internal/req/req_tracker.py +++ /dev/null @@ -1,151 +0,0 @@ -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False - -from __future__ import absolute_import - -import contextlib -import errno -import hashlib -import logging -import os - -from fetchcode.vcs.pip._vendor import contextlib2 - -from fetchcode.vcs.pip._internal.utils.temp_dir import TempDirectory -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from types import TracebackType - from typing import Dict, Iterator, Optional, Set, Type, Union - from fetchcode.vcs.pip._internal.req.req_install import InstallRequirement - from fetchcode.vcs.pip._internal.models.link import Link - -logger = logging.getLogger(__name__) - - -@contextlib.contextmanager -def update_env_context_manager(**changes): - # type: (str) -> Iterator[None] - target = os.environ - - # Save values from the target and change them. - non_existent_marker = object() - saved_values = {} # type: Dict[str, Union[object, str]] - for name, new_value in changes.items(): - try: - saved_values[name] = target[name] - except KeyError: - saved_values[name] = non_existent_marker - target[name] = new_value - - try: - yield - finally: - # Restore original values in the target. - for name, original_value in saved_values.items(): - if original_value is non_existent_marker: - del target[name] - else: - assert isinstance(original_value, str) # for mypy - target[name] = original_value - - -@contextlib.contextmanager -def get_requirement_tracker(): - # type: () -> Iterator[RequirementTracker] - root = os.environ.get('PIP_REQ_TRACKER') - with contextlib2.ExitStack() as ctx: - if root is None: - root = ctx.enter_context( - TempDirectory(kind='req-tracker') - ).path - ctx.enter_context(update_env_context_manager(PIP_REQ_TRACKER=root)) - logger.debug("Initialized build tracking at %s", root) - - with RequirementTracker(root) as tracker: - yield tracker - - -class RequirementTracker(object): - - def __init__(self, root): - # type: (str) -> None - self._root = root - self._entries = set() # type: Set[InstallRequirement] - logger.debug("Created build tracker: %s", self._root) - - def __enter__(self): - # type: () -> RequirementTracker - logger.debug("Entered build tracker: %s", self._root) - return self - - def __exit__( - self, - exc_type, # type: Optional[Type[BaseException]] - exc_val, # type: Optional[BaseException] - exc_tb # type: Optional[TracebackType] - ): - # type: (...) -> None - self.cleanup() - - def _entry_path(self, link): - # type: (Link) -> str - hashed = hashlib.sha224(link.url_without_fragment.encode()).hexdigest() - return os.path.join(self._root, hashed) - - def add(self, req): - # type: (InstallRequirement) -> None - """Add an InstallRequirement to build tracking. - """ - - # Get the file to write information about this requirement. - entry_path = self._entry_path(req.link) - - # Try reading from the file. If it exists and can be read from, a build - # is already in progress, so a LookupError is raised. - try: - with open(entry_path) as fp: - contents = fp.read() - except IOError as e: - # if the error is anything other than "file does not exist", raise. - if e.errno != errno.ENOENT: - raise - else: - message = '{} is already being built: {}'.format( - req.link, contents) - raise LookupError(message) - - # If we're here, req should really not be building already. - assert req not in self._entries - - # Start tracking this requirement. - with open(entry_path, 'w') as fp: - fp.write(str(req)) - self._entries.add(req) - - logger.debug('Added %s to build tracker %r', req, self._root) - - def remove(self, req): - # type: (InstallRequirement) -> None - """Remove an InstallRequirement from build tracking. - """ - - # Delete the created file and the corresponding entries. - os.unlink(self._entry_path(req.link)) - self._entries.remove(req) - - logger.debug('Removed %s from build tracker %r', req, self._root) - - def cleanup(self): - # type: () -> None - for req in set(self._entries): - self.remove(req) - - logger.debug("Removed build tracker: %r", self._root) - - @contextlib.contextmanager - def track(self, req): - # type: (InstallRequirement) -> Iterator[None] - self.add(req) - yield - self.remove(req) diff --git a/src/fetchcode/vcs/pip/_internal/req/req_uninstall.py b/src/fetchcode/vcs/pip/_internal/req/req_uninstall.py index b10dc398..26df2084 100644 --- a/src/fetchcode/vcs/pip/_internal/req/req_uninstall.py +++ b/src/fetchcode/vcs/pip/_internal/req/req_uninstall.py @@ -1,79 +1,58 @@ -from __future__ import absolute_import - -import csv import functools -import logging import os import sys import sysconfig +from importlib.util import cache_from_source +from typing import Any, Callable, Dict, Generator, Iterable, List, Optional, Set, Tuple + +from pip._internal.exceptions import LegacyDistutilsInstall, UninstallMissingRecord +from pip._internal.locations import get_bin_prefix, get_bin_user +from pip._internal.metadata import BaseDistribution +from pip._internal.utils.compat import WINDOWS +from pip._internal.utils.egg_link import egg_link_path_from_location +from pip._internal.utils.logging import getLogger, indent_log +from pip._internal.utils.misc import ask, normalize_path, renames, rmtree +from pip._internal.utils.temp_dir import AdjacentTempDirectory, TempDirectory +from pip._internal.utils.virtualenv import running_under_virtualenv + +logger = getLogger(__name__) -from fetchcode.vcs.pip._vendor import pkg_resources - -from fetchcode.vcs.pip._internal.exceptions import UninstallationError -from fetchcode.vcs.pip._internal.locations import bin_py, bin_user -from fetchcode.vcs.pip._internal.utils.compat import WINDOWS, cache_from_source, uses_pycache -from fetchcode.vcs.pip._internal.utils.logging import indent_log -from fetchcode.vcs.pip._internal.utils.misc import ( - FakeFile, - ask, - dist_in_usersite, - dist_is_local, - egg_link_path, - is_local, - normalize_path, - renames, - rmtree, -) -from fetchcode.vcs.pip._internal.utils.temp_dir import AdjacentTempDirectory, TempDirectory -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import ( - Any, Callable, Dict, Iterable, Iterator, List, Optional, Set, Tuple, - ) - from fetchcode.vcs.pip._vendor.pkg_resources import Distribution - -logger = logging.getLogger(__name__) - - -def _script_names(dist, script_name, is_gui): - # type: (Distribution, str, bool) -> List[str] + +def _script_names( + bin_dir: str, script_name: str, is_gui: bool +) -> Generator[str, None, None]: """Create the fully qualified name of the files created by {console,gui}_scripts for the given ``dist``. Returns the list of file names """ - if dist_in_usersite(dist): - bin_dir = bin_user - else: - bin_dir = bin_py exe_name = os.path.join(bin_dir, script_name) - paths_to_remove = [exe_name] - if WINDOWS: - paths_to_remove.append(exe_name + '.exe') - paths_to_remove.append(exe_name + '.exe.manifest') - if is_gui: - paths_to_remove.append(exe_name + '-script.pyw') - else: - paths_to_remove.append(exe_name + '-script.py') - return paths_to_remove + yield exe_name + if not WINDOWS: + return + yield f"{exe_name}.exe" + yield f"{exe_name}.exe.manifest" + if is_gui: + yield f"{exe_name}-script.pyw" + else: + yield f"{exe_name}-script.py" -def _unique(fn): - # type: (Callable[..., Iterator[Any]]) -> Callable[..., Iterator[Any]] +def _unique( + fn: Callable[..., Generator[Any, None, None]] +) -> Callable[..., Generator[Any, None, None]]: @functools.wraps(fn) - def unique(*args, **kw): - # type: (Any, Any) -> Iterator[Any] - seen = set() # type: Set[Any] + def unique(*args: Any, **kw: Any) -> Generator[Any, None, None]: + seen: Set[Any] = set() for item in fn(*args, **kw): if item not in seen: seen.add(item) yield item + return unique @_unique -def uninstallation_paths(dist): - # type: (Distribution) -> Iterator[str] +def uninstallation_paths(dist: BaseDistribution) -> Generator[str, None, None]: """ Yield all the uninstallation paths for dist based on RECORD-without-.py[co] @@ -81,33 +60,43 @@ def uninstallation_paths(dist): the .pyc and .pyo in the same directory. UninstallPathSet.add() takes care of the __pycache__ .py[co]. + + If RECORD is not found, raises an error, + with possible information from the INSTALLER file. + + https://packaging.python.org/specifications/recording-installed-packages/ """ - r = csv.reader(FakeFile(dist.get_metadata_lines('RECORD'))) - for row in r: - path = os.path.join(dist.location, row[0]) + location = dist.location + assert location is not None, "not installed" + + entries = dist.iter_declared_entries() + if entries is None: + raise UninstallMissingRecord(distribution=dist) + + for entry in entries: + path = os.path.join(location, entry) yield path - if path.endswith('.py'): + if path.endswith(".py"): dn, fn = os.path.split(path) base = fn[:-3] - path = os.path.join(dn, base + '.pyc') + path = os.path.join(dn, base + ".pyc") yield path - path = os.path.join(dn, base + '.pyo') + path = os.path.join(dn, base + ".pyo") yield path -def compact(paths): - # type: (Iterable[str]) -> Set[str] +def compact(paths: Iterable[str]) -> Set[str]: """Compact a path set to contain the minimal number of paths necessary to contain all paths in the set. If /a/path/ and /a/path/to/a/file.txt are both in the set, leave only the shorter path.""" sep = os.path.sep - short_paths = set() # type: Set[str] + short_paths: Set[str] = set() for path in sorted(paths, key=len): should_skip = any( - path.startswith(shortpath.rstrip("*")) and - path[len(shortpath.rstrip("*").rstrip(sep))] == sep + path.startswith(shortpath.rstrip("*")) + and path[len(shortpath.rstrip("*").rstrip(sep))] == sep for shortpath in short_paths ) if not should_skip: @@ -115,36 +104,30 @@ def compact(paths): return short_paths -def compress_for_rename(paths): - # type: (Iterable[str]) -> Set[str] +def compress_for_rename(paths: Iterable[str]) -> Set[str]: """Returns a set containing the paths that need to be renamed. This set may include directories when the original sequence of paths included every file on disk. """ - case_map = dict((os.path.normcase(p), p) for p in paths) + case_map = {os.path.normcase(p): p for p in paths} remaining = set(case_map) - unchecked = sorted(set(os.path.split(p)[0] - for p in case_map.values()), key=len) - wildcards = set() # type: Set[str] + unchecked = sorted({os.path.split(p)[0] for p in case_map.values()}, key=len) + wildcards: Set[str] = set() - def norm_join(*a): - # type: (str) -> str + def norm_join(*a: str) -> str: return os.path.normcase(os.path.join(*a)) for root in unchecked: - if any(os.path.normcase(root).startswith(w) - for w in wildcards): + if any(os.path.normcase(root).startswith(w) for w in wildcards): # This directory has already been handled. continue - all_files = set() # type: Set[str] - all_subdirs = set() # type: Set[str] + all_files: Set[str] = set() + all_subdirs: Set[str] = set() for dirname, subdirs, files in os.walk(root): - all_subdirs.update(norm_join(root, dirname, d) - for d in subdirs) - all_files.update(norm_join(root, dirname, f) - for f in files) + all_subdirs.update(norm_join(root, dirname, d) for d in subdirs) + all_files.update(norm_join(root, dirname, f) for f in files) # If all the files we found are in our remaining set of files to # remove, then remove them from the latter set and add a wildcard # for the directory. @@ -155,8 +138,7 @@ def norm_join(*a): return set(map(case_map.__getitem__, remaining)) | wildcards -def compress_for_output_listing(paths): - # type: (Iterable[str]) -> Tuple[Set[str], Set[str]] +def compress_for_output_listing(paths: Iterable[str]) -> Tuple[Set[str], Set[str]]: """Returns a tuple of 2 sets of which paths to display to user The first set contains paths that would be deleted. Files of a package @@ -180,8 +162,7 @@ def compress_for_output_listing(paths): folders.add(os.path.dirname(path)) files.add(path) - # probably this one https://github.com/python/mypy/issues/390 - _normcased_files = set(map(os.path.normcase, files)) # type: ignore + _normcased_files = set(map(os.path.normcase, files)) folders = compact(folders) @@ -194,47 +175,45 @@ def compress_for_output_listing(paths): continue file_ = os.path.join(dirpath, fname) - if (os.path.isfile(file_) and - os.path.normcase(file_) not in _normcased_files): + if ( + os.path.isfile(file_) + and os.path.normcase(file_) not in _normcased_files + ): # We are skipping this file. Add it to the set. will_skip.add(file_) - will_remove = files | { - os.path.join(folder, "*") for folder in folders - } + will_remove = files | {os.path.join(folder, "*") for folder in folders} return will_remove, will_skip -class StashedUninstallPathSet(object): +class StashedUninstallPathSet: """A set of file rename operations to stash files while tentatively uninstalling them.""" - def __init__(self): - # type: () -> None + + def __init__(self) -> None: # Mapping from source file root to [Adjacent]TempDirectory # for files under that directory. - self._save_dirs = {} # type: Dict[str, TempDirectory] + self._save_dirs: Dict[str, TempDirectory] = {} # (old path, new path) tuples for each move that may need # to be undone. - self._moves = [] # type: List[Tuple[str, str]] + self._moves: List[Tuple[str, str]] = [] - def _get_directory_stash(self, path): - # type: (str) -> str + def _get_directory_stash(self, path: str) -> str: """Stashes a directory. Directories are stashed adjacent to their original location if possible, or else moved/copied into the user's temp dir.""" try: - save_dir = AdjacentTempDirectory(path) # type: TempDirectory + save_dir: TempDirectory = AdjacentTempDirectory(path) except OSError: save_dir = TempDirectory(kind="uninstall") self._save_dirs[os.path.normcase(path)] = save_dir return save_dir.path - def _get_file_stash(self, path): - # type: (str) -> str + def _get_file_stash(self, path: str) -> str: """Stashes a file. If no root has been provided, one will be created for the directory @@ -253,7 +232,7 @@ def _get_file_stash(self, path): else: # Did not find any suitable root head = os.path.dirname(path) - save_dir = TempDirectory(kind='uninstall') + save_dir = TempDirectory(kind="uninstall") self._save_dirs[head] = save_dir relpath = os.path.relpath(path, head) @@ -261,8 +240,7 @@ def _get_file_stash(self, path): return os.path.join(save_dir.path, relpath) return save_dir.path - def stash(self, path): - # type: (str) -> str + def stash(self, path: str) -> str: """Stashes the directory or file and returns its new location. Handle symlinks as files to avoid modifying the symlink targets. """ @@ -273,7 +251,7 @@ def stash(self, path): new_path = self._get_file_stash(path) self._moves.append((path, new_path)) - if (path_is_dir and os.path.isdir(new_path)): + if path_is_dir and os.path.isdir(new_path): # If we're moving a directory, we need to # remove the destination first or else it will be # moved to inside the existing directory. @@ -283,23 +261,21 @@ def stash(self, path): renames(path, new_path) return new_path - def commit(self): - # type: () -> None + def commit(self) -> None: """Commits the uninstall by removing stashed files.""" - for _, save_dir in self._save_dirs.items(): + for save_dir in self._save_dirs.values(): save_dir.cleanup() self._moves = [] self._save_dirs = {} - def rollback(self): - # type: () -> None + def rollback(self) -> None: """Undoes the uninstall by moving stashed files back.""" for p in self._moves: logger.info("Moving to %s\n from %s", *p) for new_path, path in self._moves: try: - logger.debug('Replacing %s from %s', new_path, path) + logger.debug("Replacing %s from %s", new_path, path) if os.path.isfile(new_path) or os.path.islink(new_path): os.unlink(new_path) elif os.path.isdir(new_path): @@ -312,100 +288,97 @@ def rollback(self): self.commit() @property - def can_rollback(self): - # type: () -> bool + def can_rollback(self) -> bool: return bool(self._moves) -class UninstallPathSet(object): +class UninstallPathSet: """A set of file paths to be removed in the uninstallation of a requirement.""" - def __init__(self, dist): - # type: (Distribution) -> None - self.paths = set() # type: Set[str] - self._refuse = set() # type: Set[str] - self.pth = {} # type: Dict[str, UninstallPthEntries] - self.dist = dist + + def __init__(self, dist: BaseDistribution) -> None: + self._paths: Set[str] = set() + self._refuse: Set[str] = set() + self._pth: Dict[str, UninstallPthEntries] = {} + self._dist = dist self._moved_paths = StashedUninstallPathSet() + # Create local cache of normalize_path results. Creating an UninstallPathSet + # can result in hundreds/thousands of redundant calls to normalize_path with + # the same args, which hurts performance. + self._normalize_path_cached = functools.lru_cache(normalize_path) - def _permitted(self, path): - # type: (str) -> bool + def _permitted(self, path: str) -> bool: """ Return True if the given path is one we are permitted to remove/modify, False otherwise. """ - return is_local(path) + # aka is_local, but caching normalized sys.prefix + if not running_under_virtualenv(): + return True + return path.startswith(self._normalize_path_cached(sys.prefix)) - def add(self, path): - # type: (str) -> None + def add(self, path: str) -> None: head, tail = os.path.split(path) # we normalize the head to resolve parent directory symlinks, but not # the tail, since we only want to uninstall symlinks, not their targets - path = os.path.join(normalize_path(head), os.path.normcase(tail)) + path = os.path.join(self._normalize_path_cached(head), os.path.normcase(tail)) if not os.path.exists(path): return if self._permitted(path): - self.paths.add(path) + self._paths.add(path) else: self._refuse.add(path) # __pycache__ files can show up after 'installed-files.txt' is created, # due to imports - if os.path.splitext(path)[1] == '.py' and uses_pycache: + if os.path.splitext(path)[1] == ".py": self.add(cache_from_source(path)) - def add_pth(self, pth_file, entry): - # type: (str, str) -> None - pth_file = normalize_path(pth_file) + def add_pth(self, pth_file: str, entry: str) -> None: + pth_file = self._normalize_path_cached(pth_file) if self._permitted(pth_file): - if pth_file not in self.pth: - self.pth[pth_file] = UninstallPthEntries(pth_file) - self.pth[pth_file].add(entry) + if pth_file not in self._pth: + self._pth[pth_file] = UninstallPthEntries(pth_file) + self._pth[pth_file].add(entry) else: self._refuse.add(pth_file) - def remove(self, auto_confirm=False, verbose=False): - # type: (bool, bool) -> None - """Remove paths in ``self.paths`` with confirmation (unless + def remove(self, auto_confirm: bool = False, verbose: bool = False) -> None: + """Remove paths in ``self._paths`` with confirmation (unless ``auto_confirm`` is True).""" - if not self.paths: + if not self._paths: logger.info( "Can't uninstall '%s'. No files were found to uninstall.", - self.dist.project_name, + self._dist.raw_name, ) return - dist_name_version = ( - self.dist.project_name + "-" + self.dist.version - ) - logger.info('Uninstalling %s:', dist_name_version) + dist_name_version = f"{self._dist.raw_name}-{self._dist.raw_version}" + logger.info("Uninstalling %s:", dist_name_version) with indent_log(): if auto_confirm or self._allowed_to_proceed(verbose): moved = self._moved_paths - for_rename = compress_for_rename(self.paths) + for_rename = compress_for_rename(self._paths) for path in sorted(compact(for_rename)): moved.stash(path) - logger.debug('Removing file or directory %s', path) + logger.verbose("Removing file or directory %s", path) - for pth in self.pth.values(): + for pth in self._pth.values(): pth.remove() - logger.info('Successfully uninstalled %s', dist_name_version) + logger.info("Successfully uninstalled %s", dist_name_version) - def _allowed_to_proceed(self, verbose): - # type: (bool) -> bool - """Display which files would be deleted and prompt for confirmation - """ + def _allowed_to_proceed(self, verbose: bool) -> bool: + """Display which files would be deleted and prompt for confirmation""" - def _display(msg, paths): - # type: (str, Iterable[str]) -> None + def _display(msg: str, paths: Iterable[str]) -> None: if not paths: return @@ -415,182 +388,200 @@ def _display(msg, paths): logger.info(path) if not verbose: - will_remove, will_skip = compress_for_output_listing(self.paths) + will_remove, will_skip = compress_for_output_listing(self._paths) else: # In verbose mode, display all the files that are going to be # deleted. - will_remove = set(self.paths) + will_remove = set(self._paths) will_skip = set() - _display('Would remove:', will_remove) - _display('Would not remove (might be manually added):', will_skip) - _display('Would not remove (outside of prefix):', self._refuse) + _display("Would remove:", will_remove) + _display("Would not remove (might be manually added):", will_skip) + _display("Would not remove (outside of prefix):", self._refuse) if verbose: - _display('Will actually move:', compress_for_rename(self.paths)) + _display("Will actually move:", compress_for_rename(self._paths)) - return ask('Proceed (y/n)? ', ('y', 'n')) == 'y' + return ask("Proceed (Y/n)? ", ("y", "n", "")) != "n" - def rollback(self): - # type: () -> None + def rollback(self) -> None: """Rollback the changes previously made by remove().""" if not self._moved_paths.can_rollback: logger.error( "Can't roll back %s; was not uninstalled", - self.dist.project_name, + self._dist.raw_name, ) return - logger.info('Rolling back uninstall of %s', self.dist.project_name) + logger.info("Rolling back uninstall of %s", self._dist.raw_name) self._moved_paths.rollback() - for pth in self.pth.values(): + for pth in self._pth.values(): pth.rollback() - def commit(self): - # type: () -> None + def commit(self) -> None: """Remove temporary save dir: rollback will no longer be possible.""" self._moved_paths.commit() @classmethod - def from_dist(cls, dist): - # type: (Distribution) -> UninstallPathSet - dist_path = normalize_path(dist.location) - if not dist_is_local(dist): + def from_dist(cls, dist: BaseDistribution) -> "UninstallPathSet": + dist_location = dist.location + info_location = dist.info_location + if dist_location is None: + logger.info( + "Not uninstalling %s since it is not installed", + dist.canonical_name, + ) + return cls(dist) + + normalized_dist_location = normalize_path(dist_location) + if not dist.local: logger.info( "Not uninstalling %s at %s, outside environment %s", - dist.key, - dist_path, + dist.canonical_name, + normalized_dist_location, sys.prefix, ) return cls(dist) - if dist_path in {p for p in {sysconfig.get_path("stdlib"), - sysconfig.get_path("platstdlib")} - if p}: + if normalized_dist_location in { + p + for p in {sysconfig.get_path("stdlib"), sysconfig.get_path("platstdlib")} + if p + }: logger.info( "Not uninstalling %s at %s, as it is in the standard library.", - dist.key, - dist_path, + dist.canonical_name, + normalized_dist_location, ) return cls(dist) paths_to_remove = cls(dist) - develop_egg_link = egg_link_path(dist) - develop_egg_link_egg_info = '{}.egg-info'.format( - pkg_resources.to_filename(dist.project_name)) - egg_info_exists = dist.egg_info and os.path.exists(dist.egg_info) - # Special case for distutils installed package - distutils_egg_info = getattr(dist._provider, 'path', None) + develop_egg_link = egg_link_path_from_location(dist.raw_name) + + # Distribution is installed with metadata in a "flat" .egg-info + # directory. This means it is not a modern .dist-info installation, an + # egg, or legacy editable. + setuptools_flat_installation = ( + dist.installed_with_setuptools_egg_info + and info_location is not None + and os.path.exists(info_location) + # If dist is editable and the location points to a ``.egg-info``, + # we are in fact in the legacy editable case. + and not info_location.endswith(f"{dist.setuptools_filename}.egg-info") + ) # Uninstall cases order do matter as in the case of 2 installs of the # same package, pip needs to uninstall the currently detected version - if (egg_info_exists and dist.egg_info.endswith('.egg-info') and - not dist.egg_info.endswith(develop_egg_link_egg_info)): - # if dist.egg_info.endswith(develop_egg_link_egg_info), we - # are in fact in the develop_egg_link case - paths_to_remove.add(dist.egg_info) - if dist.has_metadata('installed-files.txt'): - for installed_file in dist.get_metadata( - 'installed-files.txt').splitlines(): - path = os.path.normpath( - os.path.join(dist.egg_info, installed_file) - ) - paths_to_remove.add(path) + if setuptools_flat_installation: + if info_location is not None: + paths_to_remove.add(info_location) + installed_files = dist.iter_declared_entries() + if installed_files is not None: + for installed_file in installed_files: + paths_to_remove.add(os.path.join(dist_location, installed_file)) # FIXME: need a test for this elif block # occurs with --single-version-externally-managed/--record outside # of pip - elif dist.has_metadata('top_level.txt'): - if dist.has_metadata('namespace_packages.txt'): - namespaces = dist.get_metadata('namespace_packages.txt') - else: + elif dist.is_file("top_level.txt"): + try: + namespace_packages = dist.read_text("namespace_packages.txt") + except FileNotFoundError: namespaces = [] + else: + namespaces = namespace_packages.splitlines(keepends=False) for top_level_pkg in [ - p for p - in dist.get_metadata('top_level.txt').splitlines() - if p and p not in namespaces]: - path = os.path.join(dist.location, top_level_pkg) + p + for p in dist.read_text("top_level.txt").splitlines() + if p and p not in namespaces + ]: + path = os.path.join(dist_location, top_level_pkg) paths_to_remove.add(path) - paths_to_remove.add(path + '.py') - paths_to_remove.add(path + '.pyc') - paths_to_remove.add(path + '.pyo') - - elif distutils_egg_info: - raise UninstallationError( - "Cannot uninstall {!r}. It is a distutils installed project " - "and thus we cannot accurately determine which files belong " - "to it which would lead to only a partial uninstall.".format( - dist.project_name, - ) - ) + paths_to_remove.add(f"{path}.py") + paths_to_remove.add(f"{path}.pyc") + paths_to_remove.add(f"{path}.pyo") - elif dist.location.endswith('.egg'): + elif dist.installed_by_distutils: + raise LegacyDistutilsInstall(distribution=dist) + + elif dist.installed_as_egg: # package installed by easy_install # We cannot match on dist.egg_name because it can slightly vary # i.e. setuptools-0.6c11-py2.6.egg vs setuptools-0.6rc11-py2.6.egg - paths_to_remove.add(dist.location) - easy_install_egg = os.path.split(dist.location)[1] - easy_install_pth = os.path.join(os.path.dirname(dist.location), - 'easy-install.pth') - paths_to_remove.add_pth(easy_install_pth, './' + easy_install_egg) + paths_to_remove.add(dist_location) + easy_install_egg = os.path.split(dist_location)[1] + easy_install_pth = os.path.join( + os.path.dirname(dist_location), + "easy-install.pth", + ) + paths_to_remove.add_pth(easy_install_pth, "./" + easy_install_egg) - elif egg_info_exists and dist.egg_info.endswith('.dist-info'): + elif dist.installed_with_dist_info: for path in uninstallation_paths(dist): paths_to_remove.add(path) elif develop_egg_link: - # develop egg - with open(develop_egg_link, 'r') as fh: + # PEP 660 modern editable is handled in the ``.dist-info`` case + # above, so this only covers the setuptools-style editable. + with open(develop_egg_link) as fh: link_pointer = os.path.normcase(fh.readline().strip()) - assert (link_pointer == dist.location), ( - 'Egg-link {} does not match installed location of {} ' - '(at {})'.format( - link_pointer, dist.project_name, dist.location) + normalized_link_pointer = paths_to_remove._normalize_path_cached( + link_pointer + ) + assert os.path.samefile( + normalized_link_pointer, normalized_dist_location + ), ( + f"Egg-link {develop_egg_link} (to {link_pointer}) does not match " + f"installed location of {dist.raw_name} (at {dist_location})" ) paths_to_remove.add(develop_egg_link) - easy_install_pth = os.path.join(os.path.dirname(develop_egg_link), - 'easy-install.pth') - paths_to_remove.add_pth(easy_install_pth, dist.location) + easy_install_pth = os.path.join( + os.path.dirname(develop_egg_link), "easy-install.pth" + ) + paths_to_remove.add_pth(easy_install_pth, dist_location) else: logger.debug( - 'Not sure how to uninstall: %s - Check: %s', - dist, dist.location, + "Not sure how to uninstall: %s - Check: %s", + dist, + dist_location, ) + if dist.in_usersite: + bin_dir = get_bin_user() + else: + bin_dir = get_bin_prefix() + # find distutils scripts= scripts - if dist.has_metadata('scripts') and dist.metadata_isdir('scripts'): - for script in dist.metadata_listdir('scripts'): - if dist_in_usersite(dist): - bin_dir = bin_user - else: - bin_dir = bin_py + try: + for script in dist.iter_distutils_script_names(): paths_to_remove.add(os.path.join(bin_dir, script)) if WINDOWS: - paths_to_remove.add(os.path.join(bin_dir, script) + '.bat') - - # find console_scripts - _scripts_to_remove = [] - console_scripts = dist.get_entry_map(group='console_scripts') - for name in console_scripts.keys(): - _scripts_to_remove.extend(_script_names(dist, name, False)) - # find gui_scripts - gui_scripts = dist.get_entry_map(group='gui_scripts') - for name in gui_scripts.keys(): - _scripts_to_remove.extend(_script_names(dist, name, True)) - - for s in _scripts_to_remove: + paths_to_remove.add(os.path.join(bin_dir, f"{script}.bat")) + except (FileNotFoundError, NotADirectoryError): + pass + + # find console_scripts and gui_scripts + def iter_scripts_to_remove( + dist: BaseDistribution, + bin_dir: str, + ) -> Generator[str, None, None]: + for entry_point in dist.iter_entry_points(): + if entry_point.group == "console_scripts": + yield from _script_names(bin_dir, entry_point.name, False) + elif entry_point.group == "gui_scripts": + yield from _script_names(bin_dir, entry_point.name, True) + + for s in iter_scripts_to_remove(dist, bin_dir): paths_to_remove.add(s) return paths_to_remove -class UninstallPthEntries(object): - def __init__(self, pth_file): - # type: (str) -> None +class UninstallPthEntries: + def __init__(self, pth_file: str) -> None: self.file = pth_file - self.entries = set() # type: Set[str] - self._saved_lines = None # type: Optional[List[bytes]] + self.entries: Set[str] = set() + self._saved_lines: Optional[List[bytes]] = None - def add(self, entry): - # type: (str) -> None + def add(self, entry: str) -> None: entry = os.path.normcase(entry) # On Windows, os.path.normcase converts the entry to use # backslashes. This is correct for entries that describe absolute @@ -600,50 +591,43 @@ def add(self, entry): # treats non-absolute paths with drive letter markings like c:foo\bar # as absolute paths. It also does not recognize UNC paths if they don't # have more than "\\sever\share". Valid examples: "\\server\share\" or - # "\\server\share\folder". Python 2.7.8+ support UNC in splitdrive. + # "\\server\share\folder". if WINDOWS and not os.path.splitdrive(entry)[0]: - entry = entry.replace('\\', '/') + entry = entry.replace("\\", "/") self.entries.add(entry) - def remove(self): - # type: () -> None - logger.debug('Removing pth entries from %s:', self.file) + def remove(self) -> None: + logger.verbose("Removing pth entries from %s:", self.file) # If the file doesn't exist, log a warning and return if not os.path.isfile(self.file): - logger.warning( - "Cannot remove entries from nonexistent file {}".format( - self.file) - ) + logger.warning("Cannot remove entries from nonexistent file %s", self.file) return - with open(self.file, 'rb') as fh: + with open(self.file, "rb") as fh: # windows uses '\r\n' with py3k, but uses '\n' with py2.x lines = fh.readlines() self._saved_lines = lines - if any(b'\r\n' in line for line in lines): - endline = '\r\n' + if any(b"\r\n" in line for line in lines): + endline = "\r\n" else: - endline = '\n' + endline = "\n" # handle missing trailing newline if lines and not lines[-1].endswith(endline.encode("utf-8")): lines[-1] = lines[-1] + endline.encode("utf-8") for entry in self.entries: try: - logger.debug('Removing entry: %s', entry) + logger.verbose("Removing entry: %s", entry) lines.remove((entry + endline).encode("utf-8")) except ValueError: pass - with open(self.file, 'wb') as fh: + with open(self.file, "wb") as fh: fh.writelines(lines) - def rollback(self): - # type: () -> bool + def rollback(self) -> bool: if self._saved_lines is None: - logger.error( - 'Cannot roll back changes to %s, none were made', self.file - ) + logger.error("Cannot roll back changes to %s, none were made", self.file) return False - logger.debug('Rolling %s back to previous state', self.file) - with open(self.file, 'wb') as fh: + logger.debug("Rolling %s back to previous state", self.file) + with open(self.file, "wb") as fh: fh.writelines(self._saved_lines) return True diff --git a/src/fetchcode/vcs/pip/_internal/resolution/base.py b/src/fetchcode/vcs/pip/_internal/resolution/base.py index a7585397..42dade18 100644 --- a/src/fetchcode/vcs/pip/_internal/resolution/base.py +++ b/src/fetchcode/vcs/pip/_internal/resolution/base.py @@ -1,20 +1,20 @@ -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from typing import Callable, List, Optional -if MYPY_CHECK_RUNNING: - from typing import Callable, List - from fetchcode.vcs.pip._internal.req.req_install import InstallRequirement - from fetchcode.vcs.pip._internal.req.req_set import RequirementSet +from pip._internal.req.req_install import InstallRequirement +from pip._internal.req.req_set import RequirementSet - InstallRequirementProvider = Callable[ - [str, InstallRequirement], InstallRequirement - ] +InstallRequirementProvider = Callable[ + [str, Optional[InstallRequirement]], InstallRequirement +] -class BaseResolver(object): - def resolve(self, root_reqs, check_supported_wheels): - # type: (List[InstallRequirement], bool) -> RequirementSet +class BaseResolver: + def resolve( + self, root_reqs: List[InstallRequirement], check_supported_wheels: bool + ) -> RequirementSet: raise NotImplementedError() - def get_installation_order(self, req_set): - # type: (RequirementSet) -> List[InstallRequirement] + def get_installation_order( + self, req_set: RequirementSet + ) -> List[InstallRequirement]: raise NotImplementedError() diff --git a/src/fetchcode/vcs/pip/_internal/resolution/legacy/resolver.py b/src/fetchcode/vcs/pip/_internal/resolution/legacy/resolver.py index 5f6ca236..1dd0d704 100644 --- a/src/fetchcode/vcs/pip/_internal/resolution/legacy/resolver.py +++ b/src/fetchcode/vcs/pip/_internal/resolution/legacy/resolver.py @@ -10,57 +10,53 @@ a. "first found, wins" (where the order is breadth first) """ -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False -# mypy: disallow-untyped-defs=False - import logging import sys from collections import defaultdict from itertools import chain +from typing import DefaultDict, Iterable, List, Optional, Set, Tuple -from fetchcode.vcs.pip._vendor.packaging import specifiers +from pip._vendor.packaging import specifiers +from pip._vendor.packaging.requirements import Requirement -from fetchcode.vcs.pip._internal.exceptions import ( +from pip._internal.cache import WheelCache +from pip._internal.exceptions import ( BestVersionAlreadyInstalled, DistributionNotFound, HashError, HashErrors, + InstallationError, + NoneMetadataError, UnsupportedPythonVersion, ) -from fetchcode.vcs.pip._internal.req.req_set import RequirementSet -from fetchcode.vcs.pip._internal.resolution.base import BaseResolver -from fetchcode.vcs.pip._internal.utils.compatibility_tags import get_supported -from fetchcode.vcs.pip._internal.utils.logging import indent_log -from fetchcode.vcs.pip._internal.utils.misc import dist_in_usersite, normalize_version_info -from fetchcode.vcs.pip._internal.utils.packaging import ( - check_requires_python, - get_requires_python, +from pip._internal.index.package_finder import PackageFinder +from pip._internal.metadata import BaseDistribution +from pip._internal.models.link import Link +from pip._internal.models.wheel import Wheel +from pip._internal.operations.prepare import RequirementPreparer +from pip._internal.req.req_install import ( + InstallRequirement, + check_invalid_constraint_type, ) -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import DefaultDict, List, Optional, Set, Tuple - from fetchcode.vcs.pip._vendor import pkg_resources - - from fetchcode.vcs.pip._internal.cache import WheelCache - from fetchcode.vcs.pip._internal.distributions import AbstractDistribution - from fetchcode.vcs.pip._internal.index.package_finder import PackageFinder - from fetchcode.vcs.pip._internal.operations.prepare import RequirementPreparer - from fetchcode.vcs.pip._internal.req.req_install import InstallRequirement - from fetchcode.vcs.pip._internal.resolution.base import InstallRequirementProvider - - DiscoveredDependencies = DefaultDict[str, List[InstallRequirement]] +from pip._internal.req.req_set import RequirementSet +from pip._internal.resolution.base import BaseResolver, InstallRequirementProvider +from pip._internal.utils import compatibility_tags +from pip._internal.utils.compatibility_tags import get_supported +from pip._internal.utils.direct_url_helpers import direct_url_from_link +from pip._internal.utils.logging import indent_log +from pip._internal.utils.misc import normalize_version_info +from pip._internal.utils.packaging import check_requires_python logger = logging.getLogger(__name__) +DiscoveredDependencies = DefaultDict[Optional[str], List[InstallRequirement]] + def _check_dist_requires_python( - dist, # type: pkg_resources.Distribution - version_info, # type: Tuple[int, int, int] - ignore_requires_python=False, # type: bool -): - # type: (...) -> None + dist: BaseDistribution, + version_info: Tuple[int, int, int], + ignore_requires_python: bool = False, +) -> None: """ Check whether the given Python version is compatible with a distribution's "Requires-Python" value. @@ -73,34 +69,41 @@ def _check_dist_requires_python( :raises UnsupportedPythonVersion: When the given Python version isn't compatible. """ - requires_python = get_requires_python(dist) + # This idiosyncratically converts the SpecifierSet to str and let + # check_requires_python then parse it again into SpecifierSet. But this + # is the legacy resolver so I'm just not going to bother refactoring. + try: + requires_python = str(dist.requires_python) + except FileNotFoundError as e: + raise NoneMetadataError(dist, str(e)) try: is_compatible = check_requires_python( - requires_python, version_info=version_info, + requires_python, + version_info=version_info, ) except specifiers.InvalidSpecifier as exc: logger.warning( - "Package %r has an invalid Requires-Python: %s", - dist.project_name, exc, + "Package %r has an invalid Requires-Python: %s", dist.raw_name, exc ) return if is_compatible: return - version = '.'.join(map(str, version_info)) + version = ".".join(map(str, version_info)) if ignore_requires_python: logger.debug( - 'Ignoring failed Requires-Python check for package %r: ' - '%s not in %r', - dist.project_name, version, requires_python, + "Ignoring failed Requires-Python check for package %r: %s not in %r", + dist.raw_name, + version, + requires_python, ) return raise UnsupportedPythonVersion( - 'Package {!r} requires a different Python: {} not in {!r}'.format( - dist.project_name, version, requires_python, - )) + f"Package {dist.raw_name!r} requires a different Python: " + f"{version} not in {requires_python!r}" + ) class Resolver(BaseResolver): @@ -112,20 +115,19 @@ class Resolver(BaseResolver): def __init__( self, - preparer, # type: RequirementPreparer - finder, # type: PackageFinder - wheel_cache, # type: Optional[WheelCache] - make_install_req, # type: InstallRequirementProvider - use_user_site, # type: bool - ignore_dependencies, # type: bool - ignore_installed, # type: bool - ignore_requires_python, # type: bool - force_reinstall, # type: bool - upgrade_strategy, # type: str - py_version_info=None, # type: Optional[Tuple[int, ...]] - ): - # type: (...) -> None - super(Resolver, self).__init__() + preparer: RequirementPreparer, + finder: PackageFinder, + wheel_cache: Optional[WheelCache], + make_install_req: InstallRequirementProvider, + use_user_site: bool, + ignore_dependencies: bool, + ignore_installed: bool, + ignore_requires_python: bool, + force_reinstall: bool, + upgrade_strategy: str, + py_version_info: Optional[Tuple[int, ...]] = None, + ) -> None: + super().__init__() assert upgrade_strategy in self._allowed_strategies if py_version_info is None: @@ -147,11 +149,11 @@ def __init__( self.use_user_site = use_user_site self._make_install_req = make_install_req - self._discovered_dependencies = \ - defaultdict(list) # type: DiscoveredDependencies + self._discovered_dependencies: DiscoveredDependencies = defaultdict(list) - def resolve(self, root_reqs, check_supported_wheels): - # type: (List[InstallRequirement], bool) -> RequirementSet + def resolve( + self, root_reqs: List[InstallRequirement], check_supported_wheels: bool + ) -> RequirementSet: """Resolve what operations need to be done As a side-effect of this method, the packages (and their dependencies) @@ -162,19 +164,19 @@ def resolve(self, root_reqs, check_supported_wheels): possible to move the preparation to become a step separated from dependency resolution. """ - requirement_set = RequirementSet( - check_supported_wheels=check_supported_wheels - ) + requirement_set = RequirementSet(check_supported_wheels=check_supported_wheels) for req in root_reqs: - requirement_set.add_requirement(req) + if req.constraint: + check_invalid_constraint_type(req) + self._add_requirement_to_set(requirement_set, req) # Actually prepare the files, and collect any exceptions. Most hash # exceptions cannot be checked ahead of time, because # _populate_link() needs to be called before we can make decisions # based on link type. - discovered_reqs = [] # type: List[InstallRequirement] + discovered_reqs: List[InstallRequirement] = [] hash_errors = HashErrors() - for req in chain(root_reqs, discovered_reqs): + for req in chain(requirement_set.all_requirements, discovered_reqs): try: discovered_reqs.extend(self._resolve_one(requirement_set, req)) except HashError as exc: @@ -186,29 +188,144 @@ def resolve(self, root_reqs, check_supported_wheels): return requirement_set - def _is_upgrade_allowed(self, req): - # type: (InstallRequirement) -> bool + def _add_requirement_to_set( + self, + requirement_set: RequirementSet, + install_req: InstallRequirement, + parent_req_name: Optional[str] = None, + extras_requested: Optional[Iterable[str]] = None, + ) -> Tuple[List[InstallRequirement], Optional[InstallRequirement]]: + """Add install_req as a requirement to install. + + :param parent_req_name: The name of the requirement that needed this + added. The name is used because when multiple unnamed requirements + resolve to the same name, we could otherwise end up with dependency + links that point outside the Requirements set. parent_req must + already be added. Note that None implies that this is a user + supplied requirement, vs an inferred one. + :param extras_requested: an iterable of extras used to evaluate the + environment markers. + :return: Additional requirements to scan. That is either [] if + the requirement is not applicable, or [install_req] if the + requirement is applicable and has just been added. + """ + # If the markers do not match, ignore this requirement. + if not install_req.match_markers(extras_requested): + logger.info( + "Ignoring %s: markers '%s' don't match your environment", + install_req.name, + install_req.markers, + ) + return [], None + + # If the wheel is not supported, raise an error. + # Should check this after filtering out based on environment markers to + # allow specifying different wheels based on the environment/OS, in a + # single requirements file. + if install_req.link and install_req.link.is_wheel: + wheel = Wheel(install_req.link.filename) + tags = compatibility_tags.get_supported() + if requirement_set.check_supported_wheels and not wheel.supported(tags): + raise InstallationError( + f"{wheel.filename} is not a supported wheel on this platform." + ) + + # This next bit is really a sanity check. + assert ( + not install_req.user_supplied or parent_req_name is None + ), "a user supplied req shouldn't have a parent" + + # Unnamed requirements are scanned again and the requirement won't be + # added as a dependency until after scanning. + if not install_req.name: + requirement_set.add_unnamed_requirement(install_req) + return [install_req], None + + try: + existing_req: Optional[InstallRequirement] = ( + requirement_set.get_requirement(install_req.name) + ) + except KeyError: + existing_req = None + + has_conflicting_requirement = ( + parent_req_name is None + and existing_req + and not existing_req.constraint + and existing_req.extras == install_req.extras + and existing_req.req + and install_req.req + and existing_req.req.specifier != install_req.req.specifier + ) + if has_conflicting_requirement: + raise InstallationError( + f"Double requirement given: {install_req} " + f"(already in {existing_req}, name={install_req.name!r})" + ) + + # When no existing requirement exists, add the requirement as a + # dependency and it will be scanned again after. + if not existing_req: + requirement_set.add_named_requirement(install_req) + # We'd want to rescan this requirement later + return [install_req], install_req + + # Assume there's no need to scan, and that we've already + # encountered this for scanning. + if install_req.constraint or not existing_req.constraint: + return [], existing_req + + does_not_satisfy_constraint = install_req.link and not ( + existing_req.link and install_req.link.path == existing_req.link.path + ) + if does_not_satisfy_constraint: + raise InstallationError( + f"Could not satisfy constraints for '{install_req.name}': " + "installation from path or url cannot be " + "constrained to a version" + ) + # If we're now installing a constraint, mark the existing + # object for real installation. + existing_req.constraint = False + # If we're now installing a user supplied requirement, + # mark the existing object as such. + if install_req.user_supplied: + existing_req.user_supplied = True + existing_req.extras = tuple( + sorted(set(existing_req.extras) | set(install_req.extras)) + ) + logger.debug( + "Setting %s extras to: %s", + existing_req, + existing_req.extras, + ) + # Return the existing requirement for addition to the parent and + # scanning again. + return [existing_req], existing_req + + def _is_upgrade_allowed(self, req: InstallRequirement) -> bool: if self.upgrade_strategy == "to-satisfy-only": return False elif self.upgrade_strategy == "eager": return True else: assert self.upgrade_strategy == "only-if-needed" - return req.is_direct + return req.user_supplied or req.constraint - def _set_req_to_reinstall(self, req): - # type: (InstallRequirement) -> None + def _set_req_to_reinstall(self, req: InstallRequirement) -> None: """ Set a requirement to be installed. """ # Don't uninstall the conflict if doing a user install and the # conflict is not a user install. - if not self.use_user_site or dist_in_usersite(req.satisfied_by): + assert req.satisfied_by is not None + if not self.use_user_site or req.satisfied_by.in_usersite: req.should_reinstall = True req.satisfied_by = None - def _check_skip_installed(self, req_to_install): - # type: (InstallRequirement) -> Optional[str] + def _check_skip_installed( + self, req_to_install: InstallRequirement + ) -> Optional[str]: """Check if req_to_install should be skipped. This will check if the req is installed, and whether we should upgrade @@ -239,8 +356,8 @@ def _check_skip_installed(self, req_to_install): if not self._is_upgrade_allowed(req_to_install): if self.upgrade_strategy == "only-if-needed": - return 'already satisfied, skipping upgrade' - return 'already satisfied' + return "already satisfied, skipping upgrade" + return "already satisfied" # Check for the possibility of an upgrade. For link-based # requirements we have to pull the tree down and inspect to assess @@ -250,7 +367,7 @@ def _check_skip_installed(self, req_to_install): self.finder.find_requirement(req_to_install, upgrade=True) except BestVersionAlreadyInstalled: # Then the best version is installed. - return 'already up-to-date' + return "already up-to-date" except DistributionNotFound: # No distribution found, so we squash the error. It will # be raised later when we re-try later to do the install. @@ -260,8 +377,29 @@ def _check_skip_installed(self, req_to_install): self._set_req_to_reinstall(req_to_install) return None - def _populate_link(self, req): - # type: (InstallRequirement) -> None + def _find_requirement_link(self, req: InstallRequirement) -> Optional[Link]: + upgrade = self._is_upgrade_allowed(req) + best_candidate = self.finder.find_requirement(req, upgrade) + if not best_candidate: + return None + + # Log a warning per PEP 592 if necessary before returning. + link = best_candidate.link + if link.is_yanked: + reason = link.yanked_reason or "" + msg = ( + # Mark this as a unicode string to prevent + # "UnicodeEncodeError: 'ascii' codec can't encode character" + # in Python 2 when the reason contains non-ascii characters. + "The candidate selected for download or install is a " + f"yanked version: {best_candidate}\n" + f"Reason for being yanked: {reason}" + ) + logger.warning(msg) + + return link + + def _populate_link(self, req: InstallRequirement) -> None: """Ensure that if a link can be found for this, that it is found. Note that req.link may still be None - if the requirement is already @@ -274,25 +412,33 @@ def _populate_link(self, req): mismatches. Furthermore, cached wheels at present have undeterministic contents due to file modification times. """ - upgrade = self._is_upgrade_allowed(req) if req.link is None: - req.link = self.finder.find_requirement(req, upgrade) + req.link = self._find_requirement_link(req) if self.wheel_cache is None or self.preparer.require_hashes: return + + assert req.link is not None, "_find_requirement_link unexpectedly returned None" cache_entry = self.wheel_cache.get_cache_entry( link=req.link, package_name=req.name, supported_tags=get_supported(), ) if cache_entry is not None: - logger.debug('Using cached wheel link: %s', cache_entry.link) + logger.debug("Using cached wheel link: %s", cache_entry.link) if req.link is req.original_link and cache_entry.persistent: - req.original_link_is_in_wheel_cache = True + req.cached_wheel_source_link = req.link + if cache_entry.origin is not None: + req.download_info = cache_entry.origin + else: + # Legacy cache entry that does not have origin.json. + # download_info may miss the archive_info.hashes field. + req.download_info = direct_url_from_link( + req.link, link_is_in_wheel_cache=cache_entry.persistent + ) req.link = cache_entry.link - def _get_abstract_dist_for(self, req): - # type: (InstallRequirement) -> AbstractDistribution + def _get_dist_for(self, req: InstallRequirement) -> BaseDistribution: """Takes a InstallRequirement and returns a single AbstractDist \ representing a prepared variant of the same. """ @@ -305,13 +451,11 @@ def _get_abstract_dist_for(self, req): skip_reason = self._check_skip_installed(req) if req.satisfied_by: - return self.preparer.prepare_installed_requirement( - req, skip_reason - ) + return self.preparer.prepare_installed_requirement(req, skip_reason) # We eagerly populate the link, since that's our "legacy" behavior. self._populate_link(req) - abstract_dist = self.preparer.prepare_linked_requirement(req) + dist = self.preparer.prepare_linked_requirement(req) # NOTE # The following portion is for determining if a certain package is @@ -326,27 +470,25 @@ def _get_abstract_dist_for(self, req): if req.satisfied_by: should_modify = ( - self.upgrade_strategy != "to-satisfy-only" or - self.force_reinstall or - self.ignore_installed or - req.link.scheme == 'file' + self.upgrade_strategy != "to-satisfy-only" + or self.force_reinstall + or self.ignore_installed + or req.link.scheme == "file" ) if should_modify: self._set_req_to_reinstall(req) else: logger.info( - 'Requirement already satisfied (use --upgrade to upgrade):' - ' %s', req, + "Requirement already satisfied (use --upgrade to upgrade): %s", + req, ) - - return abstract_dist + return dist def _resolve_one( self, - requirement_set, # type: RequirementSet - req_to_install, # type: InstallRequirement - ): - # type: (...) -> List[InstallRequirement] + requirement_set: RequirementSet, + req_to_install: InstallRequirement, + ) -> List[InstallRequirement]: """Prepare a single requirements file. :return: A list of additional InstallRequirements to also install. @@ -359,79 +501,75 @@ def _resolve_one( req_to_install.prepared = True - abstract_dist = self._get_abstract_dist_for(req_to_install) - # Parse and return dependencies - dist = abstract_dist.get_pkg_resources_distribution() + dist = self._get_dist_for(req_to_install) # This will raise UnsupportedPythonVersion if the given Python # version isn't compatible with the distribution's Requires-Python. _check_dist_requires_python( - dist, version_info=self._py_version_info, + dist, + version_info=self._py_version_info, ignore_requires_python=self.ignore_requires_python, ) - more_reqs = [] # type: List[InstallRequirement] + more_reqs: List[InstallRequirement] = [] - def add_req(subreq, extras_requested): - sub_install_req = self._make_install_req( - str(subreq), - req_to_install, - ) + def add_req(subreq: Requirement, extras_requested: Iterable[str]) -> None: + # This idiosyncratically converts the Requirement to str and let + # make_install_req then parse it again into Requirement. But this is + # the legacy resolver so I'm just not going to bother refactoring. + sub_install_req = self._make_install_req(str(subreq), req_to_install) parent_req_name = req_to_install.name - to_scan_again, add_to_parent = requirement_set.add_requirement( + to_scan_again, add_to_parent = self._add_requirement_to_set( + requirement_set, sub_install_req, parent_req_name=parent_req_name, extras_requested=extras_requested, ) if parent_req_name and add_to_parent: - self._discovered_dependencies[parent_req_name].append( - add_to_parent - ) + self._discovered_dependencies[parent_req_name].append(add_to_parent) more_reqs.extend(to_scan_again) with indent_log(): # We add req_to_install before its dependencies, so that we # can refer to it when adding dependencies. + assert req_to_install.name is not None if not requirement_set.has_requirement(req_to_install.name): # 'unnamed' requirements will get added here # 'unnamed' requirements can only come from being directly # provided by the user. - assert req_to_install.is_direct - requirement_set.add_requirement( - req_to_install, parent_req_name=None, + assert req_to_install.user_supplied + self._add_requirement_to_set( + requirement_set, req_to_install, parent_req_name=None ) if not self.ignore_dependencies: if req_to_install.extras: logger.debug( "Installing extra requirements: %r", - ','.join(req_to_install.extras), + ",".join(req_to_install.extras), ) missing_requested = sorted( - set(req_to_install.extras) - set(dist.extras) + set(req_to_install.extras) - set(dist.iter_provided_extras()) ) for missing in missing_requested: logger.warning( - '%s does not provide the extra \'%s\'', - dist, missing + "%s %s does not provide the extra '%s'", + dist.raw_name, + dist.version, + missing, ) available_requested = sorted( - set(dist.extras) & set(req_to_install.extras) + set(dist.iter_provided_extras()) & set(req_to_install.extras) ) - for subreq in dist.requires(available_requested): + for subreq in dist.iter_dependencies(available_requested): add_req(subreq, extras_requested=available_requested) - if not req_to_install.editable and not req_to_install.satisfied_by: - # XXX: --no-install leads this to report 'Successfully - # downloaded' for only non-editable reqs, even though we took - # action on them. - req_to_install.successfully_downloaded = True - return more_reqs - def get_installation_order(self, req_set): - # type: (RequirementSet) -> List[InstallRequirement] + def get_installation_order( + self, req_set: RequirementSet + ) -> List[InstallRequirement]: """Create the installation order. The installation order is topological - requirements are installed @@ -442,9 +580,9 @@ def get_installation_order(self, req_set): # installs the user specified things in the order given, except when # dependencies must come earlier to achieve topological order. order = [] - ordered_reqs = set() # type: Set[InstallRequirement] + ordered_reqs: Set[InstallRequirement] = set() - def schedule(req): + def schedule(req: InstallRequirement) -> None: if req.satisfied_by or req in ordered_reqs: return if req.constraint: diff --git a/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/base.py b/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/base.py index 039a0c81..0f31dc9b 100644 --- a/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/base.py +++ b/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/base.py @@ -1,52 +1,139 @@ -from fetchcode.vcs.pip._vendor.packaging.utils import canonicalize_name +from dataclasses import dataclass +from typing import FrozenSet, Iterable, Optional, Tuple -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._vendor.packaging.specifiers import SpecifierSet +from pip._vendor.packaging.utils import NormalizedName +from pip._vendor.packaging.version import Version -if MYPY_CHECK_RUNNING: - from typing import Optional, Sequence, Set +from pip._internal.models.link import Link, links_equivalent +from pip._internal.req.req_install import InstallRequirement +from pip._internal.utils.hashes import Hashes - from fetchcode.vcs.pip._internal.req.req_install import InstallRequirement - from fetchcode.vcs.pip._vendor.packaging.version import _BaseVersion +CandidateLookup = Tuple[Optional["Candidate"], Optional[InstallRequirement]] -def format_name(project, extras): - # type: (str, Set[str]) -> str +def format_name(project: NormalizedName, extras: FrozenSet[NormalizedName]) -> str: if not extras: return project - canonical_extras = sorted(canonicalize_name(e) for e in extras) - return "{}[{}]".format(project, ",".join(canonical_extras)) + extras_expr = ",".join(sorted(extras)) + return f"{project}[{extras_expr}]" -class Requirement(object): +@dataclass(frozen=True) +class Constraint: + specifier: SpecifierSet + hashes: Hashes + links: FrozenSet[Link] + + @classmethod + def empty(cls) -> "Constraint": + return Constraint(SpecifierSet(), Hashes(), frozenset()) + + @classmethod + def from_ireq(cls, ireq: InstallRequirement) -> "Constraint": + links = frozenset([ireq.link]) if ireq.link else frozenset() + return Constraint(ireq.specifier, ireq.hashes(trust_internet=False), links) + + def __bool__(self) -> bool: + return bool(self.specifier) or bool(self.hashes) or bool(self.links) + + def __and__(self, other: InstallRequirement) -> "Constraint": + if not isinstance(other, InstallRequirement): + return NotImplemented + specifier = self.specifier & other.specifier + hashes = self.hashes & other.hashes(trust_internet=False) + links = self.links + if other.link: + links = links.union([other.link]) + return Constraint(specifier, hashes, links) + + def is_satisfied_by(self, candidate: "Candidate") -> bool: + # Reject if there are any mismatched URL constraints on this package. + if self.links and not all(_match_link(link, candidate) for link in self.links): + return False + # We can safely always allow prereleases here since PackageFinder + # already implements the prerelease logic, and would have filtered out + # prerelease candidates if the user does not expect them. + return self.specifier.contains(candidate.version, prereleases=True) + + +class Requirement: @property - def name(self): - # type: () -> str + def project_name(self) -> NormalizedName: + """The "project name" of a requirement. + + This is different from ``name`` if this requirement contains extras, + in which case ``name`` would contain the ``[...]`` part, while this + refers to the name of the project. + """ raise NotImplementedError("Subclass should override") - def find_matches(self): - # type: () -> Sequence[Candidate] + @property + def name(self) -> str: + """The name identifying this requirement in the resolver. + + This is different from ``project_name`` if this requirement contains + extras, where ``project_name`` would not contain the ``[...]`` part. + """ raise NotImplementedError("Subclass should override") - def is_satisfied_by(self, candidate): - # type: (Candidate) -> bool + def is_satisfied_by(self, candidate: "Candidate") -> bool: return False + def get_candidate_lookup(self) -> CandidateLookup: + raise NotImplementedError("Subclass should override") + + def format_for_error(self) -> str: + raise NotImplementedError("Subclass should override") + + +def _match_link(link: Link, candidate: "Candidate") -> bool: + if candidate.source_link: + return links_equivalent(link, candidate.source_link) + return False + -class Candidate(object): +class Candidate: @property - def name(self): - # type: () -> str + def project_name(self) -> NormalizedName: + """The "project name" of the candidate. + + This is different from ``name`` if this candidate contains extras, + in which case ``name`` would contain the ``[...]`` part, while this + refers to the name of the project. + """ + raise NotImplementedError("Override in subclass") + + @property + def name(self) -> str: + """The name identifying this candidate in the resolver. + + This is different from ``project_name`` if this candidate contains + extras, where ``project_name`` would not contain the ``[...]`` part. + """ + raise NotImplementedError("Override in subclass") + + @property + def version(self) -> Version: raise NotImplementedError("Override in subclass") @property - def version(self): - # type: () -> _BaseVersion + def is_installed(self) -> bool: raise NotImplementedError("Override in subclass") - def get_dependencies(self): - # type: () -> Sequence[Requirement] + @property + def is_editable(self) -> bool: + raise NotImplementedError("Override in subclass") + + @property + def source_link(self) -> Optional[Link]: raise NotImplementedError("Override in subclass") - def get_install_requirement(self): - # type: () -> Optional[InstallRequirement] + def iter_dependencies(self, with_requires: bool) -> Iterable[Optional[Requirement]]: raise NotImplementedError("Override in subclass") + + def get_install_requirement(self) -> Optional[InstallRequirement]: + raise NotImplementedError("Override in subclass") + + def format_for_error(self) -> str: + raise NotImplementedError("Subclass should override") diff --git a/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/candidates.py b/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/candidates.py index 98c896e9..d30d477b 100644 --- a/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/candidates.py +++ b/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/candidates.py @@ -1,323 +1,407 @@ import logging import sys +from typing import TYPE_CHECKING, Any, FrozenSet, Iterable, Optional, Tuple, Union, cast -from fetchcode.vcs.pip._vendor.packaging.specifiers import InvalidSpecifier, SpecifierSet -from fetchcode.vcs.pip._vendor.packaging.utils import canonicalize_name -from fetchcode.vcs.pip._vendor.packaging.version import Version +from pip._vendor.packaging.requirements import InvalidRequirement +from pip._vendor.packaging.utils import NormalizedName, canonicalize_name +from pip._vendor.packaging.version import Version -from fetchcode.vcs.pip._internal.req.constructors import ( +from pip._internal.exceptions import ( + HashError, + InstallationSubprocessError, + MetadataInconsistent, + MetadataInvalid, +) +from pip._internal.metadata import BaseDistribution +from pip._internal.models.link import Link, links_equivalent +from pip._internal.models.wheel import Wheel +from pip._internal.req.constructors import ( install_req_from_editable, install_req_from_line, ) -from fetchcode.vcs.pip._internal.req.req_install import InstallRequirement -from fetchcode.vcs.pip._internal.utils.misc import normalize_version_info -from fetchcode.vcs.pip._internal.utils.packaging import get_requires_python -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -from .base import Candidate, format_name - -if MYPY_CHECK_RUNNING: - from typing import Any, Optional, Sequence, Set, Tuple, Union +from pip._internal.req.req_install import InstallRequirement +from pip._internal.utils.direct_url_helpers import direct_url_from_link +from pip._internal.utils.misc import normalize_version_info - from fetchcode.vcs.pip._vendor.packaging.version import _BaseVersion - from fetchcode.vcs.pip._vendor.pkg_resources import Distribution +from .base import Candidate, Requirement, format_name - from fetchcode.vcs.pip._internal.distributions import AbstractDistribution - from fetchcode.vcs.pip._internal.models.link import Link - - from .base import Requirement +if TYPE_CHECKING: from .factory import Factory - BaseCandidate = Union[ - "AlreadyInstalledCandidate", - "EditableCandidate", - "LinkCandidate", - ] +logger = logging.getLogger(__name__) +BaseCandidate = Union[ + "AlreadyInstalledCandidate", + "EditableCandidate", + "LinkCandidate", +] -logger = logging.getLogger(__name__) +# Avoid conflicting with the PyPI package "Python". +REQUIRES_PYTHON_IDENTIFIER = cast(NormalizedName, "") -def make_install_req_from_link(link, parent): - # type: (Link, InstallRequirement) -> InstallRequirement - assert not parent.editable, "parent is editable" - return install_req_from_line( - link.url, - comes_from=parent.comes_from, - use_pep517=parent.use_pep517, - isolated=parent.isolated, - constraint=parent.constraint, - options=dict( - install_options=parent.install_options, - global_options=parent.global_options, - hashes=parent.hash_options - ), +def as_base_candidate(candidate: Candidate) -> Optional[BaseCandidate]: + """The runtime version of BaseCandidate.""" + base_candidate_classes = ( + AlreadyInstalledCandidate, + EditableCandidate, + LinkCandidate, ) + if isinstance(candidate, base_candidate_classes): + return candidate + return None + + +def make_install_req_from_link( + link: Link, template: InstallRequirement +) -> InstallRequirement: + assert not template.editable, "template is editable" + if template.req: + line = str(template.req) + else: + line = link.url + ireq = install_req_from_line( + line, + user_supplied=template.user_supplied, + comes_from=template.comes_from, + use_pep517=template.use_pep517, + isolated=template.isolated, + constraint=template.constraint, + global_options=template.global_options, + hash_options=template.hash_options, + config_settings=template.config_settings, + ) + ireq.original_link = template.original_link + ireq.link = link + ireq.extras = template.extras + return ireq -def make_install_req_from_editable(link, parent): - # type: (Link, InstallRequirement) -> InstallRequirement - assert parent.editable, "parent not editable" - return install_req_from_editable( +def make_install_req_from_editable( + link: Link, template: InstallRequirement +) -> InstallRequirement: + assert template.editable, "template not editable" + ireq = install_req_from_editable( link.url, - comes_from=parent.comes_from, - use_pep517=parent.use_pep517, - isolated=parent.isolated, - constraint=parent.constraint, - options=dict( - install_options=parent.install_options, - global_options=parent.global_options, - hashes=parent.hash_options - ), + user_supplied=template.user_supplied, + comes_from=template.comes_from, + use_pep517=template.use_pep517, + isolated=template.isolated, + constraint=template.constraint, + permit_editable_wheels=template.permit_editable_wheels, + global_options=template.global_options, + hash_options=template.hash_options, + config_settings=template.config_settings, ) + ireq.extras = template.extras + return ireq -def make_install_req_from_dist(dist, parent): - # type: (Distribution, InstallRequirement) -> InstallRequirement +def _make_install_req_from_dist( + dist: BaseDistribution, template: InstallRequirement +) -> InstallRequirement: + if template.req: + line = str(template.req) + elif template.link: + line = f"{dist.canonical_name} @ {template.link.url}" + else: + line = f"{dist.canonical_name}=={dist.version}" ireq = install_req_from_line( - "{}=={}".format( - canonicalize_name(dist.project_name), - dist.parsed_version, - ), - comes_from=parent.comes_from, - use_pep517=parent.use_pep517, - isolated=parent.isolated, - constraint=parent.constraint, - options=dict( - install_options=parent.install_options, - global_options=parent.global_options, - hashes=parent.hash_options - ), + line, + user_supplied=template.user_supplied, + comes_from=template.comes_from, + use_pep517=template.use_pep517, + isolated=template.isolated, + constraint=template.constraint, + global_options=template.global_options, + hash_options=template.hash_options, + config_settings=template.config_settings, ) ireq.satisfied_by = dist return ireq class _InstallRequirementBackedCandidate(Candidate): + """A candidate backed by an ``InstallRequirement``. + + This represents a package request with the target not being already + in the environment, and needs to be fetched and installed. The backing + ``InstallRequirement`` is responsible for most of the leg work; this + class exposes appropriate information to the resolver. + + :param link: The link passed to the ``InstallRequirement``. The backing + ``InstallRequirement`` will use this link to fetch the distribution. + :param source_link: The link this candidate "originates" from. This is + different from ``link`` when the link is found in the wheel cache. + ``link`` would point to the wheel cache, while this points to the + found remote link (e.g. from pypi.org). + """ + + dist: BaseDistribution + is_installed = False + def __init__( self, - link, # type: Link - ireq, # type: InstallRequirement - factory, # type: Factory - name=None, # type: Optional[str] - version=None, # type: Optional[_BaseVersion] - ): - # type: (...) -> None - self.link = link + link: Link, + source_link: Link, + ireq: InstallRequirement, + factory: "Factory", + name: Optional[NormalizedName] = None, + version: Optional[Version] = None, + ) -> None: + self._link = link + self._source_link = source_link self._factory = factory self._ireq = ireq self._name = name self._version = version - self._dist = None # type: Optional[Distribution] + self.dist = self._prepare() + self._hash: Optional[int] = None - def __repr__(self): - # type: () -> str - return "{class_name}({link!r})".format( - class_name=self.__class__.__name__, - link=str(self.link), - ) + def __str__(self) -> str: + return f"{self.name} {self.version}" - def __eq__(self, other): - # type: (Any) -> bool + def __repr__(self) -> str: + return f"{self.__class__.__name__}({str(self._link)!r})" + + def __hash__(self) -> int: + if self._hash is not None: + return self._hash + + self._hash = hash((self.__class__, self._link)) + return self._hash + + def __eq__(self, other: Any) -> bool: if isinstance(other, self.__class__): - return self.link == other.link + return links_equivalent(self._link, other._link) return False - # Needed for Python 2, which does not implement this by default - def __ne__(self, other): - # type: (Any) -> bool - return not self.__eq__(other) + @property + def source_link(self) -> Optional[Link]: + return self._source_link @property - def name(self): - # type: () -> str + def project_name(self) -> NormalizedName: """The normalised name of the project the candidate refers to""" if self._name is None: - self._name = canonicalize_name(self.dist.project_name) + self._name = self.dist.canonical_name return self._name @property - def version(self): - # type: () -> _BaseVersion + def name(self) -> str: + return self.project_name + + @property + def version(self) -> Version: if self._version is None: - self._version = self.dist.parsed_version + self._version = self.dist.version return self._version - def _prepare_abstract_distribution(self): - # type: () -> AbstractDistribution - raise NotImplementedError("Override in subclass") - - def _prepare(self): - # type: () -> None - if self._dist is not None: - return - - abstract_dist = self._prepare_abstract_distribution() - self._dist = abstract_dist.get_pkg_resources_distribution() - assert self._dist is not None, "Distribution already installed" - - # TODO: Abort cleanly here, as the resolution has been - # based on the wrong name/version until now, and - # so is wrong. - # TODO: (Longer term) Rather than abort, reject this candidate - # and backtrack. This would need resolvelib support. - # These should be "proper" errors, not just asserts, as they - # can result from user errors like a requirement "foo @ URL" - # when the project at URL has a name of "bar" in its metadata. - assert ( - self._name is None or - self._name == canonicalize_name(self._dist.project_name) - ), "Name mismatch: {!r} vs {!r}".format( - self._name, canonicalize_name(self._dist.project_name), - ) - assert ( - self._version is None or - self._version == self._dist.parsed_version - ), "Version mismatch: {!r} vs {!r}".format( - self._version, self._dist.parsed_version, + def format_for_error(self) -> str: + return ( + f"{self.name} {self.version} " + f"(from {self._link.file_path if self._link.is_file else self._link})" ) - @property - def dist(self): - # type: () -> Distribution - self._prepare() - return self._dist - - def _get_requires_python_specifier(self): - # type: () -> Optional[SpecifierSet] - requires_python = get_requires_python(self.dist) - if requires_python is None: - return None - try: - spec = SpecifierSet(requires_python) - except InvalidSpecifier as e: - logger.warning( - "Package %r has an invalid Requires-Python: %s", self.name, e, + def _prepare_distribution(self) -> BaseDistribution: + raise NotImplementedError("Override in subclass") + + def _check_metadata_consistency(self, dist: BaseDistribution) -> None: + """Check for consistency of project name and version of dist.""" + if self._name is not None and self._name != dist.canonical_name: + raise MetadataInconsistent( + self._ireq, + "name", + self._name, + dist.canonical_name, ) - return None - return spec - - def get_dependencies(self): - # type: () -> Sequence[Requirement] - deps = [ - self._factory.make_requirement_from_spec(str(r), self._ireq) - for r in self.dist.requires() - ] - python_dep = self._factory.make_requires_python_requirement( - self._get_requires_python_specifier(), - ) - if python_dep: - deps.append(python_dep) - return deps + if self._version is not None and self._version != dist.version: + raise MetadataInconsistent( + self._ireq, + "version", + str(self._version), + str(dist.version), + ) + # check dependencies are valid + # TODO performance: this means we iterate the dependencies at least twice, + # we may want to cache parsed Requires-Dist + try: + list(dist.iter_dependencies(list(dist.iter_provided_extras()))) + except InvalidRequirement as e: + raise MetadataInvalid(self._ireq, str(e)) - def get_install_requirement(self): - # type: () -> Optional[InstallRequirement] - self._prepare() + def _prepare(self) -> BaseDistribution: + try: + dist = self._prepare_distribution() + except HashError as e: + # Provide HashError the underlying ireq that caused it. This + # provides context for the resulting error message to show the + # offending line to the user. + e.req = self._ireq + raise + except InstallationSubprocessError as exc: + # The output has been presented already, so don't duplicate it. + exc.context = "See above for output." + raise + + self._check_metadata_consistency(dist) + return dist + + def iter_dependencies(self, with_requires: bool) -> Iterable[Optional[Requirement]]: + requires = self.dist.iter_dependencies() if with_requires else () + for r in requires: + yield from self._factory.make_requirements_from_spec(str(r), self._ireq) + yield self._factory.make_requires_python_requirement(self.dist.requires_python) + + def get_install_requirement(self) -> Optional[InstallRequirement]: return self._ireq class LinkCandidate(_InstallRequirementBackedCandidate): + is_editable = False + def __init__( self, - link, # type: Link - parent, # type: InstallRequirement - factory, # type: Factory - name=None, # type: Optional[str] - version=None, # type: Optional[_BaseVersion] - ): - # type: (...) -> None - super(LinkCandidate, self).__init__( + link: Link, + template: InstallRequirement, + factory: "Factory", + name: Optional[NormalizedName] = None, + version: Optional[Version] = None, + ) -> None: + source_link = link + cache_entry = factory.get_wheel_cache_entry(source_link, name) + if cache_entry is not None: + logger.debug("Using cached wheel link: %s", cache_entry.link) + link = cache_entry.link + ireq = make_install_req_from_link(link, template) + assert ireq.link == link + if ireq.link.is_wheel and not ireq.link.is_file: + wheel = Wheel(ireq.link.filename) + wheel_name = canonicalize_name(wheel.name) + assert name == wheel_name, f"{name!r} != {wheel_name!r} for wheel" + # Version may not be present for PEP 508 direct URLs + if version is not None: + wheel_version = Version(wheel.version) + assert ( + version == wheel_version + ), f"{version!r} != {wheel_version!r} for wheel {name}" + + if cache_entry is not None: + assert ireq.link.is_wheel + assert ireq.link.is_file + if cache_entry.persistent and template.link is template.original_link: + ireq.cached_wheel_source_link = source_link + if cache_entry.origin is not None: + ireq.download_info = cache_entry.origin + else: + # Legacy cache entry that does not have origin.json. + # download_info may miss the archive_info.hashes field. + ireq.download_info = direct_url_from_link( + source_link, link_is_in_wheel_cache=cache_entry.persistent + ) + + super().__init__( link=link, - ireq=make_install_req_from_link(link, parent), + source_link=source_link, + ireq=ireq, factory=factory, name=name, version=version, ) - def _prepare_abstract_distribution(self): - # type: () -> AbstractDistribution - return self._factory.preparer.prepare_linked_requirement(self._ireq) + def _prepare_distribution(self) -> BaseDistribution: + preparer = self._factory.preparer + return preparer.prepare_linked_requirement(self._ireq, parallel_builds=True) class EditableCandidate(_InstallRequirementBackedCandidate): + is_editable = True + def __init__( self, - link, # type: Link - parent, # type: InstallRequirement - factory, # type: Factory - name=None, # type: Optional[str] - version=None, # type: Optional[_BaseVersion] - ): - # type: (...) -> None - super(EditableCandidate, self).__init__( + link: Link, + template: InstallRequirement, + factory: "Factory", + name: Optional[NormalizedName] = None, + version: Optional[Version] = None, + ) -> None: + super().__init__( link=link, - ireq=make_install_req_from_editable(link, parent), + source_link=link, + ireq=make_install_req_from_editable(link, template), factory=factory, name=name, version=version, ) - def _prepare_abstract_distribution(self): - # type: () -> AbstractDistribution + def _prepare_distribution(self) -> BaseDistribution: return self._factory.preparer.prepare_editable_requirement(self._ireq) class AlreadyInstalledCandidate(Candidate): + is_installed = True + source_link = None + def __init__( self, - dist, # type: Distribution - parent, # type: InstallRequirement - factory, # type: Factory - ): - # type: (...) -> None + dist: BaseDistribution, + template: InstallRequirement, + factory: "Factory", + ) -> None: self.dist = dist - self._ireq = make_install_req_from_dist(dist, parent) + self._ireq = _make_install_req_from_dist(dist, template) self._factory = factory + self._version = None # This is just logging some messages, so we can do it eagerly. # The returned dist would be exactly the same as self.dist because we - # set satisfied_by in make_install_req_from_dist. + # set satisfied_by in _make_install_req_from_dist. # TODO: Supply reason based on force_reinstall and upgrade_strategy. skip_reason = "already satisfied" factory.preparer.prepare_installed_requirement(self._ireq, skip_reason) - def __repr__(self): - # type: () -> str - return "{class_name}({distribution!r})".format( - class_name=self.__class__.__name__, - distribution=self.dist, - ) + def __str__(self) -> str: + return str(self.dist) - def __eq__(self, other): - # type: (Any) -> bool - if isinstance(other, self.__class__): - return self.name == other.name and self.version == other.version - return False + def __repr__(self) -> str: + return f"{self.__class__.__name__}({self.dist!r})" + + def __eq__(self, other: object) -> bool: + if not isinstance(other, AlreadyInstalledCandidate): + return NotImplemented + return self.name == other.name and self.version == other.version + + def __hash__(self) -> int: + return hash((self.name, self.version)) - # Needed for Python 2, which does not implement this by default - def __ne__(self, other): - # type: (Any) -> bool - return not self.__eq__(other) + @property + def project_name(self) -> NormalizedName: + return self.dist.canonical_name + + @property + def name(self) -> str: + return self.project_name @property - def name(self): - # type: () -> str - return canonicalize_name(self.dist.project_name) + def version(self) -> Version: + if self._version is None: + self._version = self.dist.version + return self._version @property - def version(self): - # type: () -> _BaseVersion - return self.dist.parsed_version - - def get_dependencies(self): - # type: () -> Sequence[Requirement] - return [ - self._factory.make_requirement_from_spec(str(r), self._ireq) - for r in self.dist.requires() - ] - - def get_install_requirement(self): - # type: () -> Optional[InstallRequirement] + def is_editable(self) -> bool: + return self.dist.editable + + def format_for_error(self) -> str: + return f"{self.name} {self.version} (Installed)" + + def iter_dependencies(self, with_requires: bool) -> Iterable[Optional[Requirement]]: + if not with_requires: + return + for r in self.dist.iter_dependencies(): + yield from self._factory.make_requirements_from_spec(str(r), self._ireq) + + def get_install_requirement(self) -> Optional[InstallRequirement]: return None @@ -335,8 +419,8 @@ class ExtrasCandidate(Candidate): to treat it as a separate node in the dependency graph. 2. When we're getting the candidate's dependencies, a) We specify that we want the extra dependencies as well. - b) We add a dependency on the base candidate (matching the name and - version). See below for why this is needed. + b) We add a dependency on the base candidate. + See below for why this is needed. 3. We return None for the underlying InstallRequirement, as the base candidate will provide it, and we don't want to end up with duplicates. @@ -345,72 +429,100 @@ class ExtrasCandidate(Candidate): version 2.0. Having those candidates depend on foo=1.0 and foo=2.0 respectively forces the resolver to recognise that this is a conflict. """ + def __init__( self, - base, # type: BaseCandidate - extras, # type: Set[str] - ): - # type: (...) -> None + base: BaseCandidate, + extras: FrozenSet[str], + *, + comes_from: Optional[InstallRequirement] = None, + ) -> None: + """ + :param comes_from: the InstallRequirement that led to this candidate if it + differs from the base's InstallRequirement. This will often be the + case in the sense that this candidate's requirement has the extras + while the base's does not. Unlike the InstallRequirement backed + candidates, this requirement is used solely for reporting purposes, + it does not do any leg work. + """ self.base = base - self.extras = extras - - def __repr__(self): - # type: () -> str - return "{class_name}(base={base!r}, extras={extras!r})".format( - class_name=self.__class__.__name__, - base=self.base, - extras=self.extras, - ) + self.extras = frozenset(canonicalize_name(e) for e in extras) + self._comes_from = comes_from if comes_from is not None else self.base._ireq + + def __str__(self) -> str: + name, rest = str(self.base).split(" ", 1) + return "{}[{}] {}".format(name, ",".join(self.extras), rest) + + def __repr__(self) -> str: + return f"{self.__class__.__name__}(base={self.base!r}, extras={self.extras!r})" + + def __hash__(self) -> int: + return hash((self.base, self.extras)) - def __eq__(self, other): - # type: (Any) -> bool + def __eq__(self, other: Any) -> bool: if isinstance(other, self.__class__): return self.base == other.base and self.extras == other.extras return False - # Needed for Python 2, which does not implement this by default - def __ne__(self, other): - # type: (Any) -> bool - return not self.__eq__(other) + @property + def project_name(self) -> NormalizedName: + return self.base.project_name @property - def name(self): - # type: () -> str + def name(self) -> str: """The normalised name of the project the candidate refers to""" - return format_name(self.base.name, self.extras) + return format_name(self.base.project_name, self.extras) @property - def version(self): - # type: () -> _BaseVersion + def version(self) -> Version: return self.base.version - def get_dependencies(self): - # type: () -> Sequence[Requirement] + def format_for_error(self) -> str: + return "{} [{}]".format( + self.base.format_for_error(), ", ".join(sorted(self.extras)) + ) + + @property + def is_installed(self) -> bool: + return self.base.is_installed + + @property + def is_editable(self) -> bool: + return self.base.is_editable + + @property + def source_link(self) -> Optional[Link]: + return self.base.source_link + + def iter_dependencies(self, with_requires: bool) -> Iterable[Optional[Requirement]]: factory = self.base._factory + # Add a dependency on the exact base + # (See note 2b in the class docstring) + yield factory.make_requirement_from_candidate(self.base) + if not with_requires: + return + # The user may have specified extras that the candidate doesn't # support. We ignore any unsupported extras here. - valid_extras = self.extras.intersection(self.base.dist.extras) - invalid_extras = self.extras.difference(self.base.dist.extras) - if invalid_extras: + valid_extras = self.extras.intersection(self.base.dist.iter_provided_extras()) + invalid_extras = self.extras.difference(self.base.dist.iter_provided_extras()) + for extra in sorted(invalid_extras): logger.warning( - "Invalid extras specified in %s: %s", - self.name, - ','.join(sorted(invalid_extras)) + "%s %s does not provide the extra '%s'", + self.base.name, + self.version, + extra, ) - deps = [ - factory.make_requirement_from_spec(str(r), self.base._ireq) - for r in self.base.dist.requires(valid_extras) - ] - # Add a dependency on the exact base. - # (See note 2b in the class docstring) - spec = "{}=={}".format(self.base.name, self.base.version) - deps.append(factory.make_requirement_from_spec(spec, self.base._ireq)) - return deps + for r in self.base.dist.iter_dependencies(valid_extras): + yield from factory.make_requirements_from_spec( + str(r), + self._comes_from, + valid_extras, + ) - def get_install_requirement(self): - # type: () -> Optional[InstallRequirement] + def get_install_requirement(self) -> Optional[InstallRequirement]: # We don't return anything here, because we always # depend on the base candidate, and we'll get the # install requirement from that. @@ -418,8 +530,10 @@ def get_install_requirement(self): class RequiresPythonCandidate(Candidate): - def __init__(self, py_version_info): - # type: (Optional[Tuple[int, ...]]) -> None + is_installed = False + source_link = None + + def __init__(self, py_version_info: Optional[Tuple[int, ...]]) -> None: if py_version_info is not None: version_info = normalize_version_info(py_version_info) else: @@ -430,21 +544,26 @@ def __init__(self, py_version_info): # only one RequiresPythonCandidate in a resolution, i.e. the host Python. # The built-in object.__eq__() and object.__ne__() do exactly what we want. + def __str__(self) -> str: + return f"Python {self._version}" + + @property + def project_name(self) -> NormalizedName: + return REQUIRES_PYTHON_IDENTIFIER + @property - def name(self): - # type: () -> str - # Avoid conflicting with the PyPI package "Python". - return "" + def name(self) -> str: + return REQUIRES_PYTHON_IDENTIFIER @property - def version(self): - # type: () -> _BaseVersion + def version(self) -> Version: return self._version - def get_dependencies(self): - # type: () -> Sequence[Requirement] - return [] + def format_for_error(self) -> str: + return f"Python {self.version}" + + def iter_dependencies(self, with_requires: bool) -> Iterable[Optional[Requirement]]: + return () - def get_install_requirement(self): - # type: () -> Optional[InstallRequirement] + def get_install_requirement(self) -> Optional[InstallRequirement]: return None diff --git a/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/factory.py b/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/factory.py index 799458ac..145bdbf7 100644 --- a/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/factory.py +++ b/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/factory.py @@ -1,201 +1,817 @@ -from fetchcode.vcs.pip._vendor.packaging.utils import canonicalize_name +import contextlib +import functools +import logging +from typing import ( + TYPE_CHECKING, + Callable, + Dict, + FrozenSet, + Iterable, + Iterator, + List, + Mapping, + NamedTuple, + Optional, + Protocol, + Sequence, + Set, + Tuple, + TypeVar, + cast, +) + +from pip._vendor.packaging.requirements import InvalidRequirement +from pip._vendor.packaging.specifiers import SpecifierSet +from pip._vendor.packaging.utils import NormalizedName, canonicalize_name +from pip._vendor.packaging.version import Version +from pip._vendor.resolvelib import ResolutionImpossible -from fetchcode.vcs.pip._internal.exceptions import ( +from pip._internal.cache import CacheEntry, WheelCache +from pip._internal.exceptions import ( + DistributionNotFound, InstallationError, + MetadataInconsistent, + MetadataInvalid, UnsupportedPythonVersion, + UnsupportedWheel, +) +from pip._internal.index.package_finder import PackageFinder +from pip._internal.metadata import BaseDistribution, get_default_environment +from pip._internal.models.link import Link +from pip._internal.models.wheel import Wheel +from pip._internal.operations.prepare import RequirementPreparer +from pip._internal.req.constructors import ( + install_req_drop_extras, + install_req_from_link_and_ireq, +) +from pip._internal.req.req_install import ( + InstallRequirement, + check_invalid_constraint_type, ) -from fetchcode.vcs.pip._internal.utils.misc import get_installed_distributions -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._internal.resolution.base import InstallRequirementProvider +from pip._internal.utils.compatibility_tags import get_supported +from pip._internal.utils.hashes import Hashes +from pip._internal.utils.packaging import get_requirement +from pip._internal.utils.virtualenv import running_under_virtualenv +from .base import Candidate, Constraint, Requirement from .candidates import ( AlreadyInstalledCandidate, + BaseCandidate, EditableCandidate, ExtrasCandidate, LinkCandidate, RequiresPythonCandidate, + as_base_candidate, ) +from .found_candidates import FoundCandidates, IndexCandidateInfo from .requirements import ( ExplicitRequirement, RequiresPythonRequirement, SpecifierRequirement, + SpecifierWithoutExtrasRequirement, + UnsatisfiableRequirement, ) -if MYPY_CHECK_RUNNING: - from typing import Dict, Iterator, Optional, Set, Tuple, TypeVar +if TYPE_CHECKING: - from fetchcode.vcs.pip._vendor.packaging.specifiers import SpecifierSet - from fetchcode.vcs.pip._vendor.packaging.version import _BaseVersion - from fetchcode.vcs.pip._vendor.pkg_resources import Distribution - from fetchcode.vcs.pip._vendor.resolvelib import ResolutionImpossible + class ConflictCause(Protocol): + requirement: RequiresPythonRequirement + parent: Candidate - from fetchcode.vcs.pip._internal.index.package_finder import PackageFinder - from fetchcode.vcs.pip._internal.models.link import Link - from fetchcode.vcs.pip._internal.operations.prepare import RequirementPreparer - from fetchcode.vcs.pip._internal.req.req_install import InstallRequirement - from fetchcode.vcs.pip._internal.resolution.base import InstallRequirementProvider - from .base import Candidate, Requirement - from .candidates import BaseCandidate +logger = logging.getLogger(__name__) - C = TypeVar("C") - Cache = Dict[Link, C] +C = TypeVar("C") +Cache = Dict[Link, C] -class Factory(object): +class CollectedRootRequirements(NamedTuple): + requirements: List[Requirement] + constraints: Dict[str, Constraint] + user_requested: Dict[str, int] + + +class Factory: def __init__( self, - finder, # type: PackageFinder - preparer, # type: RequirementPreparer - make_install_req, # type: InstallRequirementProvider - force_reinstall, # type: bool - ignore_installed, # type: bool - ignore_requires_python, # type: bool - py_version_info=None, # type: Optional[Tuple[int, ...]] - ): - # type: (...) -> None - self.finder = finder + finder: PackageFinder, + preparer: RequirementPreparer, + make_install_req: InstallRequirementProvider, + wheel_cache: Optional[WheelCache], + use_user_site: bool, + force_reinstall: bool, + ignore_installed: bool, + ignore_requires_python: bool, + py_version_info: Optional[Tuple[int, ...]] = None, + ) -> None: + self._finder = finder self.preparer = preparer + self._wheel_cache = wheel_cache self._python_candidate = RequiresPythonCandidate(py_version_info) self._make_install_req_from_spec = make_install_req + self._use_user_site = use_user_site self._force_reinstall = force_reinstall self._ignore_requires_python = ignore_requires_python - self._link_candidate_cache = {} # type: Cache[LinkCandidate] - self._editable_candidate_cache = {} # type: Cache[EditableCandidate] + self._build_failures: Cache[InstallationError] = {} + self._link_candidate_cache: Cache[LinkCandidate] = {} + self._editable_candidate_cache: Cache[EditableCandidate] = {} + self._installed_candidate_cache: Dict[str, AlreadyInstalledCandidate] = {} + self._extras_candidate_cache: Dict[ + Tuple[int, FrozenSet[NormalizedName]], ExtrasCandidate + ] = {} + self._supported_tags_cache = get_supported() if not ignore_installed: + env = get_default_environment() self._installed_dists = { - canonicalize_name(dist.project_name): dist - for dist in get_installed_distributions() + dist.canonical_name: dist + for dist in env.iter_installed_distributions(local_only=False) } else: self._installed_dists = {} + @property + def force_reinstall(self) -> bool: + return self._force_reinstall + + def _fail_if_link_is_unsupported_wheel(self, link: Link) -> None: + if not link.is_wheel: + return + wheel = Wheel(link.filename) + if wheel.supported(self._finder.target_python.get_unsorted_tags()): + return + msg = f"{link.filename} is not a supported wheel on this platform." + raise UnsupportedWheel(msg) + + def _make_extras_candidate( + self, + base: BaseCandidate, + extras: FrozenSet[str], + *, + comes_from: Optional[InstallRequirement] = None, + ) -> ExtrasCandidate: + cache_key = (id(base), frozenset(canonicalize_name(e) for e in extras)) + try: + candidate = self._extras_candidate_cache[cache_key] + except KeyError: + candidate = ExtrasCandidate(base, extras, comes_from=comes_from) + self._extras_candidate_cache[cache_key] = candidate + return candidate + def _make_candidate_from_dist( self, - dist, # type: Distribution - extras, # type: Set[str] - parent, # type: InstallRequirement - ): - # type: (...) -> Candidate - base = AlreadyInstalledCandidate(dist, parent, factory=self) - if extras: - return ExtrasCandidate(base, extras) - return base + dist: BaseDistribution, + extras: FrozenSet[str], + template: InstallRequirement, + ) -> Candidate: + try: + base = self._installed_candidate_cache[dist.canonical_name] + except KeyError: + base = AlreadyInstalledCandidate(dist, template, factory=self) + self._installed_candidate_cache[dist.canonical_name] = base + if not extras: + return base + return self._make_extras_candidate(base, extras, comes_from=template) def _make_candidate_from_link( self, - link, # type: Link - extras, # type: Set[str] - parent, # type: InstallRequirement - name=None, # type: Optional[str] - version=None, # type: Optional[_BaseVersion] - ): - # type: (...) -> Candidate + link: Link, + extras: FrozenSet[str], + template: InstallRequirement, + name: Optional[NormalizedName], + version: Optional[Version], + ) -> Optional[Candidate]: + base: Optional[BaseCandidate] = self._make_base_candidate_from_link( + link, template, name, version + ) + if not extras or base is None: + return base + return self._make_extras_candidate(base, extras, comes_from=template) + + def _make_base_candidate_from_link( + self, + link: Link, + template: InstallRequirement, + name: Optional[NormalizedName], + version: Optional[Version], + ) -> Optional[BaseCandidate]: # TODO: Check already installed candidate, and use it if the link and # editable flag match. - if parent.editable: + + if link in self._build_failures: + # We already tried this candidate before, and it does not build. + # Don't bother trying again. + return None + + if template.editable: if link not in self._editable_candidate_cache: - self._editable_candidate_cache[link] = EditableCandidate( - link, parent, factory=self, name=name, version=version, - ) - base = self._editable_candidate_cache[link] # type: BaseCandidate + try: + self._editable_candidate_cache[link] = EditableCandidate( + link, + template, + factory=self, + name=name, + version=version, + ) + except (MetadataInconsistent, MetadataInvalid) as e: + logger.info( + "Discarding [blue underline]%s[/]: [yellow]%s[reset]", + link, + e, + extra={"markup": True}, + ) + self._build_failures[link] = e + return None + + return self._editable_candidate_cache[link] else: if link not in self._link_candidate_cache: - self._link_candidate_cache[link] = LinkCandidate( - link, parent, factory=self, name=name, version=version, + try: + self._link_candidate_cache[link] = LinkCandidate( + link, + template, + factory=self, + name=name, + version=version, + ) + except MetadataInconsistent as e: + logger.info( + "Discarding [blue underline]%s[/]: [yellow]%s[reset]", + link, + e, + extra={"markup": True}, + ) + self._build_failures[link] = e + return None + return self._link_candidate_cache[link] + + def _iter_found_candidates( + self, + ireqs: Sequence[InstallRequirement], + specifier: SpecifierSet, + hashes: Hashes, + prefers_installed: bool, + incompatible_ids: Set[int], + ) -> Iterable[Candidate]: + if not ireqs: + return () + + # The InstallRequirement implementation requires us to give it a + # "template". Here we just choose the first requirement to represent + # all of them. + # Hopefully the Project model can correct this mismatch in the future. + template = ireqs[0] + assert template.req, "Candidates found on index must be PEP 508" + name = canonicalize_name(template.req.name) + + extras: FrozenSet[str] = frozenset() + for ireq in ireqs: + assert ireq.req, "Candidates found on index must be PEP 508" + specifier &= ireq.req.specifier + hashes &= ireq.hashes(trust_internet=False) + extras |= frozenset(ireq.extras) + + def _get_installed_candidate() -> Optional[Candidate]: + """Get the candidate for the currently-installed version.""" + # If --force-reinstall is set, we want the version from the index + # instead, so we "pretend" there is nothing installed. + if self._force_reinstall: + return None + try: + installed_dist = self._installed_dists[name] + except KeyError: + return None + # Don't use the installed distribution if its version does not fit + # the current dependency graph. + if not specifier.contains(installed_dist.version, prereleases=True): + return None + candidate = self._make_candidate_from_dist( + dist=installed_dist, + extras=extras, + template=template, + ) + # The candidate is a known incompatibility. Don't use it. + if id(candidate) in incompatible_ids: + return None + return candidate + + def iter_index_candidate_infos() -> Iterator[IndexCandidateInfo]: + result = self._finder.find_best_candidate( + project_name=name, + specifier=specifier, + hashes=hashes, + ) + icans = list(result.iter_applicable()) + + # PEP 592: Yanked releases are ignored unless the specifier + # explicitly pins a version (via '==' or '===') that can be + # solely satisfied by a yanked release. + all_yanked = all(ican.link.is_yanked for ican in icans) + + def is_pinned(specifier: SpecifierSet) -> bool: + for sp in specifier: + if sp.operator == "===": + return True + if sp.operator != "==": + continue + if sp.version.endswith(".*"): + continue + return True + return False + + pinned = is_pinned(specifier) + + # PackageFinder returns earlier versions first, so we reverse. + for ican in reversed(icans): + if not (all_yanked and pinned) and ican.link.is_yanked: + continue + func = functools.partial( + self._make_candidate_from_link, + link=ican.link, + extras=extras, + template=template, + name=name, + version=ican.version, ) - base = self._link_candidate_cache[link] - if extras: - return ExtrasCandidate(base, extras) - return base - - def iter_found_candidates(self, ireq, extras): - # type: (InstallRequirement, Set[str]) -> Iterator[Candidate] - name = canonicalize_name(ireq.req.name) - if not self._force_reinstall: - installed_dist = self._installed_dists.get(name) - else: - installed_dist = None + yield ican.version, func - found = self.finder.find_best_candidate( - project_name=ireq.req.name, - specifier=ireq.req.specifier, - hashes=ireq.hashes(trust_internet=False), + return FoundCandidates( + iter_index_candidate_infos, + _get_installed_candidate(), + prefers_installed, + incompatible_ids, ) - for ican in found.iter_applicable(): - if (installed_dist is not None and - installed_dist.parsed_version == ican.version): + + def _iter_explicit_candidates_from_base( + self, + base_requirements: Iterable[Requirement], + extras: FrozenSet[str], + ) -> Iterator[Candidate]: + """Produce explicit candidates from the base given an extra-ed package. + + :param base_requirements: Requirements known to the resolver. The + requirements are guaranteed to not have extras. + :param extras: The extras to inject into the explicit requirements' + candidates. + """ + for req in base_requirements: + lookup_cand, _ = req.get_candidate_lookup() + if lookup_cand is None: # Not explicit. continue - yield self._make_candidate_from_link( - link=ican.link, - extras=extras, - parent=ireq, - name=name, - version=ican.version, + # We've stripped extras from the identifier, and should always + # get a BaseCandidate here, unless there's a bug elsewhere. + base_cand = as_base_candidate(lookup_cand) + assert base_cand is not None, "no extras here" + yield self._make_extras_candidate(base_cand, extras) + + def _iter_candidates_from_constraints( + self, + identifier: str, + constraint: Constraint, + template: InstallRequirement, + ) -> Iterator[Candidate]: + """Produce explicit candidates from constraints. + + This creates "fake" InstallRequirement objects that are basically clones + of what "should" be the template, but with original_link set to link. + """ + for link in constraint.links: + self._fail_if_link_is_unsupported_wheel(link) + candidate = self._make_base_candidate_from_link( + link, + template=install_req_from_link_and_ireq(link, template), + name=canonicalize_name(identifier), + version=None, ) + if candidate: + yield candidate - # Return installed distribution if it matches the specifier. This is - # done last so the resolver will prefer it over downloading links. - if (installed_dist is not None and - installed_dist.parsed_version in ireq.req.specifier): - yield self._make_candidate_from_dist( - dist=installed_dist, - extras=extras, - parent=ireq, + def find_candidates( + self, + identifier: str, + requirements: Mapping[str, Iterable[Requirement]], + incompatibilities: Mapping[str, Iterator[Candidate]], + constraint: Constraint, + prefers_installed: bool, + is_satisfied_by: Callable[[Requirement, Candidate], bool], + ) -> Iterable[Candidate]: + # Collect basic lookup information from the requirements. + explicit_candidates: Set[Candidate] = set() + ireqs: List[InstallRequirement] = [] + for req in requirements[identifier]: + cand, ireq = req.get_candidate_lookup() + if cand is not None: + explicit_candidates.add(cand) + if ireq is not None: + ireqs.append(ireq) + + # If the current identifier contains extras, add requires and explicit + # candidates from entries from extra-less identifier. + with contextlib.suppress(InvalidRequirement): + parsed_requirement = get_requirement(identifier) + if parsed_requirement.name != identifier: + explicit_candidates.update( + self._iter_explicit_candidates_from_base( + requirements.get(parsed_requirement.name, ()), + frozenset(parsed_requirement.extras), + ), + ) + for req in requirements.get(parsed_requirement.name, []): + _, ireq = req.get_candidate_lookup() + if ireq is not None: + ireqs.append(ireq) + + # Add explicit candidates from constraints. We only do this if there are + # known ireqs, which represent requirements not already explicit. If + # there are no ireqs, we're constraining already-explicit requirements, + # which is handled later when we return the explicit candidates. + if ireqs: + try: + explicit_candidates.update( + self._iter_candidates_from_constraints( + identifier, + constraint, + template=ireqs[0], + ), + ) + except UnsupportedWheel: + # If we're constrained to install a wheel incompatible with the + # target architecture, no candidates will ever be valid. + return () + + # Since we cache all the candidates, incompatibility identification + # can be made quicker by comparing only the id() values. + incompat_ids = {id(c) for c in incompatibilities.get(identifier, ())} + + # If none of the requirements want an explicit candidate, we can ask + # the finder for candidates. + if not explicit_candidates: + return self._iter_found_candidates( + ireqs, + constraint.specifier, + constraint.hashes, + prefers_installed, + incompat_ids, ) - def make_requirement_from_install_req(self, ireq): - # type: (InstallRequirement) -> Requirement - if ireq.link: - # TODO: Get name and version from ireq, if possible? - # Specifically, this might be needed in "name @ URL" - # syntax - need to check where that syntax is handled. - cand = self._make_candidate_from_link( - ireq.link, extras=set(), parent=ireq, + return ( + c + for c in explicit_candidates + if id(c) not in incompat_ids + and constraint.is_satisfied_by(c) + and all(is_satisfied_by(req, c) for req in requirements[identifier]) + ) + + def _make_requirements_from_install_req( + self, ireq: InstallRequirement, requested_extras: Iterable[str] + ) -> Iterator[Requirement]: + """ + Returns requirement objects associated with the given InstallRequirement. In + most cases this will be a single object but the following special cases exist: + - the InstallRequirement has markers that do not apply -> result is empty + - the InstallRequirement has both a constraint (or link) and extras + -> result is split in two requirement objects: one with the constraint + (or link) and one with the extra. This allows centralized constraint + handling for the base, resulting in fewer candidate rejections. + """ + if not ireq.match_markers(requested_extras): + logger.info( + "Ignoring %s: markers '%s' don't match your environment", + ireq.name, + ireq.markers, + ) + elif not ireq.link: + if ireq.extras and ireq.req is not None and ireq.req.specifier: + yield SpecifierWithoutExtrasRequirement(ireq) + yield SpecifierRequirement(ireq) + else: + self._fail_if_link_is_unsupported_wheel(ireq.link) + # Always make the link candidate for the base requirement to make it + # available to `find_candidates` for explicit candidate lookup for any + # set of extras. + # The extras are required separately via a second requirement. + cand = self._make_base_candidate_from_link( + ireq.link, + template=install_req_drop_extras(ireq) if ireq.extras else ireq, + name=canonicalize_name(ireq.name) if ireq.name else None, + version=None, ) - return ExplicitRequirement(cand) - return SpecifierRequirement(ireq, factory=self) + if cand is None: + # There's no way we can satisfy a URL requirement if the underlying + # candidate fails to build. An unnamed URL must be user-supplied, so + # we fail eagerly. If the URL is named, an unsatisfiable requirement + # can make the resolver do the right thing, either backtrack (and + # maybe find some other requirement that's buildable) or raise a + # ResolutionImpossible eventually. + if not ireq.name: + raise self._build_failures[ireq.link] + yield UnsatisfiableRequirement(canonicalize_name(ireq.name)) + else: + # require the base from the link + yield self.make_requirement_from_candidate(cand) + if ireq.extras: + # require the extras on top of the base candidate + yield self.make_requirement_from_candidate( + self._make_extras_candidate(cand, frozenset(ireq.extras)) + ) - def make_requirement_from_spec(self, specifier, comes_from): - # type: (str, InstallRequirement) -> Requirement + def collect_root_requirements( + self, root_ireqs: List[InstallRequirement] + ) -> CollectedRootRequirements: + collected = CollectedRootRequirements([], {}, {}) + for i, ireq in enumerate(root_ireqs): + if ireq.constraint: + # Ensure we only accept valid constraints + problem = check_invalid_constraint_type(ireq) + if problem: + raise InstallationError(problem) + if not ireq.match_markers(): + continue + assert ireq.name, "Constraint must be named" + name = canonicalize_name(ireq.name) + if name in collected.constraints: + collected.constraints[name] &= ireq + else: + collected.constraints[name] = Constraint.from_ireq(ireq) + else: + reqs = list( + self._make_requirements_from_install_req( + ireq, + requested_extras=(), + ) + ) + if not reqs: + continue + template = reqs[0] + if ireq.user_supplied and template.name not in collected.user_requested: + collected.user_requested[template.name] = i + collected.requirements.extend(reqs) + # Put requirements with extras at the end of the root requires. This does not + # affect resolvelib's picking preference but it does affect its initial criteria + # population: by putting extras at the end we enable the candidate finder to + # present resolvelib with a smaller set of candidates to resolvelib, already + # taking into account any non-transient constraints on the associated base. This + # means resolvelib will have fewer candidates to visit and reject. + # Python's list sort is stable, meaning relative order is kept for objects with + # the same key. + collected.requirements.sort(key=lambda r: r.name != r.project_name) + return collected + + def make_requirement_from_candidate( + self, candidate: Candidate + ) -> ExplicitRequirement: + return ExplicitRequirement(candidate) + + def make_requirements_from_spec( + self, + specifier: str, + comes_from: Optional[InstallRequirement], + requested_extras: Iterable[str] = (), + ) -> Iterator[Requirement]: + """ + Returns requirement objects associated with the given specifier. In most cases + this will be a single object but the following special cases exist: + - the specifier has markers that do not apply -> result is empty + - the specifier has both a constraint and extras -> result is split + in two requirement objects: one with the constraint and one with the + extra. This allows centralized constraint handling for the base, + resulting in fewer candidate rejections. + """ ireq = self._make_install_req_from_spec(specifier, comes_from) - return self.make_requirement_from_install_req(ireq) + return self._make_requirements_from_install_req(ireq, requested_extras) - def make_requires_python_requirement(self, specifier): - # type: (Optional[SpecifierSet]) -> Optional[Requirement] - if self._ignore_requires_python or specifier is None: + def make_requires_python_requirement( + self, + specifier: SpecifierSet, + ) -> Optional[Requirement]: + if self._ignore_requires_python: + return None + # Don't bother creating a dependency for an empty Requires-Python. + if not str(specifier): return None return RequiresPythonRequirement(specifier, self._python_candidate) - def should_reinstall(self, candidate): - # type: (Candidate) -> bool + def get_wheel_cache_entry( + self, link: Link, name: Optional[str] + ) -> Optional[CacheEntry]: + """Look up the link in the wheel cache. + + If ``preparer.require_hashes`` is True, don't use the wheel cache, + because cached wheels, always built locally, have different hashes + than the files downloaded from the index server and thus throw false + hash mismatches. Furthermore, cached wheels at present have + nondeterministic contents due to file modification times. + """ + if self._wheel_cache is None: + return None + return self._wheel_cache.get_cache_entry( + link=link, + package_name=name, + supported_tags=self._supported_tags_cache, + ) + + def get_dist_to_uninstall(self, candidate: Candidate) -> Optional[BaseDistribution]: # TODO: Are there more cases this needs to return True? Editable? - return candidate.name in self._installed_dists + dist = self._installed_dists.get(candidate.project_name) + if dist is None: # Not installed, no uninstallation required. + return None + + # We're installing into global site. The current installation must + # be uninstalled, no matter it's in global or user site, because the + # user site installation has precedence over global. + if not self._use_user_site: + return dist + + # We're installing into user site. Remove the user site installation. + if dist.in_usersite: + return dist + + # We're installing into user site, but the installed incompatible + # package is in global site. We can't uninstall that, and would let + # the new user installation to "shadow" it. But shadowing won't work + # in virtual environments, so we error out. + if running_under_virtualenv() and dist.in_site_packages: + message = ( + f"Will not install to the user site because it will lack " + f"sys.path precedence to {dist.raw_name} in {dist.location}" + ) + raise InstallationError(message) + return None def _report_requires_python_error( + self, causes: Sequence["ConflictCause"] + ) -> UnsupportedPythonVersion: + assert causes, "Requires-Python error reported with no cause" + + version = self._python_candidate.version + + if len(causes) == 1: + specifier = str(causes[0].requirement.specifier) + message = ( + f"Package {causes[0].parent.name!r} requires a different " + f"Python: {version} not in {specifier!r}" + ) + return UnsupportedPythonVersion(message) + + message = f"Packages require a different Python. {version} not in:" + for cause in causes: + package = cause.parent.format_for_error() + specifier = str(cause.requirement.specifier) + message += f"\n{specifier!r} (required by {package})" + return UnsupportedPythonVersion(message) + + def _report_single_requirement_conflict( + self, req: Requirement, parent: Optional[Candidate] + ) -> DistributionNotFound: + if parent is None: + req_disp = str(req) + else: + req_disp = f"{req} (from {parent.name})" + + cands = self._finder.find_all_candidates(req.project_name) + skipped_by_requires_python = self._finder.requires_python_skipped_reasons() + + versions_set: Set[Version] = set() + yanked_versions_set: Set[Version] = set() + for c in cands: + is_yanked = c.link.is_yanked if c.link else False + if is_yanked: + yanked_versions_set.add(c.version) + else: + versions_set.add(c.version) + + versions = [str(v) for v in sorted(versions_set)] + yanked_versions = [str(v) for v in sorted(yanked_versions_set)] + + if yanked_versions: + # Saying "version X is yanked" isn't entirely accurate. + # https://github.com/pypa/pip/issues/11745#issuecomment-1402805842 + logger.critical( + "Ignored the following yanked versions: %s", + ", ".join(yanked_versions) or "none", + ) + if skipped_by_requires_python: + logger.critical( + "Ignored the following versions that require a different python " + "version: %s", + "; ".join(skipped_by_requires_python) or "none", + ) + logger.critical( + "Could not find a version that satisfies the requirement %s " + "(from versions: %s)", + req_disp, + ", ".join(versions) or "none", + ) + if str(req) == "requirements.txt": + logger.info( + "HINT: You are attempting to install a package literally " + 'named "requirements.txt" (which cannot exist). Consider ' + "using the '-r' flag to install the packages listed in " + "requirements.txt" + ) + + return DistributionNotFound(f"No matching distribution found for {req}") + + def get_installation_error( self, - requirement, # type: RequiresPythonRequirement - parent, # type: Candidate - ): - # type: (...) -> UnsupportedPythonVersion - template = ( - "Package {package!r} requires a different Python: " - "{version} not in {specifier!r}" + e: "ResolutionImpossible[Requirement, Candidate]", + constraints: Dict[str, Constraint], + ) -> InstallationError: + assert e.causes, "Installation error reported with no cause" + + # If one of the things we can't solve is "we need Python X.Y", + # that is what we report. + requires_python_causes = [ + cause + for cause in e.causes + if isinstance(cause.requirement, RequiresPythonRequirement) + and not cause.requirement.is_satisfied_by(self._python_candidate) + ] + if requires_python_causes: + # The comprehension above makes sure all Requirement instances are + # RequiresPythonRequirement, so let's cast for convenience. + return self._report_requires_python_error( + cast("Sequence[ConflictCause]", requires_python_causes), + ) + + # Otherwise, we have a set of causes which can't all be satisfied + # at once. + + # The simplest case is when we have *one* cause that can't be + # satisfied. We just report that case. + if len(e.causes) == 1: + req, parent = e.causes[0] + if req.name not in constraints: + return self._report_single_requirement_conflict(req, parent) + + # OK, we now have a list of requirements that can't all be + # satisfied at once. + + # A couple of formatting helpers + def text_join(parts: List[str]) -> str: + if len(parts) == 1: + return parts[0] + + return ", ".join(parts[:-1]) + " and " + parts[-1] + + def describe_trigger(parent: Candidate) -> str: + ireq = parent.get_install_requirement() + if not ireq or not ireq.comes_from: + return f"{parent.name}=={parent.version}" + if isinstance(ireq.comes_from, InstallRequirement): + return str(ireq.comes_from.name) + return str(ireq.comes_from) + + triggers = set() + for req, parent in e.causes: + if parent is None: + # This is a root requirement, so we can report it directly + trigger = req.format_for_error() + else: + trigger = describe_trigger(parent) + triggers.add(trigger) + + if triggers: + info = text_join(sorted(triggers)) + else: + info = "the requested packages" + + msg = ( + f"Cannot install {info} because these package versions " + "have conflicting dependencies." ) - message = template.format( - package=parent.name, - version=self._python_candidate.version, - specifier=str(requirement.specifier), + logger.critical(msg) + msg = "\nThe conflict is caused by:" + + relevant_constraints = set() + for req, parent in e.causes: + if req.name in constraints: + relevant_constraints.add(req.name) + msg = msg + "\n " + if parent: + msg = msg + f"{parent.name} {parent.version} depends on " + else: + msg = msg + "The user requested " + msg = msg + req.format_for_error() + for key in relevant_constraints: + spec = constraints[key].specifier + msg += f"\n The user requested (constraint) {key}{spec}" + + msg = ( + msg + + "\n\n" + + "To fix this you could try to:\n" + + "1. loosen the range of package versions you've specified\n" + + "2. remove package versions to allow pip to attempt to solve " + + "the dependency conflict\n" ) - return UnsupportedPythonVersion(message) - def get_installation_error(self, e): - # type: (ResolutionImpossible) -> Optional[InstallationError] - for cause in e.causes: - if isinstance(cause.requirement, RequiresPythonRequirement): - return self._report_requires_python_error( - cause.requirement, - cause.parent, - ) - return None + logger.info(msg) + + return DistributionNotFound( + "ResolutionImpossible: for help visit " + "https://pip.pypa.io/en/latest/topics/dependency-resolution/" + "#dealing-with-dependency-conflicts" + ) diff --git a/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/found_candidates.py b/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/found_candidates.py new file mode 100644 index 00000000..a1d57e0f --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/found_candidates.py @@ -0,0 +1,174 @@ +"""Utilities to lazily create and visit candidates found. + +Creating and visiting a candidate is a *very* costly operation. It involves +fetching, extracting, potentially building modules from source, and verifying +distribution metadata. It is therefore crucial for performance to keep +everything here lazy all the way down, so we only touch candidates that we +absolutely need, and not "download the world" when we only need one version of +something. +""" + +import functools +import logging +from collections.abc import Sequence +from typing import TYPE_CHECKING, Any, Callable, Iterator, Optional, Set, Tuple + +from pip._vendor.packaging.version import _BaseVersion + +from pip._internal.exceptions import MetadataInvalid + +from .base import Candidate + +logger = logging.getLogger(__name__) + +IndexCandidateInfo = Tuple[_BaseVersion, Callable[[], Optional[Candidate]]] + +if TYPE_CHECKING: + SequenceCandidate = Sequence[Candidate] +else: + # For compatibility: Python before 3.9 does not support using [] on the + # Sequence class. + # + # >>> from collections.abc import Sequence + # >>> Sequence[str] + # Traceback (most recent call last): + # File "", line 1, in + # TypeError: 'ABCMeta' object is not subscriptable + # + # TODO: Remove this block after dropping Python 3.8 support. + SequenceCandidate = Sequence + + +def _iter_built(infos: Iterator[IndexCandidateInfo]) -> Iterator[Candidate]: + """Iterator for ``FoundCandidates``. + + This iterator is used when the package is not already installed. Candidates + from index come later in their normal ordering. + """ + versions_found: Set[_BaseVersion] = set() + for version, func in infos: + if version in versions_found: + continue + try: + candidate = func() + except MetadataInvalid as e: + logger.warning( + "Ignoring version %s of %s since it has invalid metadata:\n" + "%s\n" + "Please use pip<24.1 if you need to use this version.", + version, + e.ireq.name, + e, + ) + # Mark version as found to avoid trying other candidates with the same + # version, since they most likely have invalid metadata as well. + versions_found.add(version) + else: + if candidate is None: + continue + yield candidate + versions_found.add(version) + + +def _iter_built_with_prepended( + installed: Candidate, infos: Iterator[IndexCandidateInfo] +) -> Iterator[Candidate]: + """Iterator for ``FoundCandidates``. + + This iterator is used when the resolver prefers the already-installed + candidate and NOT to upgrade. The installed candidate is therefore + always yielded first, and candidates from index come later in their + normal ordering, except skipped when the version is already installed. + """ + yield installed + versions_found: Set[_BaseVersion] = {installed.version} + for version, func in infos: + if version in versions_found: + continue + candidate = func() + if candidate is None: + continue + yield candidate + versions_found.add(version) + + +def _iter_built_with_inserted( + installed: Candidate, infos: Iterator[IndexCandidateInfo] +) -> Iterator[Candidate]: + """Iterator for ``FoundCandidates``. + + This iterator is used when the resolver prefers to upgrade an + already-installed package. Candidates from index are returned in their + normal ordering, except replaced when the version is already installed. + + The implementation iterates through and yields other candidates, inserting + the installed candidate exactly once before we start yielding older or + equivalent candidates, or after all other candidates if they are all newer. + """ + versions_found: Set[_BaseVersion] = set() + for version, func in infos: + if version in versions_found: + continue + # If the installed candidate is better, yield it first. + if installed.version >= version: + yield installed + versions_found.add(installed.version) + candidate = func() + if candidate is None: + continue + yield candidate + versions_found.add(version) + + # If the installed candidate is older than all other candidates. + if installed.version not in versions_found: + yield installed + + +class FoundCandidates(SequenceCandidate): + """A lazy sequence to provide candidates to the resolver. + + The intended usage is to return this from `find_matches()` so the resolver + can iterate through the sequence multiple times, but only access the index + page when remote packages are actually needed. This improve performances + when suitable candidates are already installed on disk. + """ + + def __init__( + self, + get_infos: Callable[[], Iterator[IndexCandidateInfo]], + installed: Optional[Candidate], + prefers_installed: bool, + incompatible_ids: Set[int], + ): + self._get_infos = get_infos + self._installed = installed + self._prefers_installed = prefers_installed + self._incompatible_ids = incompatible_ids + + def __getitem__(self, index: Any) -> Any: + # Implemented to satisfy the ABC check. This is not needed by the + # resolver, and should not be used by the provider either (for + # performance reasons). + raise NotImplementedError("don't do this") + + def __iter__(self) -> Iterator[Candidate]: + infos = self._get_infos() + if not self._installed: + iterator = _iter_built(infos) + elif self._prefers_installed: + iterator = _iter_built_with_prepended(self._installed, infos) + else: + iterator = _iter_built_with_inserted(self._installed, infos) + return (c for c in iterator if id(c) not in self._incompatible_ids) + + def __len__(self) -> int: + # Implemented to satisfy the ABC check. This is not needed by the + # resolver, and should not be used by the provider either (for + # performance reasons). + raise NotImplementedError("don't do this") + + @functools.lru_cache(maxsize=1) + def __bool__(self) -> bool: + if self._prefers_installed and self._installed: + return True + return any(self) diff --git a/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/provider.py b/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/provider.py index dc9e0cca..fb0dd85f 100644 --- a/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/provider.py +++ b/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/provider.py @@ -1,54 +1,258 @@ -from fetchcode.vcs.pip._vendor.resolvelib.providers import AbstractProvider +import collections +import math +from functools import lru_cache +from typing import ( + TYPE_CHECKING, + Dict, + Iterable, + Iterator, + Mapping, + Sequence, + TypeVar, + Union, +) -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._vendor.resolvelib.providers import AbstractProvider -if MYPY_CHECK_RUNNING: - from typing import Any, Optional, Sequence, Tuple, Union +from .base import Candidate, Constraint, Requirement +from .candidates import REQUIRES_PYTHON_IDENTIFIER +from .factory import Factory - from fetchcode.vcs.pip._internal.req.req_install import InstallRequirement +if TYPE_CHECKING: + from pip._vendor.resolvelib.providers import Preference + from pip._vendor.resolvelib.resolvers import RequirementInformation - from .base import Requirement, Candidate - from .factory import Factory + PreferenceInformation = RequirementInformation[Requirement, Candidate] + _ProviderBase = AbstractProvider[Requirement, Candidate, str] +else: + _ProviderBase = AbstractProvider + +# Notes on the relationship between the provider, the factory, and the +# candidate and requirement classes. +# +# The provider is a direct implementation of the resolvelib class. Its role +# is to deliver the API that resolvelib expects. +# +# Rather than work with completely abstract "requirement" and "candidate" +# concepts as resolvelib does, pip has concrete classes implementing these two +# ideas. The API of Requirement and Candidate objects are defined in the base +# classes, but essentially map fairly directly to the equivalent provider +# methods. In particular, `find_matches` and `is_satisfied_by` are +# requirement methods, and `get_dependencies` is a candidate method. +# +# The factory is the interface to pip's internal mechanisms. It is stateless, +# and is created by the resolver and held as a property of the provider. It is +# responsible for creating Requirement and Candidate objects, and provides +# services to those objects (access to pip's finder and preparer). + + +D = TypeVar("D") +V = TypeVar("V") + + +def _get_with_identifier( + mapping: Mapping[str, V], + identifier: str, + default: D, +) -> Union[D, V]: + """Get item from a package name lookup mapping with a resolver identifier. + + This extra logic is needed when the target mapping is keyed by package + name, which cannot be directly looked up with an identifier (which may + contain requested extras). Additional logic is added to also look up a value + by "cleaning up" the extras from the identifier. + """ + if identifier in mapping: + return mapping[identifier] + # HACK: Theoretically we should check whether this identifier is a valid + # "NAME[EXTRAS]" format, and parse out the name part with packaging or + # some regular expression. But since pip's resolver only spits out three + # kinds of identifiers: normalized PEP 503 names, normalized names plus + # extras, and Requires-Python, we can cheat a bit here. + name, open_bracket, _ = identifier.partition("[") + if open_bracket and name in mapping: + return mapping[name] + return default + + +class PipProvider(_ProviderBase): + """Pip's provider implementation for resolvelib. + + :params constraints: A mapping of constraints specified by the user. Keys + are canonicalized project names. + :params ignore_dependencies: Whether the user specified ``--no-deps``. + :params upgrade_strategy: The user-specified upgrade strategy. + :params user_requested: A set of canonicalized package names that the user + supplied for pip to install/upgrade. + """ -class PipProvider(AbstractProvider): def __init__( self, - factory, # type: Factory - ignore_dependencies, # type: bool - ): - # type: (...) -> None + factory: Factory, + constraints: Dict[str, Constraint], + ignore_dependencies: bool, + upgrade_strategy: str, + user_requested: Dict[str, int], + ) -> None: self._factory = factory + self._constraints = constraints self._ignore_dependencies = ignore_dependencies + self._upgrade_strategy = upgrade_strategy + self._user_requested = user_requested + self._known_depths: Dict[str, float] = collections.defaultdict(lambda: math.inf) - def get_install_requirement(self, c): - # type: (Candidate) -> Optional[InstallRequirement] - return c.get_install_requirement() - - def identify(self, dependency): - # type: (Union[Requirement, Candidate]) -> str - return dependency.name + def identify(self, requirement_or_candidate: Union[Requirement, Candidate]) -> str: + return requirement_or_candidate.name def get_preference( self, - resolution, # type: Optional[Candidate] - candidates, # type: Sequence[Candidate] - information # type: Sequence[Tuple[Requirement, Candidate]] - ): - # type: (...) -> Any - # Use the "usual" value for now - return len(candidates) - - def find_matches(self, requirement): - # type: (Requirement) -> Sequence[Candidate] - return requirement.find_matches() - - def is_satisfied_by(self, requirement, candidate): - # type: (Requirement, Candidate) -> bool + identifier: str, + resolutions: Mapping[str, Candidate], + candidates: Mapping[str, Iterator[Candidate]], + information: Mapping[str, Iterable["PreferenceInformation"]], + backtrack_causes: Sequence["PreferenceInformation"], + ) -> "Preference": + """Produce a sort key for given requirement based on preference. + + The lower the return value is, the more preferred this group of + arguments is. + + Currently pip considers the following in order: + + * Prefer if any of the known requirements is "direct", e.g. points to an + explicit URL. + * If equal, prefer if any requirement is "pinned", i.e. contains + operator ``===`` or ``==``. + * If equal, calculate an approximate "depth" and resolve requirements + closer to the user-specified requirements first. If the depth cannot + by determined (eg: due to no matching parents), it is considered + infinite. + * Order user-specified requirements by the order they are specified. + * If equal, prefers "non-free" requirements, i.e. contains at least one + operator, such as ``>=`` or ``<``. + * If equal, order alphabetically for consistency (helps debuggability). + """ + try: + next(iter(information[identifier])) + except StopIteration: + # There is no information for this identifier, so there's no known + # candidates. + has_information = False + else: + has_information = True + + if has_information: + lookups = (r.get_candidate_lookup() for r, _ in information[identifier]) + candidate, ireqs = zip(*lookups) + else: + candidate, ireqs = None, () + + operators = [ + specifier.operator + for specifier_set in (ireq.specifier for ireq in ireqs if ireq) + for specifier in specifier_set + ] + + direct = candidate is not None + pinned = any(op[:2] == "==" for op in operators) + unfree = bool(operators) + + try: + requested_order: Union[int, float] = self._user_requested[identifier] + except KeyError: + requested_order = math.inf + if has_information: + parent_depths = ( + self._known_depths[parent.name] if parent is not None else 0.0 + for _, parent in information[identifier] + ) + inferred_depth = min(d for d in parent_depths) + 1.0 + else: + inferred_depth = math.inf + else: + inferred_depth = 1.0 + self._known_depths[identifier] = inferred_depth + + requested_order = self._user_requested.get(identifier, math.inf) + + # Requires-Python has only one candidate and the check is basically + # free, so we always do it first to avoid needless work if it fails. + requires_python = identifier == REQUIRES_PYTHON_IDENTIFIER + + # Prefer the causes of backtracking on the assumption that the problem + # resolving the dependency tree is related to the failures that caused + # the backtracking + backtrack_cause = self.is_backtrack_cause(identifier, backtrack_causes) + + return ( + not requires_python, + not direct, + not pinned, + not backtrack_cause, + inferred_depth, + requested_order, + not unfree, + identifier, + ) + + def find_matches( + self, + identifier: str, + requirements: Mapping[str, Iterator[Requirement]], + incompatibilities: Mapping[str, Iterator[Candidate]], + ) -> Iterable[Candidate]: + def _eligible_for_upgrade(identifier: str) -> bool: + """Are upgrades allowed for this project? + + This checks the upgrade strategy, and whether the project was one + that the user specified in the command line, in order to decide + whether we should upgrade if there's a newer version available. + + (Note that we don't need access to the `--upgrade` flag, because + an upgrade strategy of "to-satisfy-only" means that `--upgrade` + was not specified). + """ + if self._upgrade_strategy == "eager": + return True + elif self._upgrade_strategy == "only-if-needed": + user_order = _get_with_identifier( + self._user_requested, + identifier, + default=None, + ) + return user_order is not None + return False + + constraint = _get_with_identifier( + self._constraints, + identifier, + default=Constraint.empty(), + ) + return self._factory.find_candidates( + identifier=identifier, + requirements=requirements, + constraint=constraint, + prefers_installed=(not _eligible_for_upgrade(identifier)), + incompatibilities=incompatibilities, + is_satisfied_by=self.is_satisfied_by, + ) + + @lru_cache(maxsize=None) + def is_satisfied_by(self, requirement: Requirement, candidate: Candidate) -> bool: return requirement.is_satisfied_by(candidate) - def get_dependencies(self, candidate): - # type: (Candidate) -> Sequence[Requirement] - if self._ignore_dependencies: - return [] - return candidate.get_dependencies() + def get_dependencies(self, candidate: Candidate) -> Sequence[Requirement]: + with_requires = not self._ignore_dependencies + return [r for r in candidate.iter_dependencies(with_requires) if r is not None] + + @staticmethod + def is_backtrack_cause( + identifier: str, backtrack_causes: Sequence["PreferenceInformation"] + ) -> bool: + for backtrack_cause in backtrack_causes: + if identifier == backtrack_cause.requirement.name: + return True + if backtrack_cause.parent and identifier == backtrack_cause.parent.name: + return True + return False diff --git a/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/reporter.py b/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/reporter.py new file mode 100644 index 00000000..0594569d --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/reporter.py @@ -0,0 +1,81 @@ +from collections import defaultdict +from logging import getLogger +from typing import Any, DefaultDict + +from pip._vendor.resolvelib.reporters import BaseReporter + +from .base import Candidate, Requirement + +logger = getLogger(__name__) + + +class PipReporter(BaseReporter): + def __init__(self) -> None: + self.reject_count_by_package: DefaultDict[str, int] = defaultdict(int) + + self._messages_at_reject_count = { + 1: ( + "pip is looking at multiple versions of {package_name} to " + "determine which version is compatible with other " + "requirements. This could take a while." + ), + 8: ( + "pip is still looking at multiple versions of {package_name} to " + "determine which version is compatible with other " + "requirements. This could take a while." + ), + 13: ( + "This is taking longer than usual. You might need to provide " + "the dependency resolver with stricter constraints to reduce " + "runtime. See https://pip.pypa.io/warnings/backtracking for " + "guidance. If you want to abort this run, press Ctrl + C." + ), + } + + def rejecting_candidate(self, criterion: Any, candidate: Candidate) -> None: + self.reject_count_by_package[candidate.name] += 1 + + count = self.reject_count_by_package[candidate.name] + if count not in self._messages_at_reject_count: + return + + message = self._messages_at_reject_count[count] + logger.info("INFO: %s", message.format(package_name=candidate.name)) + + msg = "Will try a different candidate, due to conflict:" + for req_info in criterion.information: + req, parent = req_info.requirement, req_info.parent + # Inspired by Factory.get_installation_error + msg += "\n " + if parent: + msg += f"{parent.name} {parent.version} depends on " + else: + msg += "The user requested " + msg += req.format_for_error() + logger.debug(msg) + + +class PipDebuggingReporter(BaseReporter): + """A reporter that does an info log for every event it sees.""" + + def starting(self) -> None: + logger.info("Reporter.starting()") + + def starting_round(self, index: int) -> None: + logger.info("Reporter.starting_round(%r)", index) + + def ending_round(self, index: int, state: Any) -> None: + logger.info("Reporter.ending_round(%r, state)", index) + logger.debug("Reporter.ending_round(%r, %r)", index, state) + + def ending(self, state: Any) -> None: + logger.info("Reporter.ending(%r)", state) + + def adding_requirement(self, requirement: Requirement, parent: Candidate) -> None: + logger.info("Reporter.adding_requirement(%r, %r)", requirement, parent) + + def rejecting_candidate(self, criterion: Any, candidate: Candidate) -> None: + logger.info("Reporter.rejecting_candidate(%r, %r)", criterion, candidate) + + def pinning(self, candidate: Candidate) -> None: + logger.info("Reporter.pinning(%r)", candidate) diff --git a/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/requirements.py b/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/requirements.py index 0d216b8d..b04f41b2 100644 --- a/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/requirements.py +++ b/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/requirements.py @@ -1,119 +1,245 @@ -from fetchcode.vcs.pip._vendor.packaging.utils import canonicalize_name +from typing import Any, Optional -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._vendor.packaging.specifiers import SpecifierSet +from pip._vendor.packaging.utils import NormalizedName, canonicalize_name -from .base import Requirement, format_name +from pip._internal.req.constructors import install_req_drop_extras +from pip._internal.req.req_install import InstallRequirement -if MYPY_CHECK_RUNNING: - from typing import Sequence +from .base import Candidate, CandidateLookup, Requirement, format_name - from fetchcode.vcs.pip._vendor.packaging.specifiers import SpecifierSet - from fetchcode.vcs.pip._internal.req.req_install import InstallRequirement +class ExplicitRequirement(Requirement): + def __init__(self, candidate: Candidate) -> None: + self.candidate = candidate - from .base import Candidate - from .factory import Factory + def __str__(self) -> str: + return str(self.candidate) + def __repr__(self) -> str: + return f"{self.__class__.__name__}({self.candidate!r})" -class ExplicitRequirement(Requirement): - def __init__(self, candidate): - # type: (Candidate) -> None - self.candidate = candidate + def __hash__(self) -> int: + return hash(self.candidate) - def __repr__(self): - # type: () -> str - return "{class_name}({candidate!r})".format( - class_name=self.__class__.__name__, - candidate=self.candidate, - ) + def __eq__(self, other: Any) -> bool: + if not isinstance(other, ExplicitRequirement): + return False + return self.candidate == other.candidate + + @property + def project_name(self) -> NormalizedName: + # No need to canonicalize - the candidate did this + return self.candidate.project_name @property - def name(self): - # type: () -> str - # No need to canonicalise - the candidate did this + def name(self) -> str: + # No need to canonicalize - the candidate did this return self.candidate.name - def find_matches(self): - # type: () -> Sequence[Candidate] - return [self.candidate] + def format_for_error(self) -> str: + return self.candidate.format_for_error() - def is_satisfied_by(self, candidate): - # type: (Candidate) -> bool + def get_candidate_lookup(self) -> CandidateLookup: + return self.candidate, None + + def is_satisfied_by(self, candidate: Candidate) -> bool: return candidate == self.candidate class SpecifierRequirement(Requirement): - def __init__(self, ireq, factory): - # type: (InstallRequirement, Factory) -> None + def __init__(self, ireq: InstallRequirement) -> None: assert ireq.link is None, "This is a link, not a specifier" self._ireq = ireq - self._factory = factory - self.extras = ireq.req.extras + self._equal_cache: Optional[str] = None + self._hash: Optional[int] = None + self._extras = frozenset(canonicalize_name(e) for e in self._ireq.extras) + + @property + def _equal(self) -> str: + if self._equal_cache is not None: + return self._equal_cache + + self._equal_cache = str(self._ireq) + return self._equal_cache - def __str__(self): - # type: () -> str + def __str__(self) -> str: return str(self._ireq.req) - def __repr__(self): - # type: () -> str - return "{class_name}({requirement!r})".format( - class_name=self.__class__.__name__, - requirement=str(self._ireq.req), - ) + def __repr__(self) -> str: + return f"{self.__class__.__name__}({str(self._ireq.req)!r})" + + def __eq__(self, other: object) -> bool: + if not isinstance(other, SpecifierRequirement): + return NotImplemented + return self._equal == other._equal + + def __hash__(self) -> int: + if self._hash is not None: + return self._hash + + self._hash = hash(self._equal) + return self._hash + + @property + def project_name(self) -> NormalizedName: + assert self._ireq.req, "Specifier-backed ireq is always PEP 508" + return canonicalize_name(self._ireq.req.name) @property - def name(self): - # type: () -> str - canonical_name = canonicalize_name(self._ireq.req.name) - return format_name(canonical_name, self.extras) - - def find_matches(self): - # type: () -> Sequence[Candidate] - it = self._factory.iter_found_candidates(self._ireq, self.extras) - return list(it) - - def is_satisfied_by(self, candidate): - # type: (Candidate) -> bool - assert candidate.name == self.name, \ - "Internal issue: Candidate is not for this requirement " \ - " {} vs {}".format(candidate.name, self.name) + def name(self) -> str: + return format_name(self.project_name, self._extras) + + def format_for_error(self) -> str: + # Convert comma-separated specifiers into "A, B, ..., F and G" + # This makes the specifier a bit more "human readable", without + # risking a change in meaning. (Hopefully! Not all edge cases have + # been checked) + parts = [s.strip() for s in str(self).split(",")] + if len(parts) == 0: + return "" + elif len(parts) == 1: + return parts[0] + + return ", ".join(parts[:-1]) + " and " + parts[-1] + + def get_candidate_lookup(self) -> CandidateLookup: + return None, self._ireq + + def is_satisfied_by(self, candidate: Candidate) -> bool: + assert candidate.name == self.name, ( + f"Internal issue: Candidate is not for this requirement " + f"{candidate.name} vs {self.name}" + ) # We can safely always allow prereleases here since PackageFinder # already implements the prerelease logic, and would have filtered out # prerelease candidates if the user does not expect them. + assert self._ireq.req, "Specifier-backed ireq is always PEP 508" spec = self._ireq.req.specifier return spec.contains(candidate.version, prereleases=True) -class RequiresPythonRequirement(Requirement): - """A requirement representing Requires-Python metadata. +class SpecifierWithoutExtrasRequirement(SpecifierRequirement): + """ + Requirement backed by an install requirement on a base package. + Trims extras from its install requirement if there are any. """ - def __init__(self, specifier, match): - # type: (SpecifierSet, Candidate) -> None + + def __init__(self, ireq: InstallRequirement) -> None: + assert ireq.link is None, "This is a link, not a specifier" + self._ireq = install_req_drop_extras(ireq) + self._equal_cache: Optional[str] = None + self._hash: Optional[int] = None + self._extras = frozenset(canonicalize_name(e) for e in self._ireq.extras) + + @property + def _equal(self) -> str: + if self._equal_cache is not None: + return self._equal_cache + + self._equal_cache = str(self._ireq) + return self._equal_cache + + def __eq__(self, other: object) -> bool: + if not isinstance(other, SpecifierWithoutExtrasRequirement): + return NotImplemented + return self._equal == other._equal + + def __hash__(self) -> int: + if self._hash is not None: + return self._hash + + self._hash = hash(self._equal) + return self._hash + + +class RequiresPythonRequirement(Requirement): + """A requirement representing Requires-Python metadata.""" + + def __init__(self, specifier: SpecifierSet, match: Candidate) -> None: self.specifier = specifier + self._specifier_string = str(specifier) # for faster __eq__ + self._hash: Optional[int] = None self._candidate = match - def __repr__(self): - # type: () -> str - return "{class_name}({specifier!r})".format( - class_name=self.__class__.__name__, - specifier=str(self.specifier), + def __str__(self) -> str: + return f"Python {self.specifier}" + + def __repr__(self) -> str: + return f"{self.__class__.__name__}({str(self.specifier)!r})" + + def __hash__(self) -> int: + if self._hash is not None: + return self._hash + + self._hash = hash((self._specifier_string, self._candidate)) + return self._hash + + def __eq__(self, other: Any) -> bool: + if not isinstance(other, RequiresPythonRequirement): + return False + return ( + self._specifier_string == other._specifier_string + and self._candidate == other._candidate ) @property - def name(self): - # type: () -> str + def project_name(self) -> NormalizedName: + return self._candidate.project_name + + @property + def name(self) -> str: return self._candidate.name - def find_matches(self): - # type: () -> Sequence[Candidate] - if self._candidate.version in self.specifier: - return [self._candidate] - return [] + def format_for_error(self) -> str: + return str(self) + + def get_candidate_lookup(self) -> CandidateLookup: + if self.specifier.contains(self._candidate.version, prereleases=True): + return self._candidate, None + return None, None - def is_satisfied_by(self, candidate): - # type: (Candidate) -> bool + def is_satisfied_by(self, candidate: Candidate) -> bool: assert candidate.name == self._candidate.name, "Not Python candidate" # We can safely always allow prereleases here since PackageFinder # already implements the prerelease logic, and would have filtered out # prerelease candidates if the user does not expect them. return self.specifier.contains(candidate.version, prereleases=True) + + +class UnsatisfiableRequirement(Requirement): + """A requirement that cannot be satisfied.""" + + def __init__(self, name: NormalizedName) -> None: + self._name = name + + def __str__(self) -> str: + return f"{self._name} (unavailable)" + + def __repr__(self) -> str: + return f"{self.__class__.__name__}({str(self._name)!r})" + + def __eq__(self, other: object) -> bool: + if not isinstance(other, UnsatisfiableRequirement): + return NotImplemented + return self._name == other._name + + def __hash__(self) -> int: + return hash(self._name) + + @property + def project_name(self) -> NormalizedName: + return self._name + + @property + def name(self) -> str: + return self._name + + def format_for_error(self) -> str: + return str(self) + + def get_candidate_lookup(self) -> CandidateLookup: + return None, None + + def is_satisfied_by(self, candidate: Candidate) -> bool: + return False diff --git a/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/resolver.py b/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/resolver.py index 55de61cd..c12beef0 100644 --- a/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/resolver.py +++ b/src/fetchcode/vcs/pip/_internal/resolution/resolvelib/resolver.py @@ -1,155 +1,210 @@ +import contextlib import functools import logging +import os +from typing import TYPE_CHECKING, Dict, List, Optional, Set, Tuple, cast -from fetchcode.vcs.pip._vendor import six -from fetchcode.vcs.pip._vendor.packaging.utils import canonicalize_name -from fetchcode.vcs.pip._vendor.resolvelib import BaseReporter, ResolutionImpossible -from fetchcode.vcs.pip._vendor.resolvelib import Resolver as RLResolver +from pip._vendor.packaging.utils import canonicalize_name +from pip._vendor.resolvelib import BaseReporter, ResolutionImpossible +from pip._vendor.resolvelib import Resolver as RLResolver +from pip._vendor.resolvelib.structs import DirectedGraph -from fetchcode.vcs.pip._internal.exceptions import InstallationError -from fetchcode.vcs.pip._internal.req.req_set import RequirementSet -from fetchcode.vcs.pip._internal.resolution.base import BaseResolver -from fetchcode.vcs.pip._internal.resolution.resolvelib.provider import PipProvider -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._internal.cache import WheelCache +from pip._internal.index.package_finder import PackageFinder +from pip._internal.operations.prepare import RequirementPreparer +from pip._internal.req.constructors import install_req_extend_extras +from pip._internal.req.req_install import InstallRequirement +from pip._internal.req.req_set import RequirementSet +from pip._internal.resolution.base import BaseResolver, InstallRequirementProvider +from pip._internal.resolution.resolvelib.provider import PipProvider +from pip._internal.resolution.resolvelib.reporter import ( + PipDebuggingReporter, + PipReporter, +) +from pip._internal.utils.packaging import get_requirement +from .base import Candidate, Requirement from .factory import Factory -if MYPY_CHECK_RUNNING: - from typing import Dict, List, Optional, Tuple +if TYPE_CHECKING: + from pip._vendor.resolvelib.resolvers import Result as RLResult - from fetchcode.vcs.pip._vendor.resolvelib.resolvers import Result - - from fetchcode.vcs.pip._internal.cache import WheelCache - from fetchcode.vcs.pip._internal.index.package_finder import PackageFinder - from fetchcode.vcs.pip._internal.operations.prepare import RequirementPreparer - from fetchcode.vcs.pip._internal.req.req_install import InstallRequirement - from fetchcode.vcs.pip._internal.resolution.base import InstallRequirementProvider + Result = RLResult[Requirement, Candidate, str] logger = logging.getLogger(__name__) class Resolver(BaseResolver): + _allowed_strategies = {"eager", "only-if-needed", "to-satisfy-only"} + def __init__( self, - preparer, # type: RequirementPreparer - finder, # type: PackageFinder - wheel_cache, # type: Optional[WheelCache] - make_install_req, # type: InstallRequirementProvider - use_user_site, # type: bool - ignore_dependencies, # type: bool - ignore_installed, # type: bool - ignore_requires_python, # type: bool - force_reinstall, # type: bool - upgrade_strategy, # type: str - py_version_info=None, # type: Optional[Tuple[int, ...]] + preparer: RequirementPreparer, + finder: PackageFinder, + wheel_cache: Optional[WheelCache], + make_install_req: InstallRequirementProvider, + use_user_site: bool, + ignore_dependencies: bool, + ignore_installed: bool, + ignore_requires_python: bool, + force_reinstall: bool, + upgrade_strategy: str, + py_version_info: Optional[Tuple[int, ...]] = None, ): - super(Resolver, self).__init__() + super().__init__() + assert upgrade_strategy in self._allowed_strategies + self.factory = Factory( finder=finder, preparer=preparer, make_install_req=make_install_req, + wheel_cache=wheel_cache, + use_user_site=use_user_site, force_reinstall=force_reinstall, ignore_installed=ignore_installed, ignore_requires_python=ignore_requires_python, py_version_info=py_version_info, ) self.ignore_dependencies = ignore_dependencies - self._result = None # type: Optional[Result] - - def resolve(self, root_reqs, check_supported_wheels): - # type: (List[InstallRequirement], bool) -> RequirementSet - - # FIXME: Implement constraints. - if any(r.constraint for r in root_reqs): - raise InstallationError("Constraints are not yet supported.") + self.upgrade_strategy = upgrade_strategy + self._result: Optional[Result] = None + def resolve( + self, root_reqs: List[InstallRequirement], check_supported_wheels: bool + ) -> RequirementSet: + collected = self.factory.collect_root_requirements(root_reqs) provider = PipProvider( factory=self.factory, + constraints=collected.constraints, ignore_dependencies=self.ignore_dependencies, + upgrade_strategy=self.upgrade_strategy, + user_requested=collected.user_requested, + ) + if "PIP_RESOLVER_DEBUG" in os.environ: + reporter: BaseReporter = PipDebuggingReporter() + else: + reporter = PipReporter() + resolver: RLResolver[Requirement, Candidate, str] = RLResolver( + provider, + reporter, ) - reporter = BaseReporter() - resolver = RLResolver(provider, reporter) - - requirements = [ - self.factory.make_requirement_from_install_req(r) - for r in root_reqs - ] try: - self._result = resolver.resolve(requirements) + limit_how_complex_resolution_can_be = 200000 + result = self._result = resolver.resolve( + collected.requirements, max_rounds=limit_how_complex_resolution_can_be + ) except ResolutionImpossible as e: - error = self.factory.get_installation_error(e) - if not error: - # TODO: This needs fixing, we need to look at the - # factory.get_installation_error infrastructure, as that - # doesn't really allow for the logger.critical calls I'm - # using here. - for req, parent in e.causes: - logger.critical( - "Could not find a version that satisfies " + - "the requirement " + - str(req) + - ("" if parent is None else " (from {})".format( - parent.name - )) - ) - raise InstallationError( - "No matching distribution found for " + - ", ".join([r.name for r, _ in e.causes]) - ) - raise - six.raise_from(error, e) + error = self.factory.get_installation_error( + cast("ResolutionImpossible[Requirement, Candidate]", e), + collected.constraints, + ) + raise error from e req_set = RequirementSet(check_supported_wheels=check_supported_wheels) - for candidate in self._result.mapping.values(): - ireq = provider.get_install_requirement(candidate) + # process candidates with extras last to ensure their base equivalent is + # already in the req_set if appropriate. + # Python's sort is stable so using a binary key function keeps relative order + # within both subsets. + for candidate in sorted( + result.mapping.values(), key=lambda c: c.name != c.project_name + ): + ireq = candidate.get_install_requirement() if ireq is None: + if candidate.name != candidate.project_name: + # extend existing req's extras + with contextlib.suppress(KeyError): + req = req_set.get_requirement(candidate.project_name) + req_set.add_named_requirement( + install_req_extend_extras( + req, get_requirement(candidate.name).extras + ) + ) + continue + + # Check if there is already an installation under the same name, + # and set a flag for later stages to uninstall it, if needed. + installed_dist = self.factory.get_dist_to_uninstall(candidate) + if installed_dist is None: + # There is no existing installation -- nothing to uninstall. + ireq.should_reinstall = False + elif self.factory.force_reinstall: + # The --force-reinstall flag is set -- reinstall. + ireq.should_reinstall = True + elif installed_dist.version != candidate.version: + # The installation is different in version -- reinstall. + ireq.should_reinstall = True + elif candidate.is_editable or installed_dist.editable: + # The incoming distribution is editable, or different in + # editable-ness to installation -- reinstall. + ireq.should_reinstall = True + elif candidate.source_link and candidate.source_link.is_file: + # The incoming distribution is under file:// + if candidate.source_link.is_wheel: + # is a local wheel -- do nothing. + logger.info( + "%s is already installed with the same version as the " + "provided wheel. Use --force-reinstall to force an " + "installation of the wheel.", + ireq.name, + ) + continue + + # is a local sdist or path -- reinstall + ireq.should_reinstall = True + else: continue - ireq.should_reinstall = self.factory.should_reinstall(candidate) + + link = candidate.source_link + if link and link.is_yanked: + # The reason can contain non-ASCII characters, Unicode + # is required for Python 2. + msg = ( + "The candidate selected for download or install is a " + "yanked version: {name!r} candidate (version {version} " + "at {link})\nReason for being yanked: {reason}" + ).format( + name=candidate.name, + version=candidate.version, + link=link, + reason=link.yanked_reason or "", + ) + logger.warning(msg) + req_set.add_named_requirement(ireq) + reqs = req_set.all_requirements + self.factory.preparer.prepare_linked_requirements_more(reqs) + for req in reqs: + req.prepared = True + req.needs_more_preparation = False return req_set - def get_installation_order(self, req_set): - # type: (RequirementSet) -> List[InstallRequirement] - """Create a list that orders given requirements for installation. + def get_installation_order( + self, req_set: RequirementSet + ) -> List[InstallRequirement]: + """Get order for installation of requirements in RequirementSet. - The returned list should contain all requirements in ``req_set``, - so the caller can loop through it and have a requirement installed - before the requiring thing. + The returned list contains a requirement before another that depends on + it. This helps ensure that the environment is kept consistent as they + get installed one-by-one. - The current implementation walks the resolved dependency graph, and - make sure every node has a greater "weight" than all its parents. + The current implementation creates a topological ordering of the + dependency graph, giving more weight to packages with less + or no dependencies, while breaking any cycles in the graph at + arbitrary points. We make no guarantees about where the cycle + would be broken, other than it *would* be broken. """ assert self._result is not None, "must call resolve() first" - weights = {} # type: Dict[Optional[str], int] + + if not req_set.requirements: + # Nothing is left to install, so we do not need an order. + return [] graph = self._result.graph - key_count = len(self._result.mapping) + 1 # Packages plus sentinal. - while len(weights) < key_count: - progressed = False - for key in graph: - if key in weights: - continue - parents = list(graph.iter_parents(key)) - if not all(p in weights for p in parents): - continue - if parents: - weight = max(weights[p] for p in parents) + 1 - else: - weight = 0 - weights[key] = weight - progressed = True - - # FIXME: This check will fail if there are unbreakable cycles. - # Implement something to forcifully break them up to continue. - if not progressed: - raise InstallationError( - "Could not determine installation order due to cicular " - "dependency." - ) + weights = get_topological_weights(graph, set(req_set.requirements.keys())) sorted_items = sorted( req_set.requirements.items(), @@ -159,11 +214,99 @@ def get_installation_order(self, req_set): return [ireq for _, ireq in sorted_items] +def get_topological_weights( + graph: "DirectedGraph[Optional[str]]", requirement_keys: Set[str] +) -> Dict[Optional[str], int]: + """Assign weights to each node based on how "deep" they are. + + This implementation may change at any point in the future without prior + notice. + + We first simplify the dependency graph by pruning any leaves and giving them + the highest weight: a package without any dependencies should be installed + first. This is done again and again in the same way, giving ever less weight + to the newly found leaves. The loop stops when no leaves are left: all + remaining packages have at least one dependency left in the graph. + + Then we continue with the remaining graph, by taking the length for the + longest path to any node from root, ignoring any paths that contain a single + node twice (i.e. cycles). This is done through a depth-first search through + the graph, while keeping track of the path to the node. + + Cycles in the graph result would result in node being revisited while also + being on its own path. In this case, take no action. This helps ensure we + don't get stuck in a cycle. + + When assigning weight, the longer path (i.e. larger length) is preferred. + + We are only interested in the weights of packages that are in the + requirement_keys. + """ + path: Set[Optional[str]] = set() + weights: Dict[Optional[str], int] = {} + + def visit(node: Optional[str]) -> None: + if node in path: + # We hit a cycle, so we'll break it here. + return + + # Time to visit the children! + path.add(node) + for child in graph.iter_children(node): + visit(child) + path.remove(node) + + if node not in requirement_keys: + return + + last_known_parent_count = weights.get(node, 0) + weights[node] = max(last_known_parent_count, len(path)) + + # Simplify the graph, pruning leaves that have no dependencies. + # This is needed for large graphs (say over 200 packages) because the + # `visit` function is exponentially slower then, taking minutes. + # See https://github.com/pypa/pip/issues/10557 + # We will loop until we explicitly break the loop. + while True: + leaves = set() + for key in graph: + if key is None: + continue + for _child in graph.iter_children(key): + # This means we have at least one child + break + else: + # No child. + leaves.add(key) + if not leaves: + # We are done simplifying. + break + # Calculate the weight for the leaves. + weight = len(graph) - 1 + for leaf in leaves: + if leaf not in requirement_keys: + continue + weights[leaf] = weight + # Remove the leaves from the graph, making it simpler. + for leaf in leaves: + graph.remove(leaf) + + # Visit the remaining graph. + # `None` is guaranteed to be the root node by resolvelib. + visit(None) + + # Sanity check: all requirement keys should be in the weights, + # and no other keys should be in the weights. + difference = set(weights.keys()).difference(requirement_keys) + assert not difference, difference + + return weights + + def _req_set_item_sorter( - item, # type: Tuple[str, InstallRequirement] - weights, # type: Dict[Optional[str], int] -): - # type: (...) -> Tuple[int, str] + item: Tuple[str, InstallRequirement], + weights: Dict[Optional[str], int], +) -> Tuple[int, str]: """Key function used to sort install requirements for installation. Based on the "weight" mapping calculated in ``get_installation_order()``. diff --git a/src/fetchcode/vcs/pip/_internal/self_outdated_check.py b/src/fetchcode/vcs/pip/_internal/self_outdated_check.py index fc15eeea..f9a91af9 100644 --- a/src/fetchcode/vcs/pip/_internal/self_outdated_check.py +++ b/src/fetchcode/vcs/pip/_internal/self_outdated_check.py @@ -1,242 +1,244 @@ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import - import datetime +import functools import hashlib import json import logging +import optparse import os.path import sys - -from fetchcode.vcs.pip._vendor import pkg_resources -from fetchcode.vcs.pip._vendor.packaging import version as packaging_version -from fetchcode.vcs.pip._vendor.six import ensure_binary - -from fetchcode.vcs.pip._internal.index.collector import LinkCollector -from fetchcode.vcs.pip._internal.index.package_finder import PackageFinder -from fetchcode.vcs.pip._internal.models.search_scope import SearchScope -from fetchcode.vcs.pip._internal.models.selection_prefs import SelectionPreferences -from fetchcode.vcs.pip._internal.utils.filesystem import ( - adjacent_tmp_file, - check_path_owner, - replace, -) -from fetchcode.vcs.pip._internal.utils.misc import ( - ensure_dir, - get_installed_version, - redact_auth_from_url, +from dataclasses import dataclass +from typing import Any, Callable, Dict, Optional + +from pip._vendor.packaging.version import Version +from pip._vendor.packaging.version import parse as parse_version +from pip._vendor.rich.console import Group +from pip._vendor.rich.markup import escape +from pip._vendor.rich.text import Text + +from pip._internal.index.collector import LinkCollector +from pip._internal.index.package_finder import PackageFinder +from pip._internal.metadata import get_default_environment +from pip._internal.models.selection_prefs import SelectionPreferences +from pip._internal.network.session import PipSession +from pip._internal.utils.compat import WINDOWS +from pip._internal.utils.entrypoints import ( + get_best_invocation_for_this_pip, + get_best_invocation_for_this_python, ) -from fetchcode.vcs.pip._internal.utils.packaging import get_installer -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - import optparse - from optparse import Values - from typing import Any, Dict, Text, Union - - from fetchcode.vcs.pip._internal.network.session import PipSession - - -SELFCHECK_DATE_FMT = "%Y-%m-%dT%H:%M:%SZ" +from pip._internal.utils.filesystem import adjacent_tmp_file, check_path_owner, replace +from pip._internal.utils.misc import ensure_dir +_WEEK = datetime.timedelta(days=7) logger = logging.getLogger(__name__) -def make_link_collector( - session, # type: PipSession - options, # type: Values - suppress_no_index=False, # type: bool -): - # type: (...) -> LinkCollector - """ - :param session: The Session to use to make requests. - :param suppress_no_index: Whether to ignore the --no-index option - when constructing the SearchScope object. - """ - index_urls = [options.index_url] + options.extra_index_urls - if options.no_index and not suppress_no_index: - logger.debug( - 'Ignoring indexes: %s', - ','.join(redact_auth_from_url(url) for url in index_urls), - ) - index_urls = [] - - # Make sure find_links is a list before passing to create(). - find_links = options.find_links or [] - - search_scope = SearchScope.create( - find_links=find_links, index_urls=index_urls, - ) - - link_collector = LinkCollector(session=session, search_scope=search_scope) +def _get_statefile_name(key: str) -> str: + key_bytes = key.encode() + name = hashlib.sha224(key_bytes).hexdigest() + return name - return link_collector +def _convert_date(isodate: str) -> datetime.datetime: + """Convert an ISO format string to a date. -def _get_statefile_name(key): - # type: (Union[str, Text]) -> str - key_bytes = ensure_binary(key) - name = hashlib.sha224(key_bytes).hexdigest() - return name + Handles the format 2020-01-22T14:24:01Z (trailing Z) + which is not supported by older versions of fromisoformat. + """ + return datetime.datetime.fromisoformat(isodate.replace("Z", "+00:00")) -class SelfCheckState(object): - def __init__(self, cache_dir): - # type: (str) -> None - self.state = {} # type: Dict[str, Any] - self.statefile_path = None +class SelfCheckState: + def __init__(self, cache_dir: str) -> None: + self._state: Dict[str, Any] = {} + self._statefile_path = None # Try to load the existing state if cache_dir: - self.statefile_path = os.path.join( + self._statefile_path = os.path.join( cache_dir, "selfcheck", _get_statefile_name(self.key) ) try: - with open(self.statefile_path) as statefile: - self.state = json.load(statefile) - except (IOError, ValueError, KeyError): + with open(self._statefile_path, encoding="utf-8") as statefile: + self._state = json.load(statefile) + except (OSError, ValueError, KeyError): # Explicitly suppressing exceptions, since we don't want to # error out if the cache file is invalid. pass @property - def key(self): + def key(self) -> str: return sys.prefix - def save(self, pypi_version, current_time): - # type: (str, datetime.datetime) -> None + def get(self, current_time: datetime.datetime) -> Optional[str]: + """Check if we have a not-outdated version loaded already.""" + if not self._state: + return None + + if "last_check" not in self._state: + return None + + if "pypi_version" not in self._state: + return None + + # Determine if we need to refresh the state + last_check = _convert_date(self._state["last_check"]) + time_since_last_check = current_time - last_check + if time_since_last_check > _WEEK: + return None + + return self._state["pypi_version"] + + def set(self, pypi_version: str, current_time: datetime.datetime) -> None: # If we do not have a path to cache in, don't bother saving. - if not self.statefile_path: + if not self._statefile_path: return # Check to make sure that we own the directory - if not check_path_owner(os.path.dirname(self.statefile_path)): + if not check_path_owner(os.path.dirname(self._statefile_path)): return # Now that we've ensured the directory is owned by this user, we'll go # ahead and make sure that all our directories are created. - ensure_dir(os.path.dirname(self.statefile_path)) + ensure_dir(os.path.dirname(self._statefile_path)) state = { # Include the key so it's easy to tell which pip wrote the # file. "key": self.key, - "last_check": current_time.strftime(SELFCHECK_DATE_FMT), + "last_check": current_time.isoformat(), "pypi_version": pypi_version, } text = json.dumps(state, sort_keys=True, separators=(",", ":")) - with adjacent_tmp_file(self.statefile_path) as f: - f.write(ensure_binary(text)) + with adjacent_tmp_file(self._statefile_path) as f: + f.write(text.encode()) try: # Since we have a prefix-specific state file, we can just # overwrite whatever is there, no need to check. - replace(f.name, self.statefile_path) + replace(f.name, self._statefile_path) except OSError: # Best effort. pass -def was_installed_by_pip(pkg): - # type: (str) -> bool +@dataclass +class UpgradePrompt: + old: str + new: str + + def __rich__(self) -> Group: + if WINDOWS: + pip_cmd = f"{get_best_invocation_for_this_python()} -m pip" + else: + pip_cmd = get_best_invocation_for_this_pip() + + notice = "[bold][[reset][blue]notice[reset][bold]][reset]" + return Group( + Text(), + Text.from_markup( + f"{notice} A new release of pip is available: " + f"[red]{self.old}[reset] -> [green]{self.new}[reset]" + ), + Text.from_markup( + f"{notice} To update, run: " + f"[green]{escape(pip_cmd)} install --upgrade pip" + ), + ) + + +def was_installed_by_pip(pkg: str) -> bool: """Checks whether pkg was installed by pip This is used not to display the upgrade message when pip is in fact installed by system package manager, such as dnf on Fedora. """ - try: - dist = pkg_resources.get_distribution(pkg) - return "pip" == get_installer(dist) - except pkg_resources.DistributionNotFound: - return False + dist = get_default_environment().get_distribution(pkg) + return dist is not None and "pip" == dist.installer + + +def _get_current_remote_pip_version( + session: PipSession, options: optparse.Values +) -> Optional[str]: + # Lets use PackageFinder to see what the latest pip version is + link_collector = LinkCollector.create( + session, + options=options, + suppress_no_index=True, + ) + + # Pass allow_yanked=False so we don't suggest upgrading to a + # yanked version. + selection_prefs = SelectionPreferences( + allow_yanked=False, + allow_all_prereleases=False, # Explicitly set to False + ) + + finder = PackageFinder.create( + link_collector=link_collector, + selection_prefs=selection_prefs, + ) + best_candidate = finder.find_best_candidate("pip").best_candidate + if best_candidate is None: + return None + + return str(best_candidate.version) + + +def _self_version_check_logic( + *, + state: SelfCheckState, + current_time: datetime.datetime, + local_version: Version, + get_remote_version: Callable[[], Optional[str]], +) -> Optional[UpgradePrompt]: + remote_version_str = state.get(current_time) + if remote_version_str is None: + remote_version_str = get_remote_version() + if remote_version_str is None: + logger.debug("No remote pip version found") + return None + state.set(remote_version_str, current_time) + + remote_version = parse_version(remote_version_str) + logger.debug("Remote version of pip: %s", remote_version) + logger.debug("Local version of pip: %s", local_version) + + pip_installed_by_pip = was_installed_by_pip("pip") + logger.debug("Was pip installed by pip? %s", pip_installed_by_pip) + if not pip_installed_by_pip: + return None # Only suggest upgrade if pip is installed by pip. + + local_version_is_older = ( + local_version < remote_version + and local_version.base_version != remote_version.base_version + ) + if local_version_is_older: + return UpgradePrompt(old=str(local_version), new=remote_version_str) + return None -def pip_self_version_check(session, options): - # type: (PipSession, optparse.Values) -> None + +def pip_self_version_check(session: PipSession, options: optparse.Values) -> None: """Check for an update for pip. Limit the frequency of checks to once per week. State is stored either in the active virtualenv or in the user's USER_CACHE_DIR keyed off the prefix of the pip script path. """ - installed_version = get_installed_version("pip") - if not installed_version: + installed_dist = get_default_environment().get_distribution("pip") + if not installed_dist: return - pip_version = packaging_version.parse(installed_version) - pypi_version = None - - try: - state = SelfCheckState(cache_dir=options.cache_dir) - - current_time = datetime.datetime.utcnow() - # Determine if we need to refresh the state - if "last_check" in state.state and "pypi_version" in state.state: - last_check = datetime.datetime.strptime( - state.state["last_check"], - SELFCHECK_DATE_FMT - ) - if (current_time - last_check).total_seconds() < 7 * 24 * 60 * 60: - pypi_version = state.state["pypi_version"] - - # Refresh the version if we need to or just see if we need to warn - if pypi_version is None: - # Lets use PackageFinder to see what the latest pip version is - link_collector = make_link_collector( - session, - options=options, - suppress_no_index=True, - ) - - # Pass allow_yanked=False so we don't suggest upgrading to a - # yanked version. - selection_prefs = SelectionPreferences( - allow_yanked=False, - allow_all_prereleases=False, # Explicitly set to False - ) - - finder = PackageFinder.create( - link_collector=link_collector, - selection_prefs=selection_prefs, - ) - best_candidate = finder.find_best_candidate("pip").best_candidate - if best_candidate is None: - return - pypi_version = str(best_candidate.version) - - # save that we've performed a check - state.save(pypi_version, current_time) - - remote_version = packaging_version.parse(pypi_version) - - local_version_is_older = ( - pip_version < remote_version and - pip_version.base_version != remote_version.base_version and - was_installed_by_pip('pip') - ) - - # Determine if our pypi_version is older - if not local_version_is_older: - return - - # We cannot tell how the current pip is available in the current - # command context, so be pragmatic here and suggest the command - # that's always available. This does not accommodate spaces in - # `sys.executable`. - pip_cmd = "{} -m pip".format(sys.executable) - logger.warning( - "You are using pip version %s; however, version %s is " - "available.\nYou should consider upgrading via the " - "'%s install --upgrade pip' command.", - pip_version, pypi_version, pip_cmd - ) - except Exception: - logger.debug( - "There was an error checking the latest version of pip", - exc_info=True, - ) + upgrade_prompt = _self_version_check_logic( + state=SelfCheckState(cache_dir=options.cache_dir), + current_time=datetime.datetime.now(datetime.timezone.utc), + local_version=installed_dist.version, + get_remote_version=functools.partial( + _get_current_remote_pip_version, session, options + ), + ) + if upgrade_prompt is not None: + logger.warning("%s", upgrade_prompt, extra={"rich": True}) diff --git a/src/fetchcode/vcs/pip/_internal/utils/_jaraco_text.py b/src/fetchcode/vcs/pip/_internal/utils/_jaraco_text.py new file mode 100644 index 00000000..6ccf53b7 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/utils/_jaraco_text.py @@ -0,0 +1,109 @@ +"""Functions brought over from jaraco.text. + +These functions are not supposed to be used within `pip._internal`. These are +helper functions brought over from `jaraco.text` to enable vendoring newer +copies of `pkg_resources` without having to vendor `jaraco.text` and its entire +dependency cone; something that our vendoring setup is not currently capable of +handling. + +License reproduced from original source below: + +Copyright Jason R. Coombs + +Permission is hereby granted, free of charge, to any person obtaining a copy +of this software and associated documentation files (the "Software"), to +deal in the Software without restriction, including without limitation the +rights to use, copy, modify, merge, publish, distribute, sublicense, and/or +sell copies of the Software, and to permit persons to whom the Software is +furnished to do so, subject to the following conditions: + +The above copyright notice and this permission notice shall be included in +all copies or substantial portions of the Software. + +THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR +IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, +FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE +AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER +LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING +FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS +IN THE SOFTWARE. +""" + +import functools +import itertools + + +def _nonblank(str): + return str and not str.startswith("#") + + +@functools.singledispatch +def yield_lines(iterable): + r""" + Yield valid lines of a string or iterable. + + >>> list(yield_lines('')) + [] + >>> list(yield_lines(['foo', 'bar'])) + ['foo', 'bar'] + >>> list(yield_lines('foo\nbar')) + ['foo', 'bar'] + >>> list(yield_lines('\nfoo\n#bar\nbaz #comment')) + ['foo', 'baz #comment'] + >>> list(yield_lines(['foo\nbar', 'baz', 'bing\n\n\n'])) + ['foo', 'bar', 'baz', 'bing'] + """ + return itertools.chain.from_iterable(map(yield_lines, iterable)) + + +@yield_lines.register(str) +def _(text): + return filter(_nonblank, map(str.strip, text.splitlines())) + + +def drop_comment(line): + """ + Drop comments. + + >>> drop_comment('foo # bar') + 'foo' + + A hash without a space may be in a URL. + + >>> drop_comment('http://example.com/foo#bar') + 'http://example.com/foo#bar' + """ + return line.partition(" #")[0] + + +def join_continuation(lines): + r""" + Join lines continued by a trailing backslash. + + >>> list(join_continuation(['foo \\', 'bar', 'baz'])) + ['foobar', 'baz'] + >>> list(join_continuation(['foo \\', 'bar', 'baz'])) + ['foobar', 'baz'] + >>> list(join_continuation(['foo \\', 'bar \\', 'baz'])) + ['foobarbaz'] + + Not sure why, but... + The character preceding the backslash is also elided. + + >>> list(join_continuation(['goo\\', 'dly'])) + ['godly'] + + A terrible idea, but... + If no line is available to continue, suppress the lines. + + >>> list(join_continuation(['foo', 'bar\\', 'baz\\'])) + ['foo'] + """ + lines = iter(lines) + for item in lines: + while item.endswith("\\"): + try: + item = item[:-2].strip() + next(lines) + except StopIteration: + return + yield item diff --git a/src/fetchcode/vcs/pip/_internal/utils/_log.py b/src/fetchcode/vcs/pip/_internal/utils/_log.py new file mode 100644 index 00000000..92c4c6a1 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/utils/_log.py @@ -0,0 +1,38 @@ +"""Customize logging + +Defines custom logger class for the `logger.verbose(...)` method. + +init_logging() must be called before any other modules that call logging.getLogger. +""" + +import logging +from typing import Any, cast + +# custom log level for `--verbose` output +# between DEBUG and INFO +VERBOSE = 15 + + +class VerboseLogger(logging.Logger): + """Custom Logger, defining a verbose log-level + + VERBOSE is between INFO and DEBUG. + """ + + def verbose(self, msg: str, *args: Any, **kwargs: Any) -> None: + return self.log(VERBOSE, msg, *args, **kwargs) + + +def getLogger(name: str) -> VerboseLogger: + """logging.getLogger, but ensures our VerboseLogger class is returned""" + return cast(VerboseLogger, logging.getLogger(name)) + + +def init_logging() -> None: + """Register our VerboseLogger and VERBOSE log level. + + Should be called before any calls to getLogger(), + i.e. in pip._internal.__init__ + """ + logging.setLoggerClass(VerboseLogger) + logging.addLevelName(VERBOSE, "VERBOSE") diff --git a/src/fetchcode/vcs/pip/_internal/utils/appdirs.py b/src/fetchcode/vcs/pip/_internal/utils/appdirs.py index 9b3bb4ea..16933bf8 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/appdirs.py +++ b/src/fetchcode/vcs/pip/_internal/utils/appdirs.py @@ -6,39 +6,47 @@ and eventually drop this after all usages are changed. """ -from __future__ import absolute_import - import os +import sys +from typing import List + +from pip._vendor import platformdirs as _appdirs -from fetchcode.vcs.pip._vendor import appdirs as _appdirs -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +def user_cache_dir(appname: str) -> str: + return _appdirs.user_cache_dir(appname, appauthor=False) -if MYPY_CHECK_RUNNING: - from typing import List +def _macos_user_config_dir(appname: str, roaming: bool = True) -> str: + # Use ~/Application Support/pip, if the directory exists. + path = _appdirs.user_data_dir(appname, appauthor=False, roaming=roaming) + if os.path.isdir(path): + return path -def user_cache_dir(appname): - # type: (str) -> str - return _appdirs.user_cache_dir(appname, appauthor=False) + # Use a Linux-like ~/.config/pip, by default. + linux_like_path = "~/.config/" + if appname: + linux_like_path = os.path.join(linux_like_path, appname) + return os.path.expanduser(linux_like_path) -def user_config_dir(appname, roaming=True): - # type: (str, bool) -> str - path = _appdirs.user_config_dir(appname, appauthor=False, roaming=roaming) - if _appdirs.system == "darwin" and not os.path.isdir(path): - path = os.path.expanduser('~/.config/') - if appname: - path = os.path.join(path, appname) - return path + +def user_config_dir(appname: str, roaming: bool = True) -> str: + if sys.platform == "darwin": + return _macos_user_config_dir(appname, roaming) + + return _appdirs.user_config_dir(appname, appauthor=False, roaming=roaming) # for the discussion regarding site_config_dir locations # see -def site_config_dirs(appname): - # type: (str) -> List[str] +def site_config_dirs(appname: str) -> List[str]: + if sys.platform == "darwin": + return [_appdirs.site_data_dir(appname, appauthor=False, multipath=True)] + dirval = _appdirs.site_config_dir(appname, appauthor=False, multipath=True) - if _appdirs.system not in ["win32", "darwin"]: - # always look in /etc directly as well - return dirval.split(os.pathsep) + ['/etc'] - return [dirval] + if sys.platform == "win32": + return [dirval] + + # Unix-y system. Look in /etc as well. + return dirval.split(os.pathsep) + ["/etc"] diff --git a/src/fetchcode/vcs/pip/_internal/utils/compat.py b/src/fetchcode/vcs/pip/_internal/utils/compat.py index b8e234c4..d8b54e4e 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/compat.py +++ b/src/fetchcode/vcs/pip/_internal/utils/compat.py @@ -1,166 +1,32 @@ """Stuff that differs in different Python versions and platform distributions.""" -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import, division - -import codecs -import locale +import importlib.resources import logging import os -import shutil import sys +from typing import IO -from fetchcode.vcs.pip._vendor.six import PY2, text_type - -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import Optional, Text, Tuple, Union - -try: - import ipaddress -except ImportError: - try: - from fetchcode.vcs.pip._vendor import ipaddress # type: ignore - except ImportError: - import ipaddr as ipaddress # type: ignore - ipaddress.ip_address = ipaddress.IPAddress # type: ignore - ipaddress.ip_network = ipaddress.IPNetwork # type: ignore - - -__all__ = [ - "ipaddress", "uses_pycache", "console_to_str", - "get_path_uid", "stdlib_pkgs", "WINDOWS", "samefile", "get_terminal_size", -] +__all__ = ["get_path_uid", "stdlib_pkgs", "WINDOWS"] logger = logging.getLogger(__name__) -if PY2: - import imp - try: - cache_from_source = imp.cache_from_source # type: ignore - except AttributeError: - # does not use __pycache__ - cache_from_source = None - - uses_pycache = cache_from_source is not None -else: - uses_pycache = True - from importlib.util import cache_from_source - - -if PY2: - # In Python 2.7, backslashreplace exists - # but does not support use for decoding. - # We implement our own replace handler for this - # situation, so that we can consistently use - # backslash replacement for all versions. - def backslashreplace_decode_fn(err): - raw_bytes = (err.object[i] for i in range(err.start, err.end)) - # Python 2 gave us characters - convert to numeric bytes - raw_bytes = (ord(b) for b in raw_bytes) - return u"".join(map(u"\\x{:x}".format, raw_bytes)), err.end - codecs.register_error( - "backslashreplace_decode", - backslashreplace_decode_fn, - ) - backslashreplace_decode = "backslashreplace_decode" -else: - backslashreplace_decode = "backslashreplace" - - -def has_tls(): - # type: () -> bool +def has_tls() -> bool: try: import _ssl # noqa: F401 # ignore unused + return True except ImportError: pass - from fetchcode.vcs.pip._vendor.urllib3.util import IS_PYOPENSSL - return IS_PYOPENSSL - - -def str_to_display(data, desc=None): - # type: (Union[bytes, Text], Optional[str]) -> Text - """ - For display or logging purposes, convert a bytes object (or text) to - text (e.g. unicode in Python 2) safe for output. - - :param desc: An optional phrase describing the input data, for use in - the log message if a warning is logged. Defaults to "Bytes object". + from pip._vendor.urllib3.util import IS_PYOPENSSL - This function should never error out and so can take a best effort - approach. It is okay to be lossy if needed since the return value is - just for display. - - We assume the data is in the locale preferred encoding. If it won't - decode properly, we warn the user but decode as best we can. - - We also ensure that the output can be safely written to standard output - without encoding errors. - """ - if isinstance(data, text_type): - return data - - # Otherwise, data is a bytes object (str in Python 2). - # First, get the encoding we assume. This is the preferred - # encoding for the locale, unless that is not found, or - # it is ASCII, in which case assume UTF-8 - encoding = locale.getpreferredencoding() - if (not encoding) or codecs.lookup(encoding).name == "ascii": - encoding = "utf-8" - - # Now try to decode the data - if we fail, warn the user and - # decode with replacement. - try: - decoded_data = data.decode(encoding) - except UnicodeDecodeError: - if desc is None: - desc = 'Bytes object' - msg_format = '{} does not appear to be encoded as %s'.format(desc) - logger.warning(msg_format, encoding) - decoded_data = data.decode(encoding, errors=backslashreplace_decode) - - # Make sure we can print the output, by encoding it to the output - # encoding with replacement of unencodable characters, and then - # decoding again. - # We use stderr's encoding because it's less likely to be - # redirected and if we don't find an encoding we skip this - # step (on the assumption that output is wrapped by something - # that won't fail). - # The double getattr is to deal with the possibility that we're - # being called in a situation where sys.__stderr__ doesn't exist, - # or doesn't have an encoding attribute. Neither of these cases - # should occur in normal pip use, but there's no harm in checking - # in case people use pip in (unsupported) unusual situations. - output_encoding = getattr(getattr(sys, "__stderr__", None), - "encoding", None) - - if output_encoding: - output_encoded = decoded_data.encode( - output_encoding, - errors="backslashreplace" - ) - decoded_data = output_encoded.decode(output_encoding) - - return decoded_data - - -def console_to_str(data): - # type: (bytes) -> Text - """Return a string, safe for output, of subprocess output. - """ - return str_to_display(data, desc='Subprocess output') + return IS_PYOPENSSL -def get_path_uid(path): - # type: (str) -> int +def get_path_uid(path: str) -> int: """ Return path's uid. @@ -172,7 +38,7 @@ def get_path_uid(path): :raises OSError: When path is a symlink or can't be read. """ - if hasattr(os, 'O_NOFOLLOW'): + if hasattr(os, "O_NOFOLLOW"): fd = os.open(path, os.O_RDONLY | os.O_NOFOLLOW) file_uid = os.fstat(fd).st_uid os.close(fd) @@ -183,24 +49,22 @@ def get_path_uid(path): file_uid = os.stat(path).st_uid else: # raise OSError for parity with os.O_NOFOLLOW above - raise OSError( - "{} is a symlink; Will not return uid for symlinks".format( - path) - ) + raise OSError(f"{path} is a symlink; Will not return uid for symlinks") return file_uid -def expanduser(path): - # type: (str) -> str - """ - Expand ~ and ~user constructions. +# The importlib.resources.open_text function was deprecated in 3.11 with suggested +# replacement we use below. +if sys.version_info < (3, 11): + open_text_resource = importlib.resources.open_text +else: - Includes a workaround for https://bugs.python.org/issue14768 - """ - expanded = os.path.expanduser(path) - if path.startswith('~/') and expanded.startswith('//'): - expanded = expanded[1:] - return expanded + def open_text_resource( + package: str, resource: str, encoding: str = "utf-8", errors: str = "strict" + ) -> IO[str]: + return (importlib.resources.files(package) / resource).open( + "r", encoding=encoding, errors=errors + ) # packages in the stdlib that may have installation metadata, but should not be @@ -212,59 +76,4 @@ def expanduser(path): # windows detection, covers cpython and ironpython -WINDOWS = (sys.platform.startswith("win") or - (sys.platform == 'cli' and os.name == 'nt')) - - -def samefile(file1, file2): - # type: (str, str) -> bool - """Provide an alternative for os.path.samefile on Windows/Python2""" - if hasattr(os.path, 'samefile'): - return os.path.samefile(file1, file2) - else: - path1 = os.path.normcase(os.path.abspath(file1)) - path2 = os.path.normcase(os.path.abspath(file2)) - return path1 == path2 - - -if hasattr(shutil, 'get_terminal_size'): - def get_terminal_size(): - # type: () -> Tuple[int, int] - """ - Returns a tuple (x, y) representing the width(x) and the height(y) - in characters of the terminal window. - """ - return tuple(shutil.get_terminal_size()) # type: ignore -else: - def get_terminal_size(): - # type: () -> Tuple[int, int] - """ - Returns a tuple (x, y) representing the width(x) and the height(y) - in characters of the terminal window. - """ - def ioctl_GWINSZ(fd): - try: - import fcntl - import termios - import struct - cr = struct.unpack_from( - 'hh', - fcntl.ioctl(fd, termios.TIOCGWINSZ, '12345678') - ) - except Exception: - return None - if cr == (0, 0): - return None - return cr - cr = ioctl_GWINSZ(0) or ioctl_GWINSZ(1) or ioctl_GWINSZ(2) - if not cr: - if sys.platform != "win32": - try: - fd = os.open(os.ctermid(), os.O_RDONLY) - cr = ioctl_GWINSZ(fd) - os.close(fd) - except Exception: - pass - if not cr: - cr = (os.environ.get('LINES', 25), os.environ.get('COLUMNS', 80)) - return int(cr[1]), int(cr[0]) +WINDOWS = sys.platform.startswith("win") or (sys.platform == "cli" and os.name == "nt") diff --git a/src/fetchcode/vcs/pip/_internal/utils/compatibility_tags.py b/src/fetchcode/vcs/pip/_internal/utils/compatibility_tags.py index 4fb79742..b6ed9a78 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/compatibility_tags.py +++ b/src/fetchcode/vcs/pip/_internal/utils/compatibility_tags.py @@ -1,12 +1,11 @@ """Generate and work with PEP 425 Compatibility Tags. """ -from __future__ import absolute_import - -import logging import re +from typing import List, Optional, Tuple -from fetchcode.vcs.pip._vendor.packaging.tags import ( +from pip._vendor.packaging.tags import ( + PythonVersion, Tag, compatible_tags, cpython_tags, @@ -16,26 +15,15 @@ mac_platforms, ) -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import List, Optional, Tuple - - from fetchcode.vcs.pip._vendor.packaging.tags import PythonVersion +_osx_arch_pat = re.compile(r"(.+)_(\d+)_(\d+)_(.+)") -logger = logging.getLogger(__name__) -_osx_arch_pat = re.compile(r'(.+)_(\d+)_(\d+)_(.+)') - - -def version_info_to_nodot(version_info): - # type: (Tuple[int, ...]) -> str +def version_info_to_nodot(version_info: Tuple[int, ...]) -> str: # Only use up to the first two numbers. - return ''.join(map(str, version_info[:2])) + return "".join(map(str, version_info[:2])) -def _mac_platforms(arch): - # type: (str) -> List[str] +def _mac_platforms(arch: str) -> List[str]: match = _osx_arch_pat.match(arch) if match: name, major, minor, actual_arch = match.groups() @@ -46,7 +34,7 @@ def _mac_platforms(arch): # actual prefix provided by the user in case they provided # something like "macosxcustom_". It may be good to remove # this as undocumented or deprecate it in the future. - '{}_{}'.format(name, arch[len('macosx_'):]) + "{}_{}".format(name, arch[len("macosx_") :]) for arch in mac_platforms(mac_version, actual_arch) ] else: @@ -55,91 +43,99 @@ def _mac_platforms(arch): return arches -def _custom_manylinux_platforms(arch): - # type: (str) -> List[str] +def _custom_manylinux_platforms(arch: str) -> List[str]: arches = [arch] - arch_prefix, arch_sep, arch_suffix = arch.partition('_') - if arch_prefix == 'manylinux2014': + arch_prefix, arch_sep, arch_suffix = arch.partition("_") + if arch_prefix == "manylinux2014": # manylinux1/manylinux2010 wheels run on most manylinux2014 systems # with the exception of wheels depending on ncurses. PEP 599 states # manylinux1/manylinux2010 wheels should be considered # manylinux2014 wheels: # https://www.python.org/dev/peps/pep-0599/#backwards-compatibility-with-manylinux2010-wheels - if arch_suffix in {'i686', 'x86_64'}: - arches.append('manylinux2010' + arch_sep + arch_suffix) - arches.append('manylinux1' + arch_sep + arch_suffix) - elif arch_prefix == 'manylinux2010': + if arch_suffix in {"i686", "x86_64"}: + arches.append("manylinux2010" + arch_sep + arch_suffix) + arches.append("manylinux1" + arch_sep + arch_suffix) + elif arch_prefix == "manylinux2010": # manylinux1 wheels run on most manylinux2010 systems with the # exception of wheels depending on ncurses. PEP 571 states # manylinux1 wheels should be considered manylinux2010 wheels: # https://www.python.org/dev/peps/pep-0571/#backwards-compatibility-with-manylinux1-wheels - arches.append('manylinux1' + arch_sep + arch_suffix) + arches.append("manylinux1" + arch_sep + arch_suffix) return arches -def _get_custom_platforms(arch): - # type: (str) -> List[str] - arch_prefix, arch_sep, arch_suffix = arch.partition('_') - if arch.startswith('macosx'): +def _get_custom_platforms(arch: str) -> List[str]: + arch_prefix, arch_sep, arch_suffix = arch.partition("_") + if arch.startswith("macosx"): arches = _mac_platforms(arch) - elif arch_prefix in ['manylinux2014', 'manylinux2010']: + elif arch_prefix in ["manylinux2014", "manylinux2010"]: arches = _custom_manylinux_platforms(arch) else: arches = [arch] return arches -def _get_python_version(version): - # type: (str) -> PythonVersion +def _expand_allowed_platforms(platforms: Optional[List[str]]) -> Optional[List[str]]: + if not platforms: + return None + + seen = set() + result = [] + + for p in platforms: + if p in seen: + continue + additions = [c for c in _get_custom_platforms(p) if c not in seen] + seen.update(additions) + result.extend(additions) + + return result + + +def _get_python_version(version: str) -> PythonVersion: if len(version) > 1: return int(version[0]), int(version[1:]) else: return (int(version[0]),) -def _get_custom_interpreter(implementation=None, version=None): - # type: (Optional[str], Optional[str]) -> str +def _get_custom_interpreter( + implementation: Optional[str] = None, version: Optional[str] = None +) -> str: if implementation is None: implementation = interpreter_name() if version is None: version = interpreter_version() - return "{}{}".format(implementation, version) + return f"{implementation}{version}" def get_supported( - version=None, # type: Optional[str] - platform=None, # type: Optional[str] - impl=None, # type: Optional[str] - abi=None # type: Optional[str] -): - # type: (...) -> List[Tag] + version: Optional[str] = None, + platforms: Optional[List[str]] = None, + impl: Optional[str] = None, + abis: Optional[List[str]] = None, +) -> List[Tag]: """Return a list of supported tags for each version specified in `versions`. :param version: a string version, of the form "33" or "32", or None. The version will be assumed to support our ABI. - :param platform: specify the exact platform you want valid + :param platform: specify a list of platforms you want valid tags for, or None. If None, use the local system platform. :param impl: specify the exact implementation you want valid tags for, or None. If None, use the local interpreter impl. - :param abi: specify the exact abi you want valid + :param abis: specify a list of abis you want valid tags for, or None. If None, use the local interpreter abi. """ - supported = [] # type: List[Tag] + supported: List[Tag] = [] - python_version = None # type: Optional[PythonVersion] + python_version: Optional[PythonVersion] = None if version is not None: python_version = _get_python_version(version) interpreter = _get_custom_interpreter(impl, version) - abis = None # type: Optional[List[str]] - if abi is not None: - abis = [abi] - - platforms = None # type: Optional[List[str]] - if platform is not None: - platforms = _get_custom_platforms(platform) + platforms = _expand_allowed_platforms(platforms) is_cpython = (impl or interpreter_name()) == "cp" if is_cpython: diff --git a/src/fetchcode/vcs/pip/_internal/utils/datetime.py b/src/fetchcode/vcs/pip/_internal/utils/datetime.py new file mode 100644 index 00000000..8668b3b0 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/utils/datetime.py @@ -0,0 +1,11 @@ +"""For when pip wants to check the date or time. +""" + +import datetime + + +def today_is_later_than(year: int, month: int, day: int) -> bool: + today = datetime.date.today() + given = datetime.date(year, month, day) + + return today > given diff --git a/src/fetchcode/vcs/pip/_internal/utils/deprecation.py b/src/fetchcode/vcs/pip/_internal/utils/deprecation.py index 399ef02a..0911147e 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/deprecation.py +++ b/src/fetchcode/vcs/pip/_internal/utils/deprecation.py @@ -2,22 +2,13 @@ A module that implements tooling to enable easy warnings about deprecations. """ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import - import logging import warnings +from typing import Any, Optional, TextIO, Type, Union -from fetchcode.vcs.pip._vendor.packaging.version import parse - -from pip import __version__ as current_version -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import Any, Optional +from pip._vendor.packaging.version import parse +from pip import __version__ as current_version # NOTE: tests patch this name. DEPRECATION_MSG_PREFIX = "DEPRECATION: " @@ -26,29 +17,31 @@ class PipDeprecationWarning(Warning): pass -_original_showwarning = None # type: Any +_original_showwarning: Any = None # Warnings <-> Logging Integration -def _showwarning(message, category, filename, lineno, file=None, line=None): +def _showwarning( + message: Union[Warning, str], + category: Type[Warning], + filename: str, + lineno: int, + file: Optional[TextIO] = None, + line: Optional[str] = None, +) -> None: if file is not None: if _original_showwarning is not None: - _original_showwarning( - message, category, filename, lineno, file, line, - ) + _original_showwarning(message, category, filename, lineno, file, line) elif issubclass(category, PipDeprecationWarning): # We use a specially named logger which will handle all of the # deprecation messages for pip. logger = logging.getLogger("pip._internal.deprecations") logger.warning(message) else: - _original_showwarning( - message, category, filename, lineno, file, line, - ) + _original_showwarning(message, category, filename, lineno, file, line) -def install_warning_logger(): - # type: () -> None +def install_warning_logger() -> None: # Enable our Deprecation Warnings warnings.simplefilter("default", PipDeprecationWarning, append=True) @@ -59,46 +52,73 @@ def install_warning_logger(): warnings.showwarning = _showwarning -def deprecated(reason, replacement, gone_in, issue=None): - # type: (str, Optional[str], Optional[str], Optional[int]) -> None +def deprecated( + *, + reason: str, + replacement: Optional[str], + gone_in: Optional[str], + feature_flag: Optional[str] = None, + issue: Optional[int] = None, +) -> None: """Helper to deprecate existing functionality. reason: Textual reason shown to the user about why this functionality has - been deprecated. + been deprecated. Should be a complete sentence. replacement: Textual suggestion shown to the user about what alternative functionality they can use. gone_in: The version of pip does this functionality should get removed in. - Raises errors if pip's current version is greater than or equal to + Raises an error if pip's current version is greater than or equal to this. + feature_flag: + Command-line flag of the form --use-feature={feature_flag} for testing + upcoming functionality. issue: Issue number on the tracker that would serve as a useful place for users to find related discussion and provide feedback. - - Always pass replacement, gone_in and issue as keyword arguments for clarity - at the call site. """ - # Construct a nice message. - # This is eagerly formatted as we want it to get logged as if someone - # typed this entire message out. - sentences = [ - (reason, DEPRECATION_MSG_PREFIX + "{}"), - (gone_in, "pip {} will remove support for this functionality."), - (replacement, "A possible replacement is {}."), - (issue, ( - "You can find discussion regarding this at " - "https://github.com/pypa/pip/issues/{}." - )), + # Determine whether or not the feature is already gone in this version. + is_gone = gone_in is not None and parse(current_version) >= parse(gone_in) + + message_parts = [ + (reason, f"{DEPRECATION_MSG_PREFIX}{{}}"), + ( + gone_in, + ( + "pip {} will enforce this behaviour change." + if not is_gone + else "Since pip {}, this is no longer supported." + ), + ), + ( + replacement, + "A possible replacement is {}.", + ), + ( + feature_flag, + ( + "You can use the flag --use-feature={} to test the upcoming behaviour." + if not is_gone + else None + ), + ), + ( + issue, + "Discussion can be found at https://github.com/pypa/pip/issues/{}", + ), ] + message = " ".join( - template.format(val) for val, template in sentences if val is not None + format_str.format(value) + for value, format_str in message_parts + if format_str is not None and value is not None ) - # Raise as an error if it has to be removed. - if gone_in is not None and parse(current_version) >= parse(gone_in): + # Raise as an error if this behaviour is deprecated. + if is_gone: raise PipDeprecationWarning(message) warnings.warn(message, category=PipDeprecationWarning, stacklevel=2) diff --git a/src/fetchcode/vcs/pip/_internal/utils/direct_url_helpers.py b/src/fetchcode/vcs/pip/_internal/utils/direct_url_helpers.py index 79be568d..66020d39 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/direct_url_helpers.py +++ b/src/fetchcode/vcs/pip/_internal/utils/direct_url_helpers.py @@ -1,41 +1,19 @@ -import logging +from typing import Optional -from fetchcode.vcs.pip._internal.models.direct_url import ( - DIRECT_URL_METADATA_NAME, - ArchiveInfo, - DirectUrl, - DirectUrlValidationError, - DirInfo, - VcsInfo, -) -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.vcs import vcs +from pip._internal.models.direct_url import ArchiveInfo, DirectUrl, DirInfo, VcsInfo +from pip._internal.models.link import Link +from pip._internal.utils.urls import path_to_url +from pip._internal.vcs import vcs -try: - from json import JSONDecodeError -except ImportError: - # PY2 - JSONDecodeError = ValueError # type: ignore -if MYPY_CHECK_RUNNING: - from typing import Optional - - from fetchcode.vcs.pip._internal.models.link import Link - - from fetchcode.vcs.pip._vendor.pkg_resources import Distribution - -logger = logging.getLogger(__name__) - - -def direct_url_as_pep440_direct_reference(direct_url, name): - # type: (DirectUrl, str) -> str +def direct_url_as_pep440_direct_reference(direct_url: DirectUrl, name: str) -> str: """Convert a DirectUrl to a pip requirement string.""" direct_url.validate() # if invalid, this is a pip bug requirement = name + " @ " fragments = [] if isinstance(direct_url.info, VcsInfo): - requirement += "{}+{}@{}".format( - direct_url.info.vcs, direct_url.url, direct_url.info.commit_id + requirement += ( + f"{direct_url.info.vcs}+{direct_url.url}@{direct_url.info.commit_id}" ) elif isinstance(direct_url.info, ArchiveInfo): requirement += direct_url.url @@ -43,10 +21,6 @@ def direct_url_as_pep440_direct_reference(direct_url, name): fragments.append(direct_url.info.hash) else: assert isinstance(direct_url.info, DirInfo) - # pip should never reach this point for editables, since - # pip freeze inspects the editable project location to produce - # the requirement string - assert not direct_url.info.editable requirement += direct_url.url if direct_url.subdirectory: fragments.append("subdirectory=" + direct_url.subdirectory) @@ -55,13 +29,21 @@ def direct_url_as_pep440_direct_reference(direct_url, name): return requirement -def direct_url_from_link(link, source_dir=None, link_is_in_wheel_cache=False): - # type: (Link, Optional[str], bool) -> DirectUrl +def direct_url_for_editable(source_dir: str) -> DirectUrl: + return DirectUrl( + url=path_to_url(source_dir), + info=DirInfo(editable=True), + ) + + +def direct_url_from_link( + link: Link, source_dir: Optional[str] = None, link_is_in_wheel_cache: bool = False +) -> DirectUrl: if link.is_vcs: vcs_backend = vcs.get_backend_for_scheme(link.scheme) assert vcs_backend - url, requested_revision, _ = ( - vcs_backend.get_url_rev_and_auth(link.url_without_fragment) + url, requested_revision, _ = vcs_backend.get_url_rev_and_auth( + link.url_without_fragment ) # For VCS links, we need to find out and add commit_id. if link_is_in_wheel_cache: @@ -97,34 +79,9 @@ def direct_url_from_link(link, source_dir=None, link_is_in_wheel_cache=False): hash = None hash_name = link.hash_name if hash_name: - hash = "{}={}".format(hash_name, link.hash) + hash = f"{hash_name}={link.hash}" return DirectUrl( url=link.url_without_fragment, info=ArchiveInfo(hash=hash), subdirectory=link.subdirectory_fragment, ) - - -def dist_get_direct_url(dist): - # type: (Distribution) -> Optional[DirectUrl] - """Obtain a DirectUrl from a pkg_resource.Distribution. - - Returns None if the distribution has no `direct_url.json` metadata, - or if `direct_url.json` is invalid. - """ - if not dist.has_metadata(DIRECT_URL_METADATA_NAME): - return None - try: - return DirectUrl.from_json(dist.get_metadata(DIRECT_URL_METADATA_NAME)) - except ( - DirectUrlValidationError, - JSONDecodeError, - UnicodeDecodeError - ) as e: - logger.warning( - "Error parsing %s for %s: %s", - DIRECT_URL_METADATA_NAME, - dist.project_name, - e, - ) - return None diff --git a/src/fetchcode/vcs/pip/_internal/utils/distutils_args.py b/src/fetchcode/vcs/pip/_internal/utils/distutils_args.py deleted file mode 100644 index 239f2c09..00000000 --- a/src/fetchcode/vcs/pip/_internal/utils/distutils_args.py +++ /dev/null @@ -1,48 +0,0 @@ -from distutils.errors import DistutilsArgError -from distutils.fancy_getopt import FancyGetopt - -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import Dict, List - - -_options = [ - ("exec-prefix=", None, ""), - ("home=", None, ""), - ("install-base=", None, ""), - ("install-data=", None, ""), - ("install-headers=", None, ""), - ("install-lib=", None, ""), - ("install-platlib=", None, ""), - ("install-purelib=", None, ""), - ("install-scripts=", None, ""), - ("prefix=", None, ""), - ("root=", None, ""), - ("user", None, ""), -] - - -# typeshed doesn't permit Tuple[str, None, str], see python/typeshed#3469. -_distutils_getopt = FancyGetopt(_options) # type: ignore - - -def parse_distutils_args(args): - # type: (List[str]) -> Dict[str, str] - """Parse provided arguments, returning an object that has the - matched arguments. - - Any unknown arguments are ignored. - """ - result = {} - for arg in args: - try: - _, match = _distutils_getopt.getopt(args=[arg]) - except DistutilsArgError: - # We don't care about any other options, which here may be - # considered unrecognized since our option list is not - # exhaustive. - pass - else: - result.update(match.__dict__) - return result diff --git a/src/fetchcode/vcs/pip/_internal/utils/egg_link.py b/src/fetchcode/vcs/pip/_internal/utils/egg_link.py new file mode 100644 index 00000000..4a384a63 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/utils/egg_link.py @@ -0,0 +1,80 @@ +import os +import re +import sys +from typing import List, Optional + +from pip._internal.locations import site_packages, user_site +from pip._internal.utils.virtualenv import ( + running_under_virtualenv, + virtualenv_no_global, +) + +__all__ = [ + "egg_link_path_from_sys_path", + "egg_link_path_from_location", +] + + +def _egg_link_names(raw_name: str) -> List[str]: + """ + Convert a Name metadata value to a .egg-link name, by applying + the same substitution as pkg_resources's safe_name function. + Note: we cannot use canonicalize_name because it has a different logic. + + We also look for the raw name (without normalization) as setuptools 69 changed + the way it names .egg-link files (https://github.com/pypa/setuptools/issues/4167). + """ + return [ + re.sub("[^A-Za-z0-9.]+", "-", raw_name) + ".egg-link", + f"{raw_name}.egg-link", + ] + + +def egg_link_path_from_sys_path(raw_name: str) -> Optional[str]: + """ + Look for a .egg-link file for project name, by walking sys.path. + """ + egg_link_names = _egg_link_names(raw_name) + for path_item in sys.path: + for egg_link_name in egg_link_names: + egg_link = os.path.join(path_item, egg_link_name) + if os.path.isfile(egg_link): + return egg_link + return None + + +def egg_link_path_from_location(raw_name: str) -> Optional[str]: + """ + Return the path for the .egg-link file if it exists, otherwise, None. + + There's 3 scenarios: + 1) not in a virtualenv + try to find in site.USER_SITE, then site_packages + 2) in a no-global virtualenv + try to find in site_packages + 3) in a yes-global virtualenv + try to find in site_packages, then site.USER_SITE + (don't look in global location) + + For #1 and #3, there could be odd cases, where there's an egg-link in 2 + locations. + + This method will just return the first one found. + """ + sites: List[str] = [] + if running_under_virtualenv(): + sites.append(site_packages) + if not virtualenv_no_global() and user_site: + sites.append(user_site) + else: + if user_site: + sites.append(user_site) + sites.append(site_packages) + + egg_link_names = _egg_link_names(raw_name) + for site in sites: + for egg_link_name in egg_link_names: + egglink = os.path.join(site, egg_link_name) + if os.path.isfile(egglink): + return egglink + return None diff --git a/src/fetchcode/vcs/pip/_internal/utils/encoding.py b/src/fetchcode/vcs/pip/_internal/utils/encoding.py index f4b972a1..008f06a7 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/encoding.py +++ b/src/fetchcode/vcs/pip/_internal/utils/encoding.py @@ -1,41 +1,35 @@ -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False - import codecs import locale import re import sys +from typing import List, Tuple -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import List, Tuple, Text - -BOMS = [ - (codecs.BOM_UTF8, 'utf-8'), - (codecs.BOM_UTF16, 'utf-16'), - (codecs.BOM_UTF16_BE, 'utf-16-be'), - (codecs.BOM_UTF16_LE, 'utf-16-le'), - (codecs.BOM_UTF32, 'utf-32'), - (codecs.BOM_UTF32_BE, 'utf-32-be'), - (codecs.BOM_UTF32_LE, 'utf-32-le'), -] # type: List[Tuple[bytes, Text]] +BOMS: List[Tuple[bytes, str]] = [ + (codecs.BOM_UTF8, "utf-8"), + (codecs.BOM_UTF16, "utf-16"), + (codecs.BOM_UTF16_BE, "utf-16-be"), + (codecs.BOM_UTF16_LE, "utf-16-le"), + (codecs.BOM_UTF32, "utf-32"), + (codecs.BOM_UTF32_BE, "utf-32-be"), + (codecs.BOM_UTF32_LE, "utf-32-le"), +] -ENCODING_RE = re.compile(br'coding[:=]\s*([-\w.]+)') +ENCODING_RE = re.compile(rb"coding[:=]\s*([-\w.]+)") -def auto_decode(data): - # type: (bytes) -> Text +def auto_decode(data: bytes) -> str: """Check a bytes string for a BOM to correctly detect the encoding Fallback to locale.getpreferredencoding(False) like open() on Python3""" for bom, encoding in BOMS: if data.startswith(bom): - return data[len(bom):].decode(encoding) + return data[len(bom) :].decode(encoding) # Lets check the first two lines as in PEP263 - for line in data.split(b'\n')[:2]: - if line[0:1] == b'#' and ENCODING_RE.search(line): - encoding = ENCODING_RE.search(line).groups()[0].decode('ascii') + for line in data.split(b"\n")[:2]: + if line[0:1] == b"#" and ENCODING_RE.search(line): + result = ENCODING_RE.search(line) + assert result is not None + encoding = result.groups()[0].decode("ascii") return data.decode(encoding) return data.decode( locale.getpreferredencoding(False) or sys.getdefaultencoding(), diff --git a/src/fetchcode/vcs/pip/_internal/utils/entrypoints.py b/src/fetchcode/vcs/pip/_internal/utils/entrypoints.py index be139101..15013693 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/entrypoints.py +++ b/src/fetchcode/vcs/pip/_internal/utils/entrypoints.py @@ -1,14 +1,26 @@ +import itertools +import os +import shutil import sys +from typing import List, Optional -from fetchcode.vcs.pip._internal.cli.main import main -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._internal.cli.main import main +from pip._internal.utils.compat import WINDOWS -if MYPY_CHECK_RUNNING: - from typing import Optional, List +_EXECUTABLE_NAMES = [ + "pip", + f"pip{sys.version_info.major}", + f"pip{sys.version_info.major}.{sys.version_info.minor}", +] +if WINDOWS: + _allowed_extensions = {"", ".exe"} + _EXECUTABLE_NAMES = [ + "".join(parts) + for parts in itertools.product(_EXECUTABLE_NAMES, _allowed_extensions) + ] -def _wrapper(args=None): - # type: (Optional[List[str]]) -> int +def _wrapper(args: Optional[List[str]] = None) -> int: """Central wrapper for all old entrypoints. Historically pip has had several entrypoints defined. Because of issues @@ -29,3 +41,44 @@ def _wrapper(args=None): "running pip directly.\n" ) return main(args) + + +def get_best_invocation_for_this_pip() -> str: + """Try to figure out the best way to invoke pip in the current environment.""" + binary_directory = "Scripts" if WINDOWS else "bin" + binary_prefix = os.path.join(sys.prefix, binary_directory) + + # Try to use pip[X[.Y]] names, if those executables for this environment are + # the first on PATH with that name. + path_parts = os.path.normcase(os.environ.get("PATH", "")).split(os.pathsep) + exe_are_in_PATH = os.path.normcase(binary_prefix) in path_parts + if exe_are_in_PATH: + for exe_name in _EXECUTABLE_NAMES: + found_executable = shutil.which(exe_name) + binary_executable = os.path.join(binary_prefix, exe_name) + if ( + found_executable + and os.path.exists(binary_executable) + and os.path.samefile( + found_executable, + binary_executable, + ) + ): + return exe_name + + # Use the `-m` invocation, if there's no "nice" invocation. + return f"{get_best_invocation_for_this_python()} -m pip" + + +def get_best_invocation_for_this_python() -> str: + """Try to figure out the best way to invoke the current Python.""" + exe = sys.executable + exe_name = os.path.basename(exe) + + # Try to use the basename, if it's the first executable. + found_executable = shutil.which(exe_name) + if found_executable and os.path.samefile(found_executable, exe): + return exe_name + + # Use the full executable name, because we couldn't find something simpler. + return exe diff --git a/src/fetchcode/vcs/pip/_internal/utils/filesystem.py b/src/fetchcode/vcs/pip/_internal/utils/filesystem.py index afd9d215..22e356cd 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/filesystem.py +++ b/src/fetchcode/vcs/pip/_internal/utils/filesystem.py @@ -1,35 +1,18 @@ -import errno import fnmatch import os import os.path import random -import shutil -import stat import sys from contextlib import contextmanager from tempfile import NamedTemporaryFile +from typing import Any, BinaryIO, Generator, List, Union, cast -# NOTE: retrying is not annotated in typeshed as on 2017-07-17, which is -# why we ignore the type on this import. -from fetchcode.vcs.pip._vendor.retrying import retry # type: ignore -from fetchcode.vcs.pip._vendor.six import PY2 +from pip._internal.utils.compat import get_path_uid +from pip._internal.utils.misc import format_size +from pip._internal.utils.retry import retry -from fetchcode.vcs.pip._internal.utils.compat import get_path_uid -from fetchcode.vcs.pip._internal.utils.misc import format_size -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING, cast -if MYPY_CHECK_RUNNING: - from typing import Any, BinaryIO, Iterator, List, Union - - class NamedTemporaryFileResult(BinaryIO): - @property - def file(self): - # type: () -> BinaryIO - pass - - -def check_path_owner(path): - # type: (str) -> bool +def check_path_owner(path: str) -> bool: # If we don't have a way to check the effective uid of this process, then # we'll just assume that we own the directory. if sys.platform == "win32" or not hasattr(os, "geteuid"): @@ -56,39 +39,8 @@ def check_path_owner(path): return False # assume we don't own the path -def copy2_fixed(src, dest): - # type: (str, str) -> None - """Wrap shutil.copy2() but map errors copying socket files to - SpecialFileError as expected. - - See also https://bugs.python.org/issue37700. - """ - try: - shutil.copy2(src, dest) - except (OSError, IOError): - for f in [src, dest]: - try: - is_socket_file = is_socket(f) - except OSError: - # An error has already occurred. Another error here is not - # a problem and we can ignore it. - pass - else: - if is_socket_file: - raise shutil.SpecialFileError( - "`{f}` is a socket".format(**locals())) - - raise - - -def is_socket(path): - # type: (str) -> bool - return stat.S_ISSOCK(os.lstat(path).st_mode) - - @contextmanager -def adjacent_tmp_file(path, **kwargs): - # type: (str, **Any) -> Iterator[NamedTemporaryFileResult] +def adjacent_tmp_file(path: str, **kwargs: Any) -> Generator[BinaryIO, None, None]: """Return a file-like object pointing to a tmp file next to path. The file is created securely and is ensured to be written to disk @@ -101,37 +53,23 @@ def adjacent_tmp_file(path, **kwargs): delete=False, dir=os.path.dirname(path), prefix=os.path.basename(path), - suffix='.tmp', - **kwargs + suffix=".tmp", + **kwargs, ) as f: - result = cast('NamedTemporaryFileResult', f) + result = cast(BinaryIO, f) try: yield result finally: - result.file.flush() - os.fsync(result.file.fileno()) + result.flush() + os.fsync(result.fileno()) -_replace_retry = retry(stop_max_delay=1000, wait_fixed=250) - -if PY2: - @_replace_retry - def replace(src, dest): - # type: (str, str) -> None - try: - os.rename(src, dest) - except OSError: - os.remove(dest) - os.rename(src, dest) - -else: - replace = _replace_retry(os.replace) +replace = retry(stop_after_delay=1, wait=0.25)(os.replace) # test_writable_dir and _test_writable_dir_win are copied from Flit, # with the author's agreement to also place them under pip's license. -def test_writable_dir(path): - # type: (str) -> bool +def test_writable_dir(path: str) -> bool: """Check if a directory is writable. Uses os.access() on POSIX, tries creating files on Windows. @@ -143,72 +81,62 @@ def test_writable_dir(path): break # Should never get here, but infinite loops are bad path = parent - if os.name == 'posix': + if os.name == "posix": return os.access(path, os.W_OK) return _test_writable_dir_win(path) -def _test_writable_dir_win(path): - # type: (str) -> bool +def _test_writable_dir_win(path: str) -> bool: # os.access doesn't work on Windows: http://bugs.python.org/issue2528 # and we can't use tempfile: http://bugs.python.org/issue22107 - basename = 'accesstest_deleteme_fishfingers_custard_' - alphabet = 'abcdefghijklmnopqrstuvwxyz0123456789' - for i in range(10): - name = basename + ''.join(random.choice(alphabet) for _ in range(6)) + basename = "accesstest_deleteme_fishfingers_custard_" + alphabet = "abcdefghijklmnopqrstuvwxyz0123456789" + for _ in range(10): + name = basename + "".join(random.choice(alphabet) for _ in range(6)) file = os.path.join(path, name) try: fd = os.open(file, os.O_RDWR | os.O_CREAT | os.O_EXCL) - # Python 2 doesn't support FileExistsError and PermissionError. - except OSError as e: - # exception FileExistsError - if e.errno == errno.EEXIST: - continue - # exception PermissionError - if e.errno == errno.EPERM or e.errno == errno.EACCES: - # This could be because there's a directory with the same name. - # But it's highly unlikely there's a directory called that, - # so we'll assume it's because the parent dir is not writable. - return False - raise + except FileExistsError: + pass + except PermissionError: + # This could be because there's a directory with the same name. + # But it's highly unlikely there's a directory called that, + # so we'll assume it's because the parent dir is not writable. + # This could as well be because the parent dir is not readable, + # due to non-privileged user access. + return False else: os.close(fd) os.unlink(file) return True # This should never be reached - raise EnvironmentError( - 'Unexpected condition testing for writable directory' - ) + raise OSError("Unexpected condition testing for writable directory") -def find_files(path, pattern): - # type: (str, str) -> List[str] +def find_files(path: str, pattern: str) -> List[str]: """Returns a list of absolute paths of files beneath path, recursively, with filenames which match the UNIX-style shell glob pattern.""" - result = [] # type: List[str] - for root, dirs, files in os.walk(path): + result: List[str] = [] + for root, _, files in os.walk(path): matches = fnmatch.filter(files, pattern) result.extend(os.path.join(root, f) for f in matches) return result -def file_size(path): - # type: (str) -> Union[int, float] +def file_size(path: str) -> Union[int, float]: # If it's a symlink, return 0. if os.path.islink(path): return 0 return os.path.getsize(path) -def format_file_size(path): - # type: (str) -> str +def format_file_size(path: str) -> str: return format_size(file_size(path)) -def directory_size(path): - # type: (str) -> Union[int, float] +def directory_size(path: str) -> Union[int, float]: size = 0.0 for root, _dirs, files in os.walk(path): for filename in files: @@ -217,6 +145,5 @@ def directory_size(path): return size -def format_directory_size(path): - # type: (str) -> str +def format_directory_size(path: str) -> str: return format_size(directory_size(path)) diff --git a/src/fetchcode/vcs/pip/_internal/utils/filetypes.py b/src/fetchcode/vcs/pip/_internal/utils/filetypes.py index cc9341e7..59485701 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/filetypes.py +++ b/src/fetchcode/vcs/pip/_internal/utils/filetypes.py @@ -1,16 +1,27 @@ """Filetype information. """ -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -if MYPY_CHECK_RUNNING: - from typing import Tuple +from typing import Tuple -WHEEL_EXTENSION = '.whl' -BZ2_EXTENSIONS = ('.tar.bz2', '.tbz') # type: Tuple[str, ...] -XZ_EXTENSIONS = ('.tar.xz', '.txz', '.tlz', - '.tar.lz', '.tar.lzma') # type: Tuple[str, ...] -ZIP_EXTENSIONS = ('.zip', WHEEL_EXTENSION) # type: Tuple[str, ...] -TAR_EXTENSIONS = ('.tar.gz', '.tgz', '.tar') # type: Tuple[str, ...] -ARCHIVE_EXTENSIONS = ( - ZIP_EXTENSIONS + BZ2_EXTENSIONS + TAR_EXTENSIONS + XZ_EXTENSIONS +from pip._internal.utils.misc import splitext + +WHEEL_EXTENSION = ".whl" +BZ2_EXTENSIONS: Tuple[str, ...] = (".tar.bz2", ".tbz") +XZ_EXTENSIONS: Tuple[str, ...] = ( + ".tar.xz", + ".txz", + ".tlz", + ".tar.lz", + ".tar.lzma", ) +ZIP_EXTENSIONS: Tuple[str, ...] = (".zip", WHEEL_EXTENSION) +TAR_EXTENSIONS: Tuple[str, ...] = (".tar.gz", ".tgz", ".tar") +ARCHIVE_EXTENSIONS = ZIP_EXTENSIONS + BZ2_EXTENSIONS + TAR_EXTENSIONS + XZ_EXTENSIONS + + +def is_archive_file(name: str) -> bool: + """Return True if `name` is a considered as an archive file.""" + ext = splitext(name)[1].lower() + if ext in ARCHIVE_EXTENSIONS: + return True + return False diff --git a/src/fetchcode/vcs/pip/_internal/utils/glibc.py b/src/fetchcode/vcs/pip/_internal/utils/glibc.py index 2bccd69a..998868ff 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/glibc.py +++ b/src/fetchcode/vcs/pip/_internal/utils/glibc.py @@ -1,25 +1,14 @@ -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False - -from __future__ import absolute_import - import os import sys - -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import Optional, Tuple +from typing import Optional, Tuple -def glibc_version_string(): - # type: () -> Optional[str] +def glibc_version_string() -> Optional[str]: "Returns glibc version string, or None if not using glibc." return glibc_version_string_confstr() or glibc_version_string_ctypes() -def glibc_version_string_confstr(): - # type: () -> Optional[str] +def glibc_version_string_confstr() -> Optional[str]: "Primary implementation of glibc_version_string using os.confstr." # os.confstr is quite a bit faster than ctypes.DLL. It's also less likely # to be broken or missing. This strategy is used in the standard library @@ -28,16 +17,18 @@ def glibc_version_string_confstr(): if sys.platform == "win32": return None try: + gnu_libc_version = os.confstr("CS_GNU_LIBC_VERSION") + if gnu_libc_version is None: + return None # os.confstr("CS_GNU_LIBC_VERSION") returns a string like "glibc 2.17": - _, version = os.confstr("CS_GNU_LIBC_VERSION").split() + _, version = gnu_libc_version.split() except (AttributeError, OSError, ValueError): # os.confstr() or CS_GNU_LIBC_VERSION not available (or a bad value)... return None return version -def glibc_version_string_ctypes(): - # type: () -> Optional[str] +def glibc_version_string_ctypes() -> Optional[str]: "Fallback implementation of glibc_version_string using ctypes." try: @@ -49,7 +40,20 @@ def glibc_version_string_ctypes(): # manpage says, "If filename is NULL, then the returned handle is for the # main program". This way we can let the linker do the work to figure out # which libc our process is actually using. - process_namespace = ctypes.CDLL(None) + # + # We must also handle the special case where the executable is not a + # dynamically linked executable. This can occur when using musl libc, + # for example. In this situation, dlopen() will error, leading to an + # OSError. Interestingly, at least in the case of musl, there is no + # errno set on the OSError. The single string argument used to construct + # OSError comes from libc itself and is therefore not portable to + # hard code here. In any case, failure to call dlopen() means we + # can't proceed, so we bail on our attempt. + try: + process_namespace = ctypes.CDLL(None) + except OSError: + return None + try: gnu_get_libc_version = process_namespace.gnu_get_libc_version except AttributeError: @@ -59,7 +63,7 @@ def glibc_version_string_ctypes(): # Call gnu_get_libc_version, which returns a string like "2.5" gnu_get_libc_version.restype = ctypes.c_char_p - version_str = gnu_get_libc_version() + version_str: str = gnu_get_libc_version() # py2 / py3 compatibility: if not isinstance(version_str, str): version_str = version_str.decode("ascii") @@ -84,8 +88,7 @@ def glibc_version_string_ctypes(): # versions that was generated by pip 8.1.2 and earlier is useless and # misleading. Solution: instead of using platform, use our code that actually # works. -def libc_ver(): - # type: () -> Tuple[str, str] +def libc_ver() -> Tuple[str, str]: """Try to determine the glibc version Returns a tuple of strings (lib, version) which default to empty strings diff --git a/src/fetchcode/vcs/pip/_internal/utils/hashes.py b/src/fetchcode/vcs/pip/_internal/utils/hashes.py index d7995948..535e94fc 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/hashes.py +++ b/src/fetchcode/vcs/pip/_internal/utils/hashes.py @@ -1,67 +1,69 @@ -from __future__ import absolute_import - import hashlib +from typing import TYPE_CHECKING, BinaryIO, Dict, Iterable, List, NoReturn, Optional -from fetchcode.vcs.pip._vendor.six import iteritems, iterkeys, itervalues - -from fetchcode.vcs.pip._internal.exceptions import ( - HashMismatch, - HashMissing, - InstallationError, -) -from fetchcode.vcs.pip._internal.utils.misc import read_chunks -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from pip._internal.exceptions import HashMismatch, HashMissing, InstallationError +from pip._internal.utils.misc import read_chunks -if MYPY_CHECK_RUNNING: - from typing import ( - Dict, List, BinaryIO, NoReturn, Iterator - ) - from fetchcode.vcs.pip._vendor.six import PY3 - if PY3: - from hashlib import _Hash - else: - from hashlib import _hash as _Hash +if TYPE_CHECKING: + from hashlib import _Hash # The recommended hash algo of the moment. Change this whenever the state of # the art changes; it won't hurt backward compatibility. -FAVORITE_HASH = 'sha256' +FAVORITE_HASH = "sha256" # Names of hashlib algorithms allowed by the --hash option and ``pip hash`` # Currently, those are the ones at least as collision-resistant as sha256. -STRONG_HASHES = ['sha256', 'sha384', 'sha512'] +STRONG_HASHES = ["sha256", "sha384", "sha512"] -class Hashes(object): +class Hashes: """A wrapper that builds multiple hashes at once and checks them against known-good values """ - def __init__(self, hashes=None): - # type: (Dict[str, List[str]]) -> None + + def __init__(self, hashes: Optional[Dict[str, List[str]]] = None) -> None: """ :param hashes: A dict of algorithm names pointing to lists of allowed hex digests """ - self._allowed = {} if hashes is None else hashes + allowed = {} + if hashes is not None: + for alg, keys in hashes.items(): + # Make sure values are always sorted (to ease equality checks) + allowed[alg] = [k.lower() for k in sorted(keys)] + self._allowed = allowed + + def __and__(self, other: "Hashes") -> "Hashes": + if not isinstance(other, Hashes): + return NotImplemented + + # If either of the Hashes object is entirely empty (i.e. no hash + # specified at all), all hashes from the other object are allowed. + if not other: + return self + if not self: + return other + + # Otherwise only hashes that present in both objects are allowed. + new = {} + for alg, values in other._allowed.items(): + if alg not in self._allowed: + continue + new[alg] = [v for v in values if v in self._allowed[alg]] + return Hashes(new) @property - def digest_count(self): - # type: () -> int + def digest_count(self) -> int: return sum(len(digests) for digests in self._allowed.values()) - def is_hash_allowed( - self, - hash_name, # type: str - hex_digest, # type: str - ): - # type: (...) -> bool + def is_hash_allowed(self, hash_name: str, hex_digest: str) -> bool: """Return whether the given hex digest is allowed.""" return hex_digest in self._allowed.get(hash_name, []) - def check_against_chunks(self, chunks): - # type: (Iterator[bytes]) -> None + def check_against_chunks(self, chunks: Iterable[bytes]) -> None: """Check good hashes against ones built from iterable of chunks of data. @@ -69,29 +71,25 @@ def check_against_chunks(self, chunks): """ gots = {} - for hash_name in iterkeys(self._allowed): + for hash_name in self._allowed.keys(): try: gots[hash_name] = hashlib.new(hash_name) except (ValueError, TypeError): - raise InstallationError( - 'Unknown hash name: {}'.format(hash_name) - ) + raise InstallationError(f"Unknown hash name: {hash_name}") for chunk in chunks: - for hash in itervalues(gots): + for hash in gots.values(): hash.update(chunk) - for hash_name, got in iteritems(gots): + for hash_name, got in gots.items(): if got.hexdigest() in self._allowed[hash_name]: return self._raise(gots) - def _raise(self, gots): - # type: (Dict[str, _Hash]) -> NoReturn + def _raise(self, gots: Dict[str, "_Hash"]) -> "NoReturn": raise HashMismatch(self._allowed, gots) - def check_against_file(self, file): - # type: (BinaryIO) -> None + def check_against_file(self, file: BinaryIO) -> None: """Check good hashes against a file-like object Raise HashMismatch if none match. @@ -99,19 +97,36 @@ def check_against_file(self, file): """ return self.check_against_chunks(read_chunks(file)) - def check_against_path(self, path): - # type: (str) -> None - with open(path, 'rb') as file: + def check_against_path(self, path: str) -> None: + with open(path, "rb") as file: return self.check_against_file(file) - def __nonzero__(self): - # type: () -> bool + def has_one_of(self, hashes: Dict[str, str]) -> bool: + """Return whether any of the given hashes are allowed.""" + for hash_name, hex_digest in hashes.items(): + if self.is_hash_allowed(hash_name, hex_digest): + return True + return False + + def __bool__(self) -> bool: """Return whether I know any known-good hashes.""" return bool(self._allowed) - def __bool__(self): - # type: () -> bool - return self.__nonzero__() + def __eq__(self, other: object) -> bool: + if not isinstance(other, Hashes): + return NotImplemented + return self._allowed == other._allowed + + def __hash__(self) -> int: + return hash( + ",".join( + sorted( + ":".join((alg, digest)) + for alg, digest_list in self._allowed.items() + for digest in digest_list + ) + ) + ) class MissingHashes(Hashes): @@ -121,13 +136,12 @@ class MissingHashes(Hashes): exception showing it to the user. """ - def __init__(self): - # type: () -> None + + def __init__(self) -> None: """Don't offer the ``hashes`` kwarg.""" # Pass our favorite hash in to generate a "gotten hash". With the # empty list, it will never match, so an error will always raise. - super(MissingHashes, self).__init__(hashes={FAVORITE_HASH: []}) + super().__init__(hashes={FAVORITE_HASH: []}) - def _raise(self, gots): - # type: (Dict[str, _Hash]) -> NoReturn + def _raise(self, gots: Dict[str, "_Hash"]) -> "NoReturn": raise HashMissing(gots[FAVORITE_HASH].hexdigest()) diff --git a/src/fetchcode/vcs/pip/_internal/utils/inject_securetransport.py b/src/fetchcode/vcs/pip/_internal/utils/inject_securetransport.py deleted file mode 100644 index 328754c7..00000000 --- a/src/fetchcode/vcs/pip/_internal/utils/inject_securetransport.py +++ /dev/null @@ -1,36 +0,0 @@ -"""A helper module that injects SecureTransport, on import. - -The import should be done as early as possible, to ensure all requests and -sessions (or whatever) are created after injecting SecureTransport. - -Note that we only do the injection on macOS, when the linked OpenSSL is too -old to handle TLSv1.2. -""" - -import sys - - -def inject_securetransport(): - # type: () -> None - # Only relevant on macOS - if sys.platform != "darwin": - return - - try: - import ssl - except ImportError: - return - - # Checks for OpenSSL 1.0.1 - if ssl.OPENSSL_VERSION_NUMBER >= 0x1000100f: - return - - try: - from fetchcode.vcs.pip._vendor.urllib3.contrib import securetransport - except (ImportError, OSError): - return - - securetransport.inject_into_urllib3() - - -inject_securetransport() diff --git a/src/fetchcode/vcs/pip/_internal/utils/logging.py b/src/fetchcode/vcs/pip/_internal/utils/logging.py index 3926bdf7..41f6eb51 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/logging.py +++ b/src/fetchcode/vcs/pip/_internal/utils/logging.py @@ -1,104 +1,58 @@ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import - import contextlib import errno import logging import logging.handlers import os import sys -from logging import Filter, getLogger - -from fetchcode.vcs.pip._vendor.six import PY2 - -from fetchcode.vcs.pip._internal.utils.compat import WINDOWS -from fetchcode.vcs.pip._internal.utils.deprecation import DEPRECATION_MSG_PREFIX -from fetchcode.vcs.pip._internal.utils.misc import ensure_dir - -try: - import threading -except ImportError: - import dummy_threading as threading # type: ignore - - -try: - # Use "import as" and set colorama in the else clause to avoid mypy - # errors and get the following correct revealed type for colorama: - # `Union[_importlib_modulespec.ModuleType, None]` - # Otherwise, we get an error like the following in the except block: - # > Incompatible types in assignment (expression has type "None", - # variable has type Module) - # TODO: eliminate the need to use "import as" once mypy addresses some - # of its issues with conditional imports. Here is an umbrella issue: - # https://github.com/python/mypy/issues/1297 - from fetchcode.vcs.pip._vendor import colorama as _colorama -# Lots of different errors can come from this, including SystemError and -# ImportError. -except Exception: - colorama = None -else: - # Import Fore explicitly rather than accessing below as colorama.Fore - # to avoid the following error running mypy: - # > Module has no attribute "Fore" - # TODO: eliminate the need to import Fore once mypy addresses some of its - # issues with conditional imports. This particular case could be an - # instance of the following issue (but also see the umbrella issue above): - # https://github.com/python/mypy/issues/3500 - from fetchcode.vcs.pip._vendor.colorama import Fore - - colorama = _colorama - +import threading +from dataclasses import dataclass +from io import TextIOWrapper +from logging import Filter +from typing import Any, ClassVar, Generator, List, Optional, TextIO, Type + +from pip._vendor.rich.console import ( + Console, + ConsoleOptions, + ConsoleRenderable, + RenderableType, + RenderResult, + RichCast, +) +from pip._vendor.rich.highlighter import NullHighlighter +from pip._vendor.rich.logging import RichHandler +from pip._vendor.rich.segment import Segment +from pip._vendor.rich.style import Style + +from pip._internal.utils._log import VERBOSE, getLogger +from pip._internal.utils.compat import WINDOWS +from pip._internal.utils.deprecation import DEPRECATION_MSG_PREFIX +from pip._internal.utils.misc import ensure_dir _log_state = threading.local() -subprocess_logger = getLogger('fetchcode.vcs.pip.subprocessor') +subprocess_logger = getLogger("pip.subprocessor") class BrokenStdoutLoggingError(Exception): """ Raised if BrokenPipeError occurs for the stdout stream while logging. """ - pass -# BrokenPipeError does not exist in Python 2 and, in addition, manifests -# differently in Windows and non-Windows. -if WINDOWS: - # In Windows, a broken pipe can show up as EINVAL rather than EPIPE: +def _is_broken_pipe_error(exc_class: Type[BaseException], exc: BaseException) -> bool: + if exc_class is BrokenPipeError: + return True + + # On Windows, a broken pipe can show up as EINVAL rather than EPIPE: # https://bugs.python.org/issue19612 # https://bugs.python.org/issue30418 - if PY2: - def _is_broken_pipe_error(exc_class, exc): - """See the docstring for non-Windows Python 3 below.""" - return (exc_class is IOError and - exc.errno in (errno.EINVAL, errno.EPIPE)) - else: - # In Windows, a broken pipe IOError became OSError in Python 3. - def _is_broken_pipe_error(exc_class, exc): - """See the docstring for non-Windows Python 3 below.""" - return ((exc_class is BrokenPipeError) or # noqa: F821 - (exc_class is OSError and - exc.errno in (errno.EINVAL, errno.EPIPE))) -elif PY2: - def _is_broken_pipe_error(exc_class, exc): - """See the docstring for non-Windows Python 3 below.""" - return (exc_class is IOError and exc.errno == errno.EPIPE) -else: - # Then we are in the non-Windows Python 3 case. - def _is_broken_pipe_error(exc_class, exc): - """ - Return whether an exception is a broken pipe error. + if not WINDOWS: + return False - Args: - exc_class: an exception class. - exc: an exception instance. - """ - return (exc_class is BrokenPipeError) # noqa: F821 + return isinstance(exc, OSError) and exc.errno in (errno.EINVAL, errno.EPIPE) @contextlib.contextmanager -def indent_log(num=2): +def indent_log(num: int = 2) -> Generator[None, None, None]: """ A context manager which will cause the log output to be indented for any log messages emitted inside it. @@ -112,188 +66,183 @@ def indent_log(num=2): _log_state.indentation -= num -def get_indentation(): - return getattr(_log_state, 'indentation', 0) +def get_indentation() -> int: + return getattr(_log_state, "indentation", 0) class IndentingFormatter(logging.Formatter): - - def __init__(self, *args, **kwargs): + default_time_format = "%Y-%m-%dT%H:%M:%S" + + def __init__( + self, + *args: Any, + add_timestamp: bool = False, + **kwargs: Any, + ) -> None: """ A logging.Formatter that obeys the indent_log() context manager. :param add_timestamp: A bool indicating output lines should be prefixed with their record's timestamp. """ - self.add_timestamp = kwargs.pop("add_timestamp", False) - super(IndentingFormatter, self).__init__(*args, **kwargs) + self.add_timestamp = add_timestamp + super().__init__(*args, **kwargs) - def get_message_start(self, formatted, levelno): + def get_message_start(self, formatted: str, levelno: int) -> str: """ Return the start of the formatted log message (not counting the prefix to add to each line). """ if levelno < logging.WARNING: - return '' + return "" if formatted.startswith(DEPRECATION_MSG_PREFIX): # Then the message already has a prefix. We don't want it to # look like "WARNING: DEPRECATION: ...." - return '' + return "" if levelno < logging.ERROR: - return 'WARNING: ' + return "WARNING: " - return 'ERROR: ' + return "ERROR: " - def format(self, record): + def format(self, record: logging.LogRecord) -> str: """ Calls the standard formatter, but will indent all of the log message lines by our current indentation level. """ - formatted = super(IndentingFormatter, self).format(record) + formatted = super().format(record) message_start = self.get_message_start(formatted, record.levelno) formatted = message_start + formatted - prefix = '' + prefix = "" if self.add_timestamp: - # TODO: Use Formatter.default_time_format after dropping PY2. - t = self.formatTime(record, "%Y-%m-%dT%H:%M:%S") - prefix = '{t},{record.msecs:03.0f} '.format(**locals()) + prefix = f"{self.formatTime(record)} " prefix += " " * get_indentation() - formatted = "".join([ - prefix + line - for line in formatted.splitlines(True) - ]) + formatted = "".join([prefix + line for line in formatted.splitlines(True)]) return formatted -def _color_wrap(*colors): - def wrapped(inp): - return "".join(list(colors) + [inp, colorama.Style.RESET_ALL]) - return wrapped - - -class ColorizedStreamHandler(logging.StreamHandler): - - # Don't build up a list of colors if we don't have colorama - if colorama: - COLORS = [ - # This needs to be in order from highest logging level to lowest. - (logging.ERROR, _color_wrap(Fore.RED)), - (logging.WARNING, _color_wrap(Fore.YELLOW)), - ] - else: - COLORS = [] - - def __init__(self, stream=None, no_color=None): - logging.StreamHandler.__init__(self, stream) - self._no_color = no_color - - if WINDOWS and colorama: - self.stream = colorama.AnsiToWin32(self.stream) - - def _using_stdout(self): - """ - Return whether the handler is using sys.stdout. - """ - if WINDOWS and colorama: - # Then self.stream is an AnsiToWin32 object. - return self.stream.wrapped is sys.stdout - - return self.stream is sys.stdout - - def should_color(self): - # Don't colorize things if we do not have colorama or if told not to - if not colorama or self._no_color: - return False - - real_stream = ( - self.stream if not isinstance(self.stream, colorama.AnsiToWin32) - else self.stream.wrapped +@dataclass +class IndentedRenderable: + renderable: RenderableType + indent: int + + def __rich_console__( + self, console: Console, options: ConsoleOptions + ) -> RenderResult: + segments = console.render(self.renderable, options) + lines = Segment.split_lines(segments) + for line in lines: + yield Segment(" " * self.indent) + yield from line + yield Segment("\n") + + +class RichPipStreamHandler(RichHandler): + KEYWORDS: ClassVar[Optional[List[str]]] = [] + + def __init__(self, stream: Optional[TextIO], no_color: bool) -> None: + super().__init__( + console=Console(file=stream, no_color=no_color, soft_wrap=True), + show_time=False, + show_level=False, + show_path=False, + highlighter=NullHighlighter(), ) - # If the stream is a tty we should color it - if hasattr(real_stream, "isatty") and real_stream.isatty(): - return True - - # If we have an ANSI term we should color it - if os.environ.get("TERM") == "ANSI": - return True - - # If anything else we should not color it - return False - - def format(self, record): - msg = logging.StreamHandler.format(self, record) - - if self.should_color(): - for level, color in self.COLORS: - if record.levelno >= level: - msg = color(msg) - break - - return msg + # Our custom override on Rich's logger, to make things work as we need them to. + def emit(self, record: logging.LogRecord) -> None: + style: Optional[Style] = None + + # If we are given a diagnostic error to present, present it with indentation. + if getattr(record, "rich", False): + assert isinstance(record.args, tuple) + (rich_renderable,) = record.args + assert isinstance( + rich_renderable, (ConsoleRenderable, RichCast, str) + ), f"{rich_renderable} is not rich-console-renderable" + + renderable: RenderableType = IndentedRenderable( + rich_renderable, indent=get_indentation() + ) + else: + message = self.format(record) + renderable = self.render_message(record, message) + if record.levelno is not None: + if record.levelno >= logging.ERROR: + style = Style(color="red") + elif record.levelno >= logging.WARNING: + style = Style(color="yellow") + + try: + self.console.print(renderable, overflow="ignore", crop=False, style=style) + except Exception: + self.handleError(record) + + def handleError(self, record: logging.LogRecord) -> None: + """Called when logging is unable to log some output.""" - # The logging module says handleError() can be customized. - def handleError(self, record): exc_class, exc = sys.exc_info()[:2] # If a broken pipe occurred while calling write() or flush() on the # stdout stream in logging's Handler.emit(), then raise our special # exception so we can handle it in main() instead of logging the # broken pipe error and continuing. - if (exc_class and self._using_stdout() and - _is_broken_pipe_error(exc_class, exc)): + if ( + exc_class + and exc + and self.console.file is sys.stdout + and _is_broken_pipe_error(exc_class, exc) + ): raise BrokenStdoutLoggingError() - return super(ColorizedStreamHandler, self).handleError(record) + return super().handleError(record) class BetterRotatingFileHandler(logging.handlers.RotatingFileHandler): - - def _open(self): + def _open(self) -> TextIOWrapper: ensure_dir(os.path.dirname(self.baseFilename)) - return logging.handlers.RotatingFileHandler._open(self) + return super()._open() class MaxLevelFilter(Filter): - - def __init__(self, level): + def __init__(self, level: int) -> None: self.level = level - def filter(self, record): + def filter(self, record: logging.LogRecord) -> bool: return record.levelno < self.level class ExcludeLoggerFilter(Filter): - """ A logging Filter that excludes records from a logger (or its children). """ - def filter(self, record): + def filter(self, record: logging.LogRecord) -> bool: # The base Filter class allows only records from a logger (or its # children). - return not super(ExcludeLoggerFilter, self).filter(record) + return not super().filter(record) -def setup_logging(verbosity, no_color, user_log_file): +def setup_logging(verbosity: int, no_color: bool, user_log_file: Optional[str]) -> int: """Configures and sets up all of the logging Returns the requested logging level, as its integer value. """ # Determine the level to be logging at. - if verbosity >= 1: - level = "DEBUG" + if verbosity >= 2: + level_number = logging.DEBUG + elif verbosity == 1: + level_number = VERBOSE elif verbosity == -1: - level = "WARNING" + level_number = logging.WARNING elif verbosity == -2: - level = "ERROR" + level_number = logging.ERROR elif verbosity <= -3: - level = "CRITICAL" + level_number = logging.CRITICAL else: - level = "INFO" + level_number = logging.INFO - level_number = getattr(logging, level) + level = logging.getLevelName(level_number) # The "root" logger should match the "console" level *unless* we also need # to log to a user log file. @@ -315,85 +264,84 @@ def setup_logging(verbosity, no_color, user_log_file): "stderr": "ext://sys.stderr", } handler_classes = { - "stream": "pip._internal.utils.logging.ColorizedStreamHandler", + "stream": "pip._internal.utils.logging.RichPipStreamHandler", "file": "pip._internal.utils.logging.BetterRotatingFileHandler", } handlers = ["console", "console_errors", "console_subprocess"] + ( ["user_log"] if include_user_log else [] ) - logging.config.dictConfig({ - "version": 1, - "disable_existing_loggers": False, - "filters": { - "exclude_warnings": { - "()": "pip._internal.utils.logging.MaxLevelFilter", - "level": logging.WARNING, - }, - "restrict_to_subprocess": { - "()": "logging.Filter", - "name": subprocess_logger.name, - }, - "exclude_subprocess": { - "()": "pip._internal.utils.logging.ExcludeLoggerFilter", - "name": subprocess_logger.name, + logging.config.dictConfig( + { + "version": 1, + "disable_existing_loggers": False, + "filters": { + "exclude_warnings": { + "()": "pip._internal.utils.logging.MaxLevelFilter", + "level": logging.WARNING, + }, + "restrict_to_subprocess": { + "()": "logging.Filter", + "name": subprocess_logger.name, + }, + "exclude_subprocess": { + "()": "pip._internal.utils.logging.ExcludeLoggerFilter", + "name": subprocess_logger.name, + }, }, - }, - "formatters": { - "indent": { - "()": IndentingFormatter, - "format": "%(message)s", + "formatters": { + "indent": { + "()": IndentingFormatter, + "format": "%(message)s", + }, + "indent_with_timestamp": { + "()": IndentingFormatter, + "format": "%(message)s", + "add_timestamp": True, + }, }, - "indent_with_timestamp": { - "()": IndentingFormatter, - "format": "%(message)s", - "add_timestamp": True, + "handlers": { + "console": { + "level": level, + "class": handler_classes["stream"], + "no_color": no_color, + "stream": log_streams["stdout"], + "filters": ["exclude_subprocess", "exclude_warnings"], + "formatter": "indent", + }, + "console_errors": { + "level": "WARNING", + "class": handler_classes["stream"], + "no_color": no_color, + "stream": log_streams["stderr"], + "filters": ["exclude_subprocess"], + "formatter": "indent", + }, + # A handler responsible for logging to the console messages + # from the "subprocessor" logger. + "console_subprocess": { + "level": level, + "class": handler_classes["stream"], + "stream": log_streams["stderr"], + "no_color": no_color, + "filters": ["restrict_to_subprocess"], + "formatter": "indent", + }, + "user_log": { + "level": "DEBUG", + "class": handler_classes["file"], + "filename": additional_log_file, + "encoding": "utf-8", + "delay": True, + "formatter": "indent_with_timestamp", + }, }, - }, - "handlers": { - "console": { - "level": level, - "class": handler_classes["stream"], - "no_color": no_color, - "stream": log_streams["stdout"], - "filters": ["exclude_subprocess", "exclude_warnings"], - "formatter": "indent", + "root": { + "level": root_level, + "handlers": handlers, }, - "console_errors": { - "level": "WARNING", - "class": handler_classes["stream"], - "no_color": no_color, - "stream": log_streams["stderr"], - "filters": ["exclude_subprocess"], - "formatter": "indent", - }, - # A handler responsible for logging to the console messages - # from the "subprocessor" logger. - "console_subprocess": { - "level": level, - "class": handler_classes["stream"], - "no_color": no_color, - "stream": log_streams["stderr"], - "filters": ["restrict_to_subprocess"], - "formatter": "indent", - }, - "user_log": { - "level": "DEBUG", - "class": handler_classes["file"], - "filename": additional_log_file, - "delay": True, - "formatter": "indent_with_timestamp", - }, - }, - "root": { - "level": root_level, - "handlers": handlers, - }, - "loggers": { - "pip._vendor": { - "level": vendored_log_level - } - }, - }) + "loggers": {"pip._vendor": {"level": vendored_log_level}}, + } + ) return level_number diff --git a/src/fetchcode/vcs/pip/_internal/utils/misc.py b/src/fetchcode/vcs/pip/_internal/utils/misc.py index 142632e0..3707e872 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/misc.py +++ b/src/fetchcode/vcs/pip/_internal/utils/misc.py @@ -1,91 +1,85 @@ -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import - -import contextlib import errno import getpass import hashlib -import io import logging import os import posixpath import shutil import stat import sys -from collections import deque - -from fetchcode.vcs.pip._vendor import pkg_resources -# NOTE: retrying is not annotated in typeshed as on 2017-07-17, which is -# why we ignore the type on this import. -from fetchcode.vcs.pip._vendor.retrying import retry # type: ignore -from fetchcode.vcs.pip._vendor.six import PY2, text_type -from fetchcode.vcs.pip._vendor.six.moves import input, map, zip_longest -from fetchcode.vcs.pip._vendor.six.moves.urllib import parse as urllib_parse -from fetchcode.vcs.pip._vendor.six.moves.urllib.parse import unquote as urllib_unquote - -from pip import __version__ -from fetchcode.vcs.pip._internal.exceptions import CommandError -from fetchcode.vcs.pip._internal.locations import ( - get_major_minor_version, - site_packages, - user_site, -) -from fetchcode.vcs.pip._internal.utils.compat import ( - WINDOWS, - expanduser, - stdlib_pkgs, - str_to_display, +import sysconfig +import urllib.parse +from dataclasses import dataclass +from functools import partial +from io import StringIO +from itertools import filterfalse, tee, zip_longest +from pathlib import Path +from types import FunctionType, TracebackType +from typing import ( + Any, + BinaryIO, + Callable, + Dict, + Generator, + Iterable, + Iterator, + List, + Optional, + TextIO, + Tuple, + Type, + TypeVar, + Union, + cast, ) -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING, cast -from fetchcode.vcs.pip._internal.utils.virtualenv import ( - running_under_virtualenv, - virtualenv_no_global, -) - -if PY2: - from io import BytesIO as StringIO -else: - from io import StringIO - -if MYPY_CHECK_RUNNING: - from typing import ( - Any, AnyStr, Container, Iterable, Iterator, List, Optional, Text, - Tuple, Union, - ) - from fetchcode.vcs.pip._vendor.pkg_resources import Distribution - VersionInfo = Tuple[int, int, int] +from pip._vendor.packaging.requirements import Requirement +from pip._vendor.pyproject_hooks import BuildBackendHookCaller +from pip import __version__ +from pip._internal.exceptions import CommandError, ExternallyManagedEnvironment +from pip._internal.locations import get_major_minor_version +from pip._internal.utils.compat import WINDOWS +from pip._internal.utils.retry import retry +from pip._internal.utils.virtualenv import running_under_virtualenv + +__all__ = [ + "rmtree", + "display_path", + "backup_dir", + "ask", + "splitext", + "format_size", + "is_installable_dir", + "normalize_path", + "renames", + "get_prog", + "ensure_dir", + "remove_auth_from_url", + "check_externally_managed", + "ConfiguredBuildBackendHookCaller", +] -__all__ = ['rmtree', 'display_path', 'backup_dir', - 'ask', 'splitext', - 'format_size', 'is_installable_dir', - 'normalize_path', - 'renames', 'get_prog', - 'captured_stdout', 'ensure_dir', - 'get_installed_version', 'remove_auth_from_url'] +logger = logging.getLogger(__name__) +T = TypeVar("T") +ExcInfo = Tuple[Type[BaseException], BaseException, TracebackType] +VersionInfo = Tuple[int, int, int] +NetlocTuple = Tuple[str, Tuple[Optional[str], Optional[str]]] +OnExc = Callable[[FunctionType, Path, BaseException], Any] +OnErr = Callable[[FunctionType, Path, ExcInfo], Any] -logger = logging.getLogger(__name__) +FILE_CHUNK_SIZE = 1024 * 1024 -def get_pip_version(): - # type: () -> str +def get_pip_version() -> str: pip_pkg_dir = os.path.join(os.path.dirname(__file__), "..", "..") pip_pkg_dir = os.path.abspath(pip_pkg_dir) - return ( - 'pip {} from {} (python {})'.format( - __version__, pip_pkg_dir, get_major_minor_version(), - ) - ) + return f"pip {__version__} from {pip_pkg_dir} (python {get_major_minor_version()})" -def normalize_version_info(py_version_info): - # type: (Tuple[int, ...]) -> Tuple[int, int, int] +def normalize_version_info(py_version_info: Tuple[int, ...]) -> Tuple[int, int, int]: """ Convert a tuple of ints representing a Python version to one of length three. @@ -101,11 +95,10 @@ def normalize_version_info(py_version_info): elif len(py_version_info) > 3: py_version_info = py_version_info[:3] - return cast('VersionInfo', py_version_info) + return cast("VersionInfo", py_version_info) -def ensure_dir(path): - # type: (AnyStr) -> None +def ensure_dir(path: str) -> None: """os.path.makedirs without EEXIST.""" try: os.makedirs(path) @@ -115,96 +108,99 @@ def ensure_dir(path): raise -def get_prog(): - # type: () -> str +def get_prog() -> str: try: prog = os.path.basename(sys.argv[0]) - if prog in ('__main__.py', '-c'): - return "{} -m pip".format(sys.executable) + if prog in ("__main__.py", "-c"): + return f"{sys.executable} -m pip" else: return prog except (AttributeError, TypeError, IndexError): pass - return 'pip' + return "pip" # Retry every half second for up to 3 seconds -@retry(stop_max_delay=3000, wait_fixed=500) -def rmtree(dir, ignore_errors=False): - # type: (str, bool) -> None - shutil.rmtree(dir, ignore_errors=ignore_errors, - onerror=rmtree_errorhandler) +@retry(stop_after_delay=3, wait=0.5) +def rmtree( + dir: str, ignore_errors: bool = False, onexc: Optional[OnExc] = None +) -> None: + if ignore_errors: + onexc = _onerror_ignore + if onexc is None: + onexc = _onerror_reraise + handler: OnErr = partial( + # `[func, path, Union[ExcInfo, BaseException]] -> Any` is equivalent to + # `Union[([func, path, ExcInfo] -> Any), ([func, path, BaseException] -> Any)]`. + cast(Union[OnExc, OnErr], rmtree_errorhandler), + onexc=onexc, + ) + if sys.version_info >= (3, 12): + # See https://docs.python.org/3.12/whatsnew/3.12.html#shutil. + shutil.rmtree(dir, onexc=handler) # type: ignore + else: + shutil.rmtree(dir, onerror=handler) # type: ignore -def rmtree_errorhandler(func, path, exc_info): - """On Windows, the files in .svn are read-only, so when rmtree() tries to - remove them, an exception is thrown. We catch that here, remove the - read-only attribute, and hopefully continue without problems.""" - try: - has_attr_readonly = not (os.stat(path).st_mode & stat.S_IWRITE) - except (IOError, OSError): - # it's equivalent to os.path.exists - return +def _onerror_ignore(*_args: Any) -> None: + pass - if has_attr_readonly: - # convert to read/write - os.chmod(path, stat.S_IWRITE) - # use the original function to repeat the operation - func(path) - return - else: - raise + +def _onerror_reraise(*_args: Any) -> None: + raise # noqa: PLE0704 - Bare exception used to reraise existing exception -def path_to_display(path): - # type: (Optional[Union[str, Text]]) -> Optional[Text] +def rmtree_errorhandler( + func: FunctionType, + path: Path, + exc_info: Union[ExcInfo, BaseException], + *, + onexc: OnExc = _onerror_reraise, +) -> None: """ - Convert a bytes (or text) path to text (unicode in Python 2) for display - and logging purposes. + `rmtree` error handler to 'force' a file remove (i.e. like `rm -f`). - This function should never error out. Also, this function is mainly needed - for Python 2 since in Python 3 str paths are already text. + * If a file is readonly then it's write flag is set and operation is + retried. + + * `onerror` is the original callback from `rmtree(... onerror=onerror)` + that is chained at the end if the "rm -f" still fails. """ - if path is None: - return None - if isinstance(path, text_type): - return path - # Otherwise, path is a bytes object (str in Python 2). try: - display_path = path.decode(sys.getfilesystemencoding(), 'strict') - except UnicodeDecodeError: - # Include the full bytes to make troubleshooting easier, even though - # it may not be very human readable. - if PY2: - # Convert the bytes to a readable str representation using - # repr(), and then convert the str to unicode. - # Also, we add the prefix "b" to the repr() return value both - # to make the Python 2 output look like the Python 3 output, and - # to signal to the user that this is a bytes representation. - display_path = str_to_display('b{!r}'.format(path)) + st_mode = os.stat(path).st_mode + except OSError: + # it's equivalent to os.path.exists + return + + if not st_mode & stat.S_IWRITE: + # convert to read/write + try: + os.chmod(path, st_mode | stat.S_IWRITE) + except OSError: + pass else: - # Silence the "F821 undefined name 'ascii'" flake8 error since - # in Python 3 ascii() is a built-in. - display_path = ascii(path) # noqa: F821 + # use the original function to repeat the operation + try: + func(path) + return + except OSError: + pass - return display_path + if not isinstance(exc_info, BaseException): + _, exc_info, _ = exc_info + onexc(func, path, exc_info) -def display_path(path): - # type: (Union[str, Text]) -> str +def display_path(path: str) -> str: """Gives the display value for a given path, making it relative to cwd if possible.""" path = os.path.normcase(os.path.abspath(path)) - if sys.version_info[0] == 2: - path = path.decode(sys.getfilesystemencoding(), 'replace') - path = path.encode(sys.getdefaultencoding(), 'replace') if path.startswith(os.getcwd() + os.path.sep): - path = '.' + path[len(os.getcwd()):] + path = "." + path[len(os.getcwd()) :] return path -def backup_dir(dir, ext='.bak'): - # type: (str, str) -> str +def backup_dir(dir: str, ext: str = ".bak") -> str: """Figure out the name of a directory to back up the given dir to (adding .bak, .bak2, etc)""" n = 1 @@ -215,26 +211,22 @@ def backup_dir(dir, ext='.bak'): return dir + extension -def ask_path_exists(message, options): - # type: (str, Iterable[str]) -> str - for action in os.environ.get('PIP_EXISTS_ACTION', '').split(): +def ask_path_exists(message: str, options: Iterable[str]) -> str: + for action in os.environ.get("PIP_EXISTS_ACTION", "").split(): if action in options: return action return ask(message, options) -def _check_no_input(message): - # type: (str) -> None +def _check_no_input(message: str) -> None: """Raise an error if no input is allowed.""" - if os.environ.get('PIP_NO_INPUT'): + if os.environ.get("PIP_NO_INPUT"): raise Exception( - 'No input was expected ($PIP_NO_INPUT set); question: {}'.format( - message) + f"No input was expected ($PIP_NO_INPUT set); question: {message}" ) -def ask(message, options): - # type: (str, Iterable[str]) -> str +def ask(message: str, options: Iterable[str]) -> str: """Ask the message interactively, with the given possible responses""" while 1: _check_no_input(message) @@ -242,41 +234,53 @@ def ask(message, options): response = response.strip().lower() if response not in options: print( - 'Your response ({!r}) was not one of the expected responses: ' - '{}'.format(response, ', '.join(options)) + "Your response ({!r}) was not one of the expected responses: " + "{}".format(response, ", ".join(options)) ) else: return response -def ask_input(message): - # type: (str) -> str +def ask_input(message: str) -> str: """Ask for input interactively.""" _check_no_input(message) return input(message) -def ask_password(message): - # type: (str) -> str +def ask_password(message: str) -> str: """Ask for a password interactively.""" _check_no_input(message) return getpass.getpass(message) -def format_size(bytes): - # type: (float) -> str +def strtobool(val: str) -> int: + """Convert a string representation of truth to true (1) or false (0). + + True values are 'y', 'yes', 't', 'true', 'on', and '1'; false values + are 'n', 'no', 'f', 'false', 'off', and '0'. Raises ValueError if + 'val' is anything else. + """ + val = val.lower() + if val in ("y", "yes", "t", "true", "on", "1"): + return 1 + elif val in ("n", "no", "f", "false", "off", "0"): + return 0 + else: + raise ValueError(f"invalid truth value {val!r}") + + +def format_size(bytes: float) -> str: if bytes > 1000 * 1000: - return '{:.1f} MB'.format(bytes / 1000.0 / 1000) + return f"{bytes / 1000.0 / 1000:.1f} MB" elif bytes > 10 * 1000: - return '{} kB'.format(int(bytes / 1000)) + return f"{int(bytes / 1000)} kB" elif bytes > 1000: - return '{:.1f} kB'.format(bytes / 1000.0) + return f"{bytes / 1000.0:.1f} kB" else: - return '{} bytes'.format(int(bytes)) + return f"{int(bytes)} bytes" -def tabulate(rows): - # type: (Iterable[Iterable[Any]]) -> Tuple[List[str], List[int]] +def tabulate(rows: Iterable[Iterable[Any]]) -> Tuple[List[str], List[int]]: """Return a list of formatted rows and a list of column sizes. For example:: @@ -285,27 +289,31 @@ def tabulate(rows): (['foobar 2000', '3735928559'], [10, 4]) """ rows = [tuple(map(str, row)) for row in rows] - sizes = [max(map(len, col)) for col in zip_longest(*rows, fillvalue='')] + sizes = [max(map(len, col)) for col in zip_longest(*rows, fillvalue="")] table = [" ".join(map(str.ljust, row, sizes)).rstrip() for row in rows] return table, sizes -def is_installable_dir(path): - # type: (str) -> bool - """Is path is a directory containing setup.py or pyproject.toml? +def is_installable_dir(path: str) -> bool: + """Is path is a directory containing pyproject.toml or setup.py? + + If pyproject.toml exists, this is a PEP 517 project. Otherwise we look for + a legacy setuptools layout by identifying setup.py. We don't check for the + setup.cfg because using it without setup.py is only available for PEP 517 + projects, which are already covered by the pyproject.toml check. """ if not os.path.isdir(path): return False - setup_py = os.path.join(path, 'setup.py') - if os.path.isfile(setup_py): + if os.path.isfile(os.path.join(path, "pyproject.toml")): return True - pyproject_toml = os.path.join(path, 'pyproject.toml') - if os.path.isfile(pyproject_toml): + if os.path.isfile(os.path.join(path, "setup.py")): return True return False -def read_chunks(file, size=io.DEFAULT_BUFFER_SIZE): +def read_chunks( + file: BinaryIO, size: int = FILE_CHUNK_SIZE +) -> Generator[bytes, None, None]: """Yield pieces of data from a file-like object until EOF.""" while True: chunk = file.read(size) @@ -314,13 +322,12 @@ def read_chunks(file, size=io.DEFAULT_BUFFER_SIZE): yield chunk -def normalize_path(path, resolve_symlinks=True): - # type: (str, bool) -> str +def normalize_path(path: str, resolve_symlinks: bool = True) -> str: """ Convert a path to its canonical, case-normalized, absolute version. """ - path = expanduser(path) + path = os.path.expanduser(path) if resolve_symlinks: path = os.path.realpath(path) else: @@ -328,18 +335,16 @@ def normalize_path(path, resolve_symlinks=True): return os.path.normcase(path) -def splitext(path): - # type: (str) -> Tuple[str, str] +def splitext(path: str) -> Tuple[str, str]: """Like os.path.splitext, but take off .tar too""" base, ext = posixpath.splitext(path) - if base.lower().endswith('.tar'): + if base.lower().endswith(".tar"): ext = base[-4:] + ext base = base[:-4] return base, ext -def renames(old, new): - # type: (str, str) -> None +def renames(old: str, new: str) -> None: """Like os.renames(), but handles renaming across devices.""" # Implementation borrowed from os.renames(). head, tail = os.path.split(new) @@ -356,8 +361,7 @@ def renames(old, new): pass -def is_local(path): - # type: (str) -> bool +def is_local(path: str) -> bool: """ Return True if path is within sys.prefix, if we're running in a virtualenv. @@ -371,353 +375,95 @@ def is_local(path): return path.startswith(normalize_path(sys.prefix)) -def dist_is_local(dist): - # type: (Distribution) -> bool - """ - Return True if given Distribution object is installed locally - (i.e. within current virtualenv). - - Always True if we're not in a virtualenv. - - """ - return is_local(dist_location(dist)) - - -def dist_in_usersite(dist): - # type: (Distribution) -> bool - """ - Return True if given Distribution is installed in user site. - """ - return dist_location(dist).startswith(normalize_path(user_site)) - - -def dist_in_site_packages(dist): - # type: (Distribution) -> bool - """ - Return True if given Distribution is installed in - sysconfig.get_python_lib(). - """ - return dist_location(dist).startswith(normalize_path(site_packages)) - - -def dist_is_editable(dist): - # type: (Distribution) -> bool - """ - Return True if given Distribution is an editable install. - """ - for path_item in sys.path: - egg_link = os.path.join(path_item, dist.project_name + '.egg-link') - if os.path.isfile(egg_link): - return True - return False - - -def get_installed_distributions( - local_only=True, # type: bool - skip=stdlib_pkgs, # type: Container[str] - include_editables=True, # type: bool - editables_only=False, # type: bool - user_only=False, # type: bool - paths=None # type: Optional[List[str]] -): - # type: (...) -> List[Distribution] - """ - Return a list of installed Distribution objects. - - If ``local_only`` is True (default), only return installations - local to the current virtualenv, if in a virtualenv. - - ``skip`` argument is an iterable of lower-case project names to - ignore; defaults to stdlib_pkgs - - If ``include_editables`` is False, don't report editables. - - If ``editables_only`` is True , only report editables. - - If ``user_only`` is True , only report installations in the user - site directory. - - If ``paths`` is set, only report the distributions present at the - specified list of locations. - """ - if paths: - working_set = pkg_resources.WorkingSet(paths) - else: - working_set = pkg_resources.working_set - - if local_only: - local_test = dist_is_local - else: - def local_test(d): - return True - - if include_editables: - def editable_test(d): - return True - else: - def editable_test(d): - return not dist_is_editable(d) - - if editables_only: - def editables_only_test(d): - return dist_is_editable(d) - else: - def editables_only_test(d): - return True - - if user_only: - user_test = dist_in_usersite - else: - def user_test(d): - return True - - return [d for d in working_set - if local_test(d) and - d.key not in skip and - editable_test(d) and - editables_only_test(d) and - user_test(d) - ] - - -def egg_link_path(dist): - # type: (Distribution) -> Optional[str] - """ - Return the path for the .egg-link file if it exists, otherwise, None. - - There's 3 scenarios: - 1) not in a virtualenv - try to find in site.USER_SITE, then site_packages - 2) in a no-global virtualenv - try to find in site_packages - 3) in a yes-global virtualenv - try to find in site_packages, then site.USER_SITE - (don't look in global location) - - For #1 and #3, there could be odd cases, where there's an egg-link in 2 - locations. - - This method will just return the first one found. - """ - sites = [] - if running_under_virtualenv(): - sites.append(site_packages) - if not virtualenv_no_global() and user_site: - sites.append(user_site) - else: - if user_site: - sites.append(user_site) - sites.append(site_packages) - - for site in sites: - egglink = os.path.join(site, dist.project_name) + '.egg-link' - if os.path.isfile(egglink): - return egglink - return None - - -def dist_location(dist): - # type: (Distribution) -> str - """ - Get the site-packages location of this distribution. Generally - this is dist.location, except in the case of develop-installed - packages, where dist.location is the source code location, and we - want to know where the egg-link file is. - - The returned location is normalized (in particular, with symlinks removed). - """ - egg_link = egg_link_path(dist) - if egg_link: - return normalize_path(egg_link) - return normalize_path(dist.location) - - -def write_output(msg, *args): - # type: (str, str) -> None +def write_output(msg: Any, *args: Any) -> None: logger.info(msg, *args) -class FakeFile(object): - """Wrap a list of lines in an object with readline() to make - ConfigParser happy.""" - def __init__(self, lines): - self._gen = (l for l in lines) - - def readline(self): - try: - try: - return next(self._gen) - except NameError: - return self._gen.next() - except StopIteration: - return '' - - def __iter__(self): - return self._gen - - class StreamWrapper(StringIO): + orig_stream: TextIO @classmethod - def from_stream(cls, orig_stream): - cls.orig_stream = orig_stream - return cls() + def from_stream(cls, orig_stream: TextIO) -> "StreamWrapper": + ret = cls() + ret.orig_stream = orig_stream + return ret # compileall.compile_dir() needs stdout.encoding to print to stdout + # type ignore is because TextIOBase.encoding is writeable @property - def encoding(self): + def encoding(self) -> str: # type: ignore return self.orig_stream.encoding -@contextlib.contextmanager -def captured_output(stream_name): - """Return a context manager used by captured_stdout/stdin/stderr - that temporarily replaces the sys stream *stream_name* with a StringIO. - - Taken from Lib/support/__init__.py in the CPython repo. - """ - orig_stdout = getattr(sys, stream_name) - setattr(sys, stream_name, StreamWrapper.from_stream(orig_stdout)) - try: - yield getattr(sys, stream_name) - finally: - setattr(sys, stream_name, orig_stdout) - - -def captured_stdout(): - """Capture the output of sys.stdout: - - with captured_stdout() as stdout: - print('hello') - self.assertEqual(stdout.getvalue(), 'hello\n') - - Taken from Lib/support/__init__.py in the CPython repo. - """ - return captured_output('stdout') - - -def captured_stderr(): - """ - See captured_stdout(). - """ - return captured_output('stderr') - - -class cached_property(object): - """A property that is only computed once per instance and then replaces - itself with an ordinary attribute. Deleting the attribute resets the - property. - - Source: https://github.com/bottlepy/bottle/blob/0.11.5/bottle.py#L175 - """ - - def __init__(self, func): - self.__doc__ = getattr(func, '__doc__') - self.func = func - - def __get__(self, obj, cls): - if obj is None: - # We're being accessed from the class itself, not from an object - return self - value = obj.__dict__[self.func.__name__] = self.func(obj) - return value - - -def get_installed_version(dist_name, working_set=None): - """Get the installed version of dist_name avoiding pkg_resources cache""" - # Create a requirement that we'll look for inside of setuptools. - req = pkg_resources.Requirement.parse(dist_name) - - if working_set is None: - # We want to avoid having this cached, so we need to construct a new - # working set each time. - working_set = pkg_resources.WorkingSet() - - # Get the installed distribution from our working set - dist = working_set.find(req) - - # Check to see if we got an installed distribution or not, if we did - # we want to return it's version. - return dist.version if dist else None - - -def consume(iterator): - """Consume an iterable at C speed.""" - deque(iterator, maxlen=0) - - # Simulates an enum -def enum(*sequential, **named): +def enum(*sequential: Any, **named: Any) -> Type[Any]: enums = dict(zip(sequential, range(len(sequential))), **named) reverse = {value: key for key, value in enums.items()} - enums['reverse_mapping'] = reverse - return type('Enum', (), enums) + enums["reverse_mapping"] = reverse + return type("Enum", (), enums) -def build_netloc(host, port): - # type: (str, Optional[int]) -> str +def build_netloc(host: str, port: Optional[int]) -> str: """ Build a netloc from a host-port pair """ if port is None: return host - if ':' in host: + if ":" in host: # Only wrap host with square brackets when it is IPv6 - host = '[{}]'.format(host) - return '{}:{}'.format(host, port) + host = f"[{host}]" + return f"{host}:{port}" -def build_url_from_netloc(netloc, scheme='https'): - # type: (str, str) -> str +def build_url_from_netloc(netloc: str, scheme: str = "https") -> str: """ Build a full URL from a netloc. """ - if netloc.count(':') >= 2 and '@' not in netloc and '[' not in netloc: + if netloc.count(":") >= 2 and "@" not in netloc and "[" not in netloc: # It must be a bare IPv6 address, so wrap it with brackets. - netloc = '[{}]'.format(netloc) - return '{}://{}'.format(scheme, netloc) + netloc = f"[{netloc}]" + return f"{scheme}://{netloc}" -def parse_netloc(netloc): - # type: (str) -> Tuple[str, Optional[int]] +def parse_netloc(netloc: str) -> Tuple[Optional[str], Optional[int]]: """ Return the host-port pair from a netloc. """ url = build_url_from_netloc(netloc) - parsed = urllib_parse.urlparse(url) + parsed = urllib.parse.urlparse(url) return parsed.hostname, parsed.port -def split_auth_from_netloc(netloc): +def split_auth_from_netloc(netloc: str) -> NetlocTuple: """ Parse out and remove the auth information from a netloc. Returns: (netloc, (username, password)). """ - if '@' not in netloc: + if "@" not in netloc: return netloc, (None, None) # Split from the right because that's how urllib.parse.urlsplit() # behaves if more than one @ is present (which can be checked using # the password attribute of urlsplit()'s return value). - auth, netloc = netloc.rsplit('@', 1) - if ':' in auth: + auth, netloc = netloc.rsplit("@", 1) + pw: Optional[str] = None + if ":" in auth: # Split from the left because that's how urllib.parse.urlsplit() # behaves if more than one : is present (which again can be checked # using the password attribute of the return value) - user_pass = auth.split(':', 1) + user, pw = auth.split(":", 1) else: - user_pass = auth, None + user, pw = auth, None - user_pass = tuple( - None if x is None else urllib_unquote(x) for x in user_pass - ) + user = urllib.parse.unquote(user) + if pw is not None: + pw = urllib.parse.unquote(pw) - return netloc, user_pass + return netloc, (user, pw) -def redact_netloc(netloc): - # type: (str) -> str +def redact_netloc(netloc: str) -> str: """ Replace the sensitive data in a netloc with "****", if it exists. @@ -729,17 +475,17 @@ def redact_netloc(netloc): if user is None: return netloc if password is None: - user = '****' - password = '' + user = "****" + password = "" else: - user = urllib_parse.quote(user) - password = ':****' - return '{user}{password}@{netloc}'.format(user=user, - password=password, - netloc=netloc) + user = urllib.parse.quote(user) + password = ":****" + return f"{user}{password}@{netloc}" -def _transform_url(url, transform_netloc): +def _transform_url( + url: str, transform_netloc: Callable[[str], Tuple[Any, ...]] +) -> Tuple[str, NetlocTuple]: """Transform and replace netloc in a url. transform_netloc is a function taking the netloc and returning a @@ -749,26 +495,25 @@ def _transform_url(url, transform_netloc): Returns a tuple containing the transformed url as item 0 and the original tuple returned by transform_netloc as item 1. """ - purl = urllib_parse.urlsplit(url) + purl = urllib.parse.urlsplit(url) netloc_tuple = transform_netloc(purl.netloc) # stripped url - url_pieces = ( - purl.scheme, netloc_tuple[0], purl.path, purl.query, purl.fragment - ) - surl = urllib_parse.urlunsplit(url_pieces) - return surl, netloc_tuple + url_pieces = (purl.scheme, netloc_tuple[0], purl.path, purl.query, purl.fragment) + surl = urllib.parse.urlunsplit(url_pieces) + return surl, cast("NetlocTuple", netloc_tuple) -def _get_netloc(netloc): +def _get_netloc(netloc: str) -> NetlocTuple: return split_auth_from_netloc(netloc) -def _redact_netloc(netloc): +def _redact_netloc(netloc: str) -> Tuple[str]: return (redact_netloc(netloc),) -def split_auth_netloc_from_url(url): - # type: (str) -> Tuple[str, str, Tuple[str, str]] +def split_auth_netloc_from_url( + url: str, +) -> Tuple[str, str, Tuple[Optional[str], Optional[str]]]: """ Parse a url into separate netloc, auth, and url with no auth. @@ -778,131 +523,114 @@ def split_auth_netloc_from_url(url): return url_without_auth, netloc, auth -def remove_auth_from_url(url): - # type: (str) -> str +def remove_auth_from_url(url: str) -> str: """Return a copy of url with 'username:password@' removed.""" # username/pass params are passed to subversion through flags # and are not recognized in the url. return _transform_url(url, _get_netloc)[0] -def redact_auth_from_url(url): - # type: (str) -> str +def redact_auth_from_url(url: str) -> str: """Replace the password in a given url with ****.""" return _transform_url(url, _redact_netloc)[0] -class HiddenText(object): - def __init__( - self, - secret, # type: str - redacted, # type: str - ): - # type: (...) -> None - self.secret = secret - self.redacted = redacted +def redact_auth_from_requirement(req: Requirement) -> str: + """Replace the password in a given requirement url with ****.""" + if not req.url: + return str(req) + return str(req).replace(req.url, redact_auth_from_url(req.url)) + - def __repr__(self): - # type: (...) -> str - return ''.format(str(self)) +@dataclass(frozen=True) +class HiddenText: + secret: str + redacted: str - def __str__(self): - # type: (...) -> str + def __repr__(self) -> str: + return f"" + + def __str__(self) -> str: return self.redacted # This is useful for testing. - def __eq__(self, other): - # type: (Any) -> bool + def __eq__(self, other: Any) -> bool: if type(self) != type(other): return False # The string being used for redaction doesn't also have to match, # just the raw, original string. - return (self.secret == other.secret) - - # We need to provide an explicit __ne__ implementation for Python 2. - # TODO: remove this when we drop PY2 support. - def __ne__(self, other): - # type: (Any) -> bool - return not self == other + return self.secret == other.secret -def hide_value(value): - # type: (str) -> HiddenText - return HiddenText(value, redacted='****') +def hide_value(value: str) -> HiddenText: + return HiddenText(value, redacted="****") -def hide_url(url): - # type: (str) -> HiddenText +def hide_url(url: str) -> HiddenText: redacted = redact_auth_from_url(url) return HiddenText(url, redacted=redacted) -def protect_pip_from_modification_on_windows(modifying_pip): - # type: (bool) -> None +def protect_pip_from_modification_on_windows(modifying_pip: bool) -> None: """Protection of pip.exe from modification on Windows On Windows, any operation modifying pip should be run as: python -m pip ... """ pip_names = [ - "pip.exe", - "pip{}.exe".format(sys.version_info[0]), - "pip{}.{}.exe".format(*sys.version_info[:2]) + "pip", + f"pip{sys.version_info.major}", + f"pip{sys.version_info.major}.{sys.version_info.minor}", ] # See https://github.com/pypa/pip/issues/1299 for more discussion should_show_use_python_msg = ( - modifying_pip and - WINDOWS and - os.path.basename(sys.argv[0]) in pip_names + modifying_pip and WINDOWS and os.path.basename(sys.argv[0]) in pip_names ) if should_show_use_python_msg: - new_command = [ - sys.executable, "-m", "pip" - ] + sys.argv[1:] + new_command = [sys.executable, "-m", "pip"] + sys.argv[1:] raise CommandError( - 'To modify pip, please run the following command:\n{}' - .format(" ".join(new_command)) + "To modify pip, please run the following command:\n{}".format( + " ".join(new_command) + ) ) -def is_console_interactive(): - # type: () -> bool - """Is this console interactive? +def check_externally_managed() -> None: + """Check whether the current environment is externally managed. + + If the ``EXTERNALLY-MANAGED`` config file is found, the current environment + is considered externally managed, and an ExternallyManagedEnvironment is + raised. """ + if running_under_virtualenv(): + return + marker = os.path.join(sysconfig.get_path("stdlib"), "EXTERNALLY-MANAGED") + if not os.path.isfile(marker): + return + raise ExternallyManagedEnvironment.from_config(marker) + + +def is_console_interactive() -> bool: + """Is this console interactive?""" return sys.stdin is not None and sys.stdin.isatty() -def hash_file(path, blocksize=1 << 20): - # type: (str, int) -> Tuple[Any, int] - """Return (hash, length) for path using hashlib.sha256() - """ +def hash_file(path: str, blocksize: int = 1 << 20) -> Tuple[Any, int]: + """Return (hash, length) for path using hashlib.sha256()""" h = hashlib.sha256() length = 0 - with open(path, 'rb') as f: + with open(path, "rb") as f: for block in read_chunks(f, size=blocksize): length += len(block) h.update(block) return h, length -def is_wheel_installed(): - """ - Return whether the wheel package is installed. - """ - try: - import wheel # noqa: F401 - except ImportError: - return False - - return True - - -def pairwise(iterable): - # type: (Iterable[Any]) -> Iterator[Tuple[Any, Any]] +def pairwise(iterable: Iterable[Any]) -> Iterator[Tuple[Any, Any]]: """ Return paired elements. @@ -911,3 +639,139 @@ def pairwise(iterable): """ iterable = iter(iterable) return zip_longest(iterable, iterable) + + +def partition( + pred: Callable[[T], bool], iterable: Iterable[T] +) -> Tuple[Iterable[T], Iterable[T]]: + """ + Use a predicate to partition entries into false entries and true entries, + like + + partition(is_odd, range(10)) --> 0 2 4 6 8 and 1 3 5 7 9 + """ + t1, t2 = tee(iterable) + return filterfalse(pred, t1), filter(pred, t2) + + +class ConfiguredBuildBackendHookCaller(BuildBackendHookCaller): + def __init__( + self, + config_holder: Any, + source_dir: str, + build_backend: str, + backend_path: Optional[str] = None, + runner: Optional[Callable[..., None]] = None, + python_executable: Optional[str] = None, + ): + super().__init__( + source_dir, build_backend, backend_path, runner, python_executable + ) + self.config_holder = config_holder + + def build_wheel( + self, + wheel_directory: str, + config_settings: Optional[Dict[str, Union[str, List[str]]]] = None, + metadata_directory: Optional[str] = None, + ) -> str: + cs = self.config_holder.config_settings + return super().build_wheel( + wheel_directory, config_settings=cs, metadata_directory=metadata_directory + ) + + def build_sdist( + self, + sdist_directory: str, + config_settings: Optional[Dict[str, Union[str, List[str]]]] = None, + ) -> str: + cs = self.config_holder.config_settings + return super().build_sdist(sdist_directory, config_settings=cs) + + def build_editable( + self, + wheel_directory: str, + config_settings: Optional[Dict[str, Union[str, List[str]]]] = None, + metadata_directory: Optional[str] = None, + ) -> str: + cs = self.config_holder.config_settings + return super().build_editable( + wheel_directory, config_settings=cs, metadata_directory=metadata_directory + ) + + def get_requires_for_build_wheel( + self, config_settings: Optional[Dict[str, Union[str, List[str]]]] = None + ) -> List[str]: + cs = self.config_holder.config_settings + return super().get_requires_for_build_wheel(config_settings=cs) + + def get_requires_for_build_sdist( + self, config_settings: Optional[Dict[str, Union[str, List[str]]]] = None + ) -> List[str]: + cs = self.config_holder.config_settings + return super().get_requires_for_build_sdist(config_settings=cs) + + def get_requires_for_build_editable( + self, config_settings: Optional[Dict[str, Union[str, List[str]]]] = None + ) -> List[str]: + cs = self.config_holder.config_settings + return super().get_requires_for_build_editable(config_settings=cs) + + def prepare_metadata_for_build_wheel( + self, + metadata_directory: str, + config_settings: Optional[Dict[str, Union[str, List[str]]]] = None, + _allow_fallback: bool = True, + ) -> str: + cs = self.config_holder.config_settings + return super().prepare_metadata_for_build_wheel( + metadata_directory=metadata_directory, + config_settings=cs, + _allow_fallback=_allow_fallback, + ) + + def prepare_metadata_for_build_editable( + self, + metadata_directory: str, + config_settings: Optional[Dict[str, Union[str, List[str]]]] = None, + _allow_fallback: bool = True, + ) -> str: + cs = self.config_holder.config_settings + return super().prepare_metadata_for_build_editable( + metadata_directory=metadata_directory, + config_settings=cs, + _allow_fallback=_allow_fallback, + ) + + +def warn_if_run_as_root() -> None: + """Output a warning for sudo users on Unix. + + In a virtual environment, sudo pip still writes to virtualenv. + On Windows, users may run pip as Administrator without issues. + This warning only applies to Unix root users outside of virtualenv. + """ + if running_under_virtualenv(): + return + if not hasattr(os, "getuid"): + return + # On Windows, there are no "system managed" Python packages. Installing as + # Administrator via pip is the correct way of updating system environments. + # + # We choose sys.platform over utils.compat.WINDOWS here to enable Mypy platform + # checks: https://mypy.readthedocs.io/en/stable/common_issues.html + if sys.platform == "win32" or sys.platform == "cygwin": + return + + if os.getuid() != 0: + return + + logger.warning( + "Running pip as the 'root' user can result in broken permissions and " + "conflicting behaviour with the system package manager, possibly " + "rendering your system unusable." + "It is recommended to use a virtual environment instead: " + "https://pip.pypa.io/warnings/venv. " + "Use the --root-user-action option if you know what you are doing and " + "want to suppress this warning." + ) diff --git a/src/fetchcode/vcs/pip/_internal/utils/models.py b/src/fetchcode/vcs/pip/_internal/utils/models.py deleted file mode 100644 index 29e14411..00000000 --- a/src/fetchcode/vcs/pip/_internal/utils/models.py +++ /dev/null @@ -1,42 +0,0 @@ -"""Utilities for defining models -""" -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - -import operator - - -class KeyBasedCompareMixin(object): - """Provides comparison capabilities that is based on a key - """ - - def __init__(self, key, defining_class): - self._compare_key = key - self._defining_class = defining_class - - def __hash__(self): - return hash(self._compare_key) - - def __lt__(self, other): - return self._compare(other, operator.__lt__) - - def __le__(self, other): - return self._compare(other, operator.__le__) - - def __gt__(self, other): - return self._compare(other, operator.__gt__) - - def __ge__(self, other): - return self._compare(other, operator.__ge__) - - def __eq__(self, other): - return self._compare(other, operator.__eq__) - - def __ne__(self, other): - return self._compare(other, operator.__ne__) - - def _compare(self, other, method): - if not isinstance(other, self._defining_class): - return NotImplemented - - return method(self._compare_key, other._compare_key) diff --git a/src/fetchcode/vcs/pip/_internal/utils/packaging.py b/src/fetchcode/vcs/pip/_internal/utils/packaging.py index 41681f44..4b8fa0fe 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/packaging.py +++ b/src/fetchcode/vcs/pip/_internal/utils/packaging.py @@ -1,26 +1,19 @@ -from __future__ import absolute_import - +import functools import logging -from email.parser import FeedParser - -from fetchcode.vcs.pip._vendor import pkg_resources -from fetchcode.vcs.pip._vendor.packaging import specifiers, version - -from fetchcode.vcs.pip._internal.exceptions import NoneMetadataError -from fetchcode.vcs.pip._internal.utils.misc import display_path -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +import re +from typing import NewType, Optional, Tuple, cast -if MYPY_CHECK_RUNNING: - from typing import Optional, Tuple - from email.message import Message - from fetchcode.vcs.pip._vendor.pkg_resources import Distribution +from pip._vendor.packaging import specifiers, version +from pip._vendor.packaging.requirements import Requirement +NormalizedExtra = NewType("NormalizedExtra", str) logger = logging.getLogger(__name__) -def check_requires_python(requires_python, version_info): - # type: (Optional[str], Tuple[int, ...]) -> bool +def check_requires_python( + requires_python: Optional[str], version_info: Tuple[int, ...] +) -> bool: """ Check if the given Python version matches a "Requires-Python" specifier. @@ -37,58 +30,28 @@ def check_requires_python(requires_python, version_info): return True requires_python_specifier = specifiers.SpecifierSet(requires_python) - python_version = version.parse('.'.join(map(str, version_info))) + python_version = version.parse(".".join(map(str, version_info))) return python_version in requires_python_specifier -def get_metadata(dist): - # type: (Distribution) -> Message - """ - :raises NoneMetadataError: if the distribution reports `has_metadata()` - True but `get_metadata()` returns None. - """ - metadata_name = 'METADATA' - if (isinstance(dist, pkg_resources.DistInfoDistribution) and - dist.has_metadata(metadata_name)): - metadata = dist.get_metadata(metadata_name) - elif dist.has_metadata('PKG-INFO'): - metadata_name = 'PKG-INFO' - metadata = dist.get_metadata(metadata_name) - else: - logger.warning("No metadata found in %s", display_path(dist.location)) - metadata = '' +@functools.lru_cache(maxsize=2048) +def get_requirement(req_string: str) -> Requirement: + """Construct a packaging.Requirement object with caching""" + # Parsing requirement strings is expensive, and is also expected to happen + # with a low diversity of different arguments (at least relative the number + # constructed). This method adds a cache to requirement object creation to + # minimize repeated parsing of the same string to construct equivalent + # Requirement objects. + return Requirement(req_string) - if metadata is None: - raise NoneMetadataError(dist, metadata_name) - feed_parser = FeedParser() - # The following line errors out if with a "NoneType" TypeError if - # passed metadata=None. - feed_parser.feed(metadata) - return feed_parser.close() +def safe_extra(extra: str) -> NormalizedExtra: + """Convert an arbitrary string to a standard 'extra' name + Any runs of non-alphanumeric characters are replaced with a single '_', + and the result is always lowercased. -def get_requires_python(dist): - # type: (pkg_resources.Distribution) -> Optional[str] + This function is duplicated from ``pkg_resources``. Note that this is not + the same to either ``canonicalize_name`` or ``_egg_link_name``. """ - Return the "Requires-Python" metadata for a distribution, or None - if not present. - """ - pkg_info_dict = get_metadata(dist) - requires_python = pkg_info_dict.get('Requires-Python') - - if requires_python is not None: - # Convert to a str to satisfy the type checker, since requires_python - # can be a Header object. - requires_python = str(requires_python) - - return requires_python - - -def get_installer(dist): - # type: (Distribution) -> str - if dist.has_metadata('INSTALLER'): - for line in dist.get_metadata_lines('INSTALLER'): - if line.strip(): - return line.strip() - return '' + return cast(NormalizedExtra, re.sub("[^A-Za-z0-9.-]+", "_", extra).lower()) diff --git a/src/fetchcode/vcs/pip/_internal/utils/pkg_resources.py b/src/fetchcode/vcs/pip/_internal/utils/pkg_resources.py deleted file mode 100644 index f72fb778..00000000 --- a/src/fetchcode/vcs/pip/_internal/utils/pkg_resources.py +++ /dev/null @@ -1,44 +0,0 @@ -from fetchcode.vcs.pip._vendor.pkg_resources import yield_lines -from fetchcode.vcs.pip._vendor.six import ensure_str - -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import Dict, Iterable, List - - -class DictMetadata(object): - """IMetadataProvider that reads metadata files from a dictionary. - """ - def __init__(self, metadata): - # type: (Dict[str, bytes]) -> None - self._metadata = metadata - - def has_metadata(self, name): - # type: (str) -> bool - return name in self._metadata - - def get_metadata(self, name): - # type: (str) -> str - try: - return ensure_str(self._metadata[name]) - except UnicodeDecodeError as e: - # Mirrors handling done in pkg_resources.NullProvider. - e.reason += " in {} file".format(name) - raise - - def get_metadata_lines(self, name): - # type: (str) -> Iterable[str] - return yield_lines(self.get_metadata(name)) - - def metadata_isdir(self, name): - # type: (str) -> bool - return False - - def metadata_listdir(self, name): - # type: (str) -> List[str] - return [] - - def run_script(self, script_name, namespace): - # type: (str, str) -> None - pass diff --git a/src/fetchcode/vcs/pip/_internal/utils/retry.py b/src/fetchcode/vcs/pip/_internal/utils/retry.py new file mode 100644 index 00000000..abfe0728 --- /dev/null +++ b/src/fetchcode/vcs/pip/_internal/utils/retry.py @@ -0,0 +1,42 @@ +import functools +from time import perf_counter, sleep +from typing import Callable, TypeVar + +from pip._vendor.typing_extensions import ParamSpec + +T = TypeVar("T") +P = ParamSpec("P") + + +def retry( + wait: float, stop_after_delay: float +) -> Callable[[Callable[P, T]], Callable[P, T]]: + """Decorator to automatically retry a function on error. + + If the function raises, the function is recalled with the same arguments + until it returns or the time limit is reached. When the time limit is + surpassed, the last exception raised is reraised. + + :param wait: The time to wait after an error before retrying, in seconds. + :param stop_after_delay: The time limit after which retries will cease, + in seconds. + """ + + def wrapper(func: Callable[P, T]) -> Callable[P, T]: + + @functools.wraps(func) + def retry_wrapped(*args: P.args, **kwargs: P.kwargs) -> T: + # The performance counter is monotonic on all platforms we care + # about and has much better resolution than time.monotonic(). + start_time = perf_counter() + while True: + try: + return func(*args, **kwargs) + except Exception: + if perf_counter() - start_time > stop_after_delay: + raise + sleep(wait) + + return retry_wrapped + + return wrapper diff --git a/src/fetchcode/vcs/pip/_internal/utils/setuptools_build.py b/src/fetchcode/vcs/pip/_internal/utils/setuptools_build.py index 8338e3bc..96d1b246 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/setuptools_build.py +++ b/src/fetchcode/vcs/pip/_internal/utils/setuptools_build.py @@ -1,32 +1,57 @@ import sys - -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import List, Optional, Sequence +import textwrap +from typing import List, Optional, Sequence # Shim to wrap setup.py invocation with setuptools -# -# We set sys.argv[0] to the path to the underlying setup.py file so -# setuptools / distutils don't take the path to the setup.py to be "-c" when -# invoking via the shim. This avoids e.g. the following manifest_maker -# warning: "warning: manifest_maker: standard file '-c' not found". -_SETUPTOOLS_SHIM = ( - "import sys, setuptools, tokenize; sys.argv[0] = {0!r}; __file__={0!r};" - "f=getattr(tokenize, 'open', open)(__file__);" - "code=f.read().replace('\\r\\n', '\\n');" - "f.close();" - "exec(compile(code, __file__, 'exec'))" -) +# Note that __file__ is handled via two {!r} *and* %r, to ensure that paths on +# Windows are correctly handled (it should be "C:\\Users" not "C:\Users"). +_SETUPTOOLS_SHIM = textwrap.dedent( + """ + exec(compile(''' + # This is -- a caller that pip uses to run setup.py + # + # - It imports setuptools before invoking setup.py, to enable projects that directly + # import from `distutils.core` to work with newer packaging standards. + # - It provides a clear error message when setuptools is not installed. + # - It sets `sys.argv[0]` to the underlying `setup.py`, when invoking `setup.py` so + # setuptools doesn't think the script is `-c`. This avoids the following warning: + # manifest_maker: standard file '-c' not found". + # - It generates a shim setup.py, for handling setup.cfg-only projects. + import os, sys, tokenize + + try: + import setuptools + except ImportError as error: + print( + "ERROR: Can not execute `setup.py` since setuptools is not available in " + "the build environment.", + file=sys.stderr, + ) + sys.exit(1) + + __file__ = %r + sys.argv[0] = __file__ + + if os.path.exists(__file__): + filename = __file__ + with tokenize.open(__file__) as f: + setup_py_code = f.read() + else: + filename = "" + setup_py_code = "from setuptools import setup; setup()" + + exec(compile(setup_py_code, filename, "exec")) + ''' % ({!r},), "", "exec")) + """ +).rstrip() def make_setuptools_shim_args( - setup_py_path, # type: str - global_options=None, # type: Sequence[str] - no_user_config=False, # type: bool - unbuffered_output=False # type: bool -): - # type: (...) -> List[str] + setup_py_path: str, + global_options: Optional[Sequence[str]] = None, + no_user_config: bool = False, + unbuffered_output: bool = False, +) -> List[str]: """ Get setuptools command arguments with shim wrapped setup file invocation. @@ -48,20 +73,17 @@ def make_setuptools_shim_args( def make_setuptools_bdist_wheel_args( - setup_py_path, # type: str - global_options, # type: Sequence[str] - build_options, # type: Sequence[str] - destination_dir, # type: str -): - # type: (...) -> List[str] + setup_py_path: str, + global_options: Sequence[str], + build_options: Sequence[str], + destination_dir: str, +) -> List[str]: # NOTE: Eventually, we'd want to also -S to the flags here, when we're # isolating. Currently, it breaks Python in virtualenvs, because it # relies on site.py to find parts of the standard library outside the # virtualenv. args = make_setuptools_shim_args( - setup_py_path, - global_options=global_options, - unbuffered_output=True + setup_py_path, global_options=global_options, unbuffered_output=True ) args += ["bdist_wheel", "-d", destination_dir] args += build_options @@ -69,29 +91,25 @@ def make_setuptools_bdist_wheel_args( def make_setuptools_clean_args( - setup_py_path, # type: str - global_options, # type: Sequence[str] -): - # type: (...) -> List[str] + setup_py_path: str, + global_options: Sequence[str], +) -> List[str]: args = make_setuptools_shim_args( - setup_py_path, - global_options=global_options, - unbuffered_output=True + setup_py_path, global_options=global_options, unbuffered_output=True ) args += ["clean", "--all"] return args def make_setuptools_develop_args( - setup_py_path, # type: str - global_options, # type: Sequence[str] - install_options, # type: Sequence[str] - no_user_config, # type: bool - prefix, # type: Optional[str] - home, # type: Optional[str] - use_user_site, # type: bool -): - # type: (...) -> List[str] + setup_py_path: str, + *, + global_options: Sequence[str], + no_user_config: bool, + prefix: Optional[str], + home: Optional[str], + use_user_site: bool, +) -> List[str]: assert not (use_user_site and prefix) args = make_setuptools_shim_args( @@ -102,12 +120,10 @@ def make_setuptools_develop_args( args += ["develop", "--no-deps"] - args += install_options - if prefix: args += ["--prefix", prefix] if home is not None: - args += ["--home", home] + args += ["--install-dir", home] if use_user_site: args += ["--user", "--prefix="] @@ -116,14 +132,11 @@ def make_setuptools_develop_args( def make_setuptools_egg_info_args( - setup_py_path, # type: str - egg_info_dir, # type: Optional[str] - no_user_config, # type: bool -): - # type: (...) -> List[str] - args = make_setuptools_shim_args( - setup_py_path, no_user_config=no_user_config - ) + setup_py_path: str, + egg_info_dir: Optional[str], + no_user_config: bool, +) -> List[str]: + args = make_setuptools_shim_args(setup_py_path, no_user_config=no_user_config) args += ["egg_info"] @@ -131,51 +144,3 @@ def make_setuptools_egg_info_args( args += ["--egg-base", egg_info_dir] return args - - -def make_setuptools_install_args( - setup_py_path, # type: str - global_options, # type: Sequence[str] - install_options, # type: Sequence[str] - record_filename, # type: str - root, # type: Optional[str] - prefix, # type: Optional[str] - header_dir, # type: Optional[str] - home, # type: Optional[str] - use_user_site, # type: bool - no_user_config, # type: bool - pycompile # type: bool -): - # type: (...) -> List[str] - assert not (use_user_site and prefix) - assert not (use_user_site and root) - - args = make_setuptools_shim_args( - setup_py_path, - global_options=global_options, - no_user_config=no_user_config, - unbuffered_output=True - ) - args += ["install", "--record", record_filename] - args += ["--single-version-externally-managed"] - - if root is not None: - args += ["--root", root] - if prefix is not None: - args += ["--prefix", prefix] - if home is not None: - args += ["--home", home] - if use_user_site: - args += ["--user", "--prefix="] - - if pycompile: - args += ["--compile"] - else: - args += ["--no-compile"] - - if header_dir: - args += ["--install-headers", header_dir] - - args += install_options - - return args diff --git a/src/fetchcode/vcs/pip/_internal/utils/subprocess.py b/src/fetchcode/vcs/pip/_internal/utils/subprocess.py index 32cf4e8f..cb2e23f0 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/subprocess.py +++ b/src/fetchcode/vcs/pip/_internal/utils/subprocess.py @@ -1,38 +1,24 @@ -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False - -from __future__ import absolute_import - import logging import os +import shlex import subprocess +from typing import Any, Callable, Iterable, List, Literal, Mapping, Optional, Union -from fetchcode.vcs.pip._vendor.six.moves import shlex_quote - -from fetchcode.vcs.pip._internal.cli.spinners import SpinnerInterface, open_spinner -from fetchcode.vcs.pip._internal.exceptions import InstallationError -from fetchcode.vcs.pip._internal.utils.compat import console_to_str, str_to_display -from fetchcode.vcs.pip._internal.utils.logging import subprocess_logger -from fetchcode.vcs.pip._internal.utils.misc import HiddenText, path_to_display -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import ( - Any, Callable, Iterable, List, Mapping, Optional, Text, Union, - ) - - CommandArgs = List[Union[str, HiddenText]] +from pip._vendor.rich.markup import escape +from pip._internal.cli.spinners import SpinnerInterface, open_spinner +from pip._internal.exceptions import InstallationSubprocessError +from pip._internal.utils.logging import VERBOSE, subprocess_logger +from pip._internal.utils.misc import HiddenText -LOG_DIVIDER = '----------------------------------------' +CommandArgs = List[Union[str, HiddenText]] -def make_command(*args): - # type: (Union[str, HiddenText, CommandArgs]) -> CommandArgs +def make_command(*args: Union[str, HiddenText, CommandArgs]) -> CommandArgs: """ Create a CommandArgs object. """ - command_args = [] # type: CommandArgs + command_args: CommandArgs = [] for arg in args: # Check for list instead of CommandArgs since CommandArgs is # only known during type-checking. @@ -45,8 +31,7 @@ def make_command(*args): return command_args -def format_command_args(args): - # type: (Union[List[str], CommandArgs]) -> str +def format_command_args(args: Union[List[str], CommandArgs]) -> str: """ Format command arguments for display. """ @@ -55,77 +40,33 @@ def format_command_args(args): # this can trigger a UnicodeDecodeError in Python 2 if the argument # has type unicode and includes a non-ascii character. (The type # checker doesn't ensure the annotations are correct in all cases.) - return ' '.join( - shlex_quote(str(arg)) if isinstance(arg, HiddenText) - else shlex_quote(arg) for arg in args + return " ".join( + shlex.quote(str(arg)) if isinstance(arg, HiddenText) else shlex.quote(arg) + for arg in args ) -def reveal_command_args(args): - # type: (Union[List[str], CommandArgs]) -> List[str] +def reveal_command_args(args: Union[List[str], CommandArgs]) -> List[str]: """ Return the arguments in their raw, unredacted form. """ - return [ - arg.secret if isinstance(arg, HiddenText) else arg for arg in args - ] - - -def make_subprocess_output_error( - cmd_args, # type: Union[List[str], CommandArgs] - cwd, # type: Optional[str] - lines, # type: List[Text] - exit_status, # type: int -): - # type: (...) -> Text - """ - Create and return the error message to use to log a subprocess error - with command output. - - :param lines: A list of lines, each ending with a newline. - """ - command = format_command_args(cmd_args) - # Convert `command` and `cwd` to text (unicode in Python 2) so we can use - # them as arguments in the unicode format string below. This avoids - # "UnicodeDecodeError: 'ascii' codec can't decode byte ..." in Python 2 - # if either contains a non-ascii character. - command_display = str_to_display(command, desc='command bytes') - cwd_display = path_to_display(cwd) - - # We know the joined output value ends in a newline. - output = ''.join(lines) - msg = ( - # Use a unicode string to avoid "UnicodeEncodeError: 'ascii' - # codec can't encode character ..." in Python 2 when a format - # argument (e.g. `output`) has a non-ascii character. - u'Command errored out with exit status {exit_status}:\n' - ' command: {command_display}\n' - ' cwd: {cwd_display}\n' - 'Complete output ({line_count} lines):\n{output}{divider}' - ).format( - exit_status=exit_status, - command_display=command_display, - cwd_display=cwd_display, - line_count=len(lines), - output=output, - divider=LOG_DIVIDER, - ) - return msg + return [arg.secret if isinstance(arg, HiddenText) else arg for arg in args] def call_subprocess( - cmd, # type: Union[List[str], CommandArgs] - show_stdout=False, # type: bool - cwd=None, # type: Optional[str] - on_returncode='raise', # type: str - extra_ok_returncodes=None, # type: Optional[Iterable[int]] - command_desc=None, # type: Optional[str] - extra_environ=None, # type: Optional[Mapping[str, Any]] - unset_environ=None, # type: Optional[Iterable[str]] - spinner=None, # type: Optional[SpinnerInterface] - log_failed_cmd=True # type: Optional[bool] -): - # type: (...) -> Text + cmd: Union[List[str], CommandArgs], + show_stdout: bool = False, + cwd: Optional[str] = None, + on_returncode: 'Literal["raise", "warn", "ignore"]' = "raise", + extra_ok_returncodes: Optional[Iterable[int]] = None, + extra_environ: Optional[Mapping[str, Any]] = None, + unset_environ: Optional[Iterable[str]] = None, + spinner: Optional[SpinnerInterface] = None, + log_failed_cmd: Optional[bool] = True, + stdout_only: Optional[bool] = False, + *, + command_desc: str, +) -> str: """ Args: show_stdout: if true, use INFO to log the subprocess's stderr and @@ -135,6 +76,9 @@ def call_subprocess( unset_environ: an iterable of environment variable names to unset prior to calling subprocess.Popen(). log_failed_cmd: if false, failed commands are not logged, only raised. + stdout_only: if true, return only stdout, else return both. When true, + logging of both stdout and stderr occurs when the subprocess has + terminated, else logging occurs as subprocess output is produced. """ if extra_ok_returncodes is None: extra_ok_returncodes = [] @@ -157,13 +101,13 @@ def call_subprocess( # replaced by INFO. if show_stdout: # Then log the subprocess output at INFO level. - log_subprocess = subprocess_logger.info + log_subprocess: Callable[..., None] = subprocess_logger.info used_level = logging.INFO else: - # Then log the subprocess output using DEBUG. This also ensures + # Then log the subprocess output using VERBOSE. This also ensures # it will be logged to the log file (aka user_log), if enabled. - log_subprocess = subprocess_logger.debug - used_level = logging.DEBUG + log_subprocess = subprocess_logger.verbose + used_level = VERBOSE # Whether the subprocess will be visible in the console. showing_subprocess = subprocess_logger.getEffectiveLevel() <= used_level @@ -172,9 +116,6 @@ def call_subprocess( # and we have a spinner. use_spinner = not showing_subprocess and spinner is not None - if command_desc is None: - command_desc = format_command_args(cmd) - log_subprocess("Running command %s", command_desc) env = os.environ.copy() if extra_environ: @@ -185,90 +126,117 @@ def call_subprocess( proc = subprocess.Popen( # Convert HiddenText objects to the underlying str. reveal_command_args(cmd), - stderr=subprocess.STDOUT, stdin=subprocess.PIPE, - stdout=subprocess.PIPE, cwd=cwd, env=env, + stdin=subprocess.PIPE, + stdout=subprocess.PIPE, + stderr=subprocess.STDOUT if not stdout_only else subprocess.PIPE, + cwd=cwd, + env=env, + errors="backslashreplace", ) - proc.stdin.close() except Exception as exc: if log_failed_cmd: subprocess_logger.critical( - "Error %s while executing command %s", exc, command_desc, + "Error %s while executing command %s", + exc, + command_desc, ) raise all_output = [] - while True: - # The "line" value is a unicode string in Python 2. - line = console_to_str(proc.stdout.readline()) - if not line: - break - line = line.rstrip() - all_output.append(line + '\n') - - # Show the line immediately. - log_subprocess(line) - # Update the spinner. - if use_spinner: - spinner.spin() - try: - proc.wait() - finally: - if proc.stdout: - proc.stdout.close() - proc_had_error = ( - proc.returncode and proc.returncode not in extra_ok_returncodes - ) + if not stdout_only: + assert proc.stdout + assert proc.stdin + proc.stdin.close() + # In this mode, stdout and stderr are in the same pipe. + while True: + line: str = proc.stdout.readline() + if not line: + break + line = line.rstrip() + all_output.append(line + "\n") + + # Show the line immediately. + log_subprocess(line) + # Update the spinner. + if use_spinner: + assert spinner + spinner.spin() + try: + proc.wait() + finally: + if proc.stdout: + proc.stdout.close() + output = "".join(all_output) + else: + # In this mode, stdout and stderr are in different pipes. + # We must use communicate() which is the only safe way to read both. + out, err = proc.communicate() + # log line by line to preserve pip log indenting + for out_line in out.splitlines(): + log_subprocess(out_line) + all_output.append(out) + for err_line in err.splitlines(): + log_subprocess(err_line) + all_output.append(err) + output = out + + proc_had_error = proc.returncode and proc.returncode not in extra_ok_returncodes if use_spinner: + assert spinner if proc_had_error: spinner.finish("error") else: spinner.finish("done") if proc_had_error: - if on_returncode == 'raise': - if not showing_subprocess and log_failed_cmd: - # Then the subprocess streams haven't been logged to the - # console yet. - msg = make_subprocess_output_error( - cmd_args=cmd, - cwd=cwd, - lines=all_output, - exit_status=proc.returncode, + if on_returncode == "raise": + error = InstallationSubprocessError( + command_description=command_desc, + exit_code=proc.returncode, + output_lines=all_output if not showing_subprocess else None, + ) + if log_failed_cmd: + subprocess_logger.error("%s", error, extra={"rich": True}) + subprocess_logger.verbose( + "[bold magenta]full command[/]: [blue]%s[/]", + escape(format_command_args(cmd)), + extra={"markup": True}, ) - subprocess_logger.error(msg) - exc_msg = ( - 'Command errored out with exit status {}: {} ' - 'Check the logs for full command output.' - ).format(proc.returncode, command_desc) - raise InstallationError(exc_msg) - elif on_returncode == 'warn': + subprocess_logger.verbose( + "[bold magenta]cwd[/]: %s", + escape(cwd or "[inherit]"), + extra={"markup": True}, + ) + + raise error + elif on_returncode == "warn": subprocess_logger.warning( - 'Command "{}" had error code {} in {}'.format( - command_desc, proc.returncode, cwd) + 'Command "%s" had error code %s in %s', + command_desc, + proc.returncode, + cwd, ) - elif on_returncode == 'ignore': + elif on_returncode == "ignore": pass else: - raise ValueError('Invalid value: on_returncode={!r}'.format( - on_returncode)) - return ''.join(all_output) + raise ValueError(f"Invalid value: on_returncode={on_returncode!r}") + return output -def runner_with_spinner_message(message): - # type: (str) -> Callable[..., None] +def runner_with_spinner_message(message: str) -> Callable[..., None]: """Provide a subprocess_runner that shows a spinner message. - Intended for use with for pep517's Pep517HookCaller. Thus, the runner has - an API that matches what's expected by Pep517HookCaller.subprocess_runner. + Intended for use with for BuildBackendHookCaller. Thus, the runner has + an API that matches what's expected by BuildBackendHookCaller.subprocess_runner. """ def runner( - cmd, # type: List[str] - cwd=None, # type: Optional[str] - extra_environ=None # type: Optional[Mapping[str, Any]] - ): - # type: (...) -> None + cmd: List[str], + cwd: Optional[str] = None, + extra_environ: Optional[Mapping[str, Any]] = None, + ) -> None: with open_spinner(message) as spinner: call_subprocess( cmd, + command_desc=message, cwd=cwd, extra_environ=extra_environ, spinner=spinner, diff --git a/src/fetchcode/vcs/pip/_internal/utils/temp_dir.py b/src/fetchcode/vcs/pip/_internal/utils/temp_dir.py index beca877d..06668e8a 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/temp_dir.py +++ b/src/fetchcode/vcs/pip/_internal/utils/temp_dir.py @@ -1,25 +1,28 @@ -from __future__ import absolute_import - import errno import itertools import logging import os.path import tempfile -from contextlib import contextmanager - -from fetchcode.vcs.pip._vendor.contextlib2 import ExitStack - -from fetchcode.vcs.pip._internal.utils.misc import enum, rmtree -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import Any, Dict, Iterator, Optional, TypeVar, Union - - _T = TypeVar('_T', bound='TempDirectory') +import traceback +from contextlib import ExitStack, contextmanager +from pathlib import Path +from typing import ( + Any, + Callable, + Dict, + Generator, + List, + Optional, + TypeVar, + Union, +) +from pip._internal.utils.misc import enum, rmtree logger = logging.getLogger(__name__) +_T = TypeVar("_T", bound="TempDirectory") + # Kinds of temporary directories. Only needed for ones that are # globally-managed. @@ -30,12 +33,11 @@ ) -_tempdir_manager = None # type: Optional[ExitStack] +_tempdir_manager: Optional[ExitStack] = None @contextmanager -def global_tempdir_manager(): - # type: () -> Iterator[None] +def global_tempdir_manager() -> Generator[None, None, None]: global _tempdir_manager with ExitStack() as stack: old_tempdir_manager, _tempdir_manager = _tempdir_manager, stack @@ -45,35 +47,30 @@ def global_tempdir_manager(): _tempdir_manager = old_tempdir_manager -class TempDirectoryTypeRegistry(object): - """Manages temp directory behavior - """ +class TempDirectoryTypeRegistry: + """Manages temp directory behavior""" - def __init__(self): - # type: () -> None - self._should_delete = {} # type: Dict[str, bool] + def __init__(self) -> None: + self._should_delete: Dict[str, bool] = {} - def set_delete(self, kind, value): - # type: (str, bool) -> None + def set_delete(self, kind: str, value: bool) -> None: """Indicate whether a TempDirectory of the given kind should be auto-deleted. """ self._should_delete[kind] = value - def get_delete(self, kind): - # type: (str) -> bool + def get_delete(self, kind: str) -> bool: """Get configured auto-delete flag for a given TempDirectory type, default True. """ return self._should_delete.get(kind, True) -_tempdir_registry = None # type: Optional[TempDirectoryTypeRegistry] +_tempdir_registry: Optional[TempDirectoryTypeRegistry] = None @contextmanager -def tempdir_registry(): - # type: () -> Iterator[TempDirectoryTypeRegistry] +def tempdir_registry() -> Generator[TempDirectoryTypeRegistry, None, None]: """Provides a scoped global tempdir registry that can be used to dictate whether directories should be deleted. """ @@ -86,14 +83,14 @@ def tempdir_registry(): _tempdir_registry = old_tempdir_registry -class _Default(object): +class _Default: pass _default = _Default() -class TempDirectory(object): +class TempDirectory: """Helper class that owns and cleans up a temporary directory. This class can be used as a context manager or as an OO representation of a @@ -116,12 +113,13 @@ class TempDirectory(object): def __init__( self, - path=None, # type: Optional[str] - delete=_default, # type: Union[bool, None, _Default] - kind="temp", # type: str - globally_managed=False, # type: bool + path: Optional[str] = None, + delete: Union[bool, None, _Default] = _default, + kind: str = "temp", + globally_managed: bool = False, + ignore_cleanup_errors: bool = True, ): - super(TempDirectory, self).__init__() + super().__init__() if delete is _default: if path is not None: @@ -133,6 +131,8 @@ def __init__( # tempdir_registry says. delete = None + # The only time we specify path is in for editables where it + # is the value of the --src option. if path is None: path = self._create(kind) @@ -140,29 +140,24 @@ def __init__( self._deleted = False self.delete = delete self.kind = kind + self.ignore_cleanup_errors = ignore_cleanup_errors if globally_managed: assert _tempdir_manager is not None _tempdir_manager.enter_context(self) @property - def path(self): - # type: () -> str - assert not self._deleted, ( - "Attempted to access deleted path: {}".format(self._path) - ) + def path(self) -> str: + assert not self._deleted, f"Attempted to access deleted path: {self._path}" return self._path - def __repr__(self): - # type: () -> str - return "<{} {!r}>".format(self.__class__.__name__, self.path) + def __repr__(self) -> str: + return f"<{self.__class__.__name__} {self.path!r}>" - def __enter__(self): - # type: (_T) -> _T + def __enter__(self: _T) -> _T: return self - def __exit__(self, exc, value, tb): - # type: (Any, Any, Any) -> None + def __exit__(self, exc: Any, value: Any, tb: Any) -> None: if self.delete is not None: delete = self.delete elif _tempdir_registry: @@ -173,26 +168,58 @@ def __exit__(self, exc, value, tb): if delete: self.cleanup() - def _create(self, kind): - # type: (str) -> str - """Create a temporary directory and store its path in self.path - """ + def _create(self, kind: str) -> str: + """Create a temporary directory and store its path in self.path""" # We realpath here because some systems have their default tmpdir # symlinked to another directory. This tends to confuse build # scripts, so we canonicalize the path by traversing potential # symlinks here. - path = os.path.realpath( - tempfile.mkdtemp(prefix="pip-{}-".format(kind)) - ) - logger.debug("Created temporary directory: {}".format(path)) + path = os.path.realpath(tempfile.mkdtemp(prefix=f"pip-{kind}-")) + logger.debug("Created temporary directory: %s", path) return path - def cleanup(self): - # type: () -> None - """Remove the temporary directory created and reset state - """ + def cleanup(self) -> None: + """Remove the temporary directory created and reset state""" self._deleted = True - if os.path.exists(self._path): + if not os.path.exists(self._path): + return + + errors: List[BaseException] = [] + + def onerror( + func: Callable[..., Any], + path: Path, + exc_val: BaseException, + ) -> None: + """Log a warning for a `rmtree` error and continue""" + formatted_exc = "\n".join( + traceback.format_exception_only(type(exc_val), exc_val) + ) + formatted_exc = formatted_exc.rstrip() # remove trailing new line + if func in (os.unlink, os.remove, os.rmdir): + logger.debug( + "Failed to remove a temporary file '%s' due to %s.\n", + path, + formatted_exc, + ) + else: + logger.debug("%s failed with %s.", func.__qualname__, formatted_exc) + errors.append(exc_val) + + if self.ignore_cleanup_errors: + try: + # first try with @retry; retrying to handle ephemeral errors + rmtree(self._path, ignore_errors=False) + except OSError: + # last pass ignore/log all errors + rmtree(self._path, onexc=onerror) + if errors: + logger.warning( + "Failed to remove contents in a temporary directory '%s'.\n" + "You can safely remove it manually.", + self._path, + ) + else: rmtree(self._path) @@ -210,6 +237,7 @@ class AdjacentTempDirectory(TempDirectory): (when used as a contextmanager) """ + # The characters that may be used to name the temp directory # We always prepend a ~ and then rotate through these until # a usable name is found. @@ -217,14 +245,12 @@ class AdjacentTempDirectory(TempDirectory): # with leading '-' and invalid metadata LEADING_CHARS = "-~.=%0123456789" - def __init__(self, original, delete=None): - # type: (str, Optional[bool]) -> None - self.original = original.rstrip('/\\') - super(AdjacentTempDirectory, self).__init__(delete=delete) + def __init__(self, original: str, delete: Optional[bool] = None) -> None: + self.original = original.rstrip("/\\") + super().__init__(delete=delete) @classmethod - def _generate_names(cls, name): - # type: (str) -> Iterator[str] + def _generate_names(cls, name: str) -> Generator[str, None, None]: """Generates a series of temporary names. The algorithm replaces the leading characters in the name @@ -234,21 +260,22 @@ def _generate_names(cls, name): """ for i in range(1, len(name)): for candidate in itertools.combinations_with_replacement( - cls.LEADING_CHARS, i - 1): - new_name = '~' + ''.join(candidate) + name[i:] + cls.LEADING_CHARS, i - 1 + ): + new_name = "~" + "".join(candidate) + name[i:] if new_name != name: yield new_name # If we make it this far, we will have to make a longer name for i in range(len(cls.LEADING_CHARS)): for candidate in itertools.combinations_with_replacement( - cls.LEADING_CHARS, i): - new_name = '~' + ''.join(candidate) + name + cls.LEADING_CHARS, i + ): + new_name = "~" + "".join(candidate) + name if new_name != name: yield new_name - def _create(self, kind): - # type: (str) -> str + def _create(self, kind: str) -> str: root, name = os.path.split(self.original) for candidate in self._generate_names(name): path = os.path.join(root, candidate) @@ -263,9 +290,7 @@ def _create(self, kind): break else: # Final fallback on the default behavior. - path = os.path.realpath( - tempfile.mkdtemp(prefix="pip-{}-".format(kind)) - ) + path = os.path.realpath(tempfile.mkdtemp(prefix=f"pip-{kind}-")) - logger.debug("Created temporary directory: {}".format(path)) + logger.debug("Created temporary directory: %s", path) return path diff --git a/src/fetchcode/vcs/pip/_internal/utils/typing.py b/src/fetchcode/vcs/pip/_internal/utils/typing.py deleted file mode 100644 index 99cfc03f..00000000 --- a/src/fetchcode/vcs/pip/_internal/utils/typing.py +++ /dev/null @@ -1,38 +0,0 @@ -"""For neatly implementing static typing in pip. - -`mypy` - the static type analysis tool we use - uses the `typing` module, which -provides core functionality fundamental to mypy's functioning. - -Generally, `typing` would be imported at runtime and used in that fashion - -it acts as a no-op at runtime and does not have any run-time overhead by -design. - -As it turns out, `typing` is not vendorable - it uses separate sources for -Python 2/Python 3. Thus, this codebase can not expect it to be present. -To work around this, mypy allows the typing import to be behind a False-y -optional to prevent it from running at runtime and type-comments can be used -to remove the need for the types to be accessible directly during runtime. - -This module provides the False-y guard in a nicely named fashion so that a -curious maintainer can reach here to read this. - -In pip, all static-typing related imports should be guarded as follows: - - from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - - if MYPY_CHECK_RUNNING: - from typing import ... - -Ref: https://github.com/python/mypy/issues/3216 -""" - -MYPY_CHECK_RUNNING = False - - -if MYPY_CHECK_RUNNING: - from typing import cast -else: - # typing's cast() is needed at runtime, but we don't want to import typing. - # Thus, we use a dummy no-op version, which we tell mypy to ignore. - def cast(type_, value): # type: ignore - return value diff --git a/src/fetchcode/vcs/pip/_internal/utils/unpacking.py b/src/fetchcode/vcs/pip/_internal/utils/unpacking.py index 0452bbbc..875e30e1 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/unpacking.py +++ b/src/fetchcode/vcs/pip/_internal/utils/unpacking.py @@ -1,32 +1,24 @@ """Utilities related archives. """ -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import - import logging import os import shutil import stat +import sys import tarfile import zipfile +from typing import Iterable, List, Optional +from zipfile import ZipInfo -from fetchcode.vcs.pip._internal.exceptions import InstallationError -from fetchcode.vcs.pip._internal.utils.filetypes import ( +from pip._internal.exceptions import InstallationError +from pip._internal.utils.filetypes import ( BZ2_EXTENSIONS, TAR_EXTENSIONS, XZ_EXTENSIONS, ZIP_EXTENSIONS, ) -from fetchcode.vcs.pip._internal.utils.misc import ensure_dir -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import Iterable, List, Optional, Text, Union - +from pip._internal.utils.misc import ensure_dir logger = logging.getLogger(__name__) @@ -35,43 +27,40 @@ try: import bz2 # noqa + SUPPORTED_EXTENSIONS += BZ2_EXTENSIONS except ImportError: - logger.debug('bz2 module is not available') + logger.debug("bz2 module is not available") try: # Only for Python 3.3+ import lzma # noqa + SUPPORTED_EXTENSIONS += XZ_EXTENSIONS except ImportError: - logger.debug('lzma module is not available') + logger.debug("lzma module is not available") -def current_umask(): +def current_umask() -> int: """Get the current umask which involves having to set it temporarily.""" mask = os.umask(0) os.umask(mask) return mask -def split_leading_dir(path): - # type: (Union[str, Text]) -> List[Union[str, Text]] - path = path.lstrip('/').lstrip('\\') - if ( - '/' in path and ( - ('\\' in path and path.find('/') < path.find('\\')) or - '\\' not in path - ) +def split_leading_dir(path: str) -> List[str]: + path = path.lstrip("/").lstrip("\\") + if "/" in path and ( + ("\\" in path and path.find("/") < path.find("\\")) or "\\" not in path ): - return path.split('/', 1) - elif '\\' in path: - return path.split('\\', 1) + return path.split("/", 1) + elif "\\" in path: + return path.split("\\", 1) else: - return [path, ''] + return [path, ""] -def has_leading_dir(paths): - # type: (Iterable[Union[str, Text]]) -> bool +def has_leading_dir(paths: Iterable[str]) -> bool: """Returns true if all the paths have the same leading path name (i.e., everything is in one subdirectory in an archive)""" common_prefix = None @@ -86,8 +75,7 @@ def has_leading_dir(paths): return True -def is_within_directory(directory, target): - # type: ((Union[str, Text]), (Union[str, Text])) -> bool +def is_within_directory(directory: str, target: str) -> bool: """ Return true if the absolute path of target is within the directory """ @@ -98,8 +86,26 @@ def is_within_directory(directory, target): return prefix == abs_directory -def unzip_file(filename, location, flatten=True): - # type: (str, str, bool) -> None +def _get_default_mode_plus_executable() -> int: + return 0o777 & ~current_umask() | 0o111 + + +def set_extracted_file_to_default_mode_plus_executable(path: str) -> None: + """ + Make file present at path have execute for user/group/world + (chmod +x) is no-op on windows per python docs + """ + os.chmod(path, _get_default_mode_plus_executable()) + + +def zip_item_is_executable(info: ZipInfo) -> bool: + mode = info.external_attr >> 16 + # if mode and regular file and any execute permissions for + # user/group/world? + return bool(mode and stat.S_ISREG(mode) and mode & 0o111) + + +def unzip_file(filename: str, location: str, flatten: bool = True) -> None: """ Unzip the file (with path `filename`) to the destination `location`. All files are written based on system defaults and umask (i.e. permissions are @@ -109,7 +115,7 @@ def unzip_file(filename, location, flatten=True): no-ops per the python docs. """ ensure_dir(location) - zipfp = open(filename, 'rb') + zipfp = open(filename, "rb") try: zip = zipfile.ZipFile(zipfp, allowZip64=True) leading = has_leading_dir(zip.namelist()) and flatten @@ -122,11 +128,11 @@ def unzip_file(filename, location, flatten=True): dir = os.path.dirname(fn) if not is_within_directory(location, fn): message = ( - 'The zip file ({}) has a file ({}) trying to install ' - 'outside target directory ({})' + "The zip file ({}) has a file ({}) trying to install " + "outside target directory ({})" ) raise InstallationError(message.format(filename, fn, location)) - if fn.endswith('/') or fn.endswith('\\'): + if fn.endswith("/") or fn.endswith("\\"): # A directory ensure_dir(fn) else: @@ -135,138 +141,197 @@ def unzip_file(filename, location, flatten=True): # chunk of memory for the file's content fp = zip.open(name) try: - with open(fn, 'wb') as destfp: + with open(fn, "wb") as destfp: shutil.copyfileobj(fp, destfp) finally: fp.close() - mode = info.external_attr >> 16 - # if mode and regular file and any execute permissions for - # user/group/world? - if mode and stat.S_ISREG(mode) and mode & 0o111: - # make dest file have execute for user/group/world - # (chmod +x) no-op on windows per python docs - os.chmod(fn, (0o777 - current_umask() | 0o111)) + if zip_item_is_executable(info): + set_extracted_file_to_default_mode_plus_executable(fn) finally: zipfp.close() -def untar_file(filename, location): - # type: (str, str) -> None +def untar_file(filename: str, location: str) -> None: """ Untar the file (with path `filename`) to the destination `location`. All files are written based on system defaults and umask (i.e. permissions are not preserved), except that regular file members with any execute - permissions (user, group, or world) have "chmod +x" applied after being - written. Note that for windows, any execute changes using os.chmod are + permissions (user, group, or world) have "chmod +x" applied on top of the + default. Note that for windows, any execute changes using os.chmod are no-ops per the python docs. """ ensure_dir(location) - if filename.lower().endswith('.gz') or filename.lower().endswith('.tgz'): - mode = 'r:gz' + if filename.lower().endswith(".gz") or filename.lower().endswith(".tgz"): + mode = "r:gz" elif filename.lower().endswith(BZ2_EXTENSIONS): - mode = 'r:bz2' + mode = "r:bz2" elif filename.lower().endswith(XZ_EXTENSIONS): - mode = 'r:xz' - elif filename.lower().endswith('.tar'): - mode = 'r' + mode = "r:xz" + elif filename.lower().endswith(".tar"): + mode = "r" else: logger.warning( - 'Cannot determine compression type for file %s', filename, + "Cannot determine compression type for file %s", + filename, ) - mode = 'r:*' - tar = tarfile.open(filename, mode) + mode = "r:*" + + tar = tarfile.open(filename, mode, encoding="utf-8") try: - leading = has_leading_dir([ - member.name for member in tar.getmembers() - ]) - for member in tar.getmembers(): - fn = member.name + leading = has_leading_dir([member.name for member in tar.getmembers()]) + + # PEP 706 added `tarfile.data_filter`, and made some other changes to + # Python's tarfile module (see below). The features were backported to + # security releases. + try: + data_filter = tarfile.data_filter + except AttributeError: + _untar_without_filter(filename, location, tar, leading) + else: + default_mode_plus_executable = _get_default_mode_plus_executable() + if leading: - # https://github.com/python/mypy/issues/1174 - fn = split_leading_dir(fn)[1] # type: ignore - path = os.path.join(location, fn) - if not is_within_directory(location, path): - message = ( - 'The tar file ({}) has a file ({}) trying to install ' - 'outside target directory ({})' - ) - raise InstallationError( - message.format(filename, path, location) - ) - if member.isdir(): - ensure_dir(path) - elif member.issym(): - try: - # https://github.com/python/typeshed/issues/2673 - tar._extract_member(member, path) # type: ignore - except Exception as exc: - # Some corrupt tar files seem to produce this - # (specifically bad symlinks) - logger.warning( - 'In the tar file %s the member %s is invalid: %s', - filename, member.name, exc, - ) - continue - else: + # Strip the leading directory from all files in the archive, + # including hardlink targets (which are relative to the + # unpack location). + for member in tar.getmembers(): + name_lead, name_rest = split_leading_dir(member.name) + member.name = name_rest + if member.islnk(): + lnk_lead, lnk_rest = split_leading_dir(member.linkname) + if lnk_lead == name_lead: + member.linkname = lnk_rest + + def pip_filter(member: tarfile.TarInfo, path: str) -> tarfile.TarInfo: + orig_mode = member.mode try: - fp = tar.extractfile(member) - except (KeyError, AttributeError) as exc: - # Some corrupt tar files seem to produce this - # (specifically bad symlinks) - logger.warning( - 'In the tar file %s the member %s is invalid: %s', - filename, member.name, exc, + try: + member = data_filter(member, location) + except tarfile.LinkOutsideDestinationError: + if sys.version_info[:3] in { + (3, 8, 17), + (3, 9, 17), + (3, 10, 12), + (3, 11, 4), + }: + # The tarfile filter in specific Python versions + # raises LinkOutsideDestinationError on valid input + # (https://github.com/python/cpython/issues/107845) + # Ignore the error there, but do use the + # more lax `tar_filter` + member = tarfile.tar_filter(member, location) + else: + raise + except tarfile.TarError as exc: + message = "Invalid member in the tar file {}: {}" + # Filter error messages mention the member name. + # No need to add it here. + raise InstallationError( + message.format( + filename, + exc, + ) ) - continue - ensure_dir(os.path.dirname(path)) - with open(path, 'wb') as destfp: - shutil.copyfileobj(fp, destfp) - fp.close() - # Update the timestamp (useful for cython compiled files) - # https://github.com/python/typeshed/issues/2673 - tar.utime(member, path) # type: ignore - # member have any execute permissions for user/group/world? - if member.mode & 0o111: - # make dest file have execute for user/group/world - # no-op on windows per python docs - os.chmod(path, (0o777 - current_umask() | 0o111)) + if member.isfile() and orig_mode & 0o111: + member.mode = default_mode_plus_executable + else: + # See PEP 706 note above. + # The PEP changed this from `int` to `Optional[int]`, + # where None means "use the default". Mypy doesn't + # know this yet. + member.mode = None # type: ignore [assignment] + return member + + tar.extractall(location, filter=pip_filter) + finally: tar.close() +def _untar_without_filter( + filename: str, + location: str, + tar: tarfile.TarFile, + leading: bool, +) -> None: + """Fallback for Python without tarfile.data_filter""" + for member in tar.getmembers(): + fn = member.name + if leading: + fn = split_leading_dir(fn)[1] + path = os.path.join(location, fn) + if not is_within_directory(location, path): + message = ( + "The tar file ({}) has a file ({}) trying to install " + "outside target directory ({})" + ) + raise InstallationError(message.format(filename, path, location)) + if member.isdir(): + ensure_dir(path) + elif member.issym(): + try: + tar._extract_member(member, path) + except Exception as exc: + # Some corrupt tar files seem to produce this + # (specifically bad symlinks) + logger.warning( + "In the tar file %s the member %s is invalid: %s", + filename, + member.name, + exc, + ) + continue + else: + try: + fp = tar.extractfile(member) + except (KeyError, AttributeError) as exc: + # Some corrupt tar files seem to produce this + # (specifically bad symlinks) + logger.warning( + "In the tar file %s the member %s is invalid: %s", + filename, + member.name, + exc, + ) + continue + ensure_dir(os.path.dirname(path)) + assert fp is not None + with open(path, "wb") as destfp: + shutil.copyfileobj(fp, destfp) + fp.close() + # Update the timestamp (useful for cython compiled files) + tar.utime(member, path) + # member have any execute permissions for user/group/world? + if member.mode & 0o111: + set_extracted_file_to_default_mode_plus_executable(path) + + def unpack_file( - filename, # type: str - location, # type: str - content_type=None, # type: Optional[str] -): - # type: (...) -> None + filename: str, + location: str, + content_type: Optional[str] = None, +) -> None: filename = os.path.realpath(filename) if ( - content_type == 'application/zip' or - filename.lower().endswith(ZIP_EXTENSIONS) or - zipfile.is_zipfile(filename) + content_type == "application/zip" + or filename.lower().endswith(ZIP_EXTENSIONS) + or zipfile.is_zipfile(filename) ): - unzip_file( - filename, - location, - flatten=not filename.endswith('.whl') - ) + unzip_file(filename, location, flatten=not filename.endswith(".whl")) elif ( - content_type == 'application/x-gzip' or - tarfile.is_tarfile(filename) or - filename.lower().endswith( - TAR_EXTENSIONS + BZ2_EXTENSIONS + XZ_EXTENSIONS - ) + content_type == "application/x-gzip" + or tarfile.is_tarfile(filename) + or filename.lower().endswith(TAR_EXTENSIONS + BZ2_EXTENSIONS + XZ_EXTENSIONS) ): untar_file(filename, location) else: # FIXME: handle? # FIXME: magic signatures? logger.critical( - 'Cannot unpack file %s (downloaded from %s, content-type: %s); ' - 'cannot detect archive format', - filename, location, content_type, - ) - raise InstallationError( - 'Cannot determine archive format of {}'.format(location) + "Cannot unpack file %s (downloaded from %s, content-type: %s); " + "cannot detect archive format", + filename, + location, + content_type, ) + raise InstallationError(f"Cannot determine archive format of {location}") diff --git a/src/fetchcode/vcs/pip/_internal/utils/urls.py b/src/fetchcode/vcs/pip/_internal/utils/urls.py index b2a452ab..9f34f882 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/urls.py +++ b/src/fetchcode/vcs/pip/_internal/utils/urls.py @@ -1,55 +1,55 @@ import os -import sys +import string +import urllib.parse +import urllib.request -from fetchcode.vcs.pip._vendor.six.moves.urllib import parse as urllib_parse -from fetchcode.vcs.pip._vendor.six.moves.urllib import request as urllib_request +from .compat import WINDOWS -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -if MYPY_CHECK_RUNNING: - from typing import Optional, Text, Union - - -def get_url_scheme(url): - # type: (Union[str, Text]) -> Optional[Text] - if ':' not in url: - return None - return url.split(':', 1)[0].lower() - - -def path_to_url(path): - # type: (Union[str, Text]) -> str +def path_to_url(path: str) -> str: """ Convert a path to a file: URL. The path will be made absolute and have quoted path parts. """ path = os.path.normpath(os.path.abspath(path)) - url = urllib_parse.urljoin('file:', urllib_request.pathname2url(path)) + url = urllib.parse.urljoin("file:", urllib.request.pathname2url(path)) return url -def url_to_path(url): - # type: (str) -> str +def url_to_path(url: str) -> str: """ Convert a file: URL to a path. """ - assert url.startswith('file:'), ( - "You can only turn file: urls into filenames (not {url!r})" - .format(**locals())) + assert url.startswith( + "file:" + ), f"You can only turn file: urls into filenames (not {url!r})" - _, netloc, path, _, _ = urllib_parse.urlsplit(url) + _, netloc, path, _, _ = urllib.parse.urlsplit(url) - if not netloc or netloc == 'localhost': + if not netloc or netloc == "localhost": # According to RFC 8089, same as empty authority. - netloc = '' - elif sys.platform == 'win32': + netloc = "" + elif WINDOWS: # If we have a UNC path, prepend UNC share notation. - netloc = '\\\\' + netloc + netloc = "\\\\" + netloc else: raise ValueError( - 'non-local file URIs are not supported on this platform: {url!r}' - .format(**locals()) + f"non-local file URIs are not supported on this platform: {url!r}" ) - path = urllib_request.url2pathname(netloc + path) + path = urllib.request.url2pathname(netloc + path) + + # On Windows, urlsplit parses the path as something like "/C:/Users/foo". + # This creates issues for path-related functions like io.open(), so we try + # to detect and strip the leading slash. + if ( + WINDOWS + and not netloc # Not UNC. + and len(path) >= 3 + and path[0] == "/" # Leading slash to strip. + and path[1] in string.ascii_letters # Drive letter. + and path[2:4] in (":", ":/") # Colon + end of string, or colon + absolute path. + ): + path = path[1:] + return path diff --git a/src/fetchcode/vcs/pip/_internal/utils/virtualenv.py b/src/fetchcode/vcs/pip/_internal/utils/virtualenv.py index 95a37514..882e36f5 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/virtualenv.py +++ b/src/fetchcode/vcs/pip/_internal/utils/virtualenv.py @@ -1,15 +1,9 @@ -from __future__ import absolute_import - import logging import os import re import site import sys - -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING - -if MYPY_CHECK_RUNNING: - from typing import List, Optional +from typing import List, Optional logger = logging.getLogger(__name__) _INCLUDE_SYSTEM_SITE_PACKAGES_REGEX = re.compile( @@ -17,8 +11,7 @@ ) -def _running_under_venv(): - # type: () -> bool +def _running_under_venv() -> bool: """Checks if sys.base_prefix and sys.prefix match. This handles PEP 405 compliant virtual environments. @@ -26,39 +19,36 @@ def _running_under_venv(): return sys.prefix != getattr(sys, "base_prefix", sys.prefix) -def _running_under_regular_virtualenv(): - # type: () -> bool +def _running_under_legacy_virtualenv() -> bool: """Checks if sys.real_prefix is set. This handles virtual environments created with pypa's virtualenv. """ # pypa/virtualenv case - return hasattr(sys, 'real_prefix') + return hasattr(sys, "real_prefix") -def running_under_virtualenv(): - # type: () -> bool - """Return True if we're running inside a virtualenv, False otherwise. - """ - return _running_under_venv() or _running_under_regular_virtualenv() +def running_under_virtualenv() -> bool: + """True if we're running inside a virtual environment, False otherwise.""" + return _running_under_venv() or _running_under_legacy_virtualenv() -def _get_pyvenv_cfg_lines(): - # type: () -> Optional[List[str]] +def _get_pyvenv_cfg_lines() -> Optional[List[str]]: """Reads {sys.prefix}/pyvenv.cfg and returns its contents as list of lines Returns None, if it could not read/access the file. """ - pyvenv_cfg_file = os.path.join(sys.prefix, 'pyvenv.cfg') + pyvenv_cfg_file = os.path.join(sys.prefix, "pyvenv.cfg") try: - with open(pyvenv_cfg_file) as f: + # Although PEP 405 does not specify, the built-in venv module always + # writes with UTF-8. (pypa/pip#8717) + with open(pyvenv_cfg_file, encoding="utf-8") as f: return f.read().splitlines() # avoids trailing newlines - except IOError: + except OSError: return None -def _no_global_under_venv(): - # type: () -> bool +def _no_global_under_venv() -> bool: """Check `{sys.prefix}/pyvenv.cfg` for system site-packages inclusion PEP 405 specifies that when system site-packages are not supposed to be @@ -82,13 +72,12 @@ def _no_global_under_venv(): for line in cfg_lines: match = _INCLUDE_SYSTEM_SITE_PACKAGES_REGEX.match(line) - if match is not None and match.group('value') == 'false': + if match is not None and match.group("value") == "false": return True return False -def _no_global_under_regular_virtualenv(): - # type: () -> bool +def _no_global_under_legacy_virtualenv() -> bool: """Check if "no-global-site-packages.txt" exists beside site.py This mirrors logic in pypa/virtualenv for determining whether system @@ -96,21 +85,20 @@ def _no_global_under_regular_virtualenv(): """ site_mod_dir = os.path.dirname(os.path.abspath(site.__file__)) no_global_site_packages_file = os.path.join( - site_mod_dir, 'no-global-site-packages.txt', + site_mod_dir, + "no-global-site-packages.txt", ) return os.path.exists(no_global_site_packages_file) -def virtualenv_no_global(): - # type: () -> bool - """Returns a boolean, whether running in venv with no system site-packages. - """ +def virtualenv_no_global() -> bool: + """Returns a boolean, whether running in venv with no system site-packages.""" # PEP 405 compliance needs to be checked first since virtualenv >=20 would # return True for both checks, but is only able to use the PEP 405 config. if _running_under_venv(): return _no_global_under_venv() - if _running_under_regular_virtualenv(): - return _no_global_under_regular_virtualenv() + if _running_under_legacy_virtualenv(): + return _no_global_under_legacy_virtualenv() return False diff --git a/src/fetchcode/vcs/pip/_internal/utils/wheel.py b/src/fetchcode/vcs/pip/_internal/utils/wheel.py index ad7df75b..f85aee8a 100644 --- a/src/fetchcode/vcs/pip/_internal/utils/wheel.py +++ b/src/fetchcode/vcs/pip/_internal/utils/wheel.py @@ -1,31 +1,15 @@ """Support functions for working with wheel files. """ -from __future__ import absolute_import - import logging +from email.message import Message from email.parser import Parser -from zipfile import ZipFile - -from fetchcode.vcs.pip._vendor.packaging.utils import canonicalize_name -from fetchcode.vcs.pip._vendor.pkg_resources import DistInfoDistribution -from fetchcode.vcs.pip._vendor.six import PY2, ensure_str - -from fetchcode.vcs.pip._internal.exceptions import UnsupportedWheel -from fetchcode.vcs.pip._internal.utils.pkg_resources import DictMetadata -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING +from typing import Tuple +from zipfile import BadZipFile, ZipFile -if MYPY_CHECK_RUNNING: - from email.message import Message - from typing import Dict, Tuple - - from fetchcode.vcs.pip._vendor.pkg_resources import Distribution - -if PY2: - from zipfile import BadZipfile as BadZipFile -else: - from zipfile import BadZipFile +from pip._vendor.packaging.utils import canonicalize_name +from pip._internal.exceptions import UnsupportedWheel VERSION_COMPATIBLE = (1, 0) @@ -33,67 +17,7 @@ logger = logging.getLogger(__name__) -class WheelMetadata(DictMetadata): - """Metadata provider that maps metadata decoding exceptions to our - internal exception type. - """ - def __init__(self, metadata, wheel_name): - # type: (Dict[str, bytes], str) -> None - super(WheelMetadata, self).__init__(metadata) - self._wheel_name = wheel_name - - def get_metadata(self, name): - # type: (str) -> str - try: - return super(WheelMetadata, self).get_metadata(name) - except UnicodeDecodeError as e: - # Augment the default error with the origin of the file. - raise UnsupportedWheel( - "Error decoding metadata for {}: {}".format( - self._wheel_name, e - ) - ) - - -def pkg_resources_distribution_for_wheel(wheel_zip, name, location): - # type: (ZipFile, str, str) -> Distribution - """Get a pkg_resources distribution given a wheel. - - :raises UnsupportedWheel: on any errors - """ - info_dir, _ = parse_wheel(wheel_zip, name) - - metadata_files = [ - p for p in wheel_zip.namelist() if p.startswith("{}/".format(info_dir)) - ] - - metadata_text = {} # type: Dict[str, bytes] - for path in metadata_files: - # If a flag is set, namelist entries may be unicode in Python 2. - # We coerce them to native str type to match the types used in the rest - # of the code. This cannot fail because unicode can always be encoded - # with UTF-8. - full_path = ensure_str(path) - _, metadata_name = full_path.split("/", 1) - - try: - metadata_text[metadata_name] = read_wheel_metadata_file( - wheel_zip, full_path - ) - except UnsupportedWheel as e: - raise UnsupportedWheel( - "{} has an invalid wheel, {}".format(name, str(e)) - ) - - metadata = WheelMetadata(metadata_text, location) - - return DistInfoDistribution( - location=location, metadata=metadata, project_name=name - ) - - -def parse_wheel(wheel_zip, name): - # type: (ZipFile, str) -> Tuple[str, Message] +def parse_wheel(wheel_zip: ZipFile, name: str) -> Tuple[str, Message]: """Extract information from the provided wheel, ensuring it meets basic standards. @@ -104,35 +28,30 @@ def parse_wheel(wheel_zip, name): metadata = wheel_metadata(wheel_zip, info_dir) version = wheel_version(metadata) except UnsupportedWheel as e: - raise UnsupportedWheel( - "{} has an invalid wheel, {}".format(name, str(e)) - ) + raise UnsupportedWheel(f"{name} has an invalid wheel, {e}") check_compatibility(version, name) return info_dir, metadata -def wheel_dist_info_dir(source, name): - # type: (ZipFile, str) -> str +def wheel_dist_info_dir(source: ZipFile, name: str) -> str: """Returns the name of the contained .dist-info directory. Raises AssertionError or UnsupportedWheel if not found, >1 found, or it doesn't match the provided name. """ # Zip file path separators must be / - subdirs = list(set(p.split("/")[0] for p in source.namelist())) + subdirs = {p.split("/", 1)[0] for p in source.namelist()} - info_dirs = [s for s in subdirs if s.endswith('.dist-info')] + info_dirs = [s for s in subdirs if s.endswith(".dist-info")] if not info_dirs: raise UnsupportedWheel(".dist-info directory not found") if len(info_dirs) > 1: raise UnsupportedWheel( - "multiple .dist-info directories found: {}".format( - ", ".join(info_dirs) - ) + "multiple .dist-info directories found: {}".format(", ".join(info_dirs)) ) info_dir = info_dirs[0] @@ -141,41 +60,33 @@ def wheel_dist_info_dir(source, name): canonical_name = canonicalize_name(name) if not info_dir_name.startswith(canonical_name): raise UnsupportedWheel( - ".dist-info directory {!r} does not start with {!r}".format( - info_dir, canonical_name - ) + f".dist-info directory {info_dir!r} does not start with {canonical_name!r}" ) - # Zip file paths can be unicode or str depending on the zip entry flags, - # so normalize it. - return ensure_str(info_dir) + return info_dir -def read_wheel_metadata_file(source, path): - # type: (ZipFile, str) -> bytes +def read_wheel_metadata_file(source: ZipFile, path: str) -> bytes: try: return source.read(path) # BadZipFile for general corruption, KeyError for missing entry, # and RuntimeError for password-protected files except (BadZipFile, KeyError, RuntimeError) as e: - raise UnsupportedWheel( - "could not read {!r} file: {!r}".format(path, e) - ) + raise UnsupportedWheel(f"could not read {path!r} file: {e!r}") -def wheel_metadata(source, dist_info_dir): - # type: (ZipFile, str) -> Message +def wheel_metadata(source: ZipFile, dist_info_dir: str) -> Message: """Return the WHEEL metadata of an extracted wheel, if possible. Otherwise, raise UnsupportedWheel. """ - path = "{}/WHEEL".format(dist_info_dir) + path = f"{dist_info_dir}/WHEEL" # Zip file path separators must be / wheel_contents = read_wheel_metadata_file(source, path) try: - wheel_text = ensure_str(wheel_contents) + wheel_text = wheel_contents.decode() except UnicodeDecodeError as e: - raise UnsupportedWheel("error decoding {!r}: {!r}".format(path, e)) + raise UnsupportedWheel(f"error decoding {path!r}: {e!r}") # FeedParser (used by Parser) does not raise any exceptions. The returned # message may have .defects populated, but for backwards-compatibility we @@ -183,8 +94,7 @@ def wheel_metadata(source, dist_info_dir): return Parser().parsestr(wheel_text) -def wheel_version(wheel_data): - # type: (Message) -> Tuple[int, ...] +def wheel_version(wheel_data: Message) -> Tuple[int, ...]: """Given WHEEL metadata, return the parsed Wheel-Version. Otherwise, raise UnsupportedWheel. """ @@ -195,13 +105,12 @@ def wheel_version(wheel_data): version = version_text.strip() try: - return tuple(map(int, version.split('.'))) + return tuple(map(int, version.split("."))) except ValueError: - raise UnsupportedWheel("invalid Wheel-Version: {!r}".format(version)) + raise UnsupportedWheel(f"invalid Wheel-Version: {version!r}") -def check_compatibility(version, name): - # type: (Tuple[int, ...], str) -> None +def check_compatibility(version: Tuple[int, ...], name: str) -> None: """Raises errors or warns if called with an incompatible Wheel-Version. pip should refuse to install a Wheel-Version that's a major series @@ -216,10 +125,10 @@ def check_compatibility(version, name): if version[0] > VERSION_COMPATIBLE[0]: raise UnsupportedWheel( "{}'s Wheel-Version ({}) is not compatible with this version " - "of pip".format(name, '.'.join(map(str, version))) + "of pip".format(name, ".".join(map(str, version))) ) elif version > VERSION_COMPATIBLE: logger.warning( - 'Installing from a newer Wheel-Version (%s)', - '.'.join(map(str, version)), + "Installing from a newer Wheel-Version (%s)", + ".".join(map(str, version)), ) diff --git a/src/fetchcode/vcs/pip/_internal/vcs/__init__.py b/src/fetchcode/vcs/pip/_internal/vcs/__init__.py index 2086fd98..b6beddbe 100644 --- a/src/fetchcode/vcs/pip/_internal/vcs/__init__.py +++ b/src/fetchcode/vcs/pip/_internal/vcs/__init__.py @@ -1,14 +1,14 @@ # Expose a limited set of classes and functions so callers outside of # the vcs package don't need to import deeper than `pip._internal.vcs`. -# (The test directory and imports protected by MYPY_CHECK_RUNNING may -# still need to import from a vcs sub-package.) +# (The test directory may still need to import from a vcs sub-package.) # Import all vcs modules to register each VCS in the VcsSupport object. -import fetchcode.vcs.pip._internal.vcs.bazaar -import fetchcode.vcs.pip._internal.vcs.git -import fetchcode.vcs.pip._internal.vcs.mercurial -import fetchcode.vcs.pip._internal.vcs.subversion # noqa: F401 -from fetchcode.vcs.pip._internal.vcs.versioncontrol import ( # noqa: F401 +import pip._internal.vcs.bazaar +import pip._internal.vcs.git +import pip._internal.vcs.mercurial +import pip._internal.vcs.subversion # noqa: F401 +from pip._internal.vcs.versioncontrol import ( # noqa: F401 RemoteNotFoundError, + RemoteNotValidError, is_url, make_vcs_requirement_url, vcs, diff --git a/src/fetchcode/vcs/pip/_internal/vcs/bazaar.py b/src/fetchcode/vcs/pip/_internal/vcs/bazaar.py index 28686245..c754b7cc 100644 --- a/src/fetchcode/vcs/pip/_internal/vcs/bazaar.py +++ b/src/fetchcode/vcs/pip/_internal/vcs/bazaar.py @@ -1,118 +1,110 @@ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import - import logging -import os - -from fetchcode.vcs.pip._vendor.six.moves.urllib import parse as urllib_parse - -from fetchcode.vcs.pip._internal.utils.misc import display_path, rmtree -from fetchcode.vcs.pip._internal.utils.subprocess import make_command -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.utils.urls import path_to_url -from fetchcode.vcs.pip._internal.vcs.versioncontrol import VersionControl, vcs - -if MYPY_CHECK_RUNNING: - from typing import Optional, Tuple - from fetchcode.vcs.pip._internal.utils.misc import HiddenText - from fetchcode.vcs.pip._internal.vcs.versioncontrol import AuthInfo, RevOptions - +from typing import List, Optional, Tuple + +from pip._internal.utils.misc import HiddenText, display_path +from pip._internal.utils.subprocess import make_command +from pip._internal.utils.urls import path_to_url +from pip._internal.vcs.versioncontrol import ( + AuthInfo, + RemoteNotFoundError, + RevOptions, + VersionControl, + vcs, +) logger = logging.getLogger(__name__) class Bazaar(VersionControl): - name = 'bzr' - dirname = '.bzr' - repo_name = 'branch' + name = "bzr" + dirname = ".bzr" + repo_name = "branch" schemes = ( - 'bzr', 'bzr+http', 'bzr+https', 'bzr+ssh', 'bzr+sftp', 'bzr+ftp', - 'bzr+lp', + "bzr+http", + "bzr+https", + "bzr+ssh", + "bzr+sftp", + "bzr+ftp", + "bzr+lp", + "bzr+file", ) - def __init__(self, *args, **kwargs): - super(Bazaar, self).__init__(*args, **kwargs) - # This is only needed for python <2.7.5 - # Register lp but do not expose as a scheme to support bzr+lp. - if getattr(urllib_parse, 'uses_fragment', None): - urllib_parse.uses_fragment.extend(['lp']) - @staticmethod - def get_base_rev_args(rev): - return ['-r', rev] - - def export(self, location, url): - # type: (str, HiddenText) -> None - """ - Export the Bazaar repository at the url to the destination location - """ - # Remove the location to make sure Bazaar can export it correctly - if os.path.exists(location): - rmtree(location) + def get_base_rev_args(rev: str) -> List[str]: + return ["-r", rev] - url, rev_options = self.get_url_rev_options(url) - self.run_command( - make_command('export', location, url, rev_options.to_args()), - show_stdout=False, - ) - - def fetch_new(self, dest, url, rev_options): - # type: (str, HiddenText, RevOptions) -> None + def fetch_new( + self, dest: str, url: HiddenText, rev_options: RevOptions, verbosity: int + ) -> None: rev_display = rev_options.to_display() logger.info( - 'Checking out %s%s to %s', + "Checking out %s%s to %s", url, rev_display, display_path(dest), ) - cmd_args = ( - make_command('branch', '-q', rev_options.to_args(), url, dest) + if verbosity <= 0: + flags = ["--quiet"] + elif verbosity == 1: + flags = [] + else: + flags = [f"-{'v'*verbosity}"] + cmd_args = make_command( + "checkout", "--lightweight", *flags, rev_options.to_args(), url, dest ) self.run_command(cmd_args) - def switch(self, dest, url, rev_options): - # type: (str, HiddenText, RevOptions) -> None - self.run_command(make_command('switch', url), cwd=dest) + def switch(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None: + self.run_command(make_command("switch", url), cwd=dest) - def update(self, dest, url, rev_options): - # type: (str, HiddenText, RevOptions) -> None - cmd_args = make_command('pull', '-q', rev_options.to_args()) + def update(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None: + output = self.run_command( + make_command("info"), show_stdout=False, stdout_only=True, cwd=dest + ) + if output.startswith("Standalone "): + # Older versions of pip used to create standalone branches. + # Convert the standalone branch to a checkout by calling "bzr bind". + cmd_args = make_command("bind", "-q", url) + self.run_command(cmd_args, cwd=dest) + + cmd_args = make_command("update", "-q", rev_options.to_args()) self.run_command(cmd_args, cwd=dest) @classmethod - def get_url_rev_and_auth(cls, url): - # type: (str) -> Tuple[str, Optional[str], AuthInfo] - # hotfix the URL scheme after removing bzr+ from bzr+ssh:// readd it - url, rev, user_pass = super(Bazaar, cls).get_url_rev_and_auth(url) - if url.startswith('ssh://'): - url = 'bzr+' + url + def get_url_rev_and_auth(cls, url: str) -> Tuple[str, Optional[str], AuthInfo]: + # hotfix the URL scheme after removing bzr+ from bzr+ssh:// re-add it + url, rev, user_pass = super().get_url_rev_and_auth(url) + if url.startswith("ssh://"): + url = "bzr+" + url return url, rev, user_pass @classmethod - def get_remote_url(cls, location): - urls = cls.run_command(['info'], show_stdout=False, cwd=location) + def get_remote_url(cls, location: str) -> str: + urls = cls.run_command( + ["info"], show_stdout=False, stdout_only=True, cwd=location + ) for line in urls.splitlines(): line = line.strip() - for x in ('checkout of branch: ', - 'parent branch: '): + for x in ("checkout of branch: ", "parent branch: "): if line.startswith(x): repo = line.split(x)[1] if cls._is_local_repository(repo): return path_to_url(repo) return repo - return None + raise RemoteNotFoundError @classmethod - def get_revision(cls, location): + def get_revision(cls, location: str) -> str: revision = cls.run_command( - ['revno'], show_stdout=False, cwd=location, + ["revno"], + show_stdout=False, + stdout_only=True, + cwd=location, ) return revision.splitlines()[-1] @classmethod - def is_commit_id_equal(cls, dest, name): + def is_commit_id_equal(cls, dest: str, name: Optional[str]) -> bool: """Always assume the versions don't match""" return False diff --git a/src/fetchcode/vcs/pip/_internal/vcs/git.py b/src/fetchcode/vcs/pip/_internal/vcs/git.py index 9f8ce125..0425debb 100644 --- a/src/fetchcode/vcs/pip/_internal/vcs/git.py +++ b/src/fetchcode/vcs/pip/_internal/vcs/git.py @@ -1,66 +1,83 @@ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import - import logging import os.path +import pathlib import re - -from fetchcode.vcs.pip._vendor.packaging.version import parse as parse_version -from fetchcode.vcs.pip._vendor.six.moves.urllib import parse as urllib_parse -from fetchcode.vcs.pip._vendor.six.moves.urllib import request as urllib_request - -from fetchcode.vcs.pip._internal.exceptions import BadCommand, InstallationError -from fetchcode.vcs.pip._internal.utils.misc import display_path, hide_url -from fetchcode.vcs.pip._internal.utils.subprocess import make_command -from fetchcode.vcs.pip._internal.utils.temp_dir import TempDirectory -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.vcs.versioncontrol import ( +import urllib.parse +import urllib.request +from dataclasses import replace +from typing import List, Optional, Tuple + +from pip._internal.exceptions import BadCommand, InstallationError +from pip._internal.utils.misc import HiddenText, display_path, hide_url +from pip._internal.utils.subprocess import make_command +from pip._internal.vcs.versioncontrol import ( + AuthInfo, RemoteNotFoundError, + RemoteNotValidError, + RevOptions, VersionControl, - find_path_to_setup_from_repo_root, + find_path_to_project_root_from_repo_root, vcs, ) -if MYPY_CHECK_RUNNING: - from typing import Optional, Tuple - from fetchcode.vcs.pip._internal.utils.misc import HiddenText - from fetchcode.vcs.pip._internal.vcs.versioncontrol import AuthInfo, RevOptions - - -urlsplit = urllib_parse.urlsplit -urlunsplit = urllib_parse.urlunsplit +urlsplit = urllib.parse.urlsplit +urlunsplit = urllib.parse.urlunsplit logger = logging.getLogger(__name__) -HASH_REGEX = re.compile('^[a-fA-F0-9]{40}$') +GIT_VERSION_REGEX = re.compile( + r"^git version " # Prefix. + r"(\d+)" # Major. + r"\.(\d+)" # Dot, minor. + r"(?:\.(\d+))?" # Optional dot, patch. + r".*$" # Suffix, including any pre- and post-release segments we don't care about. +) + +HASH_REGEX = re.compile("^[a-fA-F0-9]{40}$") + +# SCP (Secure copy protocol) shorthand. e.g. 'git@example.com:foo/bar.git' +SCP_REGEX = re.compile( + r"""^ + # Optional user, e.g. 'git@' + (\w+@)? + # Server, e.g. 'github.com'. + ([^/:]+): + # The server-side path. e.g. 'user/project.git'. Must start with an + # alphanumeric character so as not to be confusable with a Windows paths + # like 'C:/foo/bar' or 'C:\foo\bar'. + (\w[^:]*) + $""", + re.VERBOSE, +) -def looks_like_hash(sha): +def looks_like_hash(sha: str) -> bool: return bool(HASH_REGEX.match(sha)) class Git(VersionControl): - name = 'git' - dirname = '.git' - repo_name = 'clone' + name = "git" + dirname = ".git" + repo_name = "clone" schemes = ( - 'git', 'git+http', 'git+https', 'git+ssh', 'git+git', 'git+file', + "git+http", + "git+https", + "git+ssh", + "git+git", + "git+file", ) # Prevent the user's environment variables from interfering with pip: # https://github.com/pypa/pip/issues/1130 - unset_environ = ('GIT_DIR', 'GIT_WORK_TREE') - default_arg_rev = 'HEAD' + unset_environ = ("GIT_DIR", "GIT_WORK_TREE") + default_arg_rev = "HEAD" @staticmethod - def get_base_rev_args(rev): + def get_base_rev_args(rev: str) -> List[str]: return [rev] - def is_immutable_rev_checkout(self, url, dest): - # type: (str, str) -> bool + def is_immutable_rev_checkout(self, url: str, dest: str) -> bool: _, rev_options = self.get_url_rev_options(hide_url(url)) if not rev_options.rev: return False @@ -71,26 +88,24 @@ def is_immutable_rev_checkout(self, url, dest): # return False in the rare case rev is both a commit hash # and a tag or a branch; we don't want to cache in that case # because that branch/tag could point to something else in the future - is_tag_or_branch = bool( - self.get_revision_sha(dest, rev_options.rev)[0] - ) + is_tag_or_branch = bool(self.get_revision_sha(dest, rev_options.rev)[0]) return not is_tag_or_branch - def get_git_version(self): - VERSION_PFX = 'git version ' - version = self.run_command(['version'], show_stdout=False) - if version.startswith(VERSION_PFX): - version = version[len(VERSION_PFX):].split()[0] - else: - version = '' - # get first 3 positions of the git version because - # on windows it is x.y.z.windows.t, and this parses as - # LegacyVersion which always smaller than a Version. - version = '.'.join(version.split('.')[:3]) - return parse_version(version) + def get_git_version(self) -> Tuple[int, ...]: + version = self.run_command( + ["version"], + command_desc="git version", + show_stdout=False, + stdout_only=True, + ) + match = GIT_VERSION_REGEX.match(version) + if not match: + logger.warning("Can't parse git version: %s", version) + return () + return (int(match.group(1)), int(match.group(2))) @classmethod - def get_current_branch(cls, location): + def get_current_branch(cls, location: str) -> Optional[str]: """ Return the current branch, or None if HEAD isn't at a branch (e.g. detached HEAD). @@ -99,32 +114,23 @@ def get_current_branch(cls, location): # HEAD rather than a symbolic ref. In addition, the -q causes the # command to exit with status code 1 instead of 128 in this case # and to suppress the message to stderr. - args = ['symbolic-ref', '-q', 'HEAD'] + args = ["symbolic-ref", "-q", "HEAD"] output = cls.run_command( - args, extra_ok_returncodes=(1, ), show_stdout=False, cwd=location, + args, + extra_ok_returncodes=(1,), + show_stdout=False, + stdout_only=True, + cwd=location, ) ref = output.strip() - if ref.startswith('refs/heads/'): - return ref[len('refs/heads/'):] + if ref.startswith("refs/heads/"): + return ref[len("refs/heads/") :] return None - def export(self, location, url): - # type: (str, HiddenText) -> None - """Export the Git repository at the url to the destination location""" - if not location.endswith('/'): - location = location + '/' - - with TempDirectory(kind="export") as temp_dir: - self.unpack(temp_dir.path, url=url) - self.run_command( - ['checkout-index', '-a', '-f', '--prefix', location], - show_stdout=False, cwd=temp_dir.path - ) - @classmethod - def get_revision_sha(cls, dest, rev): + def get_revision_sha(cls, dest: str, rev: str) -> Tuple[Optional[str], bool]: """ Return (sha_or_none, is_branch), where sha_or_none is a commit hash if the revision names a remote branch or tag, otherwise None. @@ -134,21 +140,32 @@ def get_revision_sha(cls, dest, rev): rev: the revision name. """ # Pass rev to pre-filter the list. - output = cls.run_command(['show-ref', rev], cwd=dest, - show_stdout=False, on_returncode='ignore') + output = cls.run_command( + ["show-ref", rev], + cwd=dest, + show_stdout=False, + stdout_only=True, + on_returncode="ignore", + ) refs = {} - for line in output.strip().splitlines(): + # NOTE: We do not use splitlines here since that would split on other + # unicode separators, which can be maliciously used to install a + # different revision. + for line in output.strip().split("\n"): + line = line.rstrip("\r") + if not line: + continue try: - sha, ref = line.split() + ref_sha, ref_name = line.split(" ", maxsplit=2) except ValueError: # Include the offending line to simplify troubleshooting if # this error ever occurs. - raise ValueError('unexpected show-ref line: {!r}'.format(line)) + raise ValueError(f"unexpected show-ref line: {line!r}") - refs[ref] = sha + refs[ref_name] = ref_sha - branch_ref = 'refs/remotes/origin/{}'.format(rev) - tag_ref = 'refs/tags/{}'.format(rev) + branch_ref = f"refs/remotes/origin/{rev}" + tag_ref = f"refs/tags/{rev}" sha = refs.get(branch_ref) if sha is not None: @@ -159,8 +176,32 @@ def get_revision_sha(cls, dest, rev): return (sha, False) @classmethod - def resolve_revision(cls, dest, url, rev_options): - # type: (str, HiddenText, RevOptions) -> RevOptions + def _should_fetch(cls, dest: str, rev: str) -> bool: + """ + Return true if rev is a ref or is a commit that we don't have locally. + + Branches and tags are not considered in this method because they are + assumed to be always available locally (which is a normal outcome of + ``git clone`` and ``git fetch --tags``). + """ + if rev.startswith("refs/"): + # Always fetch remote refs. + return True + + if not looks_like_hash(rev): + # Git fetch would fail with abbreviated commits. + return False + + if cls.has_commit(dest, rev): + # Don't fetch if we have the commit locally. + return False + + return True + + @classmethod + def resolve_revision( + cls, dest: str, url: HiddenText, rev_options: RevOptions + ) -> RevOptions: """ Resolve a revision to a new RevOptions object with the SHA1 of the branch, tag, or ref if found. @@ -177,7 +218,7 @@ def resolve_revision(cls, dest, url, rev_options): if sha is not None: rev_options = rev_options.make_new(sha) - rev_options.branch_name = rev if is_branch else None + rev_options = replace(rev_options, branch_name=(rev if is_branch else None)) return rev_options @@ -189,22 +230,22 @@ def resolve_revision(cls, dest, url, rev_options): rev, ) - if not rev.startswith('refs/'): + if not cls._should_fetch(dest, rev): return rev_options - # If it looks like a ref, we have to fetch it explicitly. + # fetch the requested revision cls.run_command( - make_command('fetch', '-q', url, rev_options.to_args()), + make_command("fetch", "-q", url, rev_options.to_args()), cwd=dest, ) # Change the revision to the SHA of the ref we fetched - sha = cls.get_revision(dest, rev='FETCH_HEAD') + sha = cls.get_revision(dest, rev="FETCH_HEAD") rev_options = rev_options.make_new(sha) return rev_options @classmethod - def is_commit_id_equal(cls, dest, name): + def is_commit_id_equal(cls, dest: str, name: Optional[str]) -> bool: """ Return whether the current commit hash equals the given name. @@ -218,64 +259,95 @@ def is_commit_id_equal(cls, dest, name): return cls.get_revision(dest) == name - def fetch_new(self, dest, url, rev_options): - # type: (str, HiddenText, RevOptions) -> None + def fetch_new( + self, dest: str, url: HiddenText, rev_options: RevOptions, verbosity: int + ) -> None: rev_display = rev_options.to_display() - logger.info('Cloning %s%s to %s', url, rev_display, display_path(dest)) - self.run_command(make_command('clone', '-q', url, dest)) + logger.info("Cloning %s%s to %s", url, rev_display, display_path(dest)) + if verbosity <= 0: + flags: Tuple[str, ...] = ("--quiet",) + elif verbosity == 1: + flags = () + else: + flags = ("--verbose", "--progress") + if self.get_git_version() >= (2, 17): + # Git added support for partial clone in 2.17 + # https://git-scm.com/docs/partial-clone + # Speeds up cloning by functioning without a complete copy of repository + self.run_command( + make_command( + "clone", + "--filter=blob:none", + *flags, + url, + dest, + ) + ) + else: + self.run_command(make_command("clone", *flags, url, dest)) if rev_options.rev: # Then a specific revision was requested. rev_options = self.resolve_revision(dest, url, rev_options) - branch_name = getattr(rev_options, 'branch_name', None) + branch_name = getattr(rev_options, "branch_name", None) + logger.debug("Rev options %s, branch_name %s", rev_options, branch_name) if branch_name is None: # Only do a checkout if the current commit id doesn't match # the requested revision. if not self.is_commit_id_equal(dest, rev_options.rev): cmd_args = make_command( - 'checkout', '-q', rev_options.to_args(), + "checkout", + "-q", + rev_options.to_args(), ) self.run_command(cmd_args, cwd=dest) elif self.get_current_branch(dest) != branch_name: # Then a specific branch was requested, and that branch # is not yet checked out. - track_branch = 'origin/{}'.format(branch_name) + track_branch = f"origin/{branch_name}" cmd_args = [ - 'checkout', '-b', branch_name, '--track', track_branch, + "checkout", + "-b", + branch_name, + "--track", + track_branch, ] self.run_command(cmd_args, cwd=dest) + else: + sha = self.get_revision(dest) + rev_options = rev_options.make_new(sha) + + logger.info("Resolved %s to commit %s", url, rev_options.rev) #: repo may contain submodules self.update_submodules(dest) - def switch(self, dest, url, rev_options): - # type: (str, HiddenText, RevOptions) -> None + def switch(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None: self.run_command( - make_command('config', 'remote.origin.url', url), + make_command("config", "remote.origin.url", url), cwd=dest, ) - cmd_args = make_command('checkout', '-q', rev_options.to_args()) + cmd_args = make_command("checkout", "-q", rev_options.to_args()) self.run_command(cmd_args, cwd=dest) self.update_submodules(dest) - def update(self, dest, url, rev_options): - # type: (str, HiddenText, RevOptions) -> None + def update(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None: # First fetch changes from the default remote - if self.get_git_version() >= parse_version('1.9.0'): + if self.get_git_version() >= (1, 9): # fetch tags in addition to everything else - self.run_command(['fetch', '-q', '--tags'], cwd=dest) + self.run_command(["fetch", "-q", "--tags"], cwd=dest) else: - self.run_command(['fetch', '-q'], cwd=dest) + self.run_command(["fetch", "-q"], cwd=dest) # Then reset to wanted revision (maybe even origin/master) rev_options = self.resolve_revision(dest, url, rev_options) - cmd_args = make_command('reset', '--hard', '-q', rev_options.to_args()) + cmd_args = make_command("reset", "--hard", "-q", rev_options.to_args()) self.run_command(cmd_args, cwd=dest) #: update submodules self.update_submodules(dest) @classmethod - def get_remote_url(cls, location): + def get_remote_url(cls, location: str) -> str: """ Return URL of the first remote encountered. @@ -285,8 +357,11 @@ def get_remote_url(cls, location): # We need to pass 1 for extra_ok_returncodes since the command # exits with return code 1 if there are no matching lines. stdout = cls.run_command( - ['config', '--get-regexp', r'remote\..*\.url'], - extra_ok_returncodes=(1, ), show_stdout=False, cwd=location, + ["config", "--get-regexp", r"remote\..*\.url"], + extra_ok_returncodes=(1,), + show_stdout=False, + stdout_only=True, + cwd=location, ) remotes = stdout.splitlines() try: @@ -295,39 +370,91 @@ def get_remote_url(cls, location): raise RemoteNotFoundError for remote in remotes: - if remote.startswith('remote.origin.url '): + if remote.startswith("remote.origin.url "): found_remote = remote break - url = found_remote.split(' ')[1] - return url.strip() + url = found_remote.split(" ")[1] + return cls._git_remote_to_pip_url(url.strip()) + + @staticmethod + def _git_remote_to_pip_url(url: str) -> str: + """ + Convert a remote url from what git uses to what pip accepts. + + There are 3 legal forms **url** may take: + + 1. A fully qualified url: ssh://git@example.com/foo/bar.git + 2. A local project.git folder: /path/to/bare/repository.git + 3. SCP shorthand for form 1: git@example.com:foo/bar.git + + Form 1 is output as-is. Form 2 must be converted to URI and form 3 must + be converted to form 1. + + See the corresponding test test_git_remote_url_to_pip() for examples of + sample inputs/outputs. + """ + if re.match(r"\w+://", url): + # This is already valid. Pass it though as-is. + return url + if os.path.exists(url): + # A local bare remote (git clone --mirror). + # Needs a file:// prefix. + return pathlib.PurePath(url).as_uri() + scp_match = SCP_REGEX.match(url) + if scp_match: + # Add an ssh:// prefix and replace the ':' with a '/'. + return scp_match.expand(r"ssh://\1\2/\3") + # Otherwise, bail out. + raise RemoteNotValidError(url) + + @classmethod + def has_commit(cls, location: str, rev: str) -> bool: + """ + Check if rev is a commit that is available in the local repository. + """ + try: + cls.run_command( + ["rev-parse", "-q", "--verify", "sha^" + rev], + cwd=location, + log_failed_cmd=False, + ) + except InstallationError: + return False + else: + return True @classmethod - def get_revision(cls, location, rev=None): + def get_revision(cls, location: str, rev: Optional[str] = None) -> str: if rev is None: - rev = 'HEAD' + rev = "HEAD" current_rev = cls.run_command( - ['rev-parse', rev], show_stdout=False, cwd=location, + ["rev-parse", rev], + show_stdout=False, + stdout_only=True, + cwd=location, ) return current_rev.strip() @classmethod - def get_subdirectory(cls, location): + def get_subdirectory(cls, location: str) -> Optional[str]: """ - Return the path to setup.py, relative to the repo root. - Return None if setup.py is in the repo root. + Return the path to Python project root, relative to the repo root. + Return None if the project root is in the repo root. """ # find the repo root git_dir = cls.run_command( - ['rev-parse', '--git-dir'], - show_stdout=False, cwd=location).strip() + ["rev-parse", "--git-dir"], + show_stdout=False, + stdout_only=True, + cwd=location, + ).strip() if not os.path.isabs(git_dir): git_dir = os.path.join(location, git_dir) - repo_root = os.path.abspath(os.path.join(git_dir, '..')) - return find_path_to_setup_from_repo_root(location, repo_root) + repo_root = os.path.abspath(os.path.join(git_dir, "..")) + return find_path_to_project_root_from_repo_root(location, repo_root) @classmethod - def get_url_rev_and_auth(cls, url): - # type: (str) -> Tuple[str, Optional[str], AuthInfo] + def get_url_rev_and_auth(cls, url: str) -> Tuple[str, Optional[str], AuthInfo]: """ Prefixes stub URLs like 'user@hostname:user/repo.git' with 'ssh://'. That's required because although they use SSH they sometimes don't @@ -337,58 +464,64 @@ def get_url_rev_and_auth(cls, url): # Works around an apparent Git bug # (see https://article.gmane.org/gmane.comp.version-control.git/146500) scheme, netloc, path, query, fragment = urlsplit(url) - if scheme.endswith('file'): - initial_slashes = path[:-len(path.lstrip('/'))] - newpath = ( - initial_slashes + - urllib_request.url2pathname(path) - .replace('\\', '/').lstrip('/') - ) - url = urlunsplit((scheme, netloc, newpath, query, fragment)) - after_plus = scheme.find('+') + 1 + if scheme.endswith("file"): + initial_slashes = path[: -len(path.lstrip("/"))] + newpath = initial_slashes + urllib.request.url2pathname(path).replace( + "\\", "/" + ).lstrip("/") + after_plus = scheme.find("+") + 1 url = scheme[:after_plus] + urlunsplit( (scheme[after_plus:], netloc, newpath, query, fragment), ) - if '://' not in url: - assert 'file:' not in url - url = url.replace('git+', 'git+ssh://') - url, rev, user_pass = super(Git, cls).get_url_rev_and_auth(url) - url = url.replace('ssh://', '') + if "://" not in url: + assert "file:" not in url + url = url.replace("git+", "git+ssh://") + url, rev, user_pass = super().get_url_rev_and_auth(url) + url = url.replace("ssh://", "") else: - url, rev, user_pass = super(Git, cls).get_url_rev_and_auth(url) + url, rev, user_pass = super().get_url_rev_and_auth(url) return url, rev, user_pass @classmethod - def update_submodules(cls, location): - if not os.path.exists(os.path.join(location, '.gitmodules')): + def update_submodules(cls, location: str) -> None: + if not os.path.exists(os.path.join(location, ".gitmodules")): return cls.run_command( - ['submodule', 'update', '--init', '--recursive', '-q'], + ["submodule", "update", "--init", "--recursive", "-q"], cwd=location, ) @classmethod - def get_repository_root(cls, location): - loc = super(Git, cls).get_repository_root(location) + def get_repository_root(cls, location: str) -> Optional[str]: + loc = super().get_repository_root(location) if loc: return loc try: r = cls.run_command( - ['rev-parse', '--show-toplevel'], + ["rev-parse", "--show-toplevel"], cwd=location, show_stdout=False, - on_returncode='raise', + stdout_only=True, + on_returncode="raise", log_failed_cmd=False, ) except BadCommand: - logger.debug("could not determine if %s is under git control " - "because git is not available", location) + logger.debug( + "could not determine if %s is under git control " + "because git is not available", + location, + ) return None except InstallationError: return None - return os.path.normpath(r.rstrip('\r\n')) + return os.path.normpath(r.rstrip("\r\n")) + + @staticmethod + def should_add_vcs_url_prefix(repo_url: str) -> bool: + """In either https or ssh form, requirements must be prefixed with git+.""" + return True vcs.register(Git) diff --git a/src/fetchcode/vcs/pip/_internal/vcs/mercurial.py b/src/fetchcode/vcs/pip/_internal/vcs/mercurial.py index e41f2155..c183d41d 100644 --- a/src/fetchcode/vcs/pip/_internal/vcs/mercurial.py +++ b/src/fetchcode/vcs/pip/_internal/vcs/mercurial.py @@ -1,161 +1,163 @@ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import - +import configparser import logging import os - -from fetchcode.vcs.pip._vendor.six.moves import configparser - -from fetchcode.vcs.pip._internal.exceptions import BadCommand, InstallationError -from fetchcode.vcs.pip._internal.utils.misc import display_path -from fetchcode.vcs.pip._internal.utils.subprocess import make_command -from fetchcode.vcs.pip._internal.utils.temp_dir import TempDirectory -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.utils.urls import path_to_url -from fetchcode.vcs.pip._internal.vcs.versioncontrol import ( +from typing import List, Optional, Tuple + +from pip._internal.exceptions import BadCommand, InstallationError +from pip._internal.utils.misc import HiddenText, display_path +from pip._internal.utils.subprocess import make_command +from pip._internal.utils.urls import path_to_url +from pip._internal.vcs.versioncontrol import ( + RevOptions, VersionControl, - find_path_to_setup_from_repo_root, + find_path_to_project_root_from_repo_root, vcs, ) -if MYPY_CHECK_RUNNING: - from fetchcode.vcs.pip._internal.utils.misc import HiddenText - from fetchcode.vcs.pip._internal.vcs.versioncontrol import RevOptions - - logger = logging.getLogger(__name__) class Mercurial(VersionControl): - name = 'hg' - dirname = '.hg' - repo_name = 'clone' + name = "hg" + dirname = ".hg" + repo_name = "clone" schemes = ( - 'hg', 'hg+file', 'hg+http', 'hg+https', 'hg+ssh', 'hg+static-http', + "hg+file", + "hg+http", + "hg+https", + "hg+ssh", + "hg+static-http", ) @staticmethod - def get_base_rev_args(rev): - return [rev] + def get_base_rev_args(rev: str) -> List[str]: + return [f"--rev={rev}"] - def export(self, location, url): - # type: (str, HiddenText) -> None - """Export the Hg repository at the url to the destination location""" - with TempDirectory(kind="export") as temp_dir: - self.unpack(temp_dir.path, url=url) - - self.run_command( - ['archive', location], show_stdout=False, cwd=temp_dir.path - ) - - def fetch_new(self, dest, url, rev_options): - # type: (str, HiddenText, RevOptions) -> None + def fetch_new( + self, dest: str, url: HiddenText, rev_options: RevOptions, verbosity: int + ) -> None: rev_display = rev_options.to_display() logger.info( - 'Cloning hg %s%s to %s', + "Cloning hg %s%s to %s", url, rev_display, display_path(dest), ) - self.run_command(make_command('clone', '--noupdate', '-q', url, dest)) + if verbosity <= 0: + flags: Tuple[str, ...] = ("--quiet",) + elif verbosity == 1: + flags = () + elif verbosity == 2: + flags = ("--verbose",) + else: + flags = ("--verbose", "--debug") + self.run_command(make_command("clone", "--noupdate", *flags, url, dest)) self.run_command( - make_command('update', '-q', rev_options.to_args()), + make_command("update", *flags, rev_options.to_args()), cwd=dest, ) - def switch(self, dest, url, rev_options): - # type: (str, HiddenText, RevOptions) -> None - repo_config = os.path.join(dest, self.dirname, 'hgrc') + def switch(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None: + repo_config = os.path.join(dest, self.dirname, "hgrc") config = configparser.RawConfigParser() try: config.read(repo_config) - config.set('paths', 'default', url.secret) - with open(repo_config, 'w') as config_file: + config.set("paths", "default", url.secret) + with open(repo_config, "w") as config_file: config.write(config_file) except (OSError, configparser.NoSectionError) as exc: - logger.warning( - 'Could not switch Mercurial repository to %s: %s', url, exc, - ) + logger.warning("Could not switch Mercurial repository to %s: %s", url, exc) else: - cmd_args = make_command('update', '-q', rev_options.to_args()) + cmd_args = make_command("update", "-q", rev_options.to_args()) self.run_command(cmd_args, cwd=dest) - def update(self, dest, url, rev_options): - # type: (str, HiddenText, RevOptions) -> None - self.run_command(['pull', '-q'], cwd=dest) - cmd_args = make_command('update', '-q', rev_options.to_args()) + def update(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None: + self.run_command(["pull", "-q"], cwd=dest) + cmd_args = make_command("update", "-q", rev_options.to_args()) self.run_command(cmd_args, cwd=dest) @classmethod - def get_remote_url(cls, location): + def get_remote_url(cls, location: str) -> str: url = cls.run_command( - ['showconfig', 'paths.default'], - show_stdout=False, cwd=location).strip() + ["showconfig", "paths.default"], + show_stdout=False, + stdout_only=True, + cwd=location, + ).strip() if cls._is_local_repository(url): url = path_to_url(url) return url.strip() @classmethod - def get_revision(cls, location): + def get_revision(cls, location: str) -> str: """ Return the repository-local changeset revision number, as an integer. """ current_revision = cls.run_command( - ['parents', '--template={rev}'], - show_stdout=False, cwd=location).strip() + ["parents", "--template={rev}"], + show_stdout=False, + stdout_only=True, + cwd=location, + ).strip() return current_revision @classmethod - def get_requirement_revision(cls, location): + def get_requirement_revision(cls, location: str) -> str: """ Return the changeset identification hash, as a 40-character hexadecimal string """ current_rev_hash = cls.run_command( - ['parents', '--template={node}'], - show_stdout=False, cwd=location).strip() + ["parents", "--template={node}"], + show_stdout=False, + stdout_only=True, + cwd=location, + ).strip() return current_rev_hash @classmethod - def is_commit_id_equal(cls, dest, name): + def is_commit_id_equal(cls, dest: str, name: Optional[str]) -> bool: """Always assume the versions don't match""" return False @classmethod - def get_subdirectory(cls, location): + def get_subdirectory(cls, location: str) -> Optional[str]: """ - Return the path to setup.py, relative to the repo root. - Return None if setup.py is in the repo root. + Return the path to Python project root, relative to the repo root. + Return None if the project root is in the repo root. """ # find the repo root repo_root = cls.run_command( - ['root'], show_stdout=False, cwd=location).strip() + ["root"], show_stdout=False, stdout_only=True, cwd=location + ).strip() if not os.path.isabs(repo_root): repo_root = os.path.abspath(os.path.join(location, repo_root)) - return find_path_to_setup_from_repo_root(location, repo_root) + return find_path_to_project_root_from_repo_root(location, repo_root) @classmethod - def get_repository_root(cls, location): - loc = super(Mercurial, cls).get_repository_root(location) + def get_repository_root(cls, location: str) -> Optional[str]: + loc = super().get_repository_root(location) if loc: return loc try: r = cls.run_command( - ['root'], + ["root"], cwd=location, show_stdout=False, - on_returncode='raise', + stdout_only=True, + on_returncode="raise", log_failed_cmd=False, ) except BadCommand: - logger.debug("could not determine if %s is under hg control " - "because hg is not available", location) + logger.debug( + "could not determine if %s is under hg control " + "because hg is not available", + location, + ) return None except InstallationError: return None - return os.path.normpath(r.rstrip('\r\n')) + return os.path.normpath(r.rstrip("\r\n")) vcs.register(Mercurial) diff --git a/src/fetchcode/vcs/pip/_internal/vcs/subversion.py b/src/fetchcode/vcs/pip/_internal/vcs/subversion.py index e6eabbd0..f359266d 100644 --- a/src/fetchcode/vcs/pip/_internal/vcs/subversion.py +++ b/src/fetchcode/vcs/pip/_internal/vcs/subversion.py @@ -1,67 +1,60 @@ -# The following comment should be removed at some point in the future. -# mypy: disallow-untyped-defs=False - -from __future__ import absolute_import - import logging import os import re +from typing import List, Optional, Tuple -from fetchcode.vcs.pip._internal.utils.logging import indent_log -from fetchcode.vcs.pip._internal.utils.misc import ( +from pip._internal.utils.misc import ( + HiddenText, display_path, is_console_interactive, - rmtree, + is_installable_dir, split_auth_from_netloc, ) -from fetchcode.vcs.pip._internal.utils.subprocess import make_command -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.vcs.versioncontrol import VersionControl, vcs +from pip._internal.utils.subprocess import CommandArgs, make_command +from pip._internal.vcs.versioncontrol import ( + AuthInfo, + RemoteNotFoundError, + RevOptions, + VersionControl, + vcs, +) + +logger = logging.getLogger(__name__) _svn_xml_url_re = re.compile('url="([^"]+)"') _svn_rev_re = re.compile(r'committed-rev="(\d+)"') _svn_info_xml_rev_re = re.compile(r'\s*revision="(\d+)"') -_svn_info_xml_url_re = re.compile(r'(.*)') - - -if MYPY_CHECK_RUNNING: - from typing import Optional, Tuple - from fetchcode.vcs.pip._internal.utils.subprocess import CommandArgs - from fetchcode.vcs.pip._internal.utils.misc import HiddenText - from fetchcode.vcs.pip._internal.vcs.versioncontrol import AuthInfo, RevOptions - - -logger = logging.getLogger(__name__) +_svn_info_xml_url_re = re.compile(r"(.*)") class Subversion(VersionControl): - name = 'svn' - dirname = '.svn' - repo_name = 'checkout' - schemes = ('svn', 'svn+ssh', 'svn+http', 'svn+https', 'svn+svn') + name = "svn" + dirname = ".svn" + repo_name = "checkout" + schemes = ("svn+ssh", "svn+http", "svn+https", "svn+svn", "svn+file") @classmethod - def should_add_vcs_url_prefix(cls, remote_url): + def should_add_vcs_url_prefix(cls, remote_url: str) -> bool: return True @staticmethod - def get_base_rev_args(rev): - return ['-r', rev] + def get_base_rev_args(rev: str) -> List[str]: + return ["-r", rev] @classmethod - def get_revision(cls, location): + def get_revision(cls, location: str) -> str: """ Return the maximum revision for all files under a given location """ # Note: taken from setuptools.command.egg_info revision = 0 - for base, dirs, files in os.walk(location): + for base, dirs, _ in os.walk(location): if cls.dirname not in dirs: dirs[:] = [] - continue # no sense walking uncontrolled subdirs + continue # no sense walking uncontrolled subdirs dirs.remove(cls.dirname) - entries_fn = os.path.join(base, cls.dirname, 'entries') + entries_fn = os.path.join(base, cls.dirname, "entries") if not os.path.exists(entries_fn): # FIXME: should we warn? continue @@ -69,91 +62,95 @@ def get_revision(cls, location): dirurl, localrev = cls._get_svn_url_rev(base) if base == location: - base = dirurl + '/' # save the root url + assert dirurl is not None + base = dirurl + "/" # save the root url elif not dirurl or not dirurl.startswith(base): dirs[:] = [] - continue # not part of the same svn tree, skip it + continue # not part of the same svn tree, skip it revision = max(revision, localrev) - return revision + return str(revision) @classmethod - def get_netloc_and_auth(cls, netloc, scheme): + def get_netloc_and_auth( + cls, netloc: str, scheme: str + ) -> Tuple[str, Tuple[Optional[str], Optional[str]]]: """ This override allows the auth information to be passed to svn via the --username and --password options instead of via the URL. """ - if scheme == 'ssh': + if scheme == "ssh": # The --username and --password options can't be used for # svn+ssh URLs, so keep the auth information in the URL. - return super(Subversion, cls).get_netloc_and_auth(netloc, scheme) + return super().get_netloc_and_auth(netloc, scheme) return split_auth_from_netloc(netloc) @classmethod - def get_url_rev_and_auth(cls, url): - # type: (str) -> Tuple[str, Optional[str], AuthInfo] - # hotfix the URL scheme after removing svn+ from svn+ssh:// readd it - url, rev, user_pass = super(Subversion, cls).get_url_rev_and_auth(url) - if url.startswith('ssh://'): - url = 'svn+' + url + def get_url_rev_and_auth(cls, url: str) -> Tuple[str, Optional[str], AuthInfo]: + # hotfix the URL scheme after removing svn+ from svn+ssh:// re-add it + url, rev, user_pass = super().get_url_rev_and_auth(url) + if url.startswith("ssh://"): + url = "svn+" + url return url, rev, user_pass @staticmethod - def make_rev_args(username, password): - # type: (Optional[str], Optional[HiddenText]) -> CommandArgs - extra_args = [] # type: CommandArgs + def make_rev_args( + username: Optional[str], password: Optional[HiddenText] + ) -> CommandArgs: + extra_args: CommandArgs = [] if username: - extra_args += ['--username', username] + extra_args += ["--username", username] if password: - extra_args += ['--password', password] + extra_args += ["--password", password] return extra_args @classmethod - def get_remote_url(cls, location): - # In cases where the source is in a subdirectory, not alongside - # setup.py we have to look up in the location until we find a real - # setup.py + def get_remote_url(cls, location: str) -> str: + # In cases where the source is in a subdirectory, we have to look up in + # the location until we find a valid project root. orig_location = location - while not os.path.exists(os.path.join(location, 'setup.py')): + while not is_installable_dir(location): last_location = location location = os.path.dirname(location) if location == last_location: # We've traversed up to the root of the filesystem without - # finding setup.py + # finding a Python project. logger.warning( - "Could not find setup.py for directory %s (tried all " + "Could not find Python project for directory %s (tried all " "parent directories)", orig_location, ) - return None + raise RemoteNotFoundError - return cls._get_svn_url_rev(location)[0] + url, _rev = cls._get_svn_url_rev(location) + if url is None: + raise RemoteNotFoundError + + return url @classmethod - def _get_svn_url_rev(cls, location): - from fetchcode.vcs.pip._internal.exceptions import InstallationError + def _get_svn_url_rev(cls, location: str) -> Tuple[Optional[str], int]: + from pip._internal.exceptions import InstallationError - entries_path = os.path.join(location, cls.dirname, 'entries') + entries_path = os.path.join(location, cls.dirname, "entries") if os.path.exists(entries_path): with open(entries_path) as f: data = f.read() else: # subversion >= 1.7 does not have the 'entries' file - data = '' - - if (data.startswith('8') or - data.startswith('9') or - data.startswith('10')): - data = list(map(str.splitlines, data.split('\n\x0c\n'))) - del data[0][0] # get rid of the '8' - url = data[0][3] - revs = [int(d[9]) for d in data if len(d) > 9 and d[9]] + [0] - elif data.startswith(' 9 and d[9]] + [0] + elif data.startswith(" bool: """Always assume the versions don't match""" return False - def __init__(self, use_interactive=None): - # type: (bool) -> None + def __init__(self, use_interactive: Optional[bool] = None) -> None: if use_interactive is None: use_interactive = is_console_interactive() self.use_interactive = use_interactive @@ -197,12 +194,11 @@ def __init__(self, use_interactive=None): # Special value definitions: # None: Not evaluated yet. # Empty tuple: Could not parse version. - self._vcs_version = None # type: Optional[Tuple[int, ...]] + self._vcs_version: Optional[Tuple[int, ...]] = None - super(Subversion, self).__init__() + super().__init__() - def call_vcs_version(self): - # type: () -> Tuple[int, ...] + def call_vcs_version(self) -> Tuple[int, ...]: """Query the version of the currently installed Subversion client. :return: A tuple containing the parts of the version information or @@ -214,13 +210,15 @@ def call_vcs_version(self): # compiled Feb 25 2019, 14:20:39 on x86_64-apple-darwin17.0.0 # svn, version 1.7.14 (r1542130) # compiled Mar 28 2018, 08:49:13 on x86_64-pc-linux-gnu - version_prefix = 'svn, version ' - version = self.run_command(['--version'], show_stdout=False) + # svn, version 1.12.0-SlikSvn (SlikSvn/1.12.0) + # compiled May 28 2019, 13:44:56 on x86_64-microsoft-windows6.2 + version_prefix = "svn, version " + version = self.run_command(["--version"], show_stdout=False, stdout_only=True) if not version.startswith(version_prefix): return () - version = version[len(version_prefix):].split()[0] - version_list = version.split('.') + version = version[len(version_prefix) :].split()[0] + version_list = version.partition("-")[0].split(".") try: parsed_version = tuple(map(int, version_list)) except ValueError: @@ -228,8 +226,7 @@ def call_vcs_version(self): return parsed_version - def get_vcs_version(self): - # type: () -> Tuple[int, ...] + def get_vcs_version(self) -> Tuple[int, ...]: """Return the version of the currently installed Subversion client. If the version of the Subversion client has already been queried, @@ -249,15 +246,13 @@ def get_vcs_version(self): self._vcs_version = vcs_version return vcs_version - def get_remote_call_options(self): - # type: () -> CommandArgs + def get_remote_call_options(self) -> CommandArgs: """Return options to be used on calls to Subversion that contact the server. These options are applicable for the following ``svn`` subcommands used in this class. - checkout - - export - switch - update @@ -266,7 +261,7 @@ def get_remote_call_options(self): if not self.use_interactive: # --non-interactive switch is available since Subversion 0.14.4. # Subversion < 1.8 runs in interactive mode by default. - return ['--non-interactive'] + return ["--non-interactive"] svn_version = self.get_vcs_version() # By default, Subversion >= 1.8 runs in non-interactive mode if @@ -278,54 +273,49 @@ def get_remote_call_options(self): # SVN 1.7, pip should continue to support SVN 1.7. Therefore, pip # can't safely add the option if the SVN version is < 1.8 (or unknown). if svn_version >= (1, 8): - return ['--force-interactive'] + return ["--force-interactive"] return [] - def export(self, location, url): - # type: (str, HiddenText) -> None - """Export the svn repository at the url to the destination location""" - url, rev_options = self.get_url_rev_options(url) - - logger.info('Exporting svn repository %s to %s', url, location) - with indent_log(): - if os.path.exists(location): - # Subversion doesn't like to check out over an existing - # directory --force fixes this, but was only added in svn 1.5 - rmtree(location) - cmd_args = make_command( - 'export', self.get_remote_call_options(), - rev_options.to_args(), url, location, - ) - self.run_command(cmd_args, show_stdout=False) - - def fetch_new(self, dest, url, rev_options): - # type: (str, HiddenText, RevOptions) -> None + def fetch_new( + self, dest: str, url: HiddenText, rev_options: RevOptions, verbosity: int + ) -> None: rev_display = rev_options.to_display() logger.info( - 'Checking out %s%s to %s', + "Checking out %s%s to %s", url, rev_display, display_path(dest), ) + if verbosity <= 0: + flags = ["--quiet"] + else: + flags = [] cmd_args = make_command( - 'checkout', '-q', self.get_remote_call_options(), - rev_options.to_args(), url, dest, + "checkout", + *flags, + self.get_remote_call_options(), + rev_options.to_args(), + url, + dest, ) self.run_command(cmd_args) - def switch(self, dest, url, rev_options): - # type: (str, HiddenText, RevOptions) -> None + def switch(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None: cmd_args = make_command( - 'switch', self.get_remote_call_options(), rev_options.to_args(), - url, dest, + "switch", + self.get_remote_call_options(), + rev_options.to_args(), + url, + dest, ) self.run_command(cmd_args) - def update(self, dest, url, rev_options): - # type: (str, HiddenText, RevOptions) -> None + def update(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None: cmd_args = make_command( - 'update', self.get_remote_call_options(), rev_options.to_args(), + "update", + self.get_remote_call_options(), + rev_options.to_args(), dest, ) self.run_command(cmd_args) diff --git a/src/fetchcode/vcs/pip/_internal/vcs/versioncontrol.py b/src/fetchcode/vcs/pip/_internal/vcs/versioncontrol.py index ee93b49a..a4133165 100644 --- a/src/fetchcode/vcs/pip/_internal/vcs/versioncontrol.py +++ b/src/fetchcode/vcs/pip/_internal/vcs/versioncontrol.py @@ -1,61 +1,65 @@ """Handles all VCS (version control) support""" -from __future__ import absolute_import - -import errno import logging import os import shutil import sys +import urllib.parse +from dataclasses import dataclass, field +from typing import ( + Any, + Dict, + Iterable, + Iterator, + List, + Literal, + Mapping, + Optional, + Tuple, + Type, + Union, +) -from fetchcode.vcs.pip._vendor import pkg_resources -from fetchcode.vcs.pip._vendor.six.moves.urllib import parse as urllib_parse - -from fetchcode.vcs.pip._internal.exceptions import BadCommand, InstallationError -from fetchcode.vcs.pip._internal.utils.compat import samefile -from fetchcode.vcs.pip._internal.utils.misc import ( +from pip._internal.cli.spinners import SpinnerInterface +from pip._internal.exceptions import BadCommand, InstallationError +from pip._internal.utils.misc import ( + HiddenText, ask_path_exists, backup_dir, display_path, hide_url, hide_value, + is_installable_dir, rmtree, ) -from fetchcode.vcs.pip._internal.utils.subprocess import call_subprocess, make_command -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.utils.urls import get_url_scheme - -if MYPY_CHECK_RUNNING: - from typing import ( - Any, Dict, Iterable, Iterator, List, Mapping, Optional, Text, Tuple, - Type, Union - ) - from fetchcode.vcs.pip._internal.cli.spinners import SpinnerInterface - from fetchcode.vcs.pip._internal.utils.misc import HiddenText - from fetchcode.vcs.pip._internal.utils.subprocess import CommandArgs - - AuthInfo = Tuple[Optional[str], Optional[str]] - +from pip._internal.utils.subprocess import ( + CommandArgs, + call_subprocess, + format_command_args, + make_command, +) -__all__ = ['vcs'] +__all__ = ["vcs"] logger = logging.getLogger(__name__) +AuthInfo = Tuple[Optional[str], Optional[str]] -def is_url(name): - # type: (Union[str, Text]) -> bool + +def is_url(name: str) -> bool: """ Return true if the name looks like a URL. """ - scheme = get_url_scheme(name) - if scheme is None: + scheme = urllib.parse.urlsplit(name).scheme + if not scheme: return False - return scheme in ['http', 'https', 'file', 'ftp'] + vcs.all_schemes + return scheme in ["http", "https", "file", "ftp"] + vcs.all_schemes -def make_vcs_requirement_url(repo_url, rev, project_name, subdir=None): - # type: (str, str, str, Optional[str]) -> str +def make_vcs_requirement_url( + repo_url: str, rev: str, project_name: str, subdir: Optional[str] = None +) -> str: """ Return the URL for a VCS requirement. @@ -63,37 +67,38 @@ def make_vcs_requirement_url(repo_url, rev, project_name, subdir=None): repo_url: the remote VCS url, with any needed VCS prefix (e.g. "git+"). project_name: the (unescaped) project name. """ - egg_project_name = pkg_resources.to_filename(project_name) - req = '{}@{}#egg={}'.format(repo_url, rev, egg_project_name) + egg_project_name = project_name.replace("-", "_") + req = f"{repo_url}@{rev}#egg={egg_project_name}" if subdir: - req += '&subdirectory={}'.format(subdir) + req += f"&subdirectory={subdir}" return req -def find_path_to_setup_from_repo_root(location, repo_root): - # type: (str, str) -> Optional[str] +def find_path_to_project_root_from_repo_root( + location: str, repo_root: str +) -> Optional[str]: """ - Find the path to `setup.py` by searching up the filesystem from `location`. - Return the path to `setup.py` relative to `repo_root`. - Return None if `setup.py` is in `repo_root` or cannot be found. + Find the the Python project's root by searching up the filesystem from + `location`. Return the path to project root relative to `repo_root`. + Return None if the project root is `repo_root`, or cannot be found. """ - # find setup.py + # find project root. orig_location = location - while not os.path.exists(os.path.join(location, 'setup.py')): + while not is_installable_dir(location): last_location = location location = os.path.dirname(location) if location == last_location: # We've traversed up to the root of the filesystem without - # finding setup.py + # finding a Python project. logger.warning( - "Could not find setup.py for directory %s (tried all " + "Could not find a Python project for directory %s (tried all " "parent directories)", orig_location, ) return None - if samefile(repo_root, location): + if os.path.samefile(repo_root, location): return None return os.path.relpath(location, repo_root) @@ -103,54 +108,44 @@ class RemoteNotFoundError(Exception): pass -class RevOptions(object): +class RemoteNotValidError(Exception): + def __init__(self, url: str): + super().__init__(url) + self.url = url + +@dataclass(frozen=True) +class RevOptions: """ Encapsulates a VCS-specific revision to install, along with any VCS install options. - Instances of this class should be treated as if immutable. + Args: + vc_class: a VersionControl subclass. + rev: the name of the revision to install. + extra_args: a list of extra options. """ - def __init__( - self, - vc_class, # type: Type[VersionControl] - rev=None, # type: Optional[str] - extra_args=None, # type: Optional[CommandArgs] - ): - # type: (...) -> None - """ - Args: - vc_class: a VersionControl subclass. - rev: the name of the revision to install. - extra_args: a list of extra options. - """ - if extra_args is None: - extra_args = [] + vc_class: Type["VersionControl"] + rev: Optional[str] = None + extra_args: CommandArgs = field(default_factory=list) + branch_name: Optional[str] = None - self.extra_args = extra_args - self.rev = rev - self.vc_class = vc_class - self.branch_name = None # type: Optional[str] - - def __repr__(self): - # type: () -> str - return ''.format(self.vc_class.name, self.rev) + def __repr__(self) -> str: + return f"" @property - def arg_rev(self): - # type: () -> Optional[str] + def arg_rev(self) -> Optional[str]: if self.rev is None: return self.vc_class.default_arg_rev return self.rev - def to_args(self): - # type: () -> CommandArgs + def to_args(self) -> CommandArgs: """ Return the VCS-specific command arguments. """ - args = [] # type: CommandArgs + args: CommandArgs = [] rev = self.arg_rev if rev is not None: args += self.vc_class.get_base_rev_args(rev) @@ -158,15 +153,13 @@ def to_args(self): return args - def to_display(self): - # type: () -> str + def to_display(self) -> str: if not self.rev: - return '' + return "" - return ' (to revision {})'.format(self.rev) + return f" (to revision {self.rev})" - def make_new(self, rev): - # type: (str) -> RevOptions + def make_new(self, rev: str) -> "RevOptions": """ Make a copy of the current instance, but with a new rev. @@ -176,58 +169,47 @@ def make_new(self, rev): return self.vc_class.make_rev_options(rev, extra_args=self.extra_args) -class VcsSupport(object): - _registry = {} # type: Dict[str, VersionControl] - schemes = ['ssh', 'git', 'hg', 'bzr', 'sftp', 'svn'] +class VcsSupport: + _registry: Dict[str, "VersionControl"] = {} + schemes = ["ssh", "git", "hg", "bzr", "sftp", "svn"] - def __init__(self): - # type: () -> None + def __init__(self) -> None: # Register more schemes with urlparse for various version control # systems - urllib_parse.uses_netloc.extend(self.schemes) - # Python >= 2.7.4, 3.3 doesn't have uses_fragment - if getattr(urllib_parse, 'uses_fragment', None): - urllib_parse.uses_fragment.extend(self.schemes) - super(VcsSupport, self).__init__() - - def __iter__(self): - # type: () -> Iterator[str] + urllib.parse.uses_netloc.extend(self.schemes) + super().__init__() + + def __iter__(self) -> Iterator[str]: return self._registry.__iter__() @property - def backends(self): - # type: () -> List[VersionControl] + def backends(self) -> List["VersionControl"]: return list(self._registry.values()) @property - def dirnames(self): - # type: () -> List[str] + def dirnames(self) -> List[str]: return [backend.dirname for backend in self.backends] @property - def all_schemes(self): - # type: () -> List[str] - schemes = [] # type: List[str] + def all_schemes(self) -> List[str]: + schemes: List[str] = [] for backend in self.backends: schemes.extend(backend.schemes) return schemes - def register(self, cls): - # type: (Type[VersionControl]) -> None - if not hasattr(cls, 'name'): - logger.warning('Cannot register VCS %s', cls.__name__) + def register(self, cls: Type["VersionControl"]) -> None: + if not hasattr(cls, "name"): + logger.warning("Cannot register VCS %s", cls.__name__) return if cls.name not in self._registry: self._registry[cls.name] = cls() - logger.debug('Registered VCS backend: %s', cls.name) + logger.debug("Registered VCS backend: %s", cls.name) - def unregister(self, name): - # type: (str) -> None + def unregister(self, name: str) -> None: if name in self._registry: del self._registry[name] - def get_backend_for_dir(self, location): - # type: (str) -> Optional[VersionControl] + def get_backend_for_dir(self, location: str) -> Optional["VersionControl"]: """ Return a VersionControl object if a repository of that type is found at the given directory. @@ -237,8 +219,7 @@ def get_backend_for_dir(self, location): repo_path = vcs_backend.get_repository_root(location) if not repo_path: continue - logger.debug('Determine that %s uses VCS: %s', - location, vcs_backend.name) + logger.debug("Determine that %s uses VCS: %s", location, vcs_backend.name) vcs_backends[repo_path] = vcs_backend if not vcs_backends: @@ -251,8 +232,7 @@ def get_backend_for_dir(self, location): inner_most_repo_path = max(vcs_backends, key=len) return vcs_backends[inner_most_repo_path] - def get_backend_for_scheme(self, scheme): - # type: (str) -> Optional[VersionControl] + def get_backend_for_scheme(self, scheme: str) -> Optional["VersionControl"]: """ Return a VersionControl object or None. """ @@ -261,8 +241,7 @@ def get_backend_for_scheme(self, scheme): return vcs_backend return None - def get_backend(self, name): - # type: (str) -> Optional[VersionControl] + def get_backend(self, name: str) -> Optional["VersionControl"]: """ Return a VersionControl object or None. """ @@ -273,45 +252,41 @@ def get_backend(self, name): vcs = VcsSupport() -class VersionControl(object): - name = '' - dirname = '' - repo_name = '' +class VersionControl: + name = "" + dirname = "" + repo_name = "" # List of supported schemes for this Version Control - schemes = () # type: Tuple[str, ...] + schemes: Tuple[str, ...] = () # Iterable of environment variable names to pass to call_subprocess(). - unset_environ = () # type: Tuple[str, ...] - default_arg_rev = None # type: Optional[str] + unset_environ: Tuple[str, ...] = () + default_arg_rev: Optional[str] = None @classmethod - def should_add_vcs_url_prefix(cls, remote_url): - # type: (str) -> bool + def should_add_vcs_url_prefix(cls, remote_url: str) -> bool: """ Return whether the vcs prefix (e.g. "git+") should be added to a repository's remote url when used in a requirement. """ - return not remote_url.lower().startswith('{}:'.format(cls.name)) + return not remote_url.lower().startswith(f"{cls.name}:") @classmethod - def get_subdirectory(cls, location): - # type: (str) -> Optional[str] + def get_subdirectory(cls, location: str) -> Optional[str]: """ - Return the path to setup.py, relative to the repo root. - Return None if setup.py is in the repo root. + Return the path to Python project root, relative to the repo root. + Return None if the project root is in the repo root. """ return None @classmethod - def get_requirement_revision(cls, repo_dir): - # type: (str) -> str + def get_requirement_revision(cls, repo_dir: str) -> str: """ Return the revision string that should be used in a requirement. """ return cls.get_revision(repo_dir) @classmethod - def get_src_requirement(cls, repo_dir, project_name): - # type: (str, str) -> Optional[str] + def get_src_requirement(cls, repo_dir: str, project_name: str) -> str: """ Return the requirement string to use to redownload the files currently at the given repository directory. @@ -324,22 +299,18 @@ def get_src_requirement(cls, repo_dir, project_name): {repository_url}@{revision}#egg={project_name} """ repo_url = cls.get_remote_url(repo_dir) - if repo_url is None: - return None if cls.should_add_vcs_url_prefix(repo_url): - repo_url = '{}+{}'.format(cls.name, repo_url) + repo_url = f"{cls.name}+{repo_url}" revision = cls.get_requirement_revision(repo_dir) subdir = cls.get_subdirectory(repo_dir) - req = make_vcs_requirement_url(repo_url, revision, project_name, - subdir=subdir) + req = make_vcs_requirement_url(repo_url, revision, project_name, subdir=subdir) return req @staticmethod - def get_base_rev_args(rev): - # type: (str) -> List[str] + def get_base_rev_args(rev: str) -> List[str]: """ Return the base revision arguments for a vcs command. @@ -348,8 +319,7 @@ def get_base_rev_args(rev): """ raise NotImplementedError - def is_immutable_rev_checkout(self, url, dest): - # type: (str, str) -> bool + def is_immutable_rev_checkout(self, url: str, dest: str) -> bool: """ Return true if the commit hash checked out at dest matches the revision in url. @@ -363,8 +333,9 @@ def is_immutable_rev_checkout(self, url, dest): return False @classmethod - def make_rev_options(cls, rev=None, extra_args=None): - # type: (Optional[str], Optional[CommandArgs]) -> RevOptions + def make_rev_options( + cls, rev: Optional[str] = None, extra_args: Optional[CommandArgs] = None + ) -> RevOptions: """ Return a RevOptions object. @@ -372,31 +343,21 @@ def make_rev_options(cls, rev=None, extra_args=None): rev: the name of a revision to install. extra_args: a list of extra options. """ - return RevOptions(cls, rev, extra_args=extra_args) + return RevOptions(cls, rev, extra_args=extra_args or []) @classmethod - def _is_local_repository(cls, repo): - # type: (str) -> bool + def _is_local_repository(cls, repo: str) -> bool: """ - posix absolute paths start with os.path.sep, - win32 ones start with drive (like c:\\folder) + posix absolute paths start with os.path.sep, + win32 ones start with drive (like c:\\folder) """ drive, tail = os.path.splitdrive(repo) return repo.startswith(os.path.sep) or bool(drive) - def export(self, location, url): - # type: (str, HiddenText) -> None - """ - Export the repository at the url to the destination location - i.e. only download the files, without vcs informations - - :param url: the repository URL starting with a vcs prefix. - """ - raise NotImplementedError - @classmethod - def get_netloc_and_auth(cls, netloc, scheme): - # type: (str, str) -> Tuple[str, Tuple[Optional[str], Optional[str]]] + def get_netloc_and_auth( + cls, netloc: str, scheme: str + ) -> Tuple[str, Tuple[Optional[str], Optional[str]]]: """ Parse the repository URL's netloc, and return the new netloc to use along with auth information. @@ -415,53 +376,52 @@ def get_netloc_and_auth(cls, netloc, scheme): return netloc, (None, None) @classmethod - def get_url_rev_and_auth(cls, url): - # type: (str) -> Tuple[str, Optional[str], AuthInfo] + def get_url_rev_and_auth(cls, url: str) -> Tuple[str, Optional[str], AuthInfo]: """ Parse the repository URL to use, and return the URL, revision, and auth info to use. Returns: (url, rev, (username, password)). """ - scheme, netloc, path, query, frag = urllib_parse.urlsplit(url) - if '+' not in scheme: + scheme, netloc, path, query, frag = urllib.parse.urlsplit(url) + if "+" not in scheme: raise ValueError( - "Sorry, {!r} is a malformed VCS url. " + f"Sorry, {url!r} is a malformed VCS url. " "The format is +://, " - "e.g. svn+http://myrepo/svn/MyApp#egg=MyApp".format(url) + "e.g. svn+http://myrepo/svn/MyApp#egg=MyApp" ) # Remove the vcs prefix. - scheme = scheme.split('+', 1)[1] + scheme = scheme.split("+", 1)[1] netloc, user_pass = cls.get_netloc_and_auth(netloc, scheme) rev = None - if '@' in path: - path, rev = path.rsplit('@', 1) + if "@" in path: + path, rev = path.rsplit("@", 1) if not rev: raise InstallationError( - "The URL {!r} has an empty revision (after @) " + f"The URL {url!r} has an empty revision (after @) " "which is not supported. Include a revision after @ " - "or remove @ from the URL.".format(url) + "or remove @ from the URL." ) - url = urllib_parse.urlunsplit((scheme, netloc, path, query, '')) + url = urllib.parse.urlunsplit((scheme, netloc, path, query, "")) return url, rev, user_pass @staticmethod - def make_rev_args(username, password): - # type: (Optional[str], Optional[HiddenText]) -> CommandArgs + def make_rev_args( + username: Optional[str], password: Optional[HiddenText] + ) -> CommandArgs: """ Return the RevOptions "extra arguments" to use in obtain(). """ return [] - def get_url_rev_options(self, url): - # type: (HiddenText) -> Tuple[HiddenText, RevOptions] + def get_url_rev_options(self, url: HiddenText) -> Tuple[HiddenText, RevOptions]: """ - Return the URL and RevOptions object to use in obtain() and in - some cases export(), as a tuple (url, rev_options). + Return the URL and RevOptions object to use in obtain(), + as a tuple (url, rev_options). """ secret_url, rev, user_pass = self.get_url_rev_and_auth(url.secret) username, secret_password = user_pass - password = None # type: Optional[HiddenText] + password: Optional[HiddenText] = None if secret_password is not None: password = hide_value(secret_password) extra_args = self.make_rev_args(username, password) @@ -470,24 +430,23 @@ def get_url_rev_options(self, url): return hide_url(secret_url), rev_options @staticmethod - def normalize_url(url): - # type: (str) -> str + def normalize_url(url: str) -> str: """ Normalize a URL for comparison by unquoting it and removing any trailing slash. """ - return urllib_parse.unquote(url).rstrip('/') + return urllib.parse.unquote(url).rstrip("/") @classmethod - def compare_urls(cls, url1, url2): - # type: (str, str) -> bool + def compare_urls(cls, url1: str, url2: str) -> bool: """ Compare two repo URLs for identity, ignoring incidental differences. """ - return (cls.normalize_url(url1) == cls.normalize_url(url2)) + return cls.normalize_url(url1) == cls.normalize_url(url2) - def fetch_new(self, dest, url, rev_options): - # type: (str, HiddenText, RevOptions) -> None + def fetch_new( + self, dest: str, url: HiddenText, rev_options: RevOptions, verbosity: int + ) -> None: """ Fetch a revision from a repository, in the case that this is the first fetch from the repository. @@ -495,11 +454,11 @@ def fetch_new(self, dest, url, rev_options): Args: dest: the directory to fetch the repository to. rev_options: a RevOptions object. + verbosity: verbosity level. """ raise NotImplementedError - def switch(self, dest, url, rev_options): - # type: (str, HiddenText, RevOptions) -> None + def switch(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None: """ Switch the repo at ``dest`` to point to ``URL``. @@ -508,8 +467,7 @@ def switch(self, dest, url, rev_options): """ raise NotImplementedError - def update(self, dest, url, rev_options): - # type: (str, HiddenText, RevOptions) -> None + def update(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None: """ Update an already-existing repo to the given ``rev_options``. @@ -519,8 +477,7 @@ def update(self, dest, url, rev_options): raise NotImplementedError @classmethod - def is_commit_id_equal(cls, dest, name): - # type: (str, Optional[str]) -> bool + def is_commit_id_equal(cls, dest: str, name: Optional[str]) -> bool: """ Return whether the id of the current commit equals the given name. @@ -530,19 +487,19 @@ def is_commit_id_equal(cls, dest, name): """ raise NotImplementedError - def obtain(self, dest, url): - # type: (str, HiddenText) -> None + def obtain(self, dest: str, url: HiddenText, verbosity: int) -> None: """ Install or update in editable mode the package represented by this VersionControl object. :param dest: the repository directory in which to install or update. :param url: the repository URL starting with a vcs prefix. + :param verbosity: verbosity level. """ url, rev_options = self.get_url_rev_options(url) if not os.path.exists(dest): - self.fetch_new(dest, url, rev_options) + self.fetch_new(dest, url, rev_options, verbosity=verbosity) return rev_display = rev_options.to_display() @@ -550,95 +507,89 @@ def obtain(self, dest, url): existing_url = self.get_remote_url(dest) if self.compare_urls(existing_url, url.secret): logger.debug( - '%s in %s exists, and has correct URL (%s)', + "%s in %s exists, and has correct URL (%s)", self.repo_name.title(), display_path(dest), url, ) if not self.is_commit_id_equal(dest, rev_options.rev): logger.info( - 'Updating %s %s%s', + "Updating %s %s%s", display_path(dest), self.repo_name, rev_display, ) self.update(dest, url, rev_options) else: - logger.info('Skipping because already up-to-date.') + logger.info("Skipping because already up-to-date.") return logger.warning( - '%s %s in %s exists with URL %s', + "%s %s in %s exists with URL %s", self.name, self.repo_name, display_path(dest), existing_url, ) - prompt = ('(s)witch, (i)gnore, (w)ipe, (b)ackup ', - ('s', 'i', 'w', 'b')) + prompt = ("(s)witch, (i)gnore, (w)ipe, (b)ackup ", ("s", "i", "w", "b")) else: logger.warning( - 'Directory %s already exists, and is not a %s %s.', + "Directory %s already exists, and is not a %s %s.", dest, self.name, self.repo_name, ) # https://github.com/python/mypy/issues/1174 - prompt = ('(i)gnore, (w)ipe, (b)ackup ', # type: ignore - ('i', 'w', 'b')) + prompt = ("(i)gnore, (w)ipe, (b)ackup ", ("i", "w", "b")) # type: ignore logger.warning( - 'The plan is to install the %s repository %s', + "The plan is to install the %s repository %s", self.name, url, ) - # response = ask_path_exists('What to do? {}'.format( - # prompt[0]), prompt[1]) - - # if response == 'a': - # sys.exit(-1) - - # if response == 'w': - # logger.warning('Deleting %s', display_path(dest)) - # rmtree(dest) - # self.fetch_new(dest, url, rev_options) - # return - - # if response == 'b': - # dest_dir = backup_dir(dest) - # logger.warning( - # 'Backing up %s to %s', display_path(dest), dest_dir, - # ) - # shutil.move(dest, dest_dir) - # self.fetch_new(dest, url, rev_options) - # return - - # # Do nothing if the response is "i". - # if response == 's': - # logger.info( - # 'Switching %s %s to %s%s', - # self.repo_name, - # display_path(dest), - # url, - # rev_display, - # ) - # self.switch(dest, url, rev_options) - - def unpack(self, location, url): - # type: (str, HiddenText) -> None + response = ask_path_exists(f"What to do? {prompt[0]}", prompt[1]) + + if response == "a": + sys.exit(-1) + + if response == "w": + logger.warning("Deleting %s", display_path(dest)) + rmtree(dest) + self.fetch_new(dest, url, rev_options, verbosity=verbosity) + return + + if response == "b": + dest_dir = backup_dir(dest) + logger.warning("Backing up %s to %s", display_path(dest), dest_dir) + shutil.move(dest, dest_dir) + self.fetch_new(dest, url, rev_options, verbosity=verbosity) + return + + # Do nothing if the response is "i". + if response == "s": + logger.info( + "Switching %s %s to %s%s", + self.repo_name, + display_path(dest), + url, + rev_display, + ) + self.switch(dest, url, rev_options) + + def unpack(self, location: str, url: HiddenText, verbosity: int) -> None: """ Clean up current location and download the url repository (and vcs infos) into location :param url: the repository URL starting with a vcs prefix. + :param verbosity: verbosity level. """ if os.path.exists(location): rmtree(location) - self.obtain(location, url=url) + self.obtain(location, url=url, verbosity=verbosity) @classmethod - def get_remote_url(cls, location): - # type: (str) -> str + def get_remote_url(cls, location: str) -> str: """ Return the url used at location @@ -648,8 +599,7 @@ def get_remote_url(cls, location): raise NotImplementedError @classmethod - def get_revision(cls, location): - # type: (str) -> str + def get_revision(cls, location: str) -> str: """ Return the current commit id of the files at the given location. """ @@ -658,56 +608,71 @@ def get_revision(cls, location): @classmethod def run_command( cls, - cmd, # type: Union[List[str], CommandArgs] - show_stdout=True, # type: bool - cwd=None, # type: Optional[str] - on_returncode='raise', # type: str - extra_ok_returncodes=None, # type: Optional[Iterable[int]] - command_desc=None, # type: Optional[str] - extra_environ=None, # type: Optional[Mapping[str, Any]] - spinner=None, # type: Optional[SpinnerInterface] - log_failed_cmd=True # type: bool - ): - # type: (...) -> Text + cmd: Union[List[str], CommandArgs], + show_stdout: bool = True, + cwd: Optional[str] = None, + on_returncode: 'Literal["raise", "warn", "ignore"]' = "raise", + extra_ok_returncodes: Optional[Iterable[int]] = None, + command_desc: Optional[str] = None, + extra_environ: Optional[Mapping[str, Any]] = None, + spinner: Optional[SpinnerInterface] = None, + log_failed_cmd: bool = True, + stdout_only: bool = False, + ) -> str: """ Run a VCS subcommand This is simply a wrapper around call_subprocess that adds the VCS command name, and checks that the VCS is available """ cmd = make_command(cls.name, *cmd) + if command_desc is None: + command_desc = format_command_args(cmd) try: - return call_subprocess(cmd, show_stdout, cwd, - on_returncode=on_returncode, - extra_ok_returncodes=extra_ok_returncodes, - command_desc=command_desc, - extra_environ=extra_environ, - unset_environ=cls.unset_environ, - spinner=spinner, - log_failed_cmd=log_failed_cmd) - except OSError as e: + return call_subprocess( + cmd, + show_stdout, + cwd, + on_returncode=on_returncode, + extra_ok_returncodes=extra_ok_returncodes, + command_desc=command_desc, + extra_environ=extra_environ, + unset_environ=cls.unset_environ, + spinner=spinner, + log_failed_cmd=log_failed_cmd, + stdout_only=stdout_only, + ) + except NotADirectoryError: + raise BadCommand(f"Cannot find command {cls.name!r} - invalid PATH") + except FileNotFoundError: # errno.ENOENT = no such file or directory # In other words, the VCS executable isn't available - if e.errno == errno.ENOENT: - raise BadCommand( - 'Cannot find command {cls.name!r} - do you have ' - '{cls.name!r} installed and in your ' - 'PATH?'.format(**locals())) - else: - raise # re-raise exception if a different error occurred + raise BadCommand( + f"Cannot find command {cls.name!r} - do you have " + f"{cls.name!r} installed and in your PATH?" + ) + except PermissionError: + # errno.EACCES = Permission denied + # This error occurs, for instance, when the command is installed + # only for another user. So, the current user don't have + # permission to call the other user command. + raise BadCommand( + f"No permission to execute {cls.name!r} - install it " + f"locally, globally (ask admin), or check your PATH. " + f"See possible solutions at " + f"https://pip.pypa.io/en/latest/reference/pip_freeze/" + f"#fixing-permission-denied." + ) @classmethod - def is_repository_directory(cls, path): - # type: (str) -> bool + def is_repository_directory(cls, path: str) -> bool: """ Return whether a directory path is a repository directory. """ - logger.debug('Checking in %s for %s (%s)...', - path, cls.dirname, cls.name) + logger.debug("Checking in %s for %s (%s)...", path, cls.dirname, cls.name) return os.path.exists(os.path.join(path, cls.dirname)) @classmethod - def get_repository_root(cls, location): - # type: (str) -> Optional[str] + def get_repository_root(cls, location: str) -> Optional[str]: """ Return the "root" (top-level) directory controlled by the vcs, or `None` if the directory is not in any. diff --git a/src/fetchcode/vcs/pip/_internal/wheel_builder.py b/src/fetchcode/vcs/pip/_internal/wheel_builder.py index 329a9627..93f8e1f5 100644 --- a/src/fetchcode/vcs/pip/_internal/wheel_builder.py +++ b/src/fetchcode/vcs/pip/_internal/wheel_builder.py @@ -1,43 +1,40 @@ """Orchestrator for building wheels from InstallRequirements. """ -# The following comment should be removed at some point in the future. -# mypy: strict-optional=False - import logging import os.path import re import shutil +from typing import Iterable, List, Optional, Tuple + +from pip._vendor.packaging.utils import canonicalize_name, canonicalize_version +from pip._vendor.packaging.version import InvalidVersion, Version + +from pip._internal.cache import WheelCache +from pip._internal.exceptions import InvalidWheelFilename, UnsupportedWheel +from pip._internal.metadata import FilesystemWheel, get_wheel_distribution +from pip._internal.models.link import Link +from pip._internal.models.wheel import Wheel +from pip._internal.operations.build.wheel import build_wheel_pep517 +from pip._internal.operations.build.wheel_editable import build_wheel_editable +from pip._internal.operations.build.wheel_legacy import build_wheel_legacy +from pip._internal.req.req_install import InstallRequirement +from pip._internal.utils.logging import indent_log +from pip._internal.utils.misc import ensure_dir, hash_file +from pip._internal.utils.setuptools_build import make_setuptools_clean_args +from pip._internal.utils.subprocess import call_subprocess +from pip._internal.utils.temp_dir import TempDirectory +from pip._internal.utils.urls import path_to_url +from pip._internal.vcs import vcs -from fetchcode.vcs.pip._internal.models.link import Link -from fetchcode.vcs.pip._internal.operations.build.wheel import build_wheel_pep517 -from fetchcode.vcs.pip._internal.operations.build.wheel_legacy import build_wheel_legacy -from fetchcode.vcs.pip._internal.utils.logging import indent_log -from fetchcode.vcs.pip._internal.utils.misc import ensure_dir, hash_file, is_wheel_installed -from fetchcode.vcs.pip._internal.utils.setuptools_build import make_setuptools_clean_args -from fetchcode.vcs.pip._internal.utils.subprocess import call_subprocess -from fetchcode.vcs.pip._internal.utils.temp_dir import TempDirectory -from fetchcode.vcs.pip._internal.utils.typing import MYPY_CHECK_RUNNING -from fetchcode.vcs.pip._internal.utils.urls import path_to_url -from fetchcode.vcs.pip._internal.vcs import vcs - -if MYPY_CHECK_RUNNING: - from typing import ( - Any, Callable, Iterable, List, Optional, Pattern, Tuple, - ) - - from fetchcode.vcs.pip._internal.cache import WheelCache - from fetchcode.vcs.pip._internal.req.req_install import InstallRequirement +logger = logging.getLogger(__name__) - BinaryAllowedPredicate = Callable[[InstallRequirement], bool] - BuildResult = Tuple[List[InstallRequirement], List[InstallRequirement]] +_egg_info_re = re.compile(r"([a-z0-9_.]+)-([a-z0-9_.!+-]+)", re.IGNORECASE) -logger = logging.getLogger(__name__) +BuildResult = Tuple[List[InstallRequirement], List[InstallRequirement]] -def _contains_egg_info( - s, _egg_info_re=re.compile(r'([a-z0-9_.]+)-([a-z0-9_.!+-]+)', re.I)): - # type: (str, Pattern[str]) -> bool +def _contains_egg_info(s: str) -> bool: """Determine whether the string looks like an egg_info. :param s: The string to parse. E.g. foo-2.1 @@ -46,11 +43,9 @@ def _contains_egg_info( def _should_build( - req, # type: InstallRequirement - need_wheel, # type: bool - check_binary_allowed, # type: BinaryAllowedPredicate -): - # type: (...) -> bool + req: InstallRequirement, + need_wheel: bool, +) -> bool: """Return whether an InstallRequirement should be built into a wheel.""" if req.constraint: # never build requirements that are merely constraints @@ -58,7 +53,8 @@ def _should_build( if req.is_wheel: if need_wheel: logger.info( - 'Skipping %s, due to already being wheel.', req.name, + "Skipping %s, due to already being wheel.", + req.name, ) return False @@ -69,60 +65,38 @@ def _should_build( # From this point, this concerns the pip install command only # (need_wheel=False). - if req.editable or not req.source_dir: + if not req.source_dir: return False - if not check_binary_allowed(req): - logger.info( - "Skipping wheel build for %s, due to binaries " - "being disabled for it.", req.name, - ) - return False - - if not req.use_pep517 and not is_wheel_installed(): - # we don't build legacy requirements if wheel is not installed - logger.info( - "Using legacy setup.py install for %s, " - "since package 'wheel' is not installed.", req.name, - ) - return False + if req.editable: + # we only build PEP 660 editable requirements + return req.supports_pyproject_editable return True def should_build_for_wheel_command( - req, # type: InstallRequirement -): - # type: (...) -> bool - return _should_build( - req, need_wheel=True, check_binary_allowed=_always_true - ) + req: InstallRequirement, +) -> bool: + return _should_build(req, need_wheel=True) def should_build_for_install_command( - req, # type: InstallRequirement - check_binary_allowed, # type: BinaryAllowedPredicate -): - # type: (...) -> bool - return _should_build( - req, need_wheel=False, check_binary_allowed=check_binary_allowed - ) + req: InstallRequirement, +) -> bool: + return _should_build(req, need_wheel=False) def _should_cache( - req, # type: InstallRequirement -): - # type: (...) -> Optional[bool] + req: InstallRequirement, +) -> Optional[bool]: """ Return whether a built InstallRequirement can be stored in the persistent wheel cache, assuming the wheel cache is available, and _should_build() has determined a wheel needs to be built. """ - if not should_build_for_install_command( - req, check_binary_allowed=_always_true - ): - # never cache if pip install would not have built - # (editable mode, etc) + if req.editable or not req.source_dir: + # never cache editable requirements return False if req.link and req.link.is_vcs: @@ -136,6 +110,7 @@ def _should_cache( return True return False + assert req.link base, ext = req.link.splitext() if _contains_egg_info(base): return True @@ -145,14 +120,14 @@ def _should_cache( def _get_cache_dir( - req, # type: InstallRequirement - wheel_cache, # type: WheelCache -): - # type: (...) -> str + req: InstallRequirement, + wheel_cache: WheelCache, +) -> str: """Return the persistent or temporary cache directory where the built wheel need to be stored. """ cache_available = bool(wheel_cache.cache_dir) + assert req.link if cache_available and _should_cache(req): cache_dir = wheel_cache.get_path_for_link(req.link) else: @@ -160,54 +135,107 @@ def _get_cache_dir( return cache_dir -def _always_true(_): - # type: (Any) -> bool - return True +def _verify_one(req: InstallRequirement, wheel_path: str) -> None: + canonical_name = canonicalize_name(req.name or "") + w = Wheel(os.path.basename(wheel_path)) + if canonicalize_name(w.name) != canonical_name: + raise InvalidWheelFilename( + f"Wheel has unexpected file name: expected {canonical_name!r}, " + f"got {w.name!r}", + ) + dist = get_wheel_distribution(FilesystemWheel(wheel_path), canonical_name) + dist_verstr = str(dist.version) + if canonicalize_version(dist_verstr) != canonicalize_version(w.version): + raise InvalidWheelFilename( + f"Wheel has unexpected file name: expected {dist_verstr!r}, " + f"got {w.version!r}", + ) + metadata_version_value = dist.metadata_version + if metadata_version_value is None: + raise UnsupportedWheel("Missing Metadata-Version") + try: + metadata_version = Version(metadata_version_value) + except InvalidVersion: + msg = f"Invalid Metadata-Version: {metadata_version_value}" + raise UnsupportedWheel(msg) + if metadata_version >= Version("1.2") and not isinstance(dist.version, Version): + raise UnsupportedWheel( + f"Metadata 1.2 mandates PEP 440 version, but {dist_verstr!r} is not" + ) def _build_one( - req, # type: InstallRequirement - output_dir, # type: str - build_options, # type: List[str] - global_options, # type: List[str] -): - # type: (...) -> Optional[str] + req: InstallRequirement, + output_dir: str, + verify: bool, + build_options: List[str], + global_options: List[str], + editable: bool, +) -> Optional[str]: """Build one wheel. :return: The filename of the built wheel, or None if the build failed. """ + artifact = "editable" if editable else "wheel" try: ensure_dir(output_dir) except OSError as e: logger.warning( - "Building wheel for %s failed: %s", - req.name, e, + "Building %s for %s failed: %s", + artifact, + req.name, + e, ) return None # Install build deps into temporary directory (PEP 518) with req.build_env: - return _build_one_inside_env( - req, output_dir, build_options, global_options + wheel_path = _build_one_inside_env( + req, output_dir, build_options, global_options, editable ) + if wheel_path and verify: + try: + _verify_one(req, wheel_path) + except (InvalidWheelFilename, UnsupportedWheel) as e: + logger.warning("Built %s for %s is invalid: %s", artifact, req.name, e) + return None + return wheel_path def _build_one_inside_env( - req, # type: InstallRequirement - output_dir, # type: str - build_options, # type: List[str] - global_options, # type: List[str] -): - # type: (...) -> Optional[str] + req: InstallRequirement, + output_dir: str, + build_options: List[str], + global_options: List[str], + editable: bool, +) -> Optional[str]: with TempDirectory(kind="wheel") as temp_dir: + assert req.name if req.use_pep517: - wheel_path = build_wheel_pep517( - name=req.name, - backend=req.pep517_backend, - metadata_directory=req.metadata_directory, - build_options=build_options, - tempd=temp_dir.path, - ) + assert req.metadata_directory + assert req.pep517_backend + if global_options: + logger.warning( + "Ignoring --global-option when building %s using PEP 517", req.name + ) + if build_options: + logger.warning( + "Ignoring --build-option when building %s using PEP 517", req.name + ) + if editable: + wheel_path = build_wheel_editable( + name=req.name, + backend=req.pep517_backend, + metadata_directory=req.metadata_directory, + tempd=temp_dir.path, + ) + else: + wheel_path = build_wheel_pep517( + name=req.name, + backend=req.pep517_backend, + metadata_directory=req.metadata_directory, + tempd=temp_dir.path, + ) else: wheel_path = build_wheel_legacy( name=req.name, @@ -224,16 +252,20 @@ def _build_one_inside_env( try: wheel_hash, length = hash_file(wheel_path) shutil.move(wheel_path, dest_path) - logger.info('Created wheel for %s: ' - 'filename=%s size=%d sha256=%s', - req.name, wheel_name, length, - wheel_hash.hexdigest()) - logger.info('Stored in directory: %s', output_dir) + logger.info( + "Created wheel for %s: filename=%s size=%d sha256=%s", + req.name, + wheel_name, + length, + wheel_hash.hexdigest(), + ) + logger.info("Stored in directory: %s", output_dir) return dest_path except Exception as e: logger.warning( "Building wheel for %s failed: %s", - req.name, e, + req.name, + e, ) # Ignore return, we can't do anything else useful. if not req.use_pep517: @@ -241,29 +273,30 @@ def _build_one_inside_env( return None -def _clean_one_legacy(req, global_options): - # type: (InstallRequirement, List[str]) -> bool +def _clean_one_legacy(req: InstallRequirement, global_options: List[str]) -> bool: clean_args = make_setuptools_clean_args( req.setup_py_path, global_options=global_options, ) - logger.info('Running setup.py clean for %s', req.name) + logger.info("Running setup.py clean for %s", req.name) try: - call_subprocess(clean_args, cwd=req.source_dir) + call_subprocess( + clean_args, command_desc="python setup.py clean", cwd=req.source_dir + ) return True except Exception: - logger.error('Failed cleaning build dir for %s', req.name) + logger.error("Failed cleaning build dir for %s", req.name) return False def build( - requirements, # type: Iterable[InstallRequirement] - wheel_cache, # type: WheelCache - build_options, # type: List[str] - global_options, # type: List[str] -): - # type: (...) -> BuildResult + requirements: Iterable[InstallRequirement], + wheel_cache: WheelCache, + verify: bool, + build_options: List[str], + global_options: List[str], +) -> BuildResult: """Build wheels. :return: The list of InstallRequirement that succeeded to build and @@ -274,18 +307,30 @@ def build( # Build the wheels. logger.info( - 'Building wheels for collected packages: %s', - ', '.join(req.name for req in requirements), + "Building wheels for collected packages: %s", + ", ".join(req.name for req in requirements), # type: ignore ) with indent_log(): build_successes, build_failures = [], [] for req in requirements: + assert req.name cache_dir = _get_cache_dir(req, wheel_cache) wheel_file = _build_one( - req, cache_dir, build_options, global_options + req, + cache_dir, + verify, + build_options, + global_options, + req.editable and req.permit_editable_wheels, ) if wheel_file: + # Record the download origin in the cache + if req.download_info is not None: + # download_info is guaranteed to be set because when we build an + # InstallRequirement it has been through the preparer before, but + # let's be cautious. + wheel_cache.record_download_origin(cache_dir, req.download_info) # Update the link for this. req.link = Link(path_to_url(wheel_file)) req.local_file_path = req.link.file_path @@ -297,13 +342,13 @@ def build( # notify success/failure if build_successes: logger.info( - 'Successfully built %s', - ' '.join([req.name for req in build_successes]), + "Successfully built %s", + " ".join([req.name for req in build_successes]), # type: ignore ) if build_failures: logger.info( - 'Failed to build %s', - ' '.join([req.name for req in build_failures]), + "Failed to build %s", + " ".join([req.name for req in build_failures]), # type: ignore ) # Return a list of requirements that failed to build return build_successes, build_failures diff --git a/src/fetchcode/vcs/pip/_vendor/README.rst b/src/fetchcode/vcs/pip/_vendor/README.rst index c5ed6b0d..e3fe8041 100644 --- a/src/fetchcode/vcs/pip/_vendor/README.rst +++ b/src/fetchcode/vcs/pip/_vendor/README.rst @@ -4,23 +4,19 @@ Vendoring Policy * Vendored libraries **MUST** not be modified except as required to successfully vendor them. - * Vendored libraries **MUST** be released copies of libraries available on PyPI. - +* Vendored libraries **MUST** be available under a license that allows + them to be integrated into ``pip``, which is released under the MIT license. * Vendored libraries **MUST** be accompanied with LICENSE files. - * The versions of libraries vendored in pip **MUST** be reflected in ``pip/_vendor/vendor.txt``. - -* Vendored libraries **MUST** function without any build steps such as ``2to3`` or - compilation of C code, practically this limits to single source 2.x/3.x and +* Vendored libraries **MUST** function without any build steps such as ``2to3`` + or compilation of C code, practically this limits to single source 2.x/3.x and pure Python. - * Any modifications made to libraries **MUST** be noted in ``pip/_vendor/README.rst`` and their corresponding patches **MUST** be - included ``tools/automation/vendoring/patches``. - + included ``tools/vendoring/patches``. * Vendored libraries should have corresponding ``vendored()`` entries in ``pip/_vendor/__init__.py``. @@ -38,53 +34,53 @@ higher quality and more battle tested code, centralization of bug fixes However, there are several issues with having dependencies in the traditional way (via ``install_requires``) for pip. These issues are: -* **Fragility.** When pip depends on another library to function then if for - whatever reason that library either isn't installed or an incompatible - version is installed then pip ceases to function. This is of course true for - all Python applications, however for every application *except* for pip the - way you fix it is by re-running pip. Obviously, when pip can't run, you can't - use pip to fix pip, so you're left having to manually resolve dependencies and - installing them by hand. - -* **Making other libraries uninstallable.** One of pip's current dependencies is - the ``requests`` library, for which pip requires a fairly recent version to run. - If pip depended on ``requests`` in the traditional manner, then we'd either - have to maintain compatibility with every ``requests`` version that has ever - existed (and ever will), OR allow pip to render certain versions of ``requests`` - uninstallable. (The second issue, although technically true for any Python - application, is magnified by pip's ubiquity; pip is installed by default in - Python, in ``pyvenv``, and in ``virtualenv``.) - -* **Security.** This might seem puzzling at first glance, since vendoring - has a tendency to complicate updating dependencies for security updates, - and that holds true for pip. However, given the *other* reasons for avoiding - dependencies, the alternative is for pip to reinvent the wheel itself. - This is what pip did historically. It forced pip to re-implement its own - HTTPS verification routines as a workaround for the Python standard library's - lack of SSL validation, which resulted in similar bugs in the validation routine - in ``requests`` and ``urllib3``, except that they had to be discovered and - fixed independently. Even though we're vendoring, reusing libraries keeps pip - more secure by relying on the great work of our dependencies, *and* allowing for - faster, easier security fixes by simply pulling in newer versions of dependencies. - -* **Bootstrapping.** Currently most popular methods of installing pip rely - on pip's self-contained nature to install pip itself. These tools work by bundling - a copy of pip, adding it to ``sys.path``, and then executing that copy of pip. - This is done instead of implementing a "mini installer" (to reduce duplication); - pip already knows how to install a Python package, and is far more battle-tested - than any "mini installer" could ever possibly be. +**Fragility** + When pip depends on another library to function then if for whatever reason + that library either isn't installed or an incompatible version is installed + then pip ceases to function. This is of course true for all Python + applications, however for every application *except* for pip the way you fix + it is by re-running pip. Obviously, when pip can't run, you can't use pip to + fix pip, so you're left having to manually resolve dependencies and + installing them by hand. + +**Making other libraries uninstallable** + One of pip's current dependencies is the ``requests`` library, for which pip + requires a fairly recent version to run. If pip depended on ``requests`` in + the traditional manner, then we'd either have to maintain compatibility with + every ``requests`` version that has ever existed (and ever will), OR allow + pip to render certain versions of ``requests`` uninstallable. (The second + issue, although technically true for any Python application, is magnified by + pip's ubiquity; pip is installed by default in Python, in ``pyvenv``, and in + ``virtualenv``.) + +**Security** + This might seem puzzling at first glance, since vendoring has a tendency to + complicate updating dependencies for security updates, and that holds true + for pip. However, given the *other* reasons for avoiding dependencies, the + alternative is for pip to reinvent the wheel itself. This is what pip did + historically. It forced pip to re-implement its own HTTPS verification + routines as a workaround for the Python standard library's lack of SSL + validation, which resulted in similar bugs in the validation routine in + ``requests`` and ``urllib3``, except that they had to be discovered and + fixed independently. Even though we're vendoring, reusing libraries keeps + pip more secure by relying on the great work of our dependencies, *and* + allowing for faster, easier security fixes by simply pulling in newer + versions of dependencies. + +**Bootstrapping** + Currently most popular methods of installing pip rely on pip's + self-contained nature to install pip itself. These tools work by bundling a + copy of pip, adding it to ``sys.path``, and then executing that copy of pip. + This is done instead of implementing a "mini installer" (to reduce + duplication); pip already knows how to install a Python package, and is far + more battle-tested than any "mini installer" could ever possibly be. Many downstream redistributors have policies against this kind of bundling, and instead opt to patch the software they distribute to debundle it and make it rely on the global versions of the software that they already have packaged (which may have its own patches applied to it). We (the pip team) would prefer it if pip was *not* debundled in this manner due to the above reasons and -instead we would prefer it if pip would be left intact as it is now. The one -exception to this, is it is acceptable to remove the -``pip/_vendor/requests/cacert.pem`` file provided you ensure that the -``ssl.get_default_verify_paths().cafile`` API returns the correct CA bundle for -your system. This will ensure that pip will use your system provided CA bundle -instead of the copy bundled with pip. +instead we would prefer it if pip would be left intact as it is now. In the longer term, if someone has a *portable* solution to the above problems, other than the bundling method we currently use, that doesn't add additional @@ -97,23 +93,50 @@ such as OS packages. Modifications ============= -* ``setuptools`` is completely stripped to only keep ``pkg_resources`` -* ``pkg_resources`` has been modified to import its dependencies from ``pip._vendor`` -* ``packaging`` has been modified to import its dependencies from ``pip._vendor`` -* ``html5lib`` has been modified to import six from ``pip._vendor`` and - to prefer importing from ``collections.abc`` instead of ``collections``. -* ``CacheControl`` has been modified to import its dependencies from ``pip._vendor`` -* ``requests`` has been modified to import its other dependencies from ``pip._vendor`` - and to *not* load ``simplejson`` (all platforms) and ``pyopenssl`` (Windows). - +* ``setuptools`` is completely stripped to only keep ``pkg_resources``. +* ``pkg_resources`` has been modified to import its dependencies from + ``pip._vendor``, and to use the vendored copy of ``platformdirs`` + rather than ``appdirs``. +* ``packaging`` has been modified to import its dependencies from + ``pip._vendor``. +* ``CacheControl`` has been modified to import its dependencies from + ``pip._vendor``. +* ``requests`` has been modified to import its other dependencies from + ``pip._vendor`` and to *not* load ``simplejson`` (all platforms) and + ``pyopenssl`` (Windows). +* ``platformdirs`` has been modified to import its submodules from ``pip._vendor.platformdirs``. Automatic Vendoring =================== -Vendoring is automated via the ``vendoring`` tool from the content of +Vendoring is automated via the `vendoring `_ tool from the content of ``pip/_vendor/vendor.txt`` and the different patches in -``tools/automation/vendoring/patches``. +``tools/vendoring/patches``. Launch it via ``vendoring sync . -v`` (requires ``vendoring>=0.2.2``). +Tool configuration is done via ``pyproject.toml``. + + +Managing Local Patches +====================== + +The ``vendoring`` tool automatically applies our local patches, but updating, +the patches sometimes no longer apply cleanly. In that case, the update will +fail. To resolve this, take the following steps: + +1. Revert any incomplete changes in the revendoring branch, to ensure you have + a clean starting point. +2. Run the revendoring of the library with a problem again: ``nox -s vendoring + -- --upgrade ``. +3. This will fail again, but you will have the original source in your working + directory. Review the existing patch against the source, and modify the patch + to reflect the new version of the source. If you ``git add`` the changes the + vendoring made, you can modify the source to reflect the patch file and then + generate a new patch with ``git diff``. +4. Now, revert everything *except* the patch file changes. Leave the modified + patch file unstaged but saved in the working tree. +5. Re-run the vendoring. This time, it should pick up the changed patch file + and apply it cleanly. The patch file changes will be committed along with the + revendoring, so the new commit should be ready to test and publish as a PR. Debundling @@ -126,24 +149,19 @@ semi-supported method (that we don't test in our CI) and requires a bit of extra work on your end in order to solve the problems described above. 1. Delete everything in ``pip/_vendor/`` **except** for - ``pip/_vendor/__init__.py``. - + ``pip/_vendor/__init__.py`` and ``pip/_vendor/vendor.txt``. 2. Generate wheels for each of pip's dependencies (and any of their dependencies) using your patched copies of these libraries. These must be placed somewhere on the filesystem that pip can access (``pip/_vendor`` is the default assumption). - 3. Modify ``pip/_vendor/__init__.py`` so that the ``DEBUNDLED`` variable is ``True``. - 4. Upon installation, the ``INSTALLER`` file in pip's own ``dist-info`` directory should be set to something other than ``pip``, so that pip can detect that it wasn't installed using itself. - 5. *(optional)* If you've placed the wheels in a location other than ``pip/_vendor/``, then modify ``pip/_vendor/__init__.py`` so that the ``WHEEL_DIR`` variable points to the location you've placed them. - 6. *(optional)* Update the ``pip_self_version_check`` logic to use the appropriate logic for determining the latest available version of pip and prompt the user with the correct upgrade message. diff --git a/src/fetchcode/vcs/pip/_vendor/__init__.py b/src/fetchcode/vcs/pip/_vendor/__init__.py index 8a3471b6..561089cc 100644 --- a/src/fetchcode/vcs/pip/_vendor/__init__.py +++ b/src/fetchcode/vcs/pip/_vendor/__init__.py @@ -33,7 +33,7 @@ def vendored(modulename): __import__(modulename, globals(), locals(), level=0) except ImportError: # We can just silently allow import failures to pass here. If we - # got to this point it means that ``import fetchcode.vcs.pip._vendor.whatever`` + # got to this point it means that ``import pip._vendor.whatever`` # failed and so did ``import whatever``. Since we're importing this # upfront in an attempt to alias imports, not erroring here will # just mean we get a regular import error whenever pip *actually* @@ -58,25 +58,17 @@ def vendored(modulename): sys.path[:] = glob.glob(os.path.join(WHEEL_DIR, "*.whl")) + sys.path # Actually alias all of our vendored dependencies. - vendored("appdirs") vendored("cachecontrol") vendored("certifi") - vendored("colorama") - vendored("contextlib2") vendored("distlib") vendored("distro") - vendored("html5lib") - vendored("six") - vendored("six.moves") - vendored("six.moves.urllib") - vendored("six.moves.urllib.parse") vendored("packaging") vendored("packaging.version") vendored("packaging.specifiers") - vendored("pep517") vendored("pkg_resources") + vendored("platformdirs") vendored("progress") - vendored("retrying") + vendored("pyproject_hooks") vendored("requests") vendored("requests.exceptions") vendored("requests.packages") @@ -108,7 +100,17 @@ def vendored(modulename): vendored("requests.packages.urllib3.util.timeout") vendored("requests.packages.urllib3.util.url") vendored("resolvelib") - vendored("toml") - vendored("toml.encoder") - vendored("toml.decoder") + vendored("rich") + vendored("rich.console") + vendored("rich.highlighter") + vendored("rich.logging") + vendored("rich.markup") + vendored("rich.progress") + vendored("rich.segment") + vendored("rich.style") + vendored("rich.text") + vendored("rich.traceback") + if sys.version_info < (3, 11): + vendored("tomli") + vendored("truststore") vendored("urllib3") diff --git a/src/fetchcode/vcs/pip/_vendor/appdirs.LICENSE.txt b/src/fetchcode/vcs/pip/_vendor/appdirs.LICENSE.txt deleted file mode 100644 index 107c6140..00000000 --- a/src/fetchcode/vcs/pip/_vendor/appdirs.LICENSE.txt +++ /dev/null @@ -1,23 +0,0 @@ -# This is the MIT license - -Copyright (c) 2010 ActiveState Software Inc. - -Permission is hereby granted, free of charge, to any person obtaining a -copy of this software and associated documentation files (the -"Software"), to deal in the Software without restriction, including -without limitation the rights to use, copy, modify, merge, publish, -distribute, sublicense, and/or sell copies of the Software, and to -permit persons to whom the Software is furnished to do so, subject to -the following conditions: - -The above copyright notice and this permission notice shall be included -in all copies or substantial portions of the Software. - -THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS -OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF -MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. -IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY -CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, -TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE -SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. - diff --git a/src/fetchcode/vcs/pip/_vendor/appdirs.py b/src/fetchcode/vcs/pip/_vendor/appdirs.py deleted file mode 100644 index 8bd9c9ca..00000000 --- a/src/fetchcode/vcs/pip/_vendor/appdirs.py +++ /dev/null @@ -1,633 +0,0 @@ -#!/usr/bin/env python -# -*- coding: utf-8 -*- -# Copyright (c) 2005-2010 ActiveState Software Inc. -# Copyright (c) 2013 Eddy Petrișor - -"""Utilities for determining application-specific dirs. - -See for details and usage. -""" -# Dev Notes: -# - MSDN on where to store app data files: -# http://support.microsoft.com/default.aspx?scid=kb;en-us;310294#XSLTH3194121123120121120120 -# - Mac OS X: http://developer.apple.com/documentation/MacOSX/Conceptual/BPFileSystem/index.html -# - XDG spec for Un*x: http://standards.freedesktop.org/basedir-spec/basedir-spec-latest.html - -__version_info__ = (1, 4, 3) -__version__ = '.'.join(map(str, __version_info__)) - - -import sys -import os - -PY3 = sys.version_info[0] == 3 - -if PY3: - unicode = str - -if sys.platform.startswith('java'): - import platform - os_name = platform.java_ver()[3][0] - if os_name.startswith('Windows'): # "Windows XP", "Windows 7", etc. - system = 'win32' - elif os_name.startswith('Mac'): # "Mac OS X", etc. - system = 'darwin' - else: # "Linux", "SunOS", "FreeBSD", etc. - # Setting this to "linux2" is not ideal, but only Windows or Mac - # are actually checked for and the rest of the module expects - # *sys.platform* style strings. - system = 'linux2' -elif sys.platform == 'cli' and os.name == 'nt': - # Detect Windows in IronPython to match pip._internal.utils.compat.WINDOWS - # Discussion: - system = 'win32' -else: - system = sys.platform - - - -def user_data_dir(appname=None, appauthor=None, version=None, roaming=False): - r"""Return full path to the user-specific data dir for this application. - - "appname" is the name of application. - If None, just the system directory is returned. - "appauthor" (only used on Windows) is the name of the - appauthor or distributing body for this application. Typically - it is the owning company name. This falls back to appname. You may - pass False to disable it. - "version" is an optional version path element to append to the - path. You might want to use this if you want multiple versions - of your app to be able to run independently. If used, this - would typically be ".". - Only applied when appname is present. - "roaming" (boolean, default False) can be set True to use the Windows - roaming appdata directory. That means that for users on a Windows - network setup for roaming profiles, this user data will be - sync'd on login. See - - for a discussion of issues. - - Typical user data directories are: - Mac OS X: ~/Library/Application Support/ # or ~/.config/, if the other does not exist - Unix: ~/.local/share/ # or in $XDG_DATA_HOME, if defined - Win XP (not roaming): C:\Documents and Settings\\Application Data\\ - Win XP (roaming): C:\Documents and Settings\\Local Settings\Application Data\\ - Win 7 (not roaming): C:\Users\\AppData\Local\\ - Win 7 (roaming): C:\Users\\AppData\Roaming\\ - - For Unix, we follow the XDG spec and support $XDG_DATA_HOME. - That means, by default "~/.local/share/". - """ - if system == "win32": - if appauthor is None: - appauthor = appname - const = roaming and "CSIDL_APPDATA" or "CSIDL_LOCAL_APPDATA" - path = os.path.normpath(_get_win_folder(const)) - if appname: - if appauthor is not False: - path = os.path.join(path, appauthor, appname) - else: - path = os.path.join(path, appname) - elif system == 'darwin': - path = os.path.expanduser('~/Library/Application Support/') - if appname: - path = os.path.join(path, appname) - else: - path = os.getenv('XDG_DATA_HOME', os.path.expanduser("~/.local/share")) - if appname: - path = os.path.join(path, appname) - if appname and version: - path = os.path.join(path, version) - return path - - -def site_data_dir(appname=None, appauthor=None, version=None, multipath=False): - r"""Return full path to the user-shared data dir for this application. - - "appname" is the name of application. - If None, just the system directory is returned. - "appauthor" (only used on Windows) is the name of the - appauthor or distributing body for this application. Typically - it is the owning company name. This falls back to appname. You may - pass False to disable it. - "version" is an optional version path element to append to the - path. You might want to use this if you want multiple versions - of your app to be able to run independently. If used, this - would typically be ".". - Only applied when appname is present. - "multipath" is an optional parameter only applicable to *nix - which indicates that the entire list of data dirs should be - returned. By default, the first item from XDG_DATA_DIRS is - returned, or '/usr/local/share/', - if XDG_DATA_DIRS is not set - - Typical site data directories are: - Mac OS X: /Library/Application Support/ - Unix: /usr/local/share/ or /usr/share/ - Win XP: C:\Documents and Settings\All Users\Application Data\\ - Vista: (Fail! "C:\ProgramData" is a hidden *system* directory on Vista.) - Win 7: C:\ProgramData\\ # Hidden, but writeable on Win 7. - - For Unix, this is using the $XDG_DATA_DIRS[0] default. - - WARNING: Do not use this on Windows. See the Vista-Fail note above for why. - """ - if system == "win32": - if appauthor is None: - appauthor = appname - path = os.path.normpath(_get_win_folder("CSIDL_COMMON_APPDATA")) - if appname: - if appauthor is not False: - path = os.path.join(path, appauthor, appname) - else: - path = os.path.join(path, appname) - elif system == 'darwin': - path = os.path.expanduser('/Library/Application Support') - if appname: - path = os.path.join(path, appname) - else: - # XDG default for $XDG_DATA_DIRS - # only first, if multipath is False - path = os.getenv('XDG_DATA_DIRS', - os.pathsep.join(['/usr/local/share', '/usr/share'])) - pathlist = [os.path.expanduser(x.rstrip(os.sep)) for x in path.split(os.pathsep)] - if appname: - if version: - appname = os.path.join(appname, version) - pathlist = [os.path.join(x, appname) for x in pathlist] - - if multipath: - path = os.pathsep.join(pathlist) - else: - path = pathlist[0] - return path - - if appname and version: - path = os.path.join(path, version) - return path - - -def user_config_dir(appname=None, appauthor=None, version=None, roaming=False): - r"""Return full path to the user-specific config dir for this application. - - "appname" is the name of application. - If None, just the system directory is returned. - "appauthor" (only used on Windows) is the name of the - appauthor or distributing body for this application. Typically - it is the owning company name. This falls back to appname. You may - pass False to disable it. - "version" is an optional version path element to append to the - path. You might want to use this if you want multiple versions - of your app to be able to run independently. If used, this - would typically be ".". - Only applied when appname is present. - "roaming" (boolean, default False) can be set True to use the Windows - roaming appdata directory. That means that for users on a Windows - network setup for roaming profiles, this user data will be - sync'd on login. See - - for a discussion of issues. - - Typical user config directories are: - Mac OS X: same as user_data_dir - Unix: ~/.config/ # or in $XDG_CONFIG_HOME, if defined - Win *: same as user_data_dir - - For Unix, we follow the XDG spec and support $XDG_CONFIG_HOME. - That means, by default "~/.config/". - """ - if system in ["win32", "darwin"]: - path = user_data_dir(appname, appauthor, None, roaming) - else: - path = os.getenv('XDG_CONFIG_HOME', os.path.expanduser("~/.config")) - if appname: - path = os.path.join(path, appname) - if appname and version: - path = os.path.join(path, version) - return path - - -# for the discussion regarding site_config_dir locations -# see -def site_config_dir(appname=None, appauthor=None, version=None, multipath=False): - r"""Return full path to the user-shared data dir for this application. - - "appname" is the name of application. - If None, just the system directory is returned. - "appauthor" (only used on Windows) is the name of the - appauthor or distributing body for this application. Typically - it is the owning company name. This falls back to appname. You may - pass False to disable it. - "version" is an optional version path element to append to the - path. You might want to use this if you want multiple versions - of your app to be able to run independently. If used, this - would typically be ".". - Only applied when appname is present. - "multipath" is an optional parameter only applicable to *nix - which indicates that the entire list of config dirs should be - returned. By default, the first item from XDG_CONFIG_DIRS is - returned, or '/etc/xdg/', if XDG_CONFIG_DIRS is not set - - Typical site config directories are: - Mac OS X: same as site_data_dir - Unix: /etc/xdg/ or $XDG_CONFIG_DIRS[i]/ for each value in - $XDG_CONFIG_DIRS - Win *: same as site_data_dir - Vista: (Fail! "C:\ProgramData" is a hidden *system* directory on Vista.) - - For Unix, this is using the $XDG_CONFIG_DIRS[0] default, if multipath=False - - WARNING: Do not use this on Windows. See the Vista-Fail note above for why. - """ - if system in ["win32", "darwin"]: - path = site_data_dir(appname, appauthor) - if appname and version: - path = os.path.join(path, version) - else: - # XDG default for $XDG_CONFIG_DIRS (missing or empty) - # see - # only first, if multipath is False - path = os.getenv('XDG_CONFIG_DIRS') or '/etc/xdg' - pathlist = [os.path.expanduser(x.rstrip(os.sep)) for x in path.split(os.pathsep) if x] - if appname: - if version: - appname = os.path.join(appname, version) - pathlist = [os.path.join(x, appname) for x in pathlist] - - if multipath: - path = os.pathsep.join(pathlist) - else: - path = pathlist[0] - return path - - -def user_cache_dir(appname=None, appauthor=None, version=None, opinion=True): - r"""Return full path to the user-specific cache dir for this application. - - "appname" is the name of application. - If None, just the system directory is returned. - "appauthor" (only used on Windows) is the name of the - appauthor or distributing body for this application. Typically - it is the owning company name. This falls back to appname. You may - pass False to disable it. - "version" is an optional version path element to append to the - path. You might want to use this if you want multiple versions - of your app to be able to run independently. If used, this - would typically be ".". - Only applied when appname is present. - "opinion" (boolean) can be False to disable the appending of - "Cache" to the base app data dir for Windows. See - discussion below. - - Typical user cache directories are: - Mac OS X: ~/Library/Caches/ - Unix: ~/.cache/ (XDG default) - Win XP: C:\Documents and Settings\\Local Settings\Application Data\\\Cache - Vista: C:\Users\\AppData\Local\\\Cache - - On Windows the only suggestion in the MSDN docs is that local settings go in - the `CSIDL_LOCAL_APPDATA` directory. This is identical to the non-roaming - app data dir (the default returned by `user_data_dir` above). Apps typically - put cache data somewhere *under* the given dir here. Some examples: - ...\Mozilla\Firefox\Profiles\\Cache - ...\Acme\SuperApp\Cache\1.0 - OPINION: This function appends "Cache" to the `CSIDL_LOCAL_APPDATA` value. - This can be disabled with the `opinion=False` option. - """ - if system == "win32": - if appauthor is None: - appauthor = appname - path = os.path.normpath(_get_win_folder("CSIDL_LOCAL_APPDATA")) - # When using Python 2, return paths as bytes on Windows like we do on - # other operating systems. See helper function docs for more details. - if not PY3 and isinstance(path, unicode): - path = _win_path_to_bytes(path) - if appname: - if appauthor is not False: - path = os.path.join(path, appauthor, appname) - else: - path = os.path.join(path, appname) - if opinion: - path = os.path.join(path, "Cache") - elif system == 'darwin': - path = os.path.expanduser('~/Library/Caches') - if appname: - path = os.path.join(path, appname) - else: - path = os.getenv('XDG_CACHE_HOME', os.path.expanduser('~/.cache')) - if appname: - path = os.path.join(path, appname) - if appname and version: - path = os.path.join(path, version) - return path - - -def user_state_dir(appname=None, appauthor=None, version=None, roaming=False): - r"""Return full path to the user-specific state dir for this application. - - "appname" is the name of application. - If None, just the system directory is returned. - "appauthor" (only used on Windows) is the name of the - appauthor or distributing body for this application. Typically - it is the owning company name. This falls back to appname. You may - pass False to disable it. - "version" is an optional version path element to append to the - path. You might want to use this if you want multiple versions - of your app to be able to run independently. If used, this - would typically be ".". - Only applied when appname is present. - "roaming" (boolean, default False) can be set True to use the Windows - roaming appdata directory. That means that for users on a Windows - network setup for roaming profiles, this user data will be - sync'd on login. See - - for a discussion of issues. - - Typical user state directories are: - Mac OS X: same as user_data_dir - Unix: ~/.local/state/ # or in $XDG_STATE_HOME, if defined - Win *: same as user_data_dir - - For Unix, we follow this Debian proposal - to extend the XDG spec and support $XDG_STATE_HOME. - - That means, by default "~/.local/state/". - """ - if system in ["win32", "darwin"]: - path = user_data_dir(appname, appauthor, None, roaming) - else: - path = os.getenv('XDG_STATE_HOME', os.path.expanduser("~/.local/state")) - if appname: - path = os.path.join(path, appname) - if appname and version: - path = os.path.join(path, version) - return path - - -def user_log_dir(appname=None, appauthor=None, version=None, opinion=True): - r"""Return full path to the user-specific log dir for this application. - - "appname" is the name of application. - If None, just the system directory is returned. - "appauthor" (only used on Windows) is the name of the - appauthor or distributing body for this application. Typically - it is the owning company name. This falls back to appname. You may - pass False to disable it. - "version" is an optional version path element to append to the - path. You might want to use this if you want multiple versions - of your app to be able to run independently. If used, this - would typically be ".". - Only applied when appname is present. - "opinion" (boolean) can be False to disable the appending of - "Logs" to the base app data dir for Windows, and "log" to the - base cache dir for Unix. See discussion below. - - Typical user log directories are: - Mac OS X: ~/Library/Logs/ - Unix: ~/.cache//log # or under $XDG_CACHE_HOME if defined - Win XP: C:\Documents and Settings\\Local Settings\Application Data\\\Logs - Vista: C:\Users\\AppData\Local\\\Logs - - On Windows the only suggestion in the MSDN docs is that local settings - go in the `CSIDL_LOCAL_APPDATA` directory. (Note: I'm interested in - examples of what some windows apps use for a logs dir.) - - OPINION: This function appends "Logs" to the `CSIDL_LOCAL_APPDATA` - value for Windows and appends "log" to the user cache dir for Unix. - This can be disabled with the `opinion=False` option. - """ - if system == "darwin": - path = os.path.join( - os.path.expanduser('~/Library/Logs'), - appname) - elif system == "win32": - path = user_data_dir(appname, appauthor, version) - version = False - if opinion: - path = os.path.join(path, "Logs") - else: - path = user_cache_dir(appname, appauthor, version) - version = False - if opinion: - path = os.path.join(path, "log") - if appname and version: - path = os.path.join(path, version) - return path - - -class AppDirs(object): - """Convenience wrapper for getting application dirs.""" - def __init__(self, appname=None, appauthor=None, version=None, - roaming=False, multipath=False): - self.appname = appname - self.appauthor = appauthor - self.version = version - self.roaming = roaming - self.multipath = multipath - - @property - def user_data_dir(self): - return user_data_dir(self.appname, self.appauthor, - version=self.version, roaming=self.roaming) - - @property - def site_data_dir(self): - return site_data_dir(self.appname, self.appauthor, - version=self.version, multipath=self.multipath) - - @property - def user_config_dir(self): - return user_config_dir(self.appname, self.appauthor, - version=self.version, roaming=self.roaming) - - @property - def site_config_dir(self): - return site_config_dir(self.appname, self.appauthor, - version=self.version, multipath=self.multipath) - - @property - def user_cache_dir(self): - return user_cache_dir(self.appname, self.appauthor, - version=self.version) - - @property - def user_state_dir(self): - return user_state_dir(self.appname, self.appauthor, - version=self.version) - - @property - def user_log_dir(self): - return user_log_dir(self.appname, self.appauthor, - version=self.version) - - -#---- internal support stuff - -def _get_win_folder_from_registry(csidl_name): - """This is a fallback technique at best. I'm not sure if using the - registry for this guarantees us the correct answer for all CSIDL_* - names. - """ - if PY3: - import winreg as _winreg - else: - import _winreg - - shell_folder_name = { - "CSIDL_APPDATA": "AppData", - "CSIDL_COMMON_APPDATA": "Common AppData", - "CSIDL_LOCAL_APPDATA": "Local AppData", - }[csidl_name] - - key = _winreg.OpenKey( - _winreg.HKEY_CURRENT_USER, - r"Software\Microsoft\Windows\CurrentVersion\Explorer\Shell Folders" - ) - dir, type = _winreg.QueryValueEx(key, shell_folder_name) - return dir - - -def _get_win_folder_with_pywin32(csidl_name): - from win32com.shell import shellcon, shell - dir = shell.SHGetFolderPath(0, getattr(shellcon, csidl_name), 0, 0) - # Try to make this a unicode path because SHGetFolderPath does - # not return unicode strings when there is unicode data in the - # path. - try: - dir = unicode(dir) - - # Downgrade to short path name if have highbit chars. See - # . - has_high_char = False - for c in dir: - if ord(c) > 255: - has_high_char = True - break - if has_high_char: - try: - import win32api - dir = win32api.GetShortPathName(dir) - except ImportError: - pass - except UnicodeError: - pass - return dir - - -def _get_win_folder_with_ctypes(csidl_name): - import ctypes - - csidl_const = { - "CSIDL_APPDATA": 26, - "CSIDL_COMMON_APPDATA": 35, - "CSIDL_LOCAL_APPDATA": 28, - }[csidl_name] - - buf = ctypes.create_unicode_buffer(1024) - ctypes.windll.shell32.SHGetFolderPathW(None, csidl_const, None, 0, buf) - - # Downgrade to short path name if have highbit chars. See - # . - has_high_char = False - for c in buf: - if ord(c) > 255: - has_high_char = True - break - if has_high_char: - buf2 = ctypes.create_unicode_buffer(1024) - if ctypes.windll.kernel32.GetShortPathNameW(buf.value, buf2, 1024): - buf = buf2 - - return buf.value - -def _get_win_folder_with_jna(csidl_name): - import array - from com.sun import jna - from com.sun.jna.platform import win32 - - buf_size = win32.WinDef.MAX_PATH * 2 - buf = array.zeros('c', buf_size) - shell = win32.Shell32.INSTANCE - shell.SHGetFolderPath(None, getattr(win32.ShlObj, csidl_name), None, win32.ShlObj.SHGFP_TYPE_CURRENT, buf) - dir = jna.Native.toString(buf.tostring()).rstrip("\0") - - # Downgrade to short path name if have highbit chars. See - # . - has_high_char = False - for c in dir: - if ord(c) > 255: - has_high_char = True - break - if has_high_char: - buf = array.zeros('c', buf_size) - kernel = win32.Kernel32.INSTANCE - if kernel.GetShortPathName(dir, buf, buf_size): - dir = jna.Native.toString(buf.tostring()).rstrip("\0") - - return dir - -if system == "win32": - try: - from ctypes import windll - _get_win_folder = _get_win_folder_with_ctypes - except ImportError: - try: - import com.sun.jna - _get_win_folder = _get_win_folder_with_jna - except ImportError: - _get_win_folder = _get_win_folder_from_registry - - -def _win_path_to_bytes(path): - """Encode Windows paths to bytes. Only used on Python 2. - - Motivation is to be consistent with other operating systems where paths - are also returned as bytes. This avoids problems mixing bytes and Unicode - elsewhere in the codebase. For more details and discussion see - . - - If encoding using ASCII and MBCS fails, return the original Unicode path. - """ - for encoding in ('ASCII', 'MBCS'): - try: - return path.encode(encoding) - except (UnicodeEncodeError, LookupError): - pass - return path - - -#---- self test code - -if __name__ == "__main__": - appname = "MyApp" - appauthor = "MyCompany" - - props = ("user_data_dir", - "user_config_dir", - "user_cache_dir", - "user_state_dir", - "user_log_dir", - "site_data_dir", - "site_config_dir") - - print("-- app dirs %s --" % __version__) - - print("-- app dirs (with optional 'version')") - dirs = AppDirs(appname, appauthor, version="1.0") - for prop in props: - print("%s: %s" % (prop, getattr(dirs, prop))) - - print("\n-- app dirs (without optional 'version')") - dirs = AppDirs(appname, appauthor) - for prop in props: - print("%s: %s" % (prop, getattr(dirs, prop))) - - print("\n-- app dirs (without optional 'appauthor')") - dirs = AppDirs(appname) - for prop in props: - print("%s: %s" % (prop, getattr(dirs, prop))) - - print("\n-- app dirs (with disabled 'appauthor')") - dirs = AppDirs(appname, appauthor=False) - for prop in props: - print("%s: %s" % (prop, getattr(dirs, prop))) diff --git a/src/fetchcode/vcs/pip/_vendor/cachecontrol.pyi b/src/fetchcode/vcs/pip/_vendor/cachecontrol.pyi deleted file mode 100644 index 636a66ba..00000000 --- a/src/fetchcode/vcs/pip/_vendor/cachecontrol.pyi +++ /dev/null @@ -1 +0,0 @@ -from cachecontrol import * \ No newline at end of file diff --git a/src/fetchcode/vcs/pip/_vendor/cachecontrol/LICENSE.txt b/src/fetchcode/vcs/pip/_vendor/cachecontrol/LICENSE.txt index 1ed31ac3..d8b3b56d 100644 --- a/src/fetchcode/vcs/pip/_vendor/cachecontrol/LICENSE.txt +++ b/src/fetchcode/vcs/pip/_vendor/cachecontrol/LICENSE.txt @@ -1,4 +1,4 @@ -Copyright 2015 Eric Larson +Copyright 2012-2021 Eric Larson Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. @@ -8,8 +8,6 @@ You may obtain a copy of the License at Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or -implied. - +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. diff --git a/src/fetchcode/vcs/pip/_vendor/cachecontrol/__init__.py b/src/fetchcode/vcs/pip/_vendor/cachecontrol/__init__.py index a1bbbbe3..b34b0fcb 100644 --- a/src/fetchcode/vcs/pip/_vendor/cachecontrol/__init__.py +++ b/src/fetchcode/vcs/pip/_vendor/cachecontrol/__init__.py @@ -1,11 +1,28 @@ +# SPDX-FileCopyrightText: 2015 Eric Larson +# +# SPDX-License-Identifier: Apache-2.0 + """CacheControl import Interface. Make it easy to import from cachecontrol without long namespaces. """ __author__ = "Eric Larson" __email__ = "eric@ionrock.org" -__version__ = "0.12.6" +__version__ = "0.14.0" + +from pip._vendor.cachecontrol.adapter import CacheControlAdapter +from pip._vendor.cachecontrol.controller import CacheController +from pip._vendor.cachecontrol.wrapper import CacheControl + +__all__ = [ + "__author__", + "__email__", + "__version__", + "CacheControlAdapter", + "CacheController", + "CacheControl", +] + +import logging -from .wrapper import CacheControl -from .adapter import CacheControlAdapter -from .controller import CacheController +logging.getLogger(__name__).addHandler(logging.NullHandler()) diff --git a/src/fetchcode/vcs/pip/_vendor/cachecontrol/_cmd.py b/src/fetchcode/vcs/pip/_vendor/cachecontrol/_cmd.py index 392a5736..2c84208a 100644 --- a/src/fetchcode/vcs/pip/_vendor/cachecontrol/_cmd.py +++ b/src/fetchcode/vcs/pip/_vendor/cachecontrol/_cmd.py @@ -1,21 +1,31 @@ +# SPDX-FileCopyrightText: 2015 Eric Larson +# +# SPDX-License-Identifier: Apache-2.0 +from __future__ import annotations + import logging +from argparse import ArgumentParser +from typing import TYPE_CHECKING -from fetchcode.vcs.pip._vendor import requests +from pip._vendor import requests -from fetchcode.vcs.pip._vendor.cachecontrol.adapter import CacheControlAdapter -from fetchcode.vcs.pip._vendor.cachecontrol.cache import DictCache -from fetchcode.vcs.pip._vendor.cachecontrol.controller import logger +from pip._vendor.cachecontrol.adapter import CacheControlAdapter +from pip._vendor.cachecontrol.cache import DictCache +from pip._vendor.cachecontrol.controller import logger -from argparse import ArgumentParser +if TYPE_CHECKING: + from argparse import Namespace + from pip._vendor.cachecontrol.controller import CacheController -def setup_logging(): + +def setup_logging() -> None: logger.setLevel(logging.DEBUG) handler = logging.StreamHandler() logger.addHandler(handler) -def get_session(): +def get_session() -> requests.Session: adapter = CacheControlAdapter( DictCache(), cache_etags=True, serializer=None, heuristic=None ) @@ -23,17 +33,17 @@ def get_session(): sess.mount("http://", adapter) sess.mount("https://", adapter) - sess.cache_controller = adapter.controller + sess.cache_controller = adapter.controller # type: ignore[attr-defined] return sess -def get_args(): +def get_args() -> Namespace: parser = ArgumentParser() parser.add_argument("url", help="The URL to try and cache") return parser.parse_args() -def main(args=None): +def main() -> None: args = get_args() sess = get_session() @@ -44,10 +54,13 @@ def main(args=None): setup_logging() # try setting the cache - sess.cache_controller.cache_response(resp.request, resp.raw) + cache_controller: CacheController = ( + sess.cache_controller # type: ignore[attr-defined] + ) + cache_controller.cache_response(resp.request, resp.raw) # Now try to get it - if sess.cache_controller.cached_request(resp.request): + if cache_controller.cached_request(resp.request): print("Cached!") else: print("Not cached :(") diff --git a/src/fetchcode/vcs/pip/_vendor/cachecontrol/adapter.py b/src/fetchcode/vcs/pip/_vendor/cachecontrol/adapter.py index 9c35c105..fbb4ecc8 100644 --- a/src/fetchcode/vcs/pip/_vendor/cachecontrol/adapter.py +++ b/src/fetchcode/vcs/pip/_vendor/cachecontrol/adapter.py @@ -1,29 +1,43 @@ -import types +# SPDX-FileCopyrightText: 2015 Eric Larson +# +# SPDX-License-Identifier: Apache-2.0 +from __future__ import annotations + import functools +import types import zlib +from typing import TYPE_CHECKING, Any, Collection, Mapping + +from pip._vendor.requests.adapters import HTTPAdapter -from fetchcode.vcs.pip._vendor.requests.adapters import HTTPAdapter +from pip._vendor.cachecontrol.cache import DictCache +from pip._vendor.cachecontrol.controller import PERMANENT_REDIRECT_STATUSES, CacheController +from pip._vendor.cachecontrol.filewrapper import CallbackFileWrapper -from .controller import CacheController -from .cache import DictCache -from .filewrapper import CallbackFileWrapper +if TYPE_CHECKING: + from pip._vendor.requests import PreparedRequest, Response + from pip._vendor.urllib3 import HTTPResponse + + from pip._vendor.cachecontrol.cache import BaseCache + from pip._vendor.cachecontrol.heuristics import BaseHeuristic + from pip._vendor.cachecontrol.serialize import Serializer class CacheControlAdapter(HTTPAdapter): - invalidating_methods = {"PUT", "DELETE"} + invalidating_methods = {"PUT", "PATCH", "DELETE"} def __init__( self, - cache=None, - cache_etags=True, - controller_class=None, - serializer=None, - heuristic=None, - cacheable_methods=None, - *args, - **kw - ): - super(CacheControlAdapter, self).__init__(*args, **kw) + cache: BaseCache | None = None, + cache_etags: bool = True, + controller_class: type[CacheController] | None = None, + serializer: Serializer | None = None, + heuristic: BaseHeuristic | None = None, + cacheable_methods: Collection[str] | None = None, + *args: Any, + **kw: Any, + ) -> None: + super().__init__(*args, **kw) self.cache = DictCache() if cache is None else cache self.heuristic = heuristic self.cacheable_methods = cacheable_methods or ("GET",) @@ -33,7 +47,16 @@ def __init__( self.cache, cache_etags=cache_etags, serializer=serializer ) - def send(self, request, cacheable_methods=None, **kw): + def send( + self, + request: PreparedRequest, + stream: bool = False, + timeout: None | float | tuple[float, float] | tuple[float, None] = None, + verify: bool | str = True, + cert: (None | bytes | str | tuple[bytes | str, bytes | str]) = None, + proxies: Mapping[str, str] | None = None, + cacheable_methods: Collection[str] | None = None, + ) -> Response: """ Send a request. Use the request information to see if it exists in the cache and cache the response if we need to and can. @@ -50,13 +73,17 @@ def send(self, request, cacheable_methods=None, **kw): # check for etags and add headers if appropriate request.headers.update(self.controller.conditional_headers(request)) - resp = super(CacheControlAdapter, self).send(request, **kw) + resp = super().send(request, stream, timeout, verify, cert, proxies) return resp def build_response( - self, request, response, from_cache=False, cacheable_methods=None - ): + self, + request: PreparedRequest, + response: HTTPResponse, + from_cache: bool = False, + cacheable_methods: Collection[str] | None = None, + ) -> Response: """ Build a response by making a request or using the cache. @@ -93,13 +120,13 @@ def build_response( response = cached_response # We always cache the 301 responses - elif response.status == 301: + elif int(response.status) in PERMANENT_REDIRECT_STATUSES: self.controller.cache_response(request, response) else: # Wrap the response file with a wrapper that will cache the # response when the stream has been consumed. - response._fp = CallbackFileWrapper( - response._fp, + response._fp = CallbackFileWrapper( # type: ignore[assignment] + response._fp, # type: ignore[arg-type] functools.partial( self.controller.cache_response, request, response ), @@ -107,27 +134,28 @@ def build_response( if response.chunked: super_update_chunk_length = response._update_chunk_length - def _update_chunk_length(self): + def _update_chunk_length(self: HTTPResponse) -> None: super_update_chunk_length() if self.chunk_left == 0: - self._fp._close() + self._fp._close() # type: ignore[union-attr] - response._update_chunk_length = types.MethodType( + response._update_chunk_length = types.MethodType( # type: ignore[method-assign] _update_chunk_length, response ) - resp = super(CacheControlAdapter, self).build_response(request, response) + resp: Response = super().build_response(request, response) # type: ignore[no-untyped-call] # See if we should invalidate the cache. if request.method in self.invalidating_methods and resp.ok: + assert request.url is not None cache_url = self.controller.cache_url(request.url) self.cache.delete(cache_url) # Give the request a from_cache attr to let people use it - resp.from_cache = from_cache + resp.from_cache = from_cache # type: ignore[attr-defined] return resp - def close(self): + def close(self) -> None: self.cache.close() - super(CacheControlAdapter, self).close() + super().close() # type: ignore[no-untyped-call] diff --git a/src/fetchcode/vcs/pip/_vendor/cachecontrol/cache.py b/src/fetchcode/vcs/pip/_vendor/cachecontrol/cache.py index 94e07732..3293b005 100644 --- a/src/fetchcode/vcs/pip/_vendor/cachecontrol/cache.py +++ b/src/fetchcode/vcs/pip/_vendor/cachecontrol/cache.py @@ -1,39 +1,74 @@ +# SPDX-FileCopyrightText: 2015 Eric Larson +# +# SPDX-License-Identifier: Apache-2.0 + """ The cache object API for implementing caches. The default is a thread safe in-memory dictionary. """ +from __future__ import annotations + from threading import Lock +from typing import IO, TYPE_CHECKING, MutableMapping +if TYPE_CHECKING: + from datetime import datetime -class BaseCache(object): - def get(self, key): +class BaseCache: + def get(self, key: str) -> bytes | None: raise NotImplementedError() - def set(self, key, value): + def set( + self, key: str, value: bytes, expires: int | datetime | None = None + ) -> None: raise NotImplementedError() - def delete(self, key): + def delete(self, key: str) -> None: raise NotImplementedError() - def close(self): + def close(self) -> None: pass class DictCache(BaseCache): - - def __init__(self, init_dict=None): + def __init__(self, init_dict: MutableMapping[str, bytes] | None = None) -> None: self.lock = Lock() self.data = init_dict or {} - def get(self, key): + def get(self, key: str) -> bytes | None: return self.data.get(key, None) - def set(self, key, value): + def set( + self, key: str, value: bytes, expires: int | datetime | None = None + ) -> None: with self.lock: self.data.update({key: value}) - def delete(self, key): + def delete(self, key: str) -> None: with self.lock: if key in self.data: self.data.pop(key) + + +class SeparateBodyBaseCache(BaseCache): + """ + In this variant, the body is not stored mixed in with the metadata, but is + passed in (as a bytes-like object) in a separate call to ``set_body()``. + + That is, the expected interaction pattern is:: + + cache.set(key, serialized_metadata) + cache.set_body(key) + + Similarly, the body should be loaded separately via ``get_body()``. + """ + + def set_body(self, key: str, body: bytes) -> None: + raise NotImplementedError() + + def get_body(self, key: str) -> IO[bytes] | None: + """ + Return the body as file-like object. + """ + raise NotImplementedError() diff --git a/src/fetchcode/vcs/pip/_vendor/cachecontrol/caches/__init__.py b/src/fetchcode/vcs/pip/_vendor/cachecontrol/caches/__init__.py index 0e1658fa..24ff469f 100644 --- a/src/fetchcode/vcs/pip/_vendor/cachecontrol/caches/__init__.py +++ b/src/fetchcode/vcs/pip/_vendor/cachecontrol/caches/__init__.py @@ -1,2 +1,8 @@ -from .file_cache import FileCache # noqa -from .redis_cache import RedisCache # noqa +# SPDX-FileCopyrightText: 2015 Eric Larson +# +# SPDX-License-Identifier: Apache-2.0 + +from pip._vendor.cachecontrol.caches.file_cache import FileCache, SeparateBodyFileCache +from pip._vendor.cachecontrol.caches.redis_cache import RedisCache + +__all__ = ["FileCache", "SeparateBodyFileCache", "RedisCache"] diff --git a/src/fetchcode/vcs/pip/_vendor/cachecontrol/caches/file_cache.py b/src/fetchcode/vcs/pip/_vendor/cachecontrol/caches/file_cache.py index 607b9452..e6e3a579 100644 --- a/src/fetchcode/vcs/pip/_vendor/cachecontrol/caches/file_cache.py +++ b/src/fetchcode/vcs/pip/_vendor/cachecontrol/caches/file_cache.py @@ -1,18 +1,24 @@ +# SPDX-FileCopyrightText: 2015 Eric Larson +# +# SPDX-License-Identifier: Apache-2.0 +from __future__ import annotations + import hashlib import os from textwrap import dedent +from typing import IO, TYPE_CHECKING, Union +from pathlib import Path + +from pip._vendor.cachecontrol.cache import BaseCache, SeparateBodyBaseCache +from pip._vendor.cachecontrol.controller import CacheController -from ..cache import BaseCache -from ..controller import CacheController +if TYPE_CHECKING: + from datetime import datetime -try: - FileNotFoundError -except NameError: - # py2.X - FileNotFoundError = (IOError, OSError) + from filelock import BaseFileLock -def _secure_open_write(filename, fmode): +def _secure_open_write(filename: str, fmode: int) -> IO[bytes]: # We only want to write to this file, so open it in write only mode flags = os.O_WRONLY @@ -35,7 +41,7 @@ def _secure_open_write(filename, fmode): # there try: os.remove(filename) - except (IOError, OSError): + except OSError: # The file must not exist already, so we can just skip ahead to opening pass @@ -53,41 +59,32 @@ def _secure_open_write(filename, fmode): raise -class FileCache(BaseCache): +class _FileCacheMixin: + """Shared implementation for both FileCache variants.""" def __init__( self, - directory, - forever=False, - filemode=0o0600, - dirmode=0o0700, - use_dir_lock=None, - lock_class=None, - ): - - if use_dir_lock is not None and lock_class is not None: - raise ValueError("Cannot use use_dir_lock and lock_class together") - + directory: str | Path, + forever: bool = False, + filemode: int = 0o0600, + dirmode: int = 0o0700, + lock_class: type[BaseFileLock] | None = None, + ) -> None: try: - from lockfile import LockFile - from lockfile.mkdirlockfile import MkdirLockFile + if lock_class is None: + from filelock import FileLock + + lock_class = FileLock except ImportError: notice = dedent( """ NOTE: In order to use the FileCache you must have - lockfile installed. You can install it via pip: - pip install lockfile + filelock installed. You can install it via pip: + pip install cachecontrol[filecache] """ ) raise ImportError(notice) - else: - if use_dir_lock: - lock_class = MkdirLockFile - - elif lock_class is None: - lock_class = LockFile - self.directory = directory self.forever = forever self.filemode = filemode @@ -95,17 +92,17 @@ def __init__( self.lock_class = lock_class @staticmethod - def encode(x): + def encode(x: str) -> str: return hashlib.sha224(x.encode()).hexdigest() - def _fn(self, name): + def _fn(self, name: str) -> str: # NOTE: This method should not change as some may depend on it. # See: https://github.com/ionrock/cachecontrol/issues/63 hashed = self.encode(name) parts = list(hashed[:5]) + [hashed] return os.path.join(self.directory, *parts) - def get(self, key): + def get(self, key: str) -> bytes | None: name = self._fn(key) try: with open(name, "rb") as fh: @@ -114,22 +111,29 @@ def get(self, key): except FileNotFoundError: return None - def set(self, key, value): + def set( + self, key: str, value: bytes, expires: int | datetime | None = None + ) -> None: name = self._fn(key) + self._write(name, value) + def _write(self, path: str, data: bytes) -> None: + """ + Safely write the data to the given path. + """ # Make sure the directory exists try: - os.makedirs(os.path.dirname(name), self.dirmode) - except (IOError, OSError): + os.makedirs(os.path.dirname(path), self.dirmode) + except OSError: pass - with self.lock_class(name) as lock: + with self.lock_class(path + ".lock"): # Write our actual file - with _secure_open_write(lock.path, self.filemode) as fh: - fh.write(value) + with _secure_open_write(path, self.filemode) as fh: + fh.write(data) - def delete(self, key): - name = self._fn(key) + def _delete(self, key: str, suffix: str) -> None: + name = self._fn(key) + suffix if not self.forever: try: os.remove(name) @@ -137,7 +141,39 @@ def delete(self, key): pass -def url_to_file_path(url, filecache): +class FileCache(_FileCacheMixin, BaseCache): + """ + Traditional FileCache: body is stored in memory, so not suitable for large + downloads. + """ + + def delete(self, key: str) -> None: + self._delete(key, "") + + +class SeparateBodyFileCache(_FileCacheMixin, SeparateBodyBaseCache): + """ + Memory-efficient FileCache: body is stored in a separate file, reducing + peak memory usage. + """ + + def get_body(self, key: str) -> IO[bytes] | None: + name = self._fn(key) + ".body" + try: + return open(name, "rb") + except FileNotFoundError: + return None + + def set_body(self, key: str, body: bytes) -> None: + name = self._fn(key) + ".body" + self._write(name, body) + + def delete(self, key: str) -> None: + self._delete(key, "") + self._delete(key, ".body") + + +def url_to_file_path(url: str, filecache: FileCache) -> str: """Return the file cache path based on the URL. This does not ensure the file exists! diff --git a/src/fetchcode/vcs/pip/_vendor/cachecontrol/caches/redis_cache.py b/src/fetchcode/vcs/pip/_vendor/cachecontrol/caches/redis_cache.py index 60b13244..f4f68c47 100644 --- a/src/fetchcode/vcs/pip/_vendor/cachecontrol/caches/redis_cache.py +++ b/src/fetchcode/vcs/pip/_vendor/cachecontrol/caches/redis_cache.py @@ -1,33 +1,48 @@ -from __future__ import division +# SPDX-FileCopyrightText: 2015 Eric Larson +# +# SPDX-License-Identifier: Apache-2.0 +from __future__ import annotations -from datetime import datetime -from fetchcode.vcs.pip._vendor.cachecontrol.cache import BaseCache +from datetime import datetime, timezone +from typing import TYPE_CHECKING -class RedisCache(BaseCache): +from pip._vendor.cachecontrol.cache import BaseCache + +if TYPE_CHECKING: + from redis import Redis - def __init__(self, conn): + +class RedisCache(BaseCache): + def __init__(self, conn: Redis[bytes]) -> None: self.conn = conn - def get(self, key): + def get(self, key: str) -> bytes | None: return self.conn.get(key) - def set(self, key, value, expires=None): + def set( + self, key: str, value: bytes, expires: int | datetime | None = None + ) -> None: if not expires: self.conn.set(key, value) + elif isinstance(expires, datetime): + now_utc = datetime.now(timezone.utc) + if expires.tzinfo is None: + now_utc = now_utc.replace(tzinfo=None) + delta = expires - now_utc + self.conn.setex(key, int(delta.total_seconds()), value) else: - expires = expires - datetime.utcnow() - self.conn.setex(key, int(expires.total_seconds()), value) + self.conn.setex(key, expires, value) - def delete(self, key): + def delete(self, key: str) -> None: self.conn.delete(key) - def clear(self): + def clear(self) -> None: """Helper for clearing all the keys in a database. Use with caution!""" for key in self.conn.keys(): self.conn.delete(key) - def close(self): + def close(self) -> None: """Redis uses connection pooling, no need to close the connection.""" pass diff --git a/src/fetchcode/vcs/pip/_vendor/cachecontrol/compat.py b/src/fetchcode/vcs/pip/_vendor/cachecontrol/compat.py deleted file mode 100644 index 3c8fd2b7..00000000 --- a/src/fetchcode/vcs/pip/_vendor/cachecontrol/compat.py +++ /dev/null @@ -1,29 +0,0 @@ -try: - from urllib.parse import urljoin -except ImportError: - from urlparse import urljoin - - -try: - import cPickle as pickle -except ImportError: - import pickle - - -# Handle the case where the requests module has been patched to not have -# urllib3 bundled as part of its source. -try: - from fetchcode.vcs.pip._vendor.requests.packages.urllib3.response import HTTPResponse -except ImportError: - from fetchcode.vcs.pip._vendor.urllib3.response import HTTPResponse - -try: - from fetchcode.vcs.pip._vendor.requests.packages.urllib3.util import is_fp_closed -except ImportError: - from fetchcode.vcs.pip._vendor.urllib3.util import is_fp_closed - -# Replicate some six behaviour -try: - text_type = unicode -except NameError: - text_type = str diff --git a/src/fetchcode/vcs/pip/_vendor/cachecontrol/controller.py b/src/fetchcode/vcs/pip/_vendor/cachecontrol/controller.py index 1ecab2a0..d7dd86e5 100644 --- a/src/fetchcode/vcs/pip/_vendor/cachecontrol/controller.py +++ b/src/fetchcode/vcs/pip/_vendor/cachecontrol/controller.py @@ -1,46 +1,67 @@ +# SPDX-FileCopyrightText: 2015 Eric Larson +# +# SPDX-License-Identifier: Apache-2.0 + """ The httplib2 algorithms ported for use with requests. """ +from __future__ import annotations + +import calendar import logging import re -import calendar import time from email.utils import parsedate_tz +from typing import TYPE_CHECKING, Collection, Mapping + +from pip._vendor.requests.structures import CaseInsensitiveDict + +from pip._vendor.cachecontrol.cache import DictCache, SeparateBodyBaseCache +from pip._vendor.cachecontrol.serialize import Serializer -from fetchcode.vcs.pip._vendor.requests.structures import CaseInsensitiveDict +if TYPE_CHECKING: + from typing import Literal -from .cache import DictCache -from .serialize import Serializer + from pip._vendor.requests import PreparedRequest + from pip._vendor.urllib3 import HTTPResponse + from pip._vendor.cachecontrol.cache import BaseCache logger = logging.getLogger(__name__) URI = re.compile(r"^(([^:/?#]+):)?(//([^/?#]*))?([^?#]*)(\?([^#]*))?(#(.*))?") +PERMANENT_REDIRECT_STATUSES = (301, 308) -def parse_uri(uri): + +def parse_uri(uri: str) -> tuple[str, str, str, str, str]: """Parses a URI using the regex given in Appendix B of RFC 3986. - (scheme, authority, path, query, fragment) = parse_uri(uri) + (scheme, authority, path, query, fragment) = parse_uri(uri) """ - groups = URI.match(uri).groups() + match = URI.match(uri) + assert match is not None + groups = match.groups() return (groups[1], groups[3], groups[4], groups[6], groups[8]) -class CacheController(object): - """An interface to see if request should cached or not. - """ +class CacheController: + """An interface to see if request should cached or not.""" def __init__( - self, cache=None, cache_etags=True, serializer=None, status_codes=None + self, + cache: BaseCache | None = None, + cache_etags: bool = True, + serializer: Serializer | None = None, + status_codes: Collection[int] | None = None, ): self.cache = DictCache() if cache is None else cache self.cache_etags = cache_etags self.serializer = serializer or Serializer() - self.cacheable_status_codes = status_codes or (200, 203, 300, 301) + self.cacheable_status_codes = status_codes or (200, 203, 300, 301, 308) @classmethod - def _urlnorm(cls, uri): + def _urlnorm(cls, uri: str) -> str: """Normalize the URL to create a safe key for the cache""" (scheme, authority, path, query, fragment) = parse_uri(uri) if not scheme or not authority: @@ -60,10 +81,10 @@ def _urlnorm(cls, uri): return defrag_uri @classmethod - def cache_url(cls, uri): + def cache_url(cls, uri: str) -> str: return cls._urlnorm(uri) - def parse_cache_control(self, headers): + def parse_cache_control(self, headers: Mapping[str, str]) -> dict[str, int | None]: known_directives = { # https://tools.ietf.org/html/rfc7234#section-5.2 "max-age": (int, True), @@ -82,7 +103,7 @@ def parse_cache_control(self, headers): cc_headers = headers.get("cache-control", headers.get("Cache-Control", "")) - retval = {} + retval: dict[str, int | None] = {} for cc_directive in cc_headers.split(","): if not cc_directive.strip(): @@ -117,11 +138,38 @@ def parse_cache_control(self, headers): return retval - def cached_request(self, request): + def _load_from_cache(self, request: PreparedRequest) -> HTTPResponse | None: + """ + Load a cached response, or return None if it's not available. + """ + # We do not support caching of partial content: so if the request contains a + # Range header then we don't want to load anything from the cache. + if "Range" in request.headers: + return None + + cache_url = request.url + assert cache_url is not None + cache_data = self.cache.get(cache_url) + if cache_data is None: + logger.debug("No cache entry available") + return None + + if isinstance(self.cache, SeparateBodyBaseCache): + body_file = self.cache.get_body(cache_url) + else: + body_file = None + + result = self.serializer.loads(request, cache_data, body_file) + if result is None: + logger.warning("Cache entry deserialization failed, entry ignored") + return result + + def cached_request(self, request: PreparedRequest) -> HTTPResponse | Literal[False]: """ Return a cached response if it exists in the cache, otherwise return False. """ + assert request.url is not None cache_url = self.cache_url(request.url) logger.debug('Looking up "%s" in the cache', cache_url) cc = self.parse_cache_control(request.headers) @@ -135,35 +183,29 @@ def cached_request(self, request): logger.debug('Request header has "max_age" as 0, cache bypassed') return False - # Request allows serving from the cache, let's see if we find something - cache_data = self.cache.get(cache_url) - if cache_data is None: - logger.debug("No cache entry available") - return False - - # Check whether it can be deserialized - resp = self.serializer.loads(request, cache_data) + # Check whether we can load the response from the cache: + resp = self._load_from_cache(request) if not resp: - logger.warning("Cache entry deserialization failed, entry ignored") return False - # If we have a cached 301, return it immediately. We don't - # need to test our response for other headers b/c it is + # If we have a cached permanent redirect, return it immediately. We + # don't need to test our response for other headers b/c it is # intrinsically "cacheable" as it is Permanent. + # # See: # https://tools.ietf.org/html/rfc7231#section-6.4.2 # # Client can try to refresh the value by repeating the request # with cache busting headers as usual (ie no-cache). - if resp.status == 301: + if int(resp.status) in PERMANENT_REDIRECT_STATUSES: msg = ( - 'Returning cached "301 Moved Permanently" response ' + "Returning cached permanent redirect response " "(ignoring date and etag information)" ) logger.debug(msg) return resp - headers = CaseInsensitiveDict(resp.headers) + headers: CaseInsensitiveDict[str] = CaseInsensitiveDict(resp.headers) if not headers or "date" not in headers: if "etag" not in headers: # Without date or etag, the cached response can never be used @@ -174,7 +216,9 @@ def cached_request(self, request): return False now = time.time() - date = calendar.timegm(parsedate_tz(headers["date"])) + time_tuple = parsedate_tz(headers["date"]) + assert time_tuple is not None + date = calendar.timegm(time_tuple[:6]) current_age = max(0, now - date) logger.debug("Current age based on date: %i", current_age) @@ -188,28 +232,30 @@ def cached_request(self, request): freshness_lifetime = 0 # Check the max-age pragma in the cache control header - if "max-age" in resp_cc: - freshness_lifetime = resp_cc["max-age"] + max_age = resp_cc.get("max-age") + if max_age is not None: + freshness_lifetime = max_age logger.debug("Freshness lifetime from max-age: %i", freshness_lifetime) # If there isn't a max-age, check for an expires header elif "expires" in headers: expires = parsedate_tz(headers["expires"]) if expires is not None: - expire_time = calendar.timegm(expires) - date + expire_time = calendar.timegm(expires[:6]) - date freshness_lifetime = max(0, expire_time) logger.debug("Freshness lifetime from expires: %i", freshness_lifetime) # Determine if we are setting freshness limit in the # request. Note, this overrides what was in the response. - if "max-age" in cc: - freshness_lifetime = cc["max-age"] + max_age = cc.get("max-age") + if max_age is not None: + freshness_lifetime = max_age logger.debug( "Freshness lifetime from request max-age: %i", freshness_lifetime ) - if "min-fresh" in cc: - min_fresh = cc["min-fresh"] + min_fresh = cc.get("min-fresh") + if min_fresh is not None: # adjust our current age by our min fresh current_age += min_fresh logger.debug("Adjusted current age from min-fresh: %i", current_age) @@ -228,13 +274,12 @@ def cached_request(self, request): # return the original handler return False - def conditional_headers(self, request): - cache_url = self.cache_url(request.url) - resp = self.serializer.loads(request, self.cache.get(cache_url)) + def conditional_headers(self, request: PreparedRequest) -> dict[str, str]: + resp = self._load_from_cache(request) new_headers = {} if resp: - headers = CaseInsensitiveDict(resp.headers) + headers: CaseInsensitiveDict[str] = CaseInsensitiveDict(resp.headers) if "etag" in headers: new_headers["If-None-Match"] = headers["ETag"] @@ -244,7 +289,43 @@ def conditional_headers(self, request): return new_headers - def cache_response(self, request, response, body=None, status_codes=None): + def _cache_set( + self, + cache_url: str, + request: PreparedRequest, + response: HTTPResponse, + body: bytes | None = None, + expires_time: int | None = None, + ) -> None: + """ + Store the data in the cache. + """ + if isinstance(self.cache, SeparateBodyBaseCache): + # We pass in the body separately; just put a placeholder empty + # string in the metadata. + self.cache.set( + cache_url, + self.serializer.dumps(request, response, b""), + expires=expires_time, + ) + # body is None can happen when, for example, we're only updating + # headers, as is the case in update_cached_response(). + if body is not None: + self.cache.set_body(cache_url, body) + else: + self.cache.set( + cache_url, + self.serializer.dumps(request, response, body), + expires=expires_time, + ) + + def cache_response( + self, + request: PreparedRequest, + response: HTTPResponse, + body: bytes | None = None, + status_codes: Collection[int] | None = None, + ) -> None: """ Algorithm for caching requests. @@ -259,7 +340,16 @@ def cache_response(self, request, response, body=None, status_codes=None): ) return - response_headers = CaseInsensitiveDict(response.headers) + response_headers: CaseInsensitiveDict[str] = CaseInsensitiveDict( + response.headers + ) + + if "date" in response_headers: + time_tuple = parsedate_tz(response_headers["date"]) + assert time_tuple is not None + date = calendar.timegm(time_tuple[:6]) + else: + date = 0 # If we've been given a body, our response has a Content-Length, that # Content-Length is valid then we can check to see if the body we've @@ -276,6 +366,7 @@ def cache_response(self, request, response, body=None, status_codes=None): cc_req = self.parse_cache_control(request.headers) cc = self.parse_cache_control(response_headers) + assert request.url is not None cache_url = self.cache_url(request.url) logger.debug('Updating cache with response from "%s"', cache_url) @@ -304,47 +395,79 @@ def cache_response(self, request, response, body=None, status_codes=None): # If we've been given an etag, then keep the response if self.cache_etags and "etag" in response_headers: + expires_time = 0 + if response_headers.get("expires"): + expires = parsedate_tz(response_headers["expires"]) + if expires is not None: + expires_time = calendar.timegm(expires[:6]) - date + + expires_time = max(expires_time, 14 * 86400) + + logger.debug(f"etag object cached for {expires_time} seconds") logger.debug("Caching due to etag") - self.cache.set( - cache_url, self.serializer.dumps(request, response, body=body) - ) + self._cache_set(cache_url, request, response, body, expires_time) - # Add to the cache any 301s. We do this before looking that - # the Date headers. - elif response.status == 301: - logger.debug("Caching permanant redirect") - self.cache.set(cache_url, self.serializer.dumps(request, response)) + # Add to the cache any permanent redirects. We do this before looking + # that the Date headers. + elif int(response.status) in PERMANENT_REDIRECT_STATUSES: + logger.debug("Caching permanent redirect") + self._cache_set(cache_url, request, response, b"") # Add to the cache if the response headers demand it. If there # is no date header then we can't do anything about expiring # the cache. elif "date" in response_headers: + time_tuple = parsedate_tz(response_headers["date"]) + assert time_tuple is not None + date = calendar.timegm(time_tuple[:6]) # cache when there is a max-age > 0 - if "max-age" in cc and cc["max-age"] > 0: + max_age = cc.get("max-age") + if max_age is not None and max_age > 0: logger.debug("Caching b/c date exists and max-age > 0") - self.cache.set( - cache_url, self.serializer.dumps(request, response, body=body) + expires_time = max_age + self._cache_set( + cache_url, + request, + response, + body, + expires_time, ) # If the request can expire, it means we should cache it # in the meantime. elif "expires" in response_headers: if response_headers["expires"]: - logger.debug("Caching b/c of expires header") - self.cache.set( - cache_url, self.serializer.dumps(request, response, body=body) + expires = parsedate_tz(response_headers["expires"]) + if expires is not None: + expires_time = calendar.timegm(expires[:6]) - date + else: + expires_time = None + + logger.debug( + "Caching b/c of expires header. expires in {} seconds".format( + expires_time + ) + ) + self._cache_set( + cache_url, + request, + response, + body, + expires_time, ) - def update_cached_response(self, request, response): + def update_cached_response( + self, request: PreparedRequest, response: HTTPResponse + ) -> HTTPResponse: """On a 304 we will get a new set of headers that we want to update our cached value with, assuming we have one. This should only ever be called when we've sent an ETag and gotten a 304 as the response. """ + assert request.url is not None cache_url = self.cache_url(request.url) - - cached_response = self.serializer.loads(request, self.cache.get(cache_url)) + cached_response = self._load_from_cache(request) if not cached_response: # we didn't have a cached response @@ -360,17 +483,17 @@ def update_cached_response(self, request, response): excluded_headers = ["content-length"] cached_response.headers.update( - dict( - (k, v) + { + k: v for k, v in response.headers.items() if k.lower() not in excluded_headers - ) + } ) # we want a 200 b/c we have content via the cache cached_response.status = 200 # update our cache - self.cache.set(cache_url, self.serializer.dumps(request, cached_response)) + self._cache_set(cache_url, request, cached_response) return cached_response diff --git a/src/fetchcode/vcs/pip/_vendor/cachecontrol/filewrapper.py b/src/fetchcode/vcs/pip/_vendor/cachecontrol/filewrapper.py index 30ed4c5a..25143902 100644 --- a/src/fetchcode/vcs/pip/_vendor/cachecontrol/filewrapper.py +++ b/src/fetchcode/vcs/pip/_vendor/cachecontrol/filewrapper.py @@ -1,7 +1,17 @@ -from io import BytesIO +# SPDX-FileCopyrightText: 2015 Eric Larson +# +# SPDX-License-Identifier: Apache-2.0 +from __future__ import annotations +import mmap +from tempfile import NamedTemporaryFile +from typing import TYPE_CHECKING, Any, Callable -class CallbackFileWrapper(object): +if TYPE_CHECKING: + from http.client import HTTPResponse + + +class CallbackFileWrapper: """ Small wrapper around a fp object which will tee everything read into a buffer, and when that file is closed it will execute a callback with the @@ -11,14 +21,23 @@ class CallbackFileWrapper(object): This class uses members with a double underscore (__) leading prefix so as not to accidentally shadow an attribute. + + The data is stored in a temporary file until it is all available. As long + as the temporary files directory is disk-based (sometimes it's a + memory-backed-``tmpfs`` on Linux), data will be unloaded to disk if memory + pressure is high. For small files the disk usually won't be used at all, + it'll all be in the filesystem memory cache, so there should be no + performance impact. """ - def __init__(self, fp, callback): - self.__buf = BytesIO() + def __init__( + self, fp: HTTPResponse, callback: Callable[[bytes], None] | None + ) -> None: + self.__buf = NamedTemporaryFile("rb+", delete=True) self.__fp = fp self.__callback = callback - def __getattr__(self, name): + def __getattr__(self, name: str) -> Any: # The vaguaries of garbage collection means that self.__fp is # not always set. By using __getattribute__ and the private # name[0] allows looking up the attribute value and raising an @@ -30,7 +49,7 @@ def __getattr__(self, name): fp = self.__getattribute__("_CallbackFileWrapper__fp") return getattr(fp, name) - def __is_fp_closed(self): + def __is_fp_closed(self) -> bool: try: return self.__fp.fp is None @@ -38,7 +57,8 @@ def __is_fp_closed(self): pass try: - return self.__fp.closed + closed: bool = self.__fp.closed + return closed except AttributeError: pass @@ -47,9 +67,21 @@ def __is_fp_closed(self): # TODO: Add some logging here... return False - def _close(self): + def _close(self) -> None: if self.__callback: - self.__callback(self.__buf.getvalue()) + if self.__buf.tell() == 0: + # Empty file: + result = b"" + else: + # Return the data without actually loading it into memory, + # relying on Python's buffer API and mmap(). mmap() just gives + # a view directly into the filesystem's memory cache, so it + # doesn't result in duplicate memory use. + self.__buf.seek(0, 0) + result = memoryview( + mmap.mmap(self.__buf.fileno(), 0, access=mmap.ACCESS_READ) + ) + self.__callback(result) # We assign this to None here, because otherwise we can get into # really tricky problems where the CPython interpreter dead locks @@ -58,16 +90,23 @@ def _close(self): # and allows the garbage collector to do it's thing normally. self.__callback = None - def read(self, amt=None): - data = self.__fp.read(amt) - self.__buf.write(data) + # Closing the temporary file releases memory and frees disk space. + # Important when caching big files. + self.__buf.close() + + def read(self, amt: int | None = None) -> bytes: + data: bytes = self.__fp.read(amt) + if data: + # We may be dealing with b'', a sign that things are over: + # it's passed e.g. after we've already closed self.__buf. + self.__buf.write(data) if self.__is_fp_closed(): self._close() return data - def _safe_read(self, amt): - data = self.__fp._safe_read(amt) + def _safe_read(self, amt: int) -> bytes: + data: bytes = self.__fp._safe_read(amt) # type: ignore[attr-defined] if amt == 2 and data == b"\r\n": # urllib executes this read to toss the CRLF at the end # of the chunk. diff --git a/src/fetchcode/vcs/pip/_vendor/cachecontrol/heuristics.py b/src/fetchcode/vcs/pip/_vendor/cachecontrol/heuristics.py index 6c0e9790..f6e5634e 100644 --- a/src/fetchcode/vcs/pip/_vendor/cachecontrol/heuristics.py +++ b/src/fetchcode/vcs/pip/_vendor/cachecontrol/heuristics.py @@ -1,25 +1,31 @@ +# SPDX-FileCopyrightText: 2015 Eric Larson +# +# SPDX-License-Identifier: Apache-2.0 +from __future__ import annotations + import calendar import time - +from datetime import datetime, timedelta, timezone from email.utils import formatdate, parsedate, parsedate_tz +from typing import TYPE_CHECKING, Any, Mapping -from datetime import datetime, timedelta +if TYPE_CHECKING: + from pip._vendor.urllib3 import HTTPResponse TIME_FMT = "%a, %d %b %Y %H:%M:%S GMT" -def expire_after(delta, date=None): - date = date or datetime.utcnow() +def expire_after(delta: timedelta, date: datetime | None = None) -> datetime: + date = date or datetime.now(timezone.utc) return date + delta -def datetime_to_header(dt): +def datetime_to_header(dt: datetime) -> str: return formatdate(calendar.timegm(dt.timetuple())) -class BaseHeuristic(object): - - def warning(self, response): +class BaseHeuristic: + def warning(self, response: HTTPResponse) -> str | None: """ Return a valid 1xx warning header value describing the cache adjustments. @@ -30,7 +36,7 @@ def warning(self, response): """ return '110 - "Response is Stale"' - def update_headers(self, response): + def update_headers(self, response: HTTPResponse) -> dict[str, str]: """Update the response headers with any new headers. NOTE: This SHOULD always include some Warning header to @@ -39,7 +45,7 @@ def update_headers(self, response): """ return {} - def apply(self, response): + def apply(self, response: HTTPResponse) -> HTTPResponse: updated_headers = self.update_headers(response) if updated_headers: @@ -57,12 +63,12 @@ class OneDayCache(BaseHeuristic): future. """ - def update_headers(self, response): + def update_headers(self, response: HTTPResponse) -> dict[str, str]: headers = {} if "expires" not in response.headers: date = parsedate(response.headers["date"]) - expires = expire_after(timedelta(days=1), date=datetime(*date[:6])) + expires = expire_after(timedelta(days=1), date=datetime(*date[:6], tzinfo=timezone.utc)) # type: ignore[index,misc] headers["expires"] = datetime_to_header(expires) headers["cache-control"] = "public" return headers @@ -73,14 +79,14 @@ class ExpiresAfter(BaseHeuristic): Cache **all** requests for a defined time period. """ - def __init__(self, **kw): + def __init__(self, **kw: Any) -> None: self.delta = timedelta(**kw) - def update_headers(self, response): + def update_headers(self, response: HTTPResponse) -> dict[str, str]: expires = expire_after(self.delta) return {"expires": datetime_to_header(expires), "cache-control": "public"} - def warning(self, response): + def warning(self, response: HTTPResponse) -> str | None: tmpl = "110 - Automatically cached for %s. Response might be stale" return tmpl % self.delta @@ -97,12 +103,23 @@ class LastModified(BaseHeuristic): http://lxr.mozilla.org/mozilla-release/source/netwerk/protocol/http/nsHttpResponseHead.cpp#397 Unlike mozilla we limit this to 24-hr. """ + cacheable_by_default_statuses = { - 200, 203, 204, 206, 300, 301, 404, 405, 410, 414, 501 + 200, + 203, + 204, + 206, + 300, + 301, + 404, + 405, + 410, + 414, + 501, } - def update_headers(self, resp): - headers = resp.headers + def update_headers(self, resp: HTTPResponse) -> dict[str, str]: + headers: Mapping[str, str] = resp.headers if "expires" in headers: return {} @@ -116,9 +133,11 @@ def update_headers(self, resp): if "date" not in headers or "last-modified" not in headers: return {} - date = calendar.timegm(parsedate_tz(headers["date"])) + time_tuple = parsedate_tz(headers["date"]) + assert time_tuple is not None + date = calendar.timegm(time_tuple[:6]) last_modified = parsedate(headers["last-modified"]) - if date is None or last_modified is None: + if last_modified is None: return {} now = time.time() @@ -131,5 +150,5 @@ def update_headers(self, resp): expires = date + freshness_lifetime return {"expires": time.strftime(TIME_FMT, time.gmtime(expires))} - def warning(self, resp): + def warning(self, resp: HTTPResponse) -> str | None: return None diff --git a/src/fetchcode/vcs/pip/_vendor/cachecontrol/py.typed b/src/fetchcode/vcs/pip/_vendor/cachecontrol/py.typed new file mode 100644 index 00000000..e69de29b diff --git a/src/fetchcode/vcs/pip/_vendor/cachecontrol/serialize.py b/src/fetchcode/vcs/pip/_vendor/cachecontrol/serialize.py index cea1d0cc..a49487a1 100644 --- a/src/fetchcode/vcs/pip/_vendor/cachecontrol/serialize.py +++ b/src/fetchcode/vcs/pip/_vendor/cachecontrol/serialize.py @@ -1,107 +1,91 @@ -import base64 -import io -import json -import zlib - -from fetchcode.vcs.pip._vendor import msgpack -from fetchcode.vcs.pip._vendor.requests.structures import CaseInsensitiveDict - -from .compat import HTTPResponse, pickle, text_type - +# SPDX-FileCopyrightText: 2015 Eric Larson +# +# SPDX-License-Identifier: Apache-2.0 +from __future__ import annotations -def _b64_decode_bytes(b): - return base64.b64decode(b.encode("ascii")) +import io +from typing import IO, TYPE_CHECKING, Any, Mapping, cast +from pip._vendor import msgpack +from pip._vendor.requests.structures import CaseInsensitiveDict +from pip._vendor.urllib3 import HTTPResponse -def _b64_decode_str(s): - return _b64_decode_bytes(s).decode("utf8") +if TYPE_CHECKING: + from pip._vendor.requests import PreparedRequest -class Serializer(object): +class Serializer: + serde_version = "4" - def dumps(self, request, response, body=None): - response_headers = CaseInsensitiveDict(response.headers) + def dumps( + self, + request: PreparedRequest, + response: HTTPResponse, + body: bytes | None = None, + ) -> bytes: + response_headers: CaseInsensitiveDict[str] = CaseInsensitiveDict( + response.headers + ) if body is None: + # When a body isn't passed in, we'll read the response. We + # also update the response with a new file handler to be + # sure it acts as though it was never read. body = response.read(decode_content=False) + response._fp = io.BytesIO(body) # type: ignore[assignment] + response.length_remaining = len(body) - # NOTE: 99% sure this is dead code. I'm only leaving it - # here b/c I don't have a test yet to prove - # it. Basically, before using - # `cachecontrol.filewrapper.CallbackFileWrapper`, - # this made an effort to reset the file handle. The - # `CallbackFileWrapper` short circuits this code by - # setting the body as the content is consumed, the - # result being a `body` argument is *always* passed - # into cache_response, and in turn, - # `Serializer.dump`. - response._fp = io.BytesIO(body) - - # NOTE: This is all a bit weird, but it's really important that on - # Python 2.x these objects are unicode and not str, even when - # they contain only ascii. The problem here is that msgpack - # understands the difference between unicode and bytes and we - # have it set to differentiate between them, however Python 2 - # doesn't know the difference. Forcing these to unicode will be - # enough to have msgpack know the difference. data = { - u"response": { - u"body": body, - u"headers": dict( - (text_type(k), text_type(v)) for k, v in response.headers.items() - ), - u"status": response.status, - u"version": response.version, - u"reason": text_type(response.reason), - u"strict": response.strict, - u"decode_content": response.decode_content, + "response": { + "body": body, # Empty bytestring if body is stored separately + "headers": {str(k): str(v) for k, v in response.headers.items()}, + "status": response.status, + "version": response.version, + "reason": str(response.reason), + "decode_content": response.decode_content, } } # Construct our vary headers - data[u"vary"] = {} - if u"vary" in response_headers: - varied_headers = response_headers[u"vary"].split(",") + data["vary"] = {} + if "vary" in response_headers: + varied_headers = response_headers["vary"].split(",") for header in varied_headers: - header = text_type(header).strip() + header = str(header).strip() header_value = request.headers.get(header, None) if header_value is not None: - header_value = text_type(header_value) - data[u"vary"][header] = header_value + header_value = str(header_value) + data["vary"][header] = header_value - return b",".join([b"cc=4", msgpack.dumps(data, use_bin_type=True)]) + return b",".join([f"cc={self.serde_version}".encode(), self.serialize(data)]) - def loads(self, request, data): + def serialize(self, data: dict[str, Any]) -> bytes: + return cast(bytes, msgpack.dumps(data, use_bin_type=True)) + + def loads( + self, + request: PreparedRequest, + data: bytes, + body_file: IO[bytes] | None = None, + ) -> HTTPResponse | None: # Short circuit if we've been given an empty set of data if not data: - return - - # Determine what version of the serializer the data was serialized - # with - try: - ver, data = data.split(b",", 1) - except ValueError: - ver = b"cc=0" - - # Make sure that our "ver" is actually a version and isn't a false - # positive from a , being in the data stream. - if ver[:3] != b"cc=": - data = ver + data - ver = b"cc=0" - - # Get the version number out of the cc=N - ver = ver.split(b"=", 1)[-1].decode("ascii") - - # Dispatch to the actual load method for the given version - try: - return getattr(self, "_loads_v{}".format(ver))(request, data) - - except AttributeError: - # This is a version we don't have a loads function for, so we'll - # just treat it as a miss and return None - return - - def prepare_response(self, request, cached): + return None + + # Previous versions of this library supported other serialization + # formats, but these have all been removed. + if not data.startswith(f"cc={self.serde_version},".encode()): + return None + + data = data[5:] + return self._loads_v4(request, data, body_file) + + def prepare_response( + self, + request: PreparedRequest, + cached: Mapping[str, Any], + body_file: IO[bytes] | None = None, + ) -> HTTPResponse | None: """Verify our vary headers match and construct a real urllib3 HTTPResponse object. """ @@ -110,24 +94,30 @@ def prepare_response(self, request, cached): # This case is also handled in the controller code when creating # a cache entry, but is left here for backwards compatibility. if "*" in cached.get("vary", {}): - return + return None # Ensure that the Vary headers for the cached response match our # request for header, value in cached.get("vary", {}).items(): if request.headers.get(header, None) != value: - return + return None body_raw = cached["response"].pop("body") - headers = CaseInsensitiveDict(data=cached["response"]["headers"]) + headers: CaseInsensitiveDict[str] = CaseInsensitiveDict( + data=cached["response"]["headers"] + ) if headers.get("transfer-encoding", "") == "chunked": headers.pop("transfer-encoding") cached["response"]["headers"] = headers try: - body = io.BytesIO(body_raw) + body: IO[bytes] + if body_file is None: + body = io.BytesIO(body_raw) + else: + body = body_file except TypeError: # This can happen if cachecontrol serialized to v1 format (pickle) # using Python 2. A Python 2 str(byte string) will be unpickled as @@ -137,52 +127,20 @@ def prepare_response(self, request, cached): # TypeError: 'str' does not support the buffer interface body = io.BytesIO(body_raw.encode("utf8")) - return HTTPResponse(body=body, preload_content=False, **cached["response"]) - - def _loads_v0(self, request, data): - # The original legacy cache data. This doesn't contain enough - # information to construct everything we need, so we'll treat this as - # a miss. - return - - def _loads_v1(self, request, data): - try: - cached = pickle.loads(data) - except ValueError: - return - - return self.prepare_response(request, cached) + # Discard any `strict` parameter serialized by older version of cachecontrol. + cached["response"].pop("strict", None) - def _loads_v2(self, request, data): - try: - cached = json.loads(zlib.decompress(data).decode("utf8")) - except (ValueError, zlib.error): - return - - # We need to decode the items that we've base64 encoded - cached["response"]["body"] = _b64_decode_bytes(cached["response"]["body"]) - cached["response"]["headers"] = dict( - (_b64_decode_str(k), _b64_decode_str(v)) - for k, v in cached["response"]["headers"].items() - ) - cached["response"]["reason"] = _b64_decode_str(cached["response"]["reason"]) - cached["vary"] = dict( - (_b64_decode_str(k), _b64_decode_str(v) if v is not None else v) - for k, v in cached["vary"].items() - ) - - return self.prepare_response(request, cached) - - def _loads_v3(self, request, data): - # Due to Python 2 encoding issues, it's impossible to know for sure - # exactly how to load v3 entries, thus we'll treat these as a miss so - # that they get rewritten out as v4 entries. - return + return HTTPResponse(body=body, preload_content=False, **cached["response"]) - def _loads_v4(self, request, data): + def _loads_v4( + self, + request: PreparedRequest, + data: bytes, + body_file: IO[bytes] | None = None, + ) -> HTTPResponse | None: try: cached = msgpack.loads(data, raw=False) except ValueError: - return + return None - return self.prepare_response(request, cached) + return self.prepare_response(request, cached, body_file) diff --git a/src/fetchcode/vcs/pip/_vendor/cachecontrol/wrapper.py b/src/fetchcode/vcs/pip/_vendor/cachecontrol/wrapper.py index d8e6fc6a..f618bc36 100644 --- a/src/fetchcode/vcs/pip/_vendor/cachecontrol/wrapper.py +++ b/src/fetchcode/vcs/pip/_vendor/cachecontrol/wrapper.py @@ -1,18 +1,32 @@ -from .adapter import CacheControlAdapter -from .cache import DictCache +# SPDX-FileCopyrightText: 2015 Eric Larson +# +# SPDX-License-Identifier: Apache-2.0 +from __future__ import annotations +from typing import TYPE_CHECKING, Collection -def CacheControl( - sess, - cache=None, - cache_etags=True, - serializer=None, - heuristic=None, - controller_class=None, - adapter_class=None, - cacheable_methods=None, -): +from pip._vendor.cachecontrol.adapter import CacheControlAdapter +from pip._vendor.cachecontrol.cache import DictCache + +if TYPE_CHECKING: + from pip._vendor import requests + + from pip._vendor.cachecontrol.cache import BaseCache + from pip._vendor.cachecontrol.controller import CacheController + from pip._vendor.cachecontrol.heuristics import BaseHeuristic + from pip._vendor.cachecontrol.serialize import Serializer + +def CacheControl( + sess: requests.Session, + cache: BaseCache | None = None, + cache_etags: bool = True, + serializer: Serializer | None = None, + heuristic: BaseHeuristic | None = None, + controller_class: type[CacheController] | None = None, + adapter_class: type[CacheControlAdapter] | None = None, + cacheable_methods: Collection[str] | None = None, +) -> requests.Session: cache = DictCache() if cache is None else cache adapter_class = adapter_class or CacheControlAdapter adapter = adapter_class( diff --git a/src/fetchcode/vcs/pip/_vendor/certifi.pyi b/src/fetchcode/vcs/pip/_vendor/certifi.pyi deleted file mode 100644 index e5c4d3d2..00000000 --- a/src/fetchcode/vcs/pip/_vendor/certifi.pyi +++ /dev/null @@ -1 +0,0 @@ -from certifi import * \ No newline at end of file diff --git a/src/fetchcode/vcs/pip/_vendor/certifi/LICENSE b/src/fetchcode/vcs/pip/_vendor/certifi/LICENSE index 802b53ff..62b076cd 100644 --- a/src/fetchcode/vcs/pip/_vendor/certifi/LICENSE +++ b/src/fetchcode/vcs/pip/_vendor/certifi/LICENSE @@ -1,12 +1,11 @@ -This packge contains a modified version of ca-bundle.crt: +This package contains a modified version of ca-bundle.crt: ca-bundle.crt -- Bundle of CA Root Certificates -Certificate data from Mozilla as of: Thu Nov 3 19:04:19 2011# This is a bundle of X.509 certificates of public Certificate Authorities (CA). These were automatically extracted from Mozilla's root certificates file (certdata.txt). This file can be found in the mozilla source tree: -http://mxr.mozilla.org/mozilla/source/security/nss/lib/ckfw/builtins/certdata.txt?raw=1# +https://hg.mozilla.org/mozilla-central/file/tip/security/nss/lib/ckfw/builtins/certdata.txt It contains the certificates in PEM format and therefore can be directly used with curl / libcurl / php_curl, or with an Apache+mod_ssl webserver for SSL client authentication. diff --git a/src/fetchcode/vcs/pip/_vendor/certifi/__init__.py b/src/fetchcode/vcs/pip/_vendor/certifi/__init__.py index 1e2dfac7..d321f1bc 100644 --- a/src/fetchcode/vcs/pip/_vendor/certifi/__init__.py +++ b/src/fetchcode/vcs/pip/_vendor/certifi/__init__.py @@ -1,3 +1,4 @@ from .core import contents, where -__version__ = "2020.04.05.1" +__all__ = ["contents", "where"] +__version__ = "2024.07.04" diff --git a/src/fetchcode/vcs/pip/_vendor/certifi/__main__.py b/src/fetchcode/vcs/pip/_vendor/certifi/__main__.py index da63be02..00376349 100644 --- a/src/fetchcode/vcs/pip/_vendor/certifi/__main__.py +++ b/src/fetchcode/vcs/pip/_vendor/certifi/__main__.py @@ -1,6 +1,6 @@ import argparse -from fetchcode.vcs.pip._vendor.certifi import contents, where +from pip._vendor.certifi import contents, where parser = argparse.ArgumentParser() parser.add_argument("-c", "--contents", action="store_true") diff --git a/src/fetchcode/vcs/pip/_vendor/certifi/cacert.pem b/src/fetchcode/vcs/pip/_vendor/certifi/cacert.pem index ece147c9..a6581589 100644 --- a/src/fetchcode/vcs/pip/_vendor/certifi/cacert.pem +++ b/src/fetchcode/vcs/pip/_vendor/certifi/cacert.pem @@ -28,68 +28,6 @@ DKqC5JlR3XC321Y9YeRq4VzW9v493kHMB65jUr9TU/Qr6cf9tveCX4XSQRjbgbME HMUfpIBvFSDJ3gyICh3WZlXi/EjJKSZp4A== -----END CERTIFICATE----- -# Issuer: CN=GlobalSign O=GlobalSign OU=GlobalSign Root CA - R2 -# Subject: CN=GlobalSign O=GlobalSign OU=GlobalSign Root CA - R2 -# Label: "GlobalSign Root CA - R2" -# Serial: 4835703278459682885658125 -# MD5 Fingerprint: 94:14:77:7e:3e:5e:fd:8f:30:bd:41:b0:cf:e7:d0:30 -# SHA1 Fingerprint: 75:e0:ab:b6:13:85:12:27:1c:04:f8:5f:dd:de:38:e4:b7:24:2e:fe -# SHA256 Fingerprint: ca:42:dd:41:74:5f:d0:b8:1e:b9:02:36:2c:f9:d8:bf:71:9d:a1:bd:1b:1e:fc:94:6f:5b:4c:99:f4:2c:1b:9e ------BEGIN CERTIFICATE----- -MIIDujCCAqKgAwIBAgILBAAAAAABD4Ym5g0wDQYJKoZIhvcNAQEFBQAwTDEgMB4G -A1UECxMXR2xvYmFsU2lnbiBSb290IENBIC0gUjIxEzARBgNVBAoTCkdsb2JhbFNp -Z24xEzARBgNVBAMTCkdsb2JhbFNpZ24wHhcNMDYxMjE1MDgwMDAwWhcNMjExMjE1 -MDgwMDAwWjBMMSAwHgYDVQQLExdHbG9iYWxTaWduIFJvb3QgQ0EgLSBSMjETMBEG -A1UEChMKR2xvYmFsU2lnbjETMBEGA1UEAxMKR2xvYmFsU2lnbjCCASIwDQYJKoZI -hvcNAQEBBQADggEPADCCAQoCggEBAKbPJA6+Lm8omUVCxKs+IVSbC9N/hHD6ErPL -v4dfxn+G07IwXNb9rfF73OX4YJYJkhD10FPe+3t+c4isUoh7SqbKSaZeqKeMWhG8 -eoLrvozps6yWJQeXSpkqBy+0Hne/ig+1AnwblrjFuTosvNYSuetZfeLQBoZfXklq -tTleiDTsvHgMCJiEbKjNS7SgfQx5TfC4LcshytVsW33hoCmEofnTlEnLJGKRILzd -C9XZzPnqJworc5HGnRusyMvo4KD0L5CLTfuwNhv2GXqF4G3yYROIXJ/gkwpRl4pa -zq+r1feqCapgvdzZX99yqWATXgAByUr6P6TqBwMhAo6CygPCm48CAwEAAaOBnDCB -mTAOBgNVHQ8BAf8EBAMCAQYwDwYDVR0TAQH/BAUwAwEB/zAdBgNVHQ4EFgQUm+IH -V2ccHsBqBt5ZtJot39wZhi4wNgYDVR0fBC8wLTAroCmgJ4YlaHR0cDovL2NybC5n -bG9iYWxzaWduLm5ldC9yb290LXIyLmNybDAfBgNVHSMEGDAWgBSb4gdXZxwewGoG -3lm0mi3f3BmGLjANBgkqhkiG9w0BAQUFAAOCAQEAmYFThxxol4aR7OBKuEQLq4Gs -J0/WwbgcQ3izDJr86iw8bmEbTUsp9Z8FHSbBuOmDAGJFtqkIk7mpM0sYmsL4h4hO -291xNBrBVNpGP+DTKqttVCL1OmLNIG+6KYnX3ZHu01yiPqFbQfXf5WRDLenVOavS -ot+3i9DAgBkcRcAtjOj4LaR0VknFBbVPFd5uRHg5h6h+u/N5GJG79G+dwfCMNYxd -AfvDbbnvRG15RjF+Cv6pgsH/76tuIMRQyV+dTZsXjAzlAcmgQWpzU/qlULRuJQ/7 -TBj0/VLZjmmx6BEP3ojY+x1J96relc8geMJgEtslQIxq/H5COEBkEveegeGTLg== ------END CERTIFICATE----- - -# Issuer: CN=VeriSign Class 3 Public Primary Certification Authority - G3 O=VeriSign, Inc. OU=VeriSign Trust Network/(c) 1999 VeriSign, Inc. - For authorized use only -# Subject: CN=VeriSign Class 3 Public Primary Certification Authority - G3 O=VeriSign, Inc. OU=VeriSign Trust Network/(c) 1999 VeriSign, Inc. - For authorized use only -# Label: "Verisign Class 3 Public Primary Certification Authority - G3" -# Serial: 206684696279472310254277870180966723415 -# MD5 Fingerprint: cd:68:b6:a7:c7:c4:ce:75:e0:1d:4f:57:44:61:92:09 -# SHA1 Fingerprint: 13:2d:0d:45:53:4b:69:97:cd:b2:d5:c3:39:e2:55:76:60:9b:5c:c6 -# SHA256 Fingerprint: eb:04:cf:5e:b1:f3:9a:fa:76:2f:2b:b1:20:f2:96:cb:a5:20:c1:b9:7d:b1:58:95:65:b8:1c:b9:a1:7b:72:44 ------BEGIN CERTIFICATE----- -MIIEGjCCAwICEQCbfgZJoz5iudXukEhxKe9XMA0GCSqGSIb3DQEBBQUAMIHKMQsw -CQYDVQQGEwJVUzEXMBUGA1UEChMOVmVyaVNpZ24sIEluYy4xHzAdBgNVBAsTFlZl -cmlTaWduIFRydXN0IE5ldHdvcmsxOjA4BgNVBAsTMShjKSAxOTk5IFZlcmlTaWdu -LCBJbmMuIC0gRm9yIGF1dGhvcml6ZWQgdXNlIG9ubHkxRTBDBgNVBAMTPFZlcmlT -aWduIENsYXNzIDMgUHVibGljIFByaW1hcnkgQ2VydGlmaWNhdGlvbiBBdXRob3Jp -dHkgLSBHMzAeFw05OTEwMDEwMDAwMDBaFw0zNjA3MTYyMzU5NTlaMIHKMQswCQYD -VQQGEwJVUzEXMBUGA1UEChMOVmVyaVNpZ24sIEluYy4xHzAdBgNVBAsTFlZlcmlT -aWduIFRydXN0IE5ldHdvcmsxOjA4BgNVBAsTMShjKSAxOTk5IFZlcmlTaWduLCBJ -bmMuIC0gRm9yIGF1dGhvcml6ZWQgdXNlIG9ubHkxRTBDBgNVBAMTPFZlcmlTaWdu -IENsYXNzIDMgUHVibGljIFByaW1hcnkgQ2VydGlmaWNhdGlvbiBBdXRob3JpdHkg -LSBHMzCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMu6nFL8eB8aHm8b -N3O9+MlrlBIwT/A2R/XQkQr1F8ilYcEWQE37imGQ5XYgwREGfassbqb1EUGO+i2t -KmFZpGcmTNDovFJbcCAEWNF6yaRpvIMXZK0Fi7zQWM6NjPXr8EJJC52XJ2cybuGu -kxUccLwgTS8Y3pKI6GyFVxEa6X7jJhFUokWWVYPKMIno3Nij7SqAP395ZVc+FSBm -CC+Vk7+qRy+oRpfwEuL+wgorUeZ25rdGt+INpsyow0xZVYnm6FNcHOqd8GIWC6fJ -Xwzw3sJ2zq/3avL6QaaiMxTJ5Xpj055iN9WFZZ4O5lMkdBteHRJTW8cs54NJOxWu -imi5V5cCAwEAATANBgkqhkiG9w0BAQUFAAOCAQEAERSWwauSCPc/L8my/uRan2Te -2yFPhpk0djZX3dAVL8WtfxUfN2JzPtTnX84XA9s1+ivbrmAJXx5fj267Cz3qWhMe -DGBvtcC1IyIuBwvLqXTLR7sdwdela8wv0kL9Sd2nic9TutoAWii/gt/4uhMdUIaC -/Y4wjylGsB49Ndo4YhYYSq3mtlFs3q9i6wHQHiT+eo8SGhJouPtmmRQURVyu565p -F4ErWjfJXir0xuKhXFSbplQAz/DxwceYMBo7Nhbbo27q/a2ywtrvAkcTisDxszGt -TxzhT5yvDwyd93gN2PQ1VoDat20Xj50egWTh/sVFuq1ruQp6Tk9LhO5L8X3dEQ== ------END CERTIFICATE----- - # Issuer: CN=Entrust.net Certification Authority (2048) O=Entrust.net OU=www.entrust.net/CPS_2048 incorp. by ref. (limits liab.)/(c) 1999 Entrust.net Limited # Subject: CN=Entrust.net Certification Authority (2048) O=Entrust.net OU=www.entrust.net/CPS_2048 incorp. by ref. (limits liab.)/(c) 1999 Entrust.net Limited # Label: "Entrust.net Premium 2048 Secure Server CA" @@ -152,39 +90,6 @@ ksLi4xaNmjICq44Y3ekQEe5+NauQrz4wlHrQMz2nZQ/1/I6eYs9HRCwBXbsdtTLS R9I4LtD+gdwyah617jzV/OeBHRnDJELqYzmp -----END CERTIFICATE----- -# Issuer: CN=AddTrust External CA Root O=AddTrust AB OU=AddTrust External TTP Network -# Subject: CN=AddTrust External CA Root O=AddTrust AB OU=AddTrust External TTP Network -# Label: "AddTrust External Root" -# Serial: 1 -# MD5 Fingerprint: 1d:35:54:04:85:78:b0:3f:42:42:4d:bf:20:73:0a:3f -# SHA1 Fingerprint: 02:fa:f3:e2:91:43:54:68:60:78:57:69:4d:f5:e4:5b:68:85:18:68 -# SHA256 Fingerprint: 68:7f:a4:51:38:22:78:ff:f0:c8:b1:1f:8d:43:d5:76:67:1c:6e:b2:bc:ea:b4:13:fb:83:d9:65:d0:6d:2f:f2 ------BEGIN CERTIFICATE----- -MIIENjCCAx6gAwIBAgIBATANBgkqhkiG9w0BAQUFADBvMQswCQYDVQQGEwJTRTEU -MBIGA1UEChMLQWRkVHJ1c3QgQUIxJjAkBgNVBAsTHUFkZFRydXN0IEV4dGVybmFs -IFRUUCBOZXR3b3JrMSIwIAYDVQQDExlBZGRUcnVzdCBFeHRlcm5hbCBDQSBSb290 -MB4XDTAwMDUzMDEwNDgzOFoXDTIwMDUzMDEwNDgzOFowbzELMAkGA1UEBhMCU0Ux -FDASBgNVBAoTC0FkZFRydXN0IEFCMSYwJAYDVQQLEx1BZGRUcnVzdCBFeHRlcm5h -bCBUVFAgTmV0d29yazEiMCAGA1UEAxMZQWRkVHJ1c3QgRXh0ZXJuYWwgQ0EgUm9v -dDCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBALf3GjPm8gAELTngTlvt -H7xsD821+iO2zt6bETOXpClMfZOfvUq8k+0DGuOPz+VtUFrWlymUWoCwSXrbLpX9 -uMq/NzgtHj6RQa1wVsfwTz/oMp50ysiQVOnGXw94nZpAPA6sYapeFI+eh6FqUNzX -mk6vBbOmcZSccbNQYArHE504B4YCqOmoaSYYkKtMsE8jqzpPhNjfzp/haW+710LX -a0Tkx63ubUFfclpxCDezeWWkWaCUN/cALw3CknLa0Dhy2xSoRcRdKn23tNbE7qzN -E0S3ySvdQwAl+mG5aWpYIxG3pzOPVnVZ9c0p10a3CitlttNCbxWyuHv77+ldU9U0 -WicCAwEAAaOB3DCB2TAdBgNVHQ4EFgQUrb2YejS0Jvf6xCZU7wO94CTLVBowCwYD -VR0PBAQDAgEGMA8GA1UdEwEB/wQFMAMBAf8wgZkGA1UdIwSBkTCBjoAUrb2YejS0 -Jvf6xCZU7wO94CTLVBqhc6RxMG8xCzAJBgNVBAYTAlNFMRQwEgYDVQQKEwtBZGRU -cnVzdCBBQjEmMCQGA1UECxMdQWRkVHJ1c3QgRXh0ZXJuYWwgVFRQIE5ldHdvcmsx -IjAgBgNVBAMTGUFkZFRydXN0IEV4dGVybmFsIENBIFJvb3SCAQEwDQYJKoZIhvcN -AQEFBQADggEBALCb4IUlwtYj4g+WBpKdQZic2YR5gdkeWxQHIzZlj7DYd7usQWxH -YINRsPkyPef89iYTx4AWpb9a/IfPeHmJIZriTAcKhjW88t5RxNKWt9x+Tu5w/Rw5 -6wwCURQtjr0W4MHfRnXnJK3s9EK0hZNwEGe6nQY1ShjTK3rMUUKhemPR5ruhxSvC -Nr4TDea9Y355e6cJDUCrat2PisP29owaQgVR1EX1n6diIWgVIEM8med8vSTYqZEX -c4g/VhsxOBi0cQ+azcgOno4uG+GMmIPLHzHxREzGBHNJdmAPx/i9F4BrLunMTA5a -mnkPIAou1Z5jJh5VkpTYghdae9C8x49OhgQ= ------END CERTIFICATE----- - # Issuer: CN=Entrust Root Certification Authority O=Entrust, Inc. OU=www.entrust.net/CPS is incorporated by reference/(c) 2006 Entrust, Inc. # Subject: CN=Entrust Root Certification Authority O=Entrust, Inc. OU=www.entrust.net/CPS is incorporated by reference/(c) 2006 Entrust, Inc. # Label: "Entrust Root Certification Authority" @@ -220,112 +125,6 @@ eu6FSqdQgPCnXEqULl8FmTxSQeDNtGPPAUO6nIPcj2A781q0tHuu2guQOHXvgR1m 0vdXcDazv/wor3ElhVsT/h5/WrQ8 -----END CERTIFICATE----- -# Issuer: CN=GeoTrust Global CA O=GeoTrust Inc. -# Subject: CN=GeoTrust Global CA O=GeoTrust Inc. -# Label: "GeoTrust Global CA" -# Serial: 144470 -# MD5 Fingerprint: f7:75:ab:29:fb:51:4e:b7:77:5e:ff:05:3c:99:8e:f5 -# SHA1 Fingerprint: de:28:f4:a4:ff:e5:b9:2f:a3:c5:03:d1:a3:49:a7:f9:96:2a:82:12 -# SHA256 Fingerprint: ff:85:6a:2d:25:1d:cd:88:d3:66:56:f4:50:12:67:98:cf:ab:aa:de:40:79:9c:72:2d:e4:d2:b5:db:36:a7:3a ------BEGIN CERTIFICATE----- -MIIDVDCCAjygAwIBAgIDAjRWMA0GCSqGSIb3DQEBBQUAMEIxCzAJBgNVBAYTAlVT -MRYwFAYDVQQKEw1HZW9UcnVzdCBJbmMuMRswGQYDVQQDExJHZW9UcnVzdCBHbG9i -YWwgQ0EwHhcNMDIwNTIxMDQwMDAwWhcNMjIwNTIxMDQwMDAwWjBCMQswCQYDVQQG -EwJVUzEWMBQGA1UEChMNR2VvVHJ1c3QgSW5jLjEbMBkGA1UEAxMSR2VvVHJ1c3Qg -R2xvYmFsIENBMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA2swYYzD9 -9BcjGlZ+W988bDjkcbd4kdS8odhM+KhDtgPpTSEHCIjaWC9mOSm9BXiLnTjoBbdq -fnGk5sRgprDvgOSJKA+eJdbtg/OtppHHmMlCGDUUna2YRpIuT8rxh0PBFpVXLVDv -iS2Aelet8u5fa9IAjbkU+BQVNdnARqN7csiRv8lVK83Qlz6cJmTM386DGXHKTubU -1XupGc1V3sjs0l44U+VcT4wt/lAjNvxm5suOpDkZALeVAjmRCw7+OC7RHQWa9k0+ -bw8HHa8sHo9gOeL6NlMTOdReJivbPagUvTLrGAMoUgRx5aszPeE4uwc2hGKceeoW -MPRfwCvocWvk+QIDAQABo1MwUTAPBgNVHRMBAf8EBTADAQH/MB0GA1UdDgQWBBTA -ephojYn7qwVkDBF9qn1luMrMTjAfBgNVHSMEGDAWgBTAephojYn7qwVkDBF9qn1l -uMrMTjANBgkqhkiG9w0BAQUFAAOCAQEANeMpauUvXVSOKVCUn5kaFOSPeCpilKIn -Z57QzxpeR+nBsqTP3UEaBU6bS+5Kb1VSsyShNwrrZHYqLizz/Tt1kL/6cdjHPTfS -tQWVYrmm3ok9Nns4d0iXrKYgjy6myQzCsplFAMfOEVEiIuCl6rYVSAlk6l5PdPcF -PseKUgzbFbS9bZvlxrFUaKnjaZC2mqUPuLk/IH2uSrW4nOQdtqvmlKXBx4Ot2/Un -hw4EbNX/3aBd7YdStysVAq45pmp06drE57xNNB6pXE0zX5IJL4hmXXeXxx12E6nV -5fEWCRE11azbJHFwLJhWC9kXtNHjUStedejV0NxPNO3CBWaAocvmMw== ------END CERTIFICATE----- - -# Issuer: CN=GeoTrust Universal CA O=GeoTrust Inc. -# Subject: CN=GeoTrust Universal CA O=GeoTrust Inc. -# Label: "GeoTrust Universal CA" -# Serial: 1 -# MD5 Fingerprint: 92:65:58:8b:a2:1a:31:72:73:68:5c:b4:a5:7a:07:48 -# SHA1 Fingerprint: e6:21:f3:35:43:79:05:9a:4b:68:30:9d:8a:2f:74:22:15:87:ec:79 -# SHA256 Fingerprint: a0:45:9b:9f:63:b2:25:59:f5:fa:5d:4c:6d:b3:f9:f7:2f:f1:93:42:03:35:78:f0:73:bf:1d:1b:46:cb:b9:12 ------BEGIN CERTIFICATE----- -MIIFaDCCA1CgAwIBAgIBATANBgkqhkiG9w0BAQUFADBFMQswCQYDVQQGEwJVUzEW -MBQGA1UEChMNR2VvVHJ1c3QgSW5jLjEeMBwGA1UEAxMVR2VvVHJ1c3QgVW5pdmVy -c2FsIENBMB4XDTA0MDMwNDA1MDAwMFoXDTI5MDMwNDA1MDAwMFowRTELMAkGA1UE -BhMCVVMxFjAUBgNVBAoTDUdlb1RydXN0IEluYy4xHjAcBgNVBAMTFUdlb1RydXN0 -IFVuaXZlcnNhbCBDQTCCAiIwDQYJKoZIhvcNAQEBBQADggIPADCCAgoCggIBAKYV -VaCjxuAfjJ0hUNfBvitbtaSeodlyWL0AG0y/YckUHUWCq8YdgNY96xCcOq9tJPi8 -cQGeBvV8Xx7BDlXKg5pZMK4ZyzBIle0iN430SppyZj6tlcDgFgDgEB8rMQ7XlFTT -QjOgNB0eRXbdT8oYN+yFFXoZCPzVx5zw8qkuEKmS5j1YPakWaDwvdSEYfyh3peFh -F7em6fgemdtzbvQKoiFs7tqqhZJmr/Z6a4LauiIINQ/PQvE1+mrufislzDoR5G2v -c7J2Ha3QsnhnGqQ5HFELZ1aD/ThdDc7d8Lsrlh/eezJS/R27tQahsiFepdaVaH/w -mZ7cRQg+59IJDTWU3YBOU5fXtQlEIGQWFwMCTFMNaN7VqnJNk22CDtucvc+081xd -VHppCZbW2xHBjXWotM85yM48vCR85mLK4b19p71XZQvk/iXttmkQ3CgaRr0BHdCX -teGYO8A3ZNY9lO4L4fUorgtWv3GLIylBjobFS1J72HGrH4oVpjuDWtdYAVHGTEHZ -f9hBZ3KiKN9gg6meyHv8U3NyWfWTehd2Ds735VzZC1U0oqpbtWpU5xPKV+yXbfRe -Bi9Fi1jUIxaS5BZuKGNZMN9QAZxjiRqf2xeUgnA3wySemkfWWspOqGmJch+RbNt+ -nhutxx9z3SxPGWX9f5NAEC7S8O08ni4oPmkmM8V7AgMBAAGjYzBhMA8GA1UdEwEB -/wQFMAMBAf8wHQYDVR0OBBYEFNq7LqqwDLiIJlF0XG0D08DYj3rWMB8GA1UdIwQY -MBaAFNq7LqqwDLiIJlF0XG0D08DYj3rWMA4GA1UdDwEB/wQEAwIBhjANBgkqhkiG -9w0BAQUFAAOCAgEAMXjmx7XfuJRAyXHEqDXsRh3ChfMoWIawC/yOsjmPRFWrZIRc -aanQmjg8+uUfNeVE44B5lGiku8SfPeE0zTBGi1QrlaXv9z+ZhP015s8xxtxqv6fX -IwjhmF7DWgh2qaavdy+3YL1ERmrvl/9zlcGO6JP7/TG37FcREUWbMPEaiDnBTzyn -ANXH/KttgCJwpQzgXQQpAvvLoJHRfNbDflDVnVi+QTjruXU8FdmbyUqDWcDaU/0z -uzYYm4UPFd3uLax2k7nZAY1IEKj79TiG8dsKxr2EoyNB3tZ3b4XUhRxQ4K5RirqN -Pnbiucon8l+f725ZDQbYKxek0nxru18UGkiPGkzns0ccjkxFKyDuSN/n3QmOGKja -QI2SJhFTYXNd673nxE0pN2HrrDktZy4W1vUAg4WhzH92xH3kt0tm7wNFYGm2DFKW -koRepqO1pD4r2czYG0eq8kTaT/kD6PAUyz/zg97QwVTjt+gKN02LIFkDMBmhLMi9 -ER/frslKxfMnZmaGrGiR/9nmUxwPi1xpZQomyB40w11Re9epnAahNt3ViZS82eQt -DF4JbAiXfKM9fJP/P6EUp8+1Xevb2xzEdt+Iub1FBZUbrvxGakyvSOPOrg/Sfuvm -bJxPgWp6ZKy7PtXny3YuxadIwVyQD8vIP/rmMuGNG2+k5o7Y+SlIis5z/iw= ------END CERTIFICATE----- - -# Issuer: CN=GeoTrust Universal CA 2 O=GeoTrust Inc. -# Subject: CN=GeoTrust Universal CA 2 O=GeoTrust Inc. -# Label: "GeoTrust Universal CA 2" -# Serial: 1 -# MD5 Fingerprint: 34:fc:b8:d0:36:db:9e:14:b3:c2:f2:db:8f:e4:94:c7 -# SHA1 Fingerprint: 37:9a:19:7b:41:85:45:35:0c:a6:03:69:f3:3c:2e:af:47:4f:20:79 -# SHA256 Fingerprint: a0:23:4f:3b:c8:52:7c:a5:62:8e:ec:81:ad:5d:69:89:5d:a5:68:0d:c9:1d:1c:b8:47:7f:33:f8:78:b9:5b:0b ------BEGIN CERTIFICATE----- -MIIFbDCCA1SgAwIBAgIBATANBgkqhkiG9w0BAQUFADBHMQswCQYDVQQGEwJVUzEW -MBQGA1UEChMNR2VvVHJ1c3QgSW5jLjEgMB4GA1UEAxMXR2VvVHJ1c3QgVW5pdmVy -c2FsIENBIDIwHhcNMDQwMzA0MDUwMDAwWhcNMjkwMzA0MDUwMDAwWjBHMQswCQYD -VQQGEwJVUzEWMBQGA1UEChMNR2VvVHJ1c3QgSW5jLjEgMB4GA1UEAxMXR2VvVHJ1 -c3QgVW5pdmVyc2FsIENBIDIwggIiMA0GCSqGSIb3DQEBAQUAA4ICDwAwggIKAoIC -AQCzVFLByT7y2dyxUxpZKeexw0Uo5dfR7cXFS6GqdHtXr0om/Nj1XqduGdt0DE81 -WzILAePb63p3NeqqWuDW6KFXlPCQo3RWlEQwAx5cTiuFJnSCegx2oG9NzkEtoBUG -FF+3Qs17j1hhNNwqCPkuwwGmIkQcTAeC5lvO0Ep8BNMZcyfwqph/Lq9O64ceJHdq -XbboW0W63MOhBW9Wjo8QJqVJwy7XQYci4E+GymC16qFjwAGXEHm9ADwSbSsVsaxL -se4YuU6W3Nx2/zu+z18DwPw76L5GG//aQMJS9/7jOvdqdzXQ2o3rXhhqMcceujwb -KNZrVMaqW9eiLBsZzKIC9ptZvTdrhrVtgrrY6slWvKk2WP0+GfPtDCapkzj4T8Fd -IgbQl+rhrcZV4IErKIM6+vR7IVEAvlI4zs1meaj0gVbi0IMJR1FbUGrP20gaXT73 -y/Zl92zxlfgCOzJWgjl6W70viRu/obTo/3+NjN8D8WBOWBFM66M/ECuDmgFz2ZRt -hAAnZqzwcEAJQpKtT5MNYQlRJNiS1QuUYbKHsu3/mjX/hVTK7URDrBs8FmtISgoc -QIgfksILAAX/8sgCSqSqqcyZlpwvWOB94b67B9xfBHJcMTTD7F8t4D1kkCLm0ey4 -Lt1ZrtmhN79UNdxzMk+MBB4zsslG8dhcyFVQyWi9qLo2CQIDAQABo2MwYTAPBgNV -HRMBAf8EBTADAQH/MB0GA1UdDgQWBBR281Xh+qQ2+/CfXGJx7Tz0RzgQKzAfBgNV -HSMEGDAWgBR281Xh+qQ2+/CfXGJx7Tz0RzgQKzAOBgNVHQ8BAf8EBAMCAYYwDQYJ -KoZIhvcNAQEFBQADggIBAGbBxiPz2eAubl/oz66wsCVNK/g7WJtAJDday6sWSf+z -dXkzoS9tcBc0kf5nfo/sm+VegqlVHy/c1FEHEv6sFj4sNcZj/NwQ6w2jqtB8zNHQ -L1EuxBRa3ugZ4T7GzKQp5y6EqgYweHZUcyiYWTjgAA1i00J9IZ+uPTqM1fp3DRgr -Fg5fNuH8KrUwJM/gYwx7WBr+mbpCErGR9Hxo4sjoryzqyX6uuyo9DRXcNJW2GHSo -ag/HtPQTxORb7QrSpJdMKu0vbBKJPfEncKpqA1Ihn0CoZ1Dy81of398j9tx4TuaY -T1U6U+Pv8vSfx3zYWK8pIpe44L2RLrB27FcRz+8pRPPphXpgY+RdM4kX2TGq2tbz -GDVyz4crL2MjhF2EjD9XoIj8mZEoJmmZ1I+XRL6O1UixpCgp8RW04eWe3fiPpm8m -1wk8OhwRDqZsN/etRIcsKMfYdIKz0G9KV7s1KSegi+ghp4dkNl3M2Basx7InQJJV -OCiNUW7dFGdTbHFcJoRNdVq2fmBWqU2t+5sel/MN2dKXVHfaPRK34B7vCAas+YWH -6aLcr34YEoP9VhdBLtUpgn2Z9DH2canPLAEnpQW5qrJITirvn5NSUZU8UnOOVkwX -QMAJKOSLakhT2+zNVVXxxvjpoixMptEmX36vWkzaH6byHCx+rgIW0lbQL1dTR+iS ------END CERTIFICATE----- - # Issuer: CN=AAA Certificate Services O=Comodo CA Limited # Subject: CN=AAA Certificate Services O=Comodo CA Limited # Label: "Comodo AAA Services root" @@ -359,48 +158,6 @@ l2D4kF501KKaU73yqWjgom7C12yxow+ev+to51byrvLjKzg6CYG1a4XXvi3tPxq3 smPi9WIsgtRqAEFQ8TmDn5XpNpaYbg== -----END CERTIFICATE----- -# Issuer: CN=QuoVadis Root Certification Authority O=QuoVadis Limited OU=Root Certification Authority -# Subject: CN=QuoVadis Root Certification Authority O=QuoVadis Limited OU=Root Certification Authority -# Label: "QuoVadis Root CA" -# Serial: 985026699 -# MD5 Fingerprint: 27:de:36:fe:72:b7:00:03:00:9d:f4:f0:1e:6c:04:24 -# SHA1 Fingerprint: de:3f:40:bd:50:93:d3:9b:6c:60:f6:da:bc:07:62:01:00:89:76:c9 -# SHA256 Fingerprint: a4:5e:de:3b:bb:f0:9c:8a:e1:5c:72:ef:c0:72:68:d6:93:a2:1c:99:6f:d5:1e:67:ca:07:94:60:fd:6d:88:73 ------BEGIN CERTIFICATE----- -MIIF0DCCBLigAwIBAgIEOrZQizANBgkqhkiG9w0BAQUFADB/MQswCQYDVQQGEwJC -TTEZMBcGA1UEChMQUXVvVmFkaXMgTGltaXRlZDElMCMGA1UECxMcUm9vdCBDZXJ0 -aWZpY2F0aW9uIEF1dGhvcml0eTEuMCwGA1UEAxMlUXVvVmFkaXMgUm9vdCBDZXJ0 -aWZpY2F0aW9uIEF1dGhvcml0eTAeFw0wMTAzMTkxODMzMzNaFw0yMTAzMTcxODMz -MzNaMH8xCzAJBgNVBAYTAkJNMRkwFwYDVQQKExBRdW9WYWRpcyBMaW1pdGVkMSUw -IwYDVQQLExxSb290IENlcnRpZmljYXRpb24gQXV0aG9yaXR5MS4wLAYDVQQDEyVR -dW9WYWRpcyBSb290IENlcnRpZmljYXRpb24gQXV0aG9yaXR5MIIBIjANBgkqhkiG -9w0BAQEFAAOCAQ8AMIIBCgKCAQEAv2G1lVO6V/z68mcLOhrfEYBklbTRvM16z/Yp -li4kVEAkOPcahdxYTMukJ0KX0J+DisPkBgNbAKVRHnAEdOLB1Dqr1607BxgFjv2D -rOpm2RgbaIr1VxqYuvXtdj182d6UajtLF8HVj71lODqV0D1VNk7feVcxKh7YWWVJ -WCCYfqtffp/p1k3sg3Spx2zY7ilKhSoGFPlU5tPaZQeLYzcS19Dsw3sgQUSj7cug -F+FxZc4dZjH3dgEZyH0DWLaVSR2mEiboxgx24ONmy+pdpibu5cxfvWenAScOospU -xbF6lR1xHkopigPcakXBpBlebzbNw6Kwt/5cOOJSvPhEQ+aQuwIDAQABo4ICUjCC -Ak4wPQYIKwYBBQUHAQEEMTAvMC0GCCsGAQUFBzABhiFodHRwczovL29jc3AucXVv -dmFkaXNvZmZzaG9yZS5jb20wDwYDVR0TAQH/BAUwAwEB/zCCARoGA1UdIASCAREw -ggENMIIBCQYJKwYBBAG+WAABMIH7MIHUBggrBgEFBQcCAjCBxxqBxFJlbGlhbmNl -IG9uIHRoZSBRdW9WYWRpcyBSb290IENlcnRpZmljYXRlIGJ5IGFueSBwYXJ0eSBh -c3N1bWVzIGFjY2VwdGFuY2Ugb2YgdGhlIHRoZW4gYXBwbGljYWJsZSBzdGFuZGFy -ZCB0ZXJtcyBhbmQgY29uZGl0aW9ucyBvZiB1c2UsIGNlcnRpZmljYXRpb24gcHJh -Y3RpY2VzLCBhbmQgdGhlIFF1b1ZhZGlzIENlcnRpZmljYXRlIFBvbGljeS4wIgYI -KwYBBQUHAgEWFmh0dHA6Ly93d3cucXVvdmFkaXMuYm0wHQYDVR0OBBYEFItLbe3T -KbkGGew5Oanwl4Rqy+/fMIGuBgNVHSMEgaYwgaOAFItLbe3TKbkGGew5Oanwl4Rq -y+/foYGEpIGBMH8xCzAJBgNVBAYTAkJNMRkwFwYDVQQKExBRdW9WYWRpcyBMaW1p -dGVkMSUwIwYDVQQLExxSb290IENlcnRpZmljYXRpb24gQXV0aG9yaXR5MS4wLAYD -VQQDEyVRdW9WYWRpcyBSb290IENlcnRpZmljYXRpb24gQXV0aG9yaXR5ggQ6tlCL -MA4GA1UdDwEB/wQEAwIBBjANBgkqhkiG9w0BAQUFAAOCAQEAitQUtf70mpKnGdSk -fnIYj9lofFIk3WdvOXrEql494liwTXCYhGHoG+NpGA7O+0dQoE7/8CQfvbLO9Sf8 -7C9TqnN7Az10buYWnuulLsS/VidQK2K6vkscPFVcQR0kvoIgR13VRH56FmjffU1R -cHhXHTMe/QKZnAzNCgVPx7uOpHX6Sm2xgI4JVrmcGmD+XcHXetwReNDWXcG31a0y -mQM6isxUJTkxgXsTIlG6Rmyhu576BGxJJnSP0nPrzDCi5upZIof4l/UO/erMkqQW -xFIY6iHOsfHmhIHluqmGKPJDWl0Snawe2ajlCmqnf6CHKc/yiU3U7MXi5nrQNiOK -SnQ2+Q== ------END CERTIFICATE----- - # Issuer: CN=QuoVadis Root CA 2 O=QuoVadis Limited # Subject: CN=QuoVadis Root CA 2 O=QuoVadis Limited # Label: "QuoVadis Root CA 2" @@ -488,61 +245,6 @@ mJlglFwjz1onl14LBQaTNx47aTbrqZ5hHY8y2o4M1nQ+ewkk2gF3R8Q7zTSMmfXK 4SVhM7JZG+Ju1zdXtg2pEto= -----END CERTIFICATE----- -# Issuer: O=SECOM Trust.net OU=Security Communication RootCA1 -# Subject: O=SECOM Trust.net OU=Security Communication RootCA1 -# Label: "Security Communication Root CA" -# Serial: 0 -# MD5 Fingerprint: f1:bc:63:6a:54:e0:b5:27:f5:cd:e7:1a:e3:4d:6e:4a -# SHA1 Fingerprint: 36:b1:2b:49:f9:81:9e:d7:4c:9e:bc:38:0f:c6:56:8f:5d:ac:b2:f7 -# SHA256 Fingerprint: e7:5e:72:ed:9f:56:0e:ec:6e:b4:80:00:73:a4:3f:c3:ad:19:19:5a:39:22:82:01:78:95:97:4a:99:02:6b:6c ------BEGIN CERTIFICATE----- -MIIDWjCCAkKgAwIBAgIBADANBgkqhkiG9w0BAQUFADBQMQswCQYDVQQGEwJKUDEY -MBYGA1UEChMPU0VDT00gVHJ1c3QubmV0MScwJQYDVQQLEx5TZWN1cml0eSBDb21t -dW5pY2F0aW9uIFJvb3RDQTEwHhcNMDMwOTMwMDQyMDQ5WhcNMjMwOTMwMDQyMDQ5 -WjBQMQswCQYDVQQGEwJKUDEYMBYGA1UEChMPU0VDT00gVHJ1c3QubmV0MScwJQYD -VQQLEx5TZWN1cml0eSBDb21tdW5pY2F0aW9uIFJvb3RDQTEwggEiMA0GCSqGSIb3 -DQEBAQUAA4IBDwAwggEKAoIBAQCzs/5/022x7xZ8V6UMbXaKL0u/ZPtM7orw8yl8 -9f/uKuDp6bpbZCKamm8sOiZpUQWZJtzVHGpxxpp9Hp3dfGzGjGdnSj74cbAZJ6kJ -DKaVv0uMDPpVmDvY6CKhS3E4eayXkmmziX7qIWgGmBSWh9JhNrxtJ1aeV+7AwFb9 -Ms+k2Y7CI9eNqPPYJayX5HA49LY6tJ07lyZDo6G8SVlyTCMwhwFY9k6+HGhWZq/N -QV3Is00qVUarH9oe4kA92819uZKAnDfdDJZkndwi92SL32HeFZRSFaB9UslLqCHJ -xrHty8OVYNEP8Ktw+N/LTX7s1vqr2b1/VPKl6Xn62dZ2JChzAgMBAAGjPzA9MB0G -A1UdDgQWBBSgc0mZaNyFW2XjmygvV5+9M7wHSDALBgNVHQ8EBAMCAQYwDwYDVR0T -AQH/BAUwAwEB/zANBgkqhkiG9w0BAQUFAAOCAQEAaECpqLvkT115swW1F7NgE+vG -kl3g0dNq/vu+m22/xwVtWSDEHPC32oRYAmP6SBbvT6UL90qY8j+eG61Ha2POCEfr -Uj94nK9NrvjVT8+amCoQQTlSxN3Zmw7vkwGusi7KaEIkQmywszo+zenaSMQVy+n5 -Bw+SUEmK3TGXX8npN6o7WWWXlDLJs58+OmJYxUmtYg5xpTKqL8aJdkNAExNnPaJU -JRDL8Try2frbSVa7pv6nQTXD4IhhyYjH3zYQIphZ6rBK+1YWc26sTfcioU+tHXot -RSflMMFe8toTyyVCUZVHA4xsIcx0Qu1T/zOLjw9XARYvz6buyXAiFL39vmwLAw== ------END CERTIFICATE----- - -# Issuer: CN=Sonera Class2 CA O=Sonera -# Subject: CN=Sonera Class2 CA O=Sonera -# Label: "Sonera Class 2 Root CA" -# Serial: 29 -# MD5 Fingerprint: a3:ec:75:0f:2e:88:df:fa:48:01:4e:0b:5c:48:6f:fb -# SHA1 Fingerprint: 37:f7:6d:e6:07:7c:90:c5:b1:3e:93:1a:b7:41:10:b4:f2:e4:9a:27 -# SHA256 Fingerprint: 79:08:b4:03:14:c1:38:10:0b:51:8d:07:35:80:7f:fb:fc:f8:51:8a:00:95:33:71:05:ba:38:6b:15:3d:d9:27 ------BEGIN CERTIFICATE----- -MIIDIDCCAgigAwIBAgIBHTANBgkqhkiG9w0BAQUFADA5MQswCQYDVQQGEwJGSTEP -MA0GA1UEChMGU29uZXJhMRkwFwYDVQQDExBTb25lcmEgQ2xhc3MyIENBMB4XDTAx -MDQwNjA3Mjk0MFoXDTIxMDQwNjA3Mjk0MFowOTELMAkGA1UEBhMCRkkxDzANBgNV -BAoTBlNvbmVyYTEZMBcGA1UEAxMQU29uZXJhIENsYXNzMiBDQTCCASIwDQYJKoZI -hvcNAQEBBQADggEPADCCAQoCggEBAJAXSjWdyvANlsdE+hY3/Ei9vX+ALTU74W+o -Z6m/AxxNjG8yR9VBaKQTBME1DJqEQ/xcHf+Js+gXGM2RX/uJ4+q/Tl18GybTdXnt -5oTjV+WtKcT0OijnpXuENmmz/V52vaMtmdOQTiMofRhj8VQ7Jp12W5dCsv+u8E7s -3TmVToMGf+dJQMjFAbJUWmYdPfz56TwKnoG4cPABi+QjVHzIrviQHgCWctRUz2Ej -vOr7nQKV0ba5cTppCD8PtOFCx4j1P5iop7oc4HFx71hXgVB6XGt0Rg6DA5jDjqhu -8nYybieDwnPz3BjotJPqdURrBGAgcVeHnfO+oJAjPYok4doh28MCAwEAAaMzMDEw -DwYDVR0TAQH/BAUwAwEB/zARBgNVHQ4ECgQISqCqWITTXjwwCwYDVR0PBAQDAgEG -MA0GCSqGSIb3DQEBBQUAA4IBAQBazof5FnIVV0sd2ZvnoiYw7JNn39Yt0jSv9zil -zqsWuasvfDXLrNAPtEwr/IDva4yRXzZ299uzGxnq9LIR/WFxRL8oszodv7ND6J+/ -3DEIcbCdjdY0RzKQxmUk96BKfARzjzlvF4xytb1LyHr4e4PDKE6cCepnP7JnBBvD -FNr450kkkdAdavphOe9r5yF1BgfYErQhIHBCcYHaPJo2vqZbDWpsmh+Re/n570K6 -Tk6ezAyNlNzZRZxe7EJQY670XcSxEtzKO6gunRRaBXW37Ndj4ro1tgQIkejanZz2 -ZrUYrAqmVCY0M9IbwdR/GjqOC6oybtv8TyWf2TLHllpwrN9M ------END CERTIFICATE----- - # Issuer: CN=XRamp Global Certification Authority O=XRamp Security Services Inc OU=www.xrampsecurity.com # Subject: CN=XRamp Global Certification Authority O=XRamp Security Services Inc OU=www.xrampsecurity.com # Label: "XRamp Global CA Root" @@ -640,46 +342,6 @@ VSJYACPq4xJDKVtHCN2MQWplBqjlIapBtJUhlbl90TSrE9atvNziPTnNvT51cKEY WQPJIrSPnNVeKtelttQKbfi3QBFGmh95DmK/D5fs4C8fF5Q= -----END CERTIFICATE----- -# Issuer: O=Government Root Certification Authority -# Subject: O=Government Root Certification Authority -# Label: "Taiwan GRCA" -# Serial: 42023070807708724159991140556527066870 -# MD5 Fingerprint: 37:85:44:53:32:45:1f:20:f0:f3:95:e1:25:c4:43:4e -# SHA1 Fingerprint: f4:8b:11:bf:de:ab:be:94:54:20:71:e6:41:de:6b:be:88:2b:40:b9 -# SHA256 Fingerprint: 76:00:29:5e:ef:e8:5b:9e:1f:d6:24:db:76:06:2a:aa:ae:59:81:8a:54:d2:77:4c:d4:c0:b2:c0:11:31:e1:b3 ------BEGIN CERTIFICATE----- -MIIFcjCCA1qgAwIBAgIQH51ZWtcvwgZEpYAIaeNe9jANBgkqhkiG9w0BAQUFADA/ -MQswCQYDVQQGEwJUVzEwMC4GA1UECgwnR292ZXJubWVudCBSb290IENlcnRpZmlj -YXRpb24gQXV0aG9yaXR5MB4XDTAyMTIwNTEzMjMzM1oXDTMyMTIwNTEzMjMzM1ow -PzELMAkGA1UEBhMCVFcxMDAuBgNVBAoMJ0dvdmVybm1lbnQgUm9vdCBDZXJ0aWZp -Y2F0aW9uIEF1dGhvcml0eTCCAiIwDQYJKoZIhvcNAQEBBQADggIPADCCAgoCggIB -AJoluOzMonWoe/fOW1mKydGGEghU7Jzy50b2iPN86aXfTEc2pBsBHH8eV4qNw8XR -IePaJD9IK/ufLqGU5ywck9G/GwGHU5nOp/UKIXZ3/6m3xnOUT0b3EEk3+qhZSV1q -gQdW8or5BtD3cCJNtLdBuTK4sfCxw5w/cP1T3YGq2GN49thTbqGsaoQkclSGxtKy -yhwOeYHWtXBiCAEuTk8O1RGvqa/lmr/czIdtJuTJV6L7lvnM4T9TjGxMfptTCAts -F/tnyMKtsc2AtJfcdgEWFelq16TheEfOhtX7MfP6Mb40qij7cEwdScevLJ1tZqa2 -jWR+tSBqnTuBto9AAGdLiYa4zGX+FVPpBMHWXx1E1wovJ5pGfaENda1UhhXcSTvx -ls4Pm6Dso3pdvtUqdULle96ltqqvKKyskKw4t9VoNSZ63Pc78/1Fm9G7Q3hub/FC -VGqY8A2tl+lSXunVanLeavcbYBT0peS2cWeqH+riTcFCQP5nRhc4L0c/cZyu5SHK -YS1tB6iEfC3uUSXxY5Ce/eFXiGvviiNtsea9P63RPZYLhY3Naye7twWb7LuRqQoH -EgKXTiCQ8P8NHuJBO9NAOueNXdpm5AKwB1KYXA6OM5zCppX7VRluTI6uSw+9wThN -Xo+EHWbNxWCWtFJaBYmOlXqYwZE8lSOyDvR5tMl8wUohAgMBAAGjajBoMB0GA1Ud -DgQWBBTMzO/MKWCkO7GStjz6MmKPrCUVOzAMBgNVHRMEBTADAQH/MDkGBGcqBwAE -MTAvMC0CAQAwCQYFKw4DAhoFADAHBgVnKgMAAAQUA5vwIhP/lSg209yewDL7MTqK -UWUwDQYJKoZIhvcNAQEFBQADggIBAECASvomyc5eMN1PhnR2WPWus4MzeKR6dBcZ -TulStbngCnRiqmjKeKBMmo4sIy7VahIkv9Ro04rQ2JyftB8M3jh+Vzj8jeJPXgyf -qzvS/3WXy6TjZwj/5cAWtUgBfen5Cv8b5Wppv3ghqMKnI6mGq3ZW6A4M9hPdKmaK -ZEk9GhiHkASfQlK3T8v+R0F2Ne//AHY2RTKbxkaFXeIksB7jSJaYV0eUVXoPQbFE -JPPB/hprv4j9wabak2BegUqZIJxIZhm1AHlUD7gsL0u8qV1bYH+Mh6XgUmMqvtg7 -hUAV/h62ZT/FS9p+tXo1KaMuephgIqP0fSdOLeq0dDzpD6QzDxARvBMB1uUO07+1 -EqLhRSPAzAhuYbeJq4PjJB7mXQfnHyA+z2fI56wwbSdLaG5LKlwCCDTb+HbkZ6Mm -nD+iMsJKxYEYMRBWqoTvLQr/uB930r+lWKBi5NdLkXWNiYCYfm3LU05er/ayl4WX -udpVBrkk7tfGOB5jGxI7leFYrPLfhNVfmS8NVVvmONsuP3LpSIXLuykTjx44Vbnz -ssQwmSNOXfJIoRIM3BKQCZBUkQM8R+XVyWXgt0t97EfTsws+rZ7QdAAO671RrcDe -LMDDav7v3Aun+kbfYNucpllQdSNpc5Oy+fwC00fmcc4QAu4njIT/rEUNE1yDMuAl -pYYsfPQS ------END CERTIFICATE----- - # Issuer: CN=DigiCert Assured ID Root CA O=DigiCert Inc OU=www.digicert.com # Subject: CN=DigiCert Assured ID Root CA O=DigiCert Inc OU=www.digicert.com # Label: "DigiCert Assured ID Root CA" @@ -771,34 +433,6 @@ vEsXCS+0yx5DaMkHJ8HSXPfqIbloEpw8nL+e/IBcm2PN7EeqJSdnoDfzAIJ9VNep +OkuE6N36B9K -----END CERTIFICATE----- -# Issuer: CN=DST Root CA X3 O=Digital Signature Trust Co. -# Subject: CN=DST Root CA X3 O=Digital Signature Trust Co. -# Label: "DST Root CA X3" -# Serial: 91299735575339953335919266965803778155 -# MD5 Fingerprint: 41:03:52:dc:0f:f7:50:1b:16:f0:02:8e:ba:6f:45:c5 -# SHA1 Fingerprint: da:c9:02:4f:54:d8:f6:df:94:93:5f:b1:73:26:38:ca:6a:d7:7c:13 -# SHA256 Fingerprint: 06:87:26:03:31:a7:24:03:d9:09:f1:05:e6:9b:cf:0d:32:e1:bd:24:93:ff:c6:d9:20:6d:11:bc:d6:77:07:39 ------BEGIN CERTIFICATE----- -MIIDSjCCAjKgAwIBAgIQRK+wgNajJ7qJMDmGLvhAazANBgkqhkiG9w0BAQUFADA/ -MSQwIgYDVQQKExtEaWdpdGFsIFNpZ25hdHVyZSBUcnVzdCBDby4xFzAVBgNVBAMT -DkRTVCBSb290IENBIFgzMB4XDTAwMDkzMDIxMTIxOVoXDTIxMDkzMDE0MDExNVow -PzEkMCIGA1UEChMbRGlnaXRhbCBTaWduYXR1cmUgVHJ1c3QgQ28uMRcwFQYDVQQD -Ew5EU1QgUm9vdCBDQSBYMzCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEB -AN+v6ZdQCINXtMxiZfaQguzH0yxrMMpb7NnDfcdAwRgUi+DoM3ZJKuM/IUmTrE4O -rz5Iy2Xu/NMhD2XSKtkyj4zl93ewEnu1lcCJo6m67XMuegwGMoOifooUMM0RoOEq -OLl5CjH9UL2AZd+3UWODyOKIYepLYYHsUmu5ouJLGiifSKOeDNoJjj4XLh7dIN9b -xiqKqy69cK3FCxolkHRyxXtqqzTWMIn/5WgTe1QLyNau7Fqckh49ZLOMxt+/yUFw -7BZy1SbsOFU5Q9D8/RhcQPGX69Wam40dutolucbY38EVAjqr2m7xPi71XAicPNaD -aeQQmxkqtilX4+U9m5/wAl0CAwEAAaNCMEAwDwYDVR0TAQH/BAUwAwEB/zAOBgNV -HQ8BAf8EBAMCAQYwHQYDVR0OBBYEFMSnsaR7LHH62+FLkHX/xBVghYkQMA0GCSqG -SIb3DQEBBQUAA4IBAQCjGiybFwBcqR7uKGY3Or+Dxz9LwwmglSBd49lZRNI+DT69 -ikugdB/OEIKcdBodfpga3csTS7MgROSR6cz8faXbauX+5v3gTt23ADq1cEmv8uXr -AvHRAosZy5Q6XkjEGB5YGV8eAlrwDPGxrancWYaLbumR9YbK+rlmM6pZW87ipxZz -R8srzJmwN0jP41ZL9c8PDHIyh8bwRLtTcm1D9SZImlJnt1ir/md2cXjbDaJWFBM5 -JDGFoqgCWjBH4d1QB7wCCZAA62RjYJsWvIjJEubSfZGL+T0yjWW06XyxV3bqxbYo -Ob8VZRzI9neWagqNdwvYkQsEjgfbKbYK7p2CNTUQ ------END CERTIFICATE----- - # Issuer: CN=SwissSign Gold CA - G2 O=SwissSign AG # Subject: CN=SwissSign Gold CA - G2 O=SwissSign AG # Label: "SwissSign Gold CA - G2" @@ -881,104 +515,6 @@ hAhm0sQ2fac+EPyI4NSA5QC9qvNOBqN6avlicuMJT+ubDgEj8Z+7fNzcbBGXJbLy tGMU0gYqZ4yD9c7qB9iaah7s5Aq7KkzrCWA5zspi2C5u -----END CERTIFICATE----- -# Issuer: CN=GeoTrust Primary Certification Authority O=GeoTrust Inc. -# Subject: CN=GeoTrust Primary Certification Authority O=GeoTrust Inc. -# Label: "GeoTrust Primary Certification Authority" -# Serial: 32798226551256963324313806436981982369 -# MD5 Fingerprint: 02:26:c3:01:5e:08:30:37:43:a9:d0:7d:cf:37:e6:bf -# SHA1 Fingerprint: 32:3c:11:8e:1b:f7:b8:b6:52:54:e2:e2:10:0d:d6:02:90:37:f0:96 -# SHA256 Fingerprint: 37:d5:10:06:c5:12:ea:ab:62:64:21:f1:ec:8c:92:01:3f:c5:f8:2a:e9:8e:e5:33:eb:46:19:b8:de:b4:d0:6c ------BEGIN CERTIFICATE----- -MIIDfDCCAmSgAwIBAgIQGKy1av1pthU6Y2yv2vrEoTANBgkqhkiG9w0BAQUFADBY -MQswCQYDVQQGEwJVUzEWMBQGA1UEChMNR2VvVHJ1c3QgSW5jLjExMC8GA1UEAxMo -R2VvVHJ1c3QgUHJpbWFyeSBDZXJ0aWZpY2F0aW9uIEF1dGhvcml0eTAeFw0wNjEx -MjcwMDAwMDBaFw0zNjA3MTYyMzU5NTlaMFgxCzAJBgNVBAYTAlVTMRYwFAYDVQQK -Ew1HZW9UcnVzdCBJbmMuMTEwLwYDVQQDEyhHZW9UcnVzdCBQcmltYXJ5IENlcnRp -ZmljYXRpb24gQXV0aG9yaXR5MIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKC -AQEAvrgVe//UfH1nrYNke8hCUy3f9oQIIGHWAVlqnEQRr+92/ZV+zmEwu3qDXwK9 -AWbK7hWNb6EwnL2hhZ6UOvNWiAAxz9juapYC2e0DjPt1befquFUWBRaa9OBesYjA -ZIVcFU2Ix7e64HXprQU9nceJSOC7KMgD4TCTZF5SwFlwIjVXiIrxlQqD17wxcwE0 -7e9GceBrAqg1cmuXm2bgyxx5X9gaBGgeRwLmnWDiNpcB3841kt++Z8dtd1k7j53W -kBWUvEI0EME5+bEnPn7WinXFsq+W06Lem+SYvn3h6YGttm/81w7a4DSwDRp35+MI -mO9Y+pyEtzavwt+s0vQQBnBxNQIDAQABo0IwQDAPBgNVHRMBAf8EBTADAQH/MA4G -A1UdDwEB/wQEAwIBBjAdBgNVHQ4EFgQULNVQQZcVi/CPNmFbSvtr2ZnJM5IwDQYJ -KoZIhvcNAQEFBQADggEBAFpwfyzdtzRP9YZRqSa+S7iq8XEN3GHHoOo0Hnp3DwQ1 -6CePbJC/kRYkRj5KTs4rFtULUh38H2eiAkUxT87z+gOneZ1TatnaYzr4gNfTmeGl -4b7UVXGYNTq+k+qurUKykG/g/CFNNWMziUnWm07Kx+dOCQD32sfvmWKZd7aVIl6K -oKv0uHiYyjgZmclynnjNS6yvGaBzEi38wkG6gZHaFloxt/m0cYASSJlyc1pZU8Fj -UjPtp8nSOQJw+uCxQmYpqptR7TBUIhRf2asdweSU8Pj1K/fqynhG1riR/aYNKxoU -AT6A8EKglQdebc3MS6RFjasS6LPeWuWgfOgPIh1a6Vk= ------END CERTIFICATE----- - -# Issuer: CN=thawte Primary Root CA O=thawte, Inc. OU=Certification Services Division/(c) 2006 thawte, Inc. - For authorized use only -# Subject: CN=thawte Primary Root CA O=thawte, Inc. OU=Certification Services Division/(c) 2006 thawte, Inc. - For authorized use only -# Label: "thawte Primary Root CA" -# Serial: 69529181992039203566298953787712940909 -# MD5 Fingerprint: 8c:ca:dc:0b:22:ce:f5:be:72:ac:41:1a:11:a8:d8:12 -# SHA1 Fingerprint: 91:c6:d6:ee:3e:8a:c8:63:84:e5:48:c2:99:29:5c:75:6c:81:7b:81 -# SHA256 Fingerprint: 8d:72:2f:81:a9:c1:13:c0:79:1d:f1:36:a2:96:6d:b2:6c:95:0a:97:1d:b4:6b:41:99:f4:ea:54:b7:8b:fb:9f ------BEGIN CERTIFICATE----- -MIIEIDCCAwigAwIBAgIQNE7VVyDV7exJ9C/ON9srbTANBgkqhkiG9w0BAQUFADCB -qTELMAkGA1UEBhMCVVMxFTATBgNVBAoTDHRoYXd0ZSwgSW5jLjEoMCYGA1UECxMf -Q2VydGlmaWNhdGlvbiBTZXJ2aWNlcyBEaXZpc2lvbjE4MDYGA1UECxMvKGMpIDIw -MDYgdGhhd3RlLCBJbmMuIC0gRm9yIGF1dGhvcml6ZWQgdXNlIG9ubHkxHzAdBgNV -BAMTFnRoYXd0ZSBQcmltYXJ5IFJvb3QgQ0EwHhcNMDYxMTE3MDAwMDAwWhcNMzYw -NzE2MjM1OTU5WjCBqTELMAkGA1UEBhMCVVMxFTATBgNVBAoTDHRoYXd0ZSwgSW5j -LjEoMCYGA1UECxMfQ2VydGlmaWNhdGlvbiBTZXJ2aWNlcyBEaXZpc2lvbjE4MDYG -A1UECxMvKGMpIDIwMDYgdGhhd3RlLCBJbmMuIC0gRm9yIGF1dGhvcml6ZWQgdXNl -IG9ubHkxHzAdBgNVBAMTFnRoYXd0ZSBQcmltYXJ5IFJvb3QgQ0EwggEiMA0GCSqG -SIb3DQEBAQUAA4IBDwAwggEKAoIBAQCsoPD7gFnUnMekz52hWXMJEEUMDSxuaPFs -W0hoSVk3/AszGcJ3f8wQLZU0HObrTQmnHNK4yZc2AreJ1CRfBsDMRJSUjQJib+ta -3RGNKJpchJAQeg29dGYvajig4tVUROsdB58Hum/u6f1OCyn1PoSgAfGcq/gcfomk -6KHYcWUNo1F77rzSImANuVud37r8UVsLr5iy6S7pBOhih94ryNdOwUxkHt3Ph1i6 -Sk/KaAcdHJ1KxtUvkcx8cXIcxcBn6zL9yZJclNqFwJu/U30rCfSMnZEfl2pSy94J -NqR32HuHUETVPm4pafs5SSYeCaWAe0At6+gnhcn+Yf1+5nyXHdWdAgMBAAGjQjBA -MA8GA1UdEwEB/wQFMAMBAf8wDgYDVR0PAQH/BAQDAgEGMB0GA1UdDgQWBBR7W0XP -r87Lev0xkhpqtvNG61dIUDANBgkqhkiG9w0BAQUFAAOCAQEAeRHAS7ORtvzw6WfU -DW5FvlXok9LOAz/t2iWwHVfLHjp2oEzsUHboZHIMpKnxuIvW1oeEuzLlQRHAd9mz -YJ3rG9XRbkREqaYB7FViHXe4XI5ISXycO1cRrK1zN44veFyQaEfZYGDm/Ac9IiAX -xPcW6cTYcvnIc3zfFi8VqT79aie2oetaupgf1eNNZAqdE8hhuvU5HIe6uL17In/2 -/qxAeeWsEG89jxt5dovEN7MhGITlNgDrYyCZuen+MwS7QcjBAvlEYyCegc5C09Y/ -LHbTY5xZ3Y+m4Q6gLkH3LpVHz7z9M/P2C2F+fpErgUfCJzDupxBdN49cOSvkBPB7 -jVaMaA== ------END CERTIFICATE----- - -# Issuer: CN=VeriSign Class 3 Public Primary Certification Authority - G5 O=VeriSign, Inc. OU=VeriSign Trust Network/(c) 2006 VeriSign, Inc. - For authorized use only -# Subject: CN=VeriSign Class 3 Public Primary Certification Authority - G5 O=VeriSign, Inc. OU=VeriSign Trust Network/(c) 2006 VeriSign, Inc. - For authorized use only -# Label: "VeriSign Class 3 Public Primary Certification Authority - G5" -# Serial: 33037644167568058970164719475676101450 -# MD5 Fingerprint: cb:17:e4:31:67:3e:e2:09:fe:45:57:93:f3:0a:fa:1c -# SHA1 Fingerprint: 4e:b6:d5:78:49:9b:1c:cf:5f:58:1e:ad:56:be:3d:9b:67:44:a5:e5 -# SHA256 Fingerprint: 9a:cf:ab:7e:43:c8:d8:80:d0:6b:26:2a:94:de:ee:e4:b4:65:99:89:c3:d0:ca:f1:9b:af:64:05:e4:1a:b7:df ------BEGIN CERTIFICATE----- -MIIE0zCCA7ugAwIBAgIQGNrRniZ96LtKIVjNzGs7SjANBgkqhkiG9w0BAQUFADCB -yjELMAkGA1UEBhMCVVMxFzAVBgNVBAoTDlZlcmlTaWduLCBJbmMuMR8wHQYDVQQL -ExZWZXJpU2lnbiBUcnVzdCBOZXR3b3JrMTowOAYDVQQLEzEoYykgMjAwNiBWZXJp -U2lnbiwgSW5jLiAtIEZvciBhdXRob3JpemVkIHVzZSBvbmx5MUUwQwYDVQQDEzxW -ZXJpU2lnbiBDbGFzcyAzIFB1YmxpYyBQcmltYXJ5IENlcnRpZmljYXRpb24gQXV0 -aG9yaXR5IC0gRzUwHhcNMDYxMTA4MDAwMDAwWhcNMzYwNzE2MjM1OTU5WjCByjEL -MAkGA1UEBhMCVVMxFzAVBgNVBAoTDlZlcmlTaWduLCBJbmMuMR8wHQYDVQQLExZW -ZXJpU2lnbiBUcnVzdCBOZXR3b3JrMTowOAYDVQQLEzEoYykgMjAwNiBWZXJpU2ln -biwgSW5jLiAtIEZvciBhdXRob3JpemVkIHVzZSBvbmx5MUUwQwYDVQQDEzxWZXJp -U2lnbiBDbGFzcyAzIFB1YmxpYyBQcmltYXJ5IENlcnRpZmljYXRpb24gQXV0aG9y -aXR5IC0gRzUwggEiMA0GCSqGSIb3DQEBAQUAA4IBDwAwggEKAoIBAQCvJAgIKXo1 -nmAMqudLO07cfLw8RRy7K+D+KQL5VwijZIUVJ/XxrcgxiV0i6CqqpkKzj/i5Vbex -t0uz/o9+B1fs70PbZmIVYc9gDaTY3vjgw2IIPVQT60nKWVSFJuUrjxuf6/WhkcIz -SdhDY2pSS9KP6HBRTdGJaXvHcPaz3BJ023tdS1bTlr8Vd6Gw9KIl8q8ckmcY5fQG -BO+QueQA5N06tRn/Arr0PO7gi+s3i+z016zy9vA9r911kTMZHRxAy3QkGSGT2RT+ -rCpSx4/VBEnkjWNHiDxpg8v+R70rfk/Fla4OndTRQ8Bnc+MUCH7lP59zuDMKz10/ -NIeWiu5T6CUVAgMBAAGjgbIwga8wDwYDVR0TAQH/BAUwAwEB/zAOBgNVHQ8BAf8E -BAMCAQYwbQYIKwYBBQUHAQwEYTBfoV2gWzBZMFcwVRYJaW1hZ2UvZ2lmMCEwHzAH -BgUrDgMCGgQUj+XTGoasjY5rw8+AatRIGCx7GS4wJRYjaHR0cDovL2xvZ28udmVy -aXNpZ24uY29tL3ZzbG9nby5naWYwHQYDVR0OBBYEFH/TZafC3ey78DAJ80M5+gKv -MzEzMA0GCSqGSIb3DQEBBQUAA4IBAQCTJEowX2LP2BqYLz3q3JktvXf2pXkiOOzE -p6B4Eq1iDkVwZMXnl2YtmAl+X6/WzChl8gGqCBpH3vn5fJJaCGkgDdk+bW48DW7Y -5gaRQBi5+MHt39tBquCWIMnNZBU4gcmU7qKEKQsTb47bDN0lAtukixlE0kF6BWlK -WE9gyn6CagsCqiUXObXbf+eEZSqVir2G3l6BFoMtEMze/aiCKm0oHw0LxOXnGiYZ -4fQRbxC1lfznQgUy286dUV4otp6F01vvpX1FQHKOtw5rDgb7MzVIcbidJ4vEZV8N -hnacRHr2lVz2XTIIM6RUthg/aFzyQkqFOFSDX9HoLPKsEdao7WNq ------END CERTIFICATE----- - # Issuer: CN=SecureTrust CA O=SecureTrust Corporation # Subject: CN=SecureTrust CA O=SecureTrust Corporation # Label: "SecureTrust CA" @@ -1072,37 +608,6 @@ BA6+C4OmF4O5MBKgxTMVBbkN+8cFduPYSo38NBejxiEovjBFMR7HeL5YYTisO+IB ZQ== -----END CERTIFICATE----- -# Issuer: CN=Network Solutions Certificate Authority O=Network Solutions L.L.C. -# Subject: CN=Network Solutions Certificate Authority O=Network Solutions L.L.C. -# Label: "Network Solutions Certificate Authority" -# Serial: 116697915152937497490437556386812487904 -# MD5 Fingerprint: d3:f3:a6:16:c0:fa:6b:1d:59:b1:2d:96:4d:0e:11:2e -# SHA1 Fingerprint: 74:f8:a3:c3:ef:e7:b3:90:06:4b:83:90:3c:21:64:60:20:e5:df:ce -# SHA256 Fingerprint: 15:f0:ba:00:a3:ac:7a:f3:ac:88:4c:07:2b:10:11:a0:77:bd:77:c0:97:f4:01:64:b2:f8:59:8a:bd:83:86:0c ------BEGIN CERTIFICATE----- -MIID5jCCAs6gAwIBAgIQV8szb8JcFuZHFhfjkDFo4DANBgkqhkiG9w0BAQUFADBi -MQswCQYDVQQGEwJVUzEhMB8GA1UEChMYTmV0d29yayBTb2x1dGlvbnMgTC5MLkMu -MTAwLgYDVQQDEydOZXR3b3JrIFNvbHV0aW9ucyBDZXJ0aWZpY2F0ZSBBdXRob3Jp -dHkwHhcNMDYxMjAxMDAwMDAwWhcNMjkxMjMxMjM1OTU5WjBiMQswCQYDVQQGEwJV -UzEhMB8GA1UEChMYTmV0d29yayBTb2x1dGlvbnMgTC5MLkMuMTAwLgYDVQQDEydO -ZXR3b3JrIFNvbHV0aW9ucyBDZXJ0aWZpY2F0ZSBBdXRob3JpdHkwggEiMA0GCSqG -SIb3DQEBAQUAA4IBDwAwggEKAoIBAQDkvH6SMG3G2I4rC7xGzuAnlt7e+foS0zwz -c7MEL7xxjOWftiJgPl9dzgn/ggwbmlFQGiaJ3dVhXRncEg8tCqJDXRfQNJIg6nPP -OCwGJgl6cvf6UDL4wpPTaaIjzkGxzOTVHzbRijr4jGPiFFlp7Q3Tf2vouAPlT2rl -mGNpSAW+Lv8ztumXWWn4Zxmuk2GWRBXTcrA/vGp97Eh/jcOrqnErU2lBUzS1sLnF -BgrEsEX1QV1uiUV7PTsmjHTC5dLRfbIR1PtYMiKagMnc/Qzpf14Dl847ABSHJ3A4 -qY5usyd2mFHgBeMhqxrVhSI8KbWaFsWAqPS7azCPL0YCorEMIuDTAgMBAAGjgZcw -gZQwHQYDVR0OBBYEFCEwyfsA106Y2oeqKtCnLrFAMadMMA4GA1UdDwEB/wQEAwIB -BjAPBgNVHRMBAf8EBTADAQH/MFIGA1UdHwRLMEkwR6BFoEOGQWh0dHA6Ly9jcmwu -bmV0c29sc3NsLmNvbS9OZXR3b3JrU29sdXRpb25zQ2VydGlmaWNhdGVBdXRob3Jp -dHkuY3JsMA0GCSqGSIb3DQEBBQUAA4IBAQC7rkvnt1frf6ott3NHhWrB5KUd5Oc8 -6fRZZXe1eltajSU24HqXLjjAV2CDmAaDn7l2em5Q4LqILPxFzBiwmZVRDuwduIj/ -h1AcgsLj4DKAv6ALR8jDMe+ZZzKATxcheQxpXN5eNK4CtSbqUN9/GGUsyfJj4akH -/nxxH2szJGoeBfcFaMBqEssuXmHLrijTfsK0ZpEmXzwuJF/LWA/rKOyvEZbz3Htv -wKeI8lN3s2Berq4o2jUsbzRF0ybh3uxbTydrFny9RAQYgrOJeRcQcT16ohZO9QHN -pGxlaKFJdlxDydi8NmdspZS11My5vWo1ViHe2MPr+8ukYEywVaCge1ey ------END CERTIFICATE----- - # Issuer: CN=COMODO ECC Certification Authority O=COMODO CA Limited # Subject: CN=COMODO ECC Certification Authority O=COMODO CA Limited # Label: "COMODO ECC Certification Authority" @@ -1127,38 +632,6 @@ fQjGGoe9GKhzvSbKYAydzpmfz1wPMOG+FDHqAjAU9JM8SaczepBGR7NjfRObTrdv GDeAU/7dIOA1mjbRxwG55tzd8/8dLDoWV9mSOdY= -----END CERTIFICATE----- -# Issuer: CN=OISTE WISeKey Global Root GA CA O=WISeKey OU=Copyright (c) 2005/OISTE Foundation Endorsed -# Subject: CN=OISTE WISeKey Global Root GA CA O=WISeKey OU=Copyright (c) 2005/OISTE Foundation Endorsed -# Label: "OISTE WISeKey Global Root GA CA" -# Serial: 86718877871133159090080555911823548314 -# MD5 Fingerprint: bc:6c:51:33:a7:e9:d3:66:63:54:15:72:1b:21:92:93 -# SHA1 Fingerprint: 59:22:a1:e1:5a:ea:16:35:21:f8:98:39:6a:46:46:b0:44:1b:0f:a9 -# SHA256 Fingerprint: 41:c9:23:86:6a:b4:ca:d6:b7:ad:57:80:81:58:2e:02:07:97:a6:cb:df:4f:ff:78:ce:83:96:b3:89:37:d7:f5 ------BEGIN CERTIFICATE----- -MIID8TCCAtmgAwIBAgIQQT1yx/RrH4FDffHSKFTfmjANBgkqhkiG9w0BAQUFADCB -ijELMAkGA1UEBhMCQ0gxEDAOBgNVBAoTB1dJU2VLZXkxGzAZBgNVBAsTEkNvcHly -aWdodCAoYykgMjAwNTEiMCAGA1UECxMZT0lTVEUgRm91bmRhdGlvbiBFbmRvcnNl -ZDEoMCYGA1UEAxMfT0lTVEUgV0lTZUtleSBHbG9iYWwgUm9vdCBHQSBDQTAeFw0w -NTEyMTExNjAzNDRaFw0zNzEyMTExNjA5NTFaMIGKMQswCQYDVQQGEwJDSDEQMA4G -A1UEChMHV0lTZUtleTEbMBkGA1UECxMSQ29weXJpZ2h0IChjKSAyMDA1MSIwIAYD -VQQLExlPSVNURSBGb3VuZGF0aW9uIEVuZG9yc2VkMSgwJgYDVQQDEx9PSVNURSBX -SVNlS2V5IEdsb2JhbCBSb290IEdBIENBMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8A -MIIBCgKCAQEAy0+zAJs9Nt350UlqaxBJH+zYK7LG+DKBKUOVTJoZIyEVRd7jyBxR -VVuuk+g3/ytr6dTqvirdqFEr12bDYVxgAsj1znJ7O7jyTmUIms2kahnBAbtzptf2 -w93NvKSLtZlhuAGio9RN1AU9ka34tAhxZK9w8RxrfvbDd50kc3vkDIzh2TbhmYsF -mQvtRTEJysIA2/dyoJaqlYfQjse2YXMNdmaM3Bu0Y6Kff5MTMPGhJ9vZ/yxViJGg -4E8HsChWjBgbl0SOid3gF27nKu+POQoxhILYQBRJLnpB5Kf+42TMwVlxSywhp1t9 -4B3RLoGbw9ho972WG6xwsRYUC9tguSYBBQIDAQABo1EwTzALBgNVHQ8EBAMCAYYw -DwYDVR0TAQH/BAUwAwEB/zAdBgNVHQ4EFgQUswN+rja8sHnR3JQmthG+IbJphpQw -EAYJKwYBBAGCNxUBBAMCAQAwDQYJKoZIhvcNAQEFBQADggEBAEuh/wuHbrP5wUOx -SPMowB0uyQlB+pQAHKSkq0lPjz0e701vvbyk9vImMMkQyh2I+3QZH4VFvbBsUfk2 -ftv1TDI6QU9bR8/oCy22xBmddMVHxjtqD6wU2zz0c5ypBd8A3HR4+vg1YFkCExh8 -vPtNsCBtQ7tgMHpnM1zFmdH4LTlSc/uMqpclXHLZCB6rTjzjgTGfA6b7wP4piFXa -hNVQA7bihKOmNqoROgHhGEvWRGizPflTdISzRpFGlgC3gCy24eMQ4tui5yiPAZZi -Fj4A4xylNoEYokxSdsARo27mHbrjWr42U8U+dY+GaSlYU7Wcu2+fXMUY7N0v4ZjJ -/L7fCg0= ------END CERTIFICATE----- - # Issuer: CN=Certigna O=Dhimyotis # Subject: CN=Certigna O=Dhimyotis # Label: "Certigna" @@ -1189,36 +662,6 @@ t0QmwCbAr1UwnjvVNioZBPRcHv/PLLf/0P2HQBHVESO7SMAhqaQoLf0V+LBOK/Qw WyH8EZE0vkHve52Xdf+XlcCWWC/qu0bXu+TZLg== -----END CERTIFICATE----- -# Issuer: CN=Cybertrust Global Root O=Cybertrust, Inc -# Subject: CN=Cybertrust Global Root O=Cybertrust, Inc -# Label: "Cybertrust Global Root" -# Serial: 4835703278459682877484360 -# MD5 Fingerprint: 72:e4:4a:87:e3:69:40:80:77:ea:bc:e3:f4:ff:f0:e1 -# SHA1 Fingerprint: 5f:43:e5:b1:bf:f8:78:8c:ac:1c:c7:ca:4a:9a:c6:22:2b:cc:34:c6 -# SHA256 Fingerprint: 96:0a:df:00:63:e9:63:56:75:0c:29:65:dd:0a:08:67:da:0b:9c:bd:6e:77:71:4a:ea:fb:23:49:ab:39:3d:a3 ------BEGIN CERTIFICATE----- -MIIDoTCCAomgAwIBAgILBAAAAAABD4WqLUgwDQYJKoZIhvcNAQEFBQAwOzEYMBYG -A1UEChMPQ3liZXJ0cnVzdCwgSW5jMR8wHQYDVQQDExZDeWJlcnRydXN0IEdsb2Jh -bCBSb290MB4XDTA2MTIxNTA4MDAwMFoXDTIxMTIxNTA4MDAwMFowOzEYMBYGA1UE -ChMPQ3liZXJ0cnVzdCwgSW5jMR8wHQYDVQQDExZDeWJlcnRydXN0IEdsb2JhbCBS -b290MIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA+Mi8vRRQZhP/8NN5 -7CPytxrHjoXxEnOmGaoQ25yiZXRadz5RfVb23CO21O1fWLE3TdVJDm71aofW0ozS -J8bi/zafmGWgE07GKmSb1ZASzxQG9Dvj1Ci+6A74q05IlG2OlTEQXO2iLb3VOm2y -HLtgwEZLAfVJrn5GitB0jaEMAs7u/OePuGtm839EAL9mJRQr3RAwHQeWP032a7iP -t3sMpTjr3kfb1V05/Iin89cqdPHoWqI7n1C6poxFNcJQZZXcY4Lv3b93TZxiyWNz -FtApD0mpSPCzqrdsxacwOUBdrsTiXSZT8M4cIwhhqJQZugRiQOwfOHB3EgZxpzAY -XSUnpQIDAQABo4GlMIGiMA4GA1UdDwEB/wQEAwIBBjAPBgNVHRMBAf8EBTADAQH/ -MB0GA1UdDgQWBBS2CHsNesysIEyGVjJez6tuhS1wVzA/BgNVHR8EODA2MDSgMqAw -hi5odHRwOi8vd3d3Mi5wdWJsaWMtdHJ1c3QuY29tL2NybC9jdC9jdHJvb3QuY3Js -MB8GA1UdIwQYMBaAFLYIew16zKwgTIZWMl7Pq26FLXBXMA0GCSqGSIb3DQEBBQUA -A4IBAQBW7wojoFROlZfJ+InaRcHUowAl9B8Tq7ejhVhpwjCt2BWKLePJzYFa+HMj -Wqd8BfP9IjsO0QbE2zZMcwSO5bAi5MXzLqXZI+O4Tkogp24CJJ8iYGd7ix1yCcUx -XOl5n4BHPa2hCwcUPUf/A2kaDAtE52Mlp3+yybh2hO0j9n0Hq0V+09+zv+mKts2o -omcrUtW3ZfA5TGOgkXmTUg9U3YO7n9GPp1Nzw8v/MOx8BLjYRB+TX3EJIrduPuoc -A06dGiBh+4E37F78CkWr1+cXVdCg6mCbpvbjjFspwgZgFJ0tl0ypkxWdYcQBX0jW -WL1WMRJOEcgh4LMRkWXbtKaIOM5V ------END CERTIFICATE----- - # Issuer: O=Chunghwa Telecom Co., Ltd. OU=ePKI Root Certification Authority # Subject: O=Chunghwa Telecom Co., Ltd. OU=ePKI Root Certification Authority # Label: "ePKI Root Certification Authority" @@ -1288,185 +731,6 @@ i/nDhDwTqn6Sm1dTk/pwwpEOMfmbZ13pljheX7NzTogVZ96edhBiIL5VaZVDADlN 9u6wWk5JRFRYX0KD -----END CERTIFICATE----- -# Issuer: CN=GeoTrust Primary Certification Authority - G3 O=GeoTrust Inc. OU=(c) 2008 GeoTrust Inc. - For authorized use only -# Subject: CN=GeoTrust Primary Certification Authority - G3 O=GeoTrust Inc. OU=(c) 2008 GeoTrust Inc. - For authorized use only -# Label: "GeoTrust Primary Certification Authority - G3" -# Serial: 28809105769928564313984085209975885599 -# MD5 Fingerprint: b5:e8:34:36:c9:10:44:58:48:70:6d:2e:83:d4:b8:05 -# SHA1 Fingerprint: 03:9e:ed:b8:0b:e7:a0:3c:69:53:89:3b:20:d2:d9:32:3a:4c:2a:fd -# SHA256 Fingerprint: b4:78:b8:12:25:0d:f8:78:63:5c:2a:a7:ec:7d:15:5e:aa:62:5e:e8:29:16:e2:cd:29:43:61:88:6c:d1:fb:d4 ------BEGIN CERTIFICATE----- -MIID/jCCAuagAwIBAgIQFaxulBmyeUtB9iepwxgPHzANBgkqhkiG9w0BAQsFADCB -mDELMAkGA1UEBhMCVVMxFjAUBgNVBAoTDUdlb1RydXN0IEluYy4xOTA3BgNVBAsT -MChjKSAyMDA4IEdlb1RydXN0IEluYy4gLSBGb3IgYXV0aG9yaXplZCB1c2Ugb25s -eTE2MDQGA1UEAxMtR2VvVHJ1c3QgUHJpbWFyeSBDZXJ0aWZpY2F0aW9uIEF1dGhv -cml0eSAtIEczMB4XDTA4MDQwMjAwMDAwMFoXDTM3MTIwMTIzNTk1OVowgZgxCzAJ -BgNVBAYTAlVTMRYwFAYDVQQKEw1HZW9UcnVzdCBJbmMuMTkwNwYDVQQLEzAoYykg -MjAwOCBHZW9UcnVzdCBJbmMuIC0gRm9yIGF1dGhvcml6ZWQgdXNlIG9ubHkxNjA0 -BgNVBAMTLUdlb1RydXN0IFByaW1hcnkgQ2VydGlmaWNhdGlvbiBBdXRob3JpdHkg -LSBHMzCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBANziXmJYHTNXOTIz -+uvLh4yn1ErdBojqZI4xmKU4kB6Yzy5jK/BGvESyiaHAKAxJcCGVn2TAppMSAmUm -hsalifD614SgcK9PGpc/BkTVyetyEH3kMSj7HGHmKAdEc5IiaacDiGydY8hS2pgn -5whMcD60yRLBxWeDXTPzAxHsatBT4tG6NmCUgLthY2xbF37fQJQeqw3CIShwiP/W -JmxsYAQlTlV+fe+/lEjetx3dcI0FX4ilm/LC7urRQEFtYjgdVgbFA0dRIBn8exAL -DmKudlW/X3e+PkkBUz2YJQN2JFodtNuJ6nnltrM7P7pMKEF/BqxqjsHQ9gUdfeZC -huOl1UcCAwEAAaNCMEAwDwYDVR0TAQH/BAUwAwEB/zAOBgNVHQ8BAf8EBAMCAQYw -HQYDVR0OBBYEFMR5yo6hTgMdHNxr2zFblD4/MH8tMA0GCSqGSIb3DQEBCwUAA4IB -AQAtxRPPVoB7eni9n64smefv2t+UXglpp+duaIy9cr5HqQ6XErhK8WTTOd8lNNTB -zU6B8A8ExCSzNJbGpqow32hhc9f5joWJ7w5elShKKiePEI4ufIbEAp7aDHdlDkQN -kv39sxY2+hENHYwOB4lqKVb3cvTdFZx3NWZXqxNT2I7BQMXXExZacse3aQHEerGD -AWh9jUGhlBjBJVz88P6DAod8DQ3PLghcSkANPuyBYeYk28rgDi0Hsj5W3I31QYUH -SJsMC8tJP33st/3LjWeJGqvtux6jAAgIFyqCXDFdRootD4abdNlF+9RAsXqqaC2G -spki4cErx5z481+oghLrGREt ------END CERTIFICATE----- - -# Issuer: CN=thawte Primary Root CA - G2 O=thawte, Inc. OU=(c) 2007 thawte, Inc. - For authorized use only -# Subject: CN=thawte Primary Root CA - G2 O=thawte, Inc. OU=(c) 2007 thawte, Inc. - For authorized use only -# Label: "thawte Primary Root CA - G2" -# Serial: 71758320672825410020661621085256472406 -# MD5 Fingerprint: 74:9d:ea:60:24:c4:fd:22:53:3e:cc:3a:72:d9:29:4f -# SHA1 Fingerprint: aa:db:bc:22:23:8f:c4:01:a1:27:bb:38:dd:f4:1d:db:08:9e:f0:12 -# SHA256 Fingerprint: a4:31:0d:50:af:18:a6:44:71:90:37:2a:86:af:af:8b:95:1f:fb:43:1d:83:7f:1e:56:88:b4:59:71:ed:15:57 ------BEGIN CERTIFICATE----- -MIICiDCCAg2gAwIBAgIQNfwmXNmET8k9Jj1Xm67XVjAKBggqhkjOPQQDAzCBhDEL -MAkGA1UEBhMCVVMxFTATBgNVBAoTDHRoYXd0ZSwgSW5jLjE4MDYGA1UECxMvKGMp -IDIwMDcgdGhhd3RlLCBJbmMuIC0gRm9yIGF1dGhvcml6ZWQgdXNlIG9ubHkxJDAi -BgNVBAMTG3RoYXd0ZSBQcmltYXJ5IFJvb3QgQ0EgLSBHMjAeFw0wNzExMDUwMDAw -MDBaFw0zODAxMTgyMzU5NTlaMIGEMQswCQYDVQQGEwJVUzEVMBMGA1UEChMMdGhh -d3RlLCBJbmMuMTgwNgYDVQQLEy8oYykgMjAwNyB0aGF3dGUsIEluYy4gLSBGb3Ig -YXV0aG9yaXplZCB1c2Ugb25seTEkMCIGA1UEAxMbdGhhd3RlIFByaW1hcnkgUm9v -dCBDQSAtIEcyMHYwEAYHKoZIzj0CAQYFK4EEACIDYgAEotWcgnuVnfFSeIf+iha/ -BebfowJPDQfGAFG6DAJSLSKkQjnE/o/qycG+1E3/n3qe4rF8mq2nhglzh9HnmuN6 -papu+7qzcMBniKI11KOasf2twu8x+qi58/sIxpHR+ymVo0IwQDAPBgNVHRMBAf8E -BTADAQH/MA4GA1UdDwEB/wQEAwIBBjAdBgNVHQ4EFgQUmtgAMADna3+FGO6Lts6K -DPgR4bswCgYIKoZIzj0EAwMDaQAwZgIxAN344FdHW6fmCsO99YCKlzUNG4k8VIZ3 -KMqh9HneteY4sPBlcIx/AlTCv//YoT7ZzwIxAMSNlPzcU9LcnXgWHxUzI1NS41ox -XZ3Krr0TKUQNJ1uo52icEvdYPy5yAlejj6EULg== ------END CERTIFICATE----- - -# Issuer: CN=thawte Primary Root CA - G3 O=thawte, Inc. OU=Certification Services Division/(c) 2008 thawte, Inc. - For authorized use only -# Subject: CN=thawte Primary Root CA - G3 O=thawte, Inc. OU=Certification Services Division/(c) 2008 thawte, Inc. - For authorized use only -# Label: "thawte Primary Root CA - G3" -# Serial: 127614157056681299805556476275995414779 -# MD5 Fingerprint: fb:1b:5d:43:8a:94:cd:44:c6:76:f2:43:4b:47:e7:31 -# SHA1 Fingerprint: f1:8b:53:8d:1b:e9:03:b6:a6:f0:56:43:5b:17:15:89:ca:f3:6b:f2 -# SHA256 Fingerprint: 4b:03:f4:58:07:ad:70:f2:1b:fc:2c:ae:71:c9:fd:e4:60:4c:06:4c:f5:ff:b6:86:ba:e5:db:aa:d7:fd:d3:4c ------BEGIN CERTIFICATE----- -MIIEKjCCAxKgAwIBAgIQYAGXt0an6rS0mtZLL/eQ+zANBgkqhkiG9w0BAQsFADCB -rjELMAkGA1UEBhMCVVMxFTATBgNVBAoTDHRoYXd0ZSwgSW5jLjEoMCYGA1UECxMf -Q2VydGlmaWNhdGlvbiBTZXJ2aWNlcyBEaXZpc2lvbjE4MDYGA1UECxMvKGMpIDIw -MDggdGhhd3RlLCBJbmMuIC0gRm9yIGF1dGhvcml6ZWQgdXNlIG9ubHkxJDAiBgNV -BAMTG3RoYXd0ZSBQcmltYXJ5IFJvb3QgQ0EgLSBHMzAeFw0wODA0MDIwMDAwMDBa -Fw0zNzEyMDEyMzU5NTlaMIGuMQswCQYDVQQGEwJVUzEVMBMGA1UEChMMdGhhd3Rl -LCBJbmMuMSgwJgYDVQQLEx9DZXJ0aWZpY2F0aW9uIFNlcnZpY2VzIERpdmlzaW9u -MTgwNgYDVQQLEy8oYykgMjAwOCB0aGF3dGUsIEluYy4gLSBGb3IgYXV0aG9yaXpl -ZCB1c2Ugb25seTEkMCIGA1UEAxMbdGhhd3RlIFByaW1hcnkgUm9vdCBDQSAtIEcz -MIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAsr8nLPvb2FvdeHsbnndm -gcs+vHyu86YnmjSjaDFxODNi5PNxZnmxqWWjpYvVj2AtP0LMqmsywCPLLEHd5N/8 -YZzic7IilRFDGF/Eth9XbAoFWCLINkw6fKXRz4aviKdEAhN0cXMKQlkC+BsUa0Lf -b1+6a4KinVvnSr0eAXLbS3ToO39/fR8EtCab4LRarEc9VbjXsCZSKAExQGbY2SS9 -9irY7CFJXJv2eul/VTV+lmuNk5Mny5K76qxAwJ/C+IDPXfRa3M50hqY+bAtTyr2S -zhkGcuYMXDhpxwTWvGzOW/b3aJzcJRVIiKHpqfiYnODz1TEoYRFsZ5aNOZnLwkUk -OQIDAQABo0IwQDAPBgNVHRMBAf8EBTADAQH/MA4GA1UdDwEB/wQEAwIBBjAdBgNV -HQ4EFgQUrWyqlGCc7eT/+j4KdCtjA/e2Wb8wDQYJKoZIhvcNAQELBQADggEBABpA -2JVlrAmSicY59BDlqQ5mU1143vokkbvnRFHfxhY0Cu9qRFHqKweKA3rD6z8KLFIW -oCtDuSWQP3CpMyVtRRooOyfPqsMpQhvfO0zAMzRbQYi/aytlryjvsvXDqmbOe1bu -t8jLZ8HJnBoYuMTDSQPxYA5QzUbF83d597YV4Djbxy8ooAw/dyZ02SUS2jHaGh7c -KUGRIjxpp7sC8rZcJwOJ9Abqm+RyguOhCcHpABnTPtRwa7pxpqpYrvS76Wy274fM -m7v/OeZWYdMKp8RcTGB7BXcmer/YB1IsYvdwY9k5vG8cwnncdimvzsUsZAReiDZu -MdRAGmI0Nj81Aa6sY6A= ------END CERTIFICATE----- - -# Issuer: CN=GeoTrust Primary Certification Authority - G2 O=GeoTrust Inc. OU=(c) 2007 GeoTrust Inc. - For authorized use only -# Subject: CN=GeoTrust Primary Certification Authority - G2 O=GeoTrust Inc. OU=(c) 2007 GeoTrust Inc. - For authorized use only -# Label: "GeoTrust Primary Certification Authority - G2" -# Serial: 80682863203381065782177908751794619243 -# MD5 Fingerprint: 01:5e:d8:6b:bd:6f:3d:8e:a1:31:f8:12:e0:98:73:6a -# SHA1 Fingerprint: 8d:17:84:d5:37:f3:03:7d:ec:70:fe:57:8b:51:9a:99:e6:10:d7:b0 -# SHA256 Fingerprint: 5e:db:7a:c4:3b:82:a0:6a:87:61:e8:d7:be:49:79:eb:f2:61:1f:7d:d7:9b:f9:1c:1c:6b:56:6a:21:9e:d7:66 ------BEGIN CERTIFICATE----- -MIICrjCCAjWgAwIBAgIQPLL0SAoA4v7rJDteYD7DazAKBggqhkjOPQQDAzCBmDEL -MAkGA1UEBhMCVVMxFjAUBgNVBAoTDUdlb1RydXN0IEluYy4xOTA3BgNVBAsTMChj -KSAyMDA3IEdlb1RydXN0IEluYy4gLSBGb3IgYXV0aG9yaXplZCB1c2Ugb25seTE2 -MDQGA1UEAxMtR2VvVHJ1c3QgUHJpbWFyeSBDZXJ0aWZpY2F0aW9uIEF1dGhvcml0 -eSAtIEcyMB4XDTA3MTEwNTAwMDAwMFoXDTM4MDExODIzNTk1OVowgZgxCzAJBgNV -BAYTAlVTMRYwFAYDVQQKEw1HZW9UcnVzdCBJbmMuMTkwNwYDVQQLEzAoYykgMjAw -NyBHZW9UcnVzdCBJbmMuIC0gRm9yIGF1dGhvcml6ZWQgdXNlIG9ubHkxNjA0BgNV -BAMTLUdlb1RydXN0IFByaW1hcnkgQ2VydGlmaWNhdGlvbiBBdXRob3JpdHkgLSBH -MjB2MBAGByqGSM49AgEGBSuBBAAiA2IABBWx6P0DFUPlrOuHNxFi79KDNlJ9RVcL -So17VDs6bl8VAsBQps8lL33KSLjHUGMcKiEIfJo22Av+0SbFWDEwKCXzXV2juLal -tJLtbCyf691DiaI8S0iRHVDsJt/WYC69IaNCMEAwDwYDVR0TAQH/BAUwAwEB/zAO -BgNVHQ8BAf8EBAMCAQYwHQYDVR0OBBYEFBVfNVdRVfslsq0DafwBo/q+EVXVMAoG -CCqGSM49BAMDA2cAMGQCMGSWWaboCd6LuvpaiIjwH5HTRqjySkwCY/tsXzjbLkGT -qQ7mndwxHLKgpxgceeHHNgIwOlavmnRs9vuD4DPTCF+hnMJbn0bWtsuRBmOiBucz -rD6ogRLQy7rQkgu2npaqBA+K ------END CERTIFICATE----- - -# Issuer: CN=VeriSign Universal Root Certification Authority O=VeriSign, Inc. OU=VeriSign Trust Network/(c) 2008 VeriSign, Inc. - For authorized use only -# Subject: CN=VeriSign Universal Root Certification Authority O=VeriSign, Inc. OU=VeriSign Trust Network/(c) 2008 VeriSign, Inc. - For authorized use only -# Label: "VeriSign Universal Root Certification Authority" -# Serial: 85209574734084581917763752644031726877 -# MD5 Fingerprint: 8e:ad:b5:01:aa:4d:81:e4:8c:1d:d1:e1:14:00:95:19 -# SHA1 Fingerprint: 36:79:ca:35:66:87:72:30:4d:30:a5:fb:87:3b:0f:a7:7b:b7:0d:54 -# SHA256 Fingerprint: 23:99:56:11:27:a5:71:25:de:8c:ef:ea:61:0d:df:2f:a0:78:b5:c8:06:7f:4e:82:82:90:bf:b8:60:e8:4b:3c ------BEGIN CERTIFICATE----- -MIIEuTCCA6GgAwIBAgIQQBrEZCGzEyEDDrvkEhrFHTANBgkqhkiG9w0BAQsFADCB -vTELMAkGA1UEBhMCVVMxFzAVBgNVBAoTDlZlcmlTaWduLCBJbmMuMR8wHQYDVQQL -ExZWZXJpU2lnbiBUcnVzdCBOZXR3b3JrMTowOAYDVQQLEzEoYykgMjAwOCBWZXJp -U2lnbiwgSW5jLiAtIEZvciBhdXRob3JpemVkIHVzZSBvbmx5MTgwNgYDVQQDEy9W -ZXJpU2lnbiBVbml2ZXJzYWwgUm9vdCBDZXJ0aWZpY2F0aW9uIEF1dGhvcml0eTAe -Fw0wODA0MDIwMDAwMDBaFw0zNzEyMDEyMzU5NTlaMIG9MQswCQYDVQQGEwJVUzEX -MBUGA1UEChMOVmVyaVNpZ24sIEluYy4xHzAdBgNVBAsTFlZlcmlTaWduIFRydXN0 -IE5ldHdvcmsxOjA4BgNVBAsTMShjKSAyMDA4IFZlcmlTaWduLCBJbmMuIC0gRm9y -IGF1dGhvcml6ZWQgdXNlIG9ubHkxODA2BgNVBAMTL1ZlcmlTaWduIFVuaXZlcnNh -bCBSb290IENlcnRpZmljYXRpb24gQXV0aG9yaXR5MIIBIjANBgkqhkiG9w0BAQEF -AAOCAQ8AMIIBCgKCAQEAx2E3XrEBNNti1xWb/1hajCMj1mCOkdeQmIN65lgZOIzF -9uVkhbSicfvtvbnazU0AtMgtc6XHaXGVHzk8skQHnOgO+k1KxCHfKWGPMiJhgsWH -H26MfF8WIFFE0XBPV+rjHOPMee5Y2A7Cs0WTwCznmhcrewA3ekEzeOEz4vMQGn+H -LL729fdC4uW/h2KJXwBL38Xd5HVEMkE6HnFuacsLdUYI0crSK5XQz/u5QGtkjFdN -/BMReYTtXlT2NJ8IAfMQJQYXStrxHXpma5hgZqTZ79IugvHw7wnqRMkVauIDbjPT -rJ9VAMf2CGqUuV/c4DPxhGD5WycRtPwW8rtWaoAljQIDAQABo4GyMIGvMA8GA1Ud -EwEB/wQFMAMBAf8wDgYDVR0PAQH/BAQDAgEGMG0GCCsGAQUFBwEMBGEwX6FdoFsw -WTBXMFUWCWltYWdlL2dpZjAhMB8wBwYFKw4DAhoEFI/l0xqGrI2Oa8PPgGrUSBgs -exkuMCUWI2h0dHA6Ly9sb2dvLnZlcmlzaWduLmNvbS92c2xvZ28uZ2lmMB0GA1Ud -DgQWBBS2d/ppSEefUxLVwuoHMnYH0ZcHGTANBgkqhkiG9w0BAQsFAAOCAQEASvj4 -sAPmLGd75JR3Y8xuTPl9Dg3cyLk1uXBPY/ok+myDjEedO2Pzmvl2MpWRsXe8rJq+ -seQxIcaBlVZaDrHC1LGmWazxY8u4TB1ZkErvkBYoH1quEPuBUDgMbMzxPcP1Y+Oz -4yHJJDnp/RVmRvQbEdBNc6N9Rvk97ahfYtTxP/jgdFcrGJ2BtMQo2pSXpXDrrB2+ -BxHw1dvd5Yzw1TKwg+ZX4o+/vqGqvz0dtdQ46tewXDpPaj+PwGZsY6rp2aQW9IHR -lRQOfc2VNNnSj3BzgXucfr2YYdhFh5iQxeuGMMY1v/D/w1WIg0vvBZIGcfK4mJO3 -7M2CYfE45k+XmCpajQ== ------END CERTIFICATE----- - -# Issuer: CN=VeriSign Class 3 Public Primary Certification Authority - G4 O=VeriSign, Inc. OU=VeriSign Trust Network/(c) 2007 VeriSign, Inc. - For authorized use only -# Subject: CN=VeriSign Class 3 Public Primary Certification Authority - G4 O=VeriSign, Inc. OU=VeriSign Trust Network/(c) 2007 VeriSign, Inc. - For authorized use only -# Label: "VeriSign Class 3 Public Primary Certification Authority - G4" -# Serial: 63143484348153506665311985501458640051 -# MD5 Fingerprint: 3a:52:e1:e7:fd:6f:3a:e3:6f:f3:6f:99:1b:f9:22:41 -# SHA1 Fingerprint: 22:d5:d8:df:8f:02:31:d1:8d:f7:9d:b7:cf:8a:2d:64:c9:3f:6c:3a -# SHA256 Fingerprint: 69:dd:d7:ea:90:bb:57:c9:3e:13:5d:c8:5e:a6:fc:d5:48:0b:60:32:39:bd:c4:54:fc:75:8b:2a:26:cf:7f:79 ------BEGIN CERTIFICATE----- -MIIDhDCCAwqgAwIBAgIQL4D+I4wOIg9IZxIokYesszAKBggqhkjOPQQDAzCByjEL -MAkGA1UEBhMCVVMxFzAVBgNVBAoTDlZlcmlTaWduLCBJbmMuMR8wHQYDVQQLExZW -ZXJpU2lnbiBUcnVzdCBOZXR3b3JrMTowOAYDVQQLEzEoYykgMjAwNyBWZXJpU2ln -biwgSW5jLiAtIEZvciBhdXRob3JpemVkIHVzZSBvbmx5MUUwQwYDVQQDEzxWZXJp -U2lnbiBDbGFzcyAzIFB1YmxpYyBQcmltYXJ5IENlcnRpZmljYXRpb24gQXV0aG9y -aXR5IC0gRzQwHhcNMDcxMTA1MDAwMDAwWhcNMzgwMTE4MjM1OTU5WjCByjELMAkG -A1UEBhMCVVMxFzAVBgNVBAoTDlZlcmlTaWduLCBJbmMuMR8wHQYDVQQLExZWZXJp -U2lnbiBUcnVzdCBOZXR3b3JrMTowOAYDVQQLEzEoYykgMjAwNyBWZXJpU2lnbiwg -SW5jLiAtIEZvciBhdXRob3JpemVkIHVzZSBvbmx5MUUwQwYDVQQDEzxWZXJpU2ln -biBDbGFzcyAzIFB1YmxpYyBQcmltYXJ5IENlcnRpZmljYXRpb24gQXV0aG9yaXR5 -IC0gRzQwdjAQBgcqhkjOPQIBBgUrgQQAIgNiAASnVnp8Utpkmw4tXNherJI9/gHm -GUo9FANL+mAnINmDiWn6VMaaGF5VKmTeBvaNSjutEDxlPZCIBIngMGGzrl0Bp3ve -fLK+ymVhAIau2o970ImtTR1ZmkGxvEeA3J5iw/mjgbIwga8wDwYDVR0TAQH/BAUw -AwEB/zAOBgNVHQ8BAf8EBAMCAQYwbQYIKwYBBQUHAQwEYTBfoV2gWzBZMFcwVRYJ -aW1hZ2UvZ2lmMCEwHzAHBgUrDgMCGgQUj+XTGoasjY5rw8+AatRIGCx7GS4wJRYj -aHR0cDovL2xvZ28udmVyaXNpZ24uY29tL3ZzbG9nby5naWYwHQYDVR0OBBYEFLMW -kf3upm7ktS5Jj4d4gYDs5bG1MAoGCCqGSM49BAMDA2gAMGUCMGYhDBgmYFo4e1ZC -4Kf8NoRRkSAsdk1DPcQdhCPQrNZ8NQbOzWm9kA3bbEhCHQ6qQgIxAJw9SDkjOVga -FRJZap7v1VmyHVIsmXHNxynfGyphe3HR3vPA5Q06Sqotp9iGKt0uEA== ------END CERTIFICATE----- - # Issuer: CN=NetLock Arany (Class Gold) F\u0151tan\xfas\xedtv\xe1ny O=NetLock Kft. OU=Tan\xfas\xedtv\xe1nykiad\xf3k (Certification Services) # Subject: CN=NetLock Arany (Class Gold) F\u0151tan\xfas\xedtv\xe1ny O=NetLock Kft. OU=Tan\xfas\xedtv\xe1nykiad\xf3k (Certification Services) # Label: "NetLock Arany (Class Gold) F\u0151tan\xfas\xedtv\xe1ny" @@ -1499,75 +763,6 @@ uLjbvrW5KfnaNwUASZQDhETnv0Mxz3WLJdH0pmT1kvarBes96aULNmLazAZfNou2 XjG4Kvte9nHfRCaexOYNkbQudZWAUWpLMKawYqGT8ZvYzsRjdT9ZR7E= -----END CERTIFICATE----- -# Issuer: CN=Staat der Nederlanden Root CA - G2 O=Staat der Nederlanden -# Subject: CN=Staat der Nederlanden Root CA - G2 O=Staat der Nederlanden -# Label: "Staat der Nederlanden Root CA - G2" -# Serial: 10000012 -# MD5 Fingerprint: 7c:a5:0f:f8:5b:9a:7d:6d:30:ae:54:5a:e3:42:a2:8a -# SHA1 Fingerprint: 59:af:82:79:91:86:c7:b4:75:07:cb:cf:03:57:46:eb:04:dd:b7:16 -# SHA256 Fingerprint: 66:8c:83:94:7d:a6:3b:72:4b:ec:e1:74:3c:31:a0:e6:ae:d0:db:8e:c5:b3:1b:e3:77:bb:78:4f:91:b6:71:6f ------BEGIN CERTIFICATE----- -MIIFyjCCA7KgAwIBAgIEAJiWjDANBgkqhkiG9w0BAQsFADBaMQswCQYDVQQGEwJO -TDEeMBwGA1UECgwVU3RhYXQgZGVyIE5lZGVybGFuZGVuMSswKQYDVQQDDCJTdGFh -dCBkZXIgTmVkZXJsYW5kZW4gUm9vdCBDQSAtIEcyMB4XDTA4MDMyNjExMTgxN1oX -DTIwMDMyNTExMDMxMFowWjELMAkGA1UEBhMCTkwxHjAcBgNVBAoMFVN0YWF0IGRl -ciBOZWRlcmxhbmRlbjErMCkGA1UEAwwiU3RhYXQgZGVyIE5lZGVybGFuZGVuIFJv -b3QgQ0EgLSBHMjCCAiIwDQYJKoZIhvcNAQEBBQADggIPADCCAgoCggIBAMVZ5291 -qj5LnLW4rJ4L5PnZyqtdj7U5EILXr1HgO+EASGrP2uEGQxGZqhQlEq0i6ABtQ8Sp -uOUfiUtnvWFI7/3S4GCI5bkYYCjDdyutsDeqN95kWSpGV+RLufg3fNU254DBtvPU -Z5uW6M7XxgpT0GtJlvOjCwV3SPcl5XCsMBQgJeN/dVrlSPhOewMHBPqCYYdu8DvE -pMfQ9XQ+pV0aCPKbJdL2rAQmPlU6Yiile7Iwr/g3wtG61jj99O9JMDeZJiFIhQGp -5Rbn3JBV3w/oOM2ZNyFPXfUib2rFEhZgF1XyZWampzCROME4HYYEhLoaJXhena/M -UGDWE4dS7WMfbWV9whUYdMrhfmQpjHLYFhN9C0lK8SgbIHRrxT3dsKpICT0ugpTN -GmXZK4iambwYfp/ufWZ8Pr2UuIHOzZgweMFvZ9C+X+Bo7d7iscksWXiSqt8rYGPy -5V6548r6f1CGPqI0GAwJaCgRHOThuVw+R7oyPxjMW4T182t0xHJ04eOLoEq9jWYv -6q012iDTiIJh8BIitrzQ1aTsr1SIJSQ8p22xcik/Plemf1WvbibG/ufMQFxRRIEK -eN5KzlW/HdXZt1bv8Hb/C3m1r737qWmRRpdogBQ2HbN/uymYNqUg+oJgYjOk7Na6 -B6duxc8UpufWkjTYgfX8HV2qXB72o007uPc5AgMBAAGjgZcwgZQwDwYDVR0TAQH/ -BAUwAwEB/zBSBgNVHSAESzBJMEcGBFUdIAAwPzA9BggrBgEFBQcCARYxaHR0cDov -L3d3dy5wa2lvdmVyaGVpZC5ubC9wb2xpY2llcy9yb290LXBvbGljeS1HMjAOBgNV -HQ8BAf8EBAMCAQYwHQYDVR0OBBYEFJFoMocVHYnitfGsNig0jQt8YojrMA0GCSqG -SIb3DQEBCwUAA4ICAQCoQUpnKpKBglBu4dfYszk78wIVCVBR7y29JHuIhjv5tLyS -CZa59sCrI2AGeYwRTlHSeYAz+51IvuxBQ4EffkdAHOV6CMqqi3WtFMTC6GY8ggen -5ieCWxjmD27ZUD6KQhgpxrRW/FYQoAUXvQwjf/ST7ZwaUb7dRUG/kSS0H4zpX897 -IZmflZ85OkYcbPnNe5yQzSipx6lVu6xiNGI1E0sUOlWDuYaNkqbG9AclVMwWVxJK -gnjIFNkXgiYtXSAfea7+1HAWFpWD2DU5/1JddRwWxRNVz0fMdWVSSt7wsKfkCpYL -+63C4iWEst3kvX5ZbJvw8NjnyvLplzh+ib7M+zkXYT9y2zqR2GUBGR2tUKRXCnxL -vJxxcypFURmFzI79R6d0lR2o0a9OF7FpJsKqeFdbxU2n5Z4FF5TKsl+gSRiNNOkm -bEgeqmiSBeGCc1qb3AdbCG19ndeNIdn8FCCqwkXfP+cAslHkwvgFuXkajDTznlvk -N1trSt8sV4pAWja63XVECDdCcAz+3F4hoKOKwJCcaNpQ5kUQR3i2TtJlycM33+FC -Y7BXN0Ute4qcvwXqZVUz9zkQxSgqIXobisQk+T8VyJoVIPVVYpbtbZNQvOSqeK3Z -ywplh6ZmwcSBo3c6WB4L7oOLnR7SUqTMHW+wmG2UMbX4cQrcufx9MmDm66+KAQ== ------END CERTIFICATE----- - -# Issuer: CN=Hongkong Post Root CA 1 O=Hongkong Post -# Subject: CN=Hongkong Post Root CA 1 O=Hongkong Post -# Label: "Hongkong Post Root CA 1" -# Serial: 1000 -# MD5 Fingerprint: a8:0d:6f:39:78:b9:43:6d:77:42:6d:98:5a:cc:23:ca -# SHA1 Fingerprint: d6:da:a8:20:8d:09:d2:15:4d:24:b5:2f:cb:34:6e:b2:58:b2:8a:58 -# SHA256 Fingerprint: f9:e6:7d:33:6c:51:00:2a:c0:54:c6:32:02:2d:66:dd:a2:e7:e3:ff:f1:0a:d0:61:ed:31:d8:bb:b4:10:cf:b2 ------BEGIN CERTIFICATE----- -MIIDMDCCAhigAwIBAgICA+gwDQYJKoZIhvcNAQEFBQAwRzELMAkGA1UEBhMCSEsx -FjAUBgNVBAoTDUhvbmdrb25nIFBvc3QxIDAeBgNVBAMTF0hvbmdrb25nIFBvc3Qg -Um9vdCBDQSAxMB4XDTAzMDUxNTA1MTMxNFoXDTIzMDUxNTA0NTIyOVowRzELMAkG -A1UEBhMCSEsxFjAUBgNVBAoTDUhvbmdrb25nIFBvc3QxIDAeBgNVBAMTF0hvbmdr -b25nIFBvc3QgUm9vdCBDQSAxMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKC -AQEArP84tulmAknjorThkPlAj3n54r15/gK97iSSHSL22oVyaf7XPwnU3ZG1ApzQ -jVrhVcNQhrkpJsLj2aDxaQMoIIBFIi1WpztUlVYiWR8o3x8gPW2iNr4joLFutbEn -PzlTCeqrauh0ssJlXI6/fMN4hM2eFvz1Lk8gKgifd/PFHsSaUmYeSF7jEAaPIpjh -ZY4bXSNmO7ilMlHIhqqhqZ5/dpTCpmy3QfDVyAY45tQM4vM7TG1QjMSDJ8EThFk9 -nnV0ttgCXjqQesBCNnLsak3c78QA3xMYV18meMjWCnl3v/evt3a5pQuEF10Q6m/h -q5URX208o1xNg1vysxmKgIsLhwIDAQABoyYwJDASBgNVHRMBAf8ECDAGAQH/AgED -MA4GA1UdDwEB/wQEAwIBxjANBgkqhkiG9w0BAQUFAAOCAQEADkbVPK7ih9legYsC -mEEIjEy82tvuJxuC52pF7BaLT4Wg87JwvVqWuspube5Gi27nKi6Wsxkz67SfqLI3 -7piol7Yutmcn1KZJ/RyTZXaeQi/cImyaT/JaFTmxcdcrUehtHJjA2Sr0oYJ71clB -oiMBdDhViw+5LmeiIAQ32pwL0xch4I+XeTRvhEgCIDMb5jREn5Fw9IBehEPCKdJs -EhTkYY2sEJCehFC78JZvRZ+K88psT/oROhUVRsPNH4NbLUES7VBnQRM9IauUiqpO -fMGx+6fWtScvl6tu4B3i0RwsH0Ti/L6RoZz71ilTc4afU9hDDl3WY4JxHYB0yvbi -AmvZWg== ------END CERTIFICATE----- - # Issuer: CN=SecureSign RootCA11 O=Japan Certification Services, Inc. # Subject: CN=SecureSign RootCA11 O=Japan Certification Services, Inc. # Label: "SecureSign RootCA11" @@ -1658,49 +853,6 @@ Mx86OyXShkDOOyyGeMlhLxS67ttVb9+E7gUJTb0o2HLO02JQZR7rkpeDMdmztcpH WD9f -----END CERTIFICATE----- -# Issuer: CN=Autoridad de Certificacion Firmaprofesional CIF A62634068 -# Subject: CN=Autoridad de Certificacion Firmaprofesional CIF A62634068 -# Label: "Autoridad de Certificacion Firmaprofesional CIF A62634068" -# Serial: 6047274297262753887 -# MD5 Fingerprint: 73:3a:74:7a:ec:bb:a3:96:a6:c2:e4:e2:c8:9b:c0:c3 -# SHA1 Fingerprint: ae:c5:fb:3f:c8:e1:bf:c4:e5:4f:03:07:5a:9a:e8:00:b7:f7:b6:fa -# SHA256 Fingerprint: 04:04:80:28:bf:1f:28:64:d4:8f:9a:d4:d8:32:94:36:6a:82:88:56:55:3f:3b:14:30:3f:90:14:7f:5d:40:ef ------BEGIN CERTIFICATE----- -MIIGFDCCA/ygAwIBAgIIU+w77vuySF8wDQYJKoZIhvcNAQEFBQAwUTELMAkGA1UE -BhMCRVMxQjBABgNVBAMMOUF1dG9yaWRhZCBkZSBDZXJ0aWZpY2FjaW9uIEZpcm1h -cHJvZmVzaW9uYWwgQ0lGIEE2MjYzNDA2ODAeFw0wOTA1MjAwODM4MTVaFw0zMDEy -MzEwODM4MTVaMFExCzAJBgNVBAYTAkVTMUIwQAYDVQQDDDlBdXRvcmlkYWQgZGUg -Q2VydGlmaWNhY2lvbiBGaXJtYXByb2Zlc2lvbmFsIENJRiBBNjI2MzQwNjgwggIi -MA0GCSqGSIb3DQEBAQUAA4ICDwAwggIKAoICAQDKlmuO6vj78aI14H9M2uDDUtd9 -thDIAl6zQyrET2qyyhxdKJp4ERppWVevtSBC5IsP5t9bpgOSL/UR5GLXMnE42QQM -cas9UX4PB99jBVzpv5RvwSmCwLTaUbDBPLutN0pcyvFLNg4kq7/DhHf9qFD0sefG -L9ItWY16Ck6WaVICqjaY7Pz6FIMMNx/Jkjd/14Et5cS54D40/mf0PmbR0/RAz15i -NA9wBj4gGFrO93IbJWyTdBSTo3OxDqqHECNZXyAFGUftaI6SEspd/NYrspI8IM/h -X68gvqB2f3bl7BqGYTM+53u0P6APjqK5am+5hyZvQWyIplD9amML9ZMWGxmPsu2b -m8mQ9QEM3xk9Dz44I8kvjwzRAv4bVdZO0I08r0+k8/6vKtMFnXkIoctXMbScyJCy -Z/QYFpM6/EfY0XiWMR+6KwxfXZmtY4laJCB22N/9q06mIqqdXuYnin1oKaPnirja -EbsXLZmdEyRG98Xi2J+Of8ePdG1asuhy9azuJBCtLxTa/y2aRnFHvkLfuwHb9H/T -KI8xWVvTyQKmtFLKbpf7Q8UIJm+K9Lv9nyiqDdVF8xM6HdjAeI9BZzwelGSuewvF -6NkBiDkal4ZkQdU7hwxu+g/GvUgUvzlN1J5Bto+WHWOWk9mVBngxaJ43BjuAiUVh -OSPHG0SjFeUc+JIwuwIDAQABo4HvMIHsMBIGA1UdEwEB/wQIMAYBAf8CAQEwDgYD -VR0PAQH/BAQDAgEGMB0GA1UdDgQWBBRlzeurNR4APn7VdMActHNHDhpkLzCBpgYD -VR0gBIGeMIGbMIGYBgRVHSAAMIGPMC8GCCsGAQUFBwIBFiNodHRwOi8vd3d3LmZp -cm1hcHJvZmVzaW9uYWwuY29tL2NwczBcBggrBgEFBQcCAjBQHk4AUABhAHMAZQBv -ACAAZABlACAAbABhACAAQgBvAG4AYQBuAG8AdgBhACAANAA3ACAAQgBhAHIAYwBl -AGwAbwBuAGEAIAAwADgAMAAxADcwDQYJKoZIhvcNAQEFBQADggIBABd9oPm03cXF -661LJLWhAqvdpYhKsg9VSytXjDvlMd3+xDLx51tkljYyGOylMnfX40S2wBEqgLk9 -am58m9Ot/MPWo+ZkKXzR4Tgegiv/J2Wv+xYVxC5xhOW1//qkR71kMrv2JYSiJ0L1 -ILDCExARzRAVukKQKtJE4ZYm6zFIEv0q2skGz3QeqUvVhyj5eTSSPi5E6PaPT481 -PyWzOdxjKpBrIF/EUhJOlywqrJ2X3kjyo2bbwtKDlaZmp54lD+kLM5FlClrD2VQS -3a/DTg4fJl4N3LON7NWBcN7STyQF82xO9UxJZo3R/9ILJUFI/lGExkKvgATP0H5k -SeTy36LssUzAKh3ntLFlosS88Zj0qnAHY7S42jtM+kAiMFsRpvAFDsYCA0irhpuF -3dvd6qJ2gHN99ZwExEWN57kci57q13XRcrHedUTnQn3iV2t93Jm8PYMo6oCTjcVM -ZcFwgbg4/EMxsvYDNEeyrPsiBsse3RdHHF9mudMaotoRsaS8I8nkvof/uZS2+F0g -StRf571oe2XyFR7SOqkt6dhrJKyXWERHrVkY8SFlcN7ONGCoQPHzPKTDKCOM/icz -Q0CgFzzr6juwcqajuUpLXhZI9LK8yIySxZ2frHI2vDSANGupi5LAuBft7HZT9SQB -jLMi6Et8Vcad+qMUu2WFbm5PEn4KPJ2V ------END CERTIFICATE----- - # Issuer: CN=Izenpe.com O=IZENPE S.A. # Subject: CN=Izenpe.com O=IZENPE S.A. # Label: "Izenpe.com" @@ -1743,105 +895,6 @@ naM8THLCV8Sg1Mw4J87VBp6iSNnpn86CcDaTmjvfliHjWbcM2pE38P1ZWrOZyGls QyYBNWNgVYkDOnXYukrZVP/u3oDYLdE41V4tC5h9Pmzb/CaIxw== -----END CERTIFICATE----- -# Issuer: CN=Chambers of Commerce Root - 2008 O=AC Camerfirma S.A. -# Subject: CN=Chambers of Commerce Root - 2008 O=AC Camerfirma S.A. -# Label: "Chambers of Commerce Root - 2008" -# Serial: 11806822484801597146 -# MD5 Fingerprint: 5e:80:9e:84:5a:0e:65:0b:17:02:f3:55:18:2a:3e:d7 -# SHA1 Fingerprint: 78:6a:74:ac:76:ab:14:7f:9c:6a:30:50:ba:9e:a8:7e:fe:9a:ce:3c -# SHA256 Fingerprint: 06:3e:4a:fa:c4:91:df:d3:32:f3:08:9b:85:42:e9:46:17:d8:93:d7:fe:94:4e:10:a7:93:7e:e2:9d:96:93:c0 ------BEGIN CERTIFICATE----- -MIIHTzCCBTegAwIBAgIJAKPaQn6ksa7aMA0GCSqGSIb3DQEBBQUAMIGuMQswCQYD -VQQGEwJFVTFDMEEGA1UEBxM6TWFkcmlkIChzZWUgY3VycmVudCBhZGRyZXNzIGF0 -IHd3dy5jYW1lcmZpcm1hLmNvbS9hZGRyZXNzKTESMBAGA1UEBRMJQTgyNzQzMjg3 -MRswGQYDVQQKExJBQyBDYW1lcmZpcm1hIFMuQS4xKTAnBgNVBAMTIENoYW1iZXJz -IG9mIENvbW1lcmNlIFJvb3QgLSAyMDA4MB4XDTA4MDgwMTEyMjk1MFoXDTM4MDcz -MTEyMjk1MFowga4xCzAJBgNVBAYTAkVVMUMwQQYDVQQHEzpNYWRyaWQgKHNlZSBj -dXJyZW50IGFkZHJlc3MgYXQgd3d3LmNhbWVyZmlybWEuY29tL2FkZHJlc3MpMRIw -EAYDVQQFEwlBODI3NDMyODcxGzAZBgNVBAoTEkFDIENhbWVyZmlybWEgUy5BLjEp -MCcGA1UEAxMgQ2hhbWJlcnMgb2YgQ29tbWVyY2UgUm9vdCAtIDIwMDgwggIiMA0G -CSqGSIb3DQEBAQUAA4ICDwAwggIKAoICAQCvAMtwNyuAWko6bHiUfaN/Gh/2NdW9 -28sNRHI+JrKQUrpjOyhYb6WzbZSm891kDFX29ufyIiKAXuFixrYp4YFs8r/lfTJq -VKAyGVn+H4vXPWCGhSRv4xGzdz4gljUha7MI2XAuZPeEklPWDrCQiorjh40G072Q -DuKZoRuGDtqaCrsLYVAGUvGef3bsyw/QHg3PmTA9HMRFEFis1tPo1+XqxQEHd9ZR -5gN/ikilTWh1uem8nk4ZcfUyS5xtYBkL+8ydddy/Js2Pk3g5eXNeJQ7KXOt3EgfL -ZEFHcpOrUMPrCXZkNNI5t3YRCQ12RcSprj1qr7V9ZS+UWBDsXHyvfuK2GNnQm05a -Sd+pZgvMPMZ4fKecHePOjlO+Bd5gD2vlGts/4+EhySnB8esHnFIbAURRPHsl18Tl -UlRdJQfKFiC4reRB7noI/plvg6aRArBsNlVq5331lubKgdaX8ZSD6e2wsWsSaR6s -+12pxZjptFtYer49okQ6Y1nUCyXeG0+95QGezdIp1Z8XGQpvvwyQ0wlf2eOKNcx5 -Wk0ZN5K3xMGtr/R5JJqyAQuxr1yW84Ay+1w9mPGgP0revq+ULtlVmhduYJ1jbLhj -ya6BXBg14JC7vjxPNyK5fuvPnnchpj04gftI2jE9K+OJ9dC1vX7gUMQSibMjmhAx -hduub+84Mxh2EQIDAQABo4IBbDCCAWgwEgYDVR0TAQH/BAgwBgEB/wIBDDAdBgNV -HQ4EFgQU+SSsD7K1+HnA+mCIG8TZTQKeFxkwgeMGA1UdIwSB2zCB2IAU+SSsD7K1 -+HnA+mCIG8TZTQKeFxmhgbSkgbEwga4xCzAJBgNVBAYTAkVVMUMwQQYDVQQHEzpN -YWRyaWQgKHNlZSBjdXJyZW50IGFkZHJlc3MgYXQgd3d3LmNhbWVyZmlybWEuY29t -L2FkZHJlc3MpMRIwEAYDVQQFEwlBODI3NDMyODcxGzAZBgNVBAoTEkFDIENhbWVy -ZmlybWEgUy5BLjEpMCcGA1UEAxMgQ2hhbWJlcnMgb2YgQ29tbWVyY2UgUm9vdCAt -IDIwMDiCCQCj2kJ+pLGu2jAOBgNVHQ8BAf8EBAMCAQYwPQYDVR0gBDYwNDAyBgRV -HSAAMCowKAYIKwYBBQUHAgEWHGh0dHA6Ly9wb2xpY3kuY2FtZXJmaXJtYS5jb20w -DQYJKoZIhvcNAQEFBQADggIBAJASryI1wqM58C7e6bXpeHxIvj99RZJe6dqxGfwW -PJ+0W2aeaufDuV2I6A+tzyMP3iU6XsxPpcG1Lawk0lgH3qLPaYRgM+gQDROpI9CF -5Y57pp49chNyM/WqfcZjHwj0/gF/JM8rLFQJ3uIrbZLGOU8W6jx+ekbURWpGqOt1 -glanq6B8aBMz9p0w8G8nOSQjKpD9kCk18pPfNKXG9/jvjA9iSnyu0/VU+I22mlaH -FoI6M6taIgj3grrqLuBHmrS1RaMFO9ncLkVAO+rcf+g769HsJtg1pDDFOqxXnrN2 -pSB7+R5KBWIBpih1YJeSDW4+TTdDDZIVnBgizVGZoCkaPF+KMjNbMMeJL0eYD6MD -xvbxrN8y8NmBGuScvfaAFPDRLLmF9dijscilIeUcE5fuDr3fKanvNFNb0+RqE4QG -tjICxFKuItLcsiFCGtpA8CnJ7AoMXOLQusxI0zcKzBIKinmwPQN/aUv0NCB9szTq -jktk9T79syNnFQ0EuPAtwQlRPLJsFfClI9eDdOTlLsn+mCdCxqvGnrDQWzilm1De -fhiYtUU79nm06PcaewaD+9CL2rvHvRirCG88gGtAPxkZumWK5r7VXNM21+9AUiRg -OGcEMeyP84LG3rlV8zsxkVrctQgVrXYlCg17LofiDKYGvCYQbTed7N14jHyAxfDZ -d0jQ ------END CERTIFICATE----- - -# Issuer: CN=Global Chambersign Root - 2008 O=AC Camerfirma S.A. -# Subject: CN=Global Chambersign Root - 2008 O=AC Camerfirma S.A. -# Label: "Global Chambersign Root - 2008" -# Serial: 14541511773111788494 -# MD5 Fingerprint: 9e:80:ff:78:01:0c:2e:c1:36:bd:fe:96:90:6e:08:f3 -# SHA1 Fingerprint: 4a:bd:ee:ec:95:0d:35:9c:89:ae:c7:52:a1:2c:5b:29:f6:d6:aa:0c -# SHA256 Fingerprint: 13:63:35:43:93:34:a7:69:80:16:a0:d3:24:de:72:28:4e:07:9d:7b:52:20:bb:8f:bd:74:78:16:ee:be:ba:ca ------BEGIN CERTIFICATE----- -MIIHSTCCBTGgAwIBAgIJAMnN0+nVfSPOMA0GCSqGSIb3DQEBBQUAMIGsMQswCQYD -VQQGEwJFVTFDMEEGA1UEBxM6TWFkcmlkIChzZWUgY3VycmVudCBhZGRyZXNzIGF0 -IHd3dy5jYW1lcmZpcm1hLmNvbS9hZGRyZXNzKTESMBAGA1UEBRMJQTgyNzQzMjg3 -MRswGQYDVQQKExJBQyBDYW1lcmZpcm1hIFMuQS4xJzAlBgNVBAMTHkdsb2JhbCBD -aGFtYmVyc2lnbiBSb290IC0gMjAwODAeFw0wODA4MDExMjMxNDBaFw0zODA3MzEx -MjMxNDBaMIGsMQswCQYDVQQGEwJFVTFDMEEGA1UEBxM6TWFkcmlkIChzZWUgY3Vy -cmVudCBhZGRyZXNzIGF0IHd3dy5jYW1lcmZpcm1hLmNvbS9hZGRyZXNzKTESMBAG -A1UEBRMJQTgyNzQzMjg3MRswGQYDVQQKExJBQyBDYW1lcmZpcm1hIFMuQS4xJzAl -BgNVBAMTHkdsb2JhbCBDaGFtYmVyc2lnbiBSb290IC0gMjAwODCCAiIwDQYJKoZI -hvcNAQEBBQADggIPADCCAgoCggIBAMDfVtPkOpt2RbQT2//BthmLN0EYlVJH6xed -KYiONWwGMi5HYvNJBL99RDaxccy9Wglz1dmFRP+RVyXfXjaOcNFccUMd2drvXNL7 -G706tcuto8xEpw2uIRU/uXpbknXYpBI4iRmKt4DS4jJvVpyR1ogQC7N0ZJJ0YPP2 -zxhPYLIj0Mc7zmFLmY/CDNBAspjcDahOo7kKrmCgrUVSY7pmvWjg+b4aqIG7HkF4 -ddPB/gBVsIdU6CeQNR1MM62X/JcumIS/LMmjv9GYERTtY/jKmIhYF5ntRQOXfjyG -HoiMvvKRhI9lNNgATH23MRdaKXoKGCQwoze1eqkBfSbW+Q6OWfH9GzO1KTsXO0G2 -Id3UwD2ln58fQ1DJu7xsepeY7s2MH/ucUa6LcL0nn3HAa6x9kGbo1106DbDVwo3V -yJ2dwW3Q0L9R5OP4wzg2rtandeavhENdk5IMagfeOx2YItaswTXbo6Al/3K1dh3e -beksZixShNBFks4c5eUzHdwHU1SjqoI7mjcv3N2gZOnm3b2u/GSFHTynyQbehP9r -6GsaPMWis0L7iwk+XwhSx2LE1AVxv8Rk5Pihg+g+EpuoHtQ2TS9x9o0o9oOpE9Jh -wZG7SMA0j0GMS0zbaRL/UJScIINZc+18ofLx/d33SdNDWKBWY8o9PeU1VlnpDsog -zCtLkykPAgMBAAGjggFqMIIBZjASBgNVHRMBAf8ECDAGAQH/AgEMMB0GA1UdDgQW -BBS5CcqcHtvTbDprru1U8VuTBjUuXjCB4QYDVR0jBIHZMIHWgBS5CcqcHtvTbDpr -ru1U8VuTBjUuXqGBsqSBrzCBrDELMAkGA1UEBhMCRVUxQzBBBgNVBAcTOk1hZHJp -ZCAoc2VlIGN1cnJlbnQgYWRkcmVzcyBhdCB3d3cuY2FtZXJmaXJtYS5jb20vYWRk -cmVzcykxEjAQBgNVBAUTCUE4Mjc0MzI4NzEbMBkGA1UEChMSQUMgQ2FtZXJmaXJt -YSBTLkEuMScwJQYDVQQDEx5HbG9iYWwgQ2hhbWJlcnNpZ24gUm9vdCAtIDIwMDiC -CQDJzdPp1X0jzjAOBgNVHQ8BAf8EBAMCAQYwPQYDVR0gBDYwNDAyBgRVHSAAMCow -KAYIKwYBBQUHAgEWHGh0dHA6Ly9wb2xpY3kuY2FtZXJmaXJtYS5jb20wDQYJKoZI -hvcNAQEFBQADggIBAICIf3DekijZBZRG/5BXqfEv3xoNa/p8DhxJJHkn2EaqbylZ -UohwEurdPfWbU1Rv4WCiqAm57OtZfMY18dwY6fFn5a+6ReAJ3spED8IXDneRRXoz -X1+WLGiLwUePmJs9wOzL9dWCkoQ10b42OFZyMVtHLaoXpGNR6woBrX/sdZ7LoR/x -fxKxueRkf2fWIyr0uDldmOghp+G9PUIadJpwr2hsUF1Jz//7Dl3mLEfXgTpZALVz -a2Mg9jFFCDkO9HB+QHBaP9BrQql0PSgvAm11cpUJjUhjxsYjV5KTXjXBjfkK9yyd -Yhz2rXzdpjEetrHHfoUm+qRqtdpjMNHvkzeyZi99Bffnt0uYlDXA2TopwZ2yUDMd -SqlapskD7+3056huirRXhOukP9DuqqqHW2Pok+JrqNS4cnhrG+055F3Lm6qH1U9O -AP7Zap88MQ8oAgF9mOinsKJknnn4SPIVqczmyETrP3iZ8ntxPjzxmKfFGBI/5rso -M0LpRQp8bfKGeS/Fghl9CYl8slR2iK7ewfPM4W7bMdaTrpmg7yVqc5iJWzouE4ge -v8CSlDQb4ye3ix5vQv/n6TebUB0tovkC7stYWDpxvGjjqsGvHCgfotwjZT+B6q6Z -09gwzxMNTxXJhLynSC34MCN32EZLeW32jO06f2ARePTpm67VVMB0gNELQp/B ------END CERTIFICATE----- - # Issuer: CN=Go Daddy Root Certificate Authority - G2 O=GoDaddy.com, Inc. # Subject: CN=Go Daddy Root Certificate Authority - G2 O=GoDaddy.com, Inc. # Label: "Go Daddy Root Certificate Authority - G2" @@ -2140,78 +1193,6 @@ t/2jioSgrGK+KwmHNPBqAbubKVY8/gA3zyNs8U6qtnRGEmyR7jTV7JqR50S+kDFy SjnRBUkLp7Y3gaVdjKozXoEofKd9J+sAro03 -----END CERTIFICATE----- -# Issuer: CN=EC-ACC O=Agencia Catalana de Certificacio (NIF Q-0801176-I) OU=Serveis Publics de Certificacio/Vegeu https://www.catcert.net/verarrel (c)03/Jerarquia Entitats de Certificacio Catalanes -# Subject: CN=EC-ACC O=Agencia Catalana de Certificacio (NIF Q-0801176-I) OU=Serveis Publics de Certificacio/Vegeu https://www.catcert.net/verarrel (c)03/Jerarquia Entitats de Certificacio Catalanes -# Label: "EC-ACC" -# Serial: -23701579247955709139626555126524820479 -# MD5 Fingerprint: eb:f5:9d:29:0d:61:f9:42:1f:7c:c2:ba:6d:e3:15:09 -# SHA1 Fingerprint: 28:90:3a:63:5b:52:80:fa:e6:77:4c:0b:6d:a7:d6:ba:a6:4a:f2:e8 -# SHA256 Fingerprint: 88:49:7f:01:60:2f:31:54:24:6a:e2:8c:4d:5a:ef:10:f1:d8:7e:bb:76:62:6f:4a:e0:b7:f9:5b:a7:96:87:99 ------BEGIN CERTIFICATE----- -MIIFVjCCBD6gAwIBAgIQ7is969Qh3hSoYqwE893EATANBgkqhkiG9w0BAQUFADCB -8zELMAkGA1UEBhMCRVMxOzA5BgNVBAoTMkFnZW5jaWEgQ2F0YWxhbmEgZGUgQ2Vy -dGlmaWNhY2lvIChOSUYgUS0wODAxMTc2LUkpMSgwJgYDVQQLEx9TZXJ2ZWlzIFB1 -YmxpY3MgZGUgQ2VydGlmaWNhY2lvMTUwMwYDVQQLEyxWZWdldSBodHRwczovL3d3 -dy5jYXRjZXJ0Lm5ldC92ZXJhcnJlbCAoYykwMzE1MDMGA1UECxMsSmVyYXJxdWlh -IEVudGl0YXRzIGRlIENlcnRpZmljYWNpbyBDYXRhbGFuZXMxDzANBgNVBAMTBkVD -LUFDQzAeFw0wMzAxMDcyMzAwMDBaFw0zMTAxMDcyMjU5NTlaMIHzMQswCQYDVQQG -EwJFUzE7MDkGA1UEChMyQWdlbmNpYSBDYXRhbGFuYSBkZSBDZXJ0aWZpY2FjaW8g -KE5JRiBRLTA4MDExNzYtSSkxKDAmBgNVBAsTH1NlcnZlaXMgUHVibGljcyBkZSBD -ZXJ0aWZpY2FjaW8xNTAzBgNVBAsTLFZlZ2V1IGh0dHBzOi8vd3d3LmNhdGNlcnQu -bmV0L3ZlcmFycmVsIChjKTAzMTUwMwYDVQQLEyxKZXJhcnF1aWEgRW50aXRhdHMg -ZGUgQ2VydGlmaWNhY2lvIENhdGFsYW5lczEPMA0GA1UEAxMGRUMtQUNDMIIBIjAN -BgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAsyLHT+KXQpWIR4NA9h0X84NzJB5R -85iKw5K4/0CQBXCHYMkAqbWUZRkiFRfCQ2xmRJoNBD45b6VLeqpjt4pEndljkYRm -4CgPukLjbo73FCeTae6RDqNfDrHrZqJyTxIThmV6PttPB/SnCWDaOkKZx7J/sxaV -HMf5NLWUhdWZXqBIoH7nF2W4onW4HvPlQn2v7fOKSGRdghST2MDk/7NQcvJ29rNd -QlB50JQ+awwAvthrDk4q7D7SzIKiGGUzE3eeml0aE9jD2z3Il3rucO2n5nzbcc8t -lGLfbdb1OL4/pYUKGbio2Al1QnDE6u/LDsg0qBIimAy4E5S2S+zw0JDnJwIDAQAB -o4HjMIHgMB0GA1UdEQQWMBSBEmVjX2FjY0BjYXRjZXJ0Lm5ldDAPBgNVHRMBAf8E -BTADAQH/MA4GA1UdDwEB/wQEAwIBBjAdBgNVHQ4EFgQUoMOLRKo3pUW/l4Ba0fF4 -opvpXY0wfwYDVR0gBHgwdjB0BgsrBgEEAfV4AQMBCjBlMCwGCCsGAQUFBwIBFiBo -dHRwczovL3d3dy5jYXRjZXJ0Lm5ldC92ZXJhcnJlbDA1BggrBgEFBQcCAjApGidW -ZWdldSBodHRwczovL3d3dy5jYXRjZXJ0Lm5ldC92ZXJhcnJlbCAwDQYJKoZIhvcN -AQEFBQADggEBAKBIW4IB9k1IuDlVNZyAelOZ1Vr/sXE7zDkJlF7W2u++AVtd0x7Y -/X1PzaBB4DSTv8vihpw3kpBWHNzrKQXlxJ7HNd+KDM3FIUPpqojlNcAZQmNaAl6k -SBg6hW/cnbw/nZzBh7h6YQjpdwt/cKt63dmXLGQehb+8dJahw3oS7AwaboMMPOhy -Rp/7SNVel+axofjk70YllJyJ22k4vuxcDlbHZVHlUIiIv0LVKz3l+bqeLrPK9HOS -Agu+TGbrIP65y7WZf+a2E/rKS03Z7lNGBjvGTq2TWoF+bCpLagVFjPIhpDGQh2xl -nJ2lYJU6Un/10asIbvPuW/mIPX64b24D5EI= ------END CERTIFICATE----- - -# Issuer: CN=Hellenic Academic and Research Institutions RootCA 2011 O=Hellenic Academic and Research Institutions Cert. Authority -# Subject: CN=Hellenic Academic and Research Institutions RootCA 2011 O=Hellenic Academic and Research Institutions Cert. Authority -# Label: "Hellenic Academic and Research Institutions RootCA 2011" -# Serial: 0 -# MD5 Fingerprint: 73:9f:4c:4b:73:5b:79:e9:fa:ba:1c:ef:6e:cb:d5:c9 -# SHA1 Fingerprint: fe:45:65:9b:79:03:5b:98:a1:61:b5:51:2e:ac:da:58:09:48:22:4d -# SHA256 Fingerprint: bc:10:4f:15:a4:8b:e7:09:dc:a5:42:a7:e1:d4:b9:df:6f:05:45:27:e8:02:ea:a9:2d:59:54:44:25:8a:fe:71 ------BEGIN CERTIFICATE----- -MIIEMTCCAxmgAwIBAgIBADANBgkqhkiG9w0BAQUFADCBlTELMAkGA1UEBhMCR1Ix -RDBCBgNVBAoTO0hlbGxlbmljIEFjYWRlbWljIGFuZCBSZXNlYXJjaCBJbnN0aXR1 -dGlvbnMgQ2VydC4gQXV0aG9yaXR5MUAwPgYDVQQDEzdIZWxsZW5pYyBBY2FkZW1p -YyBhbmQgUmVzZWFyY2ggSW5zdGl0dXRpb25zIFJvb3RDQSAyMDExMB4XDTExMTIw -NjEzNDk1MloXDTMxMTIwMTEzNDk1MlowgZUxCzAJBgNVBAYTAkdSMUQwQgYDVQQK -EztIZWxsZW5pYyBBY2FkZW1pYyBhbmQgUmVzZWFyY2ggSW5zdGl0dXRpb25zIENl -cnQuIEF1dGhvcml0eTFAMD4GA1UEAxM3SGVsbGVuaWMgQWNhZGVtaWMgYW5kIFJl -c2VhcmNoIEluc3RpdHV0aW9ucyBSb290Q0EgMjAxMTCCASIwDQYJKoZIhvcNAQEB -BQADggEPADCCAQoCggEBAKlTAOMupvaO+mDYLZU++CwqVE7NuYRhlFhPjz2L5EPz -dYmNUeTDN9KKiE15HrcS3UN4SoqS5tdI1Q+kOilENbgH9mgdVc04UfCMJDGFr4PJ -fel3r+0ae50X+bOdOFAPplp5kYCvN66m0zH7tSYJnTxa71HFK9+WXesyHgLacEns -bgzImjeN9/E2YEsmLIKe0HjzDQ9jpFEw4fkrJxIH2Oq9GGKYsFk3fb7u8yBRQlqD -75O6aRXxYp2fmTmCobd0LovUxQt7L/DICto9eQqakxylKHJzkUOap9FNhYS5qXSP -FEDH3N6sQWRstBmbAmNtJGSPRLIl6s5ddAxjMlyNh+UCAwEAAaOBiTCBhjAPBgNV -HRMBAf8EBTADAQH/MAsGA1UdDwQEAwIBBjAdBgNVHQ4EFgQUppFC/RNhSiOeCKQp -5dgTBCPuQSUwRwYDVR0eBEAwPqA8MAWCAy5ncjAFggMuZXUwBoIELmVkdTAGggQu -b3JnMAWBAy5ncjAFgQMuZXUwBoEELmVkdTAGgQQub3JnMA0GCSqGSIb3DQEBBQUA -A4IBAQAf73lB4XtuP7KMhjdCSk4cNx6NZrokgclPEg8hwAOXhiVtXdMiKahsog2p -6z0GW5k6x8zDmjR/qw7IThzh+uTczQ2+vyT+bOdrwg3IBp5OjWEopmr95fZi6hg8 -TqBTnbI6nOulnJEWtk2C4AwFSKls9cz4y51JtPACpf1wA+2KIaWuE4ZJwzNzvoc7 -dIsXRSZMFpGD/md9zU1jZ/rzAxKWeAaNsWftjj++n08C9bMJL/NMh98qy5V8Acys -Nnq/onN694/BtZqhFLKPM58N7yLcZnuEvUUXBj08yrl3NI/K6s8/MT7jiOOASSXI -l7WdmplNsDz4SgCbZN2fOUvRJ9e4 ------END CERTIFICATE----- - # Issuer: CN=Actalis Authentication Root CA O=Actalis S.p.A./03358520967 # Subject: CN=Actalis Authentication Root CA O=Actalis S.p.A./03358520967 # Label: "Actalis Authentication Root CA" @@ -2253,35 +1234,6 @@ LysRJyU3eExRarDzzFhdFPFqSBX/wge2sY0PjlxQRrM9vwGYT7JZVEc+NHt4bVaT LnPqZih4zR0Uv6CPLy64Lo7yFIrM6bV8+2ydDKXhlg== -----END CERTIFICATE----- -# Issuer: O=Trustis Limited OU=Trustis FPS Root CA -# Subject: O=Trustis Limited OU=Trustis FPS Root CA -# Label: "Trustis FPS Root CA" -# Serial: 36053640375399034304724988975563710553 -# MD5 Fingerprint: 30:c9:e7:1e:6b:e6:14:eb:65:b2:16:69:20:31:67:4d -# SHA1 Fingerprint: 3b:c0:38:0b:33:c3:f6:a6:0c:86:15:22:93:d9:df:f5:4b:81:c0:04 -# SHA256 Fingerprint: c1:b4:82:99:ab:a5:20:8f:e9:63:0a:ce:55:ca:68:a0:3e:da:5a:51:9c:88:02:a0:d3:a6:73:be:8f:8e:55:7d ------BEGIN CERTIFICATE----- -MIIDZzCCAk+gAwIBAgIQGx+ttiD5JNM2a/fH8YygWTANBgkqhkiG9w0BAQUFADBF -MQswCQYDVQQGEwJHQjEYMBYGA1UEChMPVHJ1c3RpcyBMaW1pdGVkMRwwGgYDVQQL -ExNUcnVzdGlzIEZQUyBSb290IENBMB4XDTAzMTIyMzEyMTQwNloXDTI0MDEyMTEx -MzY1NFowRTELMAkGA1UEBhMCR0IxGDAWBgNVBAoTD1RydXN0aXMgTGltaXRlZDEc -MBoGA1UECxMTVHJ1c3RpcyBGUFMgUm9vdCBDQTCCASIwDQYJKoZIhvcNAQEBBQAD -ggEPADCCAQoCggEBAMVQe547NdDfxIzNjpvto8A2mfRC6qc+gIMPpqdZh8mQRUN+ -AOqGeSoDvT03mYlmt+WKVoaTnGhLaASMk5MCPjDSNzoiYYkchU59j9WvezX2fihH -iTHcDnlkH5nSW7r+f2C/revnPDgpai/lkQtV/+xvWNUtyd5MZnGPDNcE2gfmHhjj -vSkCqPoc4Vu5g6hBSLwacY3nYuUtsuvffM/bq1rKMfFMIvMFE/eC+XN5DL7XSxzA -0RU8k0Fk0ea+IxciAIleH2ulrG6nS4zto3Lmr2NNL4XSFDWaLk6M6jKYKIahkQlB -OrTh4/L68MkKokHdqeMDx4gVOxzUGpTXn2RZEm0CAwEAAaNTMFEwDwYDVR0TAQH/ -BAUwAwEB/zAfBgNVHSMEGDAWgBS6+nEleYtXQSUhhgtx67JkDoshZzAdBgNVHQ4E -FgQUuvpxJXmLV0ElIYYLceuyZA6LIWcwDQYJKoZIhvcNAQEFBQADggEBAH5Y//01 -GX2cGE+esCu8jowU/yyg2kdbw++BLa8F6nRIW/M+TgfHbcWzk88iNVy2P3UnXwmW -zaD+vkAMXBJV+JOCyinpXj9WV4s4NvdFGkwozZ5BuO1WTISkQMi4sKUraXAEasP4 -1BIy+Q7DsdwyhEQsb8tGD+pmQQ9P8Vilpg0ND2HepZ5dfWWhPBfnqFVO76DH7cZE -f1T1o+CP8HxVIo8ptoGj4W1OLBuAZ+ytIJ8MYmHVl/9D7S3B2l0pKoU/rGXuhg8F -jZBf3+6f9L/uHfuY5H+QK4R4EA5sSVPvFVtlRkpdr7r7OnIdzfYliB6XzCGcKQEN -ZetX2fNXlrtIzYE= ------END CERTIFICATE----- - # Issuer: CN=Buypass Class 2 Root CA O=Buypass AS-983163327 # Subject: CN=Buypass Class 2 Root CA O=Buypass AS-983163327 # Label: "Buypass Class 2 Root CA" @@ -2391,38 +1343,6 @@ e9eiPZaGzPImNC1qkp2aGtAw4l1OBLBfiyB+d8E9lYLRRpo7PHi4b6HQDWSieB4p TpPDpFQUWw== -----END CERTIFICATE----- -# Issuer: CN=EE Certification Centre Root CA O=AS Sertifitseerimiskeskus -# Subject: CN=EE Certification Centre Root CA O=AS Sertifitseerimiskeskus -# Label: "EE Certification Centre Root CA" -# Serial: 112324828676200291871926431888494945866 -# MD5 Fingerprint: 43:5e:88:d4:7d:1a:4a:7e:fd:84:2e:52:eb:01:d4:6f -# SHA1 Fingerprint: c9:a8:b9:e7:55:80:5e:58:e3:53:77:a7:25:eb:af:c3:7b:27:cc:d7 -# SHA256 Fingerprint: 3e:84:ba:43:42:90:85:16:e7:75:73:c0:99:2f:09:79:ca:08:4e:46:85:68:1f:f1:95:cc:ba:8a:22:9b:8a:76 ------BEGIN CERTIFICATE----- -MIIEAzCCAuugAwIBAgIQVID5oHPtPwBMyonY43HmSjANBgkqhkiG9w0BAQUFADB1 -MQswCQYDVQQGEwJFRTEiMCAGA1UECgwZQVMgU2VydGlmaXRzZWVyaW1pc2tlc2t1 -czEoMCYGA1UEAwwfRUUgQ2VydGlmaWNhdGlvbiBDZW50cmUgUm9vdCBDQTEYMBYG -CSqGSIb3DQEJARYJcGtpQHNrLmVlMCIYDzIwMTAxMDMwMTAxMDMwWhgPMjAzMDEy -MTcyMzU5NTlaMHUxCzAJBgNVBAYTAkVFMSIwIAYDVQQKDBlBUyBTZXJ0aWZpdHNl -ZXJpbWlza2Vza3VzMSgwJgYDVQQDDB9FRSBDZXJ0aWZpY2F0aW9uIENlbnRyZSBS -b290IENBMRgwFgYJKoZIhvcNAQkBFglwa2lAc2suZWUwggEiMA0GCSqGSIb3DQEB -AQUAA4IBDwAwggEKAoIBAQDIIMDs4MVLqwd4lfNE7vsLDP90jmG7sWLqI9iroWUy -euuOF0+W2Ap7kaJjbMeMTC55v6kF/GlclY1i+blw7cNRfdCT5mzrMEvhvH2/UpvO -bntl8jixwKIy72KyaOBhU8E2lf/slLo2rpwcpzIP5Xy0xm90/XsY6KxX7QYgSzIw -WFv9zajmofxwvI6Sc9uXp3whrj3B9UiHbCe9nyV0gVWw93X2PaRka9ZP585ArQ/d -MtO8ihJTmMmJ+xAdTX7Nfh9WDSFwhfYggx/2uh8Ej+p3iDXE/+pOoYtNP2MbRMNE -1CV2yreN1x5KZmTNXMWcg+HCCIia7E6j8T4cLNlsHaFLAgMBAAGjgYowgYcwDwYD -VR0TAQH/BAUwAwEB/zAOBgNVHQ8BAf8EBAMCAQYwHQYDVR0OBBYEFBLyWj7qVhy/ -zQas8fElyalL1BSZMEUGA1UdJQQ+MDwGCCsGAQUFBwMCBggrBgEFBQcDAQYIKwYB -BQUHAwMGCCsGAQUFBwMEBggrBgEFBQcDCAYIKwYBBQUHAwkwDQYJKoZIhvcNAQEF -BQADggEBAHv25MANqhlHt01Xo/6tu7Fq1Q+e2+RjxY6hUFaTlrg4wCQiZrxTFGGV -v9DHKpY5P30osxBAIWrEr7BSdxjhlthWXePdNl4dp1BUoMUq5KqMlIpPnTX/dqQG -E5Gion0ARD9V04I8GtVbvFZMIi5GQ4okQC3zErg7cBqklrkar4dBGmoYDQZPxz5u -uSlNDUmJEYcyW+ZLBMjkXOZ0c5RdFpgTlf7727FE5TpwrDdr5rMzcijJs1eg9gIW -iAYLtqZLICjU3j2LrTcFU3T+bsy8QxdxXvnFzBqpYe73dgzzcvRyrc9yAjYHR8/v -GVCJYMzpJJUPwssd8m92kMfMdcGWxZ0= ------END CERTIFICATE----- - # Issuer: CN=D-TRUST Root Class 3 CA 2 2009 O=D-Trust GmbH # Subject: CN=D-TRUST Root Class 3 CA 2 2009 O=D-Trust GmbH # Label: "D-TRUST Root Class 3 CA 2 2009" @@ -2657,50 +1577,6 @@ HL/EVlP6Y2XQ8xwOFvVrhlhNGNTkDY6lnVuR3HYkUD/GKvvZt5y11ubQ2egZixVx SK236thZiNSQvxaz2emsWWFUyBy6ysHK4bkgTI86k4mloMy/0/Z1pHWWbVY= -----END CERTIFICATE----- -# Issuer: CN=E-Tugra Certification Authority O=E-Tu\u011fra EBG Bili\u015fim Teknolojileri ve Hizmetleri A.\u015e. OU=E-Tugra Sertifikasyon Merkezi -# Subject: CN=E-Tugra Certification Authority O=E-Tu\u011fra EBG Bili\u015fim Teknolojileri ve Hizmetleri A.\u015e. OU=E-Tugra Sertifikasyon Merkezi -# Label: "E-Tugra Certification Authority" -# Serial: 7667447206703254355 -# MD5 Fingerprint: b8:a1:03:63:b0:bd:21:71:70:8a:6f:13:3a:bb:79:49 -# SHA1 Fingerprint: 51:c6:e7:08:49:06:6e:f3:92:d4:5c:a0:0d:6d:a3:62:8f:c3:52:39 -# SHA256 Fingerprint: b0:bf:d5:2b:b0:d7:d9:bd:92:bf:5d:4d:c1:3d:a2:55:c0:2c:54:2f:37:83:65:ea:89:39:11:f5:5e:55:f2:3c ------BEGIN CERTIFICATE----- -MIIGSzCCBDOgAwIBAgIIamg+nFGby1MwDQYJKoZIhvcNAQELBQAwgbIxCzAJBgNV -BAYTAlRSMQ8wDQYDVQQHDAZBbmthcmExQDA+BgNVBAoMN0UtVHXEn3JhIEVCRyBC -aWxpxZ9pbSBUZWtub2xvamlsZXJpIHZlIEhpem1ldGxlcmkgQS7Fni4xJjAkBgNV -BAsMHUUtVHVncmEgU2VydGlmaWthc3lvbiBNZXJrZXppMSgwJgYDVQQDDB9FLVR1 -Z3JhIENlcnRpZmljYXRpb24gQXV0aG9yaXR5MB4XDTEzMDMwNTEyMDk0OFoXDTIz -MDMwMzEyMDk0OFowgbIxCzAJBgNVBAYTAlRSMQ8wDQYDVQQHDAZBbmthcmExQDA+ -BgNVBAoMN0UtVHXEn3JhIEVCRyBCaWxpxZ9pbSBUZWtub2xvamlsZXJpIHZlIEhp -em1ldGxlcmkgQS7Fni4xJjAkBgNVBAsMHUUtVHVncmEgU2VydGlmaWthc3lvbiBN -ZXJrZXppMSgwJgYDVQQDDB9FLVR1Z3JhIENlcnRpZmljYXRpb24gQXV0aG9yaXR5 -MIICIjANBgkqhkiG9w0BAQEFAAOCAg8AMIICCgKCAgEA4vU/kwVRHoViVF56C/UY -B4Oufq9899SKa6VjQzm5S/fDxmSJPZQuVIBSOTkHS0vdhQd2h8y/L5VMzH2nPbxH -D5hw+IyFHnSOkm0bQNGZDbt1bsipa5rAhDGvykPL6ys06I+XawGb1Q5KCKpbknSF -Q9OArqGIW66z6l7LFpp3RMih9lRozt6Plyu6W0ACDGQXwLWTzeHxE2bODHnv0ZEo -q1+gElIwcxmOj+GMB6LDu0rw6h8VqO4lzKRG+Bsi77MOQ7osJLjFLFzUHPhdZL3D -k14opz8n8Y4e0ypQBaNV2cvnOVPAmJ6MVGKLJrD3fY185MaeZkJVgkfnsliNZvcH -fC425lAcP9tDJMW/hkd5s3kc91r0E+xs+D/iWR+V7kI+ua2oMoVJl0b+SzGPWsut -dEcf6ZG33ygEIqDUD13ieU/qbIWGvaimzuT6w+Gzrt48Ue7LE3wBf4QOXVGUnhMM -ti6lTPk5cDZvlsouDERVxcr6XQKj39ZkjFqzAQqptQpHF//vkUAqjqFGOjGY5RH8 -zLtJVor8udBhmm9lbObDyz51Sf6Pp+KJxWfXnUYTTjF2OySznhFlhqt/7x3U+Lzn -rFpct1pHXFXOVbQicVtbC/DP3KBhZOqp12gKY6fgDT+gr9Oq0n7vUaDmUStVkhUX -U8u3Zg5mTPj5dUyQ5xJwx0UCAwEAAaNjMGEwHQYDVR0OBBYEFC7j27JJ0JxUeVz6 -Jyr+zE7S6E5UMA8GA1UdEwEB/wQFMAMBAf8wHwYDVR0jBBgwFoAULuPbsknQnFR5 -XPonKv7MTtLoTlQwDgYDVR0PAQH/BAQDAgEGMA0GCSqGSIb3DQEBCwUAA4ICAQAF -Nzr0TbdF4kV1JI+2d1LoHNgQk2Xz8lkGpD4eKexd0dCrfOAKkEh47U6YA5n+KGCR -HTAduGN8qOY1tfrTYXbm1gdLymmasoR6d5NFFxWfJNCYExL/u6Au/U5Mh/jOXKqY -GwXgAEZKgoClM4so3O0409/lPun++1ndYYRP0lSWE2ETPo+Aab6TR7U1Q9Jauz1c -77NCR807VRMGsAnb/WP2OogKmW9+4c4bU2pEZiNRCHu8W1Ki/QY3OEBhj0qWuJA3 -+GbHeJAAFS6LrVE1Uweoa2iu+U48BybNCAVwzDk/dr2l02cmAYamU9JgO3xDf1WK -vJUawSg5TB9D0pH0clmKuVb8P7Sd2nCcdlqMQ1DujjByTd//SffGqWfZbawCEeI6 -FiWnWAjLb1NBnEg4R2gz0dfHj9R0IdTDBZB6/86WiLEVKV0jq9BgoRJP3vQXzTLl -yb/IQ639Lo7xr+L0mPoSHyDYwKcMhcWQ9DstliaxLL5Mq+ux0orJ23gTDx4JnW2P -AJ8C2sH6H3p6CcRK5ogql5+Ji/03X186zjhZhkuvcQu02PJwT58yE+Owp1fl2tpD -y4Q08ijE6m30Ku/Ba3ba+367hTzSU8JNvnHhRdH9I2cNE3X7z2VnIp2usAnRCf8d -NL/+I5c30jn6PQ0GC7TbO6Orb1wdtn7os4I07QZcJA== ------END CERTIFICATE----- - # Issuer: CN=T-TeleSec GlobalRoot Class 2 O=T-Systems Enterprise Services GmbH OU=T-Systems Trust Center # Subject: CN=T-TeleSec GlobalRoot Class 2 O=T-Systems Enterprise Services GmbH OU=T-Systems Trust Center # Label: "T-TeleSec GlobalRoot Class 2" @@ -3132,27 +2008,6 @@ zzuqQhFkoJ2UOQIReVx7Hfpkue4WQrO/isIJxOzksU0CMQDpKmFHjFJKS04YcPbW RNZu9YO6bVi9JNlWSOrvxKJGgYhqOkbRqZtNyWHa0V1Xahg= -----END CERTIFICATE----- -# Issuer: CN=GlobalSign O=GlobalSign OU=GlobalSign ECC Root CA - R4 -# Subject: CN=GlobalSign O=GlobalSign OU=GlobalSign ECC Root CA - R4 -# Label: "GlobalSign ECC Root CA - R4" -# Serial: 14367148294922964480859022125800977897474 -# MD5 Fingerprint: 20:f0:27:68:d1:7e:a0:9d:0e:e6:2a:ca:df:5c:89:8e -# SHA1 Fingerprint: 69:69:56:2e:40:80:f4:24:a1:e7:19:9f:14:ba:f3:ee:58:ab:6a:bb -# SHA256 Fingerprint: be:c9:49:11:c2:95:56:76:db:6c:0a:55:09:86:d7:6e:3b:a0:05:66:7c:44:2c:97:62:b4:fb:b7:73:de:22:8c ------BEGIN CERTIFICATE----- -MIIB4TCCAYegAwIBAgIRKjikHJYKBN5CsiilC+g0mAIwCgYIKoZIzj0EAwIwUDEk -MCIGA1UECxMbR2xvYmFsU2lnbiBFQ0MgUm9vdCBDQSAtIFI0MRMwEQYDVQQKEwpH -bG9iYWxTaWduMRMwEQYDVQQDEwpHbG9iYWxTaWduMB4XDTEyMTExMzAwMDAwMFoX -DTM4MDExOTAzMTQwN1owUDEkMCIGA1UECxMbR2xvYmFsU2lnbiBFQ0MgUm9vdCBD -QSAtIFI0MRMwEQYDVQQKEwpHbG9iYWxTaWduMRMwEQYDVQQDEwpHbG9iYWxTaWdu -MFkwEwYHKoZIzj0CAQYIKoZIzj0DAQcDQgAEuMZ5049sJQ6fLjkZHAOkrprlOQcJ -FspjsbmG+IpXwVfOQvpzofdlQv8ewQCybnMO/8ch5RikqtlxP6jUuc6MHaNCMEAw -DgYDVR0PAQH/BAQDAgEGMA8GA1UdEwEB/wQFMAMBAf8wHQYDVR0OBBYEFFSwe61F -uOJAf/sKbvu+M8k8o4TVMAoGCCqGSM49BAMCA0gAMEUCIQDckqGgE6bPA7DmxCGX -kPoUVy0D7O48027KqGx2vKLeuwIgJ6iFJzWbVsaj8kfSt24bAgAXqmemFZHe+pTs -ewv4n4Q= ------END CERTIFICATE----- - # Issuer: CN=GlobalSign O=GlobalSign OU=GlobalSign ECC Root CA - R5 # Subject: CN=GlobalSign O=GlobalSign OU=GlobalSign ECC Root CA - R5 # Label: "GlobalSign ECC Root CA - R5" @@ -3175,86 +2030,6 @@ KoZIzj0EAwMDaAAwZQIxAOVpEslu28YxuglB4Zf4+/2a4n0Sye18ZNPLBSWLVtmg xwy8p2Fp8fc74SrL+SvzZpA3 -----END CERTIFICATE----- -# Issuer: CN=Staat der Nederlanden Root CA - G3 O=Staat der Nederlanden -# Subject: CN=Staat der Nederlanden Root CA - G3 O=Staat der Nederlanden -# Label: "Staat der Nederlanden Root CA - G3" -# Serial: 10003001 -# MD5 Fingerprint: 0b:46:67:07:db:10:2f:19:8c:35:50:60:d1:0b:f4:37 -# SHA1 Fingerprint: d8:eb:6b:41:51:92:59:e0:f3:e7:85:00:c0:3d:b6:88:97:c9:ee:fc -# SHA256 Fingerprint: 3c:4f:b0:b9:5a:b8:b3:00:32:f4:32:b8:6f:53:5f:e1:72:c1:85:d0:fd:39:86:58:37:cf:36:18:7f:a6:f4:28 ------BEGIN CERTIFICATE----- -MIIFdDCCA1ygAwIBAgIEAJiiOTANBgkqhkiG9w0BAQsFADBaMQswCQYDVQQGEwJO -TDEeMBwGA1UECgwVU3RhYXQgZGVyIE5lZGVybGFuZGVuMSswKQYDVQQDDCJTdGFh -dCBkZXIgTmVkZXJsYW5kZW4gUm9vdCBDQSAtIEczMB4XDTEzMTExNDExMjg0MloX -DTI4MTExMzIzMDAwMFowWjELMAkGA1UEBhMCTkwxHjAcBgNVBAoMFVN0YWF0IGRl -ciBOZWRlcmxhbmRlbjErMCkGA1UEAwwiU3RhYXQgZGVyIE5lZGVybGFuZGVuIFJv -b3QgQ0EgLSBHMzCCAiIwDQYJKoZIhvcNAQEBBQADggIPADCCAgoCggIBAL4yolQP -cPssXFnrbMSkUeiFKrPMSjTysF/zDsccPVMeiAho2G89rcKezIJnByeHaHE6n3WW -IkYFsO2tx1ueKt6c/DrGlaf1F2cY5y9JCAxcz+bMNO14+1Cx3Gsy8KL+tjzk7FqX -xz8ecAgwoNzFs21v0IJyEavSgWhZghe3eJJg+szeP4TrjTgzkApyI/o1zCZxMdFy -KJLZWyNtZrVtB0LrpjPOktvA9mxjeM3KTj215VKb8b475lRgsGYeCasH/lSJEULR -9yS6YHgamPfJEf0WwTUaVHXvQ9Plrk7O53vDxk5hUUurmkVLoR9BvUhTFXFkC4az -5S6+zqQbwSmEorXLCCN2QyIkHxcE1G6cxvx/K2Ya7Irl1s9N9WMJtxU51nus6+N8 -6U78dULI7ViVDAZCopz35HCz33JvWjdAidiFpNfxC95DGdRKWCyMijmev4SH8RY7 -Ngzp07TKbBlBUgmhHbBqv4LvcFEhMtwFdozL92TkA1CvjJFnq8Xy7ljY3r735zHP -bMk7ccHViLVlvMDoFxcHErVc0qsgk7TmgoNwNsXNo42ti+yjwUOH5kPiNL6VizXt -BznaqB16nzaeErAMZRKQFWDZJkBE41ZgpRDUajz9QdwOWke275dhdU/Z/seyHdTt -XUmzqWrLZoQT1Vyg3N9udwbRcXXIV2+vD3dbAgMBAAGjQjBAMA8GA1UdEwEB/wQF -MAMBAf8wDgYDVR0PAQH/BAQDAgEGMB0GA1UdDgQWBBRUrfrHkleuyjWcLhL75Lpd -INyUVzANBgkqhkiG9w0BAQsFAAOCAgEAMJmdBTLIXg47mAE6iqTnB/d6+Oea31BD -U5cqPco8R5gu4RV78ZLzYdqQJRZlwJ9UXQ4DO1t3ApyEtg2YXzTdO2PCwyiBwpwp -LiniyMMB8jPqKqrMCQj3ZWfGzd/TtiunvczRDnBfuCPRy5FOCvTIeuXZYzbB1N/8 -Ipf3YF3qKS9Ysr1YvY2WTxB1v0h7PVGHoTx0IsL8B3+A3MSs/mrBcDCw6Y5p4ixp -gZQJut3+TcCDjJRYwEYgr5wfAvg1VUkvRtTA8KCWAg8zxXHzniN9lLf9OtMJgwYh -/WA9rjLA0u6NpvDntIJ8CsxwyXmA+P5M9zWEGYox+wrZ13+b8KKaa8MFSu1BYBQw -0aoRQm7TIwIEC8Zl3d1Sd9qBa7Ko+gE4uZbqKmxnl4mUnrzhVNXkanjvSr0rmj1A -fsbAddJu+2gw7OyLnflJNZoaLNmzlTnVHpL3prllL+U9bTpITAjc5CgSKL59NVzq -4BZ+Extq1z7XnvwtdbLBFNUjA9tbbws+eC8N3jONFrdI54OagQ97wUNNVQQXOEpR -1VmiiXTTn74eS9fGbbeIJG9gkaSChVtWQbzQRKtqE77RLFi3EjNYsjdj3BP1lB0/ -QFH1T/U67cjF68IeHRaVesd+QnGTbksVtzDfqu1XhUisHWrdOWnk4Xl4vs4Fv6EM -94B7IWcnMFk= ------END CERTIFICATE----- - -# Issuer: CN=Staat der Nederlanden EV Root CA O=Staat der Nederlanden -# Subject: CN=Staat der Nederlanden EV Root CA O=Staat der Nederlanden -# Label: "Staat der Nederlanden EV Root CA" -# Serial: 10000013 -# MD5 Fingerprint: fc:06:af:7b:e8:1a:f1:9a:b4:e8:d2:70:1f:c0:f5:ba -# SHA1 Fingerprint: 76:e2:7e:c1:4f:db:82:c1:c0:a6:75:b5:05:be:3d:29:b4:ed:db:bb -# SHA256 Fingerprint: 4d:24:91:41:4c:fe:95:67:46:ec:4c:ef:a6:cf:6f:72:e2:8a:13:29:43:2f:9d:8a:90:7a:c4:cb:5d:ad:c1:5a ------BEGIN CERTIFICATE----- -MIIFcDCCA1igAwIBAgIEAJiWjTANBgkqhkiG9w0BAQsFADBYMQswCQYDVQQGEwJO -TDEeMBwGA1UECgwVU3RhYXQgZGVyIE5lZGVybGFuZGVuMSkwJwYDVQQDDCBTdGFh -dCBkZXIgTmVkZXJsYW5kZW4gRVYgUm9vdCBDQTAeFw0xMDEyMDgxMTE5MjlaFw0y -MjEyMDgxMTEwMjhaMFgxCzAJBgNVBAYTAk5MMR4wHAYDVQQKDBVTdGFhdCBkZXIg -TmVkZXJsYW5kZW4xKTAnBgNVBAMMIFN0YWF0IGRlciBOZWRlcmxhbmRlbiBFViBS -b290IENBMIICIjANBgkqhkiG9w0BAQEFAAOCAg8AMIICCgKCAgEA48d+ifkkSzrS -M4M1LGns3Amk41GoJSt5uAg94JG6hIXGhaTK5skuU6TJJB79VWZxXSzFYGgEt9nC -UiY4iKTWO0Cmws0/zZiTs1QUWJZV1VD+hq2kY39ch/aO5ieSZxeSAgMs3NZmdO3d -Z//BYY1jTw+bbRcwJu+r0h8QoPnFfxZpgQNH7R5ojXKhTbImxrpsX23Wr9GxE46p -rfNeaXUmGD5BKyF/7otdBwadQ8QpCiv8Kj6GyzyDOvnJDdrFmeK8eEEzduG/L13l -pJhQDBXd4Pqcfzho0LKmeqfRMb1+ilgnQ7O6M5HTp5gVXJrm0w912fxBmJc+qiXb -j5IusHsMX/FjqTf5m3VpTCgmJdrV8hJwRVXj33NeN/UhbJCONVrJ0yPr08C+eKxC -KFhmpUZtcALXEPlLVPxdhkqHz3/KRawRWrUgUY0viEeXOcDPusBCAUCZSCELa6fS -/ZbV0b5GnUngC6agIk440ME8MLxwjyx1zNDFjFE7PZQIZCZhfbnDZY8UnCHQqv0X -cgOPvZuM5l5Tnrmd74K74bzickFbIZTTRTeU0d8JOV3nI6qaHcptqAqGhYqCvkIH -1vI4gnPah1vlPNOePqc7nvQDs/nxfRN0Av+7oeX6AHkcpmZBiFxgV6YuCcS6/ZrP -px9Aw7vMWgpVSzs4dlG4Y4uElBbmVvMCAwEAAaNCMEAwDwYDVR0TAQH/BAUwAwEB -/zAOBgNVHQ8BAf8EBAMCAQYwHQYDVR0OBBYEFP6rAJCYniT8qcwaivsnuL8wbqg7 -MA0GCSqGSIb3DQEBCwUAA4ICAQDPdyxuVr5Os7aEAJSrR8kN0nbHhp8dB9O2tLsI -eK9p0gtJ3jPFrK3CiAJ9Brc1AsFgyb/E6JTe1NOpEyVa/m6irn0F3H3zbPB+po3u -2dfOWBfoqSmuc0iH55vKbimhZF8ZE/euBhD/UcabTVUlT5OZEAFTdfETzsemQUHS -v4ilf0X8rLiltTMMgsT7B/Zq5SWEXwbKwYY5EdtYzXc7LMJMD16a4/CrPmEbUCTC -wPTxGfARKbalGAKb12NMcIxHowNDXLldRqANb/9Zjr7dn3LDWyvfjFvO5QxGbJKy -CqNMVEIYFRIYvdr8unRu/8G2oGTYqV9Vrp9canaW2HNnh/tNf1zuacpzEPuKqf2e -vTY4SUmH9A4U8OmHuD+nT3pajnnUk+S7aFKErGzp85hwVXIy+TSrK0m1zSBi5Dp6 -Z2Orltxtrpfs/J92VoguZs9btsmksNcFuuEnL5O7Jiqik7Ab846+HUCjuTaPPoIa -Gl6I6lD4WeKDRikL40Rc4ZW2aZCaFG+XroHPaO+Zmr615+F/+PoTRxZMzG0IQOeL -eG9QgkRQP2YGiqtDhFZKDyAthg710tvSeopLzaXoTvFeJiUBWSOgftL2fiFX1ye8 -FVdMpEbB4IMeDExNH08GGeL5qPQ6gqGyeUN51q1veieQA6TqJIc/2b3Z6fJfUEkc -7uzXLg== ------END CERTIFICATE----- - # Issuer: CN=IdenTrust Commercial Root CA 1 O=IdenTrust # Subject: CN=IdenTrust Commercial Root CA 1 O=IdenTrust # Label: "IdenTrust Commercial Root CA 1" @@ -3788,47 +2563,6 @@ CkcO8DdZEv8tmZQoTipPNU0zWgIxAOp1AE47xDqUEpHJWEadIRNyp4iciuRMStuW 1KyLa2tJElMzrdfkviT8tQp21KW8EA== -----END CERTIFICATE----- -# Issuer: CN=LuxTrust Global Root 2 O=LuxTrust S.A. -# Subject: CN=LuxTrust Global Root 2 O=LuxTrust S.A. -# Label: "LuxTrust Global Root 2" -# Serial: 59914338225734147123941058376788110305822489521 -# MD5 Fingerprint: b2:e1:09:00:61:af:f7:f1:91:6f:c4:ad:8d:5e:3b:7c -# SHA1 Fingerprint: 1e:0e:56:19:0a:d1:8b:25:98:b2:04:44:ff:66:8a:04:17:99:5f:3f -# SHA256 Fingerprint: 54:45:5f:71:29:c2:0b:14:47:c4:18:f9:97:16:8f:24:c5:8f:c5:02:3b:f5:da:5b:e2:eb:6e:1d:d8:90:2e:d5 ------BEGIN CERTIFICATE----- -MIIFwzCCA6ugAwIBAgIUCn6m30tEntpqJIWe5rgV0xZ/u7EwDQYJKoZIhvcNAQEL -BQAwRjELMAkGA1UEBhMCTFUxFjAUBgNVBAoMDUx1eFRydXN0IFMuQS4xHzAdBgNV -BAMMFkx1eFRydXN0IEdsb2JhbCBSb290IDIwHhcNMTUwMzA1MTMyMTU3WhcNMzUw -MzA1MTMyMTU3WjBGMQswCQYDVQQGEwJMVTEWMBQGA1UECgwNTHV4VHJ1c3QgUy5B -LjEfMB0GA1UEAwwWTHV4VHJ1c3QgR2xvYmFsIFJvb3QgMjCCAiIwDQYJKoZIhvcN -AQEBBQADggIPADCCAgoCggIBANeFl78RmOnwYoNMPIf5U2o3C/IPPIfOb9wmKb3F -ibrJgz337spbxm1Jc7TJRqMbNBM/wYlFV/TZsfs2ZUv7COJIcRHIbjuend+JZTem -hfY7RBi2xjcwYkSSl2l9QjAk5A0MiWtj3sXh306pFGxT4GHO9hcvHTy95iJMHZP1 -EMShduxq3sVs35a0VkBCwGKSMKEtFZSg0iAGCW5qbeXrt77U8PEVfIvmTroTzEsn -Xpk8F12PgX8zPU/TPxvsXD/wPEx1bvKm1Z3aLQdjAsZy6ZS8TEmVT4hSyNvoaYL4 -zDRbIvCGp4m9SAptZoFtyMhk+wHh9OHe2Z7d21vUKpkmFRseTJIpgp7VkoGSQXAZ -96Tlk0u8d2cx3Rz9MXANF5kM+Qw5GSoXtTBxVdUPrljhPS80m8+f9niFwpN6cj5m -j5wWEWCPnolvZ77gR1o7DJpni89Gxq44o/KnvObWhWszJHAiS8sIm7vI+AIpHb4g -DEa/a4ebsypmQjVGbKq6rfmYe+lQVRQxv7HaLe2ArWgk+2mr2HETMOZns4dA/Yl+ -8kPREd8vZS9kzl8UubG/Mb2HeFpZZYiq/FkySIbWTLkpS5XTdvN3JW1CHDiDTf2j -X5t/Lax5Gw5CMZdjpPuKadUiDTSQMC6otOBttpSsvItO13D8xTiOZCXhTTmQzsmH -hFhxAgMBAAGjgagwgaUwDwYDVR0TAQH/BAUwAwEB/zBCBgNVHSAEOzA5MDcGByuB -KwEBAQowLDAqBggrBgEFBQcCARYeaHR0cHM6Ly9yZXBvc2l0b3J5Lmx1eHRydXN0 -Lmx1MA4GA1UdDwEB/wQEAwIBBjAfBgNVHSMEGDAWgBT/GCh2+UgFLKGu8SsbK7JT -+Et8szAdBgNVHQ4EFgQU/xgodvlIBSyhrvErGyuyU/hLfLMwDQYJKoZIhvcNAQEL -BQADggIBAGoZFO1uecEsh9QNcH7X9njJCwROxLHOk3D+sFTAMs2ZMGQXvw/l4jP9 -BzZAcg4atmpZ1gDlaCDdLnINH2pkMSCEfUmmWjfrRcmF9dTHF5kH5ptV5AzoqbTO -jFu1EVzPig4N1qx3gf4ynCSecs5U89BvolbW7MM3LGVYvlcAGvI1+ut7MV3CwRI9 -loGIlonBWVx65n9wNOeD4rHh4bhY79SV5GCc8JaXcozrhAIuZY+kt9J/Z93I055c -qqmkoCUUBpvsT34tC38ddfEz2O3OuHVtPlu5mB0xDVbYQw8wkbIEa91WvpWAVWe+ -2M2D2RjuLg+GLZKecBPs3lHJQ3gCpU3I+V/EkVhGFndadKpAvAefMLmx9xIX3eP/ -JEAdemrRTxgKqpAd60Ae36EeRJIQmvKN4dFLRp7oRUKX6kWZ8+xm1QL68qZKJKre -zrnK+T+Tb/mjuuqlPpmt/f97mfVl7vBZKGfXkJWkE4SphMHozs51k2MavDzq1WQf -LSoSOcbDWjLtR5EWDrw4wVDej8oqkDQc7kGUnF4ZLvhFSZl0kbAEb+MEWrGrKqv+ -x9CWttrhSmQGbmBNvUJO/3jaJMobtNeWOWyu8Q6qp31IiyBMz2TWuJdGsE7RKlY6 -oJO9r4Ak4Ap+58rVyuiFVdw2KuGUaJPHZnJED4AhMmwlxyOAgwrr ------END CERTIFICATE----- - # Issuer: CN=TUBITAK Kamu SM SSL Kok Sertifikasi - Surum 1 O=Turkiye Bilimsel ve Teknolojik Arastirma Kurumu - TUBITAK OU=Kamu Sertifikasyon Merkezi - Kamu SM # Subject: CN=TUBITAK Kamu SM SSL Kok Sertifikasi - Surum 1 O=Turkiye Bilimsel ve Teknolojik Arastirma Kurumu - TUBITAK OU=Kamu Sertifikasyon Merkezi - Kamu SM # Label: "TUBITAK Kamu SM SSL Kok Sertifikasi - Surum 1" @@ -3903,116 +2637,6 @@ T8p+ck0LcIymSLumoRT2+1hEmRSuqguTaaApJUqlyyvdimYHFngVV3Eb7PVHhPOe MTd61X8kreS8/f3MboPoDKi3QWwH3b08hpcv0g== -----END CERTIFICATE----- -# Issuer: CN=TrustCor RootCert CA-1 O=TrustCor Systems S. de R.L. OU=TrustCor Certificate Authority -# Subject: CN=TrustCor RootCert CA-1 O=TrustCor Systems S. de R.L. OU=TrustCor Certificate Authority -# Label: "TrustCor RootCert CA-1" -# Serial: 15752444095811006489 -# MD5 Fingerprint: 6e:85:f1:dc:1a:00:d3:22:d5:b2:b2:ac:6b:37:05:45 -# SHA1 Fingerprint: ff:bd:cd:e7:82:c8:43:5e:3c:6f:26:86:5c:ca:a8:3a:45:5b:c3:0a -# SHA256 Fingerprint: d4:0e:9c:86:cd:8f:e4:68:c1:77:69:59:f4:9e:a7:74:fa:54:86:84:b6:c4:06:f3:90:92:61:f4:dc:e2:57:5c ------BEGIN CERTIFICATE----- -MIIEMDCCAxigAwIBAgIJANqb7HHzA7AZMA0GCSqGSIb3DQEBCwUAMIGkMQswCQYD -VQQGEwJQQTEPMA0GA1UECAwGUGFuYW1hMRQwEgYDVQQHDAtQYW5hbWEgQ2l0eTEk -MCIGA1UECgwbVHJ1c3RDb3IgU3lzdGVtcyBTLiBkZSBSLkwuMScwJQYDVQQLDB5U -cnVzdENvciBDZXJ0aWZpY2F0ZSBBdXRob3JpdHkxHzAdBgNVBAMMFlRydXN0Q29y -IFJvb3RDZXJ0IENBLTEwHhcNMTYwMjA0MTIzMjE2WhcNMjkxMjMxMTcyMzE2WjCB -pDELMAkGA1UEBhMCUEExDzANBgNVBAgMBlBhbmFtYTEUMBIGA1UEBwwLUGFuYW1h -IENpdHkxJDAiBgNVBAoMG1RydXN0Q29yIFN5c3RlbXMgUy4gZGUgUi5MLjEnMCUG -A1UECwweVHJ1c3RDb3IgQ2VydGlmaWNhdGUgQXV0aG9yaXR5MR8wHQYDVQQDDBZU -cnVzdENvciBSb290Q2VydCBDQS0xMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIB -CgKCAQEAv463leLCJhJrMxnHQFgKq1mqjQCj/IDHUHuO1CAmujIS2CNUSSUQIpid -RtLByZ5OGy4sDjjzGiVoHKZaBeYei0i/mJZ0PmnK6bV4pQa81QBeCQryJ3pS/C3V -seq0iWEk8xoT26nPUu0MJLq5nux+AHT6k61sKZKuUbS701e/s/OojZz0JEsq1pme -9J7+wH5COucLlVPat2gOkEz7cD+PSiyU8ybdY2mplNgQTsVHCJCZGxdNuWxu72CV -EY4hgLW9oHPY0LJ3xEXqWib7ZnZ2+AYfYW0PVcWDtxBWcgYHpfOxGgMFZA6dWorW -hnAbJN7+KIor0Gqw/Hqi3LJ5DotlDwIDAQABo2MwYTAdBgNVHQ4EFgQU7mtJPHo/ -DeOxCbeKyKsZn3MzUOcwHwYDVR0jBBgwFoAU7mtJPHo/DeOxCbeKyKsZn3MzUOcw -DwYDVR0TAQH/BAUwAwEB/zAOBgNVHQ8BAf8EBAMCAYYwDQYJKoZIhvcNAQELBQAD -ggEBACUY1JGPE+6PHh0RU9otRCkZoB5rMZ5NDp6tPVxBb5UrJKF5mDo4Nvu7Zp5I -/5CQ7z3UuJu0h3U/IJvOcs+hVcFNZKIZBqEHMwwLKeXx6quj7LUKdJDHfXLy11yf -ke+Ri7fc7Waiz45mO7yfOgLgJ90WmMCV1Aqk5IGadZQ1nJBfiDcGrVmVCrDRZ9MZ -yonnMlo2HD6CqFqTvsbQZJG2z9m2GM/bftJlo6bEjhcxwft+dtvTheNYsnd6djts -L1Ac59v2Z3kf9YKVmgenFK+P3CghZwnS1k1aHBkcjndcw5QkPTJrS37UeJSDvjdN -zl/HHk484IkzlQsPpTLWPFp5LBk= ------END CERTIFICATE----- - -# Issuer: CN=TrustCor RootCert CA-2 O=TrustCor Systems S. de R.L. OU=TrustCor Certificate Authority -# Subject: CN=TrustCor RootCert CA-2 O=TrustCor Systems S. de R.L. OU=TrustCor Certificate Authority -# Label: "TrustCor RootCert CA-2" -# Serial: 2711694510199101698 -# MD5 Fingerprint: a2:e1:f8:18:0b:ba:45:d5:c7:41:2a:bb:37:52:45:64 -# SHA1 Fingerprint: b8:be:6d:cb:56:f1:55:b9:63:d4:12:ca:4e:06:34:c7:94:b2:1c:c0 -# SHA256 Fingerprint: 07:53:e9:40:37:8c:1b:d5:e3:83:6e:39:5d:ae:a5:cb:83:9e:50:46:f1:bd:0e:ae:19:51:cf:10:fe:c7:c9:65 ------BEGIN CERTIFICATE----- -MIIGLzCCBBegAwIBAgIIJaHfyjPLWQIwDQYJKoZIhvcNAQELBQAwgaQxCzAJBgNV -BAYTAlBBMQ8wDQYDVQQIDAZQYW5hbWExFDASBgNVBAcMC1BhbmFtYSBDaXR5MSQw -IgYDVQQKDBtUcnVzdENvciBTeXN0ZW1zIFMuIGRlIFIuTC4xJzAlBgNVBAsMHlRy -dXN0Q29yIENlcnRpZmljYXRlIEF1dGhvcml0eTEfMB0GA1UEAwwWVHJ1c3RDb3Ig -Um9vdENlcnQgQ0EtMjAeFw0xNjAyMDQxMjMyMjNaFw0zNDEyMzExNzI2MzlaMIGk -MQswCQYDVQQGEwJQQTEPMA0GA1UECAwGUGFuYW1hMRQwEgYDVQQHDAtQYW5hbWEg -Q2l0eTEkMCIGA1UECgwbVHJ1c3RDb3IgU3lzdGVtcyBTLiBkZSBSLkwuMScwJQYD -VQQLDB5UcnVzdENvciBDZXJ0aWZpY2F0ZSBBdXRob3JpdHkxHzAdBgNVBAMMFlRy -dXN0Q29yIFJvb3RDZXJ0IENBLTIwggIiMA0GCSqGSIb3DQEBAQUAA4ICDwAwggIK -AoICAQCnIG7CKqJiJJWQdsg4foDSq8GbZQWU9MEKENUCrO2fk8eHyLAnK0IMPQo+ -QVqedd2NyuCb7GgypGmSaIwLgQ5WoD4a3SwlFIIvl9NkRvRUqdw6VC0xK5mC8tkq -1+9xALgxpL56JAfDQiDyitSSBBtlVkxs1Pu2YVpHI7TYabS3OtB0PAx1oYxOdqHp -2yqlO/rOsP9+aij9JxzIsekp8VduZLTQwRVtDr4uDkbIXvRR/u8OYzo7cbrPb1nK -DOObXUm4TOJXsZiKQlecdu/vvdFoqNL0Cbt3Nb4lggjEFixEIFapRBF37120Hape -az6LMvYHL1cEksr1/p3C6eizjkxLAjHZ5DxIgif3GIJ2SDpxsROhOdUuxTTCHWKF -3wP+TfSvPd9cW436cOGlfifHhi5qjxLGhF5DUVCcGZt45vz27Ud+ez1m7xMTiF88 -oWP7+ayHNZ/zgp6kPwqcMWmLmaSISo5uZk3vFsQPeSghYA2FFn3XVDjxklb9tTNM -g9zXEJ9L/cb4Qr26fHMC4P99zVvh1Kxhe1fVSntb1IVYJ12/+CtgrKAmrhQhJ8Z3 -mjOAPF5GP/fDsaOGM8boXg25NSyqRsGFAnWAoOsk+xWq5Gd/bnc/9ASKL3x74xdh -8N0JqSDIvgmk0H5Ew7IwSjiqqewYmgeCK9u4nBit2uBGF6zPXQIDAQABo2MwYTAd -BgNVHQ4EFgQU2f4hQG6UnrybPZx9mCAZ5YwwYrIwHwYDVR0jBBgwFoAU2f4hQG6U -nrybPZx9mCAZ5YwwYrIwDwYDVR0TAQH/BAUwAwEB/zAOBgNVHQ8BAf8EBAMCAYYw -DQYJKoZIhvcNAQELBQADggIBAJ5Fngw7tu/hOsh80QA9z+LqBrWyOrsGS2h60COX -dKcs8AjYeVrXWoSK2BKaG9l9XE1wxaX5q+WjiYndAfrs3fnpkpfbsEZC89NiqpX+ -MWcUaViQCqoL7jcjx1BRtPV+nuN79+TMQjItSQzL/0kMmx40/W5ulop5A7Zv2wnL -/V9lFDfhOPXzYRZY5LVtDQsEGz9QLX+zx3oaFoBg+Iof6Rsqxvm6ARppv9JYx1RX -CI/hOWB3S6xZhBqI8d3LT3jX5+EzLfzuQfogsL7L9ziUwOHQhQ+77Sxzq+3+knYa -ZH9bDTMJBzN7Bj8RpFxwPIXAz+OQqIN3+tvmxYxoZxBnpVIt8MSZj3+/0WvitUfW -2dCFmU2Umw9Lje4AWkcdEQOsQRivh7dvDDqPys/cA8GiCcjl/YBeyGBCARsaU1q7 -N6a3vLqE6R5sGtRk2tRD/pOLS/IseRYQ1JMLiI+h2IYURpFHmygk71dSTlxCnKr3 -Sewn6EAes6aJInKc9Q0ztFijMDvd1GpUk74aTfOTlPf8hAs/hCBcNANExdqtvArB -As8e5ZTZ845b2EzwnexhF7sUMlQMAimTHpKG9n/v55IFDlndmQguLvqcAFLTxWYp -5KeXRKQOKIETNcX2b2TmQcTVL8w0RSXPQQCWPUouwpaYT05KnJe32x+SMsj/D1Fu -1uwJ ------END CERTIFICATE----- - -# Issuer: CN=TrustCor ECA-1 O=TrustCor Systems S. de R.L. OU=TrustCor Certificate Authority -# Subject: CN=TrustCor ECA-1 O=TrustCor Systems S. de R.L. OU=TrustCor Certificate Authority -# Label: "TrustCor ECA-1" -# Serial: 9548242946988625984 -# MD5 Fingerprint: 27:92:23:1d:0a:f5:40:7c:e9:e6:6b:9d:d8:f5:e7:6c -# SHA1 Fingerprint: 58:d1:df:95:95:67:6b:63:c0:f0:5b:1c:17:4d:8b:84:0b:c8:78:bd -# SHA256 Fingerprint: 5a:88:5d:b1:9c:01:d9:12:c5:75:93:88:93:8c:af:bb:df:03:1a:b2:d4:8e:91:ee:15:58:9b:42:97:1d:03:9c ------BEGIN CERTIFICATE----- -MIIEIDCCAwigAwIBAgIJAISCLF8cYtBAMA0GCSqGSIb3DQEBCwUAMIGcMQswCQYD -VQQGEwJQQTEPMA0GA1UECAwGUGFuYW1hMRQwEgYDVQQHDAtQYW5hbWEgQ2l0eTEk -MCIGA1UECgwbVHJ1c3RDb3IgU3lzdGVtcyBTLiBkZSBSLkwuMScwJQYDVQQLDB5U -cnVzdENvciBDZXJ0aWZpY2F0ZSBBdXRob3JpdHkxFzAVBgNVBAMMDlRydXN0Q29y -IEVDQS0xMB4XDTE2MDIwNDEyMzIzM1oXDTI5MTIzMTE3MjgwN1owgZwxCzAJBgNV -BAYTAlBBMQ8wDQYDVQQIDAZQYW5hbWExFDASBgNVBAcMC1BhbmFtYSBDaXR5MSQw -IgYDVQQKDBtUcnVzdENvciBTeXN0ZW1zIFMuIGRlIFIuTC4xJzAlBgNVBAsMHlRy -dXN0Q29yIENlcnRpZmljYXRlIEF1dGhvcml0eTEXMBUGA1UEAwwOVHJ1c3RDb3Ig -RUNBLTEwggEiMA0GCSqGSIb3DQEBAQUAA4IBDwAwggEKAoIBAQDPj+ARtZ+odnbb -3w9U73NjKYKtR8aja+3+XzP4Q1HpGjORMRegdMTUpwHmspI+ap3tDvl0mEDTPwOA -BoJA6LHip1GnHYMma6ve+heRK9jGrB6xnhkB1Zem6g23xFUfJ3zSCNV2HykVh0A5 -3ThFEXXQmqc04L/NyFIduUd+Dbi7xgz2c1cWWn5DkR9VOsZtRASqnKmcp0yJF4Ou -owReUoCLHhIlERnXDH19MURB6tuvsBzvgdAsxZohmz3tQjtQJvLsznFhBmIhVE5/ -wZ0+fyCMgMsq2JdiyIMzkX2woloPV+g7zPIlstR8L+xNxqE6FXrntl019fZISjZF -ZtS6mFjBAgMBAAGjYzBhMB0GA1UdDgQWBBREnkj1zG1I1KBLf/5ZJC+Dl5mahjAf -BgNVHSMEGDAWgBREnkj1zG1I1KBLf/5ZJC+Dl5mahjAPBgNVHRMBAf8EBTADAQH/ -MA4GA1UdDwEB/wQEAwIBhjANBgkqhkiG9w0BAQsFAAOCAQEABT41XBVwm8nHc2Fv -civUwo/yQ10CzsSUuZQRg2dd4mdsdXa/uwyqNsatR5Nj3B5+1t4u/ukZMjgDfxT2 -AHMsWbEhBuH7rBiVDKP/mZb3Kyeb1STMHd3BOuCYRLDE5D53sXOpZCz2HAF8P11F -hcCF5yWPldwX8zyfGm6wyuMdKulMY/okYWLW2n62HGz1Ah3UKt1VkOsqEUc8Ll50 -soIipX1TH0XsJ5F95yIW6MBoNtjG8U+ARDL54dHRHareqKucBK+tIA5kmE2la8BI -WJZpTdwHjFGTot+fDz2LYLSCjaoITmJF4PkL0uDgPFveXHEnJcLmA4GLEFPjx1Wi -tJ/X5g== ------END CERTIFICATE----- - # Issuer: CN=SSL.com Root Certification Authority RSA O=SSL Corporation # Subject: CN=SSL.com Root Certification Authority RSA O=SSL Corporation # Label: "SSL.com Root Certification Authority RSA" @@ -4208,126 +2832,6 @@ rYy0UGYwEAYJKwYBBAGCNxUBBAMCAQAwCgYIKoZIzj0EAwMDaAAwZQIwJsdpW9zV Mgj/mkkCtojeFK9dbJlxjRo/i9fgojaGHAeCOnZT/cKi7e97sIBPWA9LUzm9 -----END CERTIFICATE----- -# Issuer: CN=GTS Root R1 O=Google Trust Services LLC -# Subject: CN=GTS Root R1 O=Google Trust Services LLC -# Label: "GTS Root R1" -# Serial: 146587175971765017618439757810265552097 -# MD5 Fingerprint: 82:1a:ef:d4:d2:4a:f2:9f:e2:3d:97:06:14:70:72:85 -# SHA1 Fingerprint: e1:c9:50:e6:ef:22:f8:4c:56:45:72:8b:92:20:60:d7:d5:a7:a3:e8 -# SHA256 Fingerprint: 2a:57:54:71:e3:13:40:bc:21:58:1c:bd:2c:f1:3e:15:84:63:20:3e:ce:94:bc:f9:d3:cc:19:6b:f0:9a:54:72 ------BEGIN CERTIFICATE----- -MIIFWjCCA0KgAwIBAgIQbkepxUtHDA3sM9CJuRz04TANBgkqhkiG9w0BAQwFADBH -MQswCQYDVQQGEwJVUzEiMCAGA1UEChMZR29vZ2xlIFRydXN0IFNlcnZpY2VzIExM -QzEUMBIGA1UEAxMLR1RTIFJvb3QgUjEwHhcNMTYwNjIyMDAwMDAwWhcNMzYwNjIy -MDAwMDAwWjBHMQswCQYDVQQGEwJVUzEiMCAGA1UEChMZR29vZ2xlIFRydXN0IFNl -cnZpY2VzIExMQzEUMBIGA1UEAxMLR1RTIFJvb3QgUjEwggIiMA0GCSqGSIb3DQEB -AQUAA4ICDwAwggIKAoICAQC2EQKLHuOhd5s73L+UPreVp0A8of2C+X0yBoJx9vaM -f/vo27xqLpeXo4xL+Sv2sfnOhB2x+cWX3u+58qPpvBKJXqeqUqv4IyfLpLGcY9vX -mX7wCl7raKb0xlpHDU0QM+NOsROjyBhsS+z8CZDfnWQpJSMHobTSPS5g4M/SCYe7 -zUjwTcLCeoiKu7rPWRnWr4+wB7CeMfGCwcDfLqZtbBkOtdh+JhpFAz2weaSUKK0P -fyblqAj+lug8aJRT7oM6iCsVlgmy4HqMLnXWnOunVmSPlk9orj2XwoSPwLxAwAtc -vfaHszVsrBhQf4TgTM2S0yDpM7xSma8ytSmzJSq0SPly4cpk9+aCEI3oncKKiPo4 -Zor8Y/kB+Xj9e1x3+naH+uzfsQ55lVe0vSbv1gHR6xYKu44LtcXFilWr06zqkUsp -zBmkMiVOKvFlRNACzqrOSbTqn3yDsEB750Orp2yjj32JgfpMpf/VjsPOS+C12LOO -Rc92wO1AK/1TD7Cn1TsNsYqiA94xrcx36m97PtbfkSIS5r762DL8EGMUUXLeXdYW -k70paDPvOmbsB4om3xPXV2V4J95eSRQAogB/mqghtqmxlbCluQ0WEdrHbEg8QOB+ -DVrNVjzRlwW5y0vtOUucxD/SVRNuJLDWcfr0wbrM7Rv1/oFB2ACYPTrIrnqYNxgF -lQIDAQABo0IwQDAOBgNVHQ8BAf8EBAMCAQYwDwYDVR0TAQH/BAUwAwEB/zAdBgNV -HQ4EFgQU5K8rJnEaK0gnhS9SZizv8IkTcT4wDQYJKoZIhvcNAQEMBQADggIBADiW -Cu49tJYeX++dnAsznyvgyv3SjgofQXSlfKqE1OXyHuY3UjKcC9FhHb8owbZEKTV1 -d5iyfNm9dKyKaOOpMQkpAWBz40d8U6iQSifvS9efk+eCNs6aaAyC58/UEBZvXw6Z -XPYfcX3v73svfuo21pdwCxXu11xWajOl40k4DLh9+42FpLFZXvRq4d2h9mREruZR -gyFmxhE+885H7pwoHyXa/6xmld01D1zvICxi/ZG6qcz8WpyTgYMpl0p8WnK0OdC3 -d8t5/Wk6kjftbjhlRn7pYL15iJdfOBL07q9bgsiG1eGZbYwE8na6SfZu6W0eX6Dv -J4J2QPim01hcDyxC2kLGe4g0x8HYRZvBPsVhHdljUEn2NIVq4BjFbkerQUIpm/Zg -DdIx02OYI5NaAIFItO/Nis3Jz5nu2Z6qNuFoS3FJFDYoOj0dzpqPJeaAcWErtXvM -+SUWgeExX6GjfhaknBZqlxi9dnKlC54dNuYvoS++cJEPqOba+MSSQGwlfnuzCdyy -F62ARPBopY+Udf90WuioAnwMCeKpSwughQtiue+hMZL77/ZRBIls6Kl0obsXs7X9 -SQ98POyDGCBDTtWTurQ0sR8WNh8M5mQ5Fkzc4P4dyKliPUDqysU0ArSuiYgzNdws -E3PYJ/HQcu51OyLemGhmW/HGY0dVHLqlCFF1pkgl ------END CERTIFICATE----- - -# Issuer: CN=GTS Root R2 O=Google Trust Services LLC -# Subject: CN=GTS Root R2 O=Google Trust Services LLC -# Label: "GTS Root R2" -# Serial: 146587176055767053814479386953112547951 -# MD5 Fingerprint: 44:ed:9a:0e:a4:09:3b:00:f2:ae:4c:a3:c6:61:b0:8b -# SHA1 Fingerprint: d2:73:96:2a:2a:5e:39:9f:73:3f:e1:c7:1e:64:3f:03:38:34:fc:4d -# SHA256 Fingerprint: c4:5d:7b:b0:8e:6d:67:e6:2e:42:35:11:0b:56:4e:5f:78:fd:92:ef:05:8c:84:0a:ea:4e:64:55:d7:58:5c:60 ------BEGIN CERTIFICATE----- -MIIFWjCCA0KgAwIBAgIQbkepxlqz5yDFMJo/aFLybzANBgkqhkiG9w0BAQwFADBH -MQswCQYDVQQGEwJVUzEiMCAGA1UEChMZR29vZ2xlIFRydXN0IFNlcnZpY2VzIExM -QzEUMBIGA1UEAxMLR1RTIFJvb3QgUjIwHhcNMTYwNjIyMDAwMDAwWhcNMzYwNjIy -MDAwMDAwWjBHMQswCQYDVQQGEwJVUzEiMCAGA1UEChMZR29vZ2xlIFRydXN0IFNl -cnZpY2VzIExMQzEUMBIGA1UEAxMLR1RTIFJvb3QgUjIwggIiMA0GCSqGSIb3DQEB -AQUAA4ICDwAwggIKAoICAQDO3v2m++zsFDQ8BwZabFn3GTXd98GdVarTzTukk3Lv -CvptnfbwhYBboUhSnznFt+4orO/LdmgUud+tAWyZH8QiHZ/+cnfgLFuv5AS/T3Kg -GjSY6Dlo7JUle3ah5mm5hRm9iYz+re026nO8/4Piy33B0s5Ks40FnotJk9/BW9Bu -XvAuMC6C/Pq8tBcKSOWIm8Wba96wyrQD8Nr0kLhlZPdcTK3ofmZemde4wj7I0BOd -re7kRXuJVfeKH2JShBKzwkCX44ofR5GmdFrS+LFjKBC4swm4VndAoiaYecb+3yXu -PuWgf9RhD1FLPD+M2uFwdNjCaKH5wQzpoeJ/u1U8dgbuak7MkogwTZq9TwtImoS1 -mKPV+3PBV2HdKFZ1E66HjucMUQkQdYhMvI35ezzUIkgfKtzra7tEscszcTJGr61K -8YzodDqs5xoic4DSMPclQsciOzsSrZYuxsN2B6ogtzVJV+mSSeh2FnIxZyuWfoqj -x5RWIr9qS34BIbIjMt/kmkRtWVtd9QCgHJvGeJeNkP+byKq0rxFROV7Z+2et1VsR -nTKaG73VululycslaVNVJ1zgyjbLiGH7HrfQy+4W+9OmTN6SpdTi3/UGVN4unUu0 -kzCqgc7dGtxRcw1PcOnlthYhGXmy5okLdWTK1au8CcEYof/UVKGFPP0UJAOyh9Ok -twIDAQABo0IwQDAOBgNVHQ8BAf8EBAMCAQYwDwYDVR0TAQH/BAUwAwEB/zAdBgNV -HQ4EFgQUu//KjiOfT5nK2+JopqUVJxce2Q4wDQYJKoZIhvcNAQEMBQADggIBALZp -8KZ3/p7uC4Gt4cCpx/k1HUCCq+YEtN/L9x0Pg/B+E02NjO7jMyLDOfxA325BS0JT -vhaI8dI4XsRomRyYUpOM52jtG2pzegVATX9lO9ZY8c6DR2Dj/5epnGB3GFW1fgiT -z9D2PGcDFWEJ+YF59exTpJ/JjwGLc8R3dtyDovUMSRqodt6Sm2T4syzFJ9MHwAiA -pJiS4wGWAqoC7o87xdFtCjMwc3i5T1QWvwsHoaRc5svJXISPD+AVdyx+Jn7axEvb -pxZ3B7DNdehyQtaVhJ2Gg/LkkM0JR9SLA3DaWsYDQvTtN6LwG1BUSw7YhN4ZKJmB -R64JGz9I0cNv4rBgF/XuIwKl2gBbbZCr7qLpGzvpx0QnRY5rn/WkhLx3+WuXrD5R -RaIRpsyF7gpo8j5QOHokYh4XIDdtak23CZvJ/KRY9bb7nE4Yu5UC56GtmwfuNmsk -0jmGwZODUNKBRqhfYlcsu2xkiAhu7xNUX90txGdj08+JN7+dIPT7eoOboB6BAFDC -5AwiWVIQ7UNWhwD4FFKnHYuTjKJNRn8nxnGbJN7k2oaLDX5rIMHAnuFl2GqjpuiF -izoHCBy69Y9Vmhh1fuXsgWbRIXOhNUQLgD1bnF5vKheW0YMjiGZt5obicDIvUiLn -yOd/xCxgXS/Dr55FBcOEArf9LAhST4Ldo/DUhgkC ------END CERTIFICATE----- - -# Issuer: CN=GTS Root R3 O=Google Trust Services LLC -# Subject: CN=GTS Root R3 O=Google Trust Services LLC -# Label: "GTS Root R3" -# Serial: 146587176140553309517047991083707763997 -# MD5 Fingerprint: 1a:79:5b:6b:04:52:9c:5d:c7:74:33:1b:25:9a:f9:25 -# SHA1 Fingerprint: 30:d4:24:6f:07:ff:db:91:89:8a:0b:e9:49:66:11:eb:8c:5e:46:e5 -# SHA256 Fingerprint: 15:d5:b8:77:46:19:ea:7d:54:ce:1c:a6:d0:b0:c4:03:e0:37:a9:17:f1:31:e8:a0:4e:1e:6b:7a:71:ba:bc:e5 ------BEGIN CERTIFICATE----- -MIICDDCCAZGgAwIBAgIQbkepx2ypcyRAiQ8DVd2NHTAKBggqhkjOPQQDAzBHMQsw -CQYDVQQGEwJVUzEiMCAGA1UEChMZR29vZ2xlIFRydXN0IFNlcnZpY2VzIExMQzEU -MBIGA1UEAxMLR1RTIFJvb3QgUjMwHhcNMTYwNjIyMDAwMDAwWhcNMzYwNjIyMDAw -MDAwWjBHMQswCQYDVQQGEwJVUzEiMCAGA1UEChMZR29vZ2xlIFRydXN0IFNlcnZp -Y2VzIExMQzEUMBIGA1UEAxMLR1RTIFJvb3QgUjMwdjAQBgcqhkjOPQIBBgUrgQQA -IgNiAAQfTzOHMymKoYTey8chWEGJ6ladK0uFxh1MJ7x/JlFyb+Kf1qPKzEUURout -736GjOyxfi//qXGdGIRFBEFVbivqJn+7kAHjSxm65FSWRQmx1WyRRK2EE46ajA2A -DDL24CejQjBAMA4GA1UdDwEB/wQEAwIBBjAPBgNVHRMBAf8EBTADAQH/MB0GA1Ud -DgQWBBTB8Sa6oC2uhYHP0/EqEr24Cmf9vDAKBggqhkjOPQQDAwNpADBmAjEAgFuk -fCPAlaUs3L6JbyO5o91lAFJekazInXJ0glMLfalAvWhgxeG4VDvBNhcl2MG9AjEA -njWSdIUlUfUk7GRSJFClH9voy8l27OyCbvWFGFPouOOaKaqW04MjyaR7YbPMAuhd ------END CERTIFICATE----- - -# Issuer: CN=GTS Root R4 O=Google Trust Services LLC -# Subject: CN=GTS Root R4 O=Google Trust Services LLC -# Label: "GTS Root R4" -# Serial: 146587176229350439916519468929765261721 -# MD5 Fingerprint: 5d:b6:6a:c4:60:17:24:6a:1a:99:a8:4b:ee:5e:b4:26 -# SHA1 Fingerprint: 2a:1d:60:27:d9:4a:b1:0a:1c:4d:91:5c:cd:33:a0:cb:3e:2d:54:cb -# SHA256 Fingerprint: 71:cc:a5:39:1f:9e:79:4b:04:80:25:30:b3:63:e1:21:da:8a:30:43:bb:26:66:2f:ea:4d:ca:7f:c9:51:a4:bd ------BEGIN CERTIFICATE----- -MIICCjCCAZGgAwIBAgIQbkepyIuUtui7OyrYorLBmTAKBggqhkjOPQQDAzBHMQsw -CQYDVQQGEwJVUzEiMCAGA1UEChMZR29vZ2xlIFRydXN0IFNlcnZpY2VzIExMQzEU -MBIGA1UEAxMLR1RTIFJvb3QgUjQwHhcNMTYwNjIyMDAwMDAwWhcNMzYwNjIyMDAw -MDAwWjBHMQswCQYDVQQGEwJVUzEiMCAGA1UEChMZR29vZ2xlIFRydXN0IFNlcnZp -Y2VzIExMQzEUMBIGA1UEAxMLR1RTIFJvb3QgUjQwdjAQBgcqhkjOPQIBBgUrgQQA -IgNiAATzdHOnaItgrkO4NcWBMHtLSZ37wWHO5t5GvWvVYRg1rkDdc/eJkTBa6zzu -hXyiQHY7qca4R9gq55KRanPpsXI5nymfopjTX15YhmUPoYRlBtHci8nHc8iMai/l -xKvRHYqjQjBAMA4GA1UdDwEB/wQEAwIBBjAPBgNVHRMBAf8EBTADAQH/MB0GA1Ud -DgQWBBSATNbrdP9JNqPV2Py1PsVq8JQdjDAKBggqhkjOPQQDAwNnADBkAjBqUFJ0 -CMRw3J5QdCHojXohw0+WbhXRIjVhLfoIN+4Zba3bssx9BzT1YBkstTTZbyACMANx -sbqjYAuG7ZoIapVon+Kz4ZNkfF6Tpt95LY2F45TPI11xzPKwTdb+mciUqXWi4w== ------END CERTIFICATE----- - # Issuer: CN=UCA Global G2 Root O=UniTrust # Subject: CN=UCA Global G2 Root O=UniTrust # Label: "UCA Global G2 Root" @@ -4639,3 +3143,1656 @@ IQ6SwJAfzyBfyjs4x7dtOvPmRLgOMWuIjnDrnBdSqEGULoe256YSxXXfW8AKbnuk 5F6G+TaU33fD6Q3AOfF5u0aOq0NZJ7cguyPpVkAh7DE9ZapD8j3fcEThuk0mEDuY n/PIjhs4ViFqUZPTkcpG2om3PVODLAgfi49T3f+sHw== -----END CERTIFICATE----- + +# Issuer: CN=Microsoft ECC Root Certificate Authority 2017 O=Microsoft Corporation +# Subject: CN=Microsoft ECC Root Certificate Authority 2017 O=Microsoft Corporation +# Label: "Microsoft ECC Root Certificate Authority 2017" +# Serial: 136839042543790627607696632466672567020 +# MD5 Fingerprint: dd:a1:03:e6:4a:93:10:d1:bf:f0:19:42:cb:fe:ed:67 +# SHA1 Fingerprint: 99:9a:64:c3:7f:f4:7d:9f:ab:95:f1:47:69:89:14:60:ee:c4:c3:c5 +# SHA256 Fingerprint: 35:8d:f3:9d:76:4a:f9:e1:b7:66:e9:c9:72:df:35:2e:e1:5c:fa:c2:27:af:6a:d1:d7:0e:8e:4a:6e:dc:ba:02 +-----BEGIN CERTIFICATE----- +MIICWTCCAd+gAwIBAgIQZvI9r4fei7FK6gxXMQHC7DAKBggqhkjOPQQDAzBlMQsw +CQYDVQQGEwJVUzEeMBwGA1UEChMVTWljcm9zb2Z0IENvcnBvcmF0aW9uMTYwNAYD +VQQDEy1NaWNyb3NvZnQgRUNDIFJvb3QgQ2VydGlmaWNhdGUgQXV0aG9yaXR5IDIw +MTcwHhcNMTkxMjE4MjMwNjQ1WhcNNDIwNzE4MjMxNjA0WjBlMQswCQYDVQQGEwJV +UzEeMBwGA1UEChMVTWljcm9zb2Z0IENvcnBvcmF0aW9uMTYwNAYDVQQDEy1NaWNy +b3NvZnQgRUNDIFJvb3QgQ2VydGlmaWNhdGUgQXV0aG9yaXR5IDIwMTcwdjAQBgcq +hkjOPQIBBgUrgQQAIgNiAATUvD0CQnVBEyPNgASGAlEvaqiBYgtlzPbKnR5vSmZR +ogPZnZH6thaxjG7efM3beaYvzrvOcS/lpaso7GMEZpn4+vKTEAXhgShC48Zo9OYb +hGBKia/teQ87zvH2RPUBeMCjVDBSMA4GA1UdDwEB/wQEAwIBhjAPBgNVHRMBAf8E +BTADAQH/MB0GA1UdDgQWBBTIy5lycFIM+Oa+sgRXKSrPQhDtNTAQBgkrBgEEAYI3 +FQEEAwIBADAKBggqhkjOPQQDAwNoADBlAjBY8k3qDPlfXu5gKcs68tvWMoQZP3zV +L8KxzJOuULsJMsbG7X7JNpQS5GiFBqIb0C8CMQCZ6Ra0DvpWSNSkMBaReNtUjGUB +iudQZsIxtzm6uBoiB078a1QWIP8rtedMDE2mT3M= +-----END CERTIFICATE----- + +# Issuer: CN=Microsoft RSA Root Certificate Authority 2017 O=Microsoft Corporation +# Subject: CN=Microsoft RSA Root Certificate Authority 2017 O=Microsoft Corporation +# Label: "Microsoft RSA Root Certificate Authority 2017" +# Serial: 40975477897264996090493496164228220339 +# MD5 Fingerprint: 10:ff:00:ff:cf:c9:f8:c7:7a:c0:ee:35:8e:c9:0f:47 +# SHA1 Fingerprint: 73:a5:e6:4a:3b:ff:83:16:ff:0e:dc:cc:61:8a:90:6e:4e:ae:4d:74 +# SHA256 Fingerprint: c7:41:f7:0f:4b:2a:8d:88:bf:2e:71:c1:41:22:ef:53:ef:10:eb:a0:cf:a5:e6:4c:fa:20:f4:18:85:30:73:e0 +-----BEGIN CERTIFICATE----- +MIIFqDCCA5CgAwIBAgIQHtOXCV/YtLNHcB6qvn9FszANBgkqhkiG9w0BAQwFADBl +MQswCQYDVQQGEwJVUzEeMBwGA1UEChMVTWljcm9zb2Z0IENvcnBvcmF0aW9uMTYw +NAYDVQQDEy1NaWNyb3NvZnQgUlNBIFJvb3QgQ2VydGlmaWNhdGUgQXV0aG9yaXR5 +IDIwMTcwHhcNMTkxMjE4MjI1MTIyWhcNNDIwNzE4MjMwMDIzWjBlMQswCQYDVQQG +EwJVUzEeMBwGA1UEChMVTWljcm9zb2Z0IENvcnBvcmF0aW9uMTYwNAYDVQQDEy1N +aWNyb3NvZnQgUlNBIFJvb3QgQ2VydGlmaWNhdGUgQXV0aG9yaXR5IDIwMTcwggIi +MA0GCSqGSIb3DQEBAQUAA4ICDwAwggIKAoICAQDKW76UM4wplZEWCpW9R2LBifOZ +Nt9GkMml7Xhqb0eRaPgnZ1AzHaGm++DlQ6OEAlcBXZxIQIJTELy/xztokLaCLeX0 +ZdDMbRnMlfl7rEqUrQ7eS0MdhweSE5CAg2Q1OQT85elss7YfUJQ4ZVBcF0a5toW1 +HLUX6NZFndiyJrDKxHBKrmCk3bPZ7Pw71VdyvD/IybLeS2v4I2wDwAW9lcfNcztm +gGTjGqwu+UcF8ga2m3P1eDNbx6H7JyqhtJqRjJHTOoI+dkC0zVJhUXAoP8XFWvLJ +jEm7FFtNyP9nTUwSlq31/niol4fX/V4ggNyhSyL71Imtus5Hl0dVe49FyGcohJUc +aDDv70ngNXtk55iwlNpNhTs+VcQor1fznhPbRiefHqJeRIOkpcrVE7NLP8TjwuaG +YaRSMLl6IE9vDzhTyzMMEyuP1pq9KsgtsRx9S1HKR9FIJ3Jdh+vVReZIZZ2vUpC6 +W6IYZVcSn2i51BVrlMRpIpj0M+Dt+VGOQVDJNE92kKz8OMHY4Xu54+OU4UZpyw4K +UGsTuqwPN1q3ErWQgR5WrlcihtnJ0tHXUeOrO8ZV/R4O03QK0dqq6mm4lyiPSMQH ++FJDOvTKVTUssKZqwJz58oHhEmrARdlns87/I6KJClTUFLkqqNfs+avNJVgyeY+Q +W5g5xAgGwax/Dj0ApQIDAQABo1QwUjAOBgNVHQ8BAf8EBAMCAYYwDwYDVR0TAQH/ +BAUwAwEB/zAdBgNVHQ4EFgQUCctZf4aycI8awznjwNnpv7tNsiMwEAYJKwYBBAGC +NxUBBAMCAQAwDQYJKoZIhvcNAQEMBQADggIBAKyvPl3CEZaJjqPnktaXFbgToqZC +LgLNFgVZJ8og6Lq46BrsTaiXVq5lQ7GPAJtSzVXNUzltYkyLDVt8LkS/gxCP81OC +gMNPOsduET/m4xaRhPtthH80dK2Jp86519efhGSSvpWhrQlTM93uCupKUY5vVau6 +tZRGrox/2KJQJWVggEbbMwSubLWYdFQl3JPk+ONVFT24bcMKpBLBaYVu32TxU5nh +SnUgnZUP5NbcA/FZGOhHibJXWpS2qdgXKxdJ5XbLwVaZOjex/2kskZGT4d9Mozd2 +TaGf+G0eHdP67Pv0RR0Tbc/3WeUiJ3IrhvNXuzDtJE3cfVa7o7P4NHmJweDyAmH3 +pvwPuxwXC65B2Xy9J6P9LjrRk5Sxcx0ki69bIImtt2dmefU6xqaWM/5TkshGsRGR +xpl/j8nWZjEgQRCHLQzWwa80mMpkg/sTV9HB8Dx6jKXB/ZUhoHHBk2dxEuqPiApp +GWSZI1b7rCoucL5mxAyE7+WL85MB+GqQk2dLsmijtWKP6T+MejteD+eMuMZ87zf9 +dOLITzNy4ZQ5bb0Sr74MTnB8G2+NszKTc0QWbej09+CVgI+WXTik9KveCjCHk9hN +AHFiRSdLOkKEW39lt2c0Ui2cFmuqqNh7o0JMcccMyj6D5KbvtwEwXlGjefVwaaZB +RA+GsCyRxj3qrg+E +-----END CERTIFICATE----- + +# Issuer: CN=e-Szigno Root CA 2017 O=Microsec Ltd. +# Subject: CN=e-Szigno Root CA 2017 O=Microsec Ltd. +# Label: "e-Szigno Root CA 2017" +# Serial: 411379200276854331539784714 +# MD5 Fingerprint: de:1f:f6:9e:84:ae:a7:b4:21:ce:1e:58:7d:d1:84:98 +# SHA1 Fingerprint: 89:d4:83:03:4f:9e:9a:48:80:5f:72:37:d4:a9:a6:ef:cb:7c:1f:d1 +# SHA256 Fingerprint: be:b0:0b:30:83:9b:9b:c3:2c:32:e4:44:79:05:95:06:41:f2:64:21:b1:5e:d0:89:19:8b:51:8a:e2:ea:1b:99 +-----BEGIN CERTIFICATE----- +MIICQDCCAeWgAwIBAgIMAVRI7yH9l1kN9QQKMAoGCCqGSM49BAMCMHExCzAJBgNV +BAYTAkhVMREwDwYDVQQHDAhCdWRhcGVzdDEWMBQGA1UECgwNTWljcm9zZWMgTHRk +LjEXMBUGA1UEYQwOVkFUSFUtMjM1ODQ0OTcxHjAcBgNVBAMMFWUtU3ppZ25vIFJv +b3QgQ0EgMjAxNzAeFw0xNzA4MjIxMjA3MDZaFw00MjA4MjIxMjA3MDZaMHExCzAJ +BgNVBAYTAkhVMREwDwYDVQQHDAhCdWRhcGVzdDEWMBQGA1UECgwNTWljcm9zZWMg +THRkLjEXMBUGA1UEYQwOVkFUSFUtMjM1ODQ0OTcxHjAcBgNVBAMMFWUtU3ppZ25v +IFJvb3QgQ0EgMjAxNzBZMBMGByqGSM49AgEGCCqGSM49AwEHA0IABJbcPYrYsHtv +xie+RJCxs1YVe45DJH0ahFnuY2iyxl6H0BVIHqiQrb1TotreOpCmYF9oMrWGQd+H +Wyx7xf58etqjYzBhMA8GA1UdEwEB/wQFMAMBAf8wDgYDVR0PAQH/BAQDAgEGMB0G +A1UdDgQWBBSHERUI0arBeAyxr87GyZDvvzAEwDAfBgNVHSMEGDAWgBSHERUI0arB +eAyxr87GyZDvvzAEwDAKBggqhkjOPQQDAgNJADBGAiEAtVfd14pVCzbhhkT61Nlo +jbjcI4qKDdQvfepz7L9NbKgCIQDLpbQS+ue16M9+k/zzNY9vTlp8tLxOsvxyqltZ ++efcMQ== +-----END CERTIFICATE----- + +# Issuer: O=CERTSIGN SA OU=certSIGN ROOT CA G2 +# Subject: O=CERTSIGN SA OU=certSIGN ROOT CA G2 +# Label: "certSIGN Root CA G2" +# Serial: 313609486401300475190 +# MD5 Fingerprint: 8c:f1:75:8a:c6:19:cf:94:b7:f7:65:20:87:c3:97:c7 +# SHA1 Fingerprint: 26:f9:93:b4:ed:3d:28:27:b0:b9:4b:a7:e9:15:1d:a3:8d:92:e5:32 +# SHA256 Fingerprint: 65:7c:fe:2f:a7:3f:aa:38:46:25:71:f3:32:a2:36:3a:46:fc:e7:02:09:51:71:07:02:cd:fb:b6:ee:da:33:05 +-----BEGIN CERTIFICATE----- +MIIFRzCCAy+gAwIBAgIJEQA0tk7GNi02MA0GCSqGSIb3DQEBCwUAMEExCzAJBgNV +BAYTAlJPMRQwEgYDVQQKEwtDRVJUU0lHTiBTQTEcMBoGA1UECxMTY2VydFNJR04g +Uk9PVCBDQSBHMjAeFw0xNzAyMDYwOTI3MzVaFw00MjAyMDYwOTI3MzVaMEExCzAJ +BgNVBAYTAlJPMRQwEgYDVQQKEwtDRVJUU0lHTiBTQTEcMBoGA1UECxMTY2VydFNJ +R04gUk9PVCBDQSBHMjCCAiIwDQYJKoZIhvcNAQEBBQADggIPADCCAgoCggIBAMDF +dRmRfUR0dIf+DjuW3NgBFszuY5HnC2/OOwppGnzC46+CjobXXo9X69MhWf05N0Iw +vlDqtg+piNguLWkh59E3GE59kdUWX2tbAMI5Qw02hVK5U2UPHULlj88F0+7cDBrZ +uIt4ImfkabBoxTzkbFpG583H+u/E7Eu9aqSs/cwoUe+StCmrqzWaTOTECMYmzPhp +n+Sc8CnTXPnGFiWeI8MgwT0PPzhAsP6CRDiqWhqKa2NYOLQV07YRaXseVO6MGiKs +cpc/I1mbySKEwQdPzH/iV8oScLumZfNpdWO9lfsbl83kqK/20U6o2YpxJM02PbyW +xPFsqa7lzw1uKA2wDrXKUXt4FMMgL3/7FFXhEZn91QqhngLjYl/rNUssuHLoPj1P +rCy7Lobio3aP5ZMqz6WryFyNSwb/EkaseMsUBzXgqd+L6a8VTxaJW732jcZZroiF +DsGJ6x9nxUWO/203Nit4ZoORUSs9/1F3dmKh7Gc+PoGD4FapUB8fepmrY7+EF3fx +DTvf95xhszWYijqy7DwaNz9+j5LP2RIUZNoQAhVB/0/E6xyjyfqZ90bp4RjZsbgy +LcsUDFDYg2WD7rlcz8sFWkz6GZdr1l0T08JcVLwyc6B49fFtHsufpaafItzRUZ6C +eWRgKRM+o/1Pcmqr4tTluCRVLERLiohEnMqE0yo7AgMBAAGjQjBAMA8GA1UdEwEB +/wQFMAMBAf8wDgYDVR0PAQH/BAQDAgEGMB0GA1UdDgQWBBSCIS1mxteg4BXrzkwJ +d8RgnlRuAzANBgkqhkiG9w0BAQsFAAOCAgEAYN4auOfyYILVAzOBywaK8SJJ6ejq +kX/GM15oGQOGO0MBzwdw5AgeZYWR5hEit/UCI46uuR59H35s5r0l1ZUa8gWmr4UC +b6741jH/JclKyMeKqdmfS0mbEVeZkkMR3rYzpMzXjWR91M08KCy0mpbqTfXERMQl +qiCA2ClV9+BB/AYm/7k29UMUA2Z44RGx2iBfRgB4ACGlHgAoYXhvqAEBj500mv/0 +OJD7uNGzcgbJceaBxXntC6Z58hMLnPddDnskk7RI24Zf3lCGeOdA5jGokHZwYa+c +NywRtYK3qq4kNFtyDGkNzVmf9nGvnAvRCjj5BiKDUyUM/FHE5r7iOZULJK2v0ZXk +ltd0ZGtxTgI8qoXzIKNDOXZbbFD+mpwUHmUUihW9o4JFWklWatKcsWMy5WHgUyIO +pwpJ6st+H6jiYoD2EEVSmAYY3qXNL3+q1Ok+CHLsIwMCPKaq2LxndD0UF/tUSxfj +03k9bWtJySgOLnRQvwzZRjoQhsmnP+mg7H/rpXdYaXHmgwo38oZJar55CJD2AhZk +PuXaTH4MNMn5X7azKFGnpyuqSfqNZSlO42sTp5SjLVFteAxEy9/eCG/Oo2Sr05WE +1LlSVHJ7liXMvGnjSG4N0MedJ5qq+BOS3R7fY581qRY27Iy4g/Q9iY/NtBde17MX +QRBdJ3NghVdJIgc= +-----END CERTIFICATE----- + +# Issuer: CN=Trustwave Global Certification Authority O=Trustwave Holdings, Inc. +# Subject: CN=Trustwave Global Certification Authority O=Trustwave Holdings, Inc. +# Label: "Trustwave Global Certification Authority" +# Serial: 1846098327275375458322922162 +# MD5 Fingerprint: f8:1c:18:2d:2f:ba:5f:6d:a1:6c:bc:c7:ab:91:c7:0e +# SHA1 Fingerprint: 2f:8f:36:4f:e1:58:97:44:21:59:87:a5:2a:9a:d0:69:95:26:7f:b5 +# SHA256 Fingerprint: 97:55:20:15:f5:dd:fc:3c:87:88:c0:06:94:45:55:40:88:94:45:00:84:f1:00:86:70:86:bc:1a:2b:b5:8d:c8 +-----BEGIN CERTIFICATE----- +MIIF2jCCA8KgAwIBAgIMBfcOhtpJ80Y1LrqyMA0GCSqGSIb3DQEBCwUAMIGIMQsw +CQYDVQQGEwJVUzERMA8GA1UECAwISWxsaW5vaXMxEDAOBgNVBAcMB0NoaWNhZ28x +ITAfBgNVBAoMGFRydXN0d2F2ZSBIb2xkaW5ncywgSW5jLjExMC8GA1UEAwwoVHJ1 +c3R3YXZlIEdsb2JhbCBDZXJ0aWZpY2F0aW9uIEF1dGhvcml0eTAeFw0xNzA4MjMx +OTM0MTJaFw00MjA4MjMxOTM0MTJaMIGIMQswCQYDVQQGEwJVUzERMA8GA1UECAwI +SWxsaW5vaXMxEDAOBgNVBAcMB0NoaWNhZ28xITAfBgNVBAoMGFRydXN0d2F2ZSBI +b2xkaW5ncywgSW5jLjExMC8GA1UEAwwoVHJ1c3R3YXZlIEdsb2JhbCBDZXJ0aWZp +Y2F0aW9uIEF1dGhvcml0eTCCAiIwDQYJKoZIhvcNAQEBBQADggIPADCCAgoCggIB +ALldUShLPDeS0YLOvR29zd24q88KPuFd5dyqCblXAj7mY2Hf8g+CY66j96xz0Xzn +swuvCAAJWX/NKSqIk4cXGIDtiLK0thAfLdZfVaITXdHG6wZWiYj+rDKd/VzDBcdu +7oaJuogDnXIhhpCujwOl3J+IKMujkkkP7NAP4m1ET4BqstTnoApTAbqOl5F2brz8 +1Ws25kCI1nsvXwXoLG0R8+eyvpJETNKXpP7ScoFDB5zpET71ixpZfR9oWN0EACyW +80OzfpgZdNmcc9kYvkHHNHnZ9GLCQ7mzJ7Aiy/k9UscwR7PJPrhq4ufogXBeQotP +JqX+OsIgbrv4Fo7NDKm0G2x2EOFYeUY+VM6AqFcJNykbmROPDMjWLBz7BegIlT1l +RtzuzWniTY+HKE40Cz7PFNm73bZQmq131BnW2hqIyE4bJ3XYsgjxroMwuREOzYfw +hI0Vcnyh78zyiGG69Gm7DIwLdVcEuE4qFC49DxweMqZiNu5m4iK4BUBjECLzMx10 +coos9TkpoNPnG4CELcU9402x/RpvumUHO1jsQkUm+9jaJXLE9gCxInm943xZYkqc +BW89zubWR2OZxiRvchLIrH+QtAuRcOi35hYQcRfO3gZPSEF9NUqjifLJS3tBEW1n +twiYTOURGa5CgNz7kAXU+FDKvuStx8KU1xad5hePrzb7AgMBAAGjQjBAMA8GA1Ud +EwEB/wQFMAMBAf8wHQYDVR0OBBYEFJngGWcNYtt2s9o9uFvo/ULSMQ6HMA4GA1Ud +DwEB/wQEAwIBBjANBgkqhkiG9w0BAQsFAAOCAgEAmHNw4rDT7TnsTGDZqRKGFx6W +0OhUKDtkLSGm+J1WE2pIPU/HPinbbViDVD2HfSMF1OQc3Og4ZYbFdada2zUFvXfe +uyk3QAUHw5RSn8pk3fEbK9xGChACMf1KaA0HZJDmHvUqoai7PF35owgLEQzxPy0Q +lG/+4jSHg9bP5Rs1bdID4bANqKCqRieCNqcVtgimQlRXtpla4gt5kNdXElE1GYhB +aCXUNxeEFfsBctyV3lImIJgm4nb1J2/6ADtKYdkNy1GTKv0WBpanI5ojSP5RvbbE +sLFUzt5sQa0WZ37b/TjNuThOssFgy50X31ieemKyJo90lZvkWx3SD92YHJtZuSPT +MaCm/zjdzyBP6VhWOmfD0faZmZ26NraAL4hHT4a/RDqA5Dccprrql5gR0IRiR2Qe +qu5AvzSxnI9O4fKSTx+O856X3vOmeWqJcU9LJxdI/uz0UA9PSX3MReO9ekDFQdxh +VicGaeVyQYHTtgGJoC86cnn+OjC/QezHYj6RS8fZMXZC+fc8Y+wmjHMMfRod6qh8 +h6jCJ3zhM0EPz8/8AKAigJ5Kp28AsEFFtyLKaEjFQqKu3R3y4G5OBVixwJAWKqQ9 +EEC+j2Jjg6mcgn0tAumDMHzLJ8n9HmYAsC7TIS+OMxZsmO0QqAfWzJPP29FpHOTK +yeC2nOnOcXHebD8WpHk= +-----END CERTIFICATE----- + +# Issuer: CN=Trustwave Global ECC P256 Certification Authority O=Trustwave Holdings, Inc. +# Subject: CN=Trustwave Global ECC P256 Certification Authority O=Trustwave Holdings, Inc. +# Label: "Trustwave Global ECC P256 Certification Authority" +# Serial: 4151900041497450638097112925 +# MD5 Fingerprint: 5b:44:e3:8d:5d:36:86:26:e8:0d:05:d2:59:a7:83:54 +# SHA1 Fingerprint: b4:90:82:dd:45:0c:be:8b:5b:b1:66:d3:e2:a4:08:26:cd:ed:42:cf +# SHA256 Fingerprint: 94:5b:bc:82:5e:a5:54:f4:89:d1:fd:51:a7:3d:df:2e:a6:24:ac:70:19:a0:52:05:22:5c:22:a7:8c:cf:a8:b4 +-----BEGIN CERTIFICATE----- +MIICYDCCAgegAwIBAgIMDWpfCD8oXD5Rld9dMAoGCCqGSM49BAMCMIGRMQswCQYD +VQQGEwJVUzERMA8GA1UECBMISWxsaW5vaXMxEDAOBgNVBAcTB0NoaWNhZ28xITAf +BgNVBAoTGFRydXN0d2F2ZSBIb2xkaW5ncywgSW5jLjE6MDgGA1UEAxMxVHJ1c3R3 +YXZlIEdsb2JhbCBFQ0MgUDI1NiBDZXJ0aWZpY2F0aW9uIEF1dGhvcml0eTAeFw0x +NzA4MjMxOTM1MTBaFw00MjA4MjMxOTM1MTBaMIGRMQswCQYDVQQGEwJVUzERMA8G +A1UECBMISWxsaW5vaXMxEDAOBgNVBAcTB0NoaWNhZ28xITAfBgNVBAoTGFRydXN0 +d2F2ZSBIb2xkaW5ncywgSW5jLjE6MDgGA1UEAxMxVHJ1c3R3YXZlIEdsb2JhbCBF +Q0MgUDI1NiBDZXJ0aWZpY2F0aW9uIEF1dGhvcml0eTBZMBMGByqGSM49AgEGCCqG +SM49AwEHA0IABH77bOYj43MyCMpg5lOcunSNGLB4kFKA3TjASh3RqMyTpJcGOMoN +FWLGjgEqZZ2q3zSRLoHB5DOSMcT9CTqmP62jQzBBMA8GA1UdEwEB/wQFMAMBAf8w +DwYDVR0PAQH/BAUDAwcGADAdBgNVHQ4EFgQUo0EGrJBt0UrrdaVKEJmzsaGLSvcw +CgYIKoZIzj0EAwIDRwAwRAIgB+ZU2g6gWrKuEZ+Hxbb/ad4lvvigtwjzRM4q3wgh +DDcCIC0mA6AFvWvR9lz4ZcyGbbOcNEhjhAnFjXca4syc4XR7 +-----END CERTIFICATE----- + +# Issuer: CN=Trustwave Global ECC P384 Certification Authority O=Trustwave Holdings, Inc. +# Subject: CN=Trustwave Global ECC P384 Certification Authority O=Trustwave Holdings, Inc. +# Label: "Trustwave Global ECC P384 Certification Authority" +# Serial: 2704997926503831671788816187 +# MD5 Fingerprint: ea:cf:60:c4:3b:b9:15:29:40:a1:97:ed:78:27:93:d6 +# SHA1 Fingerprint: e7:f3:a3:c8:cf:6f:c3:04:2e:6d:0e:67:32:c5:9e:68:95:0d:5e:d2 +# SHA256 Fingerprint: 55:90:38:59:c8:c0:c3:eb:b8:75:9e:ce:4e:25:57:22:5f:f5:75:8b:bd:38:eb:d4:82:76:60:1e:1b:d5:80:97 +-----BEGIN CERTIFICATE----- +MIICnTCCAiSgAwIBAgIMCL2Fl2yZJ6SAaEc7MAoGCCqGSM49BAMDMIGRMQswCQYD +VQQGEwJVUzERMA8GA1UECBMISWxsaW5vaXMxEDAOBgNVBAcTB0NoaWNhZ28xITAf +BgNVBAoTGFRydXN0d2F2ZSBIb2xkaW5ncywgSW5jLjE6MDgGA1UEAxMxVHJ1c3R3 +YXZlIEdsb2JhbCBFQ0MgUDM4NCBDZXJ0aWZpY2F0aW9uIEF1dGhvcml0eTAeFw0x +NzA4MjMxOTM2NDNaFw00MjA4MjMxOTM2NDNaMIGRMQswCQYDVQQGEwJVUzERMA8G +A1UECBMISWxsaW5vaXMxEDAOBgNVBAcTB0NoaWNhZ28xITAfBgNVBAoTGFRydXN0 +d2F2ZSBIb2xkaW5ncywgSW5jLjE6MDgGA1UEAxMxVHJ1c3R3YXZlIEdsb2JhbCBF +Q0MgUDM4NCBDZXJ0aWZpY2F0aW9uIEF1dGhvcml0eTB2MBAGByqGSM49AgEGBSuB +BAAiA2IABGvaDXU1CDFHBa5FmVXxERMuSvgQMSOjfoPTfygIOiYaOs+Xgh+AtycJ +j9GOMMQKmw6sWASr9zZ9lCOkmwqKi6vr/TklZvFe/oyujUF5nQlgziip04pt89ZF +1PKYhDhloKNDMEEwDwYDVR0TAQH/BAUwAwEB/zAPBgNVHQ8BAf8EBQMDBwYAMB0G +A1UdDgQWBBRVqYSJ0sEyvRjLbKYHTsjnnb6CkDAKBggqhkjOPQQDAwNnADBkAjA3 +AZKXRRJ+oPM+rRk6ct30UJMDEr5E0k9BpIycnR+j9sKS50gU/k6bpZFXrsY3crsC +MGclCrEMXu6pY5Jv5ZAL/mYiykf9ijH3g/56vxC+GCsej/YpHpRZ744hN8tRmKVu +Sw== +-----END CERTIFICATE----- + +# Issuer: CN=NAVER Global Root Certification Authority O=NAVER BUSINESS PLATFORM Corp. +# Subject: CN=NAVER Global Root Certification Authority O=NAVER BUSINESS PLATFORM Corp. +# Label: "NAVER Global Root Certification Authority" +# Serial: 9013692873798656336226253319739695165984492813 +# MD5 Fingerprint: c8:7e:41:f6:25:3b:f5:09:b3:17:e8:46:3d:bf:d0:9b +# SHA1 Fingerprint: 8f:6b:f2:a9:27:4a:da:14:a0:c4:f4:8e:61:27:f9:c0:1e:78:5d:d1 +# SHA256 Fingerprint: 88:f4:38:dc:f8:ff:d1:fa:8f:42:91:15:ff:e5:f8:2a:e1:e0:6e:0c:70:c3:75:fa:ad:71:7b:34:a4:9e:72:65 +-----BEGIN CERTIFICATE----- +MIIFojCCA4qgAwIBAgIUAZQwHqIL3fXFMyqxQ0Rx+NZQTQ0wDQYJKoZIhvcNAQEM +BQAwaTELMAkGA1UEBhMCS1IxJjAkBgNVBAoMHU5BVkVSIEJVU0lORVNTIFBMQVRG +T1JNIENvcnAuMTIwMAYDVQQDDClOQVZFUiBHbG9iYWwgUm9vdCBDZXJ0aWZpY2F0 +aW9uIEF1dGhvcml0eTAeFw0xNzA4MTgwODU4NDJaFw0zNzA4MTgyMzU5NTlaMGkx +CzAJBgNVBAYTAktSMSYwJAYDVQQKDB1OQVZFUiBCVVNJTkVTUyBQTEFURk9STSBD +b3JwLjEyMDAGA1UEAwwpTkFWRVIgR2xvYmFsIFJvb3QgQ2VydGlmaWNhdGlvbiBB +dXRob3JpdHkwggIiMA0GCSqGSIb3DQEBAQUAA4ICDwAwggIKAoICAQC21PGTXLVA +iQqrDZBbUGOukJR0F0Vy1ntlWilLp1agS7gvQnXp2XskWjFlqxcX0TM62RHcQDaH +38dq6SZeWYp34+hInDEW+j6RscrJo+KfziFTowI2MMtSAuXaMl3Dxeb57hHHi8lE +HoSTGEq0n+USZGnQJoViAbbJAh2+g1G7XNr4rRVqmfeSVPc0W+m/6imBEtRTkZaz +kVrd/pBzKPswRrXKCAfHcXLJZtM0l/aM9BhK4dA9WkW2aacp+yPOiNgSnABIqKYP +szuSjXEOdMWLyEz59JuOuDxp7W87UC9Y7cSw0BwbagzivESq2M0UXZR4Yb8Obtoq +vC8MC3GmsxY/nOb5zJ9TNeIDoKAYv7vxvvTWjIcNQvcGufFt7QSUqP620wbGQGHf +nZ3zVHbOUzoBppJB7ASjjw2i1QnK1sua8e9DXcCrpUHPXFNwcMmIpi3Ua2FzUCaG +YQ5fG8Ir4ozVu53BA0K6lNpfqbDKzE0K70dpAy8i+/Eozr9dUGWokG2zdLAIx6yo +0es+nPxdGoMuK8u180SdOqcXYZaicdNwlhVNt0xz7hlcxVs+Qf6sdWA7G2POAN3a +CJBitOUt7kinaxeZVL6HSuOpXgRM6xBtVNbv8ejyYhbLgGvtPe31HzClrkvJE+2K +AQHJuFFYwGY6sWZLxNUxAmLpdIQM201GLQIDAQABo0IwQDAdBgNVHQ4EFgQU0p+I +36HNLL3s9TsBAZMzJ7LrYEswDgYDVR0PAQH/BAQDAgEGMA8GA1UdEwEB/wQFMAMB +Af8wDQYJKoZIhvcNAQEMBQADggIBADLKgLOdPVQG3dLSLvCkASELZ0jKbY7gyKoN +qo0hV4/GPnrK21HUUrPUloSlWGB/5QuOH/XcChWB5Tu2tyIvCZwTFrFsDDUIbatj +cu3cvuzHV+YwIHHW1xDBE1UBjCpD5EHxzzp6U5LOogMFDTjfArsQLtk70pt6wKGm ++LUx5vR1yblTmXVHIloUFcd4G7ad6Qz4G3bxhYTeodoS76TiEJd6eN4MUZeoIUCL +hr0N8F5OSza7OyAfikJW4Qsav3vQIkMsRIz75Sq0bBwcupTgE34h5prCy8VCZLQe +lHsIJchxzIdFV4XTnyliIoNRlwAYl3dqmJLJfGBs32x9SuRwTMKeuB330DTHD8z7 +p/8Dvq1wkNoL3chtl1+afwkyQf3NosxabUzyqkn+Zvjp2DXrDige7kgvOtB5CTh8 +piKCk5XQA76+AqAF3SAi428diDRgxuYKuQl1C/AH6GmWNcf7I4GOODm4RStDeKLR +LBT/DShycpWbXgnbiUSYqqFJu3FS8r/2/yehNq+4tneI3TqkbZs0kNwUXTC/t+sX +5Ie3cdCh13cV1ELX8vMxmV2b3RZtP+oGI/hGoiLtk/bdmuYqh7GYVPEi92tF4+KO +dh2ajcQGjTa3FPOdVGm3jjzVpG2Tgbet9r1ke8LJaDmgkpzNNIaRkPpkUZ3+/uul +9XXeifdy +-----END CERTIFICATE----- + +# Issuer: CN=AC RAIZ FNMT-RCM SERVIDORES SEGUROS O=FNMT-RCM OU=Ceres +# Subject: CN=AC RAIZ FNMT-RCM SERVIDORES SEGUROS O=FNMT-RCM OU=Ceres +# Label: "AC RAIZ FNMT-RCM SERVIDORES SEGUROS" +# Serial: 131542671362353147877283741781055151509 +# MD5 Fingerprint: 19:36:9c:52:03:2f:d2:d1:bb:23:cc:dd:1e:12:55:bb +# SHA1 Fingerprint: 62:ff:d9:9e:c0:65:0d:03:ce:75:93:d2:ed:3f:2d:32:c9:e3:e5:4a +# SHA256 Fingerprint: 55:41:53:b1:3d:2c:f9:dd:b7:53:bf:be:1a:4e:0a:e0:8d:0a:a4:18:70:58:fe:60:a2:b8:62:b2:e4:b8:7b:cb +-----BEGIN CERTIFICATE----- +MIICbjCCAfOgAwIBAgIQYvYybOXE42hcG2LdnC6dlTAKBggqhkjOPQQDAzB4MQsw +CQYDVQQGEwJFUzERMA8GA1UECgwIRk5NVC1SQ00xDjAMBgNVBAsMBUNlcmVzMRgw +FgYDVQRhDA9WQVRFUy1RMjgyNjAwNEoxLDAqBgNVBAMMI0FDIFJBSVogRk5NVC1S +Q00gU0VSVklET1JFUyBTRUdVUk9TMB4XDTE4MTIyMDA5MzczM1oXDTQzMTIyMDA5 +MzczM1oweDELMAkGA1UEBhMCRVMxETAPBgNVBAoMCEZOTVQtUkNNMQ4wDAYDVQQL +DAVDZXJlczEYMBYGA1UEYQwPVkFURVMtUTI4MjYwMDRKMSwwKgYDVQQDDCNBQyBS +QUlaIEZOTVQtUkNNIFNFUlZJRE9SRVMgU0VHVVJPUzB2MBAGByqGSM49AgEGBSuB +BAAiA2IABPa6V1PIyqvfNkpSIeSX0oNnnvBlUdBeh8dHsVnyV0ebAAKTRBdp20LH +sbI6GA60XYyzZl2hNPk2LEnb80b8s0RpRBNm/dfF/a82Tc4DTQdxz69qBdKiQ1oK +Um8BA06Oi6NCMEAwDwYDVR0TAQH/BAUwAwEB/zAOBgNVHQ8BAf8EBAMCAQYwHQYD +VR0OBBYEFAG5L++/EYZg8k/QQW6rcx/n0m5JMAoGCCqGSM49BAMDA2kAMGYCMQCu +SuMrQMN0EfKVrRYj3k4MGuZdpSRea0R7/DjiT8ucRRcRTBQnJlU5dUoDzBOQn5IC +MQD6SmxgiHPz7riYYqnOK8LZiqZwMR2vsJRM60/G49HzYqc8/5MuB1xJAWdpEgJy +v+c= +-----END CERTIFICATE----- + +# Issuer: CN=GlobalSign Root R46 O=GlobalSign nv-sa +# Subject: CN=GlobalSign Root R46 O=GlobalSign nv-sa +# Label: "GlobalSign Root R46" +# Serial: 1552617688466950547958867513931858518042577 +# MD5 Fingerprint: c4:14:30:e4:fa:66:43:94:2a:6a:1b:24:5f:19:d0:ef +# SHA1 Fingerprint: 53:a2:b0:4b:ca:6b:d6:45:e6:39:8a:8e:c4:0d:d2:bf:77:c3:a2:90 +# SHA256 Fingerprint: 4f:a3:12:6d:8d:3a:11:d1:c4:85:5a:4f:80:7c:ba:d6:cf:91:9d:3a:5a:88:b0:3b:ea:2c:63:72:d9:3c:40:c9 +-----BEGIN CERTIFICATE----- +MIIFWjCCA0KgAwIBAgISEdK7udcjGJ5AXwqdLdDfJWfRMA0GCSqGSIb3DQEBDAUA +MEYxCzAJBgNVBAYTAkJFMRkwFwYDVQQKExBHbG9iYWxTaWduIG52LXNhMRwwGgYD +VQQDExNHbG9iYWxTaWduIFJvb3QgUjQ2MB4XDTE5MDMyMDAwMDAwMFoXDTQ2MDMy +MDAwMDAwMFowRjELMAkGA1UEBhMCQkUxGTAXBgNVBAoTEEdsb2JhbFNpZ24gbnYt +c2ExHDAaBgNVBAMTE0dsb2JhbFNpZ24gUm9vdCBSNDYwggIiMA0GCSqGSIb3DQEB +AQUAA4ICDwAwggIKAoICAQCsrHQy6LNl5brtQyYdpokNRbopiLKkHWPd08EsCVeJ +OaFV6Wc0dwxu5FUdUiXSE2te4R2pt32JMl8Nnp8semNgQB+msLZ4j5lUlghYruQG +vGIFAha/r6gjA7aUD7xubMLL1aa7DOn2wQL7Id5m3RerdELv8HQvJfTqa1VbkNud +316HCkD7rRlr+/fKYIje2sGP1q7Vf9Q8g+7XFkyDRTNrJ9CG0Bwta/OrffGFqfUo +0q3v84RLHIf8E6M6cqJaESvWJ3En7YEtbWaBkoe0G1h6zD8K+kZPTXhc+CtI4wSE +y132tGqzZfxCnlEmIyDLPRT5ge1lFgBPGmSXZgjPjHvjK8Cd+RTyG/FWaha/LIWF +zXg4mutCagI0GIMXTpRW+LaCtfOW3T3zvn8gdz57GSNrLNRyc0NXfeD412lPFzYE ++cCQYDdF3uYM2HSNrpyibXRdQr4G9dlkbgIQrImwTDsHTUB+JMWKmIJ5jqSngiCN +I/onccnfxkF0oE32kRbcRoxfKWMxWXEM2G/CtjJ9++ZdU6Z+Ffy7dXxd7Pj2Fxzs +x2sZy/N78CsHpdlseVR2bJ0cpm4O6XkMqCNqo98bMDGfsVR7/mrLZqrcZdCinkqa +ByFrgY/bxFn63iLABJzjqls2k+g9vXqhnQt2sQvHnf3PmKgGwvgqo6GDoLclcqUC +4wIDAQABo0IwQDAOBgNVHQ8BAf8EBAMCAYYwDwYDVR0TAQH/BAUwAwEB/zAdBgNV +HQ4EFgQUA1yrc4GHqMywptWU4jaWSf8FmSwwDQYJKoZIhvcNAQEMBQADggIBAHx4 +7PYCLLtbfpIrXTncvtgdokIzTfnvpCo7RGkerNlFo048p9gkUbJUHJNOxO97k4Vg +JuoJSOD1u8fpaNK7ajFxzHmuEajwmf3lH7wvqMxX63bEIaZHU1VNaL8FpO7XJqti +2kM3S+LGteWygxk6x9PbTZ4IevPuzz5i+6zoYMzRx6Fcg0XERczzF2sUyQQCPtIk +pnnpHs6i58FZFZ8d4kuaPp92CC1r2LpXFNqD6v6MVenQTqnMdzGxRBF6XLE+0xRF +FRhiJBPSy03OXIPBNvIQtQ6IbbjhVp+J3pZmOUdkLG5NrmJ7v2B0GbhWrJKsFjLt +rWhV/pi60zTe9Mlhww6G9kuEYO4Ne7UyWHmRVSyBQ7N0H3qqJZ4d16GLuc1CLgSk +ZoNNiTW2bKg2SnkheCLQQrzRQDGQob4Ez8pn7fXwgNNgyYMqIgXQBztSvwyeqiv5 +u+YfjyW6hY0XHgL+XVAEV8/+LbzvXMAaq7afJMbfc2hIkCwU9D9SGuTSyxTDYWnP +4vkYxboznxSjBF25cfe1lNj2M8FawTSLfJvdkzrnE6JwYZ+vj+vYxXX4M2bUdGc6 +N3ec592kD3ZDZopD8p/7DEJ4Y9HiD2971KE9dJeFt0g5QdYg/NA6s/rob8SKunE3 +vouXsXgxT7PntgMTzlSdriVZzH81Xwj3QEUxeCp6 +-----END CERTIFICATE----- + +# Issuer: CN=GlobalSign Root E46 O=GlobalSign nv-sa +# Subject: CN=GlobalSign Root E46 O=GlobalSign nv-sa +# Label: "GlobalSign Root E46" +# Serial: 1552617690338932563915843282459653771421763 +# MD5 Fingerprint: b5:b8:66:ed:de:08:83:e3:c9:e2:01:34:06:ac:51:6f +# SHA1 Fingerprint: 39:b4:6c:d5:fe:80:06:eb:e2:2f:4a:bb:08:33:a0:af:db:b9:dd:84 +# SHA256 Fingerprint: cb:b9:c4:4d:84:b8:04:3e:10:50:ea:31:a6:9f:51:49:55:d7:bf:d2:e2:c6:b4:93:01:01:9a:d6:1d:9f:50:58 +-----BEGIN CERTIFICATE----- +MIICCzCCAZGgAwIBAgISEdK7ujNu1LzmJGjFDYQdmOhDMAoGCCqGSM49BAMDMEYx +CzAJBgNVBAYTAkJFMRkwFwYDVQQKExBHbG9iYWxTaWduIG52LXNhMRwwGgYDVQQD +ExNHbG9iYWxTaWduIFJvb3QgRTQ2MB4XDTE5MDMyMDAwMDAwMFoXDTQ2MDMyMDAw +MDAwMFowRjELMAkGA1UEBhMCQkUxGTAXBgNVBAoTEEdsb2JhbFNpZ24gbnYtc2Ex +HDAaBgNVBAMTE0dsb2JhbFNpZ24gUm9vdCBFNDYwdjAQBgcqhkjOPQIBBgUrgQQA +IgNiAAScDrHPt+ieUnd1NPqlRqetMhkytAepJ8qUuwzSChDH2omwlwxwEwkBjtjq +R+q+soArzfwoDdusvKSGN+1wCAB16pMLey5SnCNoIwZD7JIvU4Tb+0cUB+hflGdd +yXqBPCCjQjBAMA4GA1UdDwEB/wQEAwIBhjAPBgNVHRMBAf8EBTADAQH/MB0GA1Ud +DgQWBBQxCpCPtsad0kRLgLWi5h+xEk8blTAKBggqhkjOPQQDAwNoADBlAjEA31SQ +7Zvvi5QCkxeCmb6zniz2C5GMn0oUsfZkvLtoURMMA/cVi4RguYv/Uo7njLwcAjA8 ++RHUjE7AwWHCFUyqqx0LMV87HOIAl0Qx5v5zli/altP+CAezNIm8BZ/3Hobui3A= +-----END CERTIFICATE----- + +# Issuer: CN=ANF Secure Server Root CA O=ANF Autoridad de Certificacion OU=ANF CA Raiz +# Subject: CN=ANF Secure Server Root CA O=ANF Autoridad de Certificacion OU=ANF CA Raiz +# Label: "ANF Secure Server Root CA" +# Serial: 996390341000653745 +# MD5 Fingerprint: 26:a6:44:5a:d9:af:4e:2f:b2:1d:b6:65:b0:4e:e8:96 +# SHA1 Fingerprint: 5b:6e:68:d0:cc:15:b6:a0:5f:1e:c1:5f:ae:02:fc:6b:2f:5d:6f:74 +# SHA256 Fingerprint: fb:8f:ec:75:91:69:b9:10:6b:1e:51:16:44:c6:18:c5:13:04:37:3f:6c:06:43:08:8d:8b:ef:fd:1b:99:75:99 +-----BEGIN CERTIFICATE----- +MIIF7zCCA9egAwIBAgIIDdPjvGz5a7EwDQYJKoZIhvcNAQELBQAwgYQxEjAQBgNV +BAUTCUc2MzI4NzUxMDELMAkGA1UEBhMCRVMxJzAlBgNVBAoTHkFORiBBdXRvcmlk +YWQgZGUgQ2VydGlmaWNhY2lvbjEUMBIGA1UECxMLQU5GIENBIFJhaXoxIjAgBgNV +BAMTGUFORiBTZWN1cmUgU2VydmVyIFJvb3QgQ0EwHhcNMTkwOTA0MTAwMDM4WhcN +MzkwODMwMTAwMDM4WjCBhDESMBAGA1UEBRMJRzYzMjg3NTEwMQswCQYDVQQGEwJF +UzEnMCUGA1UEChMeQU5GIEF1dG9yaWRhZCBkZSBDZXJ0aWZpY2FjaW9uMRQwEgYD +VQQLEwtBTkYgQ0EgUmFpejEiMCAGA1UEAxMZQU5GIFNlY3VyZSBTZXJ2ZXIgUm9v +dCBDQTCCAiIwDQYJKoZIhvcNAQEBBQADggIPADCCAgoCggIBANvrayvmZFSVgpCj +cqQZAZ2cC4Ffc0m6p6zzBE57lgvsEeBbphzOG9INgxwruJ4dfkUyYA8H6XdYfp9q +yGFOtibBTI3/TO80sh9l2Ll49a2pcbnvT1gdpd50IJeh7WhM3pIXS7yr/2WanvtH +2Vdy8wmhrnZEE26cLUQ5vPnHO6RYPUG9tMJJo8gN0pcvB2VSAKduyK9o7PQUlrZX +H1bDOZ8rbeTzPvY1ZNoMHKGESy9LS+IsJJ1tk0DrtSOOMspvRdOoiXsezx76W0OL +zc2oD2rKDF65nkeP8Nm2CgtYZRczuSPkdxl9y0oukntPLxB3sY0vaJxizOBQ+OyR +p1RMVwnVdmPF6GUe7m1qzwmd+nxPrWAI/VaZDxUse6mAq4xhj0oHdkLePfTdsiQz +W7i1o0TJrH93PB0j7IKppuLIBkwC/qxcmZkLLxCKpvR/1Yd0DVlJRfbwcVw5Kda/ +SiOL9V8BY9KHcyi1Swr1+KuCLH5zJTIdC2MKF4EA/7Z2Xue0sUDKIbvVgFHlSFJn +LNJhiQcND85Cd8BEc5xEUKDbEAotlRyBr+Qc5RQe8TZBAQIvfXOn3kLMTOmJDVb3 +n5HUA8ZsyY/b2BzgQJhdZpmYgG4t/wHFzstGH6wCxkPmrqKEPMVOHj1tyRRM4y5B +u8o5vzY8KhmqQYdOpc5LMnndkEl/AgMBAAGjYzBhMB8GA1UdIwQYMBaAFJxf0Gxj +o1+TypOYCK2Mh6UsXME3MB0GA1UdDgQWBBScX9BsY6Nfk8qTmAitjIelLFzBNzAO +BgNVHQ8BAf8EBAMCAYYwDwYDVR0TAQH/BAUwAwEB/zANBgkqhkiG9w0BAQsFAAOC +AgEATh65isagmD9uw2nAalxJUqzLK114OMHVVISfk/CHGT0sZonrDUL8zPB1hT+L +9IBdeeUXZ701guLyPI59WzbLWoAAKfLOKyzxj6ptBZNscsdW699QIyjlRRA96Gej +rw5VD5AJYu9LWaL2U/HANeQvwSS9eS9OICI7/RogsKQOLHDtdD+4E5UGUcjohybK +pFtqFiGS3XNgnhAY3jyB6ugYw3yJ8otQPr0R4hUDqDZ9MwFsSBXXiJCZBMXM5gf0 +vPSQ7RPi6ovDj6MzD8EpTBNO2hVWcXNyglD2mjN8orGoGjR0ZVzO0eurU+AagNjq +OknkJjCb5RyKqKkVMoaZkgoQI1YS4PbOTOK7vtuNknMBZi9iPrJyJ0U27U1W45eZ +/zo1PqVUSlJZS2Db7v54EX9K3BR5YLZrZAPbFYPhor72I5dQ8AkzNqdxliXzuUJ9 +2zg/LFis6ELhDtjTO0wugumDLmsx2d1Hhk9tl5EuT+IocTUW0fJz/iUrB0ckYyfI ++PbZa/wSMVYIwFNCr5zQM378BvAxRAMU8Vjq8moNqRGyg77FGr8H6lnco4g175x2 +MjxNBiLOFeXdntiP2t7SxDnlF4HPOEfrf4htWRvfn0IUrn7PqLBmZdo3r5+qPeoo +tt7VMVgWglvquxl1AnMaykgaIZOQCo6ThKd9OyMYkomgjaw= +-----END CERTIFICATE----- + +# Issuer: CN=Certum EC-384 CA O=Asseco Data Systems S.A. OU=Certum Certification Authority +# Subject: CN=Certum EC-384 CA O=Asseco Data Systems S.A. OU=Certum Certification Authority +# Label: "Certum EC-384 CA" +# Serial: 160250656287871593594747141429395092468 +# MD5 Fingerprint: b6:65:b3:96:60:97:12:a1:ec:4e:e1:3d:a3:c6:c9:f1 +# SHA1 Fingerprint: f3:3e:78:3c:ac:df:f4:a2:cc:ac:67:55:69:56:d7:e5:16:3c:e1:ed +# SHA256 Fingerprint: 6b:32:80:85:62:53:18:aa:50:d1:73:c9:8d:8b:da:09:d5:7e:27:41:3d:11:4c:f7:87:a0:f5:d0:6c:03:0c:f6 +-----BEGIN CERTIFICATE----- +MIICZTCCAeugAwIBAgIQeI8nXIESUiClBNAt3bpz9DAKBggqhkjOPQQDAzB0MQsw +CQYDVQQGEwJQTDEhMB8GA1UEChMYQXNzZWNvIERhdGEgU3lzdGVtcyBTLkEuMScw +JQYDVQQLEx5DZXJ0dW0gQ2VydGlmaWNhdGlvbiBBdXRob3JpdHkxGTAXBgNVBAMT +EENlcnR1bSBFQy0zODQgQ0EwHhcNMTgwMzI2MDcyNDU0WhcNNDMwMzI2MDcyNDU0 +WjB0MQswCQYDVQQGEwJQTDEhMB8GA1UEChMYQXNzZWNvIERhdGEgU3lzdGVtcyBT +LkEuMScwJQYDVQQLEx5DZXJ0dW0gQ2VydGlmaWNhdGlvbiBBdXRob3JpdHkxGTAX +BgNVBAMTEENlcnR1bSBFQy0zODQgQ0EwdjAQBgcqhkjOPQIBBgUrgQQAIgNiAATE +KI6rGFtqvm5kN2PkzeyrOvfMobgOgknXhimfoZTy42B4mIF4Bk3y7JoOV2CDn7Tm +Fy8as10CW4kjPMIRBSqniBMY81CE1700LCeJVf/OTOffph8oxPBUw7l8t1Ot68Kj +QjBAMA8GA1UdEwEB/wQFMAMBAf8wHQYDVR0OBBYEFI0GZnQkdjrzife81r1HfS+8 +EF9LMA4GA1UdDwEB/wQEAwIBBjAKBggqhkjOPQQDAwNoADBlAjADVS2m5hjEfO/J +UG7BJw+ch69u1RsIGL2SKcHvlJF40jocVYli5RsJHrpka/F2tNQCMQC0QoSZ/6vn +nvuRlydd3LBbMHHOXjgaatkl5+r3YZJW+OraNsKHZZYuciUvf9/DE8k= +-----END CERTIFICATE----- + +# Issuer: CN=Certum Trusted Root CA O=Asseco Data Systems S.A. OU=Certum Certification Authority +# Subject: CN=Certum Trusted Root CA O=Asseco Data Systems S.A. OU=Certum Certification Authority +# Label: "Certum Trusted Root CA" +# Serial: 40870380103424195783807378461123655149 +# MD5 Fingerprint: 51:e1:c2:e7:fe:4c:84:af:59:0e:2f:f4:54:6f:ea:29 +# SHA1 Fingerprint: c8:83:44:c0:18:ae:9f:cc:f1:87:b7:8f:22:d1:c5:d7:45:84:ba:e5 +# SHA256 Fingerprint: fe:76:96:57:38:55:77:3e:37:a9:5e:7a:d4:d9:cc:96:c3:01:57:c1:5d:31:76:5b:a9:b1:57:04:e1:ae:78:fd +-----BEGIN CERTIFICATE----- +MIIFwDCCA6igAwIBAgIQHr9ZULjJgDdMBvfrVU+17TANBgkqhkiG9w0BAQ0FADB6 +MQswCQYDVQQGEwJQTDEhMB8GA1UEChMYQXNzZWNvIERhdGEgU3lzdGVtcyBTLkEu +MScwJQYDVQQLEx5DZXJ0dW0gQ2VydGlmaWNhdGlvbiBBdXRob3JpdHkxHzAdBgNV +BAMTFkNlcnR1bSBUcnVzdGVkIFJvb3QgQ0EwHhcNMTgwMzE2MTIxMDEzWhcNNDMw +MzE2MTIxMDEzWjB6MQswCQYDVQQGEwJQTDEhMB8GA1UEChMYQXNzZWNvIERhdGEg +U3lzdGVtcyBTLkEuMScwJQYDVQQLEx5DZXJ0dW0gQ2VydGlmaWNhdGlvbiBBdXRo +b3JpdHkxHzAdBgNVBAMTFkNlcnR1bSBUcnVzdGVkIFJvb3QgQ0EwggIiMA0GCSqG +SIb3DQEBAQUAA4ICDwAwggIKAoICAQDRLY67tzbqbTeRn06TpwXkKQMlzhyC93yZ +n0EGze2jusDbCSzBfN8pfktlL5On1AFrAygYo9idBcEq2EXxkd7fO9CAAozPOA/q +p1x4EaTByIVcJdPTsuclzxFUl6s1wB52HO8AU5853BSlLCIls3Jy/I2z5T4IHhQq +NwuIPMqw9MjCoa68wb4pZ1Xi/K1ZXP69VyywkI3C7Te2fJmItdUDmj0VDT06qKhF +8JVOJVkdzZhpu9PMMsmN74H+rX2Ju7pgE8pllWeg8xn2A1bUatMn4qGtg/BKEiJ3 +HAVz4hlxQsDsdUaakFjgao4rpUYwBI4Zshfjvqm6f1bxJAPXsiEodg42MEx51UGa +mqi4NboMOvJEGyCI98Ul1z3G4z5D3Yf+xOr1Uz5MZf87Sst4WmsXXw3Hw09Omiqi +7VdNIuJGmj8PkTQkfVXjjJU30xrwCSss0smNtA0Aq2cpKNgB9RkEth2+dv5yXMSF +ytKAQd8FqKPVhJBPC/PgP5sZ0jeJP/J7UhyM9uH3PAeXjA6iWYEMspA90+NZRu0P +qafegGtaqge2Gcu8V/OXIXoMsSt0Puvap2ctTMSYnjYJdmZm/Bo/6khUHL4wvYBQ +v3y1zgD2DGHZ5yQD4OMBgQ692IU0iL2yNqh7XAjlRICMb/gv1SHKHRzQ+8S1h9E6 +Tsd2tTVItQIDAQABo0IwQDAPBgNVHRMBAf8EBTADAQH/MB0GA1UdDgQWBBSM+xx1 +vALTn04uSNn5YFSqxLNP+jAOBgNVHQ8BAf8EBAMCAQYwDQYJKoZIhvcNAQENBQAD +ggIBAEii1QALLtA/vBzVtVRJHlpr9OTy4EA34MwUe7nJ+jW1dReTagVphZzNTxl4 +WxmB82M+w85bj/UvXgF2Ez8sALnNllI5SW0ETsXpD4YN4fqzX4IS8TrOZgYkNCvo +zMrnadyHncI013nR03e4qllY/p0m+jiGPp2Kh2RX5Rc64vmNueMzeMGQ2Ljdt4NR +5MTMI9UGfOZR0800McD2RrsLrfw9EAUqO0qRJe6M1ISHgCq8CYyqOhNf6DR5UMEQ +GfnTKB7U0VEwKbOukGfWHwpjscWpxkIxYxeU72nLL/qMFH3EQxiJ2fAyQOaA4kZf +5ePBAFmo+eggvIksDkc0C+pXwlM2/KfUrzHN/gLldfq5Jwn58/U7yn2fqSLLiMmq +0Uc9NneoWWRrJ8/vJ8HjJLWG965+Mk2weWjROeiQWMODvA8s1pfrzgzhIMfatz7D +P78v3DSk+yshzWePS/Tj6tQ/50+6uaWTRRxmHyH6ZF5v4HaUMst19W7l9o/HuKTM +qJZ9ZPskWkoDbGs4xugDQ5r3V7mzKWmTOPQD8rv7gmsHINFSH5pkAnuYZttcTVoP +0ISVoDwUQwbKytu4QTbaakRnh6+v40URFWkIsr4WOZckbxJF0WddCajJFdr60qZf +E2Efv4WstK2tBZQIgx51F9NxO5NQI1mg7TyRVJ12AMXDuDjb +-----END CERTIFICATE----- + +# Issuer: CN=TunTrust Root CA O=Agence Nationale de Certification Electronique +# Subject: CN=TunTrust Root CA O=Agence Nationale de Certification Electronique +# Label: "TunTrust Root CA" +# Serial: 108534058042236574382096126452369648152337120275 +# MD5 Fingerprint: 85:13:b9:90:5b:36:5c:b6:5e:b8:5a:f8:e0:31:57:b4 +# SHA1 Fingerprint: cf:e9:70:84:0f:e0:73:0f:9d:f6:0c:7f:2c:4b:ee:20:46:34:9c:bb +# SHA256 Fingerprint: 2e:44:10:2a:b5:8c:b8:54:19:45:1c:8e:19:d9:ac:f3:66:2c:af:bc:61:4b:6a:53:96:0a:30:f7:d0:e2:eb:41 +-----BEGIN CERTIFICATE----- +MIIFszCCA5ugAwIBAgIUEwLV4kBMkkaGFmddtLu7sms+/BMwDQYJKoZIhvcNAQEL +BQAwYTELMAkGA1UEBhMCVE4xNzA1BgNVBAoMLkFnZW5jZSBOYXRpb25hbGUgZGUg +Q2VydGlmaWNhdGlvbiBFbGVjdHJvbmlxdWUxGTAXBgNVBAMMEFR1blRydXN0IFJv +b3QgQ0EwHhcNMTkwNDI2MDg1NzU2WhcNNDQwNDI2MDg1NzU2WjBhMQswCQYDVQQG +EwJUTjE3MDUGA1UECgwuQWdlbmNlIE5hdGlvbmFsZSBkZSBDZXJ0aWZpY2F0aW9u +IEVsZWN0cm9uaXF1ZTEZMBcGA1UEAwwQVHVuVHJ1c3QgUm9vdCBDQTCCAiIwDQYJ +KoZIhvcNAQEBBQADggIPADCCAgoCggIBAMPN0/y9BFPdDCA61YguBUtB9YOCfvdZ +n56eY+hz2vYGqU8ftPkLHzmMmiDQfgbU7DTZhrx1W4eI8NLZ1KMKsmwb60ksPqxd +2JQDoOw05TDENX37Jk0bbjBU2PWARZw5rZzJJQRNmpA+TkBuimvNKWfGzC3gdOgF +VwpIUPp6Q9p+7FuaDmJ2/uqdHYVy7BG7NegfJ7/Boce7SBbdVtfMTqDhuazb1YMZ +GoXRlJfXyqNlC/M4+QKu3fZnz8k/9YosRxqZbwUN/dAdgjH8KcwAWJeRTIAAHDOF +li/LQcKLEITDCSSJH7UP2dl3RxiSlGBcx5kDPP73lad9UKGAwqmDrViWVSHbhlnU +r8a83YFuB9tgYv7sEG7aaAH0gxupPqJbI9dkxt/con3YS7qC0lH4Zr8GRuR5KiY2 +eY8fTpkdso8MDhz/yV3A/ZAQprE38806JG60hZC/gLkMjNWb1sjxVj8agIl6qeIb +MlEsPvLfe/ZdeikZjuXIvTZxi11Mwh0/rViizz1wTaZQmCXcI/m4WEEIcb9PuISg +jwBUFfyRbVinljvrS5YnzWuioYasDXxU5mZMZl+QviGaAkYt5IPCgLnPSz7ofzwB +7I9ezX/SKEIBlYrilz0QIX32nRzFNKHsLA4KUiwSVXAkPcvCFDVDXSdOvsC9qnyW +5/yeYa1E0wCXAgMBAAGjYzBhMB0GA1UdDgQWBBQGmpsfU33x9aTI04Y+oXNZtPdE +ITAPBgNVHRMBAf8EBTADAQH/MB8GA1UdIwQYMBaAFAaamx9TffH1pMjThj6hc1m0 +90QhMA4GA1UdDwEB/wQEAwIBBjANBgkqhkiG9w0BAQsFAAOCAgEAqgVutt0Vyb+z +xiD2BkewhpMl0425yAA/l/VSJ4hxyXT968pk21vvHl26v9Hr7lxpuhbI87mP0zYu +QEkHDVneixCwSQXi/5E/S7fdAo74gShczNxtr18UnH1YeA32gAm56Q6XKRm4t+v4 +FstVEuTGfbvE7Pi1HE4+Z7/FXxttbUcoqgRYYdZ2vyJ/0Adqp2RT8JeNnYA/u8EH +22Wv5psymsNUk8QcCMNE+3tjEUPRahphanltkE8pjkcFwRJpadbGNjHh/PqAulxP +xOu3Mqz4dWEX1xAZufHSCe96Qp1bWgvUxpVOKs7/B9dPfhgGiPEZtdmYu65xxBzn +dFlY7wyJz4sfdZMaBBSSSFCp61cpABbjNhzI+L/wM9VBD8TMPN3pM0MBkRArHtG5 +Xc0yGYuPjCB31yLEQtyEFpslbei0VXF/sHyz03FJuc9SpAQ/3D2gu68zngowYI7b +nV2UqL1g52KAdoGDDIzMMEZJ4gzSqK/rYXHv5yJiqfdcZGyfFoxnNidF9Ql7v/YQ +CvGwjVRDjAS6oz/v4jXH+XTgbzRB0L9zZVcg+ZtnemZoJE6AZb0QmQZZ8mWvuMZH +u/2QeItBcy6vVR/cO5JyboTT0GFMDcx2V+IthSIVNg3rAZ3r2OvEhJn7wAzMMujj +d9qDRIueVSjAi1jTkD5OGwDxFa2DK5o= +-----END CERTIFICATE----- + +# Issuer: CN=HARICA TLS RSA Root CA 2021 O=Hellenic Academic and Research Institutions CA +# Subject: CN=HARICA TLS RSA Root CA 2021 O=Hellenic Academic and Research Institutions CA +# Label: "HARICA TLS RSA Root CA 2021" +# Serial: 76817823531813593706434026085292783742 +# MD5 Fingerprint: 65:47:9b:58:86:dd:2c:f0:fc:a2:84:1f:1e:96:c4:91 +# SHA1 Fingerprint: 02:2d:05:82:fa:88:ce:14:0c:06:79:de:7f:14:10:e9:45:d7:a5:6d +# SHA256 Fingerprint: d9:5d:0e:8e:da:79:52:5b:f9:be:b1:1b:14:d2:10:0d:32:94:98:5f:0c:62:d9:fa:bd:9c:d9:99:ec:cb:7b:1d +-----BEGIN CERTIFICATE----- +MIIFpDCCA4ygAwIBAgIQOcqTHO9D88aOk8f0ZIk4fjANBgkqhkiG9w0BAQsFADBs +MQswCQYDVQQGEwJHUjE3MDUGA1UECgwuSGVsbGVuaWMgQWNhZGVtaWMgYW5kIFJl +c2VhcmNoIEluc3RpdHV0aW9ucyBDQTEkMCIGA1UEAwwbSEFSSUNBIFRMUyBSU0Eg +Um9vdCBDQSAyMDIxMB4XDTIxMDIxOTEwNTUzOFoXDTQ1MDIxMzEwNTUzN1owbDEL +MAkGA1UEBhMCR1IxNzA1BgNVBAoMLkhlbGxlbmljIEFjYWRlbWljIGFuZCBSZXNl +YXJjaCBJbnN0aXR1dGlvbnMgQ0ExJDAiBgNVBAMMG0hBUklDQSBUTFMgUlNBIFJv +b3QgQ0EgMjAyMTCCAiIwDQYJKoZIhvcNAQEBBQADggIPADCCAgoCggIBAIvC569l +mwVnlskNJLnQDmT8zuIkGCyEf3dRywQRNrhe7Wlxp57kJQmXZ8FHws+RFjZiPTgE +4VGC/6zStGndLuwRo0Xua2s7TL+MjaQenRG56Tj5eg4MmOIjHdFOY9TnuEFE+2uv +a9of08WRiFukiZLRgeaMOVig1mlDqa2YUlhu2wr7a89o+uOkXjpFc5gH6l8Cct4M +pbOfrqkdtx2z/IpZ525yZa31MJQjB/OCFks1mJxTuy/K5FrZx40d/JiZ+yykgmvw +Kh+OC19xXFyuQnspiYHLA6OZyoieC0AJQTPb5lh6/a6ZcMBaD9YThnEvdmn8kN3b +LW7R8pv1GmuebxWMevBLKKAiOIAkbDakO/IwkfN4E8/BPzWr8R0RI7VDIp4BkrcY +AuUR0YLbFQDMYTfBKnya4dC6s1BG7oKsnTH4+yPiAwBIcKMJJnkVU2DzOFytOOqB +AGMUuTNe3QvboEUHGjMJ+E20pwKmafTCWQWIZYVWrkvL4N48fS0ayOn7H6NhStYq +E613TBoYm5EPWNgGVMWX+Ko/IIqmhaZ39qb8HOLubpQzKoNQhArlT4b4UEV4AIHr +W2jjJo3Me1xR9BQsQL4aYB16cmEdH2MtiKrOokWQCPxrvrNQKlr9qEgYRtaQQJKQ +CoReaDH46+0N0x3GfZkYVVYnZS6NRcUk7M7jAgMBAAGjQjBAMA8GA1UdEwEB/wQF +MAMBAf8wHQYDVR0OBBYEFApII6ZgpJIKM+qTW8VX6iVNvRLuMA4GA1UdDwEB/wQE +AwIBhjANBgkqhkiG9w0BAQsFAAOCAgEAPpBIqm5iFSVmewzVjIuJndftTgfvnNAU +X15QvWiWkKQUEapobQk1OUAJ2vQJLDSle1mESSmXdMgHHkdt8s4cUCbjnj1AUz/3 +f5Z2EMVGpdAgS1D0NTsY9FVqQRtHBmg8uwkIYtlfVUKqrFOFrJVWNlar5AWMxaja +H6NpvVMPxP/cyuN+8kyIhkdGGvMA9YCRotxDQpSbIPDRzbLrLFPCU3hKTwSUQZqP +JzLB5UkZv/HywouoCjkxKLR9YjYsTewfM7Z+d21+UPCfDtcRj88YxeMn/ibvBZ3P +zzfF0HvaO7AWhAw6k9a+F9sPPg4ZeAnHqQJyIkv3N3a6dcSFA1pj1bF1BcK5vZSt +jBWZp5N99sXzqnTPBIWUmAD04vnKJGW/4GKvyMX6ssmeVkjaef2WdhW+o45WxLM0 +/L5H9MG0qPzVMIho7suuyWPEdr6sOBjhXlzPrjoiUevRi7PzKzMHVIf6tLITe7pT +BGIBnfHAT+7hOtSLIBD6Alfm78ELt5BGnBkpjNxvoEppaZS3JGWg/6w/zgH7IS79 +aPib8qXPMThcFarmlwDB31qlpzmq6YR/PFGoOtmUW4y/Twhx5duoXNTSpv4Ao8YW +xw/ogM4cKGR0GQjTQuPOAF1/sdwTsOEFy9EgqoZ0njnnkf3/W9b3raYvAwtt41dU +63ZTGI0RmLo= +-----END CERTIFICATE----- + +# Issuer: CN=HARICA TLS ECC Root CA 2021 O=Hellenic Academic and Research Institutions CA +# Subject: CN=HARICA TLS ECC Root CA 2021 O=Hellenic Academic and Research Institutions CA +# Label: "HARICA TLS ECC Root CA 2021" +# Serial: 137515985548005187474074462014555733966 +# MD5 Fingerprint: ae:f7:4c:e5:66:35:d1:b7:9b:8c:22:93:74:d3:4b:b0 +# SHA1 Fingerprint: bc:b0:c1:9d:e9:98:92:70:19:38:57:e9:8d:a7:b4:5d:6e:ee:01:48 +# SHA256 Fingerprint: 3f:99:cc:47:4a:cf:ce:4d:fe:d5:87:94:66:5e:47:8d:15:47:73:9f:2e:78:0f:1b:b4:ca:9b:13:30:97:d4:01 +-----BEGIN CERTIFICATE----- +MIICVDCCAdugAwIBAgIQZ3SdjXfYO2rbIvT/WeK/zjAKBggqhkjOPQQDAzBsMQsw +CQYDVQQGEwJHUjE3MDUGA1UECgwuSGVsbGVuaWMgQWNhZGVtaWMgYW5kIFJlc2Vh +cmNoIEluc3RpdHV0aW9ucyBDQTEkMCIGA1UEAwwbSEFSSUNBIFRMUyBFQ0MgUm9v +dCBDQSAyMDIxMB4XDTIxMDIxOTExMDExMFoXDTQ1MDIxMzExMDEwOVowbDELMAkG +A1UEBhMCR1IxNzA1BgNVBAoMLkhlbGxlbmljIEFjYWRlbWljIGFuZCBSZXNlYXJj +aCBJbnN0aXR1dGlvbnMgQ0ExJDAiBgNVBAMMG0hBUklDQSBUTFMgRUNDIFJvb3Qg +Q0EgMjAyMTB2MBAGByqGSM49AgEGBSuBBAAiA2IABDgI/rGgltJ6rK9JOtDA4MM7 +KKrxcm1lAEeIhPyaJmuqS7psBAqIXhfyVYf8MLA04jRYVxqEU+kw2anylnTDUR9Y +STHMmE5gEYd103KUkE+bECUqqHgtvpBBWJAVcqeht6NCMEAwDwYDVR0TAQH/BAUw +AwEB/zAdBgNVHQ4EFgQUyRtTgRL+BNUW0aq8mm+3oJUZbsowDgYDVR0PAQH/BAQD +AgGGMAoGCCqGSM49BAMDA2cAMGQCMBHervjcToiwqfAircJRQO9gcS3ujwLEXQNw +SaSS6sUUiHCm0w2wqsosQJz76YJumgIwK0eaB8bRwoF8yguWGEEbo/QwCZ61IygN +nxS2PFOiTAZpffpskcYqSUXm7LcT4Tps +-----END CERTIFICATE----- + +# Issuer: CN=Autoridad de Certificacion Firmaprofesional CIF A62634068 +# Subject: CN=Autoridad de Certificacion Firmaprofesional CIF A62634068 +# Label: "Autoridad de Certificacion Firmaprofesional CIF A62634068" +# Serial: 1977337328857672817 +# MD5 Fingerprint: 4e:6e:9b:54:4c:ca:b7:fa:48:e4:90:b1:15:4b:1c:a3 +# SHA1 Fingerprint: 0b:be:c2:27:22:49:cb:39:aa:db:35:5c:53:e3:8c:ae:78:ff:b6:fe +# SHA256 Fingerprint: 57:de:05:83:ef:d2:b2:6e:03:61:da:99:da:9d:f4:64:8d:ef:7e:e8:44:1c:3b:72:8a:fa:9b:cd:e0:f9:b2:6a +-----BEGIN CERTIFICATE----- +MIIGFDCCA/ygAwIBAgIIG3Dp0v+ubHEwDQYJKoZIhvcNAQELBQAwUTELMAkGA1UE +BhMCRVMxQjBABgNVBAMMOUF1dG9yaWRhZCBkZSBDZXJ0aWZpY2FjaW9uIEZpcm1h +cHJvZmVzaW9uYWwgQ0lGIEE2MjYzNDA2ODAeFw0xNDA5MjMxNTIyMDdaFw0zNjA1 +MDUxNTIyMDdaMFExCzAJBgNVBAYTAkVTMUIwQAYDVQQDDDlBdXRvcmlkYWQgZGUg +Q2VydGlmaWNhY2lvbiBGaXJtYXByb2Zlc2lvbmFsIENJRiBBNjI2MzQwNjgwggIi +MA0GCSqGSIb3DQEBAQUAA4ICDwAwggIKAoICAQDKlmuO6vj78aI14H9M2uDDUtd9 +thDIAl6zQyrET2qyyhxdKJp4ERppWVevtSBC5IsP5t9bpgOSL/UR5GLXMnE42QQM +cas9UX4PB99jBVzpv5RvwSmCwLTaUbDBPLutN0pcyvFLNg4kq7/DhHf9qFD0sefG +L9ItWY16Ck6WaVICqjaY7Pz6FIMMNx/Jkjd/14Et5cS54D40/mf0PmbR0/RAz15i +NA9wBj4gGFrO93IbJWyTdBSTo3OxDqqHECNZXyAFGUftaI6SEspd/NYrspI8IM/h +X68gvqB2f3bl7BqGYTM+53u0P6APjqK5am+5hyZvQWyIplD9amML9ZMWGxmPsu2b +m8mQ9QEM3xk9Dz44I8kvjwzRAv4bVdZO0I08r0+k8/6vKtMFnXkIoctXMbScyJCy +Z/QYFpM6/EfY0XiWMR+6KwxfXZmtY4laJCB22N/9q06mIqqdXuYnin1oKaPnirja +EbsXLZmdEyRG98Xi2J+Of8ePdG1asuhy9azuJBCtLxTa/y2aRnFHvkLfuwHb9H/T +KI8xWVvTyQKmtFLKbpf7Q8UIJm+K9Lv9nyiqDdVF8xM6HdjAeI9BZzwelGSuewvF +6NkBiDkal4ZkQdU7hwxu+g/GvUgUvzlN1J5Bto+WHWOWk9mVBngxaJ43BjuAiUVh +OSPHG0SjFeUc+JIwuwIDAQABo4HvMIHsMB0GA1UdDgQWBBRlzeurNR4APn7VdMAc +tHNHDhpkLzASBgNVHRMBAf8ECDAGAQH/AgEBMIGmBgNVHSAEgZ4wgZswgZgGBFUd +IAAwgY8wLwYIKwYBBQUHAgEWI2h0dHA6Ly93d3cuZmlybWFwcm9mZXNpb25hbC5j +b20vY3BzMFwGCCsGAQUFBwICMFAeTgBQAGEAcwBlAG8AIABkAGUAIABsAGEAIABC +AG8AbgBhAG4AbwB2AGEAIAA0ADcAIABCAGEAcgBjAGUAbABvAG4AYQAgADAAOAAw +ADEANzAOBgNVHQ8BAf8EBAMCAQYwDQYJKoZIhvcNAQELBQADggIBAHSHKAIrdx9m +iWTtj3QuRhy7qPj4Cx2Dtjqn6EWKB7fgPiDL4QjbEwj4KKE1soCzC1HA01aajTNF +Sa9J8OA9B3pFE1r/yJfY0xgsfZb43aJlQ3CTkBW6kN/oGbDbLIpgD7dvlAceHabJ +hfa9NPhAeGIQcDq+fUs5gakQ1JZBu/hfHAsdCPKxsIl68veg4MSPi3i1O1ilI45P +Vf42O+AMt8oqMEEgtIDNrvx2ZnOorm7hfNoD6JQg5iKj0B+QXSBTFCZX2lSX3xZE +EAEeiGaPcjiT3SC3NL7X8e5jjkd5KAb881lFJWAiMxujX6i6KtoaPc1A6ozuBRWV +1aUsIC+nmCjuRfzxuIgALI9C2lHVnOUTaHFFQ4ueCyE8S1wF3BqfmI7avSKecs2t +CsvMo2ebKHTEm9caPARYpoKdrcd7b/+Alun4jWq9GJAd/0kakFI3ky88Al2CdgtR +5xbHV/g4+afNmyJU72OwFW1TZQNKXkqgsqeOSQBZONXH9IBk9W6VULgRfhVwOEqw +f9DEMnDAGf/JOC0ULGb0QkTmVXYbgBVX/8Cnp6o5qtjTcNAuuuuUavpfNIbnYrX9 +ivAwhZTJryQCL2/W3Wf+47BVTwSYT6RBVuKT0Gro1vP7ZeDOdcQxWQzugsgMYDNK +GbqEZycPvEJdvSRUDewdcAZfpLz6IHxV +-----END CERTIFICATE----- + +# Issuer: CN=vTrus ECC Root CA O=iTrusChina Co.,Ltd. +# Subject: CN=vTrus ECC Root CA O=iTrusChina Co.,Ltd. +# Label: "vTrus ECC Root CA" +# Serial: 630369271402956006249506845124680065938238527194 +# MD5 Fingerprint: de:4b:c1:f5:52:8c:9b:43:e1:3e:8f:55:54:17:8d:85 +# SHA1 Fingerprint: f6:9c:db:b0:fc:f6:02:13:b6:52:32:a6:a3:91:3f:16:70:da:c3:e1 +# SHA256 Fingerprint: 30:fb:ba:2c:32:23:8e:2a:98:54:7a:f9:79:31:e5:50:42:8b:9b:3f:1c:8e:eb:66:33:dc:fa:86:c5:b2:7d:d3 +-----BEGIN CERTIFICATE----- +MIICDzCCAZWgAwIBAgIUbmq8WapTvpg5Z6LSa6Q75m0c1towCgYIKoZIzj0EAwMw +RzELMAkGA1UEBhMCQ04xHDAaBgNVBAoTE2lUcnVzQ2hpbmEgQ28uLEx0ZC4xGjAY +BgNVBAMTEXZUcnVzIEVDQyBSb290IENBMB4XDTE4MDczMTA3MjY0NFoXDTQzMDcz +MTA3MjY0NFowRzELMAkGA1UEBhMCQ04xHDAaBgNVBAoTE2lUcnVzQ2hpbmEgQ28u +LEx0ZC4xGjAYBgNVBAMTEXZUcnVzIEVDQyBSb290IENBMHYwEAYHKoZIzj0CAQYF +K4EEACIDYgAEZVBKrox5lkqqHAjDo6LN/llWQXf9JpRCux3NCNtzslt188+cToL0 +v/hhJoVs1oVbcnDS/dtitN9Ti72xRFhiQgnH+n9bEOf+QP3A2MMrMudwpremIFUd +e4BdS49nTPEQo0IwQDAdBgNVHQ4EFgQUmDnNvtiyjPeyq+GtJK97fKHbH88wDwYD +VR0TAQH/BAUwAwEB/zAOBgNVHQ8BAf8EBAMCAQYwCgYIKoZIzj0EAwMDaAAwZQIw +V53dVvHH4+m4SVBrm2nDb+zDfSXkV5UTQJtS0zvzQBm8JsctBp61ezaf9SXUY2sA +AjEA6dPGnlaaKsyh2j/IZivTWJwghfqrkYpwcBE4YGQLYgmRWAD5Tfs0aNoJrSEG +GJTO +-----END CERTIFICATE----- + +# Issuer: CN=vTrus Root CA O=iTrusChina Co.,Ltd. +# Subject: CN=vTrus Root CA O=iTrusChina Co.,Ltd. +# Label: "vTrus Root CA" +# Serial: 387574501246983434957692974888460947164905180485 +# MD5 Fingerprint: b8:c9:37:df:fa:6b:31:84:64:c5:ea:11:6a:1b:75:fc +# SHA1 Fingerprint: 84:1a:69:fb:f5:cd:1a:25:34:13:3d:e3:f8:fc:b8:99:d0:c9:14:b7 +# SHA256 Fingerprint: 8a:71:de:65:59:33:6f:42:6c:26:e5:38:80:d0:0d:88:a1:8d:a4:c6:a9:1f:0d:cb:61:94:e2:06:c5:c9:63:87 +-----BEGIN CERTIFICATE----- +MIIFVjCCAz6gAwIBAgIUQ+NxE9izWRRdt86M/TX9b7wFjUUwDQYJKoZIhvcNAQEL +BQAwQzELMAkGA1UEBhMCQ04xHDAaBgNVBAoTE2lUcnVzQ2hpbmEgQ28uLEx0ZC4x +FjAUBgNVBAMTDXZUcnVzIFJvb3QgQ0EwHhcNMTgwNzMxMDcyNDA1WhcNNDMwNzMx +MDcyNDA1WjBDMQswCQYDVQQGEwJDTjEcMBoGA1UEChMTaVRydXNDaGluYSBDby4s +THRkLjEWMBQGA1UEAxMNdlRydXMgUm9vdCBDQTCCAiIwDQYJKoZIhvcNAQEBBQAD +ggIPADCCAgoCggIBAL1VfGHTuB0EYgWgrmy3cLRB6ksDXhA/kFocizuwZotsSKYc +IrrVQJLuM7IjWcmOvFjai57QGfIvWcaMY1q6n6MLsLOaXLoRuBLpDLvPbmyAhykU +AyyNJJrIZIO1aqwTLDPxn9wsYTwaP3BVm60AUn/PBLn+NvqcwBauYv6WTEN+VRS+ +GrPSbcKvdmaVayqwlHeFXgQPYh1jdfdr58tbmnDsPmcF8P4HCIDPKNsFxhQnL4Z9 +8Cfe/+Z+M0jnCx5Y0ScrUw5XSmXX+6KAYPxMvDVTAWqXcoKv8R1w6Jz1717CbMdH +flqUhSZNO7rrTOiwCcJlwp2dCZtOtZcFrPUGoPc2BX70kLJrxLT5ZOrpGgrIDajt +J8nU57O5q4IikCc9Kuh8kO+8T/3iCiSn3mUkpF3qwHYw03dQ+A0Em5Q2AXPKBlim +0zvc+gRGE1WKyURHuFE5Gi7oNOJ5y1lKCn+8pu8fA2dqWSslYpPZUxlmPCdiKYZN +pGvu/9ROutW04o5IWgAZCfEF2c6Rsffr6TlP9m8EQ5pV9T4FFL2/s1m02I4zhKOQ +UqqzApVg+QxMaPnu1RcN+HFXtSXkKe5lXa/R7jwXC1pDxaWG6iSe4gUH3DRCEpHW +OXSuTEGC2/KmSNGzm/MzqvOmwMVO9fSddmPmAsYiS8GVP1BkLFTltvA8Kc9XAgMB +AAGjQjBAMB0GA1UdDgQWBBRUYnBj8XWEQ1iO0RYgscasGrz2iTAPBgNVHRMBAf8E +BTADAQH/MA4GA1UdDwEB/wQEAwIBBjANBgkqhkiG9w0BAQsFAAOCAgEAKbqSSaet +8PFww+SX8J+pJdVrnjT+5hpk9jprUrIQeBqfTNqK2uwcN1LgQkv7bHbKJAs5EhWd +nxEt/Hlk3ODg9d3gV8mlsnZwUKT+twpw1aA08XXXTUm6EdGz2OyC/+sOxL9kLX1j +bhd47F18iMjrjld22VkE+rxSH0Ws8HqA7Oxvdq6R2xCOBNyS36D25q5J08FsEhvM +Kar5CKXiNxTKsbhm7xqC5PD48acWabfbqWE8n/Uxy+QARsIvdLGx14HuqCaVvIiv +TDUHKgLKeBRtRytAVunLKmChZwOgzoy8sHJnxDHO2zTlJQNgJXtxmOTAGytfdELS +S8VZCAeHvsXDf+eW2eHcKJfWjwXj9ZtOyh1QRwVTsMo554WgicEFOwE30z9J4nfr +I8iIZjs9OXYhRvHsXyO466JmdXTBQPfYaJqT4i2pLr0cox7IdMakLXogqzu4sEb9 +b91fUlV1YvCXoHzXOP0l382gmxDPi7g4Xl7FtKYCNqEeXxzP4padKar9mK5S4fNB +UvupLnKWnyfjqnN9+BojZns7q2WwMgFLFT49ok8MKzWixtlnEjUwzXYuFrOZnk1P +Ti07NEPhmg4NpGaXutIcSkwsKouLgU9xGqndXHt7CMUADTdA43x7VF8vhV929ven +sBxXVsFy6K2ir40zSbofitzmdHxghm+Hl3s= +-----END CERTIFICATE----- + +# Issuer: CN=ISRG Root X2 O=Internet Security Research Group +# Subject: CN=ISRG Root X2 O=Internet Security Research Group +# Label: "ISRG Root X2" +# Serial: 87493402998870891108772069816698636114 +# MD5 Fingerprint: d3:9e:c4:1e:23:3c:a6:df:cf:a3:7e:6d:e0:14:e6:e5 +# SHA1 Fingerprint: bd:b1:b9:3c:d5:97:8d:45:c6:26:14:55:f8:db:95:c7:5a:d1:53:af +# SHA256 Fingerprint: 69:72:9b:8e:15:a8:6e:fc:17:7a:57:af:b7:17:1d:fc:64:ad:d2:8c:2f:ca:8c:f1:50:7e:34:45:3c:cb:14:70 +-----BEGIN CERTIFICATE----- +MIICGzCCAaGgAwIBAgIQQdKd0XLq7qeAwSxs6S+HUjAKBggqhkjOPQQDAzBPMQsw +CQYDVQQGEwJVUzEpMCcGA1UEChMgSW50ZXJuZXQgU2VjdXJpdHkgUmVzZWFyY2gg +R3JvdXAxFTATBgNVBAMTDElTUkcgUm9vdCBYMjAeFw0yMDA5MDQwMDAwMDBaFw00 +MDA5MTcxNjAwMDBaME8xCzAJBgNVBAYTAlVTMSkwJwYDVQQKEyBJbnRlcm5ldCBT +ZWN1cml0eSBSZXNlYXJjaCBHcm91cDEVMBMGA1UEAxMMSVNSRyBSb290IFgyMHYw +EAYHKoZIzj0CAQYFK4EEACIDYgAEzZvVn4CDCuwJSvMWSj5cz3es3mcFDR0HttwW ++1qLFNvicWDEukWVEYmO6gbf9yoWHKS5xcUy4APgHoIYOIvXRdgKam7mAHf7AlF9 +ItgKbppbd9/w+kHsOdx1ymgHDB/qo0IwQDAOBgNVHQ8BAf8EBAMCAQYwDwYDVR0T +AQH/BAUwAwEB/zAdBgNVHQ4EFgQUfEKWrt5LSDv6kviejM9ti6lyN5UwCgYIKoZI +zj0EAwMDaAAwZQIwe3lORlCEwkSHRhtFcP9Ymd70/aTSVaYgLXTWNLxBo1BfASdW +tL4ndQavEi51mI38AjEAi/V3bNTIZargCyzuFJ0nN6T5U6VR5CmD1/iQMVtCnwr1 +/q4AaOeMSQ+2b1tbFfLn +-----END CERTIFICATE----- + +# Issuer: CN=HiPKI Root CA - G1 O=Chunghwa Telecom Co., Ltd. +# Subject: CN=HiPKI Root CA - G1 O=Chunghwa Telecom Co., Ltd. +# Label: "HiPKI Root CA - G1" +# Serial: 60966262342023497858655262305426234976 +# MD5 Fingerprint: 69:45:df:16:65:4b:e8:68:9a:8f:76:5f:ff:80:9e:d3 +# SHA1 Fingerprint: 6a:92:e4:a8:ee:1b:ec:96:45:37:e3:29:57:49:cd:96:e3:e5:d2:60 +# SHA256 Fingerprint: f0:15:ce:3c:c2:39:bf:ef:06:4b:e9:f1:d2:c4:17:e1:a0:26:4a:0a:94:be:1f:0c:8d:12:18:64:eb:69:49:cc +-----BEGIN CERTIFICATE----- +MIIFajCCA1KgAwIBAgIQLd2szmKXlKFD6LDNdmpeYDANBgkqhkiG9w0BAQsFADBP +MQswCQYDVQQGEwJUVzEjMCEGA1UECgwaQ2h1bmdod2EgVGVsZWNvbSBDby4sIEx0 +ZC4xGzAZBgNVBAMMEkhpUEtJIFJvb3QgQ0EgLSBHMTAeFw0xOTAyMjIwOTQ2MDRa +Fw0zNzEyMzExNTU5NTlaME8xCzAJBgNVBAYTAlRXMSMwIQYDVQQKDBpDaHVuZ2h3 +YSBUZWxlY29tIENvLiwgTHRkLjEbMBkGA1UEAwwSSGlQS0kgUm9vdCBDQSAtIEcx +MIICIjANBgkqhkiG9w0BAQEFAAOCAg8AMIICCgKCAgEA9B5/UnMyDHPkvRN0o9Qw +qNCuS9i233VHZvR85zkEHmpwINJaR3JnVfSl6J3VHiGh8Ge6zCFovkRTv4354twv +Vcg3Px+kwJyz5HdcoEb+d/oaoDjq7Zpy3iu9lFc6uux55199QmQ5eiY29yTw1S+6 +lZgRZq2XNdZ1AYDgr/SEYYwNHl98h5ZeQa/rh+r4XfEuiAU+TCK72h8q3VJGZDnz +Qs7ZngyzsHeXZJzA9KMuH5UHsBffMNsAGJZMoYFL3QRtU6M9/Aes1MU3guvklQgZ +KILSQjqj2FPseYlgSGDIcpJQ3AOPgz+yQlda22rpEZfdhSi8MEyr48KxRURHH+CK +FgeW0iEPU8DtqX7UTuybCeyvQqww1r/REEXgphaypcXTT3OUM3ECoWqj1jOXTyFj +HluP2cFeRXF3D4FdXyGarYPM+l7WjSNfGz1BryB1ZlpK9p/7qxj3ccC2HTHsOyDr +y+K49a6SsvfhhEvyovKTmiKe0xRvNlS9H15ZFblzqMF8b3ti6RZsR1pl8w4Rm0bZ +/W3c1pzAtH2lsN0/Vm+h+fbkEkj9Bn8SV7apI09bA8PgcSojt/ewsTu8mL3WmKgM +a/aOEmem8rJY5AIJEzypuxC00jBF8ez3ABHfZfjcK0NVvxaXxA/VLGGEqnKG/uY6 +fsI/fe78LxQ+5oXdUG+3Se0CAwEAAaNCMEAwDwYDVR0TAQH/BAUwAwEB/zAdBgNV +HQ4EFgQU8ncX+l6o/vY9cdVouslGDDjYr7AwDgYDVR0PAQH/BAQDAgGGMA0GCSqG +SIb3DQEBCwUAA4ICAQBQUfB13HAE4/+qddRxosuej6ip0691x1TPOhwEmSKsxBHi +7zNKpiMdDg1H2DfHb680f0+BazVP6XKlMeJ45/dOlBhbQH3PayFUhuaVevvGyuqc +SE5XCV0vrPSltJczWNWseanMX/mF+lLFjfiRFOs6DRfQUsJ748JzjkZ4Bjgs6Fza +ZsT0pPBWGTMpWmWSBUdGSquEwx4noR8RkpkndZMPvDY7l1ePJlsMu5wP1G4wB9Tc +XzZoZjmDlicmisjEOf6aIW/Vcobpf2Lll07QJNBAsNB1CI69aO4I1258EHBGG3zg +iLKecoaZAeO/n0kZtCW+VmWuF2PlHt/o/0elv+EmBYTksMCv5wiZqAxeJoBF1Pho +L5aPruJKHJwWDBNvOIf2u8g0X5IDUXlwpt/L9ZlNec1OvFefQ05rLisY+GpzjLrF +Ne85akEez3GoorKGB1s6yeHvP2UEgEcyRHCVTjFnanRbEEV16rCf0OY1/k6fi8wr +kkVbbiVghUbN0aqwdmaTd5a+g744tiROJgvM7XpWGuDpWsZkrUx6AEhEL7lAuxM+ +vhV4nYWBSipX3tUZQ9rbyltHhoMLP7YNdnhzeSJesYAfz77RP1YQmCuVh6EfnWQU +YDksswBVLuT1sw5XxJFBAJw/6KXf6vb/yPCtbVKoF6ubYfwSUTXkJf2vqmqGOQ== +-----END CERTIFICATE----- + +# Issuer: CN=GlobalSign O=GlobalSign OU=GlobalSign ECC Root CA - R4 +# Subject: CN=GlobalSign O=GlobalSign OU=GlobalSign ECC Root CA - R4 +# Label: "GlobalSign ECC Root CA - R4" +# Serial: 159662223612894884239637590694 +# MD5 Fingerprint: 26:29:f8:6d:e1:88:bf:a2:65:7f:aa:c4:cd:0f:7f:fc +# SHA1 Fingerprint: 6b:a0:b0:98:e1:71:ef:5a:ad:fe:48:15:80:77:10:f4:bd:6f:0b:28 +# SHA256 Fingerprint: b0:85:d7:0b:96:4f:19:1a:73:e4:af:0d:54:ae:7a:0e:07:aa:fd:af:9b:71:dd:08:62:13:8a:b7:32:5a:24:a2 +-----BEGIN CERTIFICATE----- +MIIB3DCCAYOgAwIBAgINAgPlfvU/k/2lCSGypjAKBggqhkjOPQQDAjBQMSQwIgYD +VQQLExtHbG9iYWxTaWduIEVDQyBSb290IENBIC0gUjQxEzARBgNVBAoTCkdsb2Jh +bFNpZ24xEzARBgNVBAMTCkdsb2JhbFNpZ24wHhcNMTIxMTEzMDAwMDAwWhcNMzgw +MTE5MDMxNDA3WjBQMSQwIgYDVQQLExtHbG9iYWxTaWduIEVDQyBSb290IENBIC0g +UjQxEzARBgNVBAoTCkdsb2JhbFNpZ24xEzARBgNVBAMTCkdsb2JhbFNpZ24wWTAT +BgcqhkjOPQIBBggqhkjOPQMBBwNCAAS4xnnTj2wlDp8uORkcA6SumuU5BwkWymOx +uYb4ilfBV85C+nOh92VC/x7BALJucw7/xyHlGKSq2XE/qNS5zowdo0IwQDAOBgNV +HQ8BAf8EBAMCAYYwDwYDVR0TAQH/BAUwAwEB/zAdBgNVHQ4EFgQUVLB7rUW44kB/ ++wpu+74zyTyjhNUwCgYIKoZIzj0EAwIDRwAwRAIgIk90crlgr/HmnKAWBVBfw147 +bmF0774BxL4YSFlhgjICICadVGNA3jdgUM/I2O2dgq43mLyjj0xMqTQrbO/7lZsm +-----END CERTIFICATE----- + +# Issuer: CN=GTS Root R1 O=Google Trust Services LLC +# Subject: CN=GTS Root R1 O=Google Trust Services LLC +# Label: "GTS Root R1" +# Serial: 159662320309726417404178440727 +# MD5 Fingerprint: 05:fe:d0:bf:71:a8:a3:76:63:da:01:e0:d8:52:dc:40 +# SHA1 Fingerprint: e5:8c:1c:c4:91:3b:38:63:4b:e9:10:6e:e3:ad:8e:6b:9d:d9:81:4a +# SHA256 Fingerprint: d9:47:43:2a:bd:e7:b7:fa:90:fc:2e:6b:59:10:1b:12:80:e0:e1:c7:e4:e4:0f:a3:c6:88:7f:ff:57:a7:f4:cf +-----BEGIN CERTIFICATE----- +MIIFVzCCAz+gAwIBAgINAgPlk28xsBNJiGuiFzANBgkqhkiG9w0BAQwFADBHMQsw +CQYDVQQGEwJVUzEiMCAGA1UEChMZR29vZ2xlIFRydXN0IFNlcnZpY2VzIExMQzEU +MBIGA1UEAxMLR1RTIFJvb3QgUjEwHhcNMTYwNjIyMDAwMDAwWhcNMzYwNjIyMDAw +MDAwWjBHMQswCQYDVQQGEwJVUzEiMCAGA1UEChMZR29vZ2xlIFRydXN0IFNlcnZp +Y2VzIExMQzEUMBIGA1UEAxMLR1RTIFJvb3QgUjEwggIiMA0GCSqGSIb3DQEBAQUA +A4ICDwAwggIKAoICAQC2EQKLHuOhd5s73L+UPreVp0A8of2C+X0yBoJx9vaMf/vo +27xqLpeXo4xL+Sv2sfnOhB2x+cWX3u+58qPpvBKJXqeqUqv4IyfLpLGcY9vXmX7w +Cl7raKb0xlpHDU0QM+NOsROjyBhsS+z8CZDfnWQpJSMHobTSPS5g4M/SCYe7zUjw +TcLCeoiKu7rPWRnWr4+wB7CeMfGCwcDfLqZtbBkOtdh+JhpFAz2weaSUKK0Pfybl +qAj+lug8aJRT7oM6iCsVlgmy4HqMLnXWnOunVmSPlk9orj2XwoSPwLxAwAtcvfaH +szVsrBhQf4TgTM2S0yDpM7xSma8ytSmzJSq0SPly4cpk9+aCEI3oncKKiPo4Zor8 +Y/kB+Xj9e1x3+naH+uzfsQ55lVe0vSbv1gHR6xYKu44LtcXFilWr06zqkUspzBmk +MiVOKvFlRNACzqrOSbTqn3yDsEB750Orp2yjj32JgfpMpf/VjsPOS+C12LOORc92 +wO1AK/1TD7Cn1TsNsYqiA94xrcx36m97PtbfkSIS5r762DL8EGMUUXLeXdYWk70p +aDPvOmbsB4om3xPXV2V4J95eSRQAogB/mqghtqmxlbCluQ0WEdrHbEg8QOB+DVrN +VjzRlwW5y0vtOUucxD/SVRNuJLDWcfr0wbrM7Rv1/oFB2ACYPTrIrnqYNxgFlQID +AQABo0IwQDAOBgNVHQ8BAf8EBAMCAYYwDwYDVR0TAQH/BAUwAwEB/zAdBgNVHQ4E +FgQU5K8rJnEaK0gnhS9SZizv8IkTcT4wDQYJKoZIhvcNAQEMBQADggIBAJ+qQibb +C5u+/x6Wki4+omVKapi6Ist9wTrYggoGxval3sBOh2Z5ofmmWJyq+bXmYOfg6LEe +QkEzCzc9zolwFcq1JKjPa7XSQCGYzyI0zzvFIoTgxQ6KfF2I5DUkzps+GlQebtuy +h6f88/qBVRRiClmpIgUxPoLW7ttXNLwzldMXG+gnoot7TiYaelpkttGsN/H9oPM4 +7HLwEXWdyzRSjeZ2axfG34arJ45JK3VmgRAhpuo+9K4l/3wV3s6MJT/KYnAK9y8J +ZgfIPxz88NtFMN9iiMG1D53Dn0reWVlHxYciNuaCp+0KueIHoI17eko8cdLiA6Ef +MgfdG+RCzgwARWGAtQsgWSl4vflVy2PFPEz0tv/bal8xa5meLMFrUKTX5hgUvYU/ +Z6tGn6D/Qqc6f1zLXbBwHSs09dR2CQzreExZBfMzQsNhFRAbd03OIozUhfJFfbdT +6u9AWpQKXCBfTkBdYiJ23//OYb2MI3jSNwLgjt7RETeJ9r/tSQdirpLsQBqvFAnZ +0E6yove+7u7Y/9waLd64NnHi/Hm3lCXRSHNboTXns5lndcEZOitHTtNCjv0xyBZm +2tIMPNuzjsmhDYAPexZ3FL//2wmUspO8IFgV6dtxQ/PeEMMA3KgqlbbC1j+Qa3bb +bP6MvPJwNQzcmRk13NfIRmPVNnGuV/u3gm3c +-----END CERTIFICATE----- + +# Issuer: CN=GTS Root R2 O=Google Trust Services LLC +# Subject: CN=GTS Root R2 O=Google Trust Services LLC +# Label: "GTS Root R2" +# Serial: 159662449406622349769042896298 +# MD5 Fingerprint: 1e:39:c0:53:e6:1e:29:82:0b:ca:52:55:36:5d:57:dc +# SHA1 Fingerprint: 9a:44:49:76:32:db:de:fa:d0:bc:fb:5a:7b:17:bd:9e:56:09:24:94 +# SHA256 Fingerprint: 8d:25:cd:97:22:9d:bf:70:35:6b:da:4e:b3:cc:73:40:31:e2:4c:f0:0f:af:cf:d3:2d:c7:6e:b5:84:1c:7e:a8 +-----BEGIN CERTIFICATE----- +MIIFVzCCAz+gAwIBAgINAgPlrsWNBCUaqxElqjANBgkqhkiG9w0BAQwFADBHMQsw +CQYDVQQGEwJVUzEiMCAGA1UEChMZR29vZ2xlIFRydXN0IFNlcnZpY2VzIExMQzEU +MBIGA1UEAxMLR1RTIFJvb3QgUjIwHhcNMTYwNjIyMDAwMDAwWhcNMzYwNjIyMDAw +MDAwWjBHMQswCQYDVQQGEwJVUzEiMCAGA1UEChMZR29vZ2xlIFRydXN0IFNlcnZp +Y2VzIExMQzEUMBIGA1UEAxMLR1RTIFJvb3QgUjIwggIiMA0GCSqGSIb3DQEBAQUA +A4ICDwAwggIKAoICAQDO3v2m++zsFDQ8BwZabFn3GTXd98GdVarTzTukk3LvCvpt +nfbwhYBboUhSnznFt+4orO/LdmgUud+tAWyZH8QiHZ/+cnfgLFuv5AS/T3KgGjSY +6Dlo7JUle3ah5mm5hRm9iYz+re026nO8/4Piy33B0s5Ks40FnotJk9/BW9BuXvAu +MC6C/Pq8tBcKSOWIm8Wba96wyrQD8Nr0kLhlZPdcTK3ofmZemde4wj7I0BOdre7k +RXuJVfeKH2JShBKzwkCX44ofR5GmdFrS+LFjKBC4swm4VndAoiaYecb+3yXuPuWg +f9RhD1FLPD+M2uFwdNjCaKH5wQzpoeJ/u1U8dgbuak7MkogwTZq9TwtImoS1mKPV ++3PBV2HdKFZ1E66HjucMUQkQdYhMvI35ezzUIkgfKtzra7tEscszcTJGr61K8Yzo +dDqs5xoic4DSMPclQsciOzsSrZYuxsN2B6ogtzVJV+mSSeh2FnIxZyuWfoqjx5RW +Ir9qS34BIbIjMt/kmkRtWVtd9QCgHJvGeJeNkP+byKq0rxFROV7Z+2et1VsRnTKa +G73VululycslaVNVJ1zgyjbLiGH7HrfQy+4W+9OmTN6SpdTi3/UGVN4unUu0kzCq +gc7dGtxRcw1PcOnlthYhGXmy5okLdWTK1au8CcEYof/UVKGFPP0UJAOyh9OktwID +AQABo0IwQDAOBgNVHQ8BAf8EBAMCAYYwDwYDVR0TAQH/BAUwAwEB/zAdBgNVHQ4E +FgQUu//KjiOfT5nK2+JopqUVJxce2Q4wDQYJKoZIhvcNAQEMBQADggIBAB/Kzt3H +vqGf2SdMC9wXmBFqiN495nFWcrKeGk6c1SuYJF2ba3uwM4IJvd8lRuqYnrYb/oM8 +0mJhwQTtzuDFycgTE1XnqGOtjHsB/ncw4c5omwX4Eu55MaBBRTUoCnGkJE+M3DyC +B19m3H0Q/gxhswWV7uGugQ+o+MePTagjAiZrHYNSVc61LwDKgEDg4XSsYPWHgJ2u +NmSRXbBoGOqKYcl3qJfEycel/FVL8/B/uWU9J2jQzGv6U53hkRrJXRqWbTKH7QMg +yALOWr7Z6v2yTcQvG99fevX4i8buMTolUVVnjWQye+mew4K6Ki3pHrTgSAai/Gev +HyICc/sgCq+dVEuhzf9gR7A/Xe8bVr2XIZYtCtFenTgCR2y59PYjJbigapordwj6 +xLEokCZYCDzifqrXPW+6MYgKBesntaFJ7qBFVHvmJ2WZICGoo7z7GJa7Um8M7YNR +TOlZ4iBgxcJlkoKM8xAfDoqXvneCbT+PHV28SSe9zE8P4c52hgQjxcCMElv924Sg +JPFI/2R80L5cFtHvma3AH/vLrrw4IgYmZNralw4/KBVEqE8AyvCazM90arQ+POuV +7LXTWtiBmelDGDfrs7vRWGJB82bSj6p4lVQgw1oudCvV0b4YacCs1aTPObpRhANl +6WLAYv7YTVWW4tAR+kg0Eeye7QUd5MjWHYbL +-----END CERTIFICATE----- + +# Issuer: CN=GTS Root R3 O=Google Trust Services LLC +# Subject: CN=GTS Root R3 O=Google Trust Services LLC +# Label: "GTS Root R3" +# Serial: 159662495401136852707857743206 +# MD5 Fingerprint: 3e:e7:9d:58:02:94:46:51:94:e5:e0:22:4a:8b:e7:73 +# SHA1 Fingerprint: ed:e5:71:80:2b:c8:92:b9:5b:83:3c:d2:32:68:3f:09:cd:a0:1e:46 +# SHA256 Fingerprint: 34:d8:a7:3e:e2:08:d9:bc:db:0d:95:65:20:93:4b:4e:40:e6:94:82:59:6e:8b:6f:73:c8:42:6b:01:0a:6f:48 +-----BEGIN CERTIFICATE----- +MIICCTCCAY6gAwIBAgINAgPluILrIPglJ209ZjAKBggqhkjOPQQDAzBHMQswCQYD +VQQGEwJVUzEiMCAGA1UEChMZR29vZ2xlIFRydXN0IFNlcnZpY2VzIExMQzEUMBIG +A1UEAxMLR1RTIFJvb3QgUjMwHhcNMTYwNjIyMDAwMDAwWhcNMzYwNjIyMDAwMDAw +WjBHMQswCQYDVQQGEwJVUzEiMCAGA1UEChMZR29vZ2xlIFRydXN0IFNlcnZpY2Vz +IExMQzEUMBIGA1UEAxMLR1RTIFJvb3QgUjMwdjAQBgcqhkjOPQIBBgUrgQQAIgNi +AAQfTzOHMymKoYTey8chWEGJ6ladK0uFxh1MJ7x/JlFyb+Kf1qPKzEUURout736G +jOyxfi//qXGdGIRFBEFVbivqJn+7kAHjSxm65FSWRQmx1WyRRK2EE46ajA2ADDL2 +4CejQjBAMA4GA1UdDwEB/wQEAwIBhjAPBgNVHRMBAf8EBTADAQH/MB0GA1UdDgQW +BBTB8Sa6oC2uhYHP0/EqEr24Cmf9vDAKBggqhkjOPQQDAwNpADBmAjEA9uEglRR7 +VKOQFhG/hMjqb2sXnh5GmCCbn9MN2azTL818+FsuVbu/3ZL3pAzcMeGiAjEA/Jdm +ZuVDFhOD3cffL74UOO0BzrEXGhF16b0DjyZ+hOXJYKaV11RZt+cRLInUue4X +-----END CERTIFICATE----- + +# Issuer: CN=GTS Root R4 O=Google Trust Services LLC +# Subject: CN=GTS Root R4 O=Google Trust Services LLC +# Label: "GTS Root R4" +# Serial: 159662532700760215368942768210 +# MD5 Fingerprint: 43:96:83:77:19:4d:76:b3:9d:65:52:e4:1d:22:a5:e8 +# SHA1 Fingerprint: 77:d3:03:67:b5:e0:0c:15:f6:0c:38:61:df:7c:e1:3b:92:46:4d:47 +# SHA256 Fingerprint: 34:9d:fa:40:58:c5:e2:63:12:3b:39:8a:e7:95:57:3c:4e:13:13:c8:3f:e6:8f:93:55:6c:d5:e8:03:1b:3c:7d +-----BEGIN CERTIFICATE----- +MIICCTCCAY6gAwIBAgINAgPlwGjvYxqccpBQUjAKBggqhkjOPQQDAzBHMQswCQYD +VQQGEwJVUzEiMCAGA1UEChMZR29vZ2xlIFRydXN0IFNlcnZpY2VzIExMQzEUMBIG +A1UEAxMLR1RTIFJvb3QgUjQwHhcNMTYwNjIyMDAwMDAwWhcNMzYwNjIyMDAwMDAw +WjBHMQswCQYDVQQGEwJVUzEiMCAGA1UEChMZR29vZ2xlIFRydXN0IFNlcnZpY2Vz +IExMQzEUMBIGA1UEAxMLR1RTIFJvb3QgUjQwdjAQBgcqhkjOPQIBBgUrgQQAIgNi +AATzdHOnaItgrkO4NcWBMHtLSZ37wWHO5t5GvWvVYRg1rkDdc/eJkTBa6zzuhXyi +QHY7qca4R9gq55KRanPpsXI5nymfopjTX15YhmUPoYRlBtHci8nHc8iMai/lxKvR +HYqjQjBAMA4GA1UdDwEB/wQEAwIBhjAPBgNVHRMBAf8EBTADAQH/MB0GA1UdDgQW +BBSATNbrdP9JNqPV2Py1PsVq8JQdjDAKBggqhkjOPQQDAwNpADBmAjEA6ED/g94D +9J+uHXqnLrmvT/aDHQ4thQEd0dlq7A/Cr8deVl5c1RxYIigL9zC2L7F8AjEA8GE8 +p/SgguMh1YQdc4acLa/KNJvxn7kjNuK8YAOdgLOaVsjh4rsUecrNIdSUtUlD +-----END CERTIFICATE----- + +# Issuer: CN=Telia Root CA v2 O=Telia Finland Oyj +# Subject: CN=Telia Root CA v2 O=Telia Finland Oyj +# Label: "Telia Root CA v2" +# Serial: 7288924052977061235122729490515358 +# MD5 Fingerprint: 0e:8f:ac:aa:82:df:85:b1:f4:dc:10:1c:fc:99:d9:48 +# SHA1 Fingerprint: b9:99:cd:d1:73:50:8a:c4:47:05:08:9c:8c:88:fb:be:a0:2b:40:cd +# SHA256 Fingerprint: 24:2b:69:74:2f:cb:1e:5b:2a:bf:98:89:8b:94:57:21:87:54:4e:5b:4d:99:11:78:65:73:62:1f:6a:74:b8:2c +-----BEGIN CERTIFICATE----- +MIIFdDCCA1ygAwIBAgIPAWdfJ9b+euPkrL4JWwWeMA0GCSqGSIb3DQEBCwUAMEQx +CzAJBgNVBAYTAkZJMRowGAYDVQQKDBFUZWxpYSBGaW5sYW5kIE95ajEZMBcGA1UE +AwwQVGVsaWEgUm9vdCBDQSB2MjAeFw0xODExMjkxMTU1NTRaFw00MzExMjkxMTU1 +NTRaMEQxCzAJBgNVBAYTAkZJMRowGAYDVQQKDBFUZWxpYSBGaW5sYW5kIE95ajEZ +MBcGA1UEAwwQVGVsaWEgUm9vdCBDQSB2MjCCAiIwDQYJKoZIhvcNAQEBBQADggIP +ADCCAgoCggIBALLQPwe84nvQa5n44ndp586dpAO8gm2h/oFlH0wnrI4AuhZ76zBq +AMCzdGh+sq/H1WKzej9Qyow2RCRj0jbpDIX2Q3bVTKFgcmfiKDOlyzG4OiIjNLh9 +vVYiQJ3q9HsDrWj8soFPmNB06o3lfc1jw6P23pLCWBnglrvFxKk9pXSW/q/5iaq9 +lRdU2HhE8Qx3FZLgmEKnpNaqIJLNwaCzlrI6hEKNfdWV5Nbb6WLEWLN5xYzTNTOD +n3WhUidhOPFZPY5Q4L15POdslv5e2QJltI5c0BE0312/UqeBAMN/mUWZFdUXyApT +7GPzmX3MaRKGwhfwAZ6/hLzRUssbkmbOpFPlob/E2wnW5olWK8jjfN7j/4nlNW4o +6GwLI1GpJQXrSPjdscr6bAhR77cYbETKJuFzxokGgeWKrLDiKca5JLNrRBH0pUPC +TEPlcDaMtjNXepUugqD0XBCzYYP2AgWGLnwtbNwDRm41k9V6lS/eINhbfpSQBGq6 +WT0EBXWdN6IOLj3rwaRSg/7Qa9RmjtzG6RJOHSpXqhC8fF6CfaamyfItufUXJ63R +DolUK5X6wK0dmBR4M0KGCqlztft0DbcbMBnEWg4cJ7faGND/isgFuvGqHKI3t+ZI +pEYslOqodmJHixBTB0hXbOKSTbauBcvcwUpej6w9GU7C7WB1K9vBykLVAgMBAAGj +YzBhMB8GA1UdIwQYMBaAFHKs5DN5qkWH9v2sHZ7Wxy+G2CQ5MB0GA1UdDgQWBBRy +rOQzeapFh/b9rB2e1scvhtgkOTAOBgNVHQ8BAf8EBAMCAQYwDwYDVR0TAQH/BAUw +AwEB/zANBgkqhkiG9w0BAQsFAAOCAgEAoDtZpwmUPjaE0n4vOaWWl/oRrfxn83EJ +8rKJhGdEr7nv7ZbsnGTbMjBvZ5qsfl+yqwE2foH65IRe0qw24GtixX1LDoJt0nZi +0f6X+J8wfBj5tFJ3gh1229MdqfDBmgC9bXXYfef6xzijnHDoRnkDry5023X4blMM +A8iZGok1GTzTyVR8qPAs5m4HeW9q4ebqkYJpCh3DflminmtGFZhb069GHWLIzoBS +SRE/yQQSwxN8PzuKlts8oB4KtItUsiRnDe+Cy748fdHif64W1lZYudogsYMVoe+K +TTJvQS8TUoKU1xrBeKJR3Stwbbca+few4GeXVtt8YVMJAygCQMez2P2ccGrGKMOF +6eLtGpOg3kuYooQ+BXcBlj37tCAPnHICehIv1aO6UXivKitEZU61/Qrowc15h2Er +3oBXRb9n8ZuRXqWk7FlIEA04x7D6w0RtBPV4UBySllva9bguulvP5fBqnUsvWHMt +Ty3EHD70sz+rFQ47GUGKpMFXEmZxTPpT41frYpUJnlTd0cI8Vzy9OK2YZLe4A5pT +VmBds9hCG1xLEooc6+t9xnppxyd/pPiL8uSUZodL6ZQHCRJ5irLrdATczvREWeAW +ysUsWNc8e89ihmpQfTU2Zqf7N+cox9jQraVplI/owd8k+BsHMYeB2F326CjYSlKA +rBPuUBQemMc= +-----END CERTIFICATE----- + +# Issuer: CN=D-TRUST BR Root CA 1 2020 O=D-Trust GmbH +# Subject: CN=D-TRUST BR Root CA 1 2020 O=D-Trust GmbH +# Label: "D-TRUST BR Root CA 1 2020" +# Serial: 165870826978392376648679885835942448534 +# MD5 Fingerprint: b5:aa:4b:d5:ed:f7:e3:55:2e:8f:72:0a:f3:75:b8:ed +# SHA1 Fingerprint: 1f:5b:98:f0:e3:b5:f7:74:3c:ed:e6:b0:36:7d:32:cd:f4:09:41:67 +# SHA256 Fingerprint: e5:9a:aa:81:60:09:c2:2b:ff:5b:25:ba:d3:7d:f3:06:f0:49:79:7c:1f:81:d8:5a:b0:89:e6:57:bd:8f:00:44 +-----BEGIN CERTIFICATE----- +MIIC2zCCAmCgAwIBAgIQfMmPK4TX3+oPyWWa00tNljAKBggqhkjOPQQDAzBIMQsw +CQYDVQQGEwJERTEVMBMGA1UEChMMRC1UcnVzdCBHbWJIMSIwIAYDVQQDExlELVRS +VVNUIEJSIFJvb3QgQ0EgMSAyMDIwMB4XDTIwMDIxMTA5NDUwMFoXDTM1MDIxMTA5 +NDQ1OVowSDELMAkGA1UEBhMCREUxFTATBgNVBAoTDEQtVHJ1c3QgR21iSDEiMCAG +A1UEAxMZRC1UUlVTVCBCUiBSb290IENBIDEgMjAyMDB2MBAGByqGSM49AgEGBSuB +BAAiA2IABMbLxyjR+4T1mu9CFCDhQ2tuda38KwOE1HaTJddZO0Flax7mNCq7dPYS +zuht56vkPE4/RAiLzRZxy7+SmfSk1zxQVFKQhYN4lGdnoxwJGT11NIXe7WB9xwy0 +QVK5buXuQqOCAQ0wggEJMA8GA1UdEwEB/wQFMAMBAf8wHQYDVR0OBBYEFHOREKv/ +VbNafAkl1bK6CKBrqx9tMA4GA1UdDwEB/wQEAwIBBjCBxgYDVR0fBIG+MIG7MD6g +PKA6hjhodHRwOi8vY3JsLmQtdHJ1c3QubmV0L2NybC9kLXRydXN0X2JyX3Jvb3Rf +Y2FfMV8yMDIwLmNybDB5oHegdYZzbGRhcDovL2RpcmVjdG9yeS5kLXRydXN0Lm5l +dC9DTj1ELVRSVVNUJTIwQlIlMjBSb290JTIwQ0ElMjAxJTIwMjAyMCxPPUQtVHJ1 +c3QlMjBHbWJILEM9REU/Y2VydGlmaWNhdGVyZXZvY2F0aW9ubGlzdDAKBggqhkjO +PQQDAwNpADBmAjEAlJAtE/rhY/hhY+ithXhUkZy4kzg+GkHaQBZTQgjKL47xPoFW +wKrY7RjEsK70PvomAjEA8yjixtsrmfu3Ubgko6SUeho/5jbiA1czijDLgsfWFBHV +dWNbFJWcHwHP2NVypw87 +-----END CERTIFICATE----- + +# Issuer: CN=D-TRUST EV Root CA 1 2020 O=D-Trust GmbH +# Subject: CN=D-TRUST EV Root CA 1 2020 O=D-Trust GmbH +# Label: "D-TRUST EV Root CA 1 2020" +# Serial: 126288379621884218666039612629459926992 +# MD5 Fingerprint: 8c:2d:9d:70:9f:48:99:11:06:11:fb:e9:cb:30:c0:6e +# SHA1 Fingerprint: 61:db:8c:21:59:69:03:90:d8:7c:9c:12:86:54:cf:9d:3d:f4:dd:07 +# SHA256 Fingerprint: 08:17:0d:1a:a3:64:53:90:1a:2f:95:92:45:e3:47:db:0c:8d:37:ab:aa:bc:56:b8:1a:a1:00:dc:95:89:70:db +-----BEGIN CERTIFICATE----- +MIIC2zCCAmCgAwIBAgIQXwJB13qHfEwDo6yWjfv/0DAKBggqhkjOPQQDAzBIMQsw +CQYDVQQGEwJERTEVMBMGA1UEChMMRC1UcnVzdCBHbWJIMSIwIAYDVQQDExlELVRS +VVNUIEVWIFJvb3QgQ0EgMSAyMDIwMB4XDTIwMDIxMTEwMDAwMFoXDTM1MDIxMTA5 +NTk1OVowSDELMAkGA1UEBhMCREUxFTATBgNVBAoTDEQtVHJ1c3QgR21iSDEiMCAG +A1UEAxMZRC1UUlVTVCBFViBSb290IENBIDEgMjAyMDB2MBAGByqGSM49AgEGBSuB +BAAiA2IABPEL3YZDIBnfl4XoIkqbz52Yv7QFJsnL46bSj8WeeHsxiamJrSc8ZRCC +/N/DnU7wMyPE0jL1HLDfMxddxfCxivnvubcUyilKwg+pf3VlSSowZ/Rk99Yad9rD +wpdhQntJraOCAQ0wggEJMA8GA1UdEwEB/wQFMAMBAf8wHQYDVR0OBBYEFH8QARY3 +OqQo5FD4pPfsazK2/umLMA4GA1UdDwEB/wQEAwIBBjCBxgYDVR0fBIG+MIG7MD6g +PKA6hjhodHRwOi8vY3JsLmQtdHJ1c3QubmV0L2NybC9kLXRydXN0X2V2X3Jvb3Rf +Y2FfMV8yMDIwLmNybDB5oHegdYZzbGRhcDovL2RpcmVjdG9yeS5kLXRydXN0Lm5l +dC9DTj1ELVRSVVNUJTIwRVYlMjBSb290JTIwQ0ElMjAxJTIwMjAyMCxPPUQtVHJ1 +c3QlMjBHbWJILEM9REU/Y2VydGlmaWNhdGVyZXZvY2F0aW9ubGlzdDAKBggqhkjO +PQQDAwNpADBmAjEAyjzGKnXCXnViOTYAYFqLwZOZzNnbQTs7h5kXO9XMT8oi96CA +y/m0sRtW9XLS/BnRAjEAkfcwkz8QRitxpNA7RJvAKQIFskF3UfN5Wp6OFKBOQtJb +gfM0agPnIjhQW+0ZT0MW +-----END CERTIFICATE----- + +# Issuer: CN=DigiCert TLS ECC P384 Root G5 O=DigiCert, Inc. +# Subject: CN=DigiCert TLS ECC P384 Root G5 O=DigiCert, Inc. +# Label: "DigiCert TLS ECC P384 Root G5" +# Serial: 13129116028163249804115411775095713523 +# MD5 Fingerprint: d3:71:04:6a:43:1c:db:a6:59:e1:a8:a3:aa:c5:71:ed +# SHA1 Fingerprint: 17:f3:de:5e:9f:0f:19:e9:8e:f6:1f:32:26:6e:20:c4:07:ae:30:ee +# SHA256 Fingerprint: 01:8e:13:f0:77:25:32:cf:80:9b:d1:b1:72:81:86:72:83:fc:48:c6:e1:3b:e9:c6:98:12:85:4a:49:0c:1b:05 +-----BEGIN CERTIFICATE----- +MIICGTCCAZ+gAwIBAgIQCeCTZaz32ci5PhwLBCou8zAKBggqhkjOPQQDAzBOMQsw +CQYDVQQGEwJVUzEXMBUGA1UEChMORGlnaUNlcnQsIEluYy4xJjAkBgNVBAMTHURp +Z2lDZXJ0IFRMUyBFQ0MgUDM4NCBSb290IEc1MB4XDTIxMDExNTAwMDAwMFoXDTQ2 +MDExNDIzNTk1OVowTjELMAkGA1UEBhMCVVMxFzAVBgNVBAoTDkRpZ2lDZXJ0LCBJ +bmMuMSYwJAYDVQQDEx1EaWdpQ2VydCBUTFMgRUNDIFAzODQgUm9vdCBHNTB2MBAG +ByqGSM49AgEGBSuBBAAiA2IABMFEoc8Rl1Ca3iOCNQfN0MsYndLxf3c1TzvdlHJS +7cI7+Oz6e2tYIOyZrsn8aLN1udsJ7MgT9U7GCh1mMEy7H0cKPGEQQil8pQgO4CLp +0zVozptjn4S1mU1YoI71VOeVyaNCMEAwHQYDVR0OBBYEFMFRRVBZqz7nLFr6ICIS +B4CIfBFqMA4GA1UdDwEB/wQEAwIBhjAPBgNVHRMBAf8EBTADAQH/MAoGCCqGSM49 +BAMDA2gAMGUCMQCJao1H5+z8blUD2WdsJk6Dxv3J+ysTvLd6jLRl0mlpYxNjOyZQ +LgGheQaRnUi/wr4CMEfDFXuxoJGZSZOoPHzoRgaLLPIxAJSdYsiJvRmEFOml+wG4 +DXZDjC5Ty3zfDBeWUA== +-----END CERTIFICATE----- + +# Issuer: CN=DigiCert TLS RSA4096 Root G5 O=DigiCert, Inc. +# Subject: CN=DigiCert TLS RSA4096 Root G5 O=DigiCert, Inc. +# Label: "DigiCert TLS RSA4096 Root G5" +# Serial: 11930366277458970227240571539258396554 +# MD5 Fingerprint: ac:fe:f7:34:96:a9:f2:b3:b4:12:4b:e4:27:41:6f:e1 +# SHA1 Fingerprint: a7:88:49:dc:5d:7c:75:8c:8c:de:39:98:56:b3:aa:d0:b2:a5:71:35 +# SHA256 Fingerprint: 37:1a:00:dc:05:33:b3:72:1a:7e:eb:40:e8:41:9e:70:79:9d:2b:0a:0f:2c:1d:80:69:31:65:f7:ce:c4:ad:75 +-----BEGIN CERTIFICATE----- +MIIFZjCCA06gAwIBAgIQCPm0eKj6ftpqMzeJ3nzPijANBgkqhkiG9w0BAQwFADBN +MQswCQYDVQQGEwJVUzEXMBUGA1UEChMORGlnaUNlcnQsIEluYy4xJTAjBgNVBAMT +HERpZ2lDZXJ0IFRMUyBSU0E0MDk2IFJvb3QgRzUwHhcNMjEwMTE1MDAwMDAwWhcN +NDYwMTE0MjM1OTU5WjBNMQswCQYDVQQGEwJVUzEXMBUGA1UEChMORGlnaUNlcnQs +IEluYy4xJTAjBgNVBAMTHERpZ2lDZXJ0IFRMUyBSU0E0MDk2IFJvb3QgRzUwggIi +MA0GCSqGSIb3DQEBAQUAA4ICDwAwggIKAoICAQCz0PTJeRGd/fxmgefM1eS87IE+ +ajWOLrfn3q/5B03PMJ3qCQuZvWxX2hhKuHisOjmopkisLnLlvevxGs3npAOpPxG0 +2C+JFvuUAT27L/gTBaF4HI4o4EXgg/RZG5Wzrn4DReW+wkL+7vI8toUTmDKdFqgp +wgscONyfMXdcvyej/Cestyu9dJsXLfKB2l2w4SMXPohKEiPQ6s+d3gMXsUJKoBZM +pG2T6T867jp8nVid9E6P/DsjyG244gXazOvswzH016cpVIDPRFtMbzCe88zdH5RD +nU1/cHAN1DrRN/BsnZvAFJNY781BOHW8EwOVfH/jXOnVDdXifBBiqmvwPXbzP6Po +sMH976pXTayGpxi0KcEsDr9kvimM2AItzVwv8n/vFfQMFawKsPHTDU9qTXeXAaDx +Zre3zu/O7Oyldcqs4+Fj97ihBMi8ez9dLRYiVu1ISf6nL3kwJZu6ay0/nTvEF+cd +Lvvyz6b84xQslpghjLSR6Rlgg/IwKwZzUNWYOwbpx4oMYIwo+FKbbuH2TbsGJJvX +KyY//SovcfXWJL5/MZ4PbeiPT02jP/816t9JXkGPhvnxd3lLG7SjXi/7RgLQZhNe +XoVPzthwiHvOAbWWl9fNff2C+MIkwcoBOU+NosEUQB+cZtUMCUbW8tDRSHZWOkPL +tgoRObqME2wGtZ7P6wIDAQABo0IwQDAdBgNVHQ4EFgQUUTMc7TZArxfTJc1paPKv +TiM+s0EwDgYDVR0PAQH/BAQDAgGGMA8GA1UdEwEB/wQFMAMBAf8wDQYJKoZIhvcN +AQEMBQADggIBAGCmr1tfV9qJ20tQqcQjNSH/0GEwhJG3PxDPJY7Jv0Y02cEhJhxw +GXIeo8mH/qlDZJY6yFMECrZBu8RHANmfGBg7sg7zNOok992vIGCukihfNudd5N7H +PNtQOa27PShNlnx2xlv0wdsUpasZYgcYQF+Xkdycx6u1UQ3maVNVzDl92sURVXLF +O4uJ+DQtpBflF+aZfTCIITfNMBc9uPK8qHWgQ9w+iUuQrm0D4ByjoJYJu32jtyoQ +REtGBzRj7TG5BO6jm5qu5jF49OokYTurWGT/u4cnYiWB39yhL/btp/96j1EuMPik +AdKFOV8BmZZvWltwGUb+hmA+rYAQCd05JS9Yf7vSdPD3Rh9GOUrYU9DzLjtxpdRv +/PNn5AeP3SYZ4Y1b+qOTEZvpyDrDVWiakuFSdjjo4bq9+0/V77PnSIMx8IIh47a+ +p6tv75/fTM8BuGJqIz3nCU2AG3swpMPdB380vqQmsvZB6Akd4yCYqjdP//fx4ilw +MUc/dNAUFvohigLVigmUdy7yWSiLfFCSCmZ4OIN1xLVaqBHG5cGdZlXPU8Sv13WF +qUITVuwhd4GTWgzqltlJyqEI8pc7bZsEGCREjnwB8twl2F6GmrE52/WRMmrRpnCK +ovfepEWFJqgejF0pW8hL2JpqA15w8oVPbEtoL8pU9ozaMv7Da4M/OMZ+ +-----END CERTIFICATE----- + +# Issuer: CN=Certainly Root R1 O=Certainly +# Subject: CN=Certainly Root R1 O=Certainly +# Label: "Certainly Root R1" +# Serial: 188833316161142517227353805653483829216 +# MD5 Fingerprint: 07:70:d4:3e:82:87:a0:fa:33:36:13:f4:fa:33:e7:12 +# SHA1 Fingerprint: a0:50:ee:0f:28:71:f4:27:b2:12:6d:6f:50:96:25:ba:cc:86:42:af +# SHA256 Fingerprint: 77:b8:2c:d8:64:4c:43:05:f7:ac:c5:cb:15:6b:45:67:50:04:03:3d:51:c6:0c:62:02:a8:e0:c3:34:67:d3:a0 +-----BEGIN CERTIFICATE----- +MIIFRzCCAy+gAwIBAgIRAI4P+UuQcWhlM1T01EQ5t+AwDQYJKoZIhvcNAQELBQAw +PTELMAkGA1UEBhMCVVMxEjAQBgNVBAoTCUNlcnRhaW5seTEaMBgGA1UEAxMRQ2Vy +dGFpbmx5IFJvb3QgUjEwHhcNMjEwNDAxMDAwMDAwWhcNNDYwNDAxMDAwMDAwWjA9 +MQswCQYDVQQGEwJVUzESMBAGA1UEChMJQ2VydGFpbmx5MRowGAYDVQQDExFDZXJ0 +YWlubHkgUm9vdCBSMTCCAiIwDQYJKoZIhvcNAQEBBQADggIPADCCAgoCggIBANA2 +1B/q3avk0bbm+yLA3RMNansiExyXPGhjZjKcA7WNpIGD2ngwEc/csiu+kr+O5MQT +vqRoTNoCaBZ0vrLdBORrKt03H2As2/X3oXyVtwxwhi7xOu9S98zTm/mLvg7fMbed +aFySpvXl8wo0tf97ouSHocavFwDvA5HtqRxOcT3Si2yJ9HiG5mpJoM610rCrm/b0 +1C7jcvk2xusVtyWMOvwlDbMicyF0yEqWYZL1LwsYpfSt4u5BvQF5+paMjRcCMLT5 +r3gajLQ2EBAHBXDQ9DGQilHFhiZ5shGIXsXwClTNSaa/ApzSRKft43jvRl5tcdF5 +cBxGX1HpyTfcX35pe0HfNEXgO4T0oYoKNp43zGJS4YkNKPl6I7ENPT2a/Z2B7yyQ +wHtETrtJ4A5KVpK8y7XdeReJkd5hiXSSqOMyhb5OhaRLWcsrxXiOcVTQAjeZjOVJ +6uBUcqQRBi8LjMFbvrWhsFNunLhgkR9Za/kt9JQKl7XsxXYDVBtlUrpMklZRNaBA +2CnbrlJ2Oy0wQJuK0EJWtLeIAaSHO1OWzaMWj/Nmqhexx2DgwUMFDO6bW2BvBlyH +Wyf5QBGenDPBt+U1VwV/J84XIIwc/PH72jEpSe31C4SnT8H2TsIonPru4K8H+zMR +eiFPCyEQtkA6qyI6BJyLm4SGcprSp6XEtHWRqSsjAgMBAAGjQjBAMA4GA1UdDwEB +/wQEAwIBBjAPBgNVHRMBAf8EBTADAQH/MB0GA1UdDgQWBBTgqj8ljZ9EXME66C6u +d0yEPmcM9DANBgkqhkiG9w0BAQsFAAOCAgEAuVevuBLaV4OPaAszHQNTVfSVcOQr +PbA56/qJYv331hgELyE03fFo8NWWWt7CgKPBjcZq91l3rhVkz1t5BXdm6ozTaw3d +8VkswTOlMIAVRQdFGjEitpIAq5lNOo93r6kiyi9jyhXWx8bwPWz8HA2YEGGeEaIi +1wrykXprOQ4vMMM2SZ/g6Q8CRFA3lFV96p/2O7qUpUzpvD5RtOjKkjZUbVwlKNrd +rRT90+7iIgXr0PK3aBLXWopBGsaSpVo7Y0VPv+E6dyIvXL9G+VoDhRNCX8reU9di +taY1BMJH/5n9hN9czulegChB8n3nHpDYT3Y+gjwN/KUD+nsa2UUeYNrEjvn8K8l7 +lcUq/6qJ34IxD3L/DCfXCh5WAFAeDJDBlrXYFIW7pw0WwfgHJBu6haEaBQmAupVj +yTrsJZ9/nbqkRxWbRHDxakvWOF5D8xh+UG7pWijmZeZ3Gzr9Hb4DJqPb1OG7fpYn +Kx3upPvaJVQTA945xsMfTZDsjxtK0hzthZU4UHlG1sGQUDGpXJpuHfUzVounmdLy +yCwzk5Iwx06MZTMQZBf9JBeW0Y3COmor6xOLRPIh80oat3df1+2IpHLlOR+Vnb5n +wXARPbv0+Em34yaXOp/SX3z7wJl8OSngex2/DaeP0ik0biQVy96QXr8axGbqwua6 +OV+KmalBWQewLK8= +-----END CERTIFICATE----- + +# Issuer: CN=Certainly Root E1 O=Certainly +# Subject: CN=Certainly Root E1 O=Certainly +# Label: "Certainly Root E1" +# Serial: 8168531406727139161245376702891150584 +# MD5 Fingerprint: 0a:9e:ca:cd:3e:52:50:c6:36:f3:4b:a3:ed:a7:53:e9 +# SHA1 Fingerprint: f9:e1:6d:dc:01:89:cf:d5:82:45:63:3e:c5:37:7d:c2:eb:93:6f:2b +# SHA256 Fingerprint: b4:58:5f:22:e4:ac:75:6a:4e:86:12:a1:36:1c:5d:9d:03:1a:93:fd:84:fe:bb:77:8f:a3:06:8b:0f:c4:2d:c2 +-----BEGIN CERTIFICATE----- +MIIB9zCCAX2gAwIBAgIQBiUzsUcDMydc+Y2aub/M+DAKBggqhkjOPQQDAzA9MQsw +CQYDVQQGEwJVUzESMBAGA1UEChMJQ2VydGFpbmx5MRowGAYDVQQDExFDZXJ0YWlu +bHkgUm9vdCBFMTAeFw0yMTA0MDEwMDAwMDBaFw00NjA0MDEwMDAwMDBaMD0xCzAJ +BgNVBAYTAlVTMRIwEAYDVQQKEwlDZXJ0YWlubHkxGjAYBgNVBAMTEUNlcnRhaW5s +eSBSb290IEUxMHYwEAYHKoZIzj0CAQYFK4EEACIDYgAE3m/4fxzf7flHh4axpMCK ++IKXgOqPyEpeKn2IaKcBYhSRJHpcnqMXfYqGITQYUBsQ3tA3SybHGWCA6TS9YBk2 +QNYphwk8kXr2vBMj3VlOBF7PyAIcGFPBMdjaIOlEjeR2o0IwQDAOBgNVHQ8BAf8E +BAMCAQYwDwYDVR0TAQH/BAUwAwEB/zAdBgNVHQ4EFgQU8ygYy2R17ikq6+2uI1g4 +hevIIgcwCgYIKoZIzj0EAwMDaAAwZQIxALGOWiDDshliTd6wT99u0nCK8Z9+aozm +ut6Dacpps6kFtZaSF4fC0urQe87YQVt8rgIwRt7qy12a7DLCZRawTDBcMPPaTnOG +BtjOiQRINzf43TNRnXCve1XYAS59BWQOhriR +-----END CERTIFICATE----- + +# Issuer: CN=Security Communication RootCA3 O=SECOM Trust Systems CO.,LTD. +# Subject: CN=Security Communication RootCA3 O=SECOM Trust Systems CO.,LTD. +# Label: "Security Communication RootCA3" +# Serial: 16247922307909811815 +# MD5 Fingerprint: 1c:9a:16:ff:9e:5c:e0:4d:8a:14:01:f4:35:5d:29:26 +# SHA1 Fingerprint: c3:03:c8:22:74:92:e5:61:a2:9c:5f:79:91:2b:1e:44:13:91:30:3a +# SHA256 Fingerprint: 24:a5:5c:2a:b0:51:44:2d:06:17:76:65:41:23:9a:4a:d0:32:d7:c5:51:75:aa:34:ff:de:2f:bc:4f:5c:52:94 +-----BEGIN CERTIFICATE----- +MIIFfzCCA2egAwIBAgIJAOF8N0D9G/5nMA0GCSqGSIb3DQEBDAUAMF0xCzAJBgNV +BAYTAkpQMSUwIwYDVQQKExxTRUNPTSBUcnVzdCBTeXN0ZW1zIENPLixMVEQuMScw +JQYDVQQDEx5TZWN1cml0eSBDb21tdW5pY2F0aW9uIFJvb3RDQTMwHhcNMTYwNjE2 +MDYxNzE2WhcNMzgwMTE4MDYxNzE2WjBdMQswCQYDVQQGEwJKUDElMCMGA1UEChMc +U0VDT00gVHJ1c3QgU3lzdGVtcyBDTy4sTFRELjEnMCUGA1UEAxMeU2VjdXJpdHkg +Q29tbXVuaWNhdGlvbiBSb290Q0EzMIICIjANBgkqhkiG9w0BAQEFAAOCAg8AMIIC +CgKCAgEA48lySfcw3gl8qUCBWNO0Ot26YQ+TUG5pPDXC7ltzkBtnTCHsXzW7OT4r +CmDvu20rhvtxosis5FaU+cmvsXLUIKx00rgVrVH+hXShuRD+BYD5UpOzQD11EKzA +lrenfna84xtSGc4RHwsENPXY9Wk8d/Nk9A2qhd7gCVAEF5aEt8iKvE1y/By7z/MG +TfmfZPd+pmaGNXHIEYBMwXFAWB6+oHP2/D5Q4eAvJj1+XCO1eXDe+uDRpdYMQXF7 +9+qMHIjH7Iv10S9VlkZ8WjtYO/u62C21Jdp6Ts9EriGmnpjKIG58u4iFW/vAEGK7 +8vknR+/RiTlDxN/e4UG/VHMgly1s2vPUB6PmudhvrvyMGS7TZ2crldtYXLVqAvO4 +g160a75BflcJdURQVc1aEWEhCmHCqYj9E7wtiS/NYeCVvsq1e+F7NGcLH7YMx3we +GVPKp7FKFSBWFHA9K4IsD50VHUeAR/94mQ4xr28+j+2GaR57GIgUssL8gjMunEst ++3A7caoreyYn8xrC3PsXuKHqy6C0rtOUfnrQq8PsOC0RLoi/1D+tEjtCrI8Cbn3M +0V9hvqG8OmpI6iZVIhZdXw3/JzOfGAN0iltSIEdrRU0id4xVJ/CvHozJgyJUt5rQ +T9nO/NkuHJYosQLTA70lUhw0Zk8jq/R3gpYd0VcwCBEF/VfR2ccCAwEAAaNCMEAw +HQYDVR0OBBYEFGQUfPxYchamCik0FW8qy7z8r6irMA4GA1UdDwEB/wQEAwIBBjAP +BgNVHRMBAf8EBTADAQH/MA0GCSqGSIb3DQEBDAUAA4ICAQDcAiMI4u8hOscNtybS +YpOnpSNyByCCYN8Y11StaSWSntkUz5m5UoHPrmyKO1o5yGwBQ8IibQLwYs1OY0PA +FNr0Y/Dq9HHuTofjcan0yVflLl8cebsjqodEV+m9NU1Bu0soo5iyG9kLFwfl9+qd +9XbXv8S2gVj/yP9kaWJ5rW4OH3/uHWnlt3Jxs/6lATWUVCvAUm2PVcTJ0rjLyjQI +UYWg9by0F1jqClx6vWPGOi//lkkZhOpn2ASxYfQAW0q3nHE3GYV5v4GwxxMOdnE+ +OoAGrgYWp421wsTL/0ClXI2lyTrtcoHKXJg80jQDdwj98ClZXSEIx2C/pHF7uNke +gr4Jr2VvKKu/S7XuPghHJ6APbw+LP6yVGPO5DtxnVW5inkYO0QR4ynKudtml+LLf +iAlhi+8kTtFZP1rUPcmTPCtk9YENFpb3ksP+MW/oKjJ0DvRMmEoYDjBU1cXrvMUV +nuiZIesnKwkK2/HmcBhWuwzkvvnoEKQTkrgc4NtnHVMDpCKn3F2SEDzq//wbEBrD +2NCcnWXL0CsnMQMeNuE9dnUM/0Umud1RvCPHX9jYhxBAEg09ODfnRDwYwFMJZI// +1ZqmfHAuc1Uh6N//g7kdPjIe1qZ9LPFm6Vwdp6POXiUyK+OVrCoHzrQoeIY8Laad +TdJ0MN1kURXbg4NR16/9M51NZg== +-----END CERTIFICATE----- + +# Issuer: CN=Security Communication ECC RootCA1 O=SECOM Trust Systems CO.,LTD. +# Subject: CN=Security Communication ECC RootCA1 O=SECOM Trust Systems CO.,LTD. +# Label: "Security Communication ECC RootCA1" +# Serial: 15446673492073852651 +# MD5 Fingerprint: 7e:43:b0:92:68:ec:05:43:4c:98:ab:5d:35:2e:7e:86 +# SHA1 Fingerprint: b8:0e:26:a9:bf:d2:b2:3b:c0:ef:46:c9:ba:c7:bb:f6:1d:0d:41:41 +# SHA256 Fingerprint: e7:4f:bd:a5:5b:d5:64:c4:73:a3:6b:44:1a:a7:99:c8:a6:8e:07:74:40:e8:28:8b:9f:a1:e5:0e:4b:ba:ca:11 +-----BEGIN CERTIFICATE----- +MIICODCCAb6gAwIBAgIJANZdm7N4gS7rMAoGCCqGSM49BAMDMGExCzAJBgNVBAYT +AkpQMSUwIwYDVQQKExxTRUNPTSBUcnVzdCBTeXN0ZW1zIENPLixMVEQuMSswKQYD +VQQDEyJTZWN1cml0eSBDb21tdW5pY2F0aW9uIEVDQyBSb290Q0ExMB4XDTE2MDYx +NjA1MTUyOFoXDTM4MDExODA1MTUyOFowYTELMAkGA1UEBhMCSlAxJTAjBgNVBAoT +HFNFQ09NIFRydXN0IFN5c3RlbXMgQ08uLExURC4xKzApBgNVBAMTIlNlY3VyaXR5 +IENvbW11bmljYXRpb24gRUNDIFJvb3RDQTEwdjAQBgcqhkjOPQIBBgUrgQQAIgNi +AASkpW9gAwPDvTH00xecK4R1rOX9PVdu12O/5gSJko6BnOPpR27KkBLIE+Cnnfdl +dB9sELLo5OnvbYUymUSxXv3MdhDYW72ixvnWQuRXdtyQwjWpS4g8EkdtXP9JTxpK +ULGjQjBAMB0GA1UdDgQWBBSGHOf+LaVKiwj+KBH6vqNm+GBZLzAOBgNVHQ8BAf8E +BAMCAQYwDwYDVR0TAQH/BAUwAwEB/zAKBggqhkjOPQQDAwNoADBlAjAVXUI9/Lbu +9zuxNuie9sRGKEkz0FhDKmMpzE2xtHqiuQ04pV1IKv3LsnNdo4gIxwwCMQDAqy0O +be0YottT6SXbVQjgUMzfRGEWgqtJsLKB7HOHeLRMsmIbEvoWTSVLY70eN9k= +-----END CERTIFICATE----- + +# Issuer: CN=BJCA Global Root CA1 O=BEIJING CERTIFICATE AUTHORITY +# Subject: CN=BJCA Global Root CA1 O=BEIJING CERTIFICATE AUTHORITY +# Label: "BJCA Global Root CA1" +# Serial: 113562791157148395269083148143378328608 +# MD5 Fingerprint: 42:32:99:76:43:33:36:24:35:07:82:9b:28:f9:d0:90 +# SHA1 Fingerprint: d5:ec:8d:7b:4c:ba:79:f4:e7:e8:cb:9d:6b:ae:77:83:10:03:21:6a +# SHA256 Fingerprint: f3:89:6f:88:fe:7c:0a:88:27:66:a7:fa:6a:d2:74:9f:b5:7a:7f:3e:98:fb:76:9c:1f:a7:b0:9c:2c:44:d5:ae +-----BEGIN CERTIFICATE----- +MIIFdDCCA1ygAwIBAgIQVW9l47TZkGobCdFsPsBsIDANBgkqhkiG9w0BAQsFADBU +MQswCQYDVQQGEwJDTjEmMCQGA1UECgwdQkVJSklORyBDRVJUSUZJQ0FURSBBVVRI +T1JJVFkxHTAbBgNVBAMMFEJKQ0EgR2xvYmFsIFJvb3QgQ0ExMB4XDTE5MTIxOTAz +MTYxN1oXDTQ0MTIxMjAzMTYxN1owVDELMAkGA1UEBhMCQ04xJjAkBgNVBAoMHUJF +SUpJTkcgQ0VSVElGSUNBVEUgQVVUSE9SSVRZMR0wGwYDVQQDDBRCSkNBIEdsb2Jh +bCBSb290IENBMTCCAiIwDQYJKoZIhvcNAQEBBQADggIPADCCAgoCggIBAPFmCL3Z +xRVhy4QEQaVpN3cdwbB7+sN3SJATcmTRuHyQNZ0YeYjjlwE8R4HyDqKYDZ4/N+AZ +spDyRhySsTphzvq3Rp4Dhtczbu33RYx2N95ulpH3134rhxfVizXuhJFyV9xgw8O5 +58dnJCNPYwpj9mZ9S1WnP3hkSWkSl+BMDdMJoDIwOvqfwPKcxRIqLhy1BDPapDgR +at7GGPZHOiJBhyL8xIkoVNiMpTAK+BcWyqw3/XmnkRd4OJmtWO2y3syJfQOcs4ll +5+M7sSKGjwZteAf9kRJ/sGsciQ35uMt0WwfCyPQ10WRjeulumijWML3mG90Vr4Tq +nMfK9Q7q8l0ph49pczm+LiRvRSGsxdRpJQaDrXpIhRMsDQa4bHlW/KNnMoH1V6XK +V0Jp6VwkYe/iMBhORJhVb3rCk9gZtt58R4oRTklH2yiUAguUSiz5EtBP6DF+bHq/ +pj+bOT0CFqMYs2esWz8sgytnOYFcuX6U1WTdno9uruh8W7TXakdI136z1C2OVnZO +z2nxbkRs1CTqjSShGL+9V/6pmTW12xB3uD1IutbB5/EjPtffhZ0nPNRAvQoMvfXn +jSXWgXSHRtQpdaJCbPdzied9v3pKH9MiyRVVz99vfFXQpIsHETdfg6YmV6YBW37+ +WGgHqel62bno/1Afq8K0wM7o6v0PvY1NuLxxAgMBAAGjQjBAMB0GA1UdDgQWBBTF +7+3M2I0hxkjk49cULqcWk+WYATAPBgNVHRMBAf8EBTADAQH/MA4GA1UdDwEB/wQE +AwIBBjANBgkqhkiG9w0BAQsFAAOCAgEAUoKsITQfI/Ki2Pm4rzc2IInRNwPWaZ+4 +YRC6ojGYWUfo0Q0lHhVBDOAqVdVXUsv45Mdpox1NcQJeXyFFYEhcCY5JEMEE3Kli +awLwQ8hOnThJdMkycFRtwUf8jrQ2ntScvd0g1lPJGKm1Vrl2i5VnZu69mP6u775u ++2D2/VnGKhs/I0qUJDAnyIm860Qkmss9vk/Ves6OF8tiwdneHg56/0OGNFK8YT88 +X7vZdrRTvJez/opMEi4r89fO4aL/3Xtw+zuhTaRjAv04l5U/BXCga99igUOLtFkN +SoxUnMW7gZ/NfaXvCyUeOiDbHPwfmGcCCtRzRBPbUYQaVQNW4AB+dAb/OMRyHdOo +P2gxXdMJxy6MW2Pg6Nwe0uxhHvLe5e/2mXZgLR6UcnHGCyoyx5JO1UbXHfmpGQrI ++pXObSOYqgs4rZpWDW+N8TEAiMEXnM0ZNjX+VVOg4DwzX5Ze4jLp3zO7Bkqp2IRz +znfSxqxx4VyjHQy7Ct9f4qNx2No3WqB4K/TUfet27fJhcKVlmtOJNBir+3I+17Q9 +eVzYH6Eze9mCUAyTF6ps3MKCuwJXNq+YJyo5UOGwifUll35HaBC07HPKs5fRJNz2 +YqAo07WjuGS3iGJCz51TzZm+ZGiPTx4SSPfSKcOYKMryMguTjClPPGAyzQWWYezy +r/6zcCwupvI= +-----END CERTIFICATE----- + +# Issuer: CN=BJCA Global Root CA2 O=BEIJING CERTIFICATE AUTHORITY +# Subject: CN=BJCA Global Root CA2 O=BEIJING CERTIFICATE AUTHORITY +# Label: "BJCA Global Root CA2" +# Serial: 58605626836079930195615843123109055211 +# MD5 Fingerprint: 5e:0a:f6:47:5f:a6:14:e8:11:01:95:3f:4d:01:eb:3c +# SHA1 Fingerprint: f4:27:86:eb:6e:b8:6d:88:31:67:02:fb:ba:66:a4:53:00:aa:7a:a6 +# SHA256 Fingerprint: 57:4d:f6:93:1e:27:80:39:66:7b:72:0a:fd:c1:60:0f:c2:7e:b6:6d:d3:09:29:79:fb:73:85:64:87:21:28:82 +-----BEGIN CERTIFICATE----- +MIICJTCCAaugAwIBAgIQLBcIfWQqwP6FGFkGz7RK6zAKBggqhkjOPQQDAzBUMQsw +CQYDVQQGEwJDTjEmMCQGA1UECgwdQkVJSklORyBDRVJUSUZJQ0FURSBBVVRIT1JJ +VFkxHTAbBgNVBAMMFEJKQ0EgR2xvYmFsIFJvb3QgQ0EyMB4XDTE5MTIxOTAzMTgy +MVoXDTQ0MTIxMjAzMTgyMVowVDELMAkGA1UEBhMCQ04xJjAkBgNVBAoMHUJFSUpJ +TkcgQ0VSVElGSUNBVEUgQVVUSE9SSVRZMR0wGwYDVQQDDBRCSkNBIEdsb2JhbCBS +b290IENBMjB2MBAGByqGSM49AgEGBSuBBAAiA2IABJ3LgJGNU2e1uVCxA/jlSR9B +IgmwUVJY1is0j8USRhTFiy8shP8sbqjV8QnjAyEUxEM9fMEsxEtqSs3ph+B99iK+ ++kpRuDCK/eHeGBIK9ke35xe/J4rUQUyWPGCWwf0VHKNCMEAwHQYDVR0OBBYEFNJK +sVF/BvDRgh9Obl+rg/xI1LCRMA8GA1UdEwEB/wQFMAMBAf8wDgYDVR0PAQH/BAQD +AgEGMAoGCCqGSM49BAMDA2gAMGUCMBq8W9f+qdJUDkpd0m2xQNz0Q9XSSpkZElaA +94M04TVOSG0ED1cxMDAtsaqdAzjbBgIxAMvMh1PLet8gUXOQwKhbYdDFUDn9hf7B +43j4ptZLvZuHjw/l1lOWqzzIQNph91Oj9w== +-----END CERTIFICATE----- + +# Issuer: CN=Sectigo Public Server Authentication Root E46 O=Sectigo Limited +# Subject: CN=Sectigo Public Server Authentication Root E46 O=Sectigo Limited +# Label: "Sectigo Public Server Authentication Root E46" +# Serial: 88989738453351742415770396670917916916 +# MD5 Fingerprint: 28:23:f8:b2:98:5c:37:16:3b:3e:46:13:4e:b0:b3:01 +# SHA1 Fingerprint: ec:8a:39:6c:40:f0:2e:bc:42:75:d4:9f:ab:1c:1a:5b:67:be:d2:9a +# SHA256 Fingerprint: c9:0f:26:f0:fb:1b:40:18:b2:22:27:51:9b:5c:a2:b5:3e:2c:a5:b3:be:5c:f1:8e:fe:1b:ef:47:38:0c:53:83 +-----BEGIN CERTIFICATE----- +MIICOjCCAcGgAwIBAgIQQvLM2htpN0RfFf51KBC49DAKBggqhkjOPQQDAzBfMQsw +CQYDVQQGEwJHQjEYMBYGA1UEChMPU2VjdGlnbyBMaW1pdGVkMTYwNAYDVQQDEy1T +ZWN0aWdvIFB1YmxpYyBTZXJ2ZXIgQXV0aGVudGljYXRpb24gUm9vdCBFNDYwHhcN +MjEwMzIyMDAwMDAwWhcNNDYwMzIxMjM1OTU5WjBfMQswCQYDVQQGEwJHQjEYMBYG +A1UEChMPU2VjdGlnbyBMaW1pdGVkMTYwNAYDVQQDEy1TZWN0aWdvIFB1YmxpYyBT +ZXJ2ZXIgQXV0aGVudGljYXRpb24gUm9vdCBFNDYwdjAQBgcqhkjOPQIBBgUrgQQA +IgNiAAR2+pmpbiDt+dd34wc7qNs9Xzjoq1WmVk/WSOrsfy2qw7LFeeyZYX8QeccC +WvkEN/U0NSt3zn8gj1KjAIns1aeibVvjS5KToID1AZTc8GgHHs3u/iVStSBDHBv+ +6xnOQ6OjQjBAMB0GA1UdDgQWBBTRItpMWfFLXyY4qp3W7usNw/upYTAOBgNVHQ8B +Af8EBAMCAYYwDwYDVR0TAQH/BAUwAwEB/zAKBggqhkjOPQQDAwNnADBkAjAn7qRa +qCG76UeXlImldCBteU/IvZNeWBj7LRoAasm4PdCkT0RHlAFWovgzJQxC36oCMB3q +4S6ILuH5px0CMk7yn2xVdOOurvulGu7t0vzCAxHrRVxgED1cf5kDW21USAGKcw== +-----END CERTIFICATE----- + +# Issuer: CN=Sectigo Public Server Authentication Root R46 O=Sectigo Limited +# Subject: CN=Sectigo Public Server Authentication Root R46 O=Sectigo Limited +# Label: "Sectigo Public Server Authentication Root R46" +# Serial: 156256931880233212765902055439220583700 +# MD5 Fingerprint: 32:10:09:52:00:d5:7e:6c:43:df:15:c0:b1:16:93:e5 +# SHA1 Fingerprint: ad:98:f9:f3:e4:7d:75:3b:65:d4:82:b3:a4:52:17:bb:6e:f5:e4:38 +# SHA256 Fingerprint: 7b:b6:47:a6:2a:ee:ac:88:bf:25:7a:a5:22:d0:1f:fe:a3:95:e0:ab:45:c7:3f:93:f6:56:54:ec:38:f2:5a:06 +-----BEGIN CERTIFICATE----- +MIIFijCCA3KgAwIBAgIQdY39i658BwD6qSWn4cetFDANBgkqhkiG9w0BAQwFADBf +MQswCQYDVQQGEwJHQjEYMBYGA1UEChMPU2VjdGlnbyBMaW1pdGVkMTYwNAYDVQQD +Ey1TZWN0aWdvIFB1YmxpYyBTZXJ2ZXIgQXV0aGVudGljYXRpb24gUm9vdCBSNDYw +HhcNMjEwMzIyMDAwMDAwWhcNNDYwMzIxMjM1OTU5WjBfMQswCQYDVQQGEwJHQjEY +MBYGA1UEChMPU2VjdGlnbyBMaW1pdGVkMTYwNAYDVQQDEy1TZWN0aWdvIFB1Ymxp +YyBTZXJ2ZXIgQXV0aGVudGljYXRpb24gUm9vdCBSNDYwggIiMA0GCSqGSIb3DQEB +AQUAA4ICDwAwggIKAoICAQCTvtU2UnXYASOgHEdCSe5jtrch/cSV1UgrJnwUUxDa +ef0rty2k1Cz66jLdScK5vQ9IPXtamFSvnl0xdE8H/FAh3aTPaE8bEmNtJZlMKpnz +SDBh+oF8HqcIStw+KxwfGExxqjWMrfhu6DtK2eWUAtaJhBOqbchPM8xQljeSM9xf +iOefVNlI8JhD1mb9nxc4Q8UBUQvX4yMPFF1bFOdLvt30yNoDN9HWOaEhUTCDsG3X +ME6WW5HwcCSrv0WBZEMNvSE6Lzzpng3LILVCJ8zab5vuZDCQOc2TZYEhMbUjUDM3 +IuM47fgxMMxF/mL50V0yeUKH32rMVhlATc6qu/m1dkmU8Sf4kaWD5QazYw6A3OAS +VYCmO2a0OYctyPDQ0RTp5A1NDvZdV3LFOxxHVp3i1fuBYYzMTYCQNFu31xR13NgE +SJ/AwSiItOkcyqex8Va3e0lMWeUgFaiEAin6OJRpmkkGj80feRQXEgyDet4fsZfu ++Zd4KKTIRJLpfSYFplhym3kT2BFfrsU4YjRosoYwjviQYZ4ybPUHNs2iTG7sijbt +8uaZFURww3y8nDnAtOFr94MlI1fZEoDlSfB1D++N6xybVCi0ITz8fAr/73trdf+L +HaAZBav6+CuBQug4urv7qv094PPK306Xlynt8xhW6aWWrL3DkJiy4Pmi1KZHQ3xt +zwIDAQABo0IwQDAdBgNVHQ4EFgQUVnNYZJX5khqwEioEYnmhQBWIIUkwDgYDVR0P +AQH/BAQDAgGGMA8GA1UdEwEB/wQFMAMBAf8wDQYJKoZIhvcNAQEMBQADggIBAC9c +mTz8Bl6MlC5w6tIyMY208FHVvArzZJ8HXtXBc2hkeqK5Duj5XYUtqDdFqij0lgVQ +YKlJfp/imTYpE0RHap1VIDzYm/EDMrraQKFz6oOht0SmDpkBm+S8f74TlH7Kph52 +gDY9hAaLMyZlbcp+nv4fjFg4exqDsQ+8FxG75gbMY/qB8oFM2gsQa6H61SilzwZA +Fv97fRheORKkU55+MkIQpiGRqRxOF3yEvJ+M0ejf5lG5Nkc/kLnHvALcWxxPDkjB +JYOcCj+esQMzEhonrPcibCTRAUH4WAP+JWgiH5paPHxsnnVI84HxZmduTILA7rpX +DhjvLpr3Etiga+kFpaHpaPi8TD8SHkXoUsCjvxInebnMMTzD9joiFgOgyY9mpFui +TdaBJQbpdqQACj7LzTWb4OE4y2BThihCQRxEV+ioratF4yUQvNs+ZUH7G6aXD+u5 +dHn5HrwdVw1Hr8Mvn4dGp+smWg9WY7ViYG4A++MnESLn/pmPNPW56MORcr3Ywx65 +LvKRRFHQV80MNNVIIb/bE/FmJUNS0nAiNs2fxBx1IK1jcmMGDw4nztJqDby1ORrp +0XZ60Vzk50lJLVU3aPAaOpg+VBeHVOmmJ1CJeyAvP/+/oYtKR5j/K3tJPsMpRmAY +QqszKbrAKbkTidOIijlBO8n9pu0f9GBj39ItVQGL +-----END CERTIFICATE----- + +# Issuer: CN=SSL.com TLS RSA Root CA 2022 O=SSL Corporation +# Subject: CN=SSL.com TLS RSA Root CA 2022 O=SSL Corporation +# Label: "SSL.com TLS RSA Root CA 2022" +# Serial: 148535279242832292258835760425842727825 +# MD5 Fingerprint: d8:4e:c6:59:30:d8:fe:a0:d6:7a:5a:2c:2c:69:78:da +# SHA1 Fingerprint: ec:2c:83:40:72:af:26:95:10:ff:0e:f2:03:ee:31:70:f6:78:9d:ca +# SHA256 Fingerprint: 8f:af:7d:2e:2c:b4:70:9b:b8:e0:b3:36:66:bf:75:a5:dd:45:b5:de:48:0f:8e:a8:d4:bf:e6:be:bc:17:f2:ed +-----BEGIN CERTIFICATE----- +MIIFiTCCA3GgAwIBAgIQb77arXO9CEDii02+1PdbkTANBgkqhkiG9w0BAQsFADBO +MQswCQYDVQQGEwJVUzEYMBYGA1UECgwPU1NMIENvcnBvcmF0aW9uMSUwIwYDVQQD +DBxTU0wuY29tIFRMUyBSU0EgUm9vdCBDQSAyMDIyMB4XDTIyMDgyNTE2MzQyMloX +DTQ2MDgxOTE2MzQyMVowTjELMAkGA1UEBhMCVVMxGDAWBgNVBAoMD1NTTCBDb3Jw +b3JhdGlvbjElMCMGA1UEAwwcU1NMLmNvbSBUTFMgUlNBIFJvb3QgQ0EgMjAyMjCC +AiIwDQYJKoZIhvcNAQEBBQADggIPADCCAgoCggIBANCkCXJPQIgSYT41I57u9nTP +L3tYPc48DRAokC+X94xI2KDYJbFMsBFMF3NQ0CJKY7uB0ylu1bUJPiYYf7ISf5OY +t6/wNr/y7hienDtSxUcZXXTzZGbVXcdotL8bHAajvI9AI7YexoS9UcQbOcGV0ins +S657Lb85/bRi3pZ7QcacoOAGcvvwB5cJOYF0r/c0WRFXCsJbwST0MXMwgsadugL3 +PnxEX4MN8/HdIGkWCVDi1FW24IBydm5MR7d1VVm0U3TZlMZBrViKMWYPHqIbKUBO +L9975hYsLfy/7PO0+r4Y9ptJ1O4Fbtk085zx7AGL0SDGD6C1vBdOSHtRwvzpXGk3 +R2azaPgVKPC506QVzFpPulJwoxJF3ca6TvvC0PeoUidtbnm1jPx7jMEWTO6Af77w +dr5BUxIzrlo4QqvXDz5BjXYHMtWrifZOZ9mxQnUjbvPNQrL8VfVThxc7wDNY8VLS ++YCk8OjwO4s4zKTGkH8PnP2L0aPP2oOnaclQNtVcBdIKQXTbYxE3waWglksejBYS +d66UNHsef8JmAOSqg+qKkK3ONkRN0VHpvB/zagX9wHQfJRlAUW7qglFA35u5CCoG +AtUjHBPW6dvbxrB6y3snm/vg1UYk7RBLY0ulBY+6uB0rpvqR4pJSvezrZ5dtmi2f +gTIFZzL7SAg/2SW4BCUvAgMBAAGjYzBhMA8GA1UdEwEB/wQFMAMBAf8wHwYDVR0j +BBgwFoAU+y437uOEeicuzRk1sTN8/9REQrkwHQYDVR0OBBYEFPsuN+7jhHonLs0Z +NbEzfP/UREK5MA4GA1UdDwEB/wQEAwIBhjANBgkqhkiG9w0BAQsFAAOCAgEAjYlt +hEUY8U+zoO9opMAdrDC8Z2awms22qyIZZtM7QbUQnRC6cm4pJCAcAZli05bg4vsM +QtfhWsSWTVTNj8pDU/0quOr4ZcoBwq1gaAafORpR2eCNJvkLTqVTJXojpBzOCBvf +R4iyrT7gJ4eLSYwfqUdYe5byiB0YrrPRpgqU+tvT5TgKa3kSM/tKWTcWQA673vWJ +DPFs0/dRa1419dvAJuoSc06pkZCmF8NsLzjUo3KUQyxi4U5cMj29TH0ZR6LDSeeW +P4+a0zvkEdiLA9z2tmBVGKaBUfPhqBVq6+AL8BQx1rmMRTqoENjwuSfr98t67wVy +lrXEj5ZzxOhWc5y8aVFjvO9nHEMaX3cZHxj4HCUp+UmZKbaSPaKDN7EgkaibMOlq +bLQjk2UEqxHzDh1TJElTHaE/nUiSEeJ9DU/1172iWD54nR4fK/4huxoTtrEoZP2w +AgDHbICivRZQIA9ygV/MlP+7mea6kMvq+cYMwq7FGc4zoWtcu358NFcXrfA/rs3q +r5nsLFR+jM4uElZI7xc7P0peYNLcdDa8pUNjyw9bowJWCZ4kLOGGgYz+qxcs+sji +Mho6/4UIyYOf8kpIEFR3N+2ivEC+5BB09+Rbu7nzifmPQdjH5FCQNYA+HLhNkNPU +98OwoX6EyneSMSy4kLGCenROmxMmtNVQZlR4rmA= +-----END CERTIFICATE----- + +# Issuer: CN=SSL.com TLS ECC Root CA 2022 O=SSL Corporation +# Subject: CN=SSL.com TLS ECC Root CA 2022 O=SSL Corporation +# Label: "SSL.com TLS ECC Root CA 2022" +# Serial: 26605119622390491762507526719404364228 +# MD5 Fingerprint: 99:d7:5c:f1:51:36:cc:e9:ce:d9:19:2e:77:71:56:c5 +# SHA1 Fingerprint: 9f:5f:d9:1a:54:6d:f5:0c:71:f0:ee:7a:bd:17:49:98:84:73:e2:39 +# SHA256 Fingerprint: c3:2f:fd:9f:46:f9:36:d1:6c:36:73:99:09:59:43:4b:9a:d6:0a:af:bb:9e:7c:f3:36:54:f1:44:cc:1b:a1:43 +-----BEGIN CERTIFICATE----- +MIICOjCCAcCgAwIBAgIQFAP1q/s3ixdAW+JDsqXRxDAKBggqhkjOPQQDAzBOMQsw +CQYDVQQGEwJVUzEYMBYGA1UECgwPU1NMIENvcnBvcmF0aW9uMSUwIwYDVQQDDBxT +U0wuY29tIFRMUyBFQ0MgUm9vdCBDQSAyMDIyMB4XDTIyMDgyNTE2MzM0OFoXDTQ2 +MDgxOTE2MzM0N1owTjELMAkGA1UEBhMCVVMxGDAWBgNVBAoMD1NTTCBDb3Jwb3Jh +dGlvbjElMCMGA1UEAwwcU1NMLmNvbSBUTFMgRUNDIFJvb3QgQ0EgMjAyMjB2MBAG +ByqGSM49AgEGBSuBBAAiA2IABEUpNXP6wrgjzhR9qLFNoFs27iosU8NgCTWyJGYm +acCzldZdkkAZDsalE3D07xJRKF3nzL35PIXBz5SQySvOkkJYWWf9lCcQZIxPBLFN +SeR7T5v15wj4A4j3p8OSSxlUgaNjMGEwDwYDVR0TAQH/BAUwAwEB/zAfBgNVHSME +GDAWgBSJjy+j6CugFFR781a4Jl9nOAuc0DAdBgNVHQ4EFgQUiY8vo+groBRUe/NW +uCZfZzgLnNAwDgYDVR0PAQH/BAQDAgGGMAoGCCqGSM49BAMDA2gAMGUCMFXjIlbp +15IkWE8elDIPDAI2wv2sdDJO4fscgIijzPvX6yv/N33w7deedWo1dlJF4AIxAMeN +b0Igj762TVntd00pxCAgRWSGOlDGxK0tk/UYfXLtqc/ErFc2KAhl3zx5Zn6g6g== +-----END CERTIFICATE----- + +# Issuer: CN=Atos TrustedRoot Root CA ECC TLS 2021 O=Atos +# Subject: CN=Atos TrustedRoot Root CA ECC TLS 2021 O=Atos +# Label: "Atos TrustedRoot Root CA ECC TLS 2021" +# Serial: 81873346711060652204712539181482831616 +# MD5 Fingerprint: 16:9f:ad:f1:70:ad:79:d6:ed:29:b4:d1:c5:79:70:a8 +# SHA1 Fingerprint: 9e:bc:75:10:42:b3:02:f3:81:f4:f7:30:62:d4:8f:c3:a7:51:b2:dd +# SHA256 Fingerprint: b2:fa:e5:3e:14:cc:d7:ab:92:12:06:47:01:ae:27:9c:1d:89:88:fa:cb:77:5f:a8:a0:08:91:4e:66:39:88:a8 +-----BEGIN CERTIFICATE----- +MIICFTCCAZugAwIBAgIQPZg7pmY9kGP3fiZXOATvADAKBggqhkjOPQQDAzBMMS4w +LAYDVQQDDCVBdG9zIFRydXN0ZWRSb290IFJvb3QgQ0EgRUNDIFRMUyAyMDIxMQ0w +CwYDVQQKDARBdG9zMQswCQYDVQQGEwJERTAeFw0yMTA0MjIwOTI2MjNaFw00MTA0 +MTcwOTI2MjJaMEwxLjAsBgNVBAMMJUF0b3MgVHJ1c3RlZFJvb3QgUm9vdCBDQSBF +Q0MgVExTIDIwMjExDTALBgNVBAoMBEF0b3MxCzAJBgNVBAYTAkRFMHYwEAYHKoZI +zj0CAQYFK4EEACIDYgAEloZYKDcKZ9Cg3iQZGeHkBQcfl+3oZIK59sRxUM6KDP/X +tXa7oWyTbIOiaG6l2b4siJVBzV3dscqDY4PMwL502eCdpO5KTlbgmClBk1IQ1SQ4 +AjJn8ZQSb+/Xxd4u/RmAo0IwQDAPBgNVHRMBAf8EBTADAQH/MB0GA1UdDgQWBBR2 +KCXWfeBmmnoJsmo7jjPXNtNPojAOBgNVHQ8BAf8EBAMCAYYwCgYIKoZIzj0EAwMD +aAAwZQIwW5kp85wxtolrbNa9d+F851F+uDrNozZffPc8dz7kUK2o59JZDCaOMDtu +CCrCp1rIAjEAmeMM56PDr9NJLkaCI2ZdyQAUEv049OGYa3cpetskz2VAv9LcjBHo +9H1/IISpQuQo +-----END CERTIFICATE----- + +# Issuer: CN=Atos TrustedRoot Root CA RSA TLS 2021 O=Atos +# Subject: CN=Atos TrustedRoot Root CA RSA TLS 2021 O=Atos +# Label: "Atos TrustedRoot Root CA RSA TLS 2021" +# Serial: 111436099570196163832749341232207667876 +# MD5 Fingerprint: d4:d3:46:b8:9a:c0:9c:76:5d:9e:3a:c3:b9:99:31:d2 +# SHA1 Fingerprint: 18:52:3b:0d:06:37:e4:d6:3a:df:23:e4:98:fb:5b:16:fb:86:74:48 +# SHA256 Fingerprint: 81:a9:08:8e:a5:9f:b3:64:c5:48:a6:f8:55:59:09:9b:6f:04:05:ef:bf:18:e5:32:4e:c9:f4:57:ba:00:11:2f +-----BEGIN CERTIFICATE----- +MIIFZDCCA0ygAwIBAgIQU9XP5hmTC/srBRLYwiqipDANBgkqhkiG9w0BAQwFADBM +MS4wLAYDVQQDDCVBdG9zIFRydXN0ZWRSb290IFJvb3QgQ0EgUlNBIFRMUyAyMDIx +MQ0wCwYDVQQKDARBdG9zMQswCQYDVQQGEwJERTAeFw0yMTA0MjIwOTIxMTBaFw00 +MTA0MTcwOTIxMDlaMEwxLjAsBgNVBAMMJUF0b3MgVHJ1c3RlZFJvb3QgUm9vdCBD +QSBSU0EgVExTIDIwMjExDTALBgNVBAoMBEF0b3MxCzAJBgNVBAYTAkRFMIICIjAN +BgkqhkiG9w0BAQEFAAOCAg8AMIICCgKCAgEAtoAOxHm9BYx9sKOdTSJNy/BBl01Z +4NH+VoyX8te9j2y3I49f1cTYQcvyAh5x5en2XssIKl4w8i1mx4QbZFc4nXUtVsYv +Ye+W/CBGvevUez8/fEc4BKkbqlLfEzfTFRVOvV98r61jx3ncCHvVoOX3W3WsgFWZ +kmGbzSoXfduP9LVq6hdKZChmFSlsAvFr1bqjM9xaZ6cF4r9lthawEO3NUDPJcFDs +GY6wx/J0W2tExn2WuZgIWWbeKQGb9Cpt0xU6kGpn8bRrZtkh68rZYnxGEFzedUln +nkL5/nWpo63/dgpnQOPF943HhZpZnmKaau1Fh5hnstVKPNe0OwANwI8f4UDErmwh +3El+fsqyjW22v5MvoVw+j8rtgI5Y4dtXz4U2OLJxpAmMkokIiEjxQGMYsluMWuPD +0xeqqxmjLBvk1cbiZnrXghmmOxYsL3GHX0WelXOTwkKBIROW1527k2gV+p2kHYzy +geBYBr3JtuP2iV2J+axEoctr+hbxx1A9JNr3w+SH1VbxT5Aw+kUJWdo0zuATHAR8 +ANSbhqRAvNncTFd+rrcztl524WWLZt+NyteYr842mIycg5kDcPOvdO3GDjbnvezB +c6eUWsuSZIKmAMFwoW4sKeFYV+xafJlrJaSQOoD0IJ2azsct+bJLKZWD6TWNp0lI +pw9MGZHQ9b8Q4HECAwEAAaNCMEAwDwYDVR0TAQH/BAUwAwEB/zAdBgNVHQ4EFgQU +dEmZ0f+0emhFdcN+tNzMzjkz2ggwDgYDVR0PAQH/BAQDAgGGMA0GCSqGSIb3DQEB +DAUAA4ICAQAjQ1MkYlxt/T7Cz1UAbMVWiLkO3TriJQ2VSpfKgInuKs1l+NsW4AmS +4BjHeJi78+xCUvuppILXTdiK/ORO/auQxDh1MoSf/7OwKwIzNsAQkG8dnK/haZPs +o0UvFJ/1TCplQ3IM98P4lYsU84UgYt1UU90s3BiVaU+DR3BAM1h3Egyi61IxHkzJ +qM7F78PRreBrAwA0JrRUITWXAdxfG/F851X6LWh3e9NpzNMOa7pNdkTWwhWaJuyw +xfW70Xp0wmzNxbVe9kzmWy2B27O3Opee7c9GslA9hGCZcbUztVdF5kJHdWoOsAgM +rr3e97sPWD2PAzHoPYJQyi9eDF20l74gNAf0xBLh7tew2VktafcxBPTy+av5EzH4 +AXcOPUIjJsyacmdRIXrMPIWo6iFqO9taPKU0nprALN+AnCng33eU0aKAQv9qTFsR +0PXNor6uzFFcw9VUewyu1rkGd4Di7wcaaMxZUa1+XGdrudviB0JbuAEFWDlN5LuY +o7Ey7Nmj1m+UI/87tyll5gfp77YZ6ufCOB0yiJA8EytuzO+rdwY0d4RPcuSBhPm5 +dDTedk+SKlOxJTnbPP/lPqYO5Wue/9vsL3SD3460s6neFE3/MaNFcyT6lSnMEpcE +oji2jbDwN/zIIX8/syQbPYtuzE2wFg2WHYMfRsCbvUOZ58SWLs5fyQ== +-----END CERTIFICATE----- + +# Issuer: CN=TrustAsia Global Root CA G3 O=TrustAsia Technologies, Inc. +# Subject: CN=TrustAsia Global Root CA G3 O=TrustAsia Technologies, Inc. +# Label: "TrustAsia Global Root CA G3" +# Serial: 576386314500428537169965010905813481816650257167 +# MD5 Fingerprint: 30:42:1b:b7:bb:81:75:35:e4:16:4f:53:d2:94:de:04 +# SHA1 Fingerprint: 63:cf:b6:c1:27:2b:56:e4:88:8e:1c:23:9a:b6:2e:81:47:24:c3:c7 +# SHA256 Fingerprint: e0:d3:22:6a:eb:11:63:c2:e4:8f:f9:be:3b:50:b4:c6:43:1b:e7:bb:1e:ac:c5:c3:6b:5d:5e:c5:09:03:9a:08 +-----BEGIN CERTIFICATE----- +MIIFpTCCA42gAwIBAgIUZPYOZXdhaqs7tOqFhLuxibhxkw8wDQYJKoZIhvcNAQEM +BQAwWjELMAkGA1UEBhMCQ04xJTAjBgNVBAoMHFRydXN0QXNpYSBUZWNobm9sb2dp +ZXMsIEluYy4xJDAiBgNVBAMMG1RydXN0QXNpYSBHbG9iYWwgUm9vdCBDQSBHMzAe +Fw0yMTA1MjAwMjEwMTlaFw00NjA1MTkwMjEwMTlaMFoxCzAJBgNVBAYTAkNOMSUw +IwYDVQQKDBxUcnVzdEFzaWEgVGVjaG5vbG9naWVzLCBJbmMuMSQwIgYDVQQDDBtU +cnVzdEFzaWEgR2xvYmFsIFJvb3QgQ0EgRzMwggIiMA0GCSqGSIb3DQEBAQUAA4IC +DwAwggIKAoICAQDAMYJhkuSUGwoqZdC+BqmHO1ES6nBBruL7dOoKjbmzTNyPtxNS +T1QY4SxzlZHFZjtqz6xjbYdT8PfxObegQ2OwxANdV6nnRM7EoYNl9lA+sX4WuDqK +AtCWHwDNBSHvBm3dIZwZQ0WhxeiAysKtQGIXBsaqvPPW5vxQfmZCHzyLpnl5hkA1 +nyDvP+uLRx+PjsXUjrYsyUQE49RDdT/VP68czH5GX6zfZBCK70bwkPAPLfSIC7Ep +qq+FqklYqL9joDiR5rPmd2jE+SoZhLsO4fWvieylL1AgdB4SQXMeJNnKziyhWTXA +yB1GJ2Faj/lN03J5Zh6fFZAhLf3ti1ZwA0pJPn9pMRJpxx5cynoTi+jm9WAPzJMs +hH/x/Gr8m0ed262IPfN2dTPXS6TIi/n1Q1hPy8gDVI+lhXgEGvNz8teHHUGf59gX +zhqcD0r83ERoVGjiQTz+LISGNzzNPy+i2+f3VANfWdP3kXjHi3dqFuVJhZBFcnAv +kV34PmVACxmZySYgWmjBNb9Pp1Hx2BErW+Canig7CjoKH8GB5S7wprlppYiU5msT +f9FkPz2ccEblooV7WIQn3MSAPmeamseaMQ4w7OYXQJXZRe0Blqq/DPNL0WP3E1jA +uPP6Z92bfW1K/zJMtSU7/xxnD4UiWQWRkUF3gdCFTIcQcf+eQxuulXUtgQIDAQAB +o2MwYTAPBgNVHRMBAf8EBTADAQH/MB8GA1UdIwQYMBaAFEDk5PIj7zjKsK5Xf/Ih +MBY027ySMB0GA1UdDgQWBBRA5OTyI+84yrCuV3/yITAWNNu8kjAOBgNVHQ8BAf8E +BAMCAQYwDQYJKoZIhvcNAQEMBQADggIBACY7UeFNOPMyGLS0XuFlXsSUT9SnYaP4 +wM8zAQLpw6o1D/GUE3d3NZ4tVlFEbuHGLige/9rsR82XRBf34EzC4Xx8MnpmyFq2 +XFNFV1pF1AWZLy4jVe5jaN/TG3inEpQGAHUNcoTpLrxaatXeL1nHo+zSh2bbt1S1 +JKv0Q3jbSwTEb93mPmY+KfJLaHEih6D4sTNjduMNhXJEIlU/HHzp/LgV6FL6qj6j +ITk1dImmasI5+njPtqzn59ZW/yOSLlALqbUHM/Q4X6RJpstlcHboCoWASzY9M/eV +VHUl2qzEc4Jl6VL1XP04lQJqaTDFHApXB64ipCz5xUG3uOyfT0gA+QEEVcys+TIx +xHWVBqB/0Y0n3bOppHKH/lmLmnp0Ft0WpWIp6zqW3IunaFnT63eROfjXy9mPX1on +AX1daBli2MjN9LdyR75bl87yraKZk62Uy5P2EgmVtqvXO9A/EcswFi55gORngS1d +7XB4tmBZrOFdRWOPyN9yaFvqHbgB8X7754qz41SgOAngPN5C8sLtLpvzHzW2Ntjj +gKGLzZlkD8Kqq7HK9W+eQ42EVJmzbsASZthwEPEGNTNDqJwuuhQxzhB/HIbjj9LV ++Hfsm6vxL2PZQl/gZ4FkkfGXL/xuJvYz+NO1+MRiqzFRJQJ6+N1rZdVtTTDIZbpo +FGWsJwt0ivKH +-----END CERTIFICATE----- + +# Issuer: CN=TrustAsia Global Root CA G4 O=TrustAsia Technologies, Inc. +# Subject: CN=TrustAsia Global Root CA G4 O=TrustAsia Technologies, Inc. +# Label: "TrustAsia Global Root CA G4" +# Serial: 451799571007117016466790293371524403291602933463 +# MD5 Fingerprint: 54:dd:b2:d7:5f:d8:3e:ed:7c:e0:0b:2e:cc:ed:eb:eb +# SHA1 Fingerprint: 57:73:a5:61:5d:80:b2:e6:ac:38:82:fc:68:07:31:ac:9f:b5:92:5a +# SHA256 Fingerprint: be:4b:56:cb:50:56:c0:13:6a:52:6d:f4:44:50:8d:aa:36:a0:b5:4f:42:e4:ac:38:f7:2a:f4:70:e4:79:65:4c +-----BEGIN CERTIFICATE----- +MIICVTCCAdygAwIBAgIUTyNkuI6XY57GU4HBdk7LKnQV1tcwCgYIKoZIzj0EAwMw +WjELMAkGA1UEBhMCQ04xJTAjBgNVBAoMHFRydXN0QXNpYSBUZWNobm9sb2dpZXMs +IEluYy4xJDAiBgNVBAMMG1RydXN0QXNpYSBHbG9iYWwgUm9vdCBDQSBHNDAeFw0y +MTA1MjAwMjEwMjJaFw00NjA1MTkwMjEwMjJaMFoxCzAJBgNVBAYTAkNOMSUwIwYD +VQQKDBxUcnVzdEFzaWEgVGVjaG5vbG9naWVzLCBJbmMuMSQwIgYDVQQDDBtUcnVz +dEFzaWEgR2xvYmFsIFJvb3QgQ0EgRzQwdjAQBgcqhkjOPQIBBgUrgQQAIgNiAATx +s8045CVD5d4ZCbuBeaIVXxVjAd7Cq92zphtnS4CDr5nLrBfbK5bKfFJV4hrhPVbw +LxYI+hW8m7tH5j/uqOFMjPXTNvk4XatwmkcN4oFBButJ+bAp3TPsUKV/eSm4IJij +YzBhMA8GA1UdEwEB/wQFMAMBAf8wHwYDVR0jBBgwFoAUpbtKl86zK3+kMd6Xg1mD +pm9xy94wHQYDVR0OBBYEFKW7SpfOsyt/pDHel4NZg6ZvccveMA4GA1UdDwEB/wQE +AwIBBjAKBggqhkjOPQQDAwNnADBkAjBe8usGzEkxn0AAbbd+NvBNEU/zy4k6LHiR +UKNbwMp1JvK/kF0LgoxgKJ/GcJpo5PECMFxYDlZ2z1jD1xCMuo6u47xkdUfFVZDj +/bpV6wfEU6s3qe4hsiFbYI89MvHVI5TWWA== +-----END CERTIFICATE----- + +# Issuer: CN=CommScope Public Trust ECC Root-01 O=CommScope +# Subject: CN=CommScope Public Trust ECC Root-01 O=CommScope +# Label: "CommScope Public Trust ECC Root-01" +# Serial: 385011430473757362783587124273108818652468453534 +# MD5 Fingerprint: 3a:40:a7:fc:03:8c:9c:38:79:2f:3a:a2:6c:b6:0a:16 +# SHA1 Fingerprint: 07:86:c0:d8:dd:8e:c0:80:98:06:98:d0:58:7a:ef:de:a6:cc:a2:5d +# SHA256 Fingerprint: 11:43:7c:da:7b:b4:5e:41:36:5f:45:b3:9a:38:98:6b:0d:e0:0d:ef:34:8e:0c:7b:b0:87:36:33:80:0b:c3:8b +-----BEGIN CERTIFICATE----- +MIICHTCCAaOgAwIBAgIUQ3CCd89NXTTxyq4yLzf39H91oJ4wCgYIKoZIzj0EAwMw +TjELMAkGA1UEBhMCVVMxEjAQBgNVBAoMCUNvbW1TY29wZTErMCkGA1UEAwwiQ29t +bVNjb3BlIFB1YmxpYyBUcnVzdCBFQ0MgUm9vdC0wMTAeFw0yMTA0MjgxNzM1NDNa +Fw00NjA0MjgxNzM1NDJaME4xCzAJBgNVBAYTAlVTMRIwEAYDVQQKDAlDb21tU2Nv +cGUxKzApBgNVBAMMIkNvbW1TY29wZSBQdWJsaWMgVHJ1c3QgRUNDIFJvb3QtMDEw +djAQBgcqhkjOPQIBBgUrgQQAIgNiAARLNumuV16ocNfQj3Rid8NeeqrltqLxeP0C +flfdkXmcbLlSiFS8LwS+uM32ENEp7LXQoMPwiXAZu1FlxUOcw5tjnSCDPgYLpkJE +hRGnSjot6dZoL0hOUysHP029uax3OVejQjBAMA8GA1UdEwEB/wQFMAMBAf8wDgYD +VR0PAQH/BAQDAgEGMB0GA1UdDgQWBBSOB2LAUN3GGQYARnQE9/OufXVNMDAKBggq +hkjOPQQDAwNoADBlAjEAnDPfQeMjqEI2Jpc1XHvr20v4qotzVRVcrHgpD7oh2MSg +2NED3W3ROT3Ek2DS43KyAjB8xX6I01D1HiXo+k515liWpDVfG2XqYZpwI7UNo5uS +Um9poIyNStDuiw7LR47QjRE= +-----END CERTIFICATE----- + +# Issuer: CN=CommScope Public Trust ECC Root-02 O=CommScope +# Subject: CN=CommScope Public Trust ECC Root-02 O=CommScope +# Label: "CommScope Public Trust ECC Root-02" +# Serial: 234015080301808452132356021271193974922492992893 +# MD5 Fingerprint: 59:b0:44:d5:65:4d:b8:5c:55:19:92:02:b6:d1:94:b2 +# SHA1 Fingerprint: 3c:3f:ef:57:0f:fe:65:93:86:9e:a0:fe:b0:f6:ed:8e:d1:13:c7:e5 +# SHA256 Fingerprint: 2f:fb:7f:81:3b:bb:b3:c8:9a:b4:e8:16:2d:0f:16:d7:15:09:a8:30:cc:9d:73:c2:62:e5:14:08:75:d1:ad:4a +-----BEGIN CERTIFICATE----- +MIICHDCCAaOgAwIBAgIUKP2ZYEFHpgE6yhR7H+/5aAiDXX0wCgYIKoZIzj0EAwMw +TjELMAkGA1UEBhMCVVMxEjAQBgNVBAoMCUNvbW1TY29wZTErMCkGA1UEAwwiQ29t +bVNjb3BlIFB1YmxpYyBUcnVzdCBFQ0MgUm9vdC0wMjAeFw0yMTA0MjgxNzQ0NTRa +Fw00NjA0MjgxNzQ0NTNaME4xCzAJBgNVBAYTAlVTMRIwEAYDVQQKDAlDb21tU2Nv +cGUxKzApBgNVBAMMIkNvbW1TY29wZSBQdWJsaWMgVHJ1c3QgRUNDIFJvb3QtMDIw +djAQBgcqhkjOPQIBBgUrgQQAIgNiAAR4MIHoYx7l63FRD/cHB8o5mXxO1Q/MMDAL +j2aTPs+9xYa9+bG3tD60B8jzljHz7aRP+KNOjSkVWLjVb3/ubCK1sK9IRQq9qEmU +v4RDsNuESgMjGWdqb8FuvAY5N9GIIvejQjBAMA8GA1UdEwEB/wQFMAMBAf8wDgYD +VR0PAQH/BAQDAgEGMB0GA1UdDgQWBBTmGHX/72DehKT1RsfeSlXjMjZ59TAKBggq +hkjOPQQDAwNnADBkAjAmc0l6tqvmSfR9Uj/UQQSugEODZXW5hYA4O9Zv5JOGq4/n +ich/m35rChJVYaoR4HkCMHfoMXGsPHED1oQmHhS48zs73u1Z/GtMMH9ZzkXpc2AV +mkzw5l4lIhVtwodZ0LKOag== +-----END CERTIFICATE----- + +# Issuer: CN=CommScope Public Trust RSA Root-01 O=CommScope +# Subject: CN=CommScope Public Trust RSA Root-01 O=CommScope +# Label: "CommScope Public Trust RSA Root-01" +# Serial: 354030733275608256394402989253558293562031411421 +# MD5 Fingerprint: 0e:b4:15:bc:87:63:5d:5d:02:73:d4:26:38:68:73:d8 +# SHA1 Fingerprint: 6d:0a:5f:f7:b4:23:06:b4:85:b3:b7:97:64:fc:ac:75:f5:33:f2:93 +# SHA256 Fingerprint: 02:bd:f9:6e:2a:45:dd:9b:f1:8f:c7:e1:db:df:21:a0:37:9b:a3:c9:c2:61:03:44:cf:d8:d6:06:fe:c1:ed:81 +-----BEGIN CERTIFICATE----- +MIIFbDCCA1SgAwIBAgIUPgNJgXUWdDGOTKvVxZAplsU5EN0wDQYJKoZIhvcNAQEL +BQAwTjELMAkGA1UEBhMCVVMxEjAQBgNVBAoMCUNvbW1TY29wZTErMCkGA1UEAwwi +Q29tbVNjb3BlIFB1YmxpYyBUcnVzdCBSU0EgUm9vdC0wMTAeFw0yMTA0MjgxNjQ1 +NTRaFw00NjA0MjgxNjQ1NTNaME4xCzAJBgNVBAYTAlVTMRIwEAYDVQQKDAlDb21t +U2NvcGUxKzApBgNVBAMMIkNvbW1TY29wZSBQdWJsaWMgVHJ1c3QgUlNBIFJvb3Qt +MDEwggIiMA0GCSqGSIb3DQEBAQUAA4ICDwAwggIKAoICAQCwSGWjDR1C45FtnYSk +YZYSwu3D2iM0GXb26v1VWvZVAVMP8syMl0+5UMuzAURWlv2bKOx7dAvnQmtVzslh +suitQDy6uUEKBU8bJoWPQ7VAtYXR1HHcg0Hz9kXHgKKEUJdGzqAMxGBWBB0HW0al +DrJLpA6lfO741GIDuZNqihS4cPgugkY4Iw50x2tBt9Apo52AsH53k2NC+zSDO3Oj +WiE260f6GBfZumbCk6SP/F2krfxQapWsvCQz0b2If4b19bJzKo98rwjyGpg/qYFl +P8GMicWWMJoKz/TUyDTtnS+8jTiGU+6Xn6myY5QXjQ/cZip8UlF1y5mO6D1cv547 +KI2DAg+pn3LiLCuz3GaXAEDQpFSOm117RTYm1nJD68/A6g3czhLmfTifBSeolz7p +UcZsBSjBAg/pGG3svZwG1KdJ9FQFa2ww8esD1eo9anbCyxooSU1/ZOD6K9pzg4H/ +kQO9lLvkuI6cMmPNn7togbGEW682v3fuHX/3SZtS7NJ3Wn2RnU3COS3kuoL4b/JO +Hg9O5j9ZpSPcPYeoKFgo0fEbNttPxP/hjFtyjMcmAyejOQoBqsCyMWCDIqFPEgkB +Ea801M/XrmLTBQe0MXXgDW1XT2mH+VepuhX2yFJtocucH+X8eKg1mp9BFM6ltM6U +CBwJrVbl2rZJmkrqYxhTnCwuwwIDAQABo0IwQDAPBgNVHRMBAf8EBTADAQH/MA4G +A1UdDwEB/wQEAwIBBjAdBgNVHQ4EFgQUN12mmnQywsL5x6YVEFm45P3luG0wDQYJ +KoZIhvcNAQELBQADggIBAK+nz97/4L1CjU3lIpbfaOp9TSp90K09FlxD533Ahuh6 +NWPxzIHIxgvoLlI1pKZJkGNRrDSsBTtXAOnTYtPZKdVUvhwQkZyybf5Z/Xn36lbQ +nmhUQo8mUuJM3y+Xpi/SB5io82BdS5pYV4jvguX6r2yBS5KPQJqTRlnLX3gWsWc+ +QgvfKNmwrZggvkN80V4aCRckjXtdlemrwWCrWxhkgPut4AZ9HcpZuPN4KWfGVh2v +trV0KnahP/t1MJ+UXjulYPPLXAziDslg+MkfFoom3ecnf+slpoq9uC02EJqxWE2a +aE9gVOX2RhOOiKy8IUISrcZKiX2bwdgt6ZYD9KJ0DLwAHb/WNyVntHKLr4W96ioD +j8z7PEQkguIBpQtZtjSNMgsSDesnwv1B10A8ckYpwIzqug/xBpMu95yo9GA+o/E4 +Xo4TwbM6l4c/ksp4qRyv0LAbJh6+cOx69TOY6lz/KwsETkPdY34Op054A5U+1C0w +lREQKC6/oAI+/15Z0wUOlV9TRe9rh9VIzRamloPh37MG88EU26fsHItdkJANclHn +YfkUyq+Dj7+vsQpZXdxc1+SWrVtgHdqul7I52Qb1dgAT+GhMIbA1xNxVssnBQVoc +icCMb3SgazNNtQEo/a2tiRc7ppqEvOuM6sRxJKi6KfkIsidWNTJf6jn7MZrVGczw +-----END CERTIFICATE----- + +# Issuer: CN=CommScope Public Trust RSA Root-02 O=CommScope +# Subject: CN=CommScope Public Trust RSA Root-02 O=CommScope +# Label: "CommScope Public Trust RSA Root-02" +# Serial: 480062499834624527752716769107743131258796508494 +# MD5 Fingerprint: e1:29:f9:62:7b:76:e2:96:6d:f3:d4:d7:0f:ae:1f:aa +# SHA1 Fingerprint: ea:b0:e2:52:1b:89:93:4c:11:68:f2:d8:9a:ac:22:4c:a3:8a:57:ae +# SHA256 Fingerprint: ff:e9:43:d7:93:42:4b:4f:7c:44:0c:1c:3d:64:8d:53:63:f3:4b:82:dc:87:aa:7a:9f:11:8f:c5:de:e1:01:f1 +-----BEGIN CERTIFICATE----- +MIIFbDCCA1SgAwIBAgIUVBa/O345lXGN0aoApYYNK496BU4wDQYJKoZIhvcNAQEL +BQAwTjELMAkGA1UEBhMCVVMxEjAQBgNVBAoMCUNvbW1TY29wZTErMCkGA1UEAwwi +Q29tbVNjb3BlIFB1YmxpYyBUcnVzdCBSU0EgUm9vdC0wMjAeFw0yMTA0MjgxNzE2 +NDNaFw00NjA0MjgxNzE2NDJaME4xCzAJBgNVBAYTAlVTMRIwEAYDVQQKDAlDb21t +U2NvcGUxKzApBgNVBAMMIkNvbW1TY29wZSBQdWJsaWMgVHJ1c3QgUlNBIFJvb3Qt +MDIwggIiMA0GCSqGSIb3DQEBAQUAA4ICDwAwggIKAoICAQDh+g77aAASyE3VrCLE +NQE7xVTlWXZjpX/rwcRqmL0yjReA61260WI9JSMZNRTpf4mnG2I81lDnNJUDMrG0 +kyI9p+Kx7eZ7Ti6Hmw0zdQreqjXnfuU2mKKuJZ6VszKWpCtYHu8//mI0SFHRtI1C +rWDaSWqVcN3SAOLMV2MCe5bdSZdbkk6V0/nLKR8YSvgBKtJjCW4k6YnS5cciTNxz +hkcAqg2Ijq6FfUrpuzNPDlJwnZXjfG2WWy09X6GDRl224yW4fKcZgBzqZUPckXk2 +LHR88mcGyYnJ27/aaL8j7dxrrSiDeS/sOKUNNwFnJ5rpM9kzXzehxfCrPfp4sOcs +n/Y+n2Dg70jpkEUeBVF4GiwSLFworA2iI540jwXmojPOEXcT1A6kHkIfhs1w/tku +FT0du7jyU1fbzMZ0KZwYszZ1OC4PVKH4kh+Jlk+71O6d6Ts2QrUKOyrUZHk2EOH5 +kQMreyBUzQ0ZGshBMjTRsJnhkB4BQDa1t/qp5Xd1pCKBXbCL5CcSD1SIxtuFdOa3 +wNemKfrb3vOTlycEVS8KbzfFPROvCgCpLIscgSjX74Yxqa7ybrjKaixUR9gqiC6v +wQcQeKwRoi9C8DfF8rhW3Q5iLc4tVn5V8qdE9isy9COoR+jUKgF4z2rDN6ieZdIs +5fq6M8EGRPbmz6UNp2YINIos8wIDAQABo0IwQDAPBgNVHRMBAf8EBTADAQH/MA4G +A1UdDwEB/wQEAwIBBjAdBgNVHQ4EFgQUR9DnsSL/nSz12Vdgs7GxcJXvYXowDQYJ +KoZIhvcNAQELBQADggIBAIZpsU0v6Z9PIpNojuQhmaPORVMbc0RTAIFhzTHjCLqB +KCh6krm2qMhDnscTJk3C2OVVnJJdUNjCK9v+5qiXz1I6JMNlZFxHMaNlNRPDk7n3 ++VGXu6TwYofF1gbTl4MgqX67tiHCpQ2EAOHyJxCDut0DgdXdaMNmEMjRdrSzbyme +APnCKfWxkxlSaRosTKCL4BWaMS/TiJVZbuXEs1DIFAhKm4sTg7GkcrI7djNB3Nyq +pgdvHSQSn8h2vS/ZjvQs7rfSOBAkNlEv41xdgSGn2rtO/+YHqP65DSdsu3BaVXoT +6fEqSWnHX4dXTEN5bTpl6TBcQe7rd6VzEojov32u5cSoHw2OHG1QAk8mGEPej1WF +sQs3BWDJVTkSBKEqz3EWnzZRSb9wO55nnPt7eck5HHisd5FUmrh1CoFSl+NmYWvt +PjgelmFV4ZFUjO2MJB+ByRCac5krFk5yAD9UG/iNuovnFNa2RU9g7Jauwy8CTl2d +lklyALKrdVwPaFsdZcJfMw8eD/A7hvWwTruc9+olBdytoptLFwG+Qt81IR2tq670 +v64fG9PiO/yzcnMcmyiQiRM9HcEARwmWmjgb3bHPDcK0RPOWlc4yOo80nOAXx17O +rg3bhzjlP1v9mxnhMUF6cKojawHhRUzNlM47ni3niAIi9G7oyOzWPPO5std3eqx7 +-----END CERTIFICATE----- + +# Issuer: CN=Telekom Security TLS ECC Root 2020 O=Deutsche Telekom Security GmbH +# Subject: CN=Telekom Security TLS ECC Root 2020 O=Deutsche Telekom Security GmbH +# Label: "Telekom Security TLS ECC Root 2020" +# Serial: 72082518505882327255703894282316633856 +# MD5 Fingerprint: c1:ab:fe:6a:10:2c:03:8d:bc:1c:22:32:c0:85:a7:fd +# SHA1 Fingerprint: c0:f8:96:c5:a9:3b:01:06:21:07:da:18:42:48:bc:e9:9d:88:d5:ec +# SHA256 Fingerprint: 57:8a:f4:de:d0:85:3f:4e:59:98:db:4a:ea:f9:cb:ea:8d:94:5f:60:b6:20:a3:8d:1a:3c:13:b2:bc:7b:a8:e1 +-----BEGIN CERTIFICATE----- +MIICQjCCAcmgAwIBAgIQNjqWjMlcsljN0AFdxeVXADAKBggqhkjOPQQDAzBjMQsw +CQYDVQQGEwJERTEnMCUGA1UECgweRGV1dHNjaGUgVGVsZWtvbSBTZWN1cml0eSBH +bWJIMSswKQYDVQQDDCJUZWxla29tIFNlY3VyaXR5IFRMUyBFQ0MgUm9vdCAyMDIw +MB4XDTIwMDgyNTA3NDgyMFoXDTQ1MDgyNTIzNTk1OVowYzELMAkGA1UEBhMCREUx +JzAlBgNVBAoMHkRldXRzY2hlIFRlbGVrb20gU2VjdXJpdHkgR21iSDErMCkGA1UE +AwwiVGVsZWtvbSBTZWN1cml0eSBUTFMgRUNDIFJvb3QgMjAyMDB2MBAGByqGSM49 +AgEGBSuBBAAiA2IABM6//leov9Wq9xCazbzREaK9Z0LMkOsVGJDZos0MKiXrPk/O +tdKPD/M12kOLAoC+b1EkHQ9rK8qfwm9QMuU3ILYg/4gND21Ju9sGpIeQkpT0CdDP +f8iAC8GXs7s1J8nCG6NCMEAwHQYDVR0OBBYEFONyzG6VmUex5rNhTNHLq+O6zd6f +MA8GA1UdEwEB/wQFMAMBAf8wDgYDVR0PAQH/BAQDAgEGMAoGCCqGSM49BAMDA2cA +MGQCMHVSi7ekEE+uShCLsoRbQuHmKjYC2qBuGT8lv9pZMo7k+5Dck2TOrbRBR2Di +z6fLHgIwN0GMZt9Ba9aDAEH9L1r3ULRn0SyocddDypwnJJGDSA3PzfdUga/sf+Rn +27iQ7t0l +-----END CERTIFICATE----- + +# Issuer: CN=Telekom Security TLS RSA Root 2023 O=Deutsche Telekom Security GmbH +# Subject: CN=Telekom Security TLS RSA Root 2023 O=Deutsche Telekom Security GmbH +# Label: "Telekom Security TLS RSA Root 2023" +# Serial: 44676229530606711399881795178081572759 +# MD5 Fingerprint: bf:5b:eb:54:40:cd:48:71:c4:20:8d:7d:de:0a:42:f2 +# SHA1 Fingerprint: 54:d3:ac:b3:bd:57:56:f6:85:9d:ce:e5:c3:21:e2:d4:ad:83:d0:93 +# SHA256 Fingerprint: ef:c6:5c:ad:bb:59:ad:b6:ef:e8:4d:a2:23:11:b3:56:24:b7:1b:3b:1e:a0:da:8b:66:55:17:4e:c8:97:86:46 +-----BEGIN CERTIFICATE----- +MIIFszCCA5ugAwIBAgIQIZxULej27HF3+k7ow3BXlzANBgkqhkiG9w0BAQwFADBj +MQswCQYDVQQGEwJERTEnMCUGA1UECgweRGV1dHNjaGUgVGVsZWtvbSBTZWN1cml0 +eSBHbWJIMSswKQYDVQQDDCJUZWxla29tIFNlY3VyaXR5IFRMUyBSU0EgUm9vdCAy +MDIzMB4XDTIzMDMyODEyMTY0NVoXDTQ4MDMyNzIzNTk1OVowYzELMAkGA1UEBhMC +REUxJzAlBgNVBAoMHkRldXRzY2hlIFRlbGVrb20gU2VjdXJpdHkgR21iSDErMCkG +A1UEAwwiVGVsZWtvbSBTZWN1cml0eSBUTFMgUlNBIFJvb3QgMjAyMzCCAiIwDQYJ +KoZIhvcNAQEBBQADggIPADCCAgoCggIBAO01oYGA88tKaVvC+1GDrib94W7zgRJ9 +cUD/h3VCKSHtgVIs3xLBGYSJwb3FKNXVS2xE1kzbB5ZKVXrKNoIENqil/Cf2SfHV +cp6R+SPWcHu79ZvB7JPPGeplfohwoHP89v+1VmLhc2o0mD6CuKyVU/QBoCcHcqMA +U6DksquDOFczJZSfvkgdmOGjup5czQRxUX11eKvzWarE4GC+j4NSuHUaQTXtvPM6 +Y+mpFEXX5lLRbtLevOP1Czvm4MS9Q2QTps70mDdsipWol8hHD/BeEIvnHRz+sTug +BTNoBUGCwQMrAcjnj02r6LX2zWtEtefdi+zqJbQAIldNsLGyMcEWzv/9FIS3R/qy +8XDe24tsNlikfLMR0cN3f1+2JeANxdKz+bi4d9s3cXFH42AYTyS2dTd4uaNir73J +co4vzLuu2+QVUhkHM/tqty1LkCiCc/4YizWN26cEar7qwU02OxY2kTLvtkCJkUPg +8qKrBC7m8kwOFjQgrIfBLX7JZkcXFBGk8/ehJImr2BrIoVyxo/eMbcgByU/J7MT8 +rFEz0ciD0cmfHdRHNCk+y7AO+oMLKFjlKdw/fKifybYKu6boRhYPluV75Gp6SG12 +mAWl3G0eQh5C2hrgUve1g8Aae3g1LDj1H/1Joy7SWWO/gLCMk3PLNaaZlSJhZQNg ++y+TS/qanIA7AgMBAAGjYzBhMA4GA1UdDwEB/wQEAwIBBjAdBgNVHQ4EFgQUtqeX +gj10hZv3PJ+TmpV5dVKMbUcwDwYDVR0TAQH/BAUwAwEB/zAfBgNVHSMEGDAWgBS2 +p5eCPXSFm/c8n5OalXl1UoxtRzANBgkqhkiG9w0BAQwFAAOCAgEAqMxhpr51nhVQ +pGv7qHBFfLp+sVr8WyP6Cnf4mHGCDG3gXkaqk/QeoMPhk9tLrbKmXauw1GLLXrtm +9S3ul0A8Yute1hTWjOKWi0FpkzXmuZlrYrShF2Y0pmtjxrlO8iLpWA1WQdH6DErw +M807u20hOq6OcrXDSvvpfeWxm4bu4uB9tPcy/SKE8YXJN3nptT+/XOR0so8RYgDd +GGah2XsjX/GO1WfoVNpbOms2b/mBsTNHM3dA+VKq3dSDz4V4mZqTuXNnQkYRIer+ +CqkbGmVps4+uFrb2S1ayLfmlyOw7YqPta9BO1UAJpB+Y1zqlklkg5LB9zVtzaL1t +xKITDmcZuI1CfmwMmm6gJC3VRRvcxAIU/oVbZZfKTpBQCHpCNfnqwmbU+AGuHrS+ +w6jv/naaoqYfRvaE7fzbzsQCzndILIyy7MMAo+wsVRjBfhnu4S/yrYObnqsZ38aK +L4x35bcF7DvB7L6Gs4a8wPfc5+pbrrLMtTWGS9DiP7bY+A4A7l3j941Y/8+LN+lj +X273CXE2whJdV/LItM3z7gLfEdxquVeEHVlNjM7IDiPCtyaaEBRx/pOyiriA8A4Q +ntOoUAw3gi/q4Iqd4Sw5/7W0cwDk90imc6y/st53BIe0o82bNSQ3+pCTE4FCxpgm +dTdmQRCsu/WU48IxK63nI1bMNSWSs1A= +-----END CERTIFICATE----- + +# Issuer: CN=FIRMAPROFESIONAL CA ROOT-A WEB O=Firmaprofesional SA +# Subject: CN=FIRMAPROFESIONAL CA ROOT-A WEB O=Firmaprofesional SA +# Label: "FIRMAPROFESIONAL CA ROOT-A WEB" +# Serial: 65916896770016886708751106294915943533 +# MD5 Fingerprint: 82:b2:ad:45:00:82:b0:66:63:f8:5f:c3:67:4e:ce:a3 +# SHA1 Fingerprint: a8:31:11:74:a6:14:15:0d:ca:77:dd:0e:e4:0c:5d:58:fc:a0:72:a5 +# SHA256 Fingerprint: be:f2:56:da:f2:6e:9c:69:bd:ec:16:02:35:97:98:f3:ca:f7:18:21:a0:3e:01:82:57:c5:3c:65:61:7f:3d:4a +-----BEGIN CERTIFICATE----- +MIICejCCAgCgAwIBAgIQMZch7a+JQn81QYehZ1ZMbTAKBggqhkjOPQQDAzBuMQsw +CQYDVQQGEwJFUzEcMBoGA1UECgwTRmlybWFwcm9mZXNpb25hbCBTQTEYMBYGA1UE +YQwPVkFURVMtQTYyNjM0MDY4MScwJQYDVQQDDB5GSVJNQVBST0ZFU0lPTkFMIENB +IFJPT1QtQSBXRUIwHhcNMjIwNDA2MDkwMTM2WhcNNDcwMzMxMDkwMTM2WjBuMQsw +CQYDVQQGEwJFUzEcMBoGA1UECgwTRmlybWFwcm9mZXNpb25hbCBTQTEYMBYGA1UE +YQwPVkFURVMtQTYyNjM0MDY4MScwJQYDVQQDDB5GSVJNQVBST0ZFU0lPTkFMIENB +IFJPT1QtQSBXRUIwdjAQBgcqhkjOPQIBBgUrgQQAIgNiAARHU+osEaR3xyrq89Zf +e9MEkVz6iMYiuYMQYneEMy3pA4jU4DP37XcsSmDq5G+tbbT4TIqk5B/K6k84Si6C +cyvHZpsKjECcfIr28jlgst7L7Ljkb+qbXbdTkBgyVcUgt5SjYzBhMA8GA1UdEwEB +/wQFMAMBAf8wHwYDVR0jBBgwFoAUk+FDY1w8ndYn81LsF7Kpryz3dvgwHQYDVR0O +BBYEFJPhQ2NcPJ3WJ/NS7Beyqa8s93b4MA4GA1UdDwEB/wQEAwIBBjAKBggqhkjO +PQQDAwNoADBlAjAdfKR7w4l1M+E7qUW/Runpod3JIha3RxEL2Jq68cgLcFBTApFw +hVmpHqTm6iMxoAACMQD94vizrxa5HnPEluPBMBnYfubDl94cT7iJLzPrSA8Z94dG +XSaQpYXFuXqUPoeovQA= +-----END CERTIFICATE----- diff --git a/src/fetchcode/vcs/pip/_vendor/certifi/core.py b/src/fetchcode/vcs/pip/_vendor/certifi/core.py index 56b52a3c..70e0c3bd 100644 --- a/src/fetchcode/vcs/pip/_vendor/certifi/core.py +++ b/src/fetchcode/vcs/pip/_vendor/certifi/core.py @@ -1,30 +1,114 @@ -# -*- coding: utf-8 -*- - """ certifi.py ~~~~~~~~~~ This module returns the installation location of cacert.pem or its contents. """ -import os +import sys +import atexit + +def exit_cacert_ctx() -> None: + _CACERT_CTX.__exit__(None, None, None) # type: ignore[union-attr] + + +if sys.version_info >= (3, 11): + + from importlib.resources import as_file, files + + _CACERT_CTX = None + _CACERT_PATH = None + + def where() -> str: + # This is slightly terrible, but we want to delay extracting the file + # in cases where we're inside of a zipimport situation until someone + # actually calls where(), but we don't want to re-extract the file + # on every call of where(), so we'll do it once then store it in a + # global variable. + global _CACERT_CTX + global _CACERT_PATH + if _CACERT_PATH is None: + # This is slightly janky, the importlib.resources API wants you to + # manage the cleanup of this file, so it doesn't actually return a + # path, it returns a context manager that will give you the path + # when you enter it and will do any cleanup when you leave it. In + # the common case of not needing a temporary file, it will just + # return the file system location and the __exit__() is a no-op. + # + # We also have to hold onto the actual context manager, because + # it will do the cleanup whenever it gets garbage collected, so + # we will also store that at the global level as well. + _CACERT_CTX = as_file(files("pip._vendor.certifi").joinpath("cacert.pem")) + _CACERT_PATH = str(_CACERT_CTX.__enter__()) + atexit.register(exit_cacert_ctx) + + return _CACERT_PATH + + def contents() -> str: + return files("pip._vendor.certifi").joinpath("cacert.pem").read_text(encoding="ascii") + +elif sys.version_info >= (3, 7): + + from importlib.resources import path as get_path, read_text + + _CACERT_CTX = None + _CACERT_PATH = None + + def where() -> str: + # This is slightly terrible, but we want to delay extracting the + # file in cases where we're inside of a zipimport situation until + # someone actually calls where(), but we don't want to re-extract + # the file on every call of where(), so we'll do it once then store + # it in a global variable. + global _CACERT_CTX + global _CACERT_PATH + if _CACERT_PATH is None: + # This is slightly janky, the importlib.resources API wants you + # to manage the cleanup of this file, so it doesn't actually + # return a path, it returns a context manager that will give + # you the path when you enter it and will do any cleanup when + # you leave it. In the common case of not needing a temporary + # file, it will just return the file system location and the + # __exit__() is a no-op. + # + # We also have to hold onto the actual context manager, because + # it will do the cleanup whenever it gets garbage collected, so + # we will also store that at the global level as well. + _CACERT_CTX = get_path("pip._vendor.certifi", "cacert.pem") + _CACERT_PATH = str(_CACERT_CTX.__enter__()) + atexit.register(exit_cacert_ctx) + + return _CACERT_PATH + + def contents() -> str: + return read_text("pip._vendor.certifi", "cacert.pem", encoding="ascii") + +else: + import os + import types + from typing import Union + + Package = Union[types.ModuleType, str] + Resource = Union[str, "os.PathLike"] -try: - from importlib.resources import read_text -except ImportError: # This fallback will work for Python versions prior to 3.7 that lack the # importlib.resources module but relies on the existing `where` function # so won't address issues with environments like PyOxidizer that don't set # __file__ on modules. - def read_text(_module, _path, encoding="ascii"): - with open(where(), "r", encoding=encoding) as data: + def read_text( + package: Package, + resource: Resource, + encoding: str = 'utf-8', + errors: str = 'strict' + ) -> str: + with open(where(), encoding=encoding) as data: return data.read() + # If we don't have importlib.resources, then we will just do the old logic + # of assuming we're on the filesystem and munge the path directly. + def where() -> str: + f = os.path.dirname(__file__) -def where(): - f = os.path.dirname(__file__) - - return os.path.join(f, "cacert.pem") - + return os.path.join(f, "cacert.pem") -def contents(): - return read_text("certifi", "cacert.pem", encoding="ascii") + def contents() -> str: + return read_text("pip._vendor.certifi", "cacert.pem", encoding="ascii") diff --git a/src/fetchcode/vcs/pip/_vendor/certifi/py.typed b/src/fetchcode/vcs/pip/_vendor/certifi/py.typed new file mode 100644 index 00000000..e69de29b diff --git a/src/fetchcode/vcs/pip/_vendor/chardet.pyi b/src/fetchcode/vcs/pip/_vendor/chardet.pyi deleted file mode 100644 index 29e87e33..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet.pyi +++ /dev/null @@ -1 +0,0 @@ -from chardet import * \ No newline at end of file diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/LICENSE b/src/fetchcode/vcs/pip/_vendor/chardet/LICENSE deleted file mode 100644 index 8add30ad..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/LICENSE +++ /dev/null @@ -1,504 +0,0 @@ - GNU LESSER GENERAL PUBLIC LICENSE - Version 2.1, February 1999 - - Copyright (C) 1991, 1999 Free Software Foundation, Inc. - 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA - Everyone is permitted to copy and distribute verbatim copies - of this license document, but changing it is not allowed. - -[This is the first released version of the Lesser GPL. It also counts - as the successor of the GNU Library Public License, version 2, hence - the version number 2.1.] - - Preamble - - The licenses for most software are designed to take away your -freedom to share and change it. By contrast, the GNU General Public -Licenses are intended to guarantee your freedom to share and change -free software--to make sure the software is free for all its users. - - This license, the Lesser General Public License, applies to some -specially designated software packages--typically libraries--of the -Free Software Foundation and other authors who decide to use it. You -can use it too, but we suggest you first think carefully about whether -this license or the ordinary General Public License is the better -strategy to use in any particular case, based on the explanations below. - - When we speak of free software, we are referring to freedom of use, -not price. Our General Public Licenses are designed to make sure that -you have the freedom to distribute copies of free software (and charge -for this service if you wish); that you receive source code or can get -it if you want it; that you can change the software and use pieces of -it in new free programs; and that you are informed that you can do -these things. - - To protect your rights, we need to make restrictions that forbid -distributors to deny you these rights or to ask you to surrender these -rights. These restrictions translate to certain responsibilities for -you if you distribute copies of the library or if you modify it. - - For example, if you distribute copies of the library, whether gratis -or for a fee, you must give the recipients all the rights that we gave -you. You must make sure that they, too, receive or can get the source -code. If you link other code with the library, you must provide -complete object files to the recipients, so that they can relink them -with the library after making changes to the library and recompiling -it. And you must show them these terms so they know their rights. - - We protect your rights with a two-step method: (1) we copyright the -library, and (2) we offer you this license, which gives you legal -permission to copy, distribute and/or modify the library. - - To protect each distributor, we want to make it very clear that -there is no warranty for the free library. Also, if the library is -modified by someone else and passed on, the recipients should know -that what they have is not the original version, so that the original -author's reputation will not be affected by problems that might be -introduced by others. - - Finally, software patents pose a constant threat to the existence of -any free program. We wish to make sure that a company cannot -effectively restrict the users of a free program by obtaining a -restrictive license from a patent holder. Therefore, we insist that -any patent license obtained for a version of the library must be -consistent with the full freedom of use specified in this license. - - Most GNU software, including some libraries, is covered by the -ordinary GNU General Public License. This license, the GNU Lesser -General Public License, applies to certain designated libraries, and -is quite different from the ordinary General Public License. We use -this license for certain libraries in order to permit linking those -libraries into non-free programs. - - When a program is linked with a library, whether statically or using -a shared library, the combination of the two is legally speaking a -combined work, a derivative of the original library. The ordinary -General Public License therefore permits such linking only if the -entire combination fits its criteria of freedom. The Lesser General -Public License permits more lax criteria for linking other code with -the library. - - We call this license the "Lesser" General Public License because it -does Less to protect the user's freedom than the ordinary General -Public License. It also provides other free software developers Less -of an advantage over competing non-free programs. These disadvantages -are the reason we use the ordinary General Public License for many -libraries. However, the Lesser license provides advantages in certain -special circumstances. - - For example, on rare occasions, there may be a special need to -encourage the widest possible use of a certain library, so that it becomes -a de-facto standard. To achieve this, non-free programs must be -allowed to use the library. A more frequent case is that a free -library does the same job as widely used non-free libraries. In this -case, there is little to gain by limiting the free library to free -software only, so we use the Lesser General Public License. - - In other cases, permission to use a particular library in non-free -programs enables a greater number of people to use a large body of -free software. For example, permission to use the GNU C Library in -non-free programs enables many more people to use the whole GNU -operating system, as well as its variant, the GNU/Linux operating -system. - - Although the Lesser General Public License is Less protective of the -users' freedom, it does ensure that the user of a program that is -linked with the Library has the freedom and the wherewithal to run -that program using a modified version of the Library. - - The precise terms and conditions for copying, distribution and -modification follow. Pay close attention to the difference between a -"work based on the library" and a "work that uses the library". The -former contains code derived from the library, whereas the latter must -be combined with the library in order to run. - - GNU LESSER GENERAL PUBLIC LICENSE - TERMS AND CONDITIONS FOR COPYING, DISTRIBUTION AND MODIFICATION - - 0. This License Agreement applies to any software library or other -program which contains a notice placed by the copyright holder or -other authorized party saying it may be distributed under the terms of -this Lesser General Public License (also called "this License"). -Each licensee is addressed as "you". - - A "library" means a collection of software functions and/or data -prepared so as to be conveniently linked with application programs -(which use some of those functions and data) to form executables. - - The "Library", below, refers to any such software library or work -which has been distributed under these terms. A "work based on the -Library" means either the Library or any derivative work under -copyright law: that is to say, a work containing the Library or a -portion of it, either verbatim or with modifications and/or translated -straightforwardly into another language. (Hereinafter, translation is -included without limitation in the term "modification".) - - "Source code" for a work means the preferred form of the work for -making modifications to it. For a library, complete source code means -all the source code for all modules it contains, plus any associated -interface definition files, plus the scripts used to control compilation -and installation of the library. - - Activities other than copying, distribution and modification are not -covered by this License; they are outside its scope. The act of -running a program using the Library is not restricted, and output from -such a program is covered only if its contents constitute a work based -on the Library (independent of the use of the Library in a tool for -writing it). Whether that is true depends on what the Library does -and what the program that uses the Library does. - - 1. You may copy and distribute verbatim copies of the Library's -complete source code as you receive it, in any medium, provided that -you conspicuously and appropriately publish on each copy an -appropriate copyright notice and disclaimer of warranty; keep intact -all the notices that refer to this License and to the absence of any -warranty; and distribute a copy of this License along with the -Library. - - You may charge a fee for the physical act of transferring a copy, -and you may at your option offer warranty protection in exchange for a -fee. - - 2. You may modify your copy or copies of the Library or any portion -of it, thus forming a work based on the Library, and copy and -distribute such modifications or work under the terms of Section 1 -above, provided that you also meet all of these conditions: - - a) The modified work must itself be a software library. - - b) You must cause the files modified to carry prominent notices - stating that you changed the files and the date of any change. - - c) You must cause the whole of the work to be licensed at no - charge to all third parties under the terms of this License. - - d) If a facility in the modified Library refers to a function or a - table of data to be supplied by an application program that uses - the facility, other than as an argument passed when the facility - is invoked, then you must make a good faith effort to ensure that, - in the event an application does not supply such function or - table, the facility still operates, and performs whatever part of - its purpose remains meaningful. - - (For example, a function in a library to compute square roots has - a purpose that is entirely well-defined independent of the - application. Therefore, Subsection 2d requires that any - application-supplied function or table used by this function must - be optional: if the application does not supply it, the square - root function must still compute square roots.) - -These requirements apply to the modified work as a whole. If -identifiable sections of that work are not derived from the Library, -and can be reasonably considered independent and separate works in -themselves, then this License, and its terms, do not apply to those -sections when you distribute them as separate works. But when you -distribute the same sections as part of a whole which is a work based -on the Library, the distribution of the whole must be on the terms of -this License, whose permissions for other licensees extend to the -entire whole, and thus to each and every part regardless of who wrote -it. - -Thus, it is not the intent of this section to claim rights or contest -your rights to work written entirely by you; rather, the intent is to -exercise the right to control the distribution of derivative or -collective works based on the Library. - -In addition, mere aggregation of another work not based on the Library -with the Library (or with a work based on the Library) on a volume of -a storage or distribution medium does not bring the other work under -the scope of this License. - - 3. You may opt to apply the terms of the ordinary GNU General Public -License instead of this License to a given copy of the Library. To do -this, you must alter all the notices that refer to this License, so -that they refer to the ordinary GNU General Public License, version 2, -instead of to this License. (If a newer version than version 2 of the -ordinary GNU General Public License has appeared, then you can specify -that version instead if you wish.) Do not make any other change in -these notices. - - Once this change is made in a given copy, it is irreversible for -that copy, so the ordinary GNU General Public License applies to all -subsequent copies and derivative works made from that copy. - - This option is useful when you wish to copy part of the code of -the Library into a program that is not a library. - - 4. You may copy and distribute the Library (or a portion or -derivative of it, under Section 2) in object code or executable form -under the terms of Sections 1 and 2 above provided that you accompany -it with the complete corresponding machine-readable source code, which -must be distributed under the terms of Sections 1 and 2 above on a -medium customarily used for software interchange. - - If distribution of object code is made by offering access to copy -from a designated place, then offering equivalent access to copy the -source code from the same place satisfies the requirement to -distribute the source code, even though third parties are not -compelled to copy the source along with the object code. - - 5. A program that contains no derivative of any portion of the -Library, but is designed to work with the Library by being compiled or -linked with it, is called a "work that uses the Library". Such a -work, in isolation, is not a derivative work of the Library, and -therefore falls outside the scope of this License. - - However, linking a "work that uses the Library" with the Library -creates an executable that is a derivative of the Library (because it -contains portions of the Library), rather than a "work that uses the -library". The executable is therefore covered by this License. -Section 6 states terms for distribution of such executables. - - When a "work that uses the Library" uses material from a header file -that is part of the Library, the object code for the work may be a -derivative work of the Library even though the source code is not. -Whether this is true is especially significant if the work can be -linked without the Library, or if the work is itself a library. The -threshold for this to be true is not precisely defined by law. - - If such an object file uses only numerical parameters, data -structure layouts and accessors, and small macros and small inline -functions (ten lines or less in length), then the use of the object -file is unrestricted, regardless of whether it is legally a derivative -work. (Executables containing this object code plus portions of the -Library will still fall under Section 6.) - - Otherwise, if the work is a derivative of the Library, you may -distribute the object code for the work under the terms of Section 6. -Any executables containing that work also fall under Section 6, -whether or not they are linked directly with the Library itself. - - 6. As an exception to the Sections above, you may also combine or -link a "work that uses the Library" with the Library to produce a -work containing portions of the Library, and distribute that work -under terms of your choice, provided that the terms permit -modification of the work for the customer's own use and reverse -engineering for debugging such modifications. - - You must give prominent notice with each copy of the work that the -Library is used in it and that the Library and its use are covered by -this License. You must supply a copy of this License. If the work -during execution displays copyright notices, you must include the -copyright notice for the Library among them, as well as a reference -directing the user to the copy of this License. Also, you must do one -of these things: - - a) Accompany the work with the complete corresponding - machine-readable source code for the Library including whatever - changes were used in the work (which must be distributed under - Sections 1 and 2 above); and, if the work is an executable linked - with the Library, with the complete machine-readable "work that - uses the Library", as object code and/or source code, so that the - user can modify the Library and then relink to produce a modified - executable containing the modified Library. (It is understood - that the user who changes the contents of definitions files in the - Library will not necessarily be able to recompile the application - to use the modified definitions.) - - b) Use a suitable shared library mechanism for linking with the - Library. A suitable mechanism is one that (1) uses at run time a - copy of the library already present on the user's computer system, - rather than copying library functions into the executable, and (2) - will operate properly with a modified version of the library, if - the user installs one, as long as the modified version is - interface-compatible with the version that the work was made with. - - c) Accompany the work with a written offer, valid for at - least three years, to give the same user the materials - specified in Subsection 6a, above, for a charge no more - than the cost of performing this distribution. - - d) If distribution of the work is made by offering access to copy - from a designated place, offer equivalent access to copy the above - specified materials from the same place. - - e) Verify that the user has already received a copy of these - materials or that you have already sent this user a copy. - - For an executable, the required form of the "work that uses the -Library" must include any data and utility programs needed for -reproducing the executable from it. However, as a special exception, -the materials to be distributed need not include anything that is -normally distributed (in either source or binary form) with the major -components (compiler, kernel, and so on) of the operating system on -which the executable runs, unless that component itself accompanies -the executable. - - It may happen that this requirement contradicts the license -restrictions of other proprietary libraries that do not normally -accompany the operating system. Such a contradiction means you cannot -use both them and the Library together in an executable that you -distribute. - - 7. You may place library facilities that are a work based on the -Library side-by-side in a single library together with other library -facilities not covered by this License, and distribute such a combined -library, provided that the separate distribution of the work based on -the Library and of the other library facilities is otherwise -permitted, and provided that you do these two things: - - a) Accompany the combined library with a copy of the same work - based on the Library, uncombined with any other library - facilities. This must be distributed under the terms of the - Sections above. - - b) Give prominent notice with the combined library of the fact - that part of it is a work based on the Library, and explaining - where to find the accompanying uncombined form of the same work. - - 8. You may not copy, modify, sublicense, link with, or distribute -the Library except as expressly provided under this License. Any -attempt otherwise to copy, modify, sublicense, link with, or -distribute the Library is void, and will automatically terminate your -rights under this License. However, parties who have received copies, -or rights, from you under this License will not have their licenses -terminated so long as such parties remain in full compliance. - - 9. You are not required to accept this License, since you have not -signed it. However, nothing else grants you permission to modify or -distribute the Library or its derivative works. These actions are -prohibited by law if you do not accept this License. Therefore, by -modifying or distributing the Library (or any work based on the -Library), you indicate your acceptance of this License to do so, and -all its terms and conditions for copying, distributing or modifying -the Library or works based on it. - - 10. Each time you redistribute the Library (or any work based on the -Library), the recipient automatically receives a license from the -original licensor to copy, distribute, link with or modify the Library -subject to these terms and conditions. You may not impose any further -restrictions on the recipients' exercise of the rights granted herein. -You are not responsible for enforcing compliance by third parties with -this License. - - 11. If, as a consequence of a court judgment or allegation of patent -infringement or for any other reason (not limited to patent issues), -conditions are imposed on you (whether by court order, agreement or -otherwise) that contradict the conditions of this License, they do not -excuse you from the conditions of this License. If you cannot -distribute so as to satisfy simultaneously your obligations under this -License and any other pertinent obligations, then as a consequence you -may not distribute the Library at all. For example, if a patent -license would not permit royalty-free redistribution of the Library by -all those who receive copies directly or indirectly through you, then -the only way you could satisfy both it and this License would be to -refrain entirely from distribution of the Library. - -If any portion of this section is held invalid or unenforceable under any -particular circumstance, the balance of the section is intended to apply, -and the section as a whole is intended to apply in other circumstances. - -It is not the purpose of this section to induce you to infringe any -patents or other property right claims or to contest validity of any -such claims; this section has the sole purpose of protecting the -integrity of the free software distribution system which is -implemented by public license practices. Many people have made -generous contributions to the wide range of software distributed -through that system in reliance on consistent application of that -system; it is up to the author/donor to decide if he or she is willing -to distribute software through any other system and a licensee cannot -impose that choice. - -This section is intended to make thoroughly clear what is believed to -be a consequence of the rest of this License. - - 12. If the distribution and/or use of the Library is restricted in -certain countries either by patents or by copyrighted interfaces, the -original copyright holder who places the Library under this License may add -an explicit geographical distribution limitation excluding those countries, -so that distribution is permitted only in or among countries not thus -excluded. In such case, this License incorporates the limitation as if -written in the body of this License. - - 13. The Free Software Foundation may publish revised and/or new -versions of the Lesser General Public License from time to time. -Such new versions will be similar in spirit to the present version, -but may differ in detail to address new problems or concerns. - -Each version is given a distinguishing version number. If the Library -specifies a version number of this License which applies to it and -"any later version", you have the option of following the terms and -conditions either of that version or of any later version published by -the Free Software Foundation. If the Library does not specify a -license version number, you may choose any version ever published by -the Free Software Foundation. - - 14. If you wish to incorporate parts of the Library into other free -programs whose distribution conditions are incompatible with these, -write to the author to ask for permission. For software which is -copyrighted by the Free Software Foundation, write to the Free -Software Foundation; we sometimes make exceptions for this. Our -decision will be guided by the two goals of preserving the free status -of all derivatives of our free software and of promoting the sharing -and reuse of software generally. - - NO WARRANTY - - 15. BECAUSE THE LIBRARY IS LICENSED FREE OF CHARGE, THERE IS NO -WARRANTY FOR THE LIBRARY, TO THE EXTENT PERMITTED BY APPLICABLE LAW. -EXCEPT WHEN OTHERWISE STATED IN WRITING THE COPYRIGHT HOLDERS AND/OR -OTHER PARTIES PROVIDE THE LIBRARY "AS IS" WITHOUT WARRANTY OF ANY -KIND, EITHER EXPRESSED OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE -IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR -PURPOSE. THE ENTIRE RISK AS TO THE QUALITY AND PERFORMANCE OF THE -LIBRARY IS WITH YOU. SHOULD THE LIBRARY PROVE DEFECTIVE, YOU ASSUME -THE COST OF ALL NECESSARY SERVICING, REPAIR OR CORRECTION. - - 16. IN NO EVENT UNLESS REQUIRED BY APPLICABLE LAW OR AGREED TO IN -WRITING WILL ANY COPYRIGHT HOLDER, OR ANY OTHER PARTY WHO MAY MODIFY -AND/OR REDISTRIBUTE THE LIBRARY AS PERMITTED ABOVE, BE LIABLE TO YOU -FOR DAMAGES, INCLUDING ANY GENERAL, SPECIAL, INCIDENTAL OR -CONSEQUENTIAL DAMAGES ARISING OUT OF THE USE OR INABILITY TO USE THE -LIBRARY (INCLUDING BUT NOT LIMITED TO LOSS OF DATA OR DATA BEING -RENDERED INACCURATE OR LOSSES SUSTAINED BY YOU OR THIRD PARTIES OR A -FAILURE OF THE LIBRARY TO OPERATE WITH ANY OTHER SOFTWARE), EVEN IF -SUCH HOLDER OR OTHER PARTY HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH -DAMAGES. - - END OF TERMS AND CONDITIONS - - How to Apply These Terms to Your New Libraries - - If you develop a new library, and you want it to be of the greatest -possible use to the public, we recommend making it free software that -everyone can redistribute and change. You can do so by permitting -redistribution under these terms (or, alternatively, under the terms of the -ordinary General Public License). - - To apply these terms, attach the following notices to the library. It is -safest to attach them to the start of each source file to most effectively -convey the exclusion of warranty; and each file should have at least the -"copyright" line and a pointer to where the full notice is found. - - - Copyright (C) - - This library is free software; you can redistribute it and/or - modify it under the terms of the GNU Lesser General Public - License as published by the Free Software Foundation; either - version 2.1 of the License, or (at your option) any later version. - - This library is distributed in the hope that it will be useful, - but WITHOUT ANY WARRANTY; without even the implied warranty of - MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU - Lesser General Public License for more details. - - You should have received a copy of the GNU Lesser General Public - License along with this library; if not, write to the Free Software - Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA - -Also add information on how to contact you by electronic and paper mail. - -You should also get your employer (if you work as a programmer) or your -school, if any, to sign a "copyright disclaimer" for the library, if -necessary. Here is a sample; alter the names: - - Yoyodyne, Inc., hereby disclaims all copyright interest in the - library `Frob' (a library for tweaking knobs) written by James Random Hacker. - - , 1 April 1990 - Ty Coon, President of Vice - -That's all there is to it! - - diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/__init__.py b/src/fetchcode/vcs/pip/_vendor/chardet/__init__.py deleted file mode 100644 index 0f9f820e..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/__init__.py +++ /dev/null @@ -1,39 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - - -from .compat import PY2, PY3 -from .universaldetector import UniversalDetector -from .version import __version__, VERSION - - -def detect(byte_str): - """ - Detect the encoding of the given byte string. - - :param byte_str: The byte sequence to examine. - :type byte_str: ``bytes`` or ``bytearray`` - """ - if not isinstance(byte_str, bytearray): - if not isinstance(byte_str, bytes): - raise TypeError('Expected object of type bytes or bytearray, got: ' - '{0}'.format(type(byte_str))) - else: - byte_str = bytearray(byte_str) - detector = UniversalDetector() - detector.feed(byte_str) - return detector.close() diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/big5freq.py b/src/fetchcode/vcs/pip/_vendor/chardet/big5freq.py deleted file mode 100644 index 38f32517..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/big5freq.py +++ /dev/null @@ -1,386 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Communicator client code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -# Big5 frequency table -# by Taiwan's Mandarin Promotion Council -# -# -# 128 --> 0.42261 -# 256 --> 0.57851 -# 512 --> 0.74851 -# 1024 --> 0.89384 -# 2048 --> 0.97583 -# -# Ideal Distribution Ratio = 0.74851/(1-0.74851) =2.98 -# Random Distribution Ration = 512/(5401-512)=0.105 -# -# Typical Distribution Ratio about 25% of Ideal one, still much higher than RDR - -BIG5_TYPICAL_DISTRIBUTION_RATIO = 0.75 - -#Char to FreqOrder table -BIG5_TABLE_SIZE = 5376 - -BIG5_CHAR_TO_FREQ_ORDER = ( - 1,1801,1506, 255,1431, 198, 9, 82, 6,5008, 177, 202,3681,1256,2821, 110, # 16 -3814, 33,3274, 261, 76, 44,2114, 16,2946,2187,1176, 659,3971, 26,3451,2653, # 32 -1198,3972,3350,4202, 410,2215, 302, 590, 361,1964, 8, 204, 58,4510,5009,1932, # 48 - 63,5010,5011, 317,1614, 75, 222, 159,4203,2417,1480,5012,3555,3091, 224,2822, # 64 -3682, 3, 10,3973,1471, 29,2787,1135,2866,1940, 873, 130,3275,1123, 312,5013, # 80 -4511,2052, 507, 252, 682,5014, 142,1915, 124, 206,2947, 34,3556,3204, 64, 604, # 96 -5015,2501,1977,1978, 155,1991, 645, 641,1606,5016,3452, 337, 72, 406,5017, 80, # 112 - 630, 238,3205,1509, 263, 939,1092,2654, 756,1440,1094,3453, 449, 69,2987, 591, # 128 - 179,2096, 471, 115,2035,1844, 60, 50,2988, 134, 806,1869, 734,2036,3454, 180, # 144 - 995,1607, 156, 537,2907, 688,5018, 319,1305, 779,2145, 514,2379, 298,4512, 359, # 160 -2502, 90,2716,1338, 663, 11, 906,1099,2553, 20,2441, 182, 532,1716,5019, 732, # 176 -1376,4204,1311,1420,3206, 25,2317,1056, 113, 399, 382,1950, 242,3455,2474, 529, # 192 -3276, 475,1447,3683,5020, 117, 21, 656, 810,1297,2300,2334,3557,5021, 126,4205, # 208 - 706, 456, 150, 613,4513, 71,1118,2037,4206, 145,3092, 85, 835, 486,2115,1246, # 224 -1426, 428, 727,1285,1015, 800, 106, 623, 303,1281,5022,2128,2359, 347,3815, 221, # 240 -3558,3135,5023,1956,1153,4207, 83, 296,1199,3093, 192, 624, 93,5024, 822,1898, # 256 -2823,3136, 795,2065, 991,1554,1542,1592, 27, 43,2867, 859, 139,1456, 860,4514, # 272 - 437, 712,3974, 164,2397,3137, 695, 211,3037,2097, 195,3975,1608,3559,3560,3684, # 288 -3976, 234, 811,2989,2098,3977,2233,1441,3561,1615,2380, 668,2077,1638, 305, 228, # 304 -1664,4515, 467, 415,5025, 262,2099,1593, 239, 108, 300, 200,1033, 512,1247,2078, # 320 -5026,5027,2176,3207,3685,2682, 593, 845,1062,3277, 88,1723,2038,3978,1951, 212, # 336 - 266, 152, 149, 468,1899,4208,4516, 77, 187,5028,3038, 37, 5,2990,5029,3979, # 352 -5030,5031, 39,2524,4517,2908,3208,2079, 55, 148, 74,4518, 545, 483,1474,1029, # 368 -1665, 217,1870,1531,3138,1104,2655,4209, 24, 172,3562, 900,3980,3563,3564,4519, # 384 - 32,1408,2824,1312, 329, 487,2360,2251,2717, 784,2683, 4,3039,3351,1427,1789, # 400 - 188, 109, 499,5032,3686,1717,1790, 888,1217,3040,4520,5033,3565,5034,3352,1520, # 416 -3687,3981, 196,1034, 775,5035,5036, 929,1816, 249, 439, 38,5037,1063,5038, 794, # 432 -3982,1435,2301, 46, 178,3278,2066,5039,2381,5040, 214,1709,4521, 804, 35, 707, # 448 - 324,3688,1601,2554, 140, 459,4210,5041,5042,1365, 839, 272, 978,2262,2580,3456, # 464 -2129,1363,3689,1423, 697, 100,3094, 48, 70,1231, 495,3139,2196,5043,1294,5044, # 480 -2080, 462, 586,1042,3279, 853, 256, 988, 185,2382,3457,1698, 434,1084,5045,3458, # 496 - 314,2625,2788,4522,2335,2336, 569,2285, 637,1817,2525, 757,1162,1879,1616,3459, # 512 - 287,1577,2116, 768,4523,1671,2868,3566,2526,1321,3816, 909,2418,5046,4211, 933, # 528 -3817,4212,2053,2361,1222,4524, 765,2419,1322, 786,4525,5047,1920,1462,1677,2909, # 544 -1699,5048,4526,1424,2442,3140,3690,2600,3353,1775,1941,3460,3983,4213, 309,1369, # 560 -1130,2825, 364,2234,1653,1299,3984,3567,3985,3986,2656, 525,1085,3041, 902,2001, # 576 -1475, 964,4527, 421,1845,1415,1057,2286, 940,1364,3141, 376,4528,4529,1381, 7, # 592 -2527, 983,2383, 336,1710,2684,1846, 321,3461, 559,1131,3042,2752,1809,1132,1313, # 608 - 265,1481,1858,5049, 352,1203,2826,3280, 167,1089, 420,2827, 776, 792,1724,3568, # 624 -4214,2443,3281,5050,4215,5051, 446, 229, 333,2753, 901,3818,1200,1557,4530,2657, # 640 -1921, 395,2754,2685,3819,4216,1836, 125, 916,3209,2626,4531,5052,5053,3820,5054, # 656 -5055,5056,4532,3142,3691,1133,2555,1757,3462,1510,2318,1409,3569,5057,2146, 438, # 672 -2601,2910,2384,3354,1068, 958,3043, 461, 311,2869,2686,4217,1916,3210,4218,1979, # 688 - 383, 750,2755,2627,4219, 274, 539, 385,1278,1442,5058,1154,1965, 384, 561, 210, # 704 - 98,1295,2556,3570,5059,1711,2420,1482,3463,3987,2911,1257, 129,5060,3821, 642, # 720 - 523,2789,2790,2658,5061, 141,2235,1333, 68, 176, 441, 876, 907,4220, 603,2602, # 736 - 710, 171,3464, 404, 549, 18,3143,2398,1410,3692,1666,5062,3571,4533,2912,4534, # 752 -5063,2991, 368,5064, 146, 366, 99, 871,3693,1543, 748, 807,1586,1185, 22,2263, # 768 - 379,3822,3211,5065,3212, 505,1942,2628,1992,1382,2319,5066, 380,2362, 218, 702, # 784 -1818,1248,3465,3044,3572,3355,3282,5067,2992,3694, 930,3283,3823,5068, 59,5069, # 800 - 585, 601,4221, 497,3466,1112,1314,4535,1802,5070,1223,1472,2177,5071, 749,1837, # 816 - 690,1900,3824,1773,3988,1476, 429,1043,1791,2236,2117, 917,4222, 447,1086,1629, # 832 -5072, 556,5073,5074,2021,1654, 844,1090, 105, 550, 966,1758,2828,1008,1783, 686, # 848 -1095,5075,2287, 793,1602,5076,3573,2603,4536,4223,2948,2302,4537,3825, 980,2503, # 864 - 544, 353, 527,4538, 908,2687,2913,5077, 381,2629,1943,1348,5078,1341,1252, 560, # 880 -3095,5079,3467,2870,5080,2054, 973, 886,2081, 143,4539,5081,5082, 157,3989, 496, # 896 -4224, 57, 840, 540,2039,4540,4541,3468,2118,1445, 970,2264,1748,1966,2082,4225, # 912 -3144,1234,1776,3284,2829,3695, 773,1206,2130,1066,2040,1326,3990,1738,1725,4226, # 928 - 279,3145, 51,1544,2604, 423,1578,2131,2067, 173,4542,1880,5083,5084,1583, 264, # 944 - 610,3696,4543,2444, 280, 154,5085,5086,5087,1739, 338,1282,3096, 693,2871,1411, # 960 -1074,3826,2445,5088,4544,5089,5090,1240, 952,2399,5091,2914,1538,2688, 685,1483, # 976 -4227,2475,1436, 953,4228,2055,4545, 671,2400, 79,4229,2446,3285, 608, 567,2689, # 992 -3469,4230,4231,1691, 393,1261,1792,2401,5092,4546,5093,5094,5095,5096,1383,1672, # 1008 -3827,3213,1464, 522,1119, 661,1150, 216, 675,4547,3991,1432,3574, 609,4548,2690, # 1024 -2402,5097,5098,5099,4232,3045, 0,5100,2476, 315, 231,2447, 301,3356,4549,2385, # 1040 -5101, 233,4233,3697,1819,4550,4551,5102, 96,1777,1315,2083,5103, 257,5104,1810, # 1056 -3698,2718,1139,1820,4234,2022,1124,2164,2791,1778,2659,5105,3097, 363,1655,3214, # 1072 -5106,2993,5107,5108,5109,3992,1567,3993, 718, 103,3215, 849,1443, 341,3357,2949, # 1088 -1484,5110,1712, 127, 67, 339,4235,2403, 679,1412, 821,5111,5112, 834, 738, 351, # 1104 -2994,2147, 846, 235,1497,1881, 418,1993,3828,2719, 186,1100,2148,2756,3575,1545, # 1120 -1355,2950,2872,1377, 583,3994,4236,2581,2995,5113,1298,3699,1078,2557,3700,2363, # 1136 - 78,3829,3830, 267,1289,2100,2002,1594,4237, 348, 369,1274,2197,2178,1838,4552, # 1152 -1821,2830,3701,2757,2288,2003,4553,2951,2758, 144,3358, 882,4554,3995,2759,3470, # 1168 -4555,2915,5114,4238,1726, 320,5115,3996,3046, 788,2996,5116,2831,1774,1327,2873, # 1184 -3997,2832,5117,1306,4556,2004,1700,3831,3576,2364,2660, 787,2023, 506, 824,3702, # 1200 - 534, 323,4557,1044,3359,2024,1901, 946,3471,5118,1779,1500,1678,5119,1882,4558, # 1216 - 165, 243,4559,3703,2528, 123, 683,4239, 764,4560, 36,3998,1793, 589,2916, 816, # 1232 - 626,1667,3047,2237,1639,1555,1622,3832,3999,5120,4000,2874,1370,1228,1933, 891, # 1248 -2084,2917, 304,4240,5121, 292,2997,2720,3577, 691,2101,4241,1115,4561, 118, 662, # 1264 -5122, 611,1156, 854,2386,1316,2875, 2, 386, 515,2918,5123,5124,3286, 868,2238, # 1280 -1486, 855,2661, 785,2216,3048,5125,1040,3216,3578,5126,3146, 448,5127,1525,5128, # 1296 -2165,4562,5129,3833,5130,4242,2833,3579,3147, 503, 818,4001,3148,1568, 814, 676, # 1312 -1444, 306,1749,5131,3834,1416,1030, 197,1428, 805,2834,1501,4563,5132,5133,5134, # 1328 -1994,5135,4564,5136,5137,2198, 13,2792,3704,2998,3149,1229,1917,5138,3835,2132, # 1344 -5139,4243,4565,2404,3580,5140,2217,1511,1727,1120,5141,5142, 646,3836,2448, 307, # 1360 -5143,5144,1595,3217,5145,5146,5147,3705,1113,1356,4002,1465,2529,2530,5148, 519, # 1376 -5149, 128,2133, 92,2289,1980,5150,4003,1512, 342,3150,2199,5151,2793,2218,1981, # 1392 -3360,4244, 290,1656,1317, 789, 827,2365,5152,3837,4566, 562, 581,4004,5153, 401, # 1408 -4567,2252, 94,4568,5154,1399,2794,5155,1463,2025,4569,3218,1944,5156, 828,1105, # 1424 -4245,1262,1394,5157,4246, 605,4570,5158,1784,2876,5159,2835, 819,2102, 578,2200, # 1440 -2952,5160,1502, 436,3287,4247,3288,2836,4005,2919,3472,3473,5161,2721,2320,5162, # 1456 -5163,2337,2068, 23,4571, 193, 826,3838,2103, 699,1630,4248,3098, 390,1794,1064, # 1472 -3581,5164,1579,3099,3100,1400,5165,4249,1839,1640,2877,5166,4572,4573, 137,4250, # 1488 - 598,3101,1967, 780, 104, 974,2953,5167, 278, 899, 253, 402, 572, 504, 493,1339, # 1504 -5168,4006,1275,4574,2582,2558,5169,3706,3049,3102,2253, 565,1334,2722, 863, 41, # 1520 -5170,5171,4575,5172,1657,2338, 19, 463,2760,4251, 606,5173,2999,3289,1087,2085, # 1536 -1323,2662,3000,5174,1631,1623,1750,4252,2691,5175,2878, 791,2723,2663,2339, 232, # 1552 -2421,5176,3001,1498,5177,2664,2630, 755,1366,3707,3290,3151,2026,1609, 119,1918, # 1568 -3474, 862,1026,4253,5178,4007,3839,4576,4008,4577,2265,1952,2477,5179,1125, 817, # 1584 -4254,4255,4009,1513,1766,2041,1487,4256,3050,3291,2837,3840,3152,5180,5181,1507, # 1600 -5182,2692, 733, 40,1632,1106,2879, 345,4257, 841,2531, 230,4578,3002,1847,3292, # 1616 -3475,5183,1263, 986,3476,5184, 735, 879, 254,1137, 857, 622,1300,1180,1388,1562, # 1632 -4010,4011,2954, 967,2761,2665,1349, 592,2134,1692,3361,3003,1995,4258,1679,4012, # 1648 -1902,2188,5185, 739,3708,2724,1296,1290,5186,4259,2201,2202,1922,1563,2605,2559, # 1664 -1871,2762,3004,5187, 435,5188, 343,1108, 596, 17,1751,4579,2239,3477,3709,5189, # 1680 -4580, 294,3582,2955,1693, 477, 979, 281,2042,3583, 643,2043,3710,2631,2795,2266, # 1696 -1031,2340,2135,2303,3584,4581, 367,1249,2560,5190,3585,5191,4582,1283,3362,2005, # 1712 - 240,1762,3363,4583,4584, 836,1069,3153, 474,5192,2149,2532, 268,3586,5193,3219, # 1728 -1521,1284,5194,1658,1546,4260,5195,3587,3588,5196,4261,3364,2693,1685,4262, 961, # 1744 -1673,2632, 190,2006,2203,3841,4585,4586,5197, 570,2504,3711,1490,5198,4587,2633, # 1760 -3293,1957,4588, 584,1514, 396,1045,1945,5199,4589,1968,2449,5200,5201,4590,4013, # 1776 - 619,5202,3154,3294, 215,2007,2796,2561,3220,4591,3221,4592, 763,4263,3842,4593, # 1792 -5203,5204,1958,1767,2956,3365,3712,1174, 452,1477,4594,3366,3155,5205,2838,1253, # 1808 -2387,2189,1091,2290,4264, 492,5206, 638,1169,1825,2136,1752,4014, 648, 926,1021, # 1824 -1324,4595, 520,4596, 997, 847,1007, 892,4597,3843,2267,1872,3713,2405,1785,4598, # 1840 -1953,2957,3103,3222,1728,4265,2044,3714,4599,2008,1701,3156,1551, 30,2268,4266, # 1856 -5207,2027,4600,3589,5208, 501,5209,4267, 594,3478,2166,1822,3590,3479,3591,3223, # 1872 - 829,2839,4268,5210,1680,3157,1225,4269,5211,3295,4601,4270,3158,2341,5212,4602, # 1888 -4271,5213,4015,4016,5214,1848,2388,2606,3367,5215,4603, 374,4017, 652,4272,4273, # 1904 - 375,1140, 798,5216,5217,5218,2366,4604,2269, 546,1659, 138,3051,2450,4605,5219, # 1920 -2254, 612,1849, 910, 796,3844,1740,1371, 825,3845,3846,5220,2920,2562,5221, 692, # 1936 - 444,3052,2634, 801,4606,4274,5222,1491, 244,1053,3053,4275,4276, 340,5223,4018, # 1952 -1041,3005, 293,1168, 87,1357,5224,1539, 959,5225,2240, 721, 694,4277,3847, 219, # 1968 -1478, 644,1417,3368,2666,1413,1401,1335,1389,4019,5226,5227,3006,2367,3159,1826, # 1984 - 730,1515, 184,2840, 66,4607,5228,1660,2958, 246,3369, 378,1457, 226,3480, 975, # 2000 -4020,2959,1264,3592, 674, 696,5229, 163,5230,1141,2422,2167, 713,3593,3370,4608, # 2016 -4021,5231,5232,1186, 15,5233,1079,1070,5234,1522,3224,3594, 276,1050,2725, 758, # 2032 -1126, 653,2960,3296,5235,2342, 889,3595,4022,3104,3007, 903,1250,4609,4023,3481, # 2048 -3596,1342,1681,1718, 766,3297, 286, 89,2961,3715,5236,1713,5237,2607,3371,3008, # 2064 -5238,2962,2219,3225,2880,5239,4610,2505,2533, 181, 387,1075,4024, 731,2190,3372, # 2080 -5240,3298, 310, 313,3482,2304, 770,4278, 54,3054, 189,4611,3105,3848,4025,5241, # 2096 -1230,1617,1850, 355,3597,4279,4612,3373, 111,4280,3716,1350,3160,3483,3055,4281, # 2112 -2150,3299,3598,5242,2797,4026,4027,3009, 722,2009,5243,1071, 247,1207,2343,2478, # 2128 -1378,4613,2010, 864,1437,1214,4614, 373,3849,1142,2220, 667,4615, 442,2763,2563, # 2144 -3850,4028,1969,4282,3300,1840, 837, 170,1107, 934,1336,1883,5244,5245,2119,4283, # 2160 -2841, 743,1569,5246,4616,4284, 582,2389,1418,3484,5247,1803,5248, 357,1395,1729, # 2176 -3717,3301,2423,1564,2241,5249,3106,3851,1633,4617,1114,2086,4285,1532,5250, 482, # 2192 -2451,4618,5251,5252,1492, 833,1466,5253,2726,3599,1641,2842,5254,1526,1272,3718, # 2208 -4286,1686,1795, 416,2564,1903,1954,1804,5255,3852,2798,3853,1159,2321,5256,2881, # 2224 -4619,1610,1584,3056,2424,2764, 443,3302,1163,3161,5257,5258,4029,5259,4287,2506, # 2240 -3057,4620,4030,3162,2104,1647,3600,2011,1873,4288,5260,4289, 431,3485,5261, 250, # 2256 - 97, 81,4290,5262,1648,1851,1558, 160, 848,5263, 866, 740,1694,5264,2204,2843, # 2272 -3226,4291,4621,3719,1687, 950,2479, 426, 469,3227,3720,3721,4031,5265,5266,1188, # 2288 - 424,1996, 861,3601,4292,3854,2205,2694, 168,1235,3602,4293,5267,2087,1674,4622, # 2304 -3374,3303, 220,2565,1009,5268,3855, 670,3010, 332,1208, 717,5269,5270,3603,2452, # 2320 -4032,3375,5271, 513,5272,1209,2882,3376,3163,4623,1080,5273,5274,5275,5276,2534, # 2336 -3722,3604, 815,1587,4033,4034,5277,3605,3486,3856,1254,4624,1328,3058,1390,4035, # 2352 -1741,4036,3857,4037,5278, 236,3858,2453,3304,5279,5280,3723,3859,1273,3860,4625, # 2368 -5281, 308,5282,4626, 245,4627,1852,2480,1307,2583, 430, 715,2137,2454,5283, 270, # 2384 - 199,2883,4038,5284,3606,2727,1753, 761,1754, 725,1661,1841,4628,3487,3724,5285, # 2400 -5286, 587, 14,3305, 227,2608, 326, 480,2270, 943,2765,3607, 291, 650,1884,5287, # 2416 -1702,1226, 102,1547, 62,3488, 904,4629,3489,1164,4294,5288,5289,1224,1548,2766, # 2432 - 391, 498,1493,5290,1386,1419,5291,2056,1177,4630, 813, 880,1081,2368, 566,1145, # 2448 -4631,2291,1001,1035,2566,2609,2242, 394,1286,5292,5293,2069,5294, 86,1494,1730, # 2464 -4039, 491,1588, 745, 897,2963, 843,3377,4040,2767,2884,3306,1768, 998,2221,2070, # 2480 - 397,1827,1195,1970,3725,3011,3378, 284,5295,3861,2507,2138,2120,1904,5296,4041, # 2496 -2151,4042,4295,1036,3490,1905, 114,2567,4296, 209,1527,5297,5298,2964,2844,2635, # 2512 -2390,2728,3164, 812,2568,5299,3307,5300,1559, 737,1885,3726,1210, 885, 28,2695, # 2528 -3608,3862,5301,4297,1004,1780,4632,5302, 346,1982,2222,2696,4633,3863,1742, 797, # 2544 -1642,4043,1934,1072,1384,2152, 896,4044,3308,3727,3228,2885,3609,5303,2569,1959, # 2560 -4634,2455,1786,5304,5305,5306,4045,4298,1005,1308,3728,4299,2729,4635,4636,1528, # 2576 -2610, 161,1178,4300,1983, 987,4637,1101,4301, 631,4046,1157,3229,2425,1343,1241, # 2592 -1016,2243,2570, 372, 877,2344,2508,1160, 555,1935, 911,4047,5307, 466,1170, 169, # 2608 -1051,2921,2697,3729,2481,3012,1182,2012,2571,1251,2636,5308, 992,2345,3491,1540, # 2624 -2730,1201,2071,2406,1997,2482,5309,4638, 528,1923,2191,1503,1874,1570,2369,3379, # 2640 -3309,5310, 557,1073,5311,1828,3492,2088,2271,3165,3059,3107, 767,3108,2799,4639, # 2656 -1006,4302,4640,2346,1267,2179,3730,3230, 778,4048,3231,2731,1597,2667,5312,4641, # 2672 -5313,3493,5314,5315,5316,3310,2698,1433,3311, 131, 95,1504,4049, 723,4303,3166, # 2688 -1842,3610,2768,2192,4050,2028,2105,3731,5317,3013,4051,1218,5318,3380,3232,4052, # 2704 -4304,2584, 248,1634,3864, 912,5319,2845,3732,3060,3865, 654, 53,5320,3014,5321, # 2720 -1688,4642, 777,3494,1032,4053,1425,5322, 191, 820,2121,2846, 971,4643, 931,3233, # 2736 - 135, 664, 783,3866,1998, 772,2922,1936,4054,3867,4644,2923,3234, 282,2732, 640, # 2752 -1372,3495,1127, 922, 325,3381,5323,5324, 711,2045,5325,5326,4055,2223,2800,1937, # 2768 -4056,3382,2224,2255,3868,2305,5327,4645,3869,1258,3312,4057,3235,2139,2965,4058, # 2784 -4059,5328,2225, 258,3236,4646, 101,1227,5329,3313,1755,5330,1391,3314,5331,2924, # 2800 -2057, 893,5332,5333,5334,1402,4305,2347,5335,5336,3237,3611,5337,5338, 878,1325, # 2816 -1781,2801,4647, 259,1385,2585, 744,1183,2272,4648,5339,4060,2509,5340, 684,1024, # 2832 -4306,5341, 472,3612,3496,1165,3315,4061,4062, 322,2153, 881, 455,1695,1152,1340, # 2848 - 660, 554,2154,4649,1058,4650,4307, 830,1065,3383,4063,4651,1924,5342,1703,1919, # 2864 -5343, 932,2273, 122,5344,4652, 947, 677,5345,3870,2637, 297,1906,1925,2274,4653, # 2880 -2322,3316,5346,5347,4308,5348,4309, 84,4310, 112, 989,5349, 547,1059,4064, 701, # 2896 -3613,1019,5350,4311,5351,3497, 942, 639, 457,2306,2456, 993,2966, 407, 851, 494, # 2912 -4654,3384, 927,5352,1237,5353,2426,3385, 573,4312, 680, 921,2925,1279,1875, 285, # 2928 - 790,1448,1984, 719,2168,5354,5355,4655,4065,4066,1649,5356,1541, 563,5357,1077, # 2944 -5358,3386,3061,3498, 511,3015,4067,4068,3733,4069,1268,2572,3387,3238,4656,4657, # 2960 -5359, 535,1048,1276,1189,2926,2029,3167,1438,1373,2847,2967,1134,2013,5360,4313, # 2976 -1238,2586,3109,1259,5361, 700,5362,2968,3168,3734,4314,5363,4315,1146,1876,1907, # 2992 -4658,2611,4070, 781,2427, 132,1589, 203, 147, 273,2802,2407, 898,1787,2155,4071, # 3008 -4072,5364,3871,2803,5365,5366,4659,4660,5367,3239,5368,1635,3872, 965,5369,1805, # 3024 -2699,1516,3614,1121,1082,1329,3317,4073,1449,3873, 65,1128,2848,2927,2769,1590, # 3040 -3874,5370,5371, 12,2668, 45, 976,2587,3169,4661, 517,2535,1013,1037,3240,5372, # 3056 -3875,2849,5373,3876,5374,3499,5375,2612, 614,1999,2323,3877,3110,2733,2638,5376, # 3072 -2588,4316, 599,1269,5377,1811,3735,5378,2700,3111, 759,1060, 489,1806,3388,3318, # 3088 -1358,5379,5380,2391,1387,1215,2639,2256, 490,5381,5382,4317,1759,2392,2348,5383, # 3104 -4662,3878,1908,4074,2640,1807,3241,4663,3500,3319,2770,2349, 874,5384,5385,3501, # 3120 -3736,1859, 91,2928,3737,3062,3879,4664,5386,3170,4075,2669,5387,3502,1202,1403, # 3136 -3880,2969,2536,1517,2510,4665,3503,2511,5388,4666,5389,2701,1886,1495,1731,4076, # 3152 -2370,4667,5390,2030,5391,5392,4077,2702,1216, 237,2589,4318,2324,4078,3881,4668, # 3168 -4669,2703,3615,3504, 445,4670,5393,5394,5395,5396,2771, 61,4079,3738,1823,4080, # 3184 -5397, 687,2046, 935, 925, 405,2670, 703,1096,1860,2734,4671,4081,1877,1367,2704, # 3200 -3389, 918,2106,1782,2483, 334,3320,1611,1093,4672, 564,3171,3505,3739,3390, 945, # 3216 -2641,2058,4673,5398,1926, 872,4319,5399,3506,2705,3112, 349,4320,3740,4082,4674, # 3232 -3882,4321,3741,2156,4083,4675,4676,4322,4677,2408,2047, 782,4084, 400, 251,4323, # 3248 -1624,5400,5401, 277,3742, 299,1265, 476,1191,3883,2122,4324,4325,1109, 205,5402, # 3264 -2590,1000,2157,3616,1861,5403,5404,5405,4678,5406,4679,2573, 107,2484,2158,4085, # 3280 -3507,3172,5407,1533, 541,1301, 158, 753,4326,2886,3617,5408,1696, 370,1088,4327, # 3296 -4680,3618, 579, 327, 440, 162,2244, 269,1938,1374,3508, 968,3063, 56,1396,3113, # 3312 -2107,3321,3391,5409,1927,2159,4681,3016,5410,3619,5411,5412,3743,4682,2485,5413, # 3328 -2804,5414,1650,4683,5415,2613,5416,5417,4086,2671,3392,1149,3393,4087,3884,4088, # 3344 -5418,1076, 49,5419, 951,3242,3322,3323, 450,2850, 920,5420,1812,2805,2371,4328, # 3360 -1909,1138,2372,3885,3509,5421,3243,4684,1910,1147,1518,2428,4685,3886,5422,4686, # 3376 -2393,2614, 260,1796,3244,5423,5424,3887,3324, 708,5425,3620,1704,5426,3621,1351, # 3392 -1618,3394,3017,1887, 944,4329,3395,4330,3064,3396,4331,5427,3744, 422, 413,1714, # 3408 -3325, 500,2059,2350,4332,2486,5428,1344,1911, 954,5429,1668,5430,5431,4089,2409, # 3424 -4333,3622,3888,4334,5432,2307,1318,2512,3114, 133,3115,2887,4687, 629, 31,2851, # 3440 -2706,3889,4688, 850, 949,4689,4090,2970,1732,2089,4335,1496,1853,5433,4091, 620, # 3456 -3245, 981,1242,3745,3397,1619,3746,1643,3326,2140,2457,1971,1719,3510,2169,5434, # 3472 -3246,5435,5436,3398,1829,5437,1277,4690,1565,2048,5438,1636,3623,3116,5439, 869, # 3488 -2852, 655,3890,3891,3117,4092,3018,3892,1310,3624,4691,5440,5441,5442,1733, 558, # 3504 -4692,3747, 335,1549,3065,1756,4336,3748,1946,3511,1830,1291,1192, 470,2735,2108, # 3520 -2806, 913,1054,4093,5443,1027,5444,3066,4094,4693, 982,2672,3399,3173,3512,3247, # 3536 -3248,1947,2807,5445, 571,4694,5446,1831,5447,3625,2591,1523,2429,5448,2090, 984, # 3552 -4695,3749,1960,5449,3750, 852, 923,2808,3513,3751, 969,1519, 999,2049,2325,1705, # 3568 -5450,3118, 615,1662, 151, 597,4095,2410,2326,1049, 275,4696,3752,4337, 568,3753, # 3584 -3626,2487,4338,3754,5451,2430,2275, 409,3249,5452,1566,2888,3514,1002, 769,2853, # 3600 - 194,2091,3174,3755,2226,3327,4339, 628,1505,5453,5454,1763,2180,3019,4096, 521, # 3616 -1161,2592,1788,2206,2411,4697,4097,1625,4340,4341, 412, 42,3119, 464,5455,2642, # 3632 -4698,3400,1760,1571,2889,3515,2537,1219,2207,3893,2643,2141,2373,4699,4700,3328, # 3648 -1651,3401,3627,5456,5457,3628,2488,3516,5458,3756,5459,5460,2276,2092, 460,5461, # 3664 -4701,5462,3020, 962, 588,3629, 289,3250,2644,1116, 52,5463,3067,1797,5464,5465, # 3680 -5466,1467,5467,1598,1143,3757,4342,1985,1734,1067,4702,1280,3402, 465,4703,1572, # 3696 - 510,5468,1928,2245,1813,1644,3630,5469,4704,3758,5470,5471,2673,1573,1534,5472, # 3712 -5473, 536,1808,1761,3517,3894,3175,2645,5474,5475,5476,4705,3518,2929,1912,2809, # 3728 -5477,3329,1122, 377,3251,5478, 360,5479,5480,4343,1529, 551,5481,2060,3759,1769, # 3744 -2431,5482,2930,4344,3330,3120,2327,2109,2031,4706,1404, 136,1468,1479, 672,1171, # 3760 -3252,2308, 271,3176,5483,2772,5484,2050, 678,2736, 865,1948,4707,5485,2014,4098, # 3776 -2971,5486,2737,2227,1397,3068,3760,4708,4709,1735,2931,3403,3631,5487,3895, 509, # 3792 -2854,2458,2890,3896,5488,5489,3177,3178,4710,4345,2538,4711,2309,1166,1010, 552, # 3808 - 681,1888,5490,5491,2972,2973,4099,1287,1596,1862,3179, 358, 453, 736, 175, 478, # 3824 -1117, 905,1167,1097,5492,1854,1530,5493,1706,5494,2181,3519,2292,3761,3520,3632, # 3840 -4346,2093,4347,5495,3404,1193,2489,4348,1458,2193,2208,1863,1889,1421,3331,2932, # 3856 -3069,2182,3521, 595,2123,5496,4100,5497,5498,4349,1707,2646, 223,3762,1359, 751, # 3872 -3121, 183,3522,5499,2810,3021, 419,2374, 633, 704,3897,2394, 241,5500,5501,5502, # 3888 - 838,3022,3763,2277,2773,2459,3898,1939,2051,4101,1309,3122,2246,1181,5503,1136, # 3904 -2209,3899,2375,1446,4350,2310,4712,5504,5505,4351,1055,2615, 484,3764,5506,4102, # 3920 - 625,4352,2278,3405,1499,4353,4103,5507,4104,4354,3253,2279,2280,3523,5508,5509, # 3936 -2774, 808,2616,3765,3406,4105,4355,3123,2539, 526,3407,3900,4356, 955,5510,1620, # 3952 -4357,2647,2432,5511,1429,3766,1669,1832, 994, 928,5512,3633,1260,5513,5514,5515, # 3968 -1949,2293, 741,2933,1626,4358,2738,2460, 867,1184, 362,3408,1392,5516,5517,4106, # 3984 -4359,1770,1736,3254,2934,4713,4714,1929,2707,1459,1158,5518,3070,3409,2891,1292, # 4000 -1930,2513,2855,3767,1986,1187,2072,2015,2617,4360,5519,2574,2514,2170,3768,2490, # 4016 -3332,5520,3769,4715,5521,5522, 666,1003,3023,1022,3634,4361,5523,4716,1814,2257, # 4032 - 574,3901,1603, 295,1535, 705,3902,4362, 283, 858, 417,5524,5525,3255,4717,4718, # 4048 -3071,1220,1890,1046,2281,2461,4107,1393,1599, 689,2575, 388,4363,5526,2491, 802, # 4064 -5527,2811,3903,2061,1405,2258,5528,4719,3904,2110,1052,1345,3256,1585,5529, 809, # 4080 -5530,5531,5532, 575,2739,3524, 956,1552,1469,1144,2328,5533,2329,1560,2462,3635, # 4096 -3257,4108, 616,2210,4364,3180,2183,2294,5534,1833,5535,3525,4720,5536,1319,3770, # 4112 -3771,1211,3636,1023,3258,1293,2812,5537,5538,5539,3905, 607,2311,3906, 762,2892, # 4128 -1439,4365,1360,4721,1485,3072,5540,4722,1038,4366,1450,2062,2648,4367,1379,4723, # 4144 -2593,5541,5542,4368,1352,1414,2330,2935,1172,5543,5544,3907,3908,4724,1798,1451, # 4160 -5545,5546,5547,5548,2936,4109,4110,2492,2351, 411,4111,4112,3637,3333,3124,4725, # 4176 -1561,2674,1452,4113,1375,5549,5550, 47,2974, 316,5551,1406,1591,2937,3181,5552, # 4192 -1025,2142,3125,3182, 354,2740, 884,2228,4369,2412, 508,3772, 726,3638, 996,2433, # 4208 -3639, 729,5553, 392,2194,1453,4114,4726,3773,5554,5555,2463,3640,2618,1675,2813, # 4224 - 919,2352,2975,2353,1270,4727,4115, 73,5556,5557, 647,5558,3259,2856,2259,1550, # 4240 -1346,3024,5559,1332, 883,3526,5560,5561,5562,5563,3334,2775,5564,1212, 831,1347, # 4256 -4370,4728,2331,3909,1864,3073, 720,3910,4729,4730,3911,5565,4371,5566,5567,4731, # 4272 -5568,5569,1799,4732,3774,2619,4733,3641,1645,2376,4734,5570,2938, 669,2211,2675, # 4288 -2434,5571,2893,5572,5573,1028,3260,5574,4372,2413,5575,2260,1353,5576,5577,4735, # 4304 -3183, 518,5578,4116,5579,4373,1961,5580,2143,4374,5581,5582,3025,2354,2355,3912, # 4320 - 516,1834,1454,4117,2708,4375,4736,2229,2620,1972,1129,3642,5583,2776,5584,2976, # 4336 -1422, 577,1470,3026,1524,3410,5585,5586, 432,4376,3074,3527,5587,2594,1455,2515, # 4352 -2230,1973,1175,5588,1020,2741,4118,3528,4737,5589,2742,5590,1743,1361,3075,3529, # 4368 -2649,4119,4377,4738,2295, 895, 924,4378,2171, 331,2247,3076, 166,1627,3077,1098, # 4384 -5591,1232,2894,2231,3411,4739, 657, 403,1196,2377, 542,3775,3412,1600,4379,3530, # 4400 -5592,4740,2777,3261, 576, 530,1362,4741,4742,2540,2676,3776,4120,5593, 842,3913, # 4416 -5594,2814,2032,1014,4121, 213,2709,3413, 665, 621,4380,5595,3777,2939,2435,5596, # 4432 -2436,3335,3643,3414,4743,4381,2541,4382,4744,3644,1682,4383,3531,1380,5597, 724, # 4448 -2282, 600,1670,5598,1337,1233,4745,3126,2248,5599,1621,4746,5600, 651,4384,5601, # 4464 -1612,4385,2621,5602,2857,5603,2743,2312,3078,5604, 716,2464,3079, 174,1255,2710, # 4480 -4122,3645, 548,1320,1398, 728,4123,1574,5605,1891,1197,3080,4124,5606,3081,3082, # 4496 -3778,3646,3779, 747,5607, 635,4386,4747,5608,5609,5610,4387,5611,5612,4748,5613, # 4512 -3415,4749,2437, 451,5614,3780,2542,2073,4388,2744,4389,4125,5615,1764,4750,5616, # 4528 -4390, 350,4751,2283,2395,2493,5617,4391,4126,2249,1434,4127, 488,4752, 458,4392, # 4544 -4128,3781, 771,1330,2396,3914,2576,3184,2160,2414,1553,2677,3185,4393,5618,2494, # 4560 -2895,2622,1720,2711,4394,3416,4753,5619,2543,4395,5620,3262,4396,2778,5621,2016, # 4576 -2745,5622,1155,1017,3782,3915,5623,3336,2313, 201,1865,4397,1430,5624,4129,5625, # 4592 -5626,5627,5628,5629,4398,1604,5630, 414,1866, 371,2595,4754,4755,3532,2017,3127, # 4608 -4756,1708, 960,4399, 887, 389,2172,1536,1663,1721,5631,2232,4130,2356,2940,1580, # 4624 -5632,5633,1744,4757,2544,4758,4759,5634,4760,5635,2074,5636,4761,3647,3417,2896, # 4640 -4400,5637,4401,2650,3418,2815, 673,2712,2465, 709,3533,4131,3648,4402,5638,1148, # 4656 - 502, 634,5639,5640,1204,4762,3649,1575,4763,2623,3783,5641,3784,3128, 948,3263, # 4672 - 121,1745,3916,1110,5642,4403,3083,2516,3027,4132,3785,1151,1771,3917,1488,4133, # 4688 -1987,5643,2438,3534,5644,5645,2094,5646,4404,3918,1213,1407,2816, 531,2746,2545, # 4704 -3264,1011,1537,4764,2779,4405,3129,1061,5647,3786,3787,1867,2897,5648,2018, 120, # 4720 -4406,4407,2063,3650,3265,2314,3919,2678,3419,1955,4765,4134,5649,3535,1047,2713, # 4736 -1266,5650,1368,4766,2858, 649,3420,3920,2546,2747,1102,2859,2679,5651,5652,2000, # 4752 -5653,1111,3651,2977,5654,2495,3921,3652,2817,1855,3421,3788,5655,5656,3422,2415, # 4768 -2898,3337,3266,3653,5657,2577,5658,3654,2818,4135,1460, 856,5659,3655,5660,2899, # 4784 -2978,5661,2900,3922,5662,4408, 632,2517, 875,3923,1697,3924,2296,5663,5664,4767, # 4800 -3028,1239, 580,4768,4409,5665, 914, 936,2075,1190,4136,1039,2124,5666,5667,5668, # 4816 -5669,3423,1473,5670,1354,4410,3925,4769,2173,3084,4137, 915,3338,4411,4412,3339, # 4832 -1605,1835,5671,2748, 398,3656,4413,3926,4138, 328,1913,2860,4139,3927,1331,4414, # 4848 -3029, 937,4415,5672,3657,4140,4141,3424,2161,4770,3425, 524, 742, 538,3085,1012, # 4864 -5673,5674,3928,2466,5675, 658,1103, 225,3929,5676,5677,4771,5678,4772,5679,3267, # 4880 -1243,5680,4142, 963,2250,4773,5681,2714,3658,3186,5682,5683,2596,2332,5684,4774, # 4896 -5685,5686,5687,3536, 957,3426,2547,2033,1931,2941,2467, 870,2019,3659,1746,2780, # 4912 -2781,2439,2468,5688,3930,5689,3789,3130,3790,3537,3427,3791,5690,1179,3086,5691, # 4928 -3187,2378,4416,3792,2548,3188,3131,2749,4143,5692,3428,1556,2549,2297, 977,2901, # 4944 -2034,4144,1205,3429,5693,1765,3430,3189,2125,1271, 714,1689,4775,3538,5694,2333, # 4960 -3931, 533,4417,3660,2184, 617,5695,2469,3340,3539,2315,5696,5697,3190,5698,5699, # 4976 -3932,1988, 618, 427,2651,3540,3431,5700,5701,1244,1690,5702,2819,4418,4776,5703, # 4992 -3541,4777,5704,2284,1576, 473,3661,4419,3432, 972,5705,3662,5706,3087,5707,5708, # 5008 -4778,4779,5709,3793,4145,4146,5710, 153,4780, 356,5711,1892,2902,4420,2144, 408, # 5024 - 803,2357,5712,3933,5713,4421,1646,2578,2518,4781,4782,3934,5714,3935,4422,5715, # 5040 -2416,3433, 752,5716,5717,1962,3341,2979,5718, 746,3030,2470,4783,4423,3794, 698, # 5056 -4784,1893,4424,3663,2550,4785,3664,3936,5719,3191,3434,5720,1824,1302,4147,2715, # 5072 -3937,1974,4425,5721,4426,3192, 823,1303,1288,1236,2861,3542,4148,3435, 774,3938, # 5088 -5722,1581,4786,1304,2862,3939,4787,5723,2440,2162,1083,3268,4427,4149,4428, 344, # 5104 -1173, 288,2316, 454,1683,5724,5725,1461,4788,4150,2597,5726,5727,4789, 985, 894, # 5120 -5728,3436,3193,5729,1914,2942,3795,1989,5730,2111,1975,5731,4151,5732,2579,1194, # 5136 - 425,5733,4790,3194,1245,3796,4429,5734,5735,2863,5736, 636,4791,1856,3940, 760, # 5152 -1800,5737,4430,2212,1508,4792,4152,1894,1684,2298,5738,5739,4793,4431,4432,2213, # 5168 - 479,5740,5741, 832,5742,4153,2496,5743,2980,2497,3797, 990,3132, 627,1815,2652, # 5184 -4433,1582,4434,2126,2112,3543,4794,5744, 799,4435,3195,5745,4795,2113,1737,3031, # 5200 -1018, 543, 754,4436,3342,1676,4796,4797,4154,4798,1489,5746,3544,5747,2624,2903, # 5216 -4155,5748,5749,2981,5750,5751,5752,5753,3196,4799,4800,2185,1722,5754,3269,3270, # 5232 -1843,3665,1715, 481, 365,1976,1857,5755,5756,1963,2498,4801,5757,2127,3666,3271, # 5248 - 433,1895,2064,2076,5758, 602,2750,5759,5760,5761,5762,5763,3032,1628,3437,5764, # 5264 -3197,4802,4156,2904,4803,2519,5765,2551,2782,5766,5767,5768,3343,4804,2905,5769, # 5280 -4805,5770,2864,4806,4807,1221,2982,4157,2520,5771,5772,5773,1868,1990,5774,5775, # 5296 -5776,1896,5777,5778,4808,1897,4158, 318,5779,2095,4159,4437,5780,5781, 485,5782, # 5312 - 938,3941, 553,2680, 116,5783,3942,3667,5784,3545,2681,2783,3438,3344,2820,5785, # 5328 -3668,2943,4160,1747,2944,2983,5786,5787, 207,5788,4809,5789,4810,2521,5790,3033, # 5344 - 890,3669,3943,5791,1878,3798,3439,5792,2186,2358,3440,1652,5793,5794,5795, 941, # 5360 -2299, 208,3546,4161,2020, 330,4438,3944,2906,2499,3799,4439,4811,5796,5797,5798, # 5376 -) - diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/big5prober.py b/src/fetchcode/vcs/pip/_vendor/chardet/big5prober.py deleted file mode 100644 index 98f99701..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/big5prober.py +++ /dev/null @@ -1,47 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Communicator client code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -from .mbcharsetprober import MultiByteCharSetProber -from .codingstatemachine import CodingStateMachine -from .chardistribution import Big5DistributionAnalysis -from .mbcssm import BIG5_SM_MODEL - - -class Big5Prober(MultiByteCharSetProber): - def __init__(self): - super(Big5Prober, self).__init__() - self.coding_sm = CodingStateMachine(BIG5_SM_MODEL) - self.distribution_analyzer = Big5DistributionAnalysis() - self.reset() - - @property - def charset_name(self): - return "Big5" - - @property - def language(self): - return "Chinese" diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/chardistribution.py b/src/fetchcode/vcs/pip/_vendor/chardet/chardistribution.py deleted file mode 100644 index c0395f4a..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/chardistribution.py +++ /dev/null @@ -1,233 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Communicator client code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -from .euctwfreq import (EUCTW_CHAR_TO_FREQ_ORDER, EUCTW_TABLE_SIZE, - EUCTW_TYPICAL_DISTRIBUTION_RATIO) -from .euckrfreq import (EUCKR_CHAR_TO_FREQ_ORDER, EUCKR_TABLE_SIZE, - EUCKR_TYPICAL_DISTRIBUTION_RATIO) -from .gb2312freq import (GB2312_CHAR_TO_FREQ_ORDER, GB2312_TABLE_SIZE, - GB2312_TYPICAL_DISTRIBUTION_RATIO) -from .big5freq import (BIG5_CHAR_TO_FREQ_ORDER, BIG5_TABLE_SIZE, - BIG5_TYPICAL_DISTRIBUTION_RATIO) -from .jisfreq import (JIS_CHAR_TO_FREQ_ORDER, JIS_TABLE_SIZE, - JIS_TYPICAL_DISTRIBUTION_RATIO) - - -class CharDistributionAnalysis(object): - ENOUGH_DATA_THRESHOLD = 1024 - SURE_YES = 0.99 - SURE_NO = 0.01 - MINIMUM_DATA_THRESHOLD = 3 - - def __init__(self): - # Mapping table to get frequency order from char order (get from - # GetOrder()) - self._char_to_freq_order = None - self._table_size = None # Size of above table - # This is a constant value which varies from language to language, - # used in calculating confidence. See - # http://www.mozilla.org/projects/intl/UniversalCharsetDetection.html - # for further detail. - self.typical_distribution_ratio = None - self._done = None - self._total_chars = None - self._freq_chars = None - self.reset() - - def reset(self): - """reset analyser, clear any state""" - # If this flag is set to True, detection is done and conclusion has - # been made - self._done = False - self._total_chars = 0 # Total characters encountered - # The number of characters whose frequency order is less than 512 - self._freq_chars = 0 - - def feed(self, char, char_len): - """feed a character with known length""" - if char_len == 2: - # we only care about 2-bytes character in our distribution analysis - order = self.get_order(char) - else: - order = -1 - if order >= 0: - self._total_chars += 1 - # order is valid - if order < self._table_size: - if 512 > self._char_to_freq_order[order]: - self._freq_chars += 1 - - def get_confidence(self): - """return confidence based on existing data""" - # if we didn't receive any character in our consideration range, - # return negative answer - if self._total_chars <= 0 or self._freq_chars <= self.MINIMUM_DATA_THRESHOLD: - return self.SURE_NO - - if self._total_chars != self._freq_chars: - r = (self._freq_chars / ((self._total_chars - self._freq_chars) - * self.typical_distribution_ratio)) - if r < self.SURE_YES: - return r - - # normalize confidence (we don't want to be 100% sure) - return self.SURE_YES - - def got_enough_data(self): - # It is not necessary to receive all data to draw conclusion. - # For charset detection, certain amount of data is enough - return self._total_chars > self.ENOUGH_DATA_THRESHOLD - - def get_order(self, byte_str): - # We do not handle characters based on the original encoding string, - # but convert this encoding string to a number, here called order. - # This allows multiple encodings of a language to share one frequency - # table. - return -1 - - -class EUCTWDistributionAnalysis(CharDistributionAnalysis): - def __init__(self): - super(EUCTWDistributionAnalysis, self).__init__() - self._char_to_freq_order = EUCTW_CHAR_TO_FREQ_ORDER - self._table_size = EUCTW_TABLE_SIZE - self.typical_distribution_ratio = EUCTW_TYPICAL_DISTRIBUTION_RATIO - - def get_order(self, byte_str): - # for euc-TW encoding, we are interested - # first byte range: 0xc4 -- 0xfe - # second byte range: 0xa1 -- 0xfe - # no validation needed here. State machine has done that - first_char = byte_str[0] - if first_char >= 0xC4: - return 94 * (first_char - 0xC4) + byte_str[1] - 0xA1 - else: - return -1 - - -class EUCKRDistributionAnalysis(CharDistributionAnalysis): - def __init__(self): - super(EUCKRDistributionAnalysis, self).__init__() - self._char_to_freq_order = EUCKR_CHAR_TO_FREQ_ORDER - self._table_size = EUCKR_TABLE_SIZE - self.typical_distribution_ratio = EUCKR_TYPICAL_DISTRIBUTION_RATIO - - def get_order(self, byte_str): - # for euc-KR encoding, we are interested - # first byte range: 0xb0 -- 0xfe - # second byte range: 0xa1 -- 0xfe - # no validation needed here. State machine has done that - first_char = byte_str[0] - if first_char >= 0xB0: - return 94 * (first_char - 0xB0) + byte_str[1] - 0xA1 - else: - return -1 - - -class GB2312DistributionAnalysis(CharDistributionAnalysis): - def __init__(self): - super(GB2312DistributionAnalysis, self).__init__() - self._char_to_freq_order = GB2312_CHAR_TO_FREQ_ORDER - self._table_size = GB2312_TABLE_SIZE - self.typical_distribution_ratio = GB2312_TYPICAL_DISTRIBUTION_RATIO - - def get_order(self, byte_str): - # for GB2312 encoding, we are interested - # first byte range: 0xb0 -- 0xfe - # second byte range: 0xa1 -- 0xfe - # no validation needed here. State machine has done that - first_char, second_char = byte_str[0], byte_str[1] - if (first_char >= 0xB0) and (second_char >= 0xA1): - return 94 * (first_char - 0xB0) + second_char - 0xA1 - else: - return -1 - - -class Big5DistributionAnalysis(CharDistributionAnalysis): - def __init__(self): - super(Big5DistributionAnalysis, self).__init__() - self._char_to_freq_order = BIG5_CHAR_TO_FREQ_ORDER - self._table_size = BIG5_TABLE_SIZE - self.typical_distribution_ratio = BIG5_TYPICAL_DISTRIBUTION_RATIO - - def get_order(self, byte_str): - # for big5 encoding, we are interested - # first byte range: 0xa4 -- 0xfe - # second byte range: 0x40 -- 0x7e , 0xa1 -- 0xfe - # no validation needed here. State machine has done that - first_char, second_char = byte_str[0], byte_str[1] - if first_char >= 0xA4: - if second_char >= 0xA1: - return 157 * (first_char - 0xA4) + second_char - 0xA1 + 63 - else: - return 157 * (first_char - 0xA4) + second_char - 0x40 - else: - return -1 - - -class SJISDistributionAnalysis(CharDistributionAnalysis): - def __init__(self): - super(SJISDistributionAnalysis, self).__init__() - self._char_to_freq_order = JIS_CHAR_TO_FREQ_ORDER - self._table_size = JIS_TABLE_SIZE - self.typical_distribution_ratio = JIS_TYPICAL_DISTRIBUTION_RATIO - - def get_order(self, byte_str): - # for sjis encoding, we are interested - # first byte range: 0x81 -- 0x9f , 0xe0 -- 0xfe - # second byte range: 0x40 -- 0x7e, 0x81 -- oxfe - # no validation needed here. State machine has done that - first_char, second_char = byte_str[0], byte_str[1] - if (first_char >= 0x81) and (first_char <= 0x9F): - order = 188 * (first_char - 0x81) - elif (first_char >= 0xE0) and (first_char <= 0xEF): - order = 188 * (first_char - 0xE0 + 31) - else: - return -1 - order = order + second_char - 0x40 - if second_char > 0x7F: - order = -1 - return order - - -class EUCJPDistributionAnalysis(CharDistributionAnalysis): - def __init__(self): - super(EUCJPDistributionAnalysis, self).__init__() - self._char_to_freq_order = JIS_CHAR_TO_FREQ_ORDER - self._table_size = JIS_TABLE_SIZE - self.typical_distribution_ratio = JIS_TYPICAL_DISTRIBUTION_RATIO - - def get_order(self, byte_str): - # for euc-JP encoding, we are interested - # first byte range: 0xa0 -- 0xfe - # second byte range: 0xa1 -- 0xfe - # no validation needed here. State machine has done that - char = byte_str[0] - if char >= 0xA0: - return 94 * (char - 0xA1) + byte_str[1] - 0xa1 - else: - return -1 diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/charsetgroupprober.py b/src/fetchcode/vcs/pip/_vendor/chardet/charsetgroupprober.py deleted file mode 100644 index 8b3738ef..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/charsetgroupprober.py +++ /dev/null @@ -1,106 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Communicator client code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -from .enums import ProbingState -from .charsetprober import CharSetProber - - -class CharSetGroupProber(CharSetProber): - def __init__(self, lang_filter=None): - super(CharSetGroupProber, self).__init__(lang_filter=lang_filter) - self._active_num = 0 - self.probers = [] - self._best_guess_prober = None - - def reset(self): - super(CharSetGroupProber, self).reset() - self._active_num = 0 - for prober in self.probers: - if prober: - prober.reset() - prober.active = True - self._active_num += 1 - self._best_guess_prober = None - - @property - def charset_name(self): - if not self._best_guess_prober: - self.get_confidence() - if not self._best_guess_prober: - return None - return self._best_guess_prober.charset_name - - @property - def language(self): - if not self._best_guess_prober: - self.get_confidence() - if not self._best_guess_prober: - return None - return self._best_guess_prober.language - - def feed(self, byte_str): - for prober in self.probers: - if not prober: - continue - if not prober.active: - continue - state = prober.feed(byte_str) - if not state: - continue - if state == ProbingState.FOUND_IT: - self._best_guess_prober = prober - return self.state - elif state == ProbingState.NOT_ME: - prober.active = False - self._active_num -= 1 - if self._active_num <= 0: - self._state = ProbingState.NOT_ME - return self.state - return self.state - - def get_confidence(self): - state = self.state - if state == ProbingState.FOUND_IT: - return 0.99 - elif state == ProbingState.NOT_ME: - return 0.01 - best_conf = 0.0 - self._best_guess_prober = None - for prober in self.probers: - if not prober: - continue - if not prober.active: - self.logger.debug('%s not active', prober.charset_name) - continue - conf = prober.get_confidence() - self.logger.debug('%s %s confidence = %s', prober.charset_name, prober.language, conf) - if best_conf < conf: - best_conf = conf - self._best_guess_prober = prober - if not self._best_guess_prober: - return 0.0 - return best_conf diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/charsetprober.py b/src/fetchcode/vcs/pip/_vendor/chardet/charsetprober.py deleted file mode 100644 index eac4e598..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/charsetprober.py +++ /dev/null @@ -1,145 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Universal charset detector code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 2001 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# Shy Shalom - original C code -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -import logging -import re - -from .enums import ProbingState - - -class CharSetProber(object): - - SHORTCUT_THRESHOLD = 0.95 - - def __init__(self, lang_filter=None): - self._state = None - self.lang_filter = lang_filter - self.logger = logging.getLogger(__name__) - - def reset(self): - self._state = ProbingState.DETECTING - - @property - def charset_name(self): - return None - - def feed(self, buf): - pass - - @property - def state(self): - return self._state - - def get_confidence(self): - return 0.0 - - @staticmethod - def filter_high_byte_only(buf): - buf = re.sub(b'([\x00-\x7F])+', b' ', buf) - return buf - - @staticmethod - def filter_international_words(buf): - """ - We define three types of bytes: - alphabet: english alphabets [a-zA-Z] - international: international characters [\x80-\xFF] - marker: everything else [^a-zA-Z\x80-\xFF] - - The input buffer can be thought to contain a series of words delimited - by markers. This function works to filter all words that contain at - least one international character. All contiguous sequences of markers - are replaced by a single space ascii character. - - This filter applies to all scripts which do not use English characters. - """ - filtered = bytearray() - - # This regex expression filters out only words that have at-least one - # international character. The word may include one marker character at - # the end. - words = re.findall(b'[a-zA-Z]*[\x80-\xFF]+[a-zA-Z]*[^a-zA-Z\x80-\xFF]?', - buf) - - for word in words: - filtered.extend(word[:-1]) - - # If the last character in the word is a marker, replace it with a - # space as markers shouldn't affect our analysis (they are used - # similarly across all languages and may thus have similar - # frequencies). - last_char = word[-1:] - if not last_char.isalpha() and last_char < b'\x80': - last_char = b' ' - filtered.extend(last_char) - - return filtered - - @staticmethod - def filter_with_english_letters(buf): - """ - Returns a copy of ``buf`` that retains only the sequences of English - alphabet and high byte characters that are not between <> characters. - Also retains English alphabet and high byte characters immediately - before occurrences of >. - - This filter can be applied to all scripts which contain both English - characters and extended ASCII characters, but is currently only used by - ``Latin1Prober``. - """ - filtered = bytearray() - in_tag = False - prev = 0 - - for curr in range(len(buf)): - # Slice here to get bytes instead of an int with Python 3 - buf_char = buf[curr:curr + 1] - # Check if we're coming out of or entering an HTML tag - if buf_char == b'>': - in_tag = False - elif buf_char == b'<': - in_tag = True - - # If current character is not extended-ASCII and not alphabetic... - if buf_char < b'\x80' and not buf_char.isalpha(): - # ...and we're not in a tag - if curr > prev and not in_tag: - # Keep everything after last non-extended-ASCII, - # non-alphabetic character - filtered.extend(buf[prev:curr]) - # Output a space to delimit stretch we kept - filtered.extend(b' ') - prev = curr + 1 - - # If we're not in a tag... - if not in_tag: - # Keep everything after last non-extended-ASCII, non-alphabetic - # character - filtered.extend(buf[prev:]) - - return filtered diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/cli/__init__.py b/src/fetchcode/vcs/pip/_vendor/chardet/cli/__init__.py deleted file mode 100644 index 8b137891..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/cli/__init__.py +++ /dev/null @@ -1 +0,0 @@ - diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/cli/chardetect.py b/src/fetchcode/vcs/pip/_vendor/chardet/cli/chardetect.py deleted file mode 100644 index 1679e38f..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/cli/chardetect.py +++ /dev/null @@ -1,85 +0,0 @@ -#!/usr/bin/env python -""" -Script which takes one or more file paths and reports on their detected -encodings - -Example:: - - % chardetect somefile someotherfile - somefile: windows-1252 with confidence 0.5 - someotherfile: ascii with confidence 1.0 - -If no paths are provided, it takes its input from stdin. - -""" - -from __future__ import absolute_import, print_function, unicode_literals - -import argparse -import sys - -from fetchcode.vcs.pip._vendor.chardet import __version__ -from fetchcode.vcs.pip._vendor.chardet.compat import PY2 -from fetchcode.vcs.pip._vendor.chardet.universaldetector import UniversalDetector - - -def description_of(lines, name='stdin'): - """ - Return a string describing the probable encoding of a file or - list of strings. - - :param lines: The lines to get the encoding of. - :type lines: Iterable of bytes - :param name: Name of file or collection of lines - :type name: str - """ - u = UniversalDetector() - for line in lines: - line = bytearray(line) - u.feed(line) - # shortcut out of the loop to save reading further - particularly useful if we read a BOM. - if u.done: - break - u.close() - result = u.result - if PY2: - name = name.decode(sys.getfilesystemencoding(), 'ignore') - if result['encoding']: - return '{0}: {1} with confidence {2}'.format(name, result['encoding'], - result['confidence']) - else: - return '{0}: no result'.format(name) - - -def main(argv=None): - """ - Handles command line arguments and gets things started. - - :param argv: List of arguments, as if specified on the command-line. - If None, ``sys.argv[1:]`` is used instead. - :type argv: list of str - """ - # Get command line arguments - parser = argparse.ArgumentParser( - description="Takes one or more file paths and reports their detected \ - encodings") - parser.add_argument('input', - help='File whose encoding we would like to determine. \ - (default: stdin)', - type=argparse.FileType('rb'), nargs='*', - default=[sys.stdin if PY2 else sys.stdin.buffer]) - parser.add_argument('--version', action='version', - version='%(prog)s {0}'.format(__version__)) - args = parser.parse_args(argv) - - for f in args.input: - if f.isatty(): - print("You are running chardetect interactively. Press " + - "CTRL-D twice at the start of a blank line to signal the " + - "end of your input. If you want help, run chardetect " + - "--help\n", file=sys.stderr) - print(description_of(f, f.name)) - - -if __name__ == '__main__': - main() diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/codingstatemachine.py b/src/fetchcode/vcs/pip/_vendor/chardet/codingstatemachine.py deleted file mode 100644 index 68fba44f..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/codingstatemachine.py +++ /dev/null @@ -1,88 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is mozilla.org code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -import logging - -from .enums import MachineState - - -class CodingStateMachine(object): - """ - A state machine to verify a byte sequence for a particular encoding. For - each byte the detector receives, it will feed that byte to every active - state machine available, one byte at a time. The state machine changes its - state based on its previous state and the byte it receives. There are 3 - states in a state machine that are of interest to an auto-detector: - - START state: This is the state to start with, or a legal byte sequence - (i.e. a valid code point) for character has been identified. - - ME state: This indicates that the state machine identified a byte sequence - that is specific to the charset it is designed for and that - there is no other possible encoding which can contain this byte - sequence. This will to lead to an immediate positive answer for - the detector. - - ERROR state: This indicates the state machine identified an illegal byte - sequence for that encoding. This will lead to an immediate - negative answer for this encoding. Detector will exclude this - encoding from consideration from here on. - """ - def __init__(self, sm): - self._model = sm - self._curr_byte_pos = 0 - self._curr_char_len = 0 - self._curr_state = None - self.logger = logging.getLogger(__name__) - self.reset() - - def reset(self): - self._curr_state = MachineState.START - - def next_state(self, c): - # for each byte we get its class - # if it is first byte, we also get byte length - byte_class = self._model['class_table'][c] - if self._curr_state == MachineState.START: - self._curr_byte_pos = 0 - self._curr_char_len = self._model['char_len_table'][byte_class] - # from byte's class and state_table, we get its next state - curr_state = (self._curr_state * self._model['class_factor'] - + byte_class) - self._curr_state = self._model['state_table'][curr_state] - self._curr_byte_pos += 1 - return self._curr_state - - def get_current_charlen(self): - return self._curr_char_len - - def get_coding_state_machine(self): - return self._model['name'] - - @property - def language(self): - return self._model['language'] diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/compat.py b/src/fetchcode/vcs/pip/_vendor/chardet/compat.py deleted file mode 100644 index ddd74687..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/compat.py +++ /dev/null @@ -1,34 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# Contributor(s): -# Dan Blanchard -# Ian Cordasco -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -import sys - - -if sys.version_info < (3, 0): - PY2 = True - PY3 = False - base_str = (str, unicode) - text_type = unicode -else: - PY2 = False - PY3 = True - base_str = (bytes, str) - text_type = str diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/cp949prober.py b/src/fetchcode/vcs/pip/_vendor/chardet/cp949prober.py deleted file mode 100644 index efd793ab..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/cp949prober.py +++ /dev/null @@ -1,49 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is mozilla.org code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -from .chardistribution import EUCKRDistributionAnalysis -from .codingstatemachine import CodingStateMachine -from .mbcharsetprober import MultiByteCharSetProber -from .mbcssm import CP949_SM_MODEL - - -class CP949Prober(MultiByteCharSetProber): - def __init__(self): - super(CP949Prober, self).__init__() - self.coding_sm = CodingStateMachine(CP949_SM_MODEL) - # NOTE: CP949 is a superset of EUC-KR, so the distribution should be - # not different. - self.distribution_analyzer = EUCKRDistributionAnalysis() - self.reset() - - @property - def charset_name(self): - return "CP949" - - @property - def language(self): - return "Korean" diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/enums.py b/src/fetchcode/vcs/pip/_vendor/chardet/enums.py deleted file mode 100644 index 04512072..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/enums.py +++ /dev/null @@ -1,76 +0,0 @@ -""" -All of the Enums that are used throughout the chardet package. - -:author: Dan Blanchard (dan.blanchard@gmail.com) -""" - - -class InputState(object): - """ - This enum represents the different states a universal detector can be in. - """ - PURE_ASCII = 0 - ESC_ASCII = 1 - HIGH_BYTE = 2 - - -class LanguageFilter(object): - """ - This enum represents the different language filters we can apply to a - ``UniversalDetector``. - """ - CHINESE_SIMPLIFIED = 0x01 - CHINESE_TRADITIONAL = 0x02 - JAPANESE = 0x04 - KOREAN = 0x08 - NON_CJK = 0x10 - ALL = 0x1F - CHINESE = CHINESE_SIMPLIFIED | CHINESE_TRADITIONAL - CJK = CHINESE | JAPANESE | KOREAN - - -class ProbingState(object): - """ - This enum represents the different states a prober can be in. - """ - DETECTING = 0 - FOUND_IT = 1 - NOT_ME = 2 - - -class MachineState(object): - """ - This enum represents the different states a state machine can be in. - """ - START = 0 - ERROR = 1 - ITS_ME = 2 - - -class SequenceLikelihood(object): - """ - This enum represents the likelihood of a character following the previous one. - """ - NEGATIVE = 0 - UNLIKELY = 1 - LIKELY = 2 - POSITIVE = 3 - - @classmethod - def get_num_categories(cls): - """:returns: The number of likelihood categories in the enum.""" - return 4 - - -class CharacterCategory(object): - """ - This enum represents the different categories language models for - ``SingleByteCharsetProber`` put characters into. - - Anything less than CONTROL is considered a letter. - """ - UNDEFINED = 255 - LINE_BREAK = 254 - SYMBOL = 253 - DIGIT = 252 - CONTROL = 251 diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/escprober.py b/src/fetchcode/vcs/pip/_vendor/chardet/escprober.py deleted file mode 100644 index c70493f2..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/escprober.py +++ /dev/null @@ -1,101 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is mozilla.org code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -from .charsetprober import CharSetProber -from .codingstatemachine import CodingStateMachine -from .enums import LanguageFilter, ProbingState, MachineState -from .escsm import (HZ_SM_MODEL, ISO2022CN_SM_MODEL, ISO2022JP_SM_MODEL, - ISO2022KR_SM_MODEL) - - -class EscCharSetProber(CharSetProber): - """ - This CharSetProber uses a "code scheme" approach for detecting encodings, - whereby easily recognizable escape or shift sequences are relied on to - identify these encodings. - """ - - def __init__(self, lang_filter=None): - super(EscCharSetProber, self).__init__(lang_filter=lang_filter) - self.coding_sm = [] - if self.lang_filter & LanguageFilter.CHINESE_SIMPLIFIED: - self.coding_sm.append(CodingStateMachine(HZ_SM_MODEL)) - self.coding_sm.append(CodingStateMachine(ISO2022CN_SM_MODEL)) - if self.lang_filter & LanguageFilter.JAPANESE: - self.coding_sm.append(CodingStateMachine(ISO2022JP_SM_MODEL)) - if self.lang_filter & LanguageFilter.KOREAN: - self.coding_sm.append(CodingStateMachine(ISO2022KR_SM_MODEL)) - self.active_sm_count = None - self._detected_charset = None - self._detected_language = None - self._state = None - self.reset() - - def reset(self): - super(EscCharSetProber, self).reset() - for coding_sm in self.coding_sm: - if not coding_sm: - continue - coding_sm.active = True - coding_sm.reset() - self.active_sm_count = len(self.coding_sm) - self._detected_charset = None - self._detected_language = None - - @property - def charset_name(self): - return self._detected_charset - - @property - def language(self): - return self._detected_language - - def get_confidence(self): - if self._detected_charset: - return 0.99 - else: - return 0.00 - - def feed(self, byte_str): - for c in byte_str: - for coding_sm in self.coding_sm: - if not coding_sm or not coding_sm.active: - continue - coding_state = coding_sm.next_state(c) - if coding_state == MachineState.ERROR: - coding_sm.active = False - self.active_sm_count -= 1 - if self.active_sm_count <= 0: - self._state = ProbingState.NOT_ME - return self.state - elif coding_state == MachineState.ITS_ME: - self._state = ProbingState.FOUND_IT - self._detected_charset = coding_sm.get_coding_state_machine() - self._detected_language = coding_sm.language - return self.state - - return self.state diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/escsm.py b/src/fetchcode/vcs/pip/_vendor/chardet/escsm.py deleted file mode 100644 index 0069523a..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/escsm.py +++ /dev/null @@ -1,246 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is mozilla.org code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -from .enums import MachineState - -HZ_CLS = ( -1,0,0,0,0,0,0,0, # 00 - 07 -0,0,0,0,0,0,0,0, # 08 - 0f -0,0,0,0,0,0,0,0, # 10 - 17 -0,0,0,1,0,0,0,0, # 18 - 1f -0,0,0,0,0,0,0,0, # 20 - 27 -0,0,0,0,0,0,0,0, # 28 - 2f -0,0,0,0,0,0,0,0, # 30 - 37 -0,0,0,0,0,0,0,0, # 38 - 3f -0,0,0,0,0,0,0,0, # 40 - 47 -0,0,0,0,0,0,0,0, # 48 - 4f -0,0,0,0,0,0,0,0, # 50 - 57 -0,0,0,0,0,0,0,0, # 58 - 5f -0,0,0,0,0,0,0,0, # 60 - 67 -0,0,0,0,0,0,0,0, # 68 - 6f -0,0,0,0,0,0,0,0, # 70 - 77 -0,0,0,4,0,5,2,0, # 78 - 7f -1,1,1,1,1,1,1,1, # 80 - 87 -1,1,1,1,1,1,1,1, # 88 - 8f -1,1,1,1,1,1,1,1, # 90 - 97 -1,1,1,1,1,1,1,1, # 98 - 9f -1,1,1,1,1,1,1,1, # a0 - a7 -1,1,1,1,1,1,1,1, # a8 - af -1,1,1,1,1,1,1,1, # b0 - b7 -1,1,1,1,1,1,1,1, # b8 - bf -1,1,1,1,1,1,1,1, # c0 - c7 -1,1,1,1,1,1,1,1, # c8 - cf -1,1,1,1,1,1,1,1, # d0 - d7 -1,1,1,1,1,1,1,1, # d8 - df -1,1,1,1,1,1,1,1, # e0 - e7 -1,1,1,1,1,1,1,1, # e8 - ef -1,1,1,1,1,1,1,1, # f0 - f7 -1,1,1,1,1,1,1,1, # f8 - ff -) - -HZ_ST = ( -MachineState.START,MachineState.ERROR, 3,MachineState.START,MachineState.START,MachineState.START,MachineState.ERROR,MachineState.ERROR,# 00-07 -MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,# 08-0f -MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START, 4,MachineState.ERROR,# 10-17 - 5,MachineState.ERROR, 6,MachineState.ERROR, 5, 5, 4,MachineState.ERROR,# 18-1f - 4,MachineState.ERROR, 4, 4, 4,MachineState.ERROR, 4,MachineState.ERROR,# 20-27 - 4,MachineState.ITS_ME,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,# 28-2f -) - -HZ_CHAR_LEN_TABLE = (0, 0, 0, 0, 0, 0) - -HZ_SM_MODEL = {'class_table': HZ_CLS, - 'class_factor': 6, - 'state_table': HZ_ST, - 'char_len_table': HZ_CHAR_LEN_TABLE, - 'name': "HZ-GB-2312", - 'language': 'Chinese'} - -ISO2022CN_CLS = ( -2,0,0,0,0,0,0,0, # 00 - 07 -0,0,0,0,0,0,0,0, # 08 - 0f -0,0,0,0,0,0,0,0, # 10 - 17 -0,0,0,1,0,0,0,0, # 18 - 1f -0,0,0,0,0,0,0,0, # 20 - 27 -0,3,0,0,0,0,0,0, # 28 - 2f -0,0,0,0,0,0,0,0, # 30 - 37 -0,0,0,0,0,0,0,0, # 38 - 3f -0,0,0,4,0,0,0,0, # 40 - 47 -0,0,0,0,0,0,0,0, # 48 - 4f -0,0,0,0,0,0,0,0, # 50 - 57 -0,0,0,0,0,0,0,0, # 58 - 5f -0,0,0,0,0,0,0,0, # 60 - 67 -0,0,0,0,0,0,0,0, # 68 - 6f -0,0,0,0,0,0,0,0, # 70 - 77 -0,0,0,0,0,0,0,0, # 78 - 7f -2,2,2,2,2,2,2,2, # 80 - 87 -2,2,2,2,2,2,2,2, # 88 - 8f -2,2,2,2,2,2,2,2, # 90 - 97 -2,2,2,2,2,2,2,2, # 98 - 9f -2,2,2,2,2,2,2,2, # a0 - a7 -2,2,2,2,2,2,2,2, # a8 - af -2,2,2,2,2,2,2,2, # b0 - b7 -2,2,2,2,2,2,2,2, # b8 - bf -2,2,2,2,2,2,2,2, # c0 - c7 -2,2,2,2,2,2,2,2, # c8 - cf -2,2,2,2,2,2,2,2, # d0 - d7 -2,2,2,2,2,2,2,2, # d8 - df -2,2,2,2,2,2,2,2, # e0 - e7 -2,2,2,2,2,2,2,2, # e8 - ef -2,2,2,2,2,2,2,2, # f0 - f7 -2,2,2,2,2,2,2,2, # f8 - ff -) - -ISO2022CN_ST = ( -MachineState.START, 3,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,# 00-07 -MachineState.START,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,# 08-0f -MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,# 10-17 -MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR, 4,MachineState.ERROR,# 18-1f -MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,# 20-27 - 5, 6,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,# 28-2f -MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,# 30-37 -MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ERROR,MachineState.START,# 38-3f -) - -ISO2022CN_CHAR_LEN_TABLE = (0, 0, 0, 0, 0, 0, 0, 0, 0) - -ISO2022CN_SM_MODEL = {'class_table': ISO2022CN_CLS, - 'class_factor': 9, - 'state_table': ISO2022CN_ST, - 'char_len_table': ISO2022CN_CHAR_LEN_TABLE, - 'name': "ISO-2022-CN", - 'language': 'Chinese'} - -ISO2022JP_CLS = ( -2,0,0,0,0,0,0,0, # 00 - 07 -0,0,0,0,0,0,2,2, # 08 - 0f -0,0,0,0,0,0,0,0, # 10 - 17 -0,0,0,1,0,0,0,0, # 18 - 1f -0,0,0,0,7,0,0,0, # 20 - 27 -3,0,0,0,0,0,0,0, # 28 - 2f -0,0,0,0,0,0,0,0, # 30 - 37 -0,0,0,0,0,0,0,0, # 38 - 3f -6,0,4,0,8,0,0,0, # 40 - 47 -0,9,5,0,0,0,0,0, # 48 - 4f -0,0,0,0,0,0,0,0, # 50 - 57 -0,0,0,0,0,0,0,0, # 58 - 5f -0,0,0,0,0,0,0,0, # 60 - 67 -0,0,0,0,0,0,0,0, # 68 - 6f -0,0,0,0,0,0,0,0, # 70 - 77 -0,0,0,0,0,0,0,0, # 78 - 7f -2,2,2,2,2,2,2,2, # 80 - 87 -2,2,2,2,2,2,2,2, # 88 - 8f -2,2,2,2,2,2,2,2, # 90 - 97 -2,2,2,2,2,2,2,2, # 98 - 9f -2,2,2,2,2,2,2,2, # a0 - a7 -2,2,2,2,2,2,2,2, # a8 - af -2,2,2,2,2,2,2,2, # b0 - b7 -2,2,2,2,2,2,2,2, # b8 - bf -2,2,2,2,2,2,2,2, # c0 - c7 -2,2,2,2,2,2,2,2, # c8 - cf -2,2,2,2,2,2,2,2, # d0 - d7 -2,2,2,2,2,2,2,2, # d8 - df -2,2,2,2,2,2,2,2, # e0 - e7 -2,2,2,2,2,2,2,2, # e8 - ef -2,2,2,2,2,2,2,2, # f0 - f7 -2,2,2,2,2,2,2,2, # f8 - ff -) - -ISO2022JP_ST = ( -MachineState.START, 3,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,# 00-07 -MachineState.START,MachineState.START,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,# 08-0f -MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,# 10-17 -MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ERROR,MachineState.ERROR,# 18-1f -MachineState.ERROR, 5,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR, 4,MachineState.ERROR,MachineState.ERROR,# 20-27 -MachineState.ERROR,MachineState.ERROR,MachineState.ERROR, 6,MachineState.ITS_ME,MachineState.ERROR,MachineState.ITS_ME,MachineState.ERROR,# 28-2f -MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,# 30-37 -MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,# 38-3f -MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ERROR,MachineState.START,MachineState.START,# 40-47 -) - -ISO2022JP_CHAR_LEN_TABLE = (0, 0, 0, 0, 0, 0, 0, 0, 0, 0) - -ISO2022JP_SM_MODEL = {'class_table': ISO2022JP_CLS, - 'class_factor': 10, - 'state_table': ISO2022JP_ST, - 'char_len_table': ISO2022JP_CHAR_LEN_TABLE, - 'name': "ISO-2022-JP", - 'language': 'Japanese'} - -ISO2022KR_CLS = ( -2,0,0,0,0,0,0,0, # 00 - 07 -0,0,0,0,0,0,0,0, # 08 - 0f -0,0,0,0,0,0,0,0, # 10 - 17 -0,0,0,1,0,0,0,0, # 18 - 1f -0,0,0,0,3,0,0,0, # 20 - 27 -0,4,0,0,0,0,0,0, # 28 - 2f -0,0,0,0,0,0,0,0, # 30 - 37 -0,0,0,0,0,0,0,0, # 38 - 3f -0,0,0,5,0,0,0,0, # 40 - 47 -0,0,0,0,0,0,0,0, # 48 - 4f -0,0,0,0,0,0,0,0, # 50 - 57 -0,0,0,0,0,0,0,0, # 58 - 5f -0,0,0,0,0,0,0,0, # 60 - 67 -0,0,0,0,0,0,0,0, # 68 - 6f -0,0,0,0,0,0,0,0, # 70 - 77 -0,0,0,0,0,0,0,0, # 78 - 7f -2,2,2,2,2,2,2,2, # 80 - 87 -2,2,2,2,2,2,2,2, # 88 - 8f -2,2,2,2,2,2,2,2, # 90 - 97 -2,2,2,2,2,2,2,2, # 98 - 9f -2,2,2,2,2,2,2,2, # a0 - a7 -2,2,2,2,2,2,2,2, # a8 - af -2,2,2,2,2,2,2,2, # b0 - b7 -2,2,2,2,2,2,2,2, # b8 - bf -2,2,2,2,2,2,2,2, # c0 - c7 -2,2,2,2,2,2,2,2, # c8 - cf -2,2,2,2,2,2,2,2, # d0 - d7 -2,2,2,2,2,2,2,2, # d8 - df -2,2,2,2,2,2,2,2, # e0 - e7 -2,2,2,2,2,2,2,2, # e8 - ef -2,2,2,2,2,2,2,2, # f0 - f7 -2,2,2,2,2,2,2,2, # f8 - ff -) - -ISO2022KR_ST = ( -MachineState.START, 3,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.ERROR,MachineState.ERROR,# 00-07 -MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,# 08-0f -MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR, 4,MachineState.ERROR,MachineState.ERROR,# 10-17 -MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR, 5,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,# 18-1f -MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.START,MachineState.START,MachineState.START,MachineState.START,# 20-27 -) - -ISO2022KR_CHAR_LEN_TABLE = (0, 0, 0, 0, 0, 0) - -ISO2022KR_SM_MODEL = {'class_table': ISO2022KR_CLS, - 'class_factor': 6, - 'state_table': ISO2022KR_ST, - 'char_len_table': ISO2022KR_CHAR_LEN_TABLE, - 'name': "ISO-2022-KR", - 'language': 'Korean'} - - diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/eucjpprober.py b/src/fetchcode/vcs/pip/_vendor/chardet/eucjpprober.py deleted file mode 100644 index 20ce8f7d..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/eucjpprober.py +++ /dev/null @@ -1,92 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is mozilla.org code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -from .enums import ProbingState, MachineState -from .mbcharsetprober import MultiByteCharSetProber -from .codingstatemachine import CodingStateMachine -from .chardistribution import EUCJPDistributionAnalysis -from .jpcntx import EUCJPContextAnalysis -from .mbcssm import EUCJP_SM_MODEL - - -class EUCJPProber(MultiByteCharSetProber): - def __init__(self): - super(EUCJPProber, self).__init__() - self.coding_sm = CodingStateMachine(EUCJP_SM_MODEL) - self.distribution_analyzer = EUCJPDistributionAnalysis() - self.context_analyzer = EUCJPContextAnalysis() - self.reset() - - def reset(self): - super(EUCJPProber, self).reset() - self.context_analyzer.reset() - - @property - def charset_name(self): - return "EUC-JP" - - @property - def language(self): - return "Japanese" - - def feed(self, byte_str): - for i in range(len(byte_str)): - # PY3K: byte_str is a byte array, so byte_str[i] is an int, not a byte - coding_state = self.coding_sm.next_state(byte_str[i]) - if coding_state == MachineState.ERROR: - self.logger.debug('%s %s prober hit error at byte %s', - self.charset_name, self.language, i) - self._state = ProbingState.NOT_ME - break - elif coding_state == MachineState.ITS_ME: - self._state = ProbingState.FOUND_IT - break - elif coding_state == MachineState.START: - char_len = self.coding_sm.get_current_charlen() - if i == 0: - self._last_char[1] = byte_str[0] - self.context_analyzer.feed(self._last_char, char_len) - self.distribution_analyzer.feed(self._last_char, char_len) - else: - self.context_analyzer.feed(byte_str[i - 1:i + 1], - char_len) - self.distribution_analyzer.feed(byte_str[i - 1:i + 1], - char_len) - - self._last_char[0] = byte_str[-1] - - if self.state == ProbingState.DETECTING: - if (self.context_analyzer.got_enough_data() and - (self.get_confidence() > self.SHORTCUT_THRESHOLD)): - self._state = ProbingState.FOUND_IT - - return self.state - - def get_confidence(self): - context_conf = self.context_analyzer.get_confidence() - distrib_conf = self.distribution_analyzer.get_confidence() - return max(context_conf, distrib_conf) diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/euckrfreq.py b/src/fetchcode/vcs/pip/_vendor/chardet/euckrfreq.py deleted file mode 100644 index b68078cb..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/euckrfreq.py +++ /dev/null @@ -1,195 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Communicator client code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -# Sampling from about 20M text materials include literature and computer technology - -# 128 --> 0.79 -# 256 --> 0.92 -# 512 --> 0.986 -# 1024 --> 0.99944 -# 2048 --> 0.99999 -# -# Idea Distribution Ratio = 0.98653 / (1-0.98653) = 73.24 -# Random Distribution Ration = 512 / (2350-512) = 0.279. -# -# Typical Distribution Ratio - -EUCKR_TYPICAL_DISTRIBUTION_RATIO = 6.0 - -EUCKR_TABLE_SIZE = 2352 - -# Char to FreqOrder table , -EUCKR_CHAR_TO_FREQ_ORDER = ( - 13, 130, 120,1396, 481,1719,1720, 328, 609, 212,1721, 707, 400, 299,1722, 87, -1397,1723, 104, 536,1117,1203,1724,1267, 685,1268, 508,1725,1726,1727,1728,1398, -1399,1729,1730,1731, 141, 621, 326,1057, 368,1732, 267, 488, 20,1733,1269,1734, - 945,1400,1735, 47, 904,1270,1736,1737, 773, 248,1738, 409, 313, 786, 429,1739, - 116, 987, 813,1401, 683, 75,1204, 145,1740,1741,1742,1743, 16, 847, 667, 622, - 708,1744,1745,1746, 966, 787, 304, 129,1747, 60, 820, 123, 676,1748,1749,1750, -1751, 617,1752, 626,1753,1754,1755,1756, 653,1757,1758,1759,1760,1761,1762, 856, - 344,1763,1764,1765,1766, 89, 401, 418, 806, 905, 848,1767,1768,1769, 946,1205, - 709,1770,1118,1771, 241,1772,1773,1774,1271,1775, 569,1776, 999,1777,1778,1779, -1780, 337, 751,1058, 28, 628, 254,1781, 177, 906, 270, 349, 891,1079,1782, 19, -1783, 379,1784, 315,1785, 629, 754,1402, 559,1786, 636, 203,1206,1787, 710, 567, -1788, 935, 814,1789,1790,1207, 766, 528,1791,1792,1208,1793,1794,1795,1796,1797, -1403,1798,1799, 533,1059,1404,1405,1156,1406, 936, 884,1080,1800, 351,1801,1802, -1803,1804,1805, 801,1806,1807,1808,1119,1809,1157, 714, 474,1407,1810, 298, 899, - 885,1811,1120, 802,1158,1812, 892,1813,1814,1408, 659,1815,1816,1121,1817,1818, -1819,1820,1821,1822, 319,1823, 594, 545,1824, 815, 937,1209,1825,1826, 573,1409, -1022,1827,1210,1828,1829,1830,1831,1832,1833, 556, 722, 807,1122,1060,1834, 697, -1835, 900, 557, 715,1836,1410, 540,1411, 752,1159, 294, 597,1211, 976, 803, 770, -1412,1837,1838, 39, 794,1413, 358,1839, 371, 925,1840, 453, 661, 788, 531, 723, - 544,1023,1081, 869, 91,1841, 392, 430, 790, 602,1414, 677,1082, 457,1415,1416, -1842,1843, 475, 327,1024,1417, 795, 121,1844, 733, 403,1418,1845,1846,1847, 300, - 119, 711,1212, 627,1848,1272, 207,1849,1850, 796,1213, 382,1851, 519,1852,1083, - 893,1853,1854,1855, 367, 809, 487, 671,1856, 663,1857,1858, 956, 471, 306, 857, -1859,1860,1160,1084,1861,1862,1863,1864,1865,1061,1866,1867,1868,1869,1870,1871, - 282, 96, 574,1872, 502,1085,1873,1214,1874, 907,1875,1876, 827, 977,1419,1420, -1421, 268,1877,1422,1878,1879,1880, 308,1881, 2, 537,1882,1883,1215,1884,1885, - 127, 791,1886,1273,1423,1887, 34, 336, 404, 643,1888, 571, 654, 894, 840,1889, - 0, 886,1274, 122, 575, 260, 908, 938,1890,1275, 410, 316,1891,1892, 100,1893, -1894,1123, 48,1161,1124,1025,1895, 633, 901,1276,1896,1897, 115, 816,1898, 317, -1899, 694,1900, 909, 734,1424, 572, 866,1425, 691, 85, 524,1010, 543, 394, 841, -1901,1902,1903,1026,1904,1905,1906,1907,1908,1909, 30, 451, 651, 988, 310,1910, -1911,1426, 810,1216, 93,1912,1913,1277,1217,1914, 858, 759, 45, 58, 181, 610, - 269,1915,1916, 131,1062, 551, 443,1000, 821,1427, 957, 895,1086,1917,1918, 375, -1919, 359,1920, 687,1921, 822,1922, 293,1923,1924, 40, 662, 118, 692, 29, 939, - 887, 640, 482, 174,1925, 69,1162, 728,1428, 910,1926,1278,1218,1279, 386, 870, - 217, 854,1163, 823,1927,1928,1929,1930, 834,1931, 78,1932, 859,1933,1063,1934, -1935,1936,1937, 438,1164, 208, 595,1938,1939,1940,1941,1219,1125,1942, 280, 888, -1429,1430,1220,1431,1943,1944,1945,1946,1947,1280, 150, 510,1432,1948,1949,1950, -1951,1952,1953,1954,1011,1087,1955,1433,1043,1956, 881,1957, 614, 958,1064,1065, -1221,1958, 638,1001, 860, 967, 896,1434, 989, 492, 553,1281,1165,1959,1282,1002, -1283,1222,1960,1961,1962,1963, 36, 383, 228, 753, 247, 454,1964, 876, 678,1965, -1966,1284, 126, 464, 490, 835, 136, 672, 529, 940,1088,1435, 473,1967,1968, 467, - 50, 390, 227, 587, 279, 378, 598, 792, 968, 240, 151, 160, 849, 882,1126,1285, - 639,1044, 133, 140, 288, 360, 811, 563,1027, 561, 142, 523,1969,1970,1971, 7, - 103, 296, 439, 407, 506, 634, 990,1972,1973,1974,1975, 645,1976,1977,1978,1979, -1980,1981, 236,1982,1436,1983,1984,1089, 192, 828, 618, 518,1166, 333,1127,1985, - 818,1223,1986,1987,1988,1989,1990,1991,1992,1993, 342,1128,1286, 746, 842,1994, -1995, 560, 223,1287, 98, 8, 189, 650, 978,1288,1996,1437,1997, 17, 345, 250, - 423, 277, 234, 512, 226, 97, 289, 42, 167,1998, 201,1999,2000, 843, 836, 824, - 532, 338, 783,1090, 182, 576, 436,1438,1439, 527, 500,2001, 947, 889,2002,2003, -2004,2005, 262, 600, 314, 447,2006, 547,2007, 693, 738,1129,2008, 71,1440, 745, - 619, 688,2009, 829,2010,2011, 147,2012, 33, 948,2013,2014, 74, 224,2015, 61, - 191, 918, 399, 637,2016,1028,1130, 257, 902,2017,2018,2019,2020,2021,2022,2023, -2024,2025,2026, 837,2027,2028,2029,2030, 179, 874, 591, 52, 724, 246,2031,2032, -2033,2034,1167, 969,2035,1289, 630, 605, 911,1091,1168,2036,2037,2038,1441, 912, -2039, 623,2040,2041, 253,1169,1290,2042,1442, 146, 620, 611, 577, 433,2043,1224, - 719,1170, 959, 440, 437, 534, 84, 388, 480,1131, 159, 220, 198, 679,2044,1012, - 819,1066,1443, 113,1225, 194, 318,1003,1029,2045,2046,2047,2048,1067,2049,2050, -2051,2052,2053, 59, 913, 112,2054, 632,2055, 455, 144, 739,1291,2056, 273, 681, - 499,2057, 448,2058,2059, 760,2060,2061, 970, 384, 169, 245,1132,2062,2063, 414, -1444,2064,2065, 41, 235,2066, 157, 252, 877, 568, 919, 789, 580,2067, 725,2068, -2069,1292,2070,2071,1445,2072,1446,2073,2074, 55, 588, 66,1447, 271,1092,2075, -1226,2076, 960,1013, 372,2077,2078,2079,2080,2081,1293,2082,2083,2084,2085, 850, -2086,2087,2088,2089,2090, 186,2091,1068, 180,2092,2093,2094, 109,1227, 522, 606, -2095, 867,1448,1093, 991,1171, 926, 353,1133,2096, 581,2097,2098,2099,1294,1449, -1450,2100, 596,1172,1014,1228,2101,1451,1295,1173,1229,2102,2103,1296,1134,1452, - 949,1135,2104,2105,1094,1453,1454,1455,2106,1095,2107,2108,2109,2110,2111,2112, -2113,2114,2115,2116,2117, 804,2118,2119,1230,1231, 805,1456, 405,1136,2120,2121, -2122,2123,2124, 720, 701,1297, 992,1457, 927,1004,2125,2126,2127,2128,2129,2130, - 22, 417,2131, 303,2132, 385,2133, 971, 520, 513,2134,1174, 73,1096, 231, 274, - 962,1458, 673,2135,1459,2136, 152,1137,2137,2138,2139,2140,1005,1138,1460,1139, -2141,2142,2143,2144, 11, 374, 844,2145, 154,1232, 46,1461,2146, 838, 830, 721, -1233, 106,2147, 90, 428, 462, 578, 566,1175, 352,2148,2149, 538,1234, 124,1298, -2150,1462, 761, 565,2151, 686,2152, 649,2153, 72, 173,2154, 460, 415,2155,1463, -2156,1235, 305,2157,2158,2159,2160,2161,2162, 579,2163,2164,2165,2166,2167, 747, -2168,2169,2170,2171,1464, 669,2172,2173,2174,2175,2176,1465,2177, 23, 530, 285, -2178, 335, 729,2179, 397,2180,2181,2182,1030,2183,2184, 698,2185,2186, 325,2187, -2188, 369,2189, 799,1097,1015, 348,2190,1069, 680,2191, 851,1466,2192,2193, 10, -2194, 613, 424,2195, 979, 108, 449, 589, 27, 172, 81,1031, 80, 774, 281, 350, -1032, 525, 301, 582,1176,2196, 674,1045,2197,2198,1467, 730, 762,2199,2200,2201, -2202,1468,2203, 993,2204,2205, 266,1070, 963,1140,2206,2207,2208, 664,1098, 972, -2209,2210,2211,1177,1469,1470, 871,2212,2213,2214,2215,2216,1471,2217,2218,2219, -2220,2221,2222,2223,2224,2225,2226,2227,1472,1236,2228,2229,2230,2231,2232,2233, -2234,2235,1299,2236,2237, 200,2238, 477, 373,2239,2240, 731, 825, 777,2241,2242, -2243, 521, 486, 548,2244,2245,2246,1473,1300, 53, 549, 137, 875, 76, 158,2247, -1301,1474, 469, 396,1016, 278, 712,2248, 321, 442, 503, 767, 744, 941,1237,1178, -1475,2249, 82, 178,1141,1179, 973,2250,1302,2251, 297,2252,2253, 570,2254,2255, -2256, 18, 450, 206,2257, 290, 292,1142,2258, 511, 162, 99, 346, 164, 735,2259, -1476,1477, 4, 554, 343, 798,1099,2260,1100,2261, 43, 171,1303, 139, 215,2262, -2263, 717, 775,2264,1033, 322, 216,2265, 831,2266, 149,2267,1304,2268,2269, 702, -1238, 135, 845, 347, 309,2270, 484,2271, 878, 655, 238,1006,1478,2272, 67,2273, - 295,2274,2275, 461,2276, 478, 942, 412,2277,1034,2278,2279,2280, 265,2281, 541, -2282,2283,2284,2285,2286, 70, 852,1071,2287,2288,2289,2290, 21, 56, 509, 117, - 432,2291,2292, 331, 980, 552,1101, 148, 284, 105, 393,1180,1239, 755,2293, 187, -2294,1046,1479,2295, 340,2296, 63,1047, 230,2297,2298,1305, 763,1306, 101, 800, - 808, 494,2299,2300,2301, 903,2302, 37,1072, 14, 5,2303, 79, 675,2304, 312, -2305,2306,2307,2308,2309,1480, 6,1307,2310,2311,2312, 1, 470, 35, 24, 229, -2313, 695, 210, 86, 778, 15, 784, 592, 779, 32, 77, 855, 964,2314, 259,2315, - 501, 380,2316,2317, 83, 981, 153, 689,1308,1481,1482,1483,2318,2319, 716,1484, -2320,2321,2322,2323,2324,2325,1485,2326,2327, 128, 57, 68, 261,1048, 211, 170, -1240, 31,2328, 51, 435, 742,2329,2330,2331, 635,2332, 264, 456,2333,2334,2335, - 425,2336,1486, 143, 507, 263, 943,2337, 363, 920,1487, 256,1488,1102, 243, 601, -1489,2338,2339,2340,2341,2342,2343,2344, 861,2345,2346,2347,2348,2349,2350, 395, -2351,1490,1491, 62, 535, 166, 225,2352,2353, 668, 419,1241, 138, 604, 928,2354, -1181,2355,1492,1493,2356,2357,2358,1143,2359, 696,2360, 387, 307,1309, 682, 476, -2361,2362, 332, 12, 222, 156,2363, 232,2364, 641, 276, 656, 517,1494,1495,1035, - 416, 736,1496,2365,1017, 586,2366,2367,2368,1497,2369, 242,2370,2371,2372,1498, -2373, 965, 713,2374,2375,2376,2377, 740, 982,1499, 944,1500,1007,2378,2379,1310, -1501,2380,2381,2382, 785, 329,2383,2384,1502,2385,2386,2387, 932,2388,1503,2389, -2390,2391,2392,1242,2393,2394,2395,2396,2397, 994, 950,2398,2399,2400,2401,1504, -1311,2402,2403,2404,2405,1049, 749,2406,2407, 853, 718,1144,1312,2408,1182,1505, -2409,2410, 255, 516, 479, 564, 550, 214,1506,1507,1313, 413, 239, 444, 339,1145, -1036,1508,1509,1314,1037,1510,1315,2411,1511,2412,2413,2414, 176, 703, 497, 624, - 593, 921, 302,2415, 341, 165,1103,1512,2416,1513,2417,2418,2419, 376,2420, 700, -2421,2422,2423, 258, 768,1316,2424,1183,2425, 995, 608,2426,2427,2428,2429, 221, -2430,2431,2432,2433,2434,2435,2436,2437, 195, 323, 726, 188, 897, 983,1317, 377, - 644,1050, 879,2438, 452,2439,2440,2441,2442,2443,2444, 914,2445,2446,2447,2448, - 915, 489,2449,1514,1184,2450,2451, 515, 64, 427, 495,2452, 583,2453, 483, 485, -1038, 562, 213,1515, 748, 666,2454,2455,2456,2457, 334,2458, 780, 996,1008, 705, -1243,2459,2460,2461,2462,2463, 114,2464, 493,1146, 366, 163,1516, 961,1104,2465, - 291,2466,1318,1105,2467,1517, 365,2468, 355, 951,1244,2469,1319,2470, 631,2471, -2472, 218,1320, 364, 320, 756,1518,1519,1321,1520,1322,2473,2474,2475,2476, 997, -2477,2478,2479,2480, 665,1185,2481, 916,1521,2482,2483,2484, 584, 684,2485,2486, - 797,2487,1051,1186,2488,2489,2490,1522,2491,2492, 370,2493,1039,1187, 65,2494, - 434, 205, 463,1188,2495, 125, 812, 391, 402, 826, 699, 286, 398, 155, 781, 771, - 585,2496, 590, 505,1073,2497, 599, 244, 219, 917,1018, 952, 646,1523,2498,1323, -2499,2500, 49, 984, 354, 741,2501, 625,2502,1324,2503,1019, 190, 357, 757, 491, - 95, 782, 868,2504,2505,2506,2507,2508,2509, 134,1524,1074, 422,1525, 898,2510, - 161,2511,2512,2513,2514, 769,2515,1526,2516,2517, 411,1325,2518, 472,1527,2519, -2520,2521,2522,2523,2524, 985,2525,2526,2527,2528,2529,2530, 764,2531,1245,2532, -2533, 25, 204, 311,2534, 496,2535,1052,2536,2537,2538,2539,2540,2541,2542, 199, - 704, 504, 468, 758, 657,1528, 196, 44, 839,1246, 272, 750,2543, 765, 862,2544, -2545,1326,2546, 132, 615, 933,2547, 732,2548,2549,2550,1189,1529,2551, 283,1247, -1053, 607, 929,2552,2553,2554, 930, 183, 872, 616,1040,1147,2555,1148,1020, 441, - 249,1075,2556,2557,2558, 466, 743,2559,2560,2561, 92, 514, 426, 420, 526,2562, -2563,2564,2565,2566,2567,2568, 185,2569,2570,2571,2572, 776,1530, 658,2573, 362, -2574, 361, 922,1076, 793,2575,2576,2577,2578,2579,2580,1531, 251,2581,2582,2583, -2584,1532, 54, 612, 237,1327,2585,2586, 275, 408, 647, 111,2587,1533,1106, 465, - 3, 458, 9, 38,2588, 107, 110, 890, 209, 26, 737, 498,2589,1534,2590, 431, - 202, 88,1535, 356, 287,1107, 660,1149,2591, 381,1536, 986,1150, 445,1248,1151, - 974,2592,2593, 846,2594, 446, 953, 184,1249,1250, 727,2595, 923, 193, 883,2596, -2597,2598, 102, 324, 539, 817,2599, 421,1041,2600, 832,2601, 94, 175, 197, 406, -2602, 459,2603,2604,2605,2606,2607, 330, 555,2608,2609,2610, 706,1108, 389,2611, -2612,2613,2614, 233,2615, 833, 558, 931, 954,1251,2616,2617,1537, 546,2618,2619, -1009,2620,2621,2622,1538, 690,1328,2623, 955,2624,1539,2625,2626, 772,2627,2628, -2629,2630,2631, 924, 648, 863, 603,2632,2633, 934,1540, 864, 865,2634, 642,1042, - 670,1190,2635,2636,2637,2638, 168,2639, 652, 873, 542,1054,1541,2640,2641,2642, # 512, 256 -) - diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/euckrprober.py b/src/fetchcode/vcs/pip/_vendor/chardet/euckrprober.py deleted file mode 100644 index 345a060d..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/euckrprober.py +++ /dev/null @@ -1,47 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is mozilla.org code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -from .mbcharsetprober import MultiByteCharSetProber -from .codingstatemachine import CodingStateMachine -from .chardistribution import EUCKRDistributionAnalysis -from .mbcssm import EUCKR_SM_MODEL - - -class EUCKRProber(MultiByteCharSetProber): - def __init__(self): - super(EUCKRProber, self).__init__() - self.coding_sm = CodingStateMachine(EUCKR_SM_MODEL) - self.distribution_analyzer = EUCKRDistributionAnalysis() - self.reset() - - @property - def charset_name(self): - return "EUC-KR" - - @property - def language(self): - return "Korean" diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/euctwfreq.py b/src/fetchcode/vcs/pip/_vendor/chardet/euctwfreq.py deleted file mode 100644 index ed7a995a..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/euctwfreq.py +++ /dev/null @@ -1,387 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Communicator client code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -# EUCTW frequency table -# Converted from big5 work -# by Taiwan's Mandarin Promotion Council -# - -# 128 --> 0.42261 -# 256 --> 0.57851 -# 512 --> 0.74851 -# 1024 --> 0.89384 -# 2048 --> 0.97583 -# -# Idea Distribution Ratio = 0.74851/(1-0.74851) =2.98 -# Random Distribution Ration = 512/(5401-512)=0.105 -# -# Typical Distribution Ratio about 25% of Ideal one, still much higher than RDR - -EUCTW_TYPICAL_DISTRIBUTION_RATIO = 0.75 - -# Char to FreqOrder table , -EUCTW_TABLE_SIZE = 5376 - -EUCTW_CHAR_TO_FREQ_ORDER = ( - 1,1800,1506, 255,1431, 198, 9, 82, 6,7310, 177, 202,3615,1256,2808, 110, # 2742 -3735, 33,3241, 261, 76, 44,2113, 16,2931,2184,1176, 659,3868, 26,3404,2643, # 2758 -1198,3869,3313,4060, 410,2211, 302, 590, 361,1963, 8, 204, 58,4296,7311,1931, # 2774 - 63,7312,7313, 317,1614, 75, 222, 159,4061,2412,1480,7314,3500,3068, 224,2809, # 2790 -3616, 3, 10,3870,1471, 29,2774,1135,2852,1939, 873, 130,3242,1123, 312,7315, # 2806 -4297,2051, 507, 252, 682,7316, 142,1914, 124, 206,2932, 34,3501,3173, 64, 604, # 2822 -7317,2494,1976,1977, 155,1990, 645, 641,1606,7318,3405, 337, 72, 406,7319, 80, # 2838 - 630, 238,3174,1509, 263, 939,1092,2644, 756,1440,1094,3406, 449, 69,2969, 591, # 2854 - 179,2095, 471, 115,2034,1843, 60, 50,2970, 134, 806,1868, 734,2035,3407, 180, # 2870 - 995,1607, 156, 537,2893, 688,7320, 319,1305, 779,2144, 514,2374, 298,4298, 359, # 2886 -2495, 90,2707,1338, 663, 11, 906,1099,2545, 20,2436, 182, 532,1716,7321, 732, # 2902 -1376,4062,1311,1420,3175, 25,2312,1056, 113, 399, 382,1949, 242,3408,2467, 529, # 2918 -3243, 475,1447,3617,7322, 117, 21, 656, 810,1297,2295,2329,3502,7323, 126,4063, # 2934 - 706, 456, 150, 613,4299, 71,1118,2036,4064, 145,3069, 85, 835, 486,2114,1246, # 2950 -1426, 428, 727,1285,1015, 800, 106, 623, 303,1281,7324,2127,2354, 347,3736, 221, # 2966 -3503,3110,7325,1955,1153,4065, 83, 296,1199,3070, 192, 624, 93,7326, 822,1897, # 2982 -2810,3111, 795,2064, 991,1554,1542,1592, 27, 43,2853, 859, 139,1456, 860,4300, # 2998 - 437, 712,3871, 164,2392,3112, 695, 211,3017,2096, 195,3872,1608,3504,3505,3618, # 3014 -3873, 234, 811,2971,2097,3874,2229,1441,3506,1615,2375, 668,2076,1638, 305, 228, # 3030 -1664,4301, 467, 415,7327, 262,2098,1593, 239, 108, 300, 200,1033, 512,1247,2077, # 3046 -7328,7329,2173,3176,3619,2673, 593, 845,1062,3244, 88,1723,2037,3875,1950, 212, # 3062 - 266, 152, 149, 468,1898,4066,4302, 77, 187,7330,3018, 37, 5,2972,7331,3876, # 3078 -7332,7333, 39,2517,4303,2894,3177,2078, 55, 148, 74,4304, 545, 483,1474,1029, # 3094 -1665, 217,1869,1531,3113,1104,2645,4067, 24, 172,3507, 900,3877,3508,3509,4305, # 3110 - 32,1408,2811,1312, 329, 487,2355,2247,2708, 784,2674, 4,3019,3314,1427,1788, # 3126 - 188, 109, 499,7334,3620,1717,1789, 888,1217,3020,4306,7335,3510,7336,3315,1520, # 3142 -3621,3878, 196,1034, 775,7337,7338, 929,1815, 249, 439, 38,7339,1063,7340, 794, # 3158 -3879,1435,2296, 46, 178,3245,2065,7341,2376,7342, 214,1709,4307, 804, 35, 707, # 3174 - 324,3622,1601,2546, 140, 459,4068,7343,7344,1365, 839, 272, 978,2257,2572,3409, # 3190 -2128,1363,3623,1423, 697, 100,3071, 48, 70,1231, 495,3114,2193,7345,1294,7346, # 3206 -2079, 462, 586,1042,3246, 853, 256, 988, 185,2377,3410,1698, 434,1084,7347,3411, # 3222 - 314,2615,2775,4308,2330,2331, 569,2280, 637,1816,2518, 757,1162,1878,1616,3412, # 3238 - 287,1577,2115, 768,4309,1671,2854,3511,2519,1321,3737, 909,2413,7348,4069, 933, # 3254 -3738,7349,2052,2356,1222,4310, 765,2414,1322, 786,4311,7350,1919,1462,1677,2895, # 3270 -1699,7351,4312,1424,2437,3115,3624,2590,3316,1774,1940,3413,3880,4070, 309,1369, # 3286 -1130,2812, 364,2230,1653,1299,3881,3512,3882,3883,2646, 525,1085,3021, 902,2000, # 3302 -1475, 964,4313, 421,1844,1415,1057,2281, 940,1364,3116, 376,4314,4315,1381, 7, # 3318 -2520, 983,2378, 336,1710,2675,1845, 321,3414, 559,1131,3022,2742,1808,1132,1313, # 3334 - 265,1481,1857,7352, 352,1203,2813,3247, 167,1089, 420,2814, 776, 792,1724,3513, # 3350 -4071,2438,3248,7353,4072,7354, 446, 229, 333,2743, 901,3739,1200,1557,4316,2647, # 3366 -1920, 395,2744,2676,3740,4073,1835, 125, 916,3178,2616,4317,7355,7356,3741,7357, # 3382 -7358,7359,4318,3117,3625,1133,2547,1757,3415,1510,2313,1409,3514,7360,2145, 438, # 3398 -2591,2896,2379,3317,1068, 958,3023, 461, 311,2855,2677,4074,1915,3179,4075,1978, # 3414 - 383, 750,2745,2617,4076, 274, 539, 385,1278,1442,7361,1154,1964, 384, 561, 210, # 3430 - 98,1295,2548,3515,7362,1711,2415,1482,3416,3884,2897,1257, 129,7363,3742, 642, # 3446 - 523,2776,2777,2648,7364, 141,2231,1333, 68, 176, 441, 876, 907,4077, 603,2592, # 3462 - 710, 171,3417, 404, 549, 18,3118,2393,1410,3626,1666,7365,3516,4319,2898,4320, # 3478 -7366,2973, 368,7367, 146, 366, 99, 871,3627,1543, 748, 807,1586,1185, 22,2258, # 3494 - 379,3743,3180,7368,3181, 505,1941,2618,1991,1382,2314,7369, 380,2357, 218, 702, # 3510 -1817,1248,3418,3024,3517,3318,3249,7370,2974,3628, 930,3250,3744,7371, 59,7372, # 3526 - 585, 601,4078, 497,3419,1112,1314,4321,1801,7373,1223,1472,2174,7374, 749,1836, # 3542 - 690,1899,3745,1772,3885,1476, 429,1043,1790,2232,2116, 917,4079, 447,1086,1629, # 3558 -7375, 556,7376,7377,2020,1654, 844,1090, 105, 550, 966,1758,2815,1008,1782, 686, # 3574 -1095,7378,2282, 793,1602,7379,3518,2593,4322,4080,2933,2297,4323,3746, 980,2496, # 3590 - 544, 353, 527,4324, 908,2678,2899,7380, 381,2619,1942,1348,7381,1341,1252, 560, # 3606 -3072,7382,3420,2856,7383,2053, 973, 886,2080, 143,4325,7384,7385, 157,3886, 496, # 3622 -4081, 57, 840, 540,2038,4326,4327,3421,2117,1445, 970,2259,1748,1965,2081,4082, # 3638 -3119,1234,1775,3251,2816,3629, 773,1206,2129,1066,2039,1326,3887,1738,1725,4083, # 3654 - 279,3120, 51,1544,2594, 423,1578,2130,2066, 173,4328,1879,7386,7387,1583, 264, # 3670 - 610,3630,4329,2439, 280, 154,7388,7389,7390,1739, 338,1282,3073, 693,2857,1411, # 3686 -1074,3747,2440,7391,4330,7392,7393,1240, 952,2394,7394,2900,1538,2679, 685,1483, # 3702 -4084,2468,1436, 953,4085,2054,4331, 671,2395, 79,4086,2441,3252, 608, 567,2680, # 3718 -3422,4087,4088,1691, 393,1261,1791,2396,7395,4332,7396,7397,7398,7399,1383,1672, # 3734 -3748,3182,1464, 522,1119, 661,1150, 216, 675,4333,3888,1432,3519, 609,4334,2681, # 3750 -2397,7400,7401,7402,4089,3025, 0,7403,2469, 315, 231,2442, 301,3319,4335,2380, # 3766 -7404, 233,4090,3631,1818,4336,4337,7405, 96,1776,1315,2082,7406, 257,7407,1809, # 3782 -3632,2709,1139,1819,4091,2021,1124,2163,2778,1777,2649,7408,3074, 363,1655,3183, # 3798 -7409,2975,7410,7411,7412,3889,1567,3890, 718, 103,3184, 849,1443, 341,3320,2934, # 3814 -1484,7413,1712, 127, 67, 339,4092,2398, 679,1412, 821,7414,7415, 834, 738, 351, # 3830 -2976,2146, 846, 235,1497,1880, 418,1992,3749,2710, 186,1100,2147,2746,3520,1545, # 3846 -1355,2935,2858,1377, 583,3891,4093,2573,2977,7416,1298,3633,1078,2549,3634,2358, # 3862 - 78,3750,3751, 267,1289,2099,2001,1594,4094, 348, 369,1274,2194,2175,1837,4338, # 3878 -1820,2817,3635,2747,2283,2002,4339,2936,2748, 144,3321, 882,4340,3892,2749,3423, # 3894 -4341,2901,7417,4095,1726, 320,7418,3893,3026, 788,2978,7419,2818,1773,1327,2859, # 3910 -3894,2819,7420,1306,4342,2003,1700,3752,3521,2359,2650, 787,2022, 506, 824,3636, # 3926 - 534, 323,4343,1044,3322,2023,1900, 946,3424,7421,1778,1500,1678,7422,1881,4344, # 3942 - 165, 243,4345,3637,2521, 123, 683,4096, 764,4346, 36,3895,1792, 589,2902, 816, # 3958 - 626,1667,3027,2233,1639,1555,1622,3753,3896,7423,3897,2860,1370,1228,1932, 891, # 3974 -2083,2903, 304,4097,7424, 292,2979,2711,3522, 691,2100,4098,1115,4347, 118, 662, # 3990 -7425, 611,1156, 854,2381,1316,2861, 2, 386, 515,2904,7426,7427,3253, 868,2234, # 4006 -1486, 855,2651, 785,2212,3028,7428,1040,3185,3523,7429,3121, 448,7430,1525,7431, # 4022 -2164,4348,7432,3754,7433,4099,2820,3524,3122, 503, 818,3898,3123,1568, 814, 676, # 4038 -1444, 306,1749,7434,3755,1416,1030, 197,1428, 805,2821,1501,4349,7435,7436,7437, # 4054 -1993,7438,4350,7439,7440,2195, 13,2779,3638,2980,3124,1229,1916,7441,3756,2131, # 4070 -7442,4100,4351,2399,3525,7443,2213,1511,1727,1120,7444,7445, 646,3757,2443, 307, # 4086 -7446,7447,1595,3186,7448,7449,7450,3639,1113,1356,3899,1465,2522,2523,7451, 519, # 4102 -7452, 128,2132, 92,2284,1979,7453,3900,1512, 342,3125,2196,7454,2780,2214,1980, # 4118 -3323,7455, 290,1656,1317, 789, 827,2360,7456,3758,4352, 562, 581,3901,7457, 401, # 4134 -4353,2248, 94,4354,1399,2781,7458,1463,2024,4355,3187,1943,7459, 828,1105,4101, # 4150 -1262,1394,7460,4102, 605,4356,7461,1783,2862,7462,2822, 819,2101, 578,2197,2937, # 4166 -7463,1502, 436,3254,4103,3255,2823,3902,2905,3425,3426,7464,2712,2315,7465,7466, # 4182 -2332,2067, 23,4357, 193, 826,3759,2102, 699,1630,4104,3075, 390,1793,1064,3526, # 4198 -7467,1579,3076,3077,1400,7468,4105,1838,1640,2863,7469,4358,4359, 137,4106, 598, # 4214 -3078,1966, 780, 104, 974,2938,7470, 278, 899, 253, 402, 572, 504, 493,1339,7471, # 4230 -3903,1275,4360,2574,2550,7472,3640,3029,3079,2249, 565,1334,2713, 863, 41,7473, # 4246 -7474,4361,7475,1657,2333, 19, 463,2750,4107, 606,7476,2981,3256,1087,2084,1323, # 4262 -2652,2982,7477,1631,1623,1750,4108,2682,7478,2864, 791,2714,2653,2334, 232,2416, # 4278 -7479,2983,1498,7480,2654,2620, 755,1366,3641,3257,3126,2025,1609, 119,1917,3427, # 4294 - 862,1026,4109,7481,3904,3760,4362,3905,4363,2260,1951,2470,7482,1125, 817,4110, # 4310 -4111,3906,1513,1766,2040,1487,4112,3030,3258,2824,3761,3127,7483,7484,1507,7485, # 4326 -2683, 733, 40,1632,1106,2865, 345,4113, 841,2524, 230,4364,2984,1846,3259,3428, # 4342 -7486,1263, 986,3429,7487, 735, 879, 254,1137, 857, 622,1300,1180,1388,1562,3907, # 4358 -3908,2939, 967,2751,2655,1349, 592,2133,1692,3324,2985,1994,4114,1679,3909,1901, # 4374 -2185,7488, 739,3642,2715,1296,1290,7489,4115,2198,2199,1921,1563,2595,2551,1870, # 4390 -2752,2986,7490, 435,7491, 343,1108, 596, 17,1751,4365,2235,3430,3643,7492,4366, # 4406 - 294,3527,2940,1693, 477, 979, 281,2041,3528, 643,2042,3644,2621,2782,2261,1031, # 4422 -2335,2134,2298,3529,4367, 367,1249,2552,7493,3530,7494,4368,1283,3325,2004, 240, # 4438 -1762,3326,4369,4370, 836,1069,3128, 474,7495,2148,2525, 268,3531,7496,3188,1521, # 4454 -1284,7497,1658,1546,4116,7498,3532,3533,7499,4117,3327,2684,1685,4118, 961,1673, # 4470 -2622, 190,2005,2200,3762,4371,4372,7500, 570,2497,3645,1490,7501,4373,2623,3260, # 4486 -1956,4374, 584,1514, 396,1045,1944,7502,4375,1967,2444,7503,7504,4376,3910, 619, # 4502 -7505,3129,3261, 215,2006,2783,2553,3189,4377,3190,4378, 763,4119,3763,4379,7506, # 4518 -7507,1957,1767,2941,3328,3646,1174, 452,1477,4380,3329,3130,7508,2825,1253,2382, # 4534 -2186,1091,2285,4120, 492,7509, 638,1169,1824,2135,1752,3911, 648, 926,1021,1324, # 4550 -4381, 520,4382, 997, 847,1007, 892,4383,3764,2262,1871,3647,7510,2400,1784,4384, # 4566 -1952,2942,3080,3191,1728,4121,2043,3648,4385,2007,1701,3131,1551, 30,2263,4122, # 4582 -7511,2026,4386,3534,7512, 501,7513,4123, 594,3431,2165,1821,3535,3432,3536,3192, # 4598 - 829,2826,4124,7514,1680,3132,1225,4125,7515,3262,4387,4126,3133,2336,7516,4388, # 4614 -4127,7517,3912,3913,7518,1847,2383,2596,3330,7519,4389, 374,3914, 652,4128,4129, # 4630 - 375,1140, 798,7520,7521,7522,2361,4390,2264, 546,1659, 138,3031,2445,4391,7523, # 4646 -2250, 612,1848, 910, 796,3765,1740,1371, 825,3766,3767,7524,2906,2554,7525, 692, # 4662 - 444,3032,2624, 801,4392,4130,7526,1491, 244,1053,3033,4131,4132, 340,7527,3915, # 4678 -1041,2987, 293,1168, 87,1357,7528,1539, 959,7529,2236, 721, 694,4133,3768, 219, # 4694 -1478, 644,1417,3331,2656,1413,1401,1335,1389,3916,7530,7531,2988,2362,3134,1825, # 4710 - 730,1515, 184,2827, 66,4393,7532,1660,2943, 246,3332, 378,1457, 226,3433, 975, # 4726 -3917,2944,1264,3537, 674, 696,7533, 163,7534,1141,2417,2166, 713,3538,3333,4394, # 4742 -3918,7535,7536,1186, 15,7537,1079,1070,7538,1522,3193,3539, 276,1050,2716, 758, # 4758 -1126, 653,2945,3263,7539,2337, 889,3540,3919,3081,2989, 903,1250,4395,3920,3434, # 4774 -3541,1342,1681,1718, 766,3264, 286, 89,2946,3649,7540,1713,7541,2597,3334,2990, # 4790 -7542,2947,2215,3194,2866,7543,4396,2498,2526, 181, 387,1075,3921, 731,2187,3335, # 4806 -7544,3265, 310, 313,3435,2299, 770,4134, 54,3034, 189,4397,3082,3769,3922,7545, # 4822 -1230,1617,1849, 355,3542,4135,4398,3336, 111,4136,3650,1350,3135,3436,3035,4137, # 4838 -2149,3266,3543,7546,2784,3923,3924,2991, 722,2008,7547,1071, 247,1207,2338,2471, # 4854 -1378,4399,2009, 864,1437,1214,4400, 373,3770,1142,2216, 667,4401, 442,2753,2555, # 4870 -3771,3925,1968,4138,3267,1839, 837, 170,1107, 934,1336,1882,7548,7549,2118,4139, # 4886 -2828, 743,1569,7550,4402,4140, 582,2384,1418,3437,7551,1802,7552, 357,1395,1729, # 4902 -3651,3268,2418,1564,2237,7553,3083,3772,1633,4403,1114,2085,4141,1532,7554, 482, # 4918 -2446,4404,7555,7556,1492, 833,1466,7557,2717,3544,1641,2829,7558,1526,1272,3652, # 4934 -4142,1686,1794, 416,2556,1902,1953,1803,7559,3773,2785,3774,1159,2316,7560,2867, # 4950 -4405,1610,1584,3036,2419,2754, 443,3269,1163,3136,7561,7562,3926,7563,4143,2499, # 4966 -3037,4406,3927,3137,2103,1647,3545,2010,1872,4144,7564,4145, 431,3438,7565, 250, # 4982 - 97, 81,4146,7566,1648,1850,1558, 160, 848,7567, 866, 740,1694,7568,2201,2830, # 4998 -3195,4147,4407,3653,1687, 950,2472, 426, 469,3196,3654,3655,3928,7569,7570,1188, # 5014 - 424,1995, 861,3546,4148,3775,2202,2685, 168,1235,3547,4149,7571,2086,1674,4408, # 5030 -3337,3270, 220,2557,1009,7572,3776, 670,2992, 332,1208, 717,7573,7574,3548,2447, # 5046 -3929,3338,7575, 513,7576,1209,2868,3339,3138,4409,1080,7577,7578,7579,7580,2527, # 5062 -3656,3549, 815,1587,3930,3931,7581,3550,3439,3777,1254,4410,1328,3038,1390,3932, # 5078 -1741,3933,3778,3934,7582, 236,3779,2448,3271,7583,7584,3657,3780,1273,3781,4411, # 5094 -7585, 308,7586,4412, 245,4413,1851,2473,1307,2575, 430, 715,2136,2449,7587, 270, # 5110 - 199,2869,3935,7588,3551,2718,1753, 761,1754, 725,1661,1840,4414,3440,3658,7589, # 5126 -7590, 587, 14,3272, 227,2598, 326, 480,2265, 943,2755,3552, 291, 650,1883,7591, # 5142 -1702,1226, 102,1547, 62,3441, 904,4415,3442,1164,4150,7592,7593,1224,1548,2756, # 5158 - 391, 498,1493,7594,1386,1419,7595,2055,1177,4416, 813, 880,1081,2363, 566,1145, # 5174 -4417,2286,1001,1035,2558,2599,2238, 394,1286,7596,7597,2068,7598, 86,1494,1730, # 5190 -3936, 491,1588, 745, 897,2948, 843,3340,3937,2757,2870,3273,1768, 998,2217,2069, # 5206 - 397,1826,1195,1969,3659,2993,3341, 284,7599,3782,2500,2137,2119,1903,7600,3938, # 5222 -2150,3939,4151,1036,3443,1904, 114,2559,4152, 209,1527,7601,7602,2949,2831,2625, # 5238 -2385,2719,3139, 812,2560,7603,3274,7604,1559, 737,1884,3660,1210, 885, 28,2686, # 5254 -3553,3783,7605,4153,1004,1779,4418,7606, 346,1981,2218,2687,4419,3784,1742, 797, # 5270 -1642,3940,1933,1072,1384,2151, 896,3941,3275,3661,3197,2871,3554,7607,2561,1958, # 5286 -4420,2450,1785,7608,7609,7610,3942,4154,1005,1308,3662,4155,2720,4421,4422,1528, # 5302 -2600, 161,1178,4156,1982, 987,4423,1101,4157, 631,3943,1157,3198,2420,1343,1241, # 5318 -1016,2239,2562, 372, 877,2339,2501,1160, 555,1934, 911,3944,7611, 466,1170, 169, # 5334 -1051,2907,2688,3663,2474,2994,1182,2011,2563,1251,2626,7612, 992,2340,3444,1540, # 5350 -2721,1201,2070,2401,1996,2475,7613,4424, 528,1922,2188,1503,1873,1570,2364,3342, # 5366 -3276,7614, 557,1073,7615,1827,3445,2087,2266,3140,3039,3084, 767,3085,2786,4425, # 5382 -1006,4158,4426,2341,1267,2176,3664,3199, 778,3945,3200,2722,1597,2657,7616,4427, # 5398 -7617,3446,7618,7619,7620,3277,2689,1433,3278, 131, 95,1504,3946, 723,4159,3141, # 5414 -1841,3555,2758,2189,3947,2027,2104,3665,7621,2995,3948,1218,7622,3343,3201,3949, # 5430 -4160,2576, 248,1634,3785, 912,7623,2832,3666,3040,3786, 654, 53,7624,2996,7625, # 5446 -1688,4428, 777,3447,1032,3950,1425,7626, 191, 820,2120,2833, 971,4429, 931,3202, # 5462 - 135, 664, 783,3787,1997, 772,2908,1935,3951,3788,4430,2909,3203, 282,2723, 640, # 5478 -1372,3448,1127, 922, 325,3344,7627,7628, 711,2044,7629,7630,3952,2219,2787,1936, # 5494 -3953,3345,2220,2251,3789,2300,7631,4431,3790,1258,3279,3954,3204,2138,2950,3955, # 5510 -3956,7632,2221, 258,3205,4432, 101,1227,7633,3280,1755,7634,1391,3281,7635,2910, # 5526 -2056, 893,7636,7637,7638,1402,4161,2342,7639,7640,3206,3556,7641,7642, 878,1325, # 5542 -1780,2788,4433, 259,1385,2577, 744,1183,2267,4434,7643,3957,2502,7644, 684,1024, # 5558 -4162,7645, 472,3557,3449,1165,3282,3958,3959, 322,2152, 881, 455,1695,1152,1340, # 5574 - 660, 554,2153,4435,1058,4436,4163, 830,1065,3346,3960,4437,1923,7646,1703,1918, # 5590 -7647, 932,2268, 122,7648,4438, 947, 677,7649,3791,2627, 297,1905,1924,2269,4439, # 5606 -2317,3283,7650,7651,4164,7652,4165, 84,4166, 112, 989,7653, 547,1059,3961, 701, # 5622 -3558,1019,7654,4167,7655,3450, 942, 639, 457,2301,2451, 993,2951, 407, 851, 494, # 5638 -4440,3347, 927,7656,1237,7657,2421,3348, 573,4168, 680, 921,2911,1279,1874, 285, # 5654 - 790,1448,1983, 719,2167,7658,7659,4441,3962,3963,1649,7660,1541, 563,7661,1077, # 5670 -7662,3349,3041,3451, 511,2997,3964,3965,3667,3966,1268,2564,3350,3207,4442,4443, # 5686 -7663, 535,1048,1276,1189,2912,2028,3142,1438,1373,2834,2952,1134,2012,7664,4169, # 5702 -1238,2578,3086,1259,7665, 700,7666,2953,3143,3668,4170,7667,4171,1146,1875,1906, # 5718 -4444,2601,3967, 781,2422, 132,1589, 203, 147, 273,2789,2402, 898,1786,2154,3968, # 5734 -3969,7668,3792,2790,7669,7670,4445,4446,7671,3208,7672,1635,3793, 965,7673,1804, # 5750 -2690,1516,3559,1121,1082,1329,3284,3970,1449,3794, 65,1128,2835,2913,2759,1590, # 5766 -3795,7674,7675, 12,2658, 45, 976,2579,3144,4447, 517,2528,1013,1037,3209,7676, # 5782 -3796,2836,7677,3797,7678,3452,7679,2602, 614,1998,2318,3798,3087,2724,2628,7680, # 5798 -2580,4172, 599,1269,7681,1810,3669,7682,2691,3088, 759,1060, 489,1805,3351,3285, # 5814 -1358,7683,7684,2386,1387,1215,2629,2252, 490,7685,7686,4173,1759,2387,2343,7687, # 5830 -4448,3799,1907,3971,2630,1806,3210,4449,3453,3286,2760,2344, 874,7688,7689,3454, # 5846 -3670,1858, 91,2914,3671,3042,3800,4450,7690,3145,3972,2659,7691,3455,1202,1403, # 5862 -3801,2954,2529,1517,2503,4451,3456,2504,7692,4452,7693,2692,1885,1495,1731,3973, # 5878 -2365,4453,7694,2029,7695,7696,3974,2693,1216, 237,2581,4174,2319,3975,3802,4454, # 5894 -4455,2694,3560,3457, 445,4456,7697,7698,7699,7700,2761, 61,3976,3672,1822,3977, # 5910 -7701, 687,2045, 935, 925, 405,2660, 703,1096,1859,2725,4457,3978,1876,1367,2695, # 5926 -3352, 918,2105,1781,2476, 334,3287,1611,1093,4458, 564,3146,3458,3673,3353, 945, # 5942 -2631,2057,4459,7702,1925, 872,4175,7703,3459,2696,3089, 349,4176,3674,3979,4460, # 5958 -3803,4177,3675,2155,3980,4461,4462,4178,4463,2403,2046, 782,3981, 400, 251,4179, # 5974 -1624,7704,7705, 277,3676, 299,1265, 476,1191,3804,2121,4180,4181,1109, 205,7706, # 5990 -2582,1000,2156,3561,1860,7707,7708,7709,4464,7710,4465,2565, 107,2477,2157,3982, # 6006 -3460,3147,7711,1533, 541,1301, 158, 753,4182,2872,3562,7712,1696, 370,1088,4183, # 6022 -4466,3563, 579, 327, 440, 162,2240, 269,1937,1374,3461, 968,3043, 56,1396,3090, # 6038 -2106,3288,3354,7713,1926,2158,4467,2998,7714,3564,7715,7716,3677,4468,2478,7717, # 6054 -2791,7718,1650,4469,7719,2603,7720,7721,3983,2661,3355,1149,3356,3984,3805,3985, # 6070 -7722,1076, 49,7723, 951,3211,3289,3290, 450,2837, 920,7724,1811,2792,2366,4184, # 6086 -1908,1138,2367,3806,3462,7725,3212,4470,1909,1147,1518,2423,4471,3807,7726,4472, # 6102 -2388,2604, 260,1795,3213,7727,7728,3808,3291, 708,7729,3565,1704,7730,3566,1351, # 6118 -1618,3357,2999,1886, 944,4185,3358,4186,3044,3359,4187,7731,3678, 422, 413,1714, # 6134 -3292, 500,2058,2345,4188,2479,7732,1344,1910, 954,7733,1668,7734,7735,3986,2404, # 6150 -4189,3567,3809,4190,7736,2302,1318,2505,3091, 133,3092,2873,4473, 629, 31,2838, # 6166 -2697,3810,4474, 850, 949,4475,3987,2955,1732,2088,4191,1496,1852,7737,3988, 620, # 6182 -3214, 981,1242,3679,3360,1619,3680,1643,3293,2139,2452,1970,1719,3463,2168,7738, # 6198 -3215,7739,7740,3361,1828,7741,1277,4476,1565,2047,7742,1636,3568,3093,7743, 869, # 6214 -2839, 655,3811,3812,3094,3989,3000,3813,1310,3569,4477,7744,7745,7746,1733, 558, # 6230 -4478,3681, 335,1549,3045,1756,4192,3682,1945,3464,1829,1291,1192, 470,2726,2107, # 6246 -2793, 913,1054,3990,7747,1027,7748,3046,3991,4479, 982,2662,3362,3148,3465,3216, # 6262 -3217,1946,2794,7749, 571,4480,7750,1830,7751,3570,2583,1523,2424,7752,2089, 984, # 6278 -4481,3683,1959,7753,3684, 852, 923,2795,3466,3685, 969,1519, 999,2048,2320,1705, # 6294 -7754,3095, 615,1662, 151, 597,3992,2405,2321,1049, 275,4482,3686,4193, 568,3687, # 6310 -3571,2480,4194,3688,7755,2425,2270, 409,3218,7756,1566,2874,3467,1002, 769,2840, # 6326 - 194,2090,3149,3689,2222,3294,4195, 628,1505,7757,7758,1763,2177,3001,3993, 521, # 6342 -1161,2584,1787,2203,2406,4483,3994,1625,4196,4197, 412, 42,3096, 464,7759,2632, # 6358 -4484,3363,1760,1571,2875,3468,2530,1219,2204,3814,2633,2140,2368,4485,4486,3295, # 6374 -1651,3364,3572,7760,7761,3573,2481,3469,7762,3690,7763,7764,2271,2091, 460,7765, # 6390 -4487,7766,3002, 962, 588,3574, 289,3219,2634,1116, 52,7767,3047,1796,7768,7769, # 6406 -7770,1467,7771,1598,1143,3691,4198,1984,1734,1067,4488,1280,3365, 465,4489,1572, # 6422 - 510,7772,1927,2241,1812,1644,3575,7773,4490,3692,7774,7775,2663,1573,1534,7776, # 6438 -7777,4199, 536,1807,1761,3470,3815,3150,2635,7778,7779,7780,4491,3471,2915,1911, # 6454 -2796,7781,3296,1122, 377,3220,7782, 360,7783,7784,4200,1529, 551,7785,2059,3693, # 6470 -1769,2426,7786,2916,4201,3297,3097,2322,2108,2030,4492,1404, 136,1468,1479, 672, # 6486 -1171,3221,2303, 271,3151,7787,2762,7788,2049, 678,2727, 865,1947,4493,7789,2013, # 6502 -3995,2956,7790,2728,2223,1397,3048,3694,4494,4495,1735,2917,3366,3576,7791,3816, # 6518 - 509,2841,2453,2876,3817,7792,7793,3152,3153,4496,4202,2531,4497,2304,1166,1010, # 6534 - 552, 681,1887,7794,7795,2957,2958,3996,1287,1596,1861,3154, 358, 453, 736, 175, # 6550 - 478,1117, 905,1167,1097,7796,1853,1530,7797,1706,7798,2178,3472,2287,3695,3473, # 6566 -3577,4203,2092,4204,7799,3367,1193,2482,4205,1458,2190,2205,1862,1888,1421,3298, # 6582 -2918,3049,2179,3474, 595,2122,7800,3997,7801,7802,4206,1707,2636, 223,3696,1359, # 6598 - 751,3098, 183,3475,7803,2797,3003, 419,2369, 633, 704,3818,2389, 241,7804,7805, # 6614 -7806, 838,3004,3697,2272,2763,2454,3819,1938,2050,3998,1309,3099,2242,1181,7807, # 6630 -1136,2206,3820,2370,1446,4207,2305,4498,7808,7809,4208,1055,2605, 484,3698,7810, # 6646 -3999, 625,4209,2273,3368,1499,4210,4000,7811,4001,4211,3222,2274,2275,3476,7812, # 6662 -7813,2764, 808,2606,3699,3369,4002,4212,3100,2532, 526,3370,3821,4213, 955,7814, # 6678 -1620,4214,2637,2427,7815,1429,3700,1669,1831, 994, 928,7816,3578,1260,7817,7818, # 6694 -7819,1948,2288, 741,2919,1626,4215,2729,2455, 867,1184, 362,3371,1392,7820,7821, # 6710 -4003,4216,1770,1736,3223,2920,4499,4500,1928,2698,1459,1158,7822,3050,3372,2877, # 6726 -1292,1929,2506,2842,3701,1985,1187,2071,2014,2607,4217,7823,2566,2507,2169,3702, # 6742 -2483,3299,7824,3703,4501,7825,7826, 666,1003,3005,1022,3579,4218,7827,4502,1813, # 6758 -2253, 574,3822,1603, 295,1535, 705,3823,4219, 283, 858, 417,7828,7829,3224,4503, # 6774 -4504,3051,1220,1889,1046,2276,2456,4004,1393,1599, 689,2567, 388,4220,7830,2484, # 6790 - 802,7831,2798,3824,2060,1405,2254,7832,4505,3825,2109,1052,1345,3225,1585,7833, # 6806 - 809,7834,7835,7836, 575,2730,3477, 956,1552,1469,1144,2323,7837,2324,1560,2457, # 6822 -3580,3226,4005, 616,2207,3155,2180,2289,7838,1832,7839,3478,4506,7840,1319,3704, # 6838 -3705,1211,3581,1023,3227,1293,2799,7841,7842,7843,3826, 607,2306,3827, 762,2878, # 6854 -1439,4221,1360,7844,1485,3052,7845,4507,1038,4222,1450,2061,2638,4223,1379,4508, # 6870 -2585,7846,7847,4224,1352,1414,2325,2921,1172,7848,7849,3828,3829,7850,1797,1451, # 6886 -7851,7852,7853,7854,2922,4006,4007,2485,2346, 411,4008,4009,3582,3300,3101,4509, # 6902 -1561,2664,1452,4010,1375,7855,7856, 47,2959, 316,7857,1406,1591,2923,3156,7858, # 6918 -1025,2141,3102,3157, 354,2731, 884,2224,4225,2407, 508,3706, 726,3583, 996,2428, # 6934 -3584, 729,7859, 392,2191,1453,4011,4510,3707,7860,7861,2458,3585,2608,1675,2800, # 6950 - 919,2347,2960,2348,1270,4511,4012, 73,7862,7863, 647,7864,3228,2843,2255,1550, # 6966 -1346,3006,7865,1332, 883,3479,7866,7867,7868,7869,3301,2765,7870,1212, 831,1347, # 6982 -4226,4512,2326,3830,1863,3053, 720,3831,4513,4514,3832,7871,4227,7872,7873,4515, # 6998 -7874,7875,1798,4516,3708,2609,4517,3586,1645,2371,7876,7877,2924, 669,2208,2665, # 7014 -2429,7878,2879,7879,7880,1028,3229,7881,4228,2408,7882,2256,1353,7883,7884,4518, # 7030 -3158, 518,7885,4013,7886,4229,1960,7887,2142,4230,7888,7889,3007,2349,2350,3833, # 7046 - 516,1833,1454,4014,2699,4231,4519,2225,2610,1971,1129,3587,7890,2766,7891,2961, # 7062 -1422, 577,1470,3008,1524,3373,7892,7893, 432,4232,3054,3480,7894,2586,1455,2508, # 7078 -2226,1972,1175,7895,1020,2732,4015,3481,4520,7896,2733,7897,1743,1361,3055,3482, # 7094 -2639,4016,4233,4521,2290, 895, 924,4234,2170, 331,2243,3056, 166,1627,3057,1098, # 7110 -7898,1232,2880,2227,3374,4522, 657, 403,1196,2372, 542,3709,3375,1600,4235,3483, # 7126 -7899,4523,2767,3230, 576, 530,1362,7900,4524,2533,2666,3710,4017,7901, 842,3834, # 7142 -7902,2801,2031,1014,4018, 213,2700,3376, 665, 621,4236,7903,3711,2925,2430,7904, # 7158 -2431,3302,3588,3377,7905,4237,2534,4238,4525,3589,1682,4239,3484,1380,7906, 724, # 7174 -2277, 600,1670,7907,1337,1233,4526,3103,2244,7908,1621,4527,7909, 651,4240,7910, # 7190 -1612,4241,2611,7911,2844,7912,2734,2307,3058,7913, 716,2459,3059, 174,1255,2701, # 7206 -4019,3590, 548,1320,1398, 728,4020,1574,7914,1890,1197,3060,4021,7915,3061,3062, # 7222 -3712,3591,3713, 747,7916, 635,4242,4528,7917,7918,7919,4243,7920,7921,4529,7922, # 7238 -3378,4530,2432, 451,7923,3714,2535,2072,4244,2735,4245,4022,7924,1764,4531,7925, # 7254 -4246, 350,7926,2278,2390,2486,7927,4247,4023,2245,1434,4024, 488,4532, 458,4248, # 7270 -4025,3715, 771,1330,2391,3835,2568,3159,2159,2409,1553,2667,3160,4249,7928,2487, # 7286 -2881,2612,1720,2702,4250,3379,4533,7929,2536,4251,7930,3231,4252,2768,7931,2015, # 7302 -2736,7932,1155,1017,3716,3836,7933,3303,2308, 201,1864,4253,1430,7934,4026,7935, # 7318 -7936,7937,7938,7939,4254,1604,7940, 414,1865, 371,2587,4534,4535,3485,2016,3104, # 7334 -4536,1708, 960,4255, 887, 389,2171,1536,1663,1721,7941,2228,4027,2351,2926,1580, # 7350 -7942,7943,7944,1744,7945,2537,4537,4538,7946,4539,7947,2073,7948,7949,3592,3380, # 7366 -2882,4256,7950,4257,2640,3381,2802, 673,2703,2460, 709,3486,4028,3593,4258,7951, # 7382 -1148, 502, 634,7952,7953,1204,4540,3594,1575,4541,2613,3717,7954,3718,3105, 948, # 7398 -3232, 121,1745,3837,1110,7955,4259,3063,2509,3009,4029,3719,1151,1771,3838,1488, # 7414 -4030,1986,7956,2433,3487,7957,7958,2093,7959,4260,3839,1213,1407,2803, 531,2737, # 7430 -2538,3233,1011,1537,7960,2769,4261,3106,1061,7961,3720,3721,1866,2883,7962,2017, # 7446 - 120,4262,4263,2062,3595,3234,2309,3840,2668,3382,1954,4542,7963,7964,3488,1047, # 7462 -2704,1266,7965,1368,4543,2845, 649,3383,3841,2539,2738,1102,2846,2669,7966,7967, # 7478 -1999,7968,1111,3596,2962,7969,2488,3842,3597,2804,1854,3384,3722,7970,7971,3385, # 7494 -2410,2884,3304,3235,3598,7972,2569,7973,3599,2805,4031,1460, 856,7974,3600,7975, # 7510 -2885,2963,7976,2886,3843,7977,4264, 632,2510, 875,3844,1697,3845,2291,7978,7979, # 7526 -4544,3010,1239, 580,4545,4265,7980, 914, 936,2074,1190,4032,1039,2123,7981,7982, # 7542 -7983,3386,1473,7984,1354,4266,3846,7985,2172,3064,4033, 915,3305,4267,4268,3306, # 7558 -1605,1834,7986,2739, 398,3601,4269,3847,4034, 328,1912,2847,4035,3848,1331,4270, # 7574 -3011, 937,4271,7987,3602,4036,4037,3387,2160,4546,3388, 524, 742, 538,3065,1012, # 7590 -7988,7989,3849,2461,7990, 658,1103, 225,3850,7991,7992,4547,7993,4548,7994,3236, # 7606 -1243,7995,4038, 963,2246,4549,7996,2705,3603,3161,7997,7998,2588,2327,7999,4550, # 7622 -8000,8001,8002,3489,3307, 957,3389,2540,2032,1930,2927,2462, 870,2018,3604,1746, # 7638 -2770,2771,2434,2463,8003,3851,8004,3723,3107,3724,3490,3390,3725,8005,1179,3066, # 7654 -8006,3162,2373,4272,3726,2541,3163,3108,2740,4039,8007,3391,1556,2542,2292, 977, # 7670 -2887,2033,4040,1205,3392,8008,1765,3393,3164,2124,1271,1689, 714,4551,3491,8009, # 7686 -2328,3852, 533,4273,3605,2181, 617,8010,2464,3308,3492,2310,8011,8012,3165,8013, # 7702 -8014,3853,1987, 618, 427,2641,3493,3394,8015,8016,1244,1690,8017,2806,4274,4552, # 7718 -8018,3494,8019,8020,2279,1576, 473,3606,4275,3395, 972,8021,3607,8022,3067,8023, # 7734 -8024,4553,4554,8025,3727,4041,4042,8026, 153,4555, 356,8027,1891,2888,4276,2143, # 7750 - 408, 803,2352,8028,3854,8029,4277,1646,2570,2511,4556,4557,3855,8030,3856,4278, # 7766 -8031,2411,3396, 752,8032,8033,1961,2964,8034, 746,3012,2465,8035,4279,3728, 698, # 7782 -4558,1892,4280,3608,2543,4559,3609,3857,8036,3166,3397,8037,1823,1302,4043,2706, # 7798 -3858,1973,4281,8038,4282,3167, 823,1303,1288,1236,2848,3495,4044,3398, 774,3859, # 7814 -8039,1581,4560,1304,2849,3860,4561,8040,2435,2161,1083,3237,4283,4045,4284, 344, # 7830 -1173, 288,2311, 454,1683,8041,8042,1461,4562,4046,2589,8043,8044,4563, 985, 894, # 7846 -8045,3399,3168,8046,1913,2928,3729,1988,8047,2110,1974,8048,4047,8049,2571,1194, # 7862 - 425,8050,4564,3169,1245,3730,4285,8051,8052,2850,8053, 636,4565,1855,3861, 760, # 7878 -1799,8054,4286,2209,1508,4566,4048,1893,1684,2293,8055,8056,8057,4287,4288,2210, # 7894 - 479,8058,8059, 832,8060,4049,2489,8061,2965,2490,3731, 990,3109, 627,1814,2642, # 7910 -4289,1582,4290,2125,2111,3496,4567,8062, 799,4291,3170,8063,4568,2112,1737,3013, # 7926 -1018, 543, 754,4292,3309,1676,4569,4570,4050,8064,1489,8065,3497,8066,2614,2889, # 7942 -4051,8067,8068,2966,8069,8070,8071,8072,3171,4571,4572,2182,1722,8073,3238,3239, # 7958 -1842,3610,1715, 481, 365,1975,1856,8074,8075,1962,2491,4573,8076,2126,3611,3240, # 7974 - 433,1894,2063,2075,8077, 602,2741,8078,8079,8080,8081,8082,3014,1628,3400,8083, # 7990 -3172,4574,4052,2890,4575,2512,8084,2544,2772,8085,8086,8087,3310,4576,2891,8088, # 8006 -4577,8089,2851,4578,4579,1221,2967,4053,2513,8090,8091,8092,1867,1989,8093,8094, # 8022 -8095,1895,8096,8097,4580,1896,4054, 318,8098,2094,4055,4293,8099,8100, 485,8101, # 8038 - 938,3862, 553,2670, 116,8102,3863,3612,8103,3498,2671,2773,3401,3311,2807,8104, # 8054 -3613,2929,4056,1747,2930,2968,8105,8106, 207,8107,8108,2672,4581,2514,8109,3015, # 8070 - 890,3614,3864,8110,1877,3732,3402,8111,2183,2353,3403,1652,8112,8113,8114, 941, # 8086 -2294, 208,3499,4057,2019, 330,4294,3865,2892,2492,3733,4295,8115,8116,8117,8118, # 8102 -) - diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/euctwprober.py b/src/fetchcode/vcs/pip/_vendor/chardet/euctwprober.py deleted file mode 100644 index 35669cc4..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/euctwprober.py +++ /dev/null @@ -1,46 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is mozilla.org code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -from .mbcharsetprober import MultiByteCharSetProber -from .codingstatemachine import CodingStateMachine -from .chardistribution import EUCTWDistributionAnalysis -from .mbcssm import EUCTW_SM_MODEL - -class EUCTWProber(MultiByteCharSetProber): - def __init__(self): - super(EUCTWProber, self).__init__() - self.coding_sm = CodingStateMachine(EUCTW_SM_MODEL) - self.distribution_analyzer = EUCTWDistributionAnalysis() - self.reset() - - @property - def charset_name(self): - return "EUC-TW" - - @property - def language(self): - return "Taiwan" diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/gb2312freq.py b/src/fetchcode/vcs/pip/_vendor/chardet/gb2312freq.py deleted file mode 100644 index 697837bd..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/gb2312freq.py +++ /dev/null @@ -1,283 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Communicator client code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -# GB2312 most frequently used character table -# -# Char to FreqOrder table , from hz6763 - -# 512 --> 0.79 -- 0.79 -# 1024 --> 0.92 -- 0.13 -# 2048 --> 0.98 -- 0.06 -# 6768 --> 1.00 -- 0.02 -# -# Ideal Distribution Ratio = 0.79135/(1-0.79135) = 3.79 -# Random Distribution Ration = 512 / (3755 - 512) = 0.157 -# -# Typical Distribution Ratio about 25% of Ideal one, still much higher that RDR - -GB2312_TYPICAL_DISTRIBUTION_RATIO = 0.9 - -GB2312_TABLE_SIZE = 3760 - -GB2312_CHAR_TO_FREQ_ORDER = ( -1671, 749,1443,2364,3924,3807,2330,3921,1704,3463,2691,1511,1515, 572,3191,2205, -2361, 224,2558, 479,1711, 963,3162, 440,4060,1905,2966,2947,3580,2647,3961,3842, -2204, 869,4207, 970,2678,5626,2944,2956,1479,4048, 514,3595, 588,1346,2820,3409, - 249,4088,1746,1873,2047,1774, 581,1813, 358,1174,3590,1014,1561,4844,2245, 670, -1636,3112, 889,1286, 953, 556,2327,3060,1290,3141, 613, 185,3477,1367, 850,3820, -1715,2428,2642,2303,2732,3041,2562,2648,3566,3946,1349, 388,3098,2091,1360,3585, - 152,1687,1539, 738,1559, 59,1232,2925,2267,1388,1249,1741,1679,2960, 151,1566, -1125,1352,4271, 924,4296, 385,3166,4459, 310,1245,2850, 70,3285,2729,3534,3575, -2398,3298,3466,1960,2265, 217,3647, 864,1909,2084,4401,2773,1010,3269,5152, 853, -3051,3121,1244,4251,1895, 364,1499,1540,2313,1180,3655,2268, 562, 715,2417,3061, - 544, 336,3768,2380,1752,4075, 950, 280,2425,4382, 183,2759,3272, 333,4297,2155, -1688,2356,1444,1039,4540, 736,1177,3349,2443,2368,2144,2225, 565, 196,1482,3406, - 927,1335,4147, 692, 878,1311,1653,3911,3622,1378,4200,1840,2969,3149,2126,1816, -2534,1546,2393,2760, 737,2494, 13, 447, 245,2747, 38,2765,2129,2589,1079, 606, - 360, 471,3755,2890, 404, 848, 699,1785,1236, 370,2221,1023,3746,2074,2026,2023, -2388,1581,2119, 812,1141,3091,2536,1519, 804,2053, 406,1596,1090, 784, 548,4414, -1806,2264,2936,1100, 343,4114,5096, 622,3358, 743,3668,1510,1626,5020,3567,2513, -3195,4115,5627,2489,2991, 24,2065,2697,1087,2719, 48,1634, 315, 68, 985,2052, - 198,2239,1347,1107,1439, 597,2366,2172, 871,3307, 919,2487,2790,1867, 236,2570, -1413,3794, 906,3365,3381,1701,1982,1818,1524,2924,1205, 616,2586,2072,2004, 575, - 253,3099, 32,1365,1182, 197,1714,2454,1201, 554,3388,3224,2748, 756,2587, 250, -2567,1507,1517,3529,1922,2761,2337,3416,1961,1677,2452,2238,3153, 615, 911,1506, -1474,2495,1265,1906,2749,3756,3280,2161, 898,2714,1759,3450,2243,2444, 563, 26, -3286,2266,3769,3344,2707,3677, 611,1402, 531,1028,2871,4548,1375, 261,2948, 835, -1190,4134, 353, 840,2684,1900,3082,1435,2109,1207,1674, 329,1872,2781,4055,2686, -2104, 608,3318,2423,2957,2768,1108,3739,3512,3271,3985,2203,1771,3520,1418,2054, -1681,1153, 225,1627,2929, 162,2050,2511,3687,1954, 124,1859,2431,1684,3032,2894, - 585,4805,3969,2869,2704,2088,2032,2095,3656,2635,4362,2209, 256, 518,2042,2105, -3777,3657, 643,2298,1148,1779, 190, 989,3544, 414, 11,2135,2063,2979,1471, 403, -3678, 126, 770,1563, 671,2499,3216,2877, 600,1179, 307,2805,4937,1268,1297,2694, - 252,4032,1448,1494,1331,1394, 127,2256, 222,1647,1035,1481,3056,1915,1048, 873, -3651, 210, 33,1608,2516, 200,1520, 415, 102, 0,3389,1287, 817, 91,3299,2940, - 836,1814, 549,2197,1396,1669,2987,3582,2297,2848,4528,1070, 687, 20,1819, 121, -1552,1364,1461,1968,2617,3540,2824,2083, 177, 948,4938,2291, 110,4549,2066, 648, -3359,1755,2110,2114,4642,4845,1693,3937,3308,1257,1869,2123, 208,1804,3159,2992, -2531,2549,3361,2418,1350,2347,2800,2568,1291,2036,2680, 72, 842,1990, 212,1233, -1154,1586, 75,2027,3410,4900,1823,1337,2710,2676, 728,2810,1522,3026,4995, 157, - 755,1050,4022, 710, 785,1936,2194,2085,1406,2777,2400, 150,1250,4049,1206, 807, -1910, 534, 529,3309,1721,1660, 274, 39,2827, 661,2670,1578, 925,3248,3815,1094, -4278,4901,4252, 41,1150,3747,2572,2227,4501,3658,4902,3813,3357,3617,2884,2258, - 887, 538,4187,3199,1294,2439,3042,2329,2343,2497,1255, 107, 543,1527, 521,3478, -3568, 194,5062, 15, 961,3870,1241,1192,2664, 66,5215,3260,2111,1295,1127,2152, -3805,4135, 901,1164,1976, 398,1278, 530,1460, 748, 904,1054,1966,1426, 53,2909, - 509, 523,2279,1534, 536,1019, 239,1685, 460,2353, 673,1065,2401,3600,4298,2272, -1272,2363, 284,1753,3679,4064,1695, 81, 815,2677,2757,2731,1386, 859, 500,4221, -2190,2566, 757,1006,2519,2068,1166,1455, 337,2654,3203,1863,1682,1914,3025,1252, -1409,1366, 847, 714,2834,2038,3209, 964,2970,1901, 885,2553,1078,1756,3049, 301, -1572,3326, 688,2130,1996,2429,1805,1648,2930,3421,2750,3652,3088, 262,1158,1254, - 389,1641,1812, 526,1719, 923,2073,1073,1902, 468, 489,4625,1140, 857,2375,3070, -3319,2863, 380, 116,1328,2693,1161,2244, 273,1212,1884,2769,3011,1775,1142, 461, -3066,1200,2147,2212, 790, 702,2695,4222,1601,1058, 434,2338,5153,3640, 67,2360, -4099,2502, 618,3472,1329, 416,1132, 830,2782,1807,2653,3211,3510,1662, 192,2124, - 296,3979,1739,1611,3684, 23, 118, 324, 446,1239,1225, 293,2520,3814,3795,2535, -3116, 17,1074, 467,2692,2201, 387,2922, 45,1326,3055,1645,3659,2817, 958, 243, -1903,2320,1339,2825,1784,3289, 356, 576, 865,2315,2381,3377,3916,1088,3122,1713, -1655, 935, 628,4689,1034,1327, 441, 800, 720, 894,1979,2183,1528,5289,2702,1071, -4046,3572,2399,1571,3281, 79, 761,1103, 327, 134, 758,1899,1371,1615, 879, 442, - 215,2605,2579, 173,2048,2485,1057,2975,3317,1097,2253,3801,4263,1403,1650,2946, - 814,4968,3487,1548,2644,1567,1285, 2, 295,2636, 97, 946,3576, 832, 141,4257, -3273, 760,3821,3521,3156,2607, 949,1024,1733,1516,1803,1920,2125,2283,2665,3180, -1501,2064,3560,2171,1592, 803,3518,1416, 732,3897,4258,1363,1362,2458, 119,1427, - 602,1525,2608,1605,1639,3175, 694,3064, 10, 465, 76,2000,4846,4208, 444,3781, -1619,3353,2206,1273,3796, 740,2483, 320,1723,2377,3660,2619,1359,1137,1762,1724, -2345,2842,1850,1862, 912, 821,1866, 612,2625,1735,2573,3369,1093, 844, 89, 937, - 930,1424,3564,2413,2972,1004,3046,3019,2011, 711,3171,1452,4178, 428, 801,1943, - 432, 445,2811, 206,4136,1472, 730, 349, 73, 397,2802,2547, 998,1637,1167, 789, - 396,3217, 154,1218, 716,1120,1780,2819,4826,1931,3334,3762,2139,1215,2627, 552, -3664,3628,3232,1405,2383,3111,1356,2652,3577,3320,3101,1703, 640,1045,1370,1246, -4996, 371,1575,2436,1621,2210, 984,4033,1734,2638, 16,4529, 663,2755,3255,1451, -3917,2257,1253,1955,2234,1263,2951, 214,1229, 617, 485, 359,1831,1969, 473,2310, - 750,2058, 165, 80,2864,2419, 361,4344,2416,2479,1134, 796,3726,1266,2943, 860, -2715, 938, 390,2734,1313,1384, 248, 202, 877,1064,2854, 522,3907, 279,1602, 297, -2357, 395,3740, 137,2075, 944,4089,2584,1267,3802, 62,1533,2285, 178, 176, 780, -2440, 201,3707, 590, 478,1560,4354,2117,1075, 30, 74,4643,4004,1635,1441,2745, - 776,2596, 238,1077,1692,1912,2844, 605, 499,1742,3947, 241,3053, 980,1749, 936, -2640,4511,2582, 515,1543,2162,5322,2892,2993, 890,2148,1924, 665,1827,3581,1032, - 968,3163, 339,1044,1896, 270, 583,1791,1720,4367,1194,3488,3669, 43,2523,1657, - 163,2167, 290,1209,1622,3378, 550, 634,2508,2510, 695,2634,2384,2512,1476,1414, - 220,1469,2341,2138,2852,3183,2900,4939,2865,3502,1211,3680, 854,3227,1299,2976, -3172, 186,2998,1459, 443,1067,3251,1495, 321,1932,3054, 909, 753,1410,1828, 436, -2441,1119,1587,3164,2186,1258, 227, 231,1425,1890,3200,3942, 247, 959, 725,5254, -2741, 577,2158,2079, 929, 120, 174, 838,2813, 591,1115, 417,2024, 40,3240,1536, -1037, 291,4151,2354, 632,1298,2406,2500,3535,1825,1846,3451, 205,1171, 345,4238, - 18,1163, 811, 685,2208,1217, 425,1312,1508,1175,4308,2552,1033, 587,1381,3059, -2984,3482, 340,1316,4023,3972, 792,3176, 519, 777,4690, 918, 933,4130,2981,3741, - 90,3360,2911,2200,5184,4550, 609,3079,2030, 272,3379,2736, 363,3881,1130,1447, - 286, 779, 357,1169,3350,3137,1630,1220,2687,2391, 747,1277,3688,2618,2682,2601, -1156,3196,5290,4034,3102,1689,3596,3128, 874, 219,2783, 798, 508,1843,2461, 269, -1658,1776,1392,1913,2983,3287,2866,2159,2372, 829,4076, 46,4253,2873,1889,1894, - 915,1834,1631,2181,2318, 298, 664,2818,3555,2735, 954,3228,3117, 527,3511,2173, - 681,2712,3033,2247,2346,3467,1652, 155,2164,3382, 113,1994, 450, 899, 494, 994, -1237,2958,1875,2336,1926,3727, 545,1577,1550, 633,3473, 204,1305,3072,2410,1956, -2471, 707,2134, 841,2195,2196,2663,3843,1026,4940, 990,3252,4997, 368,1092, 437, -3212,3258,1933,1829, 675,2977,2893, 412, 943,3723,4644,3294,3283,2230,2373,5154, -2389,2241,2661,2323,1404,2524, 593, 787, 677,3008,1275,2059, 438,2709,2609,2240, -2269,2246,1446, 36,1568,1373,3892,1574,2301,1456,3962, 693,2276,5216,2035,1143, -2720,1919,1797,1811,2763,4137,2597,1830,1699,1488,1198,2090, 424,1694, 312,3634, -3390,4179,3335,2252,1214, 561,1059,3243,2295,2561, 975,5155,2321,2751,3772, 472, -1537,3282,3398,1047,2077,2348,2878,1323,3340,3076, 690,2906, 51, 369, 170,3541, -1060,2187,2688,3670,2541,1083,1683, 928,3918, 459, 109,4427, 599,3744,4286, 143, -2101,2730,2490, 82,1588,3036,2121, 281,1860, 477,4035,1238,2812,3020,2716,3312, -1530,2188,2055,1317, 843, 636,1808,1173,3495, 649, 181,1002, 147,3641,1159,2414, -3750,2289,2795, 813,3123,2610,1136,4368, 5,3391,4541,2174, 420, 429,1728, 754, -1228,2115,2219, 347,2223,2733, 735,1518,3003,2355,3134,1764,3948,3329,1888,2424, -1001,1234,1972,3321,3363,1672,1021,1450,1584, 226, 765, 655,2526,3404,3244,2302, -3665, 731, 594,2184, 319,1576, 621, 658,2656,4299,2099,3864,1279,2071,2598,2739, - 795,3086,3699,3908,1707,2352,2402,1382,3136,2475,1465,4847,3496,3865,1085,3004, -2591,1084, 213,2287,1963,3565,2250, 822, 793,4574,3187,1772,1789,3050, 595,1484, -1959,2770,1080,2650, 456, 422,2996, 940,3322,4328,4345,3092,2742, 965,2784, 739, -4124, 952,1358,2498,2949,2565, 332,2698,2378, 660,2260,2473,4194,3856,2919, 535, -1260,2651,1208,1428,1300,1949,1303,2942, 433,2455,2450,1251,1946, 614,1269, 641, -1306,1810,2737,3078,2912, 564,2365,1419,1415,1497,4460,2367,2185,1379,3005,1307, -3218,2175,1897,3063, 682,1157,4040,4005,1712,1160,1941,1399, 394, 402,2952,1573, -1151,2986,2404, 862, 299,2033,1489,3006, 346, 171,2886,3401,1726,2932, 168,2533, - 47,2507,1030,3735,1145,3370,1395,1318,1579,3609,4560,2857,4116,1457,2529,1965, - 504,1036,2690,2988,2405, 745,5871, 849,2397,2056,3081, 863,2359,3857,2096, 99, -1397,1769,2300,4428,1643,3455,1978,1757,3718,1440, 35,4879,3742,1296,4228,2280, - 160,5063,1599,2013, 166, 520,3479,1646,3345,3012, 490,1937,1545,1264,2182,2505, -1096,1188,1369,1436,2421,1667,2792,2460,1270,2122, 727,3167,2143, 806,1706,1012, -1800,3037, 960,2218,1882, 805, 139,2456,1139,1521, 851,1052,3093,3089, 342,2039, - 744,5097,1468,1502,1585,2087, 223, 939, 326,2140,2577, 892,2481,1623,4077, 982, -3708, 135,2131, 87,2503,3114,2326,1106, 876,1616, 547,2997,2831,2093,3441,4530, -4314, 9,3256,4229,4148, 659,1462,1986,1710,2046,2913,2231,4090,4880,5255,3392, -3274,1368,3689,4645,1477, 705,3384,3635,1068,1529,2941,1458,3782,1509, 100,1656, -2548, 718,2339, 408,1590,2780,3548,1838,4117,3719,1345,3530, 717,3442,2778,3220, -2898,1892,4590,3614,3371,2043,1998,1224,3483, 891, 635, 584,2559,3355, 733,1766, -1729,1172,3789,1891,2307, 781,2982,2271,1957,1580,5773,2633,2005,4195,3097,1535, -3213,1189,1934,5693,3262, 586,3118,1324,1598, 517,1564,2217,1868,1893,4445,3728, -2703,3139,1526,1787,1992,3882,2875,1549,1199,1056,2224,1904,2711,5098,4287, 338, -1993,3129,3489,2689,1809,2815,1997, 957,1855,3898,2550,3275,3057,1105,1319, 627, -1505,1911,1883,3526, 698,3629,3456,1833,1431, 746, 77,1261,2017,2296,1977,1885, - 125,1334,1600, 525,1798,1109,2222,1470,1945, 559,2236,1186,3443,2476,1929,1411, -2411,3135,1777,3372,2621,1841,1613,3229, 668,1430,1839,2643,2916, 195,1989,2671, -2358,1387, 629,3205,2293,5256,4439, 123,1310, 888,1879,4300,3021,3605,1003,1162, -3192,2910,2010, 140,2395,2859, 55,1082,2012,2901, 662, 419,2081,1438, 680,2774, -4654,3912,1620,1731,1625,5035,4065,2328, 512,1344, 802,5443,2163,2311,2537, 524, -3399, 98,1155,2103,1918,2606,3925,2816,1393,2465,1504,3773,2177,3963,1478,4346, - 180,1113,4655,3461,2028,1698, 833,2696,1235,1322,1594,4408,3623,3013,3225,2040, -3022, 541,2881, 607,3632,2029,1665,1219, 639,1385,1686,1099,2803,3231,1938,3188, -2858, 427, 676,2772,1168,2025, 454,3253,2486,3556, 230,1950, 580, 791,1991,1280, -1086,1974,2034, 630, 257,3338,2788,4903,1017, 86,4790, 966,2789,1995,1696,1131, - 259,3095,4188,1308, 179,1463,5257, 289,4107,1248, 42,3413,1725,2288, 896,1947, - 774,4474,4254, 604,3430,4264, 392,2514,2588, 452, 237,1408,3018, 988,4531,1970, -3034,3310, 540,2370,1562,1288,2990, 502,4765,1147, 4,1853,2708, 207, 294,2814, -4078,2902,2509, 684, 34,3105,3532,2551, 644, 709,2801,2344, 573,1727,3573,3557, -2021,1081,3100,4315,2100,3681, 199,2263,1837,2385, 146,3484,1195,2776,3949, 997, -1939,3973,1008,1091,1202,1962,1847,1149,4209,5444,1076, 493, 117,5400,2521, 972, -1490,2934,1796,4542,2374,1512,2933,2657, 413,2888,1135,2762,2314,2156,1355,2369, - 766,2007,2527,2170,3124,2491,2593,2632,4757,2437, 234,3125,3591,1898,1750,1376, -1942,3468,3138, 570,2127,2145,3276,4131, 962, 132,1445,4196, 19, 941,3624,3480, -3366,1973,1374,4461,3431,2629, 283,2415,2275, 808,2887,3620,2112,2563,1353,3610, - 955,1089,3103,1053, 96, 88,4097, 823,3808,1583, 399, 292,4091,3313, 421,1128, - 642,4006, 903,2539,1877,2082, 596, 29,4066,1790, 722,2157, 130, 995,1569, 769, -1485, 464, 513,2213, 288,1923,1101,2453,4316, 133, 486,2445, 50, 625, 487,2207, - 57, 423, 481,2962, 159,3729,1558, 491, 303, 482, 501, 240,2837, 112,3648,2392, -1783, 362, 8,3433,3422, 610,2793,3277,1390,1284,1654, 21,3823, 734, 367, 623, - 193, 287, 374,1009,1483, 816, 476, 313,2255,2340,1262,2150,2899,1146,2581, 782, -2116,1659,2018,1880, 255,3586,3314,1110,2867,2137,2564, 986,2767,5185,2006, 650, - 158, 926, 762, 881,3157,2717,2362,3587, 306,3690,3245,1542,3077,2427,1691,2478, -2118,2985,3490,2438, 539,2305, 983, 129,1754, 355,4201,2386, 827,2923, 104,1773, -2838,2771, 411,2905,3919, 376, 767, 122,1114, 828,2422,1817,3506, 266,3460,1007, -1609,4998, 945,2612,4429,2274, 726,1247,1964,2914,2199,2070,4002,4108, 657,3323, -1422, 579, 455,2764,4737,1222,2895,1670, 824,1223,1487,2525, 558, 861,3080, 598, -2659,2515,1967, 752,2583,2376,2214,4180, 977, 704,2464,4999,2622,4109,1210,2961, - 819,1541, 142,2284, 44, 418, 457,1126,3730,4347,4626,1644,1876,3671,1864, 302, -1063,5694, 624, 723,1984,3745,1314,1676,2488,1610,1449,3558,3569,2166,2098, 409, -1011,2325,3704,2306, 818,1732,1383,1824,1844,3757, 999,2705,3497,1216,1423,2683, -2426,2954,2501,2726,2229,1475,2554,5064,1971,1794,1666,2014,1343, 783, 724, 191, -2434,1354,2220,5065,1763,2752,2472,4152, 131, 175,2885,3434, 92,1466,4920,2616, -3871,3872,3866, 128,1551,1632, 669,1854,3682,4691,4125,1230, 188,2973,3290,1302, -1213, 560,3266, 917, 763,3909,3249,1760, 868,1958, 764,1782,2097, 145,2277,3774, -4462, 64,1491,3062, 971,2132,3606,2442, 221,1226,1617, 218, 323,1185,3207,3147, - 571, 619,1473,1005,1744,2281, 449,1887,2396,3685, 275, 375,3816,1743,3844,3731, - 845,1983,2350,4210,1377, 773, 967,3499,3052,3743,2725,4007,1697,1022,3943,1464, -3264,2855,2722,1952,1029,2839,2467, 84,4383,2215, 820,1391,2015,2448,3672, 377, -1948,2168, 797,2545,3536,2578,2645, 94,2874,1678, 405,1259,3071, 771, 546,1315, - 470,1243,3083, 895,2468, 981, 969,2037, 846,4181, 653,1276,2928, 14,2594, 557, -3007,2474, 156, 902,1338,1740,2574, 537,2518, 973,2282,2216,2433,1928, 138,2903, -1293,2631,1612, 646,3457, 839,2935, 111, 496,2191,2847, 589,3186, 149,3994,2060, -4031,2641,4067,3145,1870, 37,3597,2136,1025,2051,3009,3383,3549,1121,1016,3261, -1301, 251,2446,2599,2153, 872,3246, 637, 334,3705, 831, 884, 921,3065,3140,4092, -2198,1944, 246,2964, 108,2045,1152,1921,2308,1031, 203,3173,4170,1907,3890, 810, -1401,2003,1690, 506, 647,1242,2828,1761,1649,3208,2249,1589,3709,2931,5156,1708, - 498, 666,2613, 834,3817,1231, 184,2851,1124, 883,3197,2261,3710,1765,1553,2658, -1178,2639,2351, 93,1193, 942,2538,2141,4402, 235,1821, 870,1591,2192,1709,1871, -3341,1618,4126,2595,2334, 603, 651, 69, 701, 268,2662,3411,2555,1380,1606, 503, - 448, 254,2371,2646, 574,1187,2309,1770, 322,2235,1292,1801, 305, 566,1133, 229, -2067,2057, 706, 167, 483,2002,2672,3295,1820,3561,3067, 316, 378,2746,3452,1112, - 136,1981, 507,1651,2917,1117, 285,4591, 182,2580,3522,1304, 335,3303,1835,2504, -1795,1792,2248, 674,1018,2106,2449,1857,2292,2845, 976,3047,1781,2600,2727,1389, -1281, 52,3152, 153, 265,3950, 672,3485,3951,4463, 430,1183, 365, 278,2169, 27, -1407,1336,2304, 209,1340,1730,2202,1852,2403,2883, 979,1737,1062, 631,2829,2542, -3876,2592, 825,2086,2226,3048,3625, 352,1417,3724, 542, 991, 431,1351,3938,1861, -2294, 826,1361,2927,3142,3503,1738, 463,2462,2723, 582,1916,1595,2808, 400,3845, -3891,2868,3621,2254, 58,2492,1123, 910,2160,2614,1372,1603,1196,1072,3385,1700, -3267,1980, 696, 480,2430, 920, 799,1570,2920,1951,2041,4047,2540,1321,4223,2469, -3562,2228,1271,2602, 401,2833,3351,2575,5157, 907,2312,1256, 410, 263,3507,1582, - 996, 678,1849,2316,1480, 908,3545,2237, 703,2322, 667,1826,2849,1531,2604,2999, -2407,3146,2151,2630,1786,3711, 469,3542, 497,3899,2409, 858, 837,4446,3393,1274, - 786, 620,1845,2001,3311, 484, 308,3367,1204,1815,3691,2332,1532,2557,1842,2020, -2724,1927,2333,4440, 567, 22,1673,2728,4475,1987,1858,1144,1597, 101,1832,3601, - 12, 974,3783,4391, 951,1412, 1,3720, 453,4608,4041, 528,1041,1027,3230,2628, -1129, 875,1051,3291,1203,2262,1069,2860,2799,2149,2615,3278, 144,1758,3040, 31, - 475,1680, 366,2685,3184, 311,1642,4008,2466,5036,1593,1493,2809, 216,1420,1668, - 233, 304,2128,3284, 232,1429,1768,1040,2008,3407,2740,2967,2543, 242,2133, 778, -1565,2022,2620, 505,2189,2756,1098,2273, 372,1614, 708, 553,2846,2094,2278, 169, -3626,2835,4161, 228,2674,3165, 809,1454,1309, 466,1705,1095, 900,3423, 880,2667, -3751,5258,2317,3109,2571,4317,2766,1503,1342, 866,4447,1118, 63,2076, 314,1881, -1348,1061, 172, 978,3515,1747, 532, 511,3970, 6, 601, 905,2699,3300,1751, 276, -1467,3725,2668, 65,4239,2544,2779,2556,1604, 578,2451,1802, 992,2331,2624,1320, -3446, 713,1513,1013, 103,2786,2447,1661, 886,1702, 916, 654,3574,2031,1556, 751, -2178,2821,2179,1498,1538,2176, 271, 914,2251,2080,1325, 638,1953,2937,3877,2432, -2754, 95,3265,1716, 260,1227,4083, 775, 106,1357,3254, 426,1607, 555,2480, 772, -1985, 244,2546, 474, 495,1046,2611,1851,2061, 71,2089,1675,2590, 742,3758,2843, -3222,1433, 267,2180,2576,2826,2233,2092,3913,2435, 956,1745,3075, 856,2113,1116, - 451, 3,1988,2896,1398, 993,2463,1878,2049,1341,2718,2721,2870,2108, 712,2904, -4363,2753,2324, 277,2872,2349,2649, 384, 987, 435, 691,3000, 922, 164,3939, 652, -1500,1184,4153,2482,3373,2165,4848,2335,3775,3508,3154,2806,2830,1554,2102,1664, -2530,1434,2408, 893,1547,2623,3447,2832,2242,2532,3169,2856,3223,2078, 49,3770, -3469, 462, 318, 656,2259,3250,3069, 679,1629,2758, 344,1138,1104,3120,1836,1283, -3115,2154,1437,4448, 934, 759,1999, 794,2862,1038, 533,2560,1722,2342, 855,2626, -1197,1663,4476,3127, 85,4240,2528, 25,1111,1181,3673, 407,3470,4561,2679,2713, - 768,1925,2841,3986,1544,1165, 932, 373,1240,2146,1930,2673, 721,4766, 354,4333, - 391,2963, 187, 61,3364,1442,1102, 330,1940,1767, 341,3809,4118, 393,2496,2062, -2211, 105, 331, 300, 439, 913,1332, 626, 379,3304,1557, 328, 689,3952, 309,1555, - 931, 317,2517,3027, 325, 569, 686,2107,3084, 60,1042,1333,2794, 264,3177,4014, -1628, 258,3712, 7,4464,1176,1043,1778, 683, 114,1975, 78,1492, 383,1886, 510, - 386, 645,5291,2891,2069,3305,4138,3867,2939,2603,2493,1935,1066,1848,3588,1015, -1282,1289,4609, 697,1453,3044,2666,3611,1856,2412, 54, 719,1330, 568,3778,2459, -1748, 788, 492, 551,1191,1000, 488,3394,3763, 282,1799, 348,2016,1523,3155,2390, -1049, 382,2019,1788,1170, 729,2968,3523, 897,3926,2785,2938,3292, 350,2319,3238, -1718,1717,2655,3453,3143,4465, 161,2889,2980,2009,1421, 56,1908,1640,2387,2232, -1917,1874,2477,4921, 148, 83,3438, 592,4245,2882,1822,1055, 741, 115,1496,1624, - 381,1638,4592,1020, 516,3214, 458, 947,4575,1432, 211,1514,2926,1865,2142, 189, - 852,1221,1400,1486, 882,2299,4036, 351, 28,1122, 700,6479,6480,6481,6482,6483, #last 512 -) - diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/gb2312prober.py b/src/fetchcode/vcs/pip/_vendor/chardet/gb2312prober.py deleted file mode 100644 index 8446d2dd..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/gb2312prober.py +++ /dev/null @@ -1,46 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is mozilla.org code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -from .mbcharsetprober import MultiByteCharSetProber -from .codingstatemachine import CodingStateMachine -from .chardistribution import GB2312DistributionAnalysis -from .mbcssm import GB2312_SM_MODEL - -class GB2312Prober(MultiByteCharSetProber): - def __init__(self): - super(GB2312Prober, self).__init__() - self.coding_sm = CodingStateMachine(GB2312_SM_MODEL) - self.distribution_analyzer = GB2312DistributionAnalysis() - self.reset() - - @property - def charset_name(self): - return "GB2312" - - @property - def language(self): - return "Chinese" diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/hebrewprober.py b/src/fetchcode/vcs/pip/_vendor/chardet/hebrewprober.py deleted file mode 100644 index b0e1bf49..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/hebrewprober.py +++ /dev/null @@ -1,292 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Universal charset detector code. -# -# The Initial Developer of the Original Code is -# Shy Shalom -# Portions created by the Initial Developer are Copyright (C) 2005 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -from .charsetprober import CharSetProber -from .enums import ProbingState - -# This prober doesn't actually recognize a language or a charset. -# It is a helper prober for the use of the Hebrew model probers - -### General ideas of the Hebrew charset recognition ### -# -# Four main charsets exist in Hebrew: -# "ISO-8859-8" - Visual Hebrew -# "windows-1255" - Logical Hebrew -# "ISO-8859-8-I" - Logical Hebrew -# "x-mac-hebrew" - ?? Logical Hebrew ?? -# -# Both "ISO" charsets use a completely identical set of code points, whereas -# "windows-1255" and "x-mac-hebrew" are two different proper supersets of -# these code points. windows-1255 defines additional characters in the range -# 0x80-0x9F as some misc punctuation marks as well as some Hebrew-specific -# diacritics and additional 'Yiddish' ligature letters in the range 0xc0-0xd6. -# x-mac-hebrew defines similar additional code points but with a different -# mapping. -# -# As far as an average Hebrew text with no diacritics is concerned, all four -# charsets are identical with respect to code points. Meaning that for the -# main Hebrew alphabet, all four map the same values to all 27 Hebrew letters -# (including final letters). -# -# The dominant difference between these charsets is their directionality. -# "Visual" directionality means that the text is ordered as if the renderer is -# not aware of a BIDI rendering algorithm. The renderer sees the text and -# draws it from left to right. The text itself when ordered naturally is read -# backwards. A buffer of Visual Hebrew generally looks like so: -# "[last word of first line spelled backwards] [whole line ordered backwards -# and spelled backwards] [first word of first line spelled backwards] -# [end of line] [last word of second line] ... etc' " -# adding punctuation marks, numbers and English text to visual text is -# naturally also "visual" and from left to right. -# -# "Logical" directionality means the text is ordered "naturally" according to -# the order it is read. It is the responsibility of the renderer to display -# the text from right to left. A BIDI algorithm is used to place general -# punctuation marks, numbers and English text in the text. -# -# Texts in x-mac-hebrew are almost impossible to find on the Internet. From -# what little evidence I could find, it seems that its general directionality -# is Logical. -# -# To sum up all of the above, the Hebrew probing mechanism knows about two -# charsets: -# Visual Hebrew - "ISO-8859-8" - backwards text - Words and sentences are -# backwards while line order is natural. For charset recognition purposes -# the line order is unimportant (In fact, for this implementation, even -# word order is unimportant). -# Logical Hebrew - "windows-1255" - normal, naturally ordered text. -# -# "ISO-8859-8-I" is a subset of windows-1255 and doesn't need to be -# specifically identified. -# "x-mac-hebrew" is also identified as windows-1255. A text in x-mac-hebrew -# that contain special punctuation marks or diacritics is displayed with -# some unconverted characters showing as question marks. This problem might -# be corrected using another model prober for x-mac-hebrew. Due to the fact -# that x-mac-hebrew texts are so rare, writing another model prober isn't -# worth the effort and performance hit. -# -#### The Prober #### -# -# The prober is divided between two SBCharSetProbers and a HebrewProber, -# all of which are managed, created, fed data, inquired and deleted by the -# SBCSGroupProber. The two SBCharSetProbers identify that the text is in -# fact some kind of Hebrew, Logical or Visual. The final decision about which -# one is it is made by the HebrewProber by combining final-letter scores -# with the scores of the two SBCharSetProbers to produce a final answer. -# -# The SBCSGroupProber is responsible for stripping the original text of HTML -# tags, English characters, numbers, low-ASCII punctuation characters, spaces -# and new lines. It reduces any sequence of such characters to a single space. -# The buffer fed to each prober in the SBCS group prober is pure text in -# high-ASCII. -# The two SBCharSetProbers (model probers) share the same language model: -# Win1255Model. -# The first SBCharSetProber uses the model normally as any other -# SBCharSetProber does, to recognize windows-1255, upon which this model was -# built. The second SBCharSetProber is told to make the pair-of-letter -# lookup in the language model backwards. This in practice exactly simulates -# a visual Hebrew model using the windows-1255 logical Hebrew model. -# -# The HebrewProber is not using any language model. All it does is look for -# final-letter evidence suggesting the text is either logical Hebrew or visual -# Hebrew. Disjointed from the model probers, the results of the HebrewProber -# alone are meaningless. HebrewProber always returns 0.00 as confidence -# since it never identifies a charset by itself. Instead, the pointer to the -# HebrewProber is passed to the model probers as a helper "Name Prober". -# When the Group prober receives a positive identification from any prober, -# it asks for the name of the charset identified. If the prober queried is a -# Hebrew model prober, the model prober forwards the call to the -# HebrewProber to make the final decision. In the HebrewProber, the -# decision is made according to the final-letters scores maintained and Both -# model probers scores. The answer is returned in the form of the name of the -# charset identified, either "windows-1255" or "ISO-8859-8". - -class HebrewProber(CharSetProber): - # windows-1255 / ISO-8859-8 code points of interest - FINAL_KAF = 0xea - NORMAL_KAF = 0xeb - FINAL_MEM = 0xed - NORMAL_MEM = 0xee - FINAL_NUN = 0xef - NORMAL_NUN = 0xf0 - FINAL_PE = 0xf3 - NORMAL_PE = 0xf4 - FINAL_TSADI = 0xf5 - NORMAL_TSADI = 0xf6 - - # Minimum Visual vs Logical final letter score difference. - # If the difference is below this, don't rely solely on the final letter score - # distance. - MIN_FINAL_CHAR_DISTANCE = 5 - - # Minimum Visual vs Logical model score difference. - # If the difference is below this, don't rely at all on the model score - # distance. - MIN_MODEL_DISTANCE = 0.01 - - VISUAL_HEBREW_NAME = "ISO-8859-8" - LOGICAL_HEBREW_NAME = "windows-1255" - - def __init__(self): - super(HebrewProber, self).__init__() - self._final_char_logical_score = None - self._final_char_visual_score = None - self._prev = None - self._before_prev = None - self._logical_prober = None - self._visual_prober = None - self.reset() - - def reset(self): - self._final_char_logical_score = 0 - self._final_char_visual_score = 0 - # The two last characters seen in the previous buffer, - # mPrev and mBeforePrev are initialized to space in order to simulate - # a word delimiter at the beginning of the data - self._prev = ' ' - self._before_prev = ' ' - # These probers are owned by the group prober. - - def set_model_probers(self, logicalProber, visualProber): - self._logical_prober = logicalProber - self._visual_prober = visualProber - - def is_final(self, c): - return c in [self.FINAL_KAF, self.FINAL_MEM, self.FINAL_NUN, - self.FINAL_PE, self.FINAL_TSADI] - - def is_non_final(self, c): - # The normal Tsadi is not a good Non-Final letter due to words like - # 'lechotet' (to chat) containing an apostrophe after the tsadi. This - # apostrophe is converted to a space in FilterWithoutEnglishLetters - # causing the Non-Final tsadi to appear at an end of a word even - # though this is not the case in the original text. - # The letters Pe and Kaf rarely display a related behavior of not being - # a good Non-Final letter. Words like 'Pop', 'Winamp' and 'Mubarak' - # for example legally end with a Non-Final Pe or Kaf. However, the - # benefit of these letters as Non-Final letters outweighs the damage - # since these words are quite rare. - return c in [self.NORMAL_KAF, self.NORMAL_MEM, - self.NORMAL_NUN, self.NORMAL_PE] - - def feed(self, byte_str): - # Final letter analysis for logical-visual decision. - # Look for evidence that the received buffer is either logical Hebrew - # or visual Hebrew. - # The following cases are checked: - # 1) A word longer than 1 letter, ending with a final letter. This is - # an indication that the text is laid out "naturally" since the - # final letter really appears at the end. +1 for logical score. - # 2) A word longer than 1 letter, ending with a Non-Final letter. In - # normal Hebrew, words ending with Kaf, Mem, Nun, Pe or Tsadi, - # should not end with the Non-Final form of that letter. Exceptions - # to this rule are mentioned above in isNonFinal(). This is an - # indication that the text is laid out backwards. +1 for visual - # score - # 3) A word longer than 1 letter, starting with a final letter. Final - # letters should not appear at the beginning of a word. This is an - # indication that the text is laid out backwards. +1 for visual - # score. - # - # The visual score and logical score are accumulated throughout the - # text and are finally checked against each other in GetCharSetName(). - # No checking for final letters in the middle of words is done since - # that case is not an indication for either Logical or Visual text. - # - # We automatically filter out all 7-bit characters (replace them with - # spaces) so the word boundary detection works properly. [MAP] - - if self.state == ProbingState.NOT_ME: - # Both model probers say it's not them. No reason to continue. - return ProbingState.NOT_ME - - byte_str = self.filter_high_byte_only(byte_str) - - for cur in byte_str: - if cur == ' ': - # We stand on a space - a word just ended - if self._before_prev != ' ': - # next-to-last char was not a space so self._prev is not a - # 1 letter word - if self.is_final(self._prev): - # case (1) [-2:not space][-1:final letter][cur:space] - self._final_char_logical_score += 1 - elif self.is_non_final(self._prev): - # case (2) [-2:not space][-1:Non-Final letter][ - # cur:space] - self._final_char_visual_score += 1 - else: - # Not standing on a space - if ((self._before_prev == ' ') and - (self.is_final(self._prev)) and (cur != ' ')): - # case (3) [-2:space][-1:final letter][cur:not space] - self._final_char_visual_score += 1 - self._before_prev = self._prev - self._prev = cur - - # Forever detecting, till the end or until both model probers return - # ProbingState.NOT_ME (handled above) - return ProbingState.DETECTING - - @property - def charset_name(self): - # Make the decision: is it Logical or Visual? - # If the final letter score distance is dominant enough, rely on it. - finalsub = self._final_char_logical_score - self._final_char_visual_score - if finalsub >= self.MIN_FINAL_CHAR_DISTANCE: - return self.LOGICAL_HEBREW_NAME - if finalsub <= -self.MIN_FINAL_CHAR_DISTANCE: - return self.VISUAL_HEBREW_NAME - - # It's not dominant enough, try to rely on the model scores instead. - modelsub = (self._logical_prober.get_confidence() - - self._visual_prober.get_confidence()) - if modelsub > self.MIN_MODEL_DISTANCE: - return self.LOGICAL_HEBREW_NAME - if modelsub < -self.MIN_MODEL_DISTANCE: - return self.VISUAL_HEBREW_NAME - - # Still no good, back to final letter distance, maybe it'll save the - # day. - if finalsub < 0.0: - return self.VISUAL_HEBREW_NAME - - # (finalsub > 0 - Logical) or (don't know what to do) default to - # Logical. - return self.LOGICAL_HEBREW_NAME - - @property - def language(self): - return 'Hebrew' - - @property - def state(self): - # Remain active as long as any of the model probers are active. - if (self._logical_prober.state == ProbingState.NOT_ME) and \ - (self._visual_prober.state == ProbingState.NOT_ME): - return ProbingState.NOT_ME - return ProbingState.DETECTING diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/jisfreq.py b/src/fetchcode/vcs/pip/_vendor/chardet/jisfreq.py deleted file mode 100644 index 83fc082b..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/jisfreq.py +++ /dev/null @@ -1,325 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Communicator client code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -# Sampling from about 20M text materials include literature and computer technology -# -# Japanese frequency table, applied to both S-JIS and EUC-JP -# They are sorted in order. - -# 128 --> 0.77094 -# 256 --> 0.85710 -# 512 --> 0.92635 -# 1024 --> 0.97130 -# 2048 --> 0.99431 -# -# Ideal Distribution Ratio = 0.92635 / (1-0.92635) = 12.58 -# Random Distribution Ration = 512 / (2965+62+83+86-512) = 0.191 -# -# Typical Distribution Ratio, 25% of IDR - -JIS_TYPICAL_DISTRIBUTION_RATIO = 3.0 - -# Char to FreqOrder table , -JIS_TABLE_SIZE = 4368 - -JIS_CHAR_TO_FREQ_ORDER = ( - 40, 1, 6, 182, 152, 180, 295,2127, 285, 381,3295,4304,3068,4606,3165,3510, # 16 -3511,1822,2785,4607,1193,2226,5070,4608, 171,2996,1247, 18, 179,5071, 856,1661, # 32 -1262,5072, 619, 127,3431,3512,3230,1899,1700, 232, 228,1294,1298, 284, 283,2041, # 48 -2042,1061,1062, 48, 49, 44, 45, 433, 434,1040,1041, 996, 787,2997,1255,4305, # 64 -2108,4609,1684,1648,5073,5074,5075,5076,5077,5078,3687,5079,4610,5080,3927,3928, # 80 -5081,3296,3432, 290,2285,1471,2187,5082,2580,2825,1303,2140,1739,1445,2691,3375, # 96 -1691,3297,4306,4307,4611, 452,3376,1182,2713,3688,3069,4308,5083,5084,5085,5086, # 112 -5087,5088,5089,5090,5091,5092,5093,5094,5095,5096,5097,5098,5099,5100,5101,5102, # 128 -5103,5104,5105,5106,5107,5108,5109,5110,5111,5112,4097,5113,5114,5115,5116,5117, # 144 -5118,5119,5120,5121,5122,5123,5124,5125,5126,5127,5128,5129,5130,5131,5132,5133, # 160 -5134,5135,5136,5137,5138,5139,5140,5141,5142,5143,5144,5145,5146,5147,5148,5149, # 176 -5150,5151,5152,4612,5153,5154,5155,5156,5157,5158,5159,5160,5161,5162,5163,5164, # 192 -5165,5166,5167,5168,5169,5170,5171,5172,5173,5174,5175,1472, 598, 618, 820,1205, # 208 -1309,1412,1858,1307,1692,5176,5177,5178,5179,5180,5181,5182,1142,1452,1234,1172, # 224 -1875,2043,2149,1793,1382,2973, 925,2404,1067,1241, 960,1377,2935,1491, 919,1217, # 240 -1865,2030,1406,1499,2749,4098,5183,5184,5185,5186,5187,5188,2561,4099,3117,1804, # 256 -2049,3689,4309,3513,1663,5189,3166,3118,3298,1587,1561,3433,5190,3119,1625,2998, # 272 -3299,4613,1766,3690,2786,4614,5191,5192,5193,5194,2161, 26,3377, 2,3929, 20, # 288 -3691, 47,4100, 50, 17, 16, 35, 268, 27, 243, 42, 155, 24, 154, 29, 184, # 304 - 4, 91, 14, 92, 53, 396, 33, 289, 9, 37, 64, 620, 21, 39, 321, 5, # 320 - 12, 11, 52, 13, 3, 208, 138, 0, 7, 60, 526, 141, 151,1069, 181, 275, # 336 -1591, 83, 132,1475, 126, 331, 829, 15, 69, 160, 59, 22, 157, 55,1079, 312, # 352 - 109, 38, 23, 25, 10, 19, 79,5195, 61, 382,1124, 8, 30,5196,5197,5198, # 368 -5199,5200,5201,5202,5203,5204,5205,5206, 89, 62, 74, 34,2416, 112, 139, 196, # 384 - 271, 149, 84, 607, 131, 765, 46, 88, 153, 683, 76, 874, 101, 258, 57, 80, # 400 - 32, 364, 121,1508, 169,1547, 68, 235, 145,2999, 41, 360,3027, 70, 63, 31, # 416 - 43, 259, 262,1383, 99, 533, 194, 66, 93, 846, 217, 192, 56, 106, 58, 565, # 432 - 280, 272, 311, 256, 146, 82, 308, 71, 100, 128, 214, 655, 110, 261, 104,1140, # 448 - 54, 51, 36, 87, 67,3070, 185,2618,2936,2020, 28,1066,2390,2059,5207,5208, # 464 -5209,5210,5211,5212,5213,5214,5215,5216,4615,5217,5218,5219,5220,5221,5222,5223, # 480 -5224,5225,5226,5227,5228,5229,5230,5231,5232,5233,5234,5235,5236,3514,5237,5238, # 496 -5239,5240,5241,5242,5243,5244,2297,2031,4616,4310,3692,5245,3071,5246,3598,5247, # 512 -4617,3231,3515,5248,4101,4311,4618,3808,4312,4102,5249,4103,4104,3599,5250,5251, # 528 -5252,5253,5254,5255,5256,5257,5258,5259,5260,5261,5262,5263,5264,5265,5266,5267, # 544 -5268,5269,5270,5271,5272,5273,5274,5275,5276,5277,5278,5279,5280,5281,5282,5283, # 560 -5284,5285,5286,5287,5288,5289,5290,5291,5292,5293,5294,5295,5296,5297,5298,5299, # 576 -5300,5301,5302,5303,5304,5305,5306,5307,5308,5309,5310,5311,5312,5313,5314,5315, # 592 -5316,5317,5318,5319,5320,5321,5322,5323,5324,5325,5326,5327,5328,5329,5330,5331, # 608 -5332,5333,5334,5335,5336,5337,5338,5339,5340,5341,5342,5343,5344,5345,5346,5347, # 624 -5348,5349,5350,5351,5352,5353,5354,5355,5356,5357,5358,5359,5360,5361,5362,5363, # 640 -5364,5365,5366,5367,5368,5369,5370,5371,5372,5373,5374,5375,5376,5377,5378,5379, # 656 -5380,5381, 363, 642,2787,2878,2788,2789,2316,3232,2317,3434,2011, 165,1942,3930, # 672 -3931,3932,3933,5382,4619,5383,4620,5384,5385,5386,5387,5388,5389,5390,5391,5392, # 688 -5393,5394,5395,5396,5397,5398,5399,5400,5401,5402,5403,5404,5405,5406,5407,5408, # 704 -5409,5410,5411,5412,5413,5414,5415,5416,5417,5418,5419,5420,5421,5422,5423,5424, # 720 -5425,5426,5427,5428,5429,5430,5431,5432,5433,5434,5435,5436,5437,5438,5439,5440, # 736 -5441,5442,5443,5444,5445,5446,5447,5448,5449,5450,5451,5452,5453,5454,5455,5456, # 752 -5457,5458,5459,5460,5461,5462,5463,5464,5465,5466,5467,5468,5469,5470,5471,5472, # 768 -5473,5474,5475,5476,5477,5478,5479,5480,5481,5482,5483,5484,5485,5486,5487,5488, # 784 -5489,5490,5491,5492,5493,5494,5495,5496,5497,5498,5499,5500,5501,5502,5503,5504, # 800 -5505,5506,5507,5508,5509,5510,5511,5512,5513,5514,5515,5516,5517,5518,5519,5520, # 816 -5521,5522,5523,5524,5525,5526,5527,5528,5529,5530,5531,5532,5533,5534,5535,5536, # 832 -5537,5538,5539,5540,5541,5542,5543,5544,5545,5546,5547,5548,5549,5550,5551,5552, # 848 -5553,5554,5555,5556,5557,5558,5559,5560,5561,5562,5563,5564,5565,5566,5567,5568, # 864 -5569,5570,5571,5572,5573,5574,5575,5576,5577,5578,5579,5580,5581,5582,5583,5584, # 880 -5585,5586,5587,5588,5589,5590,5591,5592,5593,5594,5595,5596,5597,5598,5599,5600, # 896 -5601,5602,5603,5604,5605,5606,5607,5608,5609,5610,5611,5612,5613,5614,5615,5616, # 912 -5617,5618,5619,5620,5621,5622,5623,5624,5625,5626,5627,5628,5629,5630,5631,5632, # 928 -5633,5634,5635,5636,5637,5638,5639,5640,5641,5642,5643,5644,5645,5646,5647,5648, # 944 -5649,5650,5651,5652,5653,5654,5655,5656,5657,5658,5659,5660,5661,5662,5663,5664, # 960 -5665,5666,5667,5668,5669,5670,5671,5672,5673,5674,5675,5676,5677,5678,5679,5680, # 976 -5681,5682,5683,5684,5685,5686,5687,5688,5689,5690,5691,5692,5693,5694,5695,5696, # 992 -5697,5698,5699,5700,5701,5702,5703,5704,5705,5706,5707,5708,5709,5710,5711,5712, # 1008 -5713,5714,5715,5716,5717,5718,5719,5720,5721,5722,5723,5724,5725,5726,5727,5728, # 1024 -5729,5730,5731,5732,5733,5734,5735,5736,5737,5738,5739,5740,5741,5742,5743,5744, # 1040 -5745,5746,5747,5748,5749,5750,5751,5752,5753,5754,5755,5756,5757,5758,5759,5760, # 1056 -5761,5762,5763,5764,5765,5766,5767,5768,5769,5770,5771,5772,5773,5774,5775,5776, # 1072 -5777,5778,5779,5780,5781,5782,5783,5784,5785,5786,5787,5788,5789,5790,5791,5792, # 1088 -5793,5794,5795,5796,5797,5798,5799,5800,5801,5802,5803,5804,5805,5806,5807,5808, # 1104 -5809,5810,5811,5812,5813,5814,5815,5816,5817,5818,5819,5820,5821,5822,5823,5824, # 1120 -5825,5826,5827,5828,5829,5830,5831,5832,5833,5834,5835,5836,5837,5838,5839,5840, # 1136 -5841,5842,5843,5844,5845,5846,5847,5848,5849,5850,5851,5852,5853,5854,5855,5856, # 1152 -5857,5858,5859,5860,5861,5862,5863,5864,5865,5866,5867,5868,5869,5870,5871,5872, # 1168 -5873,5874,5875,5876,5877,5878,5879,5880,5881,5882,5883,5884,5885,5886,5887,5888, # 1184 -5889,5890,5891,5892,5893,5894,5895,5896,5897,5898,5899,5900,5901,5902,5903,5904, # 1200 -5905,5906,5907,5908,5909,5910,5911,5912,5913,5914,5915,5916,5917,5918,5919,5920, # 1216 -5921,5922,5923,5924,5925,5926,5927,5928,5929,5930,5931,5932,5933,5934,5935,5936, # 1232 -5937,5938,5939,5940,5941,5942,5943,5944,5945,5946,5947,5948,5949,5950,5951,5952, # 1248 -5953,5954,5955,5956,5957,5958,5959,5960,5961,5962,5963,5964,5965,5966,5967,5968, # 1264 -5969,5970,5971,5972,5973,5974,5975,5976,5977,5978,5979,5980,5981,5982,5983,5984, # 1280 -5985,5986,5987,5988,5989,5990,5991,5992,5993,5994,5995,5996,5997,5998,5999,6000, # 1296 -6001,6002,6003,6004,6005,6006,6007,6008,6009,6010,6011,6012,6013,6014,6015,6016, # 1312 -6017,6018,6019,6020,6021,6022,6023,6024,6025,6026,6027,6028,6029,6030,6031,6032, # 1328 -6033,6034,6035,6036,6037,6038,6039,6040,6041,6042,6043,6044,6045,6046,6047,6048, # 1344 -6049,6050,6051,6052,6053,6054,6055,6056,6057,6058,6059,6060,6061,6062,6063,6064, # 1360 -6065,6066,6067,6068,6069,6070,6071,6072,6073,6074,6075,6076,6077,6078,6079,6080, # 1376 -6081,6082,6083,6084,6085,6086,6087,6088,6089,6090,6091,6092,6093,6094,6095,6096, # 1392 -6097,6098,6099,6100,6101,6102,6103,6104,6105,6106,6107,6108,6109,6110,6111,6112, # 1408 -6113,6114,2044,2060,4621, 997,1235, 473,1186,4622, 920,3378,6115,6116, 379,1108, # 1424 -4313,2657,2735,3934,6117,3809, 636,3233, 573,1026,3693,3435,2974,3300,2298,4105, # 1440 - 854,2937,2463, 393,2581,2417, 539, 752,1280,2750,2480, 140,1161, 440, 708,1569, # 1456 - 665,2497,1746,1291,1523,3000, 164,1603, 847,1331, 537,1997, 486, 508,1693,2418, # 1472 -1970,2227, 878,1220, 299,1030, 969, 652,2751, 624,1137,3301,2619, 65,3302,2045, # 1488 -1761,1859,3120,1930,3694,3516, 663,1767, 852, 835,3695, 269, 767,2826,2339,1305, # 1504 - 896,1150, 770,1616,6118, 506,1502,2075,1012,2519, 775,2520,2975,2340,2938,4314, # 1520 -3028,2086,1224,1943,2286,6119,3072,4315,2240,1273,1987,3935,1557, 175, 597, 985, # 1536 -3517,2419,2521,1416,3029, 585, 938,1931,1007,1052,1932,1685,6120,3379,4316,4623, # 1552 - 804, 599,3121,1333,2128,2539,1159,1554,2032,3810, 687,2033,2904, 952, 675,1467, # 1568 -3436,6121,2241,1096,1786,2440,1543,1924, 980,1813,2228, 781,2692,1879, 728,1918, # 1584 -3696,4624, 548,1950,4625,1809,1088,1356,3303,2522,1944, 502, 972, 373, 513,2827, # 1600 - 586,2377,2391,1003,1976,1631,6122,2464,1084, 648,1776,4626,2141, 324, 962,2012, # 1616 -2177,2076,1384, 742,2178,1448,1173,1810, 222, 102, 301, 445, 125,2420, 662,2498, # 1632 - 277, 200,1476,1165,1068, 224,2562,1378,1446, 450,1880, 659, 791, 582,4627,2939, # 1648 -3936,1516,1274, 555,2099,3697,1020,1389,1526,3380,1762,1723,1787,2229, 412,2114, # 1664 -1900,2392,3518, 512,2597, 427,1925,2341,3122,1653,1686,2465,2499, 697, 330, 273, # 1680 - 380,2162, 951, 832, 780, 991,1301,3073, 965,2270,3519, 668,2523,2636,1286, 535, # 1696 -1407, 518, 671, 957,2658,2378, 267, 611,2197,3030,6123, 248,2299, 967,1799,2356, # 1712 - 850,1418,3437,1876,1256,1480,2828,1718,6124,6125,1755,1664,2405,6126,4628,2879, # 1728 -2829, 499,2179, 676,4629, 557,2329,2214,2090, 325,3234, 464, 811,3001, 992,2342, # 1744 -2481,1232,1469, 303,2242, 466,1070,2163, 603,1777,2091,4630,2752,4631,2714, 322, # 1760 -2659,1964,1768, 481,2188,1463,2330,2857,3600,2092,3031,2421,4632,2318,2070,1849, # 1776 -2598,4633,1302,2254,1668,1701,2422,3811,2905,3032,3123,2046,4106,1763,1694,4634, # 1792 -1604, 943,1724,1454, 917, 868,2215,1169,2940, 552,1145,1800,1228,1823,1955, 316, # 1808 -1080,2510, 361,1807,2830,4107,2660,3381,1346,1423,1134,4108,6127, 541,1263,1229, # 1824 -1148,2540, 545, 465,1833,2880,3438,1901,3074,2482, 816,3937, 713,1788,2500, 122, # 1840 -1575, 195,1451,2501,1111,6128, 859, 374,1225,2243,2483,4317, 390,1033,3439,3075, # 1856 -2524,1687, 266, 793,1440,2599, 946, 779, 802, 507, 897,1081, 528,2189,1292, 711, # 1872 -1866,1725,1167,1640, 753, 398,2661,1053, 246, 348,4318, 137,1024,3440,1600,2077, # 1888 -2129, 825,4319, 698, 238, 521, 187,2300,1157,2423,1641,1605,1464,1610,1097,2541, # 1904 -1260,1436, 759,2255,1814,2150, 705,3235, 409,2563,3304, 561,3033,2005,2564, 726, # 1920 -1956,2343,3698,4109, 949,3812,3813,3520,1669, 653,1379,2525, 881,2198, 632,2256, # 1936 -1027, 778,1074, 733,1957, 514,1481,2466, 554,2180, 702,3938,1606,1017,1398,6129, # 1952 -1380,3521, 921, 993,1313, 594, 449,1489,1617,1166, 768,1426,1360, 495,1794,3601, # 1968 -1177,3602,1170,4320,2344, 476, 425,3167,4635,3168,1424, 401,2662,1171,3382,1998, # 1984 -1089,4110, 477,3169, 474,6130,1909, 596,2831,1842, 494, 693,1051,1028,1207,3076, # 2000 - 606,2115, 727,2790,1473,1115, 743,3522, 630, 805,1532,4321,2021, 366,1057, 838, # 2016 - 684,1114,2142,4322,2050,1492,1892,1808,2271,3814,2424,1971,1447,1373,3305,1090, # 2032 -1536,3939,3523,3306,1455,2199, 336, 369,2331,1035, 584,2393, 902, 718,2600,6131, # 2048 -2753, 463,2151,1149,1611,2467, 715,1308,3124,1268, 343,1413,3236,1517,1347,2663, # 2064 -2093,3940,2022,1131,1553,2100,2941,1427,3441,2942,1323,2484,6132,1980, 872,2368, # 2080 -2441,2943, 320,2369,2116,1082, 679,1933,3941,2791,3815, 625,1143,2023, 422,2200, # 2096 -3816,6133, 730,1695, 356,2257,1626,2301,2858,2637,1627,1778, 937, 883,2906,2693, # 2112 -3002,1769,1086, 400,1063,1325,3307,2792,4111,3077, 456,2345,1046, 747,6134,1524, # 2128 - 884,1094,3383,1474,2164,1059, 974,1688,2181,2258,1047, 345,1665,1187, 358, 875, # 2144 -3170, 305, 660,3524,2190,1334,1135,3171,1540,1649,2542,1527, 927, 968,2793, 885, # 2160 -1972,1850, 482, 500,2638,1218,1109,1085,2543,1654,2034, 876, 78,2287,1482,1277, # 2176 - 861,1675,1083,1779, 724,2754, 454, 397,1132,1612,2332, 893, 672,1237, 257,2259, # 2192 -2370, 135,3384, 337,2244, 547, 352, 340, 709,2485,1400, 788,1138,2511, 540, 772, # 2208 -1682,2260,2272,2544,2013,1843,1902,4636,1999,1562,2288,4637,2201,1403,1533, 407, # 2224 - 576,3308,1254,2071, 978,3385, 170, 136,1201,3125,2664,3172,2394, 213, 912, 873, # 2240 -3603,1713,2202, 699,3604,3699, 813,3442, 493, 531,1054, 468,2907,1483, 304, 281, # 2256 -4112,1726,1252,2094, 339,2319,2130,2639, 756,1563,2944, 748, 571,2976,1588,2425, # 2272 -2715,1851,1460,2426,1528,1392,1973,3237, 288,3309, 685,3386, 296, 892,2716,2216, # 2288 -1570,2245, 722,1747,2217, 905,3238,1103,6135,1893,1441,1965, 251,1805,2371,3700, # 2304 -2601,1919,1078, 75,2182,1509,1592,1270,2640,4638,2152,6136,3310,3817, 524, 706, # 2320 -1075, 292,3818,1756,2602, 317, 98,3173,3605,3525,1844,2218,3819,2502, 814, 567, # 2336 - 385,2908,1534,6137, 534,1642,3239, 797,6138,1670,1529, 953,4323, 188,1071, 538, # 2352 - 178, 729,3240,2109,1226,1374,2000,2357,2977, 731,2468,1116,2014,2051,6139,1261, # 2368 -1593, 803,2859,2736,3443, 556, 682, 823,1541,6140,1369,2289,1706,2794, 845, 462, # 2384 -2603,2665,1361, 387, 162,2358,1740, 739,1770,1720,1304,1401,3241,1049, 627,1571, # 2400 -2427,3526,1877,3942,1852,1500, 431,1910,1503, 677, 297,2795, 286,1433,1038,1198, # 2416 -2290,1133,1596,4113,4639,2469,1510,1484,3943,6141,2442, 108, 712,4640,2372, 866, # 2432 -3701,2755,3242,1348, 834,1945,1408,3527,2395,3243,1811, 824, 994,1179,2110,1548, # 2448 -1453, 790,3003, 690,4324,4325,2832,2909,3820,1860,3821, 225,1748, 310, 346,1780, # 2464 -2470, 821,1993,2717,2796, 828, 877,3528,2860,2471,1702,2165,2910,2486,1789, 453, # 2480 - 359,2291,1676, 73,1164,1461,1127,3311, 421, 604, 314,1037, 589, 116,2487, 737, # 2496 - 837,1180, 111, 244, 735,6142,2261,1861,1362, 986, 523, 418, 581,2666,3822, 103, # 2512 - 855, 503,1414,1867,2488,1091, 657,1597, 979, 605,1316,4641,1021,2443,2078,2001, # 2528 -1209, 96, 587,2166,1032, 260,1072,2153, 173, 94, 226,3244, 819,2006,4642,4114, # 2544 -2203, 231,1744, 782, 97,2667, 786,3387, 887, 391, 442,2219,4326,1425,6143,2694, # 2560 - 633,1544,1202, 483,2015, 592,2052,1958,2472,1655, 419, 129,4327,3444,3312,1714, # 2576 -1257,3078,4328,1518,1098, 865,1310,1019,1885,1512,1734, 469,2444, 148, 773, 436, # 2592 -1815,1868,1128,1055,4329,1245,2756,3445,2154,1934,1039,4643, 579,1238, 932,2320, # 2608 - 353, 205, 801, 115,2428, 944,2321,1881, 399,2565,1211, 678, 766,3944, 335,2101, # 2624 -1459,1781,1402,3945,2737,2131,1010, 844, 981,1326,1013, 550,1816,1545,2620,1335, # 2640 -1008, 371,2881, 936,1419,1613,3529,1456,1395,2273,1834,2604,1317,2738,2503, 416, # 2656 -1643,4330, 806,1126, 229, 591,3946,1314,1981,1576,1837,1666, 347,1790, 977,3313, # 2672 - 764,2861,1853, 688,2429,1920,1462, 77, 595, 415,2002,3034, 798,1192,4115,6144, # 2688 -2978,4331,3035,2695,2582,2072,2566, 430,2430,1727, 842,1396,3947,3702, 613, 377, # 2704 - 278, 236,1417,3388,3314,3174, 757,1869, 107,3530,6145,1194, 623,2262, 207,1253, # 2720 -2167,3446,3948, 492,1117,1935, 536,1838,2757,1246,4332, 696,2095,2406,1393,1572, # 2736 -3175,1782, 583, 190, 253,1390,2230, 830,3126,3389, 934,3245,1703,1749,2979,1870, # 2752 -2545,1656,2204, 869,2346,4116,3176,1817, 496,1764,4644, 942,1504, 404,1903,1122, # 2768 -1580,3606,2945,1022, 515, 372,1735, 955,2431,3036,6146,2797,1110,2302,2798, 617, # 2784 -6147, 441, 762,1771,3447,3607,3608,1904, 840,3037, 86, 939,1385, 572,1370,2445, # 2800 -1336, 114,3703, 898, 294, 203,3315, 703,1583,2274, 429, 961,4333,1854,1951,3390, # 2816 -2373,3704,4334,1318,1381, 966,1911,2322,1006,1155, 309, 989, 458,2718,1795,1372, # 2832 -1203, 252,1689,1363,3177, 517,1936, 168,1490, 562, 193,3823,1042,4117,1835, 551, # 2848 - 470,4645, 395, 489,3448,1871,1465,2583,2641, 417,1493, 279,1295, 511,1236,1119, # 2864 - 72,1231,1982,1812,3004, 871,1564, 984,3449,1667,2696,2096,4646,2347,2833,1673, # 2880 -3609, 695,3246,2668, 807,1183,4647, 890, 388,2333,1801,1457,2911,1765,1477,1031, # 2896 -3316,3317,1278,3391,2799,2292,2526, 163,3450,4335,2669,1404,1802,6148,2323,2407, # 2912 -1584,1728,1494,1824,1269, 298, 909,3318,1034,1632, 375, 776,1683,2061, 291, 210, # 2928 -1123, 809,1249,1002,2642,3038, 206,1011,2132, 144, 975, 882,1565, 342, 667, 754, # 2944 -1442,2143,1299,2303,2062, 447, 626,2205,1221,2739,2912,1144,1214,2206,2584, 760, # 2960 -1715, 614, 950,1281,2670,2621, 810, 577,1287,2546,4648, 242,2168, 250,2643, 691, # 2976 - 123,2644, 647, 313,1029, 689,1357,2946,1650, 216, 771,1339,1306, 808,2063, 549, # 2992 - 913,1371,2913,2914,6149,1466,1092,1174,1196,1311,2605,2396,1783,1796,3079, 406, # 3008 -2671,2117,3949,4649, 487,1825,2220,6150,2915, 448,2348,1073,6151,2397,1707, 130, # 3024 - 900,1598, 329, 176,1959,2527,1620,6152,2275,4336,3319,1983,2191,3705,3610,2155, # 3040 -3706,1912,1513,1614,6153,1988, 646, 392,2304,1589,3320,3039,1826,1239,1352,1340, # 3056 -2916, 505,2567,1709,1437,2408,2547, 906,6154,2672, 384,1458,1594,1100,1329, 710, # 3072 - 423,3531,2064,2231,2622,1989,2673,1087,1882, 333, 841,3005,1296,2882,2379, 580, # 3088 -1937,1827,1293,2585, 601, 574, 249,1772,4118,2079,1120, 645, 901,1176,1690, 795, # 3104 -2207, 478,1434, 516,1190,1530, 761,2080, 930,1264, 355, 435,1552, 644,1791, 987, # 3120 - 220,1364,1163,1121,1538, 306,2169,1327,1222, 546,2645, 218, 241, 610,1704,3321, # 3136 -1984,1839,1966,2528, 451,6155,2586,3707,2568, 907,3178, 254,2947, 186,1845,4650, # 3152 - 745, 432,1757, 428,1633, 888,2246,2221,2489,3611,2118,1258,1265, 956,3127,1784, # 3168 -4337,2490, 319, 510, 119, 457,3612, 274,2035,2007,4651,1409,3128, 970,2758, 590, # 3184 -2800, 661,2247,4652,2008,3950,1420,1549,3080,3322,3951,1651,1375,2111, 485,2491, # 3200 -1429,1156,6156,2548,2183,1495, 831,1840,2529,2446, 501,1657, 307,1894,3247,1341, # 3216 - 666, 899,2156,1539,2549,1559, 886, 349,2208,3081,2305,1736,3824,2170,2759,1014, # 3232 -1913,1386, 542,1397,2948, 490, 368, 716, 362, 159, 282,2569,1129,1658,1288,1750, # 3248 -2674, 276, 649,2016, 751,1496, 658,1818,1284,1862,2209,2087,2512,3451, 622,2834, # 3264 - 376, 117,1060,2053,1208,1721,1101,1443, 247,1250,3179,1792,3952,2760,2398,3953, # 3280 -6157,2144,3708, 446,2432,1151,2570,3452,2447,2761,2835,1210,2448,3082, 424,2222, # 3296 -1251,2449,2119,2836, 504,1581,4338, 602, 817, 857,3825,2349,2306, 357,3826,1470, # 3312 -1883,2883, 255, 958, 929,2917,3248, 302,4653,1050,1271,1751,2307,1952,1430,2697, # 3328 -2719,2359, 354,3180, 777, 158,2036,4339,1659,4340,4654,2308,2949,2248,1146,2232, # 3344 -3532,2720,1696,2623,3827,6158,3129,1550,2698,1485,1297,1428, 637, 931,2721,2145, # 3360 - 914,2550,2587, 81,2450, 612, 827,2646,1242,4655,1118,2884, 472,1855,3181,3533, # 3376 -3534, 569,1353,2699,1244,1758,2588,4119,2009,2762,2171,3709,1312,1531,6159,1152, # 3392 -1938, 134,1830, 471,3710,2276,1112,1535,3323,3453,3535, 982,1337,2950, 488, 826, # 3408 - 674,1058,1628,4120,2017, 522,2399, 211, 568,1367,3454, 350, 293,1872,1139,3249, # 3424 -1399,1946,3006,1300,2360,3324, 588, 736,6160,2606, 744, 669,3536,3828,6161,1358, # 3440 - 199, 723, 848, 933, 851,1939,1505,1514,1338,1618,1831,4656,1634,3613, 443,2740, # 3456 -3829, 717,1947, 491,1914,6162,2551,1542,4121,1025,6163,1099,1223, 198,3040,2722, # 3472 - 370, 410,1905,2589, 998,1248,3182,2380, 519,1449,4122,1710, 947, 928,1153,4341, # 3488 -2277, 344,2624,1511, 615, 105, 161,1212,1076,1960,3130,2054,1926,1175,1906,2473, # 3504 - 414,1873,2801,6164,2309, 315,1319,3325, 318,2018,2146,2157, 963, 631, 223,4342, # 3520 -4343,2675, 479,3711,1197,2625,3712,2676,2361,6165,4344,4123,6166,2451,3183,1886, # 3536 -2184,1674,1330,1711,1635,1506, 799, 219,3250,3083,3954,1677,3713,3326,2081,3614, # 3552 -1652,2073,4657,1147,3041,1752, 643,1961, 147,1974,3955,6167,1716,2037, 918,3007, # 3568 -1994, 120,1537, 118, 609,3184,4345, 740,3455,1219, 332,1615,3830,6168,1621,2980, # 3584 -1582, 783, 212, 553,2350,3714,1349,2433,2082,4124, 889,6169,2310,1275,1410, 973, # 3600 - 166,1320,3456,1797,1215,3185,2885,1846,2590,2763,4658, 629, 822,3008, 763, 940, # 3616 -1990,2862, 439,2409,1566,1240,1622, 926,1282,1907,2764, 654,2210,1607, 327,1130, # 3632 -3956,1678,1623,6170,2434,2192, 686, 608,3831,3715, 903,3957,3042,6171,2741,1522, # 3648 -1915,1105,1555,2552,1359, 323,3251,4346,3457, 738,1354,2553,2311,2334,1828,2003, # 3664 -3832,1753,2351,1227,6172,1887,4125,1478,6173,2410,1874,1712,1847, 520,1204,2607, # 3680 - 264,4659, 836,2677,2102, 600,4660,3833,2278,3084,6174,4347,3615,1342, 640, 532, # 3696 - 543,2608,1888,2400,2591,1009,4348,1497, 341,1737,3616,2723,1394, 529,3252,1321, # 3712 - 983,4661,1515,2120, 971,2592, 924, 287,1662,3186,4349,2700,4350,1519, 908,1948, # 3728 -2452, 156, 796,1629,1486,2223,2055, 694,4126,1259,1036,3392,1213,2249,2742,1889, # 3744 -1230,3958,1015, 910, 408, 559,3617,4662, 746, 725, 935,4663,3959,3009,1289, 563, # 3760 - 867,4664,3960,1567,2981,2038,2626, 988,2263,2381,4351, 143,2374, 704,1895,6175, # 3776 -1188,3716,2088, 673,3085,2362,4352, 484,1608,1921,2765,2918, 215, 904,3618,3537, # 3792 - 894, 509, 976,3043,2701,3961,4353,2837,2982, 498,6176,6177,1102,3538,1332,3393, # 3808 -1487,1636,1637, 233, 245,3962, 383, 650, 995,3044, 460,1520,1206,2352, 749,3327, # 3824 - 530, 700, 389,1438,1560,1773,3963,2264, 719,2951,2724,3834, 870,1832,1644,1000, # 3840 - 839,2474,3717, 197,1630,3394, 365,2886,3964,1285,2133, 734, 922, 818,1106, 732, # 3856 - 480,2083,1774,3458, 923,2279,1350, 221,3086, 85,2233,2234,3835,1585,3010,2147, # 3872 -1387,1705,2382,1619,2475, 133, 239,2802,1991,1016,2084,2383, 411,2838,1113, 651, # 3888 -1985,1160,3328, 990,1863,3087,1048,1276,2647, 265,2627,1599,3253,2056, 150, 638, # 3904 -2019, 656, 853, 326,1479, 680,1439,4354,1001,1759, 413,3459,3395,2492,1431, 459, # 3920 -4355,1125,3329,2265,1953,1450,2065,2863, 849, 351,2678,3131,3254,3255,1104,1577, # 3936 - 227,1351,1645,2453,2193,1421,2887, 812,2121, 634, 95,2435, 201,2312,4665,1646, # 3952 -1671,2743,1601,2554,2702,2648,2280,1315,1366,2089,3132,1573,3718,3965,1729,1189, # 3968 - 328,2679,1077,1940,1136, 558,1283, 964,1195, 621,2074,1199,1743,3460,3619,1896, # 3984 -1916,1890,3836,2952,1154,2112,1064, 862, 378,3011,2066,2113,2803,1568,2839,6178, # 4000 -3088,2919,1941,1660,2004,1992,2194, 142, 707,1590,1708,1624,1922,1023,1836,1233, # 4016 -1004,2313, 789, 741,3620,6179,1609,2411,1200,4127,3719,3720,4666,2057,3721, 593, # 4032 -2840, 367,2920,1878,6180,3461,1521, 628,1168, 692,2211,2649, 300, 720,2067,2571, # 4048 -2953,3396, 959,2504,3966,3539,3462,1977, 701,6181, 954,1043, 800, 681, 183,3722, # 4064 -1803,1730,3540,4128,2103, 815,2314, 174, 467, 230,2454,1093,2134, 755,3541,3397, # 4080 -1141,1162,6182,1738,2039, 270,3256,2513,1005,1647,2185,3837, 858,1679,1897,1719, # 4096 -2954,2324,1806, 402, 670, 167,4129,1498,2158,2104, 750,6183, 915, 189,1680,1551, # 4112 - 455,4356,1501,2455, 405,1095,2955, 338,1586,1266,1819, 570, 641,1324, 237,1556, # 4128 -2650,1388,3723,6184,1368,2384,1343,1978,3089,2436, 879,3724, 792,1191, 758,3012, # 4144 -1411,2135,1322,4357, 240,4667,1848,3725,1574,6185, 420,3045,1546,1391, 714,4358, # 4160 -1967, 941,1864, 863, 664, 426, 560,1731,2680,1785,2864,1949,2363, 403,3330,1415, # 4176 -1279,2136,1697,2335, 204, 721,2097,3838, 90,6186,2085,2505, 191,3967, 124,2148, # 4192 -1376,1798,1178,1107,1898,1405, 860,4359,1243,1272,2375,2983,1558,2456,1638, 113, # 4208 -3621, 578,1923,2609, 880, 386,4130, 784,2186,2266,1422,2956,2172,1722, 497, 263, # 4224 -2514,1267,2412,2610, 177,2703,3542, 774,1927,1344, 616,1432,1595,1018, 172,4360, # 4240 -2325, 911,4361, 438,1468,3622, 794,3968,2024,2173,1681,1829,2957, 945, 895,3090, # 4256 - 575,2212,2476, 475,2401,2681, 785,2744,1745,2293,2555,1975,3133,2865, 394,4668, # 4272 -3839, 635,4131, 639, 202,1507,2195,2766,1345,1435,2572,3726,1908,1184,1181,2457, # 4288 -3727,3134,4362, 843,2611, 437, 916,4669, 234, 769,1884,3046,3047,3623, 833,6187, # 4304 -1639,2250,2402,1355,1185,2010,2047, 999, 525,1732,1290,1488,2612, 948,1578,3728, # 4320 -2413,2477,1216,2725,2159, 334,3840,1328,3624,2921,1525,4132, 564,1056, 891,4363, # 4336 -1444,1698,2385,2251,3729,1365,2281,2235,1717,6188, 864,3841,2515, 444, 527,2767, # 4352 -2922,3625, 544, 461,6189, 566, 209,2437,3398,2098,1065,2068,3331,3626,3257,2137, # 4368 #last 512 -) - - diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/jpcntx.py b/src/fetchcode/vcs/pip/_vendor/chardet/jpcntx.py deleted file mode 100644 index 20044e4b..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/jpcntx.py +++ /dev/null @@ -1,233 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Communicator client code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - - -# This is hiragana 2-char sequence table, the number in each cell represents its frequency category -jp2CharContext = ( -(0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1), -(2,4,0,4,0,3,0,4,0,3,4,4,4,2,4,3,3,4,3,2,3,3,4,2,3,3,3,2,4,1,4,3,3,1,5,4,3,4,3,4,3,5,3,0,3,5,4,2,0,3,1,0,3,3,0,3,3,0,1,1,0,4,3,0,3,3,0,4,0,2,0,3,5,5,5,5,4,0,4,1,0,3,4), -(0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2), -(0,4,0,5,0,5,0,4,0,4,5,4,4,3,5,3,5,1,5,3,4,3,4,4,3,4,3,3,4,3,5,4,4,3,5,5,3,5,5,5,3,5,5,3,4,5,5,3,1,3,2,0,3,4,0,4,2,0,4,2,1,5,3,2,3,5,0,4,0,2,0,5,4,4,5,4,5,0,4,0,0,4,4), -(0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0), -(0,3,0,4,0,3,0,3,0,4,5,4,3,3,3,3,4,3,5,4,4,3,5,4,4,3,4,3,4,4,4,4,5,3,4,4,3,4,5,5,4,5,5,1,4,5,4,3,0,3,3,1,3,3,0,4,4,0,3,3,1,5,3,3,3,5,0,4,0,3,0,4,4,3,4,3,3,0,4,1,1,3,4), -(0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0), -(0,4,0,3,0,3,0,4,0,3,4,4,3,2,2,1,2,1,3,1,3,3,3,3,3,4,3,1,3,3,5,3,3,0,4,3,0,5,4,3,3,5,4,4,3,4,4,5,0,1,2,0,1,2,0,2,2,0,1,0,0,5,2,2,1,4,0,3,0,1,0,4,4,3,5,4,3,0,2,1,0,4,3), -(0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0), -(0,3,0,5,0,4,0,2,1,4,4,2,4,1,4,2,4,2,4,3,3,3,4,3,3,3,3,1,4,2,3,3,3,1,4,4,1,1,1,4,3,3,2,0,2,4,3,2,0,3,3,0,3,1,1,0,0,0,3,3,0,4,2,2,3,4,0,4,0,3,0,4,4,5,3,4,4,0,3,0,0,1,4), -(1,4,0,4,0,4,0,4,0,3,5,4,4,3,4,3,5,4,3,3,4,3,5,4,4,4,4,3,4,2,4,3,3,1,5,4,3,2,4,5,4,5,5,4,4,5,4,4,0,3,2,2,3,3,0,4,3,1,3,2,1,4,3,3,4,5,0,3,0,2,0,4,5,5,4,5,4,0,4,0,0,5,4), -(0,5,0,5,0,4,0,3,0,4,4,3,4,3,3,3,4,0,4,4,4,3,4,3,4,3,3,1,4,2,4,3,4,0,5,4,1,4,5,4,4,5,3,2,4,3,4,3,2,4,1,3,3,3,2,3,2,0,4,3,3,4,3,3,3,4,0,4,0,3,0,4,5,4,4,4,3,0,4,1,0,1,3), -(0,3,1,4,0,3,0,2,0,3,4,4,3,1,4,2,3,3,4,3,4,3,4,3,4,4,3,2,3,1,5,4,4,1,4,4,3,5,4,4,3,5,5,4,3,4,4,3,1,2,3,1,2,2,0,3,2,0,3,1,0,5,3,3,3,4,3,3,3,3,4,4,4,4,5,4,2,0,3,3,2,4,3), -(0,2,0,3,0,1,0,1,0,0,3,2,0,0,2,0,1,0,2,1,3,3,3,1,2,3,1,0,1,0,4,2,1,1,3,3,0,4,3,3,1,4,3,3,0,3,3,2,0,0,0,0,1,0,0,2,0,0,0,0,0,4,1,0,2,3,2,2,2,1,3,3,3,4,4,3,2,0,3,1,0,3,3), -(0,4,0,4,0,3,0,3,0,4,4,4,3,3,3,3,3,3,4,3,4,2,4,3,4,3,3,2,4,3,4,5,4,1,4,5,3,5,4,5,3,5,4,0,3,5,5,3,1,3,3,2,2,3,0,3,4,1,3,3,2,4,3,3,3,4,0,4,0,3,0,4,5,4,4,5,3,0,4,1,0,3,4), -(0,2,0,3,0,3,0,0,0,2,2,2,1,0,1,0,0,0,3,0,3,0,3,0,1,3,1,0,3,1,3,3,3,1,3,3,3,0,1,3,1,3,4,0,0,3,1,1,0,3,2,0,0,0,0,1,3,0,1,0,0,3,3,2,0,3,0,0,0,0,0,3,4,3,4,3,3,0,3,0,0,2,3), -(2,3,0,3,0,2,0,1,0,3,3,4,3,1,3,1,1,1,3,1,4,3,4,3,3,3,0,0,3,1,5,4,3,1,4,3,2,5,5,4,4,4,4,3,3,4,4,4,0,2,1,1,3,2,0,1,2,0,0,1,0,4,1,3,3,3,0,3,0,1,0,4,4,4,5,5,3,0,2,0,0,4,4), -(0,2,0,1,0,3,1,3,0,2,3,3,3,0,3,1,0,0,3,0,3,2,3,1,3,2,1,1,0,0,4,2,1,0,2,3,1,4,3,2,0,4,4,3,1,3,1,3,0,1,0,0,1,0,0,0,1,0,0,0,0,4,1,1,1,2,0,3,0,0,0,3,4,2,4,3,2,0,1,0,0,3,3), -(0,1,0,4,0,5,0,4,0,2,4,4,2,3,3,2,3,3,5,3,3,3,4,3,4,2,3,0,4,3,3,3,4,1,4,3,2,1,5,5,3,4,5,1,3,5,4,2,0,3,3,0,1,3,0,4,2,0,1,3,1,4,3,3,3,3,0,3,0,1,0,3,4,4,4,5,5,0,3,0,1,4,5), -(0,2,0,3,0,3,0,0,0,2,3,1,3,0,4,0,1,1,3,0,3,4,3,2,3,1,0,3,3,2,3,1,3,0,2,3,0,2,1,4,1,2,2,0,0,3,3,0,0,2,0,0,0,1,0,0,0,0,2,2,0,3,2,1,3,3,0,2,0,2,0,0,3,3,1,2,4,0,3,0,2,2,3), -(2,4,0,5,0,4,0,4,0,2,4,4,4,3,4,3,3,3,1,2,4,3,4,3,4,4,5,0,3,3,3,3,2,0,4,3,1,4,3,4,1,4,4,3,3,4,4,3,1,2,3,0,4,2,0,4,1,0,3,3,0,4,3,3,3,4,0,4,0,2,0,3,5,3,4,5,2,0,3,0,0,4,5), -(0,3,0,4,0,1,0,1,0,1,3,2,2,1,3,0,3,0,2,0,2,0,3,0,2,0,0,0,1,0,1,1,0,0,3,1,0,0,0,4,0,3,1,0,2,1,3,0,0,0,0,0,0,3,0,0,0,0,0,0,0,4,2,2,3,1,0,3,0,0,0,1,4,4,4,3,0,0,4,0,0,1,4), -(1,4,1,5,0,3,0,3,0,4,5,4,4,3,5,3,3,4,4,3,4,1,3,3,3,3,2,1,4,1,5,4,3,1,4,4,3,5,4,4,3,5,4,3,3,4,4,4,0,3,3,1,2,3,0,3,1,0,3,3,0,5,4,4,4,4,4,4,3,3,5,4,4,3,3,5,4,0,3,2,0,4,4), -(0,2,0,3,0,1,0,0,0,1,3,3,3,2,4,1,3,0,3,1,3,0,2,2,1,1,0,0,2,0,4,3,1,0,4,3,0,4,4,4,1,4,3,1,1,3,3,1,0,2,0,0,1,3,0,0,0,0,2,0,0,4,3,2,4,3,5,4,3,3,3,4,3,3,4,3,3,0,2,1,0,3,3), -(0,2,0,4,0,3,0,2,0,2,5,5,3,4,4,4,4,1,4,3,3,0,4,3,4,3,1,3,3,2,4,3,0,3,4,3,0,3,4,4,2,4,4,0,4,5,3,3,2,2,1,1,1,2,0,1,5,0,3,3,2,4,3,3,3,4,0,3,0,2,0,4,4,3,5,5,0,0,3,0,2,3,3), -(0,3,0,4,0,3,0,1,0,3,4,3,3,1,3,3,3,0,3,1,3,0,4,3,3,1,1,0,3,0,3,3,0,0,4,4,0,1,5,4,3,3,5,0,3,3,4,3,0,2,0,1,1,1,0,1,3,0,1,2,1,3,3,2,3,3,0,3,0,1,0,1,3,3,4,4,1,0,1,2,2,1,3), -(0,1,0,4,0,4,0,3,0,1,3,3,3,2,3,1,1,0,3,0,3,3,4,3,2,4,2,0,1,0,4,3,2,0,4,3,0,5,3,3,2,4,4,4,3,3,3,4,0,1,3,0,0,1,0,0,1,0,0,0,0,4,2,3,3,3,0,3,0,0,0,4,4,4,5,3,2,0,3,3,0,3,5), -(0,2,0,3,0,0,0,3,0,1,3,0,2,0,0,0,1,0,3,1,1,3,3,0,0,3,0,0,3,0,2,3,1,0,3,1,0,3,3,2,0,4,2,2,0,2,0,0,0,4,0,0,0,0,0,0,0,0,0,0,0,2,1,2,0,1,0,1,0,0,0,1,3,1,2,0,0,0,1,0,0,1,4), -(0,3,0,3,0,5,0,1,0,2,4,3,1,3,3,2,1,1,5,2,1,0,5,1,2,0,0,0,3,3,2,2,3,2,4,3,0,0,3,3,1,3,3,0,2,5,3,4,0,3,3,0,1,2,0,2,2,0,3,2,0,2,2,3,3,3,0,2,0,1,0,3,4,4,2,5,4,0,3,0,0,3,5), -(0,3,0,3,0,3,0,1,0,3,3,3,3,0,3,0,2,0,2,1,1,0,2,0,1,0,0,0,2,1,0,0,1,0,3,2,0,0,3,3,1,2,3,1,0,3,3,0,0,1,0,0,0,0,0,2,0,0,0,0,0,2,3,1,2,3,0,3,0,1,0,3,2,1,0,4,3,0,1,1,0,3,3), -(0,4,0,5,0,3,0,3,0,4,5,5,4,3,5,3,4,3,5,3,3,2,5,3,4,4,4,3,4,3,4,5,5,3,4,4,3,4,4,5,4,4,4,3,4,5,5,4,2,3,4,2,3,4,0,3,3,1,4,3,2,4,3,3,5,5,0,3,0,3,0,5,5,5,5,4,4,0,4,0,1,4,4), -(0,4,0,4,0,3,0,3,0,3,5,4,4,2,3,2,5,1,3,2,5,1,4,2,3,2,3,3,4,3,3,3,3,2,5,4,1,3,3,5,3,4,4,0,4,4,3,1,1,3,1,0,2,3,0,2,3,0,3,0,0,4,3,1,3,4,0,3,0,2,0,4,4,4,3,4,5,0,4,0,0,3,4), -(0,3,0,3,0,3,1,2,0,3,4,4,3,3,3,0,2,2,4,3,3,1,3,3,3,1,1,0,3,1,4,3,2,3,4,4,2,4,4,4,3,4,4,3,2,4,4,3,1,3,3,1,3,3,0,4,1,0,2,2,1,4,3,2,3,3,5,4,3,3,5,4,4,3,3,0,4,0,3,2,2,4,4), -(0,2,0,1,0,0,0,0,0,1,2,1,3,0,0,0,0,0,2,0,1,2,1,0,0,1,0,0,0,0,3,0,0,1,0,1,1,3,1,0,0,0,1,1,0,1,1,0,0,0,0,0,2,0,0,0,0,0,0,0,0,1,1,2,2,0,3,4,0,0,0,1,1,0,0,1,0,0,0,0,0,1,1), -(0,1,0,0,0,1,0,0,0,0,4,0,4,1,4,0,3,0,4,0,3,0,4,0,3,0,3,0,4,1,5,1,4,0,0,3,0,5,0,5,2,0,1,0,0,0,2,1,4,0,1,3,0,0,3,0,0,3,1,1,4,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0), -(1,4,0,5,0,3,0,2,0,3,5,4,4,3,4,3,5,3,4,3,3,0,4,3,3,3,3,3,3,2,4,4,3,1,3,4,4,5,4,4,3,4,4,1,3,5,4,3,3,3,1,2,2,3,3,1,3,1,3,3,3,5,3,3,4,5,0,3,0,3,0,3,4,3,4,4,3,0,3,0,2,4,3), -(0,1,0,4,0,0,0,0,0,1,4,0,4,1,4,2,4,0,3,0,1,0,1,0,0,0,0,0,2,0,3,1,1,1,0,3,0,0,0,1,2,1,0,0,1,1,1,1,0,1,0,0,0,1,0,0,3,0,0,0,0,3,2,0,2,2,0,1,0,0,0,2,3,2,3,3,0,0,0,0,2,1,0), -(0,5,1,5,0,3,0,3,0,5,4,4,5,1,5,3,3,0,4,3,4,3,5,3,4,3,3,2,4,3,4,3,3,0,3,3,1,4,4,3,4,4,4,3,4,5,5,3,2,3,1,1,3,3,1,3,1,1,3,3,2,4,5,3,3,5,0,4,0,3,0,4,4,3,5,3,3,0,3,4,0,4,3), -(0,5,0,5,0,3,0,2,0,4,4,3,5,2,4,3,3,3,4,4,4,3,5,3,5,3,3,1,4,0,4,3,3,0,3,3,0,4,4,4,4,5,4,3,3,5,5,3,2,3,1,2,3,2,0,1,0,0,3,2,2,4,4,3,1,5,0,4,0,3,0,4,3,1,3,2,1,0,3,3,0,3,3), -(0,4,0,5,0,5,0,4,0,4,5,5,5,3,4,3,3,2,5,4,4,3,5,3,5,3,4,0,4,3,4,4,3,2,4,4,3,4,5,4,4,5,5,0,3,5,5,4,1,3,3,2,3,3,1,3,1,0,4,3,1,4,4,3,4,5,0,4,0,2,0,4,3,4,4,3,3,0,4,0,0,5,5), -(0,4,0,4,0,5,0,1,1,3,3,4,4,3,4,1,3,0,5,1,3,0,3,1,3,1,1,0,3,0,3,3,4,0,4,3,0,4,4,4,3,4,4,0,3,5,4,1,0,3,0,0,2,3,0,3,1,0,3,1,0,3,2,1,3,5,0,3,0,1,0,3,2,3,3,4,4,0,2,2,0,4,4), -(2,4,0,5,0,4,0,3,0,4,5,5,4,3,5,3,5,3,5,3,5,2,5,3,4,3,3,4,3,4,5,3,2,1,5,4,3,2,3,4,5,3,4,1,2,5,4,3,0,3,3,0,3,2,0,2,3,0,4,1,0,3,4,3,3,5,0,3,0,1,0,4,5,5,5,4,3,0,4,2,0,3,5), -(0,5,0,4,0,4,0,2,0,5,4,3,4,3,4,3,3,3,4,3,4,2,5,3,5,3,4,1,4,3,4,4,4,0,3,5,0,4,4,4,4,5,3,1,3,4,5,3,3,3,3,3,3,3,0,2,2,0,3,3,2,4,3,3,3,5,3,4,1,3,3,5,3,2,0,0,0,0,4,3,1,3,3), -(0,1,0,3,0,3,0,1,0,1,3,3,3,2,3,3,3,0,3,0,0,0,3,1,3,0,0,0,2,2,2,3,0,0,3,2,0,1,2,4,1,3,3,0,0,3,3,3,0,1,0,0,2,1,0,0,3,0,3,1,0,3,0,0,1,3,0,2,0,1,0,3,3,1,3,3,0,0,1,1,0,3,3), -(0,2,0,3,0,2,1,4,0,2,2,3,1,1,3,1,1,0,2,0,3,1,2,3,1,3,0,0,1,0,4,3,2,3,3,3,1,4,2,3,3,3,3,1,0,3,1,4,0,1,1,0,1,2,0,1,1,0,1,1,0,3,1,3,2,2,0,1,0,0,0,2,3,3,3,1,0,0,0,0,0,2,3), -(0,5,0,4,0,5,0,2,0,4,5,5,3,3,4,3,3,1,5,4,4,2,4,4,4,3,4,2,4,3,5,5,4,3,3,4,3,3,5,5,4,5,5,1,3,4,5,3,1,4,3,1,3,3,0,3,3,1,4,3,1,4,5,3,3,5,0,4,0,3,0,5,3,3,1,4,3,0,4,0,1,5,3), -(0,5,0,5,0,4,0,2,0,4,4,3,4,3,3,3,3,3,5,4,4,4,4,4,4,5,3,3,5,2,4,4,4,3,4,4,3,3,4,4,5,5,3,3,4,3,4,3,3,4,3,3,3,3,1,2,2,1,4,3,3,5,4,4,3,4,0,4,0,3,0,4,4,4,4,4,1,0,4,2,0,2,4), -(0,4,0,4,0,3,0,1,0,3,5,2,3,0,3,0,2,1,4,2,3,3,4,1,4,3,3,2,4,1,3,3,3,0,3,3,0,0,3,3,3,5,3,3,3,3,3,2,0,2,0,0,2,0,0,2,0,0,1,0,0,3,1,2,2,3,0,3,0,2,0,4,4,3,3,4,1,0,3,0,0,2,4), -(0,0,0,4,0,0,0,0,0,0,1,0,1,0,2,0,0,0,0,0,1,0,2,0,1,0,0,0,0,0,3,1,3,0,3,2,0,0,0,1,0,3,2,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,4,0,2,0,0,0,0,0,0,2), -(0,2,1,3,0,2,0,2,0,3,3,3,3,1,3,1,3,3,3,3,3,3,4,2,2,1,2,1,4,0,4,3,1,3,3,3,2,4,3,5,4,3,3,3,3,3,3,3,0,1,3,0,2,0,0,1,0,0,1,0,0,4,2,0,2,3,0,3,3,0,3,3,4,2,3,1,4,0,1,2,0,2,3), -(0,3,0,3,0,1,0,3,0,2,3,3,3,0,3,1,2,0,3,3,2,3,3,2,3,2,3,1,3,0,4,3,2,0,3,3,1,4,3,3,2,3,4,3,1,3,3,1,1,0,1,1,0,1,0,1,0,1,0,0,0,4,1,1,0,3,0,3,1,0,2,3,3,3,3,3,1,0,0,2,0,3,3), -(0,0,0,0,0,0,0,0,0,0,3,0,2,0,3,0,0,0,0,0,0,0,3,0,0,0,0,0,0,0,3,0,3,0,3,1,0,1,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,0,2,0,2,3,0,0,0,0,0,0,0,0,3), -(0,2,0,3,1,3,0,3,0,2,3,3,3,1,3,1,3,1,3,1,3,3,3,1,3,0,2,3,1,1,4,3,3,2,3,3,1,2,2,4,1,3,3,0,1,4,2,3,0,1,3,0,3,0,0,1,3,0,2,0,0,3,3,2,1,3,0,3,0,2,0,3,4,4,4,3,1,0,3,0,0,3,3), -(0,2,0,1,0,2,0,0,0,1,3,2,2,1,3,0,1,1,3,0,3,2,3,1,2,0,2,0,1,1,3,3,3,0,3,3,1,1,2,3,2,3,3,1,2,3,2,0,0,1,0,0,0,0,0,0,3,0,1,0,0,2,1,2,1,3,0,3,0,0,0,3,4,4,4,3,2,0,2,0,0,2,4), -(0,0,0,1,0,1,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,2,2,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,3,1,0,0,0,0,0,0,0,3), -(0,3,0,3,0,2,0,3,0,3,3,3,2,3,2,2,2,0,3,1,3,3,3,2,3,3,0,0,3,0,3,2,2,0,2,3,1,4,3,4,3,3,2,3,1,5,4,4,0,3,1,2,1,3,0,3,1,1,2,0,2,3,1,3,1,3,0,3,0,1,0,3,3,4,4,2,1,0,2,1,0,2,4), -(0,1,0,3,0,1,0,2,0,1,4,2,5,1,4,0,2,0,2,1,3,1,4,0,2,1,0,0,2,1,4,1,1,0,3,3,0,5,1,3,2,3,3,1,0,3,2,3,0,1,0,0,0,0,0,0,1,0,0,0,0,4,0,1,0,3,0,2,0,1,0,3,3,3,4,3,3,0,0,0,0,2,3), -(0,0,0,1,0,0,0,0,0,0,2,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,3,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,1,0,0,1,0,0,0,0,0,3), -(0,1,0,3,0,4,0,3,0,2,4,3,1,0,3,2,2,1,3,1,2,2,3,1,1,1,2,1,3,0,1,2,0,1,3,2,1,3,0,5,5,1,0,0,1,3,2,1,0,3,0,0,1,0,0,0,0,0,3,4,0,1,1,1,3,2,0,2,0,1,0,2,3,3,1,2,3,0,1,0,1,0,4), -(0,0,0,1,0,3,0,3,0,2,2,1,0,0,4,0,3,0,3,1,3,0,3,0,3,0,1,0,3,0,3,1,3,0,3,3,0,0,1,2,1,1,1,0,1,2,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,2,2,1,2,0,0,2,0,0,0,0,2,3,3,3,3,0,0,0,0,1,4), -(0,0,0,3,0,3,0,0,0,0,3,1,1,0,3,0,1,0,2,0,1,0,0,0,0,0,0,0,1,0,3,0,2,0,2,3,0,0,2,2,3,1,2,0,0,1,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,0,0,2,0,0,0,0,2,3), -(2,4,0,5,0,5,0,4,0,3,4,3,3,3,4,3,3,3,4,3,4,4,5,4,5,5,5,2,3,0,5,5,4,1,5,4,3,1,5,4,3,4,4,3,3,4,3,3,0,3,2,0,2,3,0,3,0,0,3,3,0,5,3,2,3,3,0,3,0,3,0,3,4,5,4,5,3,0,4,3,0,3,4), -(0,3,0,3,0,3,0,3,0,3,3,4,3,2,3,2,3,0,4,3,3,3,3,3,3,3,3,0,3,2,4,3,3,1,3,4,3,4,4,4,3,4,4,3,2,4,4,1,0,2,0,0,1,1,0,2,0,0,3,1,0,5,3,2,1,3,0,3,0,1,2,4,3,2,4,3,3,0,3,2,0,4,4), -(0,3,0,3,0,1,0,0,0,1,4,3,3,2,3,1,3,1,4,2,3,2,4,2,3,4,3,0,2,2,3,3,3,0,3,3,3,0,3,4,1,3,3,0,3,4,3,3,0,1,1,0,1,0,0,0,4,0,3,0,0,3,1,2,1,3,0,4,0,1,0,4,3,3,4,3,3,0,2,0,0,3,3), -(0,3,0,4,0,1,0,3,0,3,4,3,3,0,3,3,3,1,3,1,3,3,4,3,3,3,0,0,3,1,5,3,3,1,3,3,2,5,4,3,3,4,5,3,2,5,3,4,0,1,0,0,0,0,0,2,0,0,1,1,0,4,2,2,1,3,0,3,0,2,0,4,4,3,5,3,2,0,1,1,0,3,4), -(0,5,0,4,0,5,0,2,0,4,4,3,3,2,3,3,3,1,4,3,4,1,5,3,4,3,4,0,4,2,4,3,4,1,5,4,0,4,4,4,4,5,4,1,3,5,4,2,1,4,1,1,3,2,0,3,1,0,3,2,1,4,3,3,3,4,0,4,0,3,0,4,4,4,3,3,3,0,4,2,0,3,4), -(1,4,0,4,0,3,0,1,0,3,3,3,1,1,3,3,2,2,3,3,1,0,3,2,2,1,2,0,3,1,2,1,2,0,3,2,0,2,2,3,3,4,3,0,3,3,1,2,0,1,1,3,1,2,0,0,3,0,1,1,0,3,2,2,3,3,0,3,0,0,0,2,3,3,4,3,3,0,1,0,0,1,4), -(0,4,0,4,0,4,0,0,0,3,4,4,3,1,4,2,3,2,3,3,3,1,4,3,4,0,3,0,4,2,3,3,2,2,5,4,2,1,3,4,3,4,3,1,3,3,4,2,0,2,1,0,3,3,0,0,2,0,3,1,0,4,4,3,4,3,0,4,0,1,0,2,4,4,4,4,4,0,3,2,0,3,3), -(0,0,0,1,0,4,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,1,0,3,2,0,0,1,0,0,0,1,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,2), -(0,2,0,3,0,4,0,4,0,1,3,3,3,0,4,0,2,1,2,1,1,1,2,0,3,1,1,0,1,0,3,1,0,0,3,3,2,0,1,1,0,0,0,0,0,1,0,2,0,2,2,0,3,1,0,0,1,0,1,1,0,1,2,0,3,0,0,0,0,1,0,0,3,3,4,3,1,0,1,0,3,0,2), -(0,0,0,3,0,5,0,0,0,0,1,0,2,0,3,1,0,1,3,0,0,0,2,0,0,0,1,0,0,0,1,1,0,0,4,0,0,0,2,3,0,1,4,1,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,3,0,0,0,0,0,1,0,0,0,0,0,0,0,2,0,0,3,0,0,0,0,0,3), -(0,2,0,5,0,5,0,1,0,2,4,3,3,2,5,1,3,2,3,3,3,0,4,1,2,0,3,0,4,0,2,2,1,1,5,3,0,0,1,4,2,3,2,0,3,3,3,2,0,2,4,1,1,2,0,1,1,0,3,1,0,1,3,1,2,3,0,2,0,0,0,1,3,5,4,4,4,0,3,0,0,1,3), -(0,4,0,5,0,4,0,4,0,4,5,4,3,3,4,3,3,3,4,3,4,4,5,3,4,5,4,2,4,2,3,4,3,1,4,4,1,3,5,4,4,5,5,4,4,5,5,5,2,3,3,1,4,3,1,3,3,0,3,3,1,4,3,4,4,4,0,3,0,4,0,3,3,4,4,5,0,0,4,3,0,4,5), -(0,4,0,4,0,3,0,3,0,3,4,4,4,3,3,2,4,3,4,3,4,3,5,3,4,3,2,1,4,2,4,4,3,1,3,4,2,4,5,5,3,4,5,4,1,5,4,3,0,3,2,2,3,2,1,3,1,0,3,3,3,5,3,3,3,5,4,4,2,3,3,4,3,3,3,2,1,0,3,2,1,4,3), -(0,4,0,5,0,4,0,3,0,3,5,5,3,2,4,3,4,0,5,4,4,1,4,4,4,3,3,3,4,3,5,5,2,3,3,4,1,2,5,5,3,5,5,2,3,5,5,4,0,3,2,0,3,3,1,1,5,1,4,1,0,4,3,2,3,5,0,4,0,3,0,5,4,3,4,3,0,0,4,1,0,4,4), -(1,3,0,4,0,2,0,2,0,2,5,5,3,3,3,3,3,0,4,2,3,4,4,4,3,4,0,0,3,4,5,4,3,3,3,3,2,5,5,4,5,5,5,4,3,5,5,5,1,3,1,0,1,0,0,3,2,0,4,2,0,5,2,3,2,4,1,3,0,3,0,4,5,4,5,4,3,0,4,2,0,5,4), -(0,3,0,4,0,5,0,3,0,3,4,4,3,2,3,2,3,3,3,3,3,2,4,3,3,2,2,0,3,3,3,3,3,1,3,3,3,0,4,4,3,4,4,1,1,4,4,2,0,3,1,0,1,1,0,4,1,0,2,3,1,3,3,1,3,4,0,3,0,1,0,3,1,3,0,0,1,0,2,0,0,4,4), -(0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0), -(0,3,0,3,0,2,0,3,0,1,5,4,3,3,3,1,4,2,1,2,3,4,4,2,4,4,5,0,3,1,4,3,4,0,4,3,3,3,2,3,2,5,3,4,3,2,2,3,0,0,3,0,2,1,0,1,2,0,0,0,0,2,1,1,3,1,0,2,0,4,0,3,4,4,4,5,2,0,2,0,0,1,3), -(0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,1,1,1,0,0,1,1,0,0,0,4,2,1,1,0,1,0,3,2,0,0,3,1,1,1,2,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,0,1,0,0,0,2,0,0,0,1,4,0,4,2,1,0,0,0,0,0,1), -(0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,1,0,0,0,0,0,0,1,0,1,0,0,0,0,3,1,0,0,0,2,0,2,1,0,0,1,2,1,0,1,1,0,0,3,0,0,0,0,0,0,0,0,0,0,0,1,3,1,0,0,0,0,0,1,0,0,2,1,0,0,0,0,0,0,0,0,2), -(0,4,0,4,0,4,0,3,0,4,4,3,4,2,4,3,2,0,4,4,4,3,5,3,5,3,3,2,4,2,4,3,4,3,1,4,0,2,3,4,4,4,3,3,3,4,4,4,3,4,1,3,4,3,2,1,2,1,3,3,3,4,4,3,3,5,0,4,0,3,0,4,3,3,3,2,1,0,3,0,0,3,3), -(0,4,0,3,0,3,0,3,0,3,5,5,3,3,3,3,4,3,4,3,3,3,4,4,4,3,3,3,3,4,3,5,3,3,1,3,2,4,5,5,5,5,4,3,4,5,5,3,2,2,3,3,3,3,2,3,3,1,2,3,2,4,3,3,3,4,0,4,0,2,0,4,3,2,2,1,2,0,3,0,0,4,1), -) - -class JapaneseContextAnalysis(object): - NUM_OF_CATEGORY = 6 - DONT_KNOW = -1 - ENOUGH_REL_THRESHOLD = 100 - MAX_REL_THRESHOLD = 1000 - MINIMUM_DATA_THRESHOLD = 4 - - def __init__(self): - self._total_rel = None - self._rel_sample = None - self._need_to_skip_char_num = None - self._last_char_order = None - self._done = None - self.reset() - - def reset(self): - self._total_rel = 0 # total sequence received - # category counters, each integer counts sequence in its category - self._rel_sample = [0] * self.NUM_OF_CATEGORY - # if last byte in current buffer is not the last byte of a character, - # we need to know how many bytes to skip in next buffer - self._need_to_skip_char_num = 0 - self._last_char_order = -1 # The order of previous char - # If this flag is set to True, detection is done and conclusion has - # been made - self._done = False - - def feed(self, byte_str, num_bytes): - if self._done: - return - - # The buffer we got is byte oriented, and a character may span in more than one - # buffers. In case the last one or two byte in last buffer is not - # complete, we record how many byte needed to complete that character - # and skip these bytes here. We can choose to record those bytes as - # well and analyse the character once it is complete, but since a - # character will not make much difference, by simply skipping - # this character will simply our logic and improve performance. - i = self._need_to_skip_char_num - while i < num_bytes: - order, char_len = self.get_order(byte_str[i:i + 2]) - i += char_len - if i > num_bytes: - self._need_to_skip_char_num = i - num_bytes - self._last_char_order = -1 - else: - if (order != -1) and (self._last_char_order != -1): - self._total_rel += 1 - if self._total_rel > self.MAX_REL_THRESHOLD: - self._done = True - break - self._rel_sample[jp2CharContext[self._last_char_order][order]] += 1 - self._last_char_order = order - - def got_enough_data(self): - return self._total_rel > self.ENOUGH_REL_THRESHOLD - - def get_confidence(self): - # This is just one way to calculate confidence. It works well for me. - if self._total_rel > self.MINIMUM_DATA_THRESHOLD: - return (self._total_rel - self._rel_sample[0]) / self._total_rel - else: - return self.DONT_KNOW - - def get_order(self, byte_str): - return -1, 1 - -class SJISContextAnalysis(JapaneseContextAnalysis): - def __init__(self): - super(SJISContextAnalysis, self).__init__() - self._charset_name = "SHIFT_JIS" - - @property - def charset_name(self): - return self._charset_name - - def get_order(self, byte_str): - if not byte_str: - return -1, 1 - # find out current char's byte length - first_char = byte_str[0] - if (0x81 <= first_char <= 0x9F) or (0xE0 <= first_char <= 0xFC): - char_len = 2 - if (first_char == 0x87) or (0xFA <= first_char <= 0xFC): - self._charset_name = "CP932" - else: - char_len = 1 - - # return its order if it is hiragana - if len(byte_str) > 1: - second_char = byte_str[1] - if (first_char == 202) and (0x9F <= second_char <= 0xF1): - return second_char - 0x9F, char_len - - return -1, char_len - -class EUCJPContextAnalysis(JapaneseContextAnalysis): - def get_order(self, byte_str): - if not byte_str: - return -1, 1 - # find out current char's byte length - first_char = byte_str[0] - if (first_char == 0x8E) or (0xA1 <= first_char <= 0xFE): - char_len = 2 - elif first_char == 0x8F: - char_len = 3 - else: - char_len = 1 - - # return its order if it is hiragana - if len(byte_str) > 1: - second_char = byte_str[1] - if (first_char == 0xA4) and (0xA1 <= second_char <= 0xF3): - return second_char - 0xA1, char_len - - return -1, char_len - - diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/langbulgarianmodel.py b/src/fetchcode/vcs/pip/_vendor/chardet/langbulgarianmodel.py deleted file mode 100644 index 2aa4fb2e..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/langbulgarianmodel.py +++ /dev/null @@ -1,228 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Communicator client code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -# 255: Control characters that usually does not exist in any text -# 254: Carriage/Return -# 253: symbol (punctuation) that does not belong to word -# 252: 0 - 9 - -# Character Mapping Table: -# this table is modified base on win1251BulgarianCharToOrderMap, so -# only number <64 is sure valid - -Latin5_BulgarianCharToOrderMap = ( -255,255,255,255,255,255,255,255,255,255,254,255,255,254,255,255, # 00 -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 10 -253,253,253,253,253,253,253,253,253,253,253,253,253,253,253,253, # 20 -252,252,252,252,252,252,252,252,252,252,253,253,253,253,253,253, # 30 -253, 77, 90, 99,100, 72,109,107,101, 79,185, 81,102, 76, 94, 82, # 40 -110,186,108, 91, 74,119, 84, 96,111,187,115,253,253,253,253,253, # 50 -253, 65, 69, 70, 66, 63, 68,112,103, 92,194,104, 95, 86, 87, 71, # 60 -116,195, 85, 93, 97,113,196,197,198,199,200,253,253,253,253,253, # 70 -194,195,196,197,198,199,200,201,202,203,204,205,206,207,208,209, # 80 -210,211,212,213,214,215,216,217,218,219,220,221,222,223,224,225, # 90 - 81,226,227,228,229,230,105,231,232,233,234,235,236, 45,237,238, # a0 - 31, 32, 35, 43, 37, 44, 55, 47, 40, 59, 33, 46, 38, 36, 41, 30, # b0 - 39, 28, 34, 51, 48, 49, 53, 50, 54, 57, 61,239, 67,240, 60, 56, # c0 - 1, 18, 9, 20, 11, 3, 23, 15, 2, 26, 12, 10, 14, 6, 4, 13, # d0 - 7, 8, 5, 19, 29, 25, 22, 21, 27, 24, 17, 75, 52,241, 42, 16, # e0 - 62,242,243,244, 58,245, 98,246,247,248,249,250,251, 91,252,253, # f0 -) - -win1251BulgarianCharToOrderMap = ( -255,255,255,255,255,255,255,255,255,255,254,255,255,254,255,255, # 00 -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 10 -253,253,253,253,253,253,253,253,253,253,253,253,253,253,253,253, # 20 -252,252,252,252,252,252,252,252,252,252,253,253,253,253,253,253, # 30 -253, 77, 90, 99,100, 72,109,107,101, 79,185, 81,102, 76, 94, 82, # 40 -110,186,108, 91, 74,119, 84, 96,111,187,115,253,253,253,253,253, # 50 -253, 65, 69, 70, 66, 63, 68,112,103, 92,194,104, 95, 86, 87, 71, # 60 -116,195, 85, 93, 97,113,196,197,198,199,200,253,253,253,253,253, # 70 -206,207,208,209,210,211,212,213,120,214,215,216,217,218,219,220, # 80 -221, 78, 64, 83,121, 98,117,105,222,223,224,225,226,227,228,229, # 90 - 88,230,231,232,233,122, 89,106,234,235,236,237,238, 45,239,240, # a0 - 73, 80,118,114,241,242,243,244,245, 62, 58,246,247,248,249,250, # b0 - 31, 32, 35, 43, 37, 44, 55, 47, 40, 59, 33, 46, 38, 36, 41, 30, # c0 - 39, 28, 34, 51, 48, 49, 53, 50, 54, 57, 61,251, 67,252, 60, 56, # d0 - 1, 18, 9, 20, 11, 3, 23, 15, 2, 26, 12, 10, 14, 6, 4, 13, # e0 - 7, 8, 5, 19, 29, 25, 22, 21, 27, 24, 17, 75, 52,253, 42, 16, # f0 -) - -# Model Table: -# total sequences: 100% -# first 512 sequences: 96.9392% -# first 1024 sequences:3.0618% -# rest sequences: 0.2992% -# negative sequences: 0.0020% -BulgarianLangModel = ( -0,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,2,3,3,3,3,3,3,3,3,2,3,3,3,3,3, -3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,0,3,3,3,2,2,3,2,2,1,2,2, -3,1,3,3,2,3,3,3,3,3,3,3,3,3,3,3,3,0,3,3,3,3,3,3,3,3,3,3,0,3,0,1, -0,0,0,0,0,0,0,0,0,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1, -3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,2,3,2,3,3,3,3,3,3,3,3,0,3,1,0, -0,1,0,0,0,0,0,0,0,0,1,1,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1, -3,2,2,2,3,3,3,3,3,3,3,3,3,3,3,3,3,1,3,2,3,3,3,3,3,3,3,3,0,3,0,0, -0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,2,3,3,2,3,3,3,3,3,3,3,3,3,3,3,3,1,3,2,3,3,3,3,3,3,3,3,0,3,0,0, -0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,3,3,3,3,3,3,2,3,2,2,1,3,3,3,3,2,2,2,1,1,2,0,1,0,1,0,0, -0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,1, -3,3,3,3,3,3,3,2,3,2,2,3,3,1,1,2,3,3,2,3,3,3,3,2,1,2,0,2,0,3,0,0, -0,0,0,0,0,0,0,1,0,0,2,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,1, -3,3,3,3,3,3,3,1,3,3,3,3,3,2,3,2,3,3,3,3,3,2,3,3,1,3,0,3,0,2,0,0, -0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1, -3,3,3,3,3,3,3,3,1,3,3,2,3,3,3,1,3,3,2,3,2,2,2,0,0,2,0,2,0,2,0,0, -0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,1, -3,3,3,3,3,3,3,3,3,0,3,3,3,2,2,3,3,3,1,2,2,3,2,1,1,2,0,2,0,0,0,0, -1,0,0,0,0,0,0,0,0,0,2,0,0,1,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1, -3,3,3,3,3,3,3,2,3,3,1,2,3,2,2,2,3,3,3,3,3,2,2,3,1,2,0,2,1,2,0,0, -0,0,0,0,0,0,0,0,0,0,3,0,0,1,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,1, -3,3,3,3,3,1,3,3,3,3,3,2,3,3,3,2,3,3,2,3,2,2,2,3,1,2,0,1,0,1,0,0, -0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1, -3,3,3,3,3,3,3,3,3,3,3,1,1,1,2,2,1,3,1,3,2,2,3,0,0,1,0,1,0,1,0,0, -0,0,0,1,0,0,0,0,1,0,2,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1, -3,3,3,3,3,2,2,3,2,2,3,1,2,1,1,1,2,3,1,3,1,2,2,0,1,1,1,1,0,1,0,0, -0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1, -3,3,3,3,3,1,3,2,2,3,3,1,2,3,1,1,3,3,3,3,1,2,2,1,1,1,0,2,0,2,0,1, -0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1, -3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,1,2,2,3,3,3,2,2,1,1,2,0,2,0,1,0,0, -0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1, -3,0,1,2,1,3,3,2,3,3,3,3,3,2,3,2,1,0,3,1,2,1,2,1,2,3,2,1,0,1,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -1,1,1,2,3,3,3,3,3,3,3,3,3,3,3,3,0,0,3,1,3,3,2,3,3,2,2,2,0,1,0,0, -0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,3,3,3,3,0,3,3,3,3,3,2,1,1,2,1,3,3,0,3,1,1,1,1,3,2,0,1,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1, -3,3,2,2,2,3,3,3,3,3,3,3,3,3,3,3,1,1,3,1,3,3,2,3,2,2,2,3,0,2,0,0, -0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,2,3,3,2,2,3,2,1,1,1,1,1,3,1,3,1,1,0,0,0,1,0,0,0,1,0,0, -0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,2,3,2,0,3,2,0,3,0,2,0,0,2,1,3,1,0,0,1,0,0,0,1,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1, -3,3,3,3,2,1,1,1,1,2,1,1,2,1,1,1,2,2,1,2,1,1,1,0,1,1,0,1,0,1,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1, -3,3,3,3,2,1,3,1,1,2,1,3,2,1,1,0,1,2,3,2,1,1,1,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,3,3,3,3,2,2,1,0,1,0,0,1,0,0,0,2,1,0,3,0,0,1,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1, -3,3,3,2,3,2,3,3,1,3,2,1,1,1,2,1,1,2,1,3,0,1,0,0,0,1,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,1,1,2,2,3,3,2,3,2,2,2,3,1,2,2,1,1,2,1,1,2,2,0,1,1,0,1,0,2,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,2,1,3,1,0,2,2,1,3,2,1,0,0,2,0,2,0,1,0,0,0,0,0,0,0,1,0,0, -0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1, -3,3,3,3,3,3,1,2,0,2,3,1,2,3,2,0,1,3,1,2,1,1,1,0,0,1,0,0,2,2,2,3, -2,2,2,2,1,2,1,1,2,2,1,1,2,0,1,1,1,0,0,1,1,0,0,1,1,0,0,0,1,1,0,1, -3,3,3,3,3,2,1,2,2,1,2,0,2,0,1,0,1,2,1,2,1,1,0,0,0,1,0,1,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,1, -3,3,2,3,3,1,1,3,1,0,3,2,1,0,0,0,1,2,0,2,0,1,0,0,0,1,0,1,2,1,2,2, -1,1,1,1,1,1,1,2,2,2,1,1,1,1,1,1,1,0,1,2,1,1,1,0,0,0,0,0,1,1,0,0, -3,1,0,1,0,2,3,2,2,2,3,2,2,2,2,2,1,0,2,1,2,1,1,1,0,1,2,1,2,2,2,1, -1,1,2,2,2,2,1,2,1,1,0,1,2,1,2,2,2,1,1,1,0,1,1,1,1,2,0,1,0,0,0,0, -2,3,2,3,3,0,0,2,1,0,2,1,0,0,0,0,2,3,0,2,0,0,0,0,0,1,0,0,2,0,1,2, -2,1,2,1,2,2,1,1,1,2,1,1,1,0,1,2,2,1,1,1,1,1,0,1,1,1,0,0,1,2,0,0, -3,3,2,2,3,0,2,3,1,1,2,0,0,0,1,0,0,2,0,2,0,0,0,1,0,1,0,1,2,0,2,2, -1,1,1,1,2,1,0,1,2,2,2,1,1,1,1,1,1,1,0,1,1,1,0,0,0,0,0,0,1,1,0,0, -2,3,2,3,3,0,0,3,0,1,1,0,1,0,0,0,2,2,1,2,0,0,0,0,0,0,0,0,2,0,1,2, -2,2,1,1,1,1,1,2,2,2,1,0,2,0,1,0,1,0,0,1,0,1,0,0,1,0,0,0,0,1,0,0, -3,3,3,3,2,2,2,2,2,0,2,1,1,1,1,2,1,2,1,1,0,2,0,1,0,1,0,0,2,0,1,2, -1,1,1,1,1,1,1,2,2,1,1,0,2,0,1,0,2,0,0,1,1,1,0,0,2,0,0,0,1,1,0,0, -2,3,3,3,3,1,0,0,0,0,0,0,0,0,0,0,2,0,0,1,1,0,0,0,0,0,0,1,2,0,1,2, -2,2,2,1,1,2,1,1,2,2,2,1,2,0,1,1,1,1,1,1,0,1,1,1,1,0,0,1,1,1,0,0, -2,3,3,3,3,0,2,2,0,2,1,0,0,0,1,1,1,2,0,2,0,0,0,3,0,0,0,0,2,0,2,2, -1,1,1,2,1,2,1,1,2,2,2,1,2,0,1,1,1,0,1,1,1,1,0,2,1,0,0,0,1,1,0,0, -2,3,3,3,3,0,2,1,0,0,2,0,0,0,0,0,1,2,0,2,0,0,0,0,0,0,0,0,2,0,1,2, -1,1,1,2,1,1,1,1,2,2,2,0,1,0,1,1,1,0,0,1,1,1,0,0,1,0,0,0,0,1,0,0, -3,3,2,2,3,0,1,0,1,0,0,0,0,0,0,0,1,1,0,3,0,0,0,0,0,0,0,0,1,0,2,2, -1,1,1,1,1,2,1,1,2,2,1,2,2,1,0,1,1,1,1,1,0,1,0,0,1,0,0,0,1,1,0,0, -3,1,0,1,0,2,2,2,2,3,2,1,1,1,2,3,0,0,1,0,2,1,1,0,1,1,1,1,2,1,1,1, -1,2,2,1,2,1,2,2,1,1,0,1,2,1,2,2,1,1,1,0,0,1,1,1,2,1,0,1,0,0,0,0, -2,1,0,1,0,3,1,2,2,2,2,1,2,2,1,1,1,0,2,1,2,2,1,1,2,1,1,0,2,1,1,1, -1,2,2,2,2,2,2,2,1,2,0,1,1,0,2,1,1,1,1,1,0,0,1,1,1,1,0,1,0,0,0,0, -2,1,1,1,1,2,2,2,2,1,2,2,2,1,2,2,1,1,2,1,2,3,2,2,1,1,1,1,0,1,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,2,2,3,2,0,1,2,0,1,2,1,1,0,1,0,1,2,1,2,0,0,0,1,1,0,0,0,1,0,0,2, -1,1,0,0,1,1,0,1,1,1,1,0,2,0,1,1,1,0,0,1,1,0,0,0,0,1,0,0,0,1,0,0, -2,0,0,0,0,1,2,2,2,2,2,2,2,1,2,1,1,1,1,1,1,1,0,1,1,1,1,1,2,1,1,1, -1,2,2,2,2,1,1,2,1,2,1,1,1,0,2,1,2,1,1,1,0,2,1,1,1,1,0,1,0,0,0,0, -3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0, -1,1,0,1,0,1,1,1,1,1,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,2,2,3,2,0,0,0,0,1,0,0,0,0,0,0,1,1,0,2,0,0,0,0,0,0,0,0,1,0,1,2, -1,1,1,1,1,1,0,0,2,2,2,2,2,0,1,1,0,1,1,1,1,1,0,0,1,0,0,0,1,1,0,1, -2,3,1,2,1,0,1,1,0,2,2,2,0,0,1,0,0,1,1,1,1,0,0,0,0,0,0,0,1,0,1,2, -1,1,1,1,2,1,1,1,1,1,1,1,1,0,1,1,0,1,0,1,0,1,0,0,1,0,0,0,0,1,0,0, -2,2,2,2,2,0,0,2,0,0,2,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,2,0,2,2, -1,1,1,1,1,0,0,1,2,1,1,0,1,0,1,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0, -1,2,2,2,2,0,0,2,0,1,1,0,0,0,1,0,0,2,0,2,0,0,0,0,0,0,0,0,0,0,1,1, -0,0,0,1,1,1,1,1,1,1,1,1,1,0,1,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0, -1,2,2,3,2,0,0,1,0,0,1,0,0,0,0,0,0,1,0,2,0,0,0,1,0,0,0,0,0,0,0,2, -1,1,0,0,1,0,0,0,1,1,0,0,1,0,1,1,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0, -2,1,2,2,2,1,2,1,2,2,1,1,2,1,1,1,0,1,1,1,1,2,0,1,0,1,1,1,1,0,1,1, -1,1,2,1,1,1,1,1,1,0,0,1,2,1,1,1,1,1,1,0,0,1,1,1,0,0,0,0,0,0,0,0, -1,0,0,1,3,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,2,2,2,1,0,0,1,0,2,0,0,0,0,0,1,1,1,0,1,0,0,0,0,0,0,0,0,2,0,0,1, -0,2,0,1,0,0,1,1,2,0,1,0,1,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0, -1,2,2,2,2,0,1,1,0,2,1,0,1,1,1,0,0,1,0,2,0,1,0,0,0,0,0,0,0,0,0,1, -0,1,0,0,1,0,0,0,1,1,0,0,1,0,0,1,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0, -2,2,2,2,2,0,0,1,0,0,0,1,0,1,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,1, -0,1,0,1,1,1,0,0,1,1,1,0,1,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0, -2,0,1,0,0,1,2,1,1,1,1,1,1,2,2,1,0,0,1,0,1,0,0,0,0,1,1,1,1,0,0,0, -1,1,2,1,1,1,1,0,0,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,2,1,2,1,0,0,1,0,0,0,0,0,0,0,0,1,1,0,1,0,0,0,0,0,0,0,0,0,0,0,1, -0,0,0,0,0,0,0,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -1,0,0,1,2,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1,0,0,0, -0,1,1,0,1,1,1,0,0,1,0,0,1,0,1,0,0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0, -1,0,1,0,0,1,1,1,1,1,1,1,1,1,1,1,0,0,1,0,2,0,0,2,0,1,0,0,1,0,0,1, -1,1,0,0,1,1,0,1,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0, -1,1,1,1,1,1,1,2,0,0,0,0,0,0,2,1,0,1,1,0,0,1,1,1,0,1,0,0,0,0,0,0, -2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -1,0,0,1,1,1,1,1,1,1,1,1,1,1,1,1,1,0,1,0,1,1,0,1,1,1,1,1,0,1,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1, -) - -Latin5BulgarianModel = { - 'char_to_order_map': Latin5_BulgarianCharToOrderMap, - 'precedence_matrix': BulgarianLangModel, - 'typical_positive_ratio': 0.969392, - 'keep_english_letter': False, - 'charset_name': "ISO-8859-5", - 'language': 'Bulgairan', -} - -Win1251BulgarianModel = { - 'char_to_order_map': win1251BulgarianCharToOrderMap, - 'precedence_matrix': BulgarianLangModel, - 'typical_positive_ratio': 0.969392, - 'keep_english_letter': False, - 'charset_name': "windows-1251", - 'language': 'Bulgarian', -} diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/langcyrillicmodel.py b/src/fetchcode/vcs/pip/_vendor/chardet/langcyrillicmodel.py deleted file mode 100644 index e5f9a1fd..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/langcyrillicmodel.py +++ /dev/null @@ -1,333 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Communicator client code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -# KOI8-R language model -# Character Mapping Table: -KOI8R_char_to_order_map = ( -255,255,255,255,255,255,255,255,255,255,254,255,255,254,255,255, # 00 -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 10 -253,253,253,253,253,253,253,253,253,253,253,253,253,253,253,253, # 20 -252,252,252,252,252,252,252,252,252,252,253,253,253,253,253,253, # 30 -253,142,143,144,145,146,147,148,149,150,151,152, 74,153, 75,154, # 40 -155,156,157,158,159,160,161,162,163,164,165,253,253,253,253,253, # 50 -253, 71,172, 66,173, 65,174, 76,175, 64,176,177, 77, 72,178, 69, # 60 - 67,179, 78, 73,180,181, 79,182,183,184,185,253,253,253,253,253, # 70 -191,192,193,194,195,196,197,198,199,200,201,202,203,204,205,206, # 80 -207,208,209,210,211,212,213,214,215,216,217,218,219,220,221,222, # 90 -223,224,225, 68,226,227,228,229,230,231,232,233,234,235,236,237, # a0 -238,239,240,241,242,243,244,245,246,247,248,249,250,251,252,253, # b0 - 27, 3, 21, 28, 13, 2, 39, 19, 26, 4, 23, 11, 8, 12, 5, 1, # c0 - 15, 16, 9, 7, 6, 14, 24, 10, 17, 18, 20, 25, 30, 29, 22, 54, # d0 - 59, 37, 44, 58, 41, 48, 53, 46, 55, 42, 60, 36, 49, 38, 31, 34, # e0 - 35, 43, 45, 32, 40, 52, 56, 33, 61, 62, 51, 57, 47, 63, 50, 70, # f0 -) - -win1251_char_to_order_map = ( -255,255,255,255,255,255,255,255,255,255,254,255,255,254,255,255, # 00 -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 10 -253,253,253,253,253,253,253,253,253,253,253,253,253,253,253,253, # 20 -252,252,252,252,252,252,252,252,252,252,253,253,253,253,253,253, # 30 -253,142,143,144,145,146,147,148,149,150,151,152, 74,153, 75,154, # 40 -155,156,157,158,159,160,161,162,163,164,165,253,253,253,253,253, # 50 -253, 71,172, 66,173, 65,174, 76,175, 64,176,177, 77, 72,178, 69, # 60 - 67,179, 78, 73,180,181, 79,182,183,184,185,253,253,253,253,253, # 70 -191,192,193,194,195,196,197,198,199,200,201,202,203,204,205,206, -207,208,209,210,211,212,213,214,215,216,217,218,219,220,221,222, -223,224,225,226,227,228,229,230,231,232,233,234,235,236,237,238, -239,240,241,242,243,244,245,246, 68,247,248,249,250,251,252,253, - 37, 44, 33, 46, 41, 48, 56, 51, 42, 60, 36, 49, 38, 31, 34, 35, - 45, 32, 40, 52, 53, 55, 58, 50, 57, 63, 70, 62, 61, 47, 59, 43, - 3, 21, 10, 19, 13, 2, 24, 20, 4, 23, 11, 8, 12, 5, 1, 15, - 9, 7, 6, 14, 39, 26, 28, 22, 25, 29, 54, 18, 17, 30, 27, 16, -) - -latin5_char_to_order_map = ( -255,255,255,255,255,255,255,255,255,255,254,255,255,254,255,255, # 00 -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 10 -253,253,253,253,253,253,253,253,253,253,253,253,253,253,253,253, # 20 -252,252,252,252,252,252,252,252,252,252,253,253,253,253,253,253, # 30 -253,142,143,144,145,146,147,148,149,150,151,152, 74,153, 75,154, # 40 -155,156,157,158,159,160,161,162,163,164,165,253,253,253,253,253, # 50 -253, 71,172, 66,173, 65,174, 76,175, 64,176,177, 77, 72,178, 69, # 60 - 67,179, 78, 73,180,181, 79,182,183,184,185,253,253,253,253,253, # 70 -191,192,193,194,195,196,197,198,199,200,201,202,203,204,205,206, -207,208,209,210,211,212,213,214,215,216,217,218,219,220,221,222, -223,224,225,226,227,228,229,230,231,232,233,234,235,236,237,238, - 37, 44, 33, 46, 41, 48, 56, 51, 42, 60, 36, 49, 38, 31, 34, 35, - 45, 32, 40, 52, 53, 55, 58, 50, 57, 63, 70, 62, 61, 47, 59, 43, - 3, 21, 10, 19, 13, 2, 24, 20, 4, 23, 11, 8, 12, 5, 1, 15, - 9, 7, 6, 14, 39, 26, 28, 22, 25, 29, 54, 18, 17, 30, 27, 16, -239, 68,240,241,242,243,244,245,246,247,248,249,250,251,252,255, -) - -macCyrillic_char_to_order_map = ( -255,255,255,255,255,255,255,255,255,255,254,255,255,254,255,255, # 00 -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 10 -253,253,253,253,253,253,253,253,253,253,253,253,253,253,253,253, # 20 -252,252,252,252,252,252,252,252,252,252,253,253,253,253,253,253, # 30 -253,142,143,144,145,146,147,148,149,150,151,152, 74,153, 75,154, # 40 -155,156,157,158,159,160,161,162,163,164,165,253,253,253,253,253, # 50 -253, 71,172, 66,173, 65,174, 76,175, 64,176,177, 77, 72,178, 69, # 60 - 67,179, 78, 73,180,181, 79,182,183,184,185,253,253,253,253,253, # 70 - 37, 44, 33, 46, 41, 48, 56, 51, 42, 60, 36, 49, 38, 31, 34, 35, - 45, 32, 40, 52, 53, 55, 58, 50, 57, 63, 70, 62, 61, 47, 59, 43, -191,192,193,194,195,196,197,198,199,200,201,202,203,204,205,206, -207,208,209,210,211,212,213,214,215,216,217,218,219,220,221,222, -223,224,225,226,227,228,229,230,231,232,233,234,235,236,237,238, -239,240,241,242,243,244,245,246,247,248,249,250,251,252, 68, 16, - 3, 21, 10, 19, 13, 2, 24, 20, 4, 23, 11, 8, 12, 5, 1, 15, - 9, 7, 6, 14, 39, 26, 28, 22, 25, 29, 54, 18, 17, 30, 27,255, -) - -IBM855_char_to_order_map = ( -255,255,255,255,255,255,255,255,255,255,254,255,255,254,255,255, # 00 -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 10 -253,253,253,253,253,253,253,253,253,253,253,253,253,253,253,253, # 20 -252,252,252,252,252,252,252,252,252,252,253,253,253,253,253,253, # 30 -253,142,143,144,145,146,147,148,149,150,151,152, 74,153, 75,154, # 40 -155,156,157,158,159,160,161,162,163,164,165,253,253,253,253,253, # 50 -253, 71,172, 66,173, 65,174, 76,175, 64,176,177, 77, 72,178, 69, # 60 - 67,179, 78, 73,180,181, 79,182,183,184,185,253,253,253,253,253, # 70 -191,192,193,194, 68,195,196,197,198,199,200,201,202,203,204,205, -206,207,208,209,210,211,212,213,214,215,216,217, 27, 59, 54, 70, - 3, 37, 21, 44, 28, 58, 13, 41, 2, 48, 39, 53, 19, 46,218,219, -220,221,222,223,224, 26, 55, 4, 42,225,226,227,228, 23, 60,229, -230,231,232,233,234,235, 11, 36,236,237,238,239,240,241,242,243, - 8, 49, 12, 38, 5, 31, 1, 34, 15,244,245,246,247, 35, 16,248, - 43, 9, 45, 7, 32, 6, 40, 14, 52, 24, 56, 10, 33, 17, 61,249, -250, 18, 62, 20, 51, 25, 57, 30, 47, 29, 63, 22, 50,251,252,255, -) - -IBM866_char_to_order_map = ( -255,255,255,255,255,255,255,255,255,255,254,255,255,254,255,255, # 00 -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 10 -253,253,253,253,253,253,253,253,253,253,253,253,253,253,253,253, # 20 -252,252,252,252,252,252,252,252,252,252,253,253,253,253,253,253, # 30 -253,142,143,144,145,146,147,148,149,150,151,152, 74,153, 75,154, # 40 -155,156,157,158,159,160,161,162,163,164,165,253,253,253,253,253, # 50 -253, 71,172, 66,173, 65,174, 76,175, 64,176,177, 77, 72,178, 69, # 60 - 67,179, 78, 73,180,181, 79,182,183,184,185,253,253,253,253,253, # 70 - 37, 44, 33, 46, 41, 48, 56, 51, 42, 60, 36, 49, 38, 31, 34, 35, - 45, 32, 40, 52, 53, 55, 58, 50, 57, 63, 70, 62, 61, 47, 59, 43, - 3, 21, 10, 19, 13, 2, 24, 20, 4, 23, 11, 8, 12, 5, 1, 15, -191,192,193,194,195,196,197,198,199,200,201,202,203,204,205,206, -207,208,209,210,211,212,213,214,215,216,217,218,219,220,221,222, -223,224,225,226,227,228,229,230,231,232,233,234,235,236,237,238, - 9, 7, 6, 14, 39, 26, 28, 22, 25, 29, 54, 18, 17, 30, 27, 16, -239, 68,240,241,242,243,244,245,246,247,248,249,250,251,252,255, -) - -# Model Table: -# total sequences: 100% -# first 512 sequences: 97.6601% -# first 1024 sequences: 2.3389% -# rest sequences: 0.1237% -# negative sequences: 0.0009% -RussianLangModel = ( -0,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,1,1,3,3,3,3,1,3,3,3,2,3,2,3,3, -3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,0,3,2,2,2,2,2,0,0,2, -3,3,3,2,3,3,3,3,3,3,3,3,3,3,2,3,3,0,0,3,3,3,3,3,3,3,3,3,2,3,2,0, -0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,2,2,3,3,3,3,3,3,3,3,3,2,3,3,0,0,3,3,3,3,3,3,3,3,2,3,3,1,0, -0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,2,3,2,3,3,3,3,3,3,3,3,3,3,3,3,3,0,0,3,3,3,3,3,3,3,3,3,3,3,2,1, -0,0,0,0,0,0,0,2,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,3,3,3,3,3,3,3,3,3,2,3,3,0,0,3,3,3,3,3,3,3,3,3,3,3,2,1, -0,0,0,0,0,1,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,3,3,3,2,2,2,3,1,3,3,1,3,3,3,3,2,2,3,0,2,2,2,3,3,2,1,0, -0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,3,2,3,3,3,3,3,2,2,3,2,3,3,3,2,1,2,2,0,1,2,2,2,2,2,2,0, -0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,2,2,2,3,0,2,2,3,3,2,1,2,0, -0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,1,0,0,2,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,3,2,3,3,1,2,3,2,2,3,2,3,3,3,3,2,2,3,0,3,2,2,3,1,1,1,0, -0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,3,3,3,2,2,3,3,3,3,3,2,3,3,3,3,2,2,2,0,3,3,3,2,2,2,2,0, -0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,3,3,3,3,3,2,3,2,3,3,3,3,3,3,2,3,2,2,0,1,3,2,1,2,2,1,0, -0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,3,3,3,3,3,3,2,1,1,3,0,1,1,1,1,2,1,1,0,2,2,2,1,2,0,1,0, -0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,3,2,3,3,2,2,2,2,1,3,2,3,2,3,2,1,2,2,0,1,1,2,1,2,1,2,0, -0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,3,3,3,3,3,3,3,2,2,3,2,3,3,3,2,2,2,2,0,2,2,2,2,3,1,1,0, -0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0, -3,2,3,2,2,3,3,3,3,3,3,3,3,3,1,3,2,0,0,3,3,3,3,2,3,3,3,3,2,3,2,0, -0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,3,3,3,3,3,2,2,3,3,0,2,1,0,3,2,3,2,3,0,0,1,2,0,0,1,0,1,2,1,1,0, -0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,0,3,0,2,3,3,3,3,2,3,3,3,3,1,2,2,0,0,2,3,2,2,2,3,2,3,2,2,3,0,0, -0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,2,3,0,2,3,2,3,0,1,2,3,3,2,0,2,3,0,0,2,3,2,2,0,1,3,1,3,2,2,1,0, -0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,1,3,0,2,3,3,3,3,3,3,3,3,2,1,3,2,0,0,2,2,3,3,3,2,3,3,0,2,2,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,3,2,2,3,3,2,2,2,3,3,0,0,1,1,1,1,1,2,0,0,1,1,1,1,0,1,0, -0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,3,2,2,3,3,3,3,3,3,3,0,3,2,3,3,2,3,2,0,2,1,0,1,1,0,1,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,3,2,3,3,3,2,2,2,2,3,1,3,2,3,1,1,2,1,0,2,2,2,2,1,3,1,0, -0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0, -2,2,3,3,3,3,3,1,2,2,1,3,1,0,3,0,0,3,0,0,0,1,1,0,1,2,1,0,0,0,0,0, -0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,2,2,1,1,3,3,3,2,2,1,2,2,3,1,1,2,0,0,2,2,1,3,0,0,2,1,1,2,1,1,0, -0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,2,3,3,3,3,1,2,2,2,1,2,1,3,3,1,1,2,1,2,1,2,2,0,2,0,0,1,1,0,1,0, -0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,3,3,3,3,3,2,1,3,2,2,3,2,0,3,2,0,3,0,1,0,1,1,0,0,1,1,1,1,0,1,0, -0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,2,3,3,3,2,2,2,3,3,1,2,1,2,1,0,1,0,1,1,0,1,0,0,2,1,1,1,0,1,0, -0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0, -3,1,1,2,1,2,3,3,2,2,1,2,2,3,0,2,1,0,0,2,2,3,2,1,2,2,2,2,2,3,1,0, -0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,1,1,0,1,1,2,2,1,1,3,0,0,1,3,1,1,1,0,0,0,1,0,1,1,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,1,3,3,3,2,0,0,0,2,1,0,1,0,2,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,0,1,0,0,2,3,2,2,2,1,2,2,2,1,2,1,0,0,1,1,1,0,2,0,1,1,1,0,0,1,1, -1,0,0,0,0,0,1,2,0,0,0,0,0,1,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0, -2,3,3,3,3,0,0,0,0,1,0,0,0,0,3,0,1,2,1,0,0,0,0,0,0,0,1,1,0,0,1,1, -1,0,1,0,1,2,0,0,1,1,2,1,0,1,1,1,1,0,1,1,1,1,0,1,0,0,1,0,0,1,1,0, -2,2,3,2,2,2,3,1,2,2,2,2,2,2,2,2,1,1,1,1,1,1,1,0,1,0,1,1,1,0,2,1, -1,1,1,1,1,1,1,1,2,1,1,1,1,1,1,1,1,1,1,0,1,0,1,1,0,1,1,1,0,1,1,0, -3,3,3,2,2,2,2,3,2,2,1,1,2,2,2,2,1,1,3,1,2,1,2,0,0,1,1,0,1,0,2,1, -1,1,1,1,1,2,1,0,1,1,1,1,0,1,0,0,1,1,0,0,1,0,1,0,0,1,0,0,0,1,1,0, -2,0,0,1,0,3,2,2,2,2,1,2,1,2,1,2,0,0,0,2,1,2,2,1,1,2,2,0,1,1,0,2, -1,1,1,1,1,0,1,1,1,2,1,1,1,2,1,0,1,2,1,1,1,1,0,1,1,1,0,0,1,0,0,1, -1,3,2,2,2,1,1,1,2,3,0,0,0,0,2,0,2,2,1,0,0,0,0,0,0,1,0,0,0,0,1,1, -1,0,1,1,0,1,0,1,1,0,1,1,0,2,0,0,1,1,0,0,1,0,0,0,0,0,0,0,0,1,1,0, -2,3,2,3,2,1,2,2,2,2,1,0,0,0,2,0,0,1,1,0,0,0,0,0,0,0,1,1,0,0,2,1, -1,1,2,1,0,2,0,0,1,0,1,0,0,1,0,0,1,1,0,1,1,0,0,0,0,0,1,0,0,0,0,0, -3,0,0,1,0,2,2,2,3,2,2,2,2,2,2,2,0,0,0,2,1,2,1,1,1,2,2,0,0,0,1,2, -1,1,1,1,1,0,1,2,1,1,1,1,1,1,1,0,1,1,1,1,1,1,0,1,1,1,1,1,1,0,0,1, -2,3,2,3,3,2,0,1,1,1,0,0,1,0,2,0,1,1,3,1,0,0,0,0,0,0,0,1,0,0,2,1, -1,1,1,1,1,1,1,0,1,0,1,1,1,1,0,1,1,1,0,0,1,1,0,1,0,0,0,0,0,0,1,0, -2,3,3,3,3,1,2,2,2,2,0,1,1,0,2,1,1,1,2,1,0,1,1,0,0,1,0,1,0,0,2,0, -0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,3,3,3,2,0,0,1,1,2,2,1,0,0,2,0,1,1,3,0,0,1,0,0,0,0,0,1,0,1,2,1, -1,1,2,0,1,1,1,0,1,0,1,1,0,1,0,1,1,1,1,0,1,0,0,0,0,0,0,1,0,1,1,0, -1,3,2,3,2,1,0,0,2,2,2,0,1,0,2,0,1,1,1,0,1,0,0,0,3,0,1,1,0,0,2,1, -1,1,1,0,1,1,0,0,0,0,1,1,0,1,0,0,2,1,1,0,1,0,0,0,1,0,1,0,0,1,1,0, -3,1,2,1,1,2,2,2,2,2,2,1,2,2,1,1,0,0,0,2,2,2,0,0,0,1,2,1,0,1,0,1, -2,1,1,1,1,1,1,1,1,1,1,1,1,1,1,0,2,1,1,1,0,1,0,1,1,0,1,1,1,0,0,1, -3,0,0,0,0,2,0,1,1,1,1,1,1,1,0,1,0,0,0,1,1,1,0,1,0,1,1,0,0,1,0,1, -1,1,0,0,1,0,0,0,1,0,1,1,0,0,1,0,1,0,1,0,0,0,0,1,0,0,0,1,0,0,0,1, -1,3,3,2,2,0,0,0,2,2,0,0,0,1,2,0,1,1,2,0,0,0,0,0,0,0,0,1,0,0,2,1, -0,1,1,0,0,1,1,0,0,0,1,1,0,1,1,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,1,0, -2,3,2,3,2,0,0,0,0,1,1,0,0,0,2,0,2,0,2,0,0,0,0,0,1,0,0,1,0,0,1,1, -1,1,2,0,1,2,1,0,1,1,2,1,1,1,1,1,2,1,1,0,1,0,0,1,1,1,1,1,0,1,1,0, -1,3,2,2,2,1,0,0,2,2,1,0,1,2,2,0,0,1,0,0,0,0,0,0,0,0,0,1,0,0,1,1, -0,0,1,1,0,1,1,0,0,1,1,0,1,1,0,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0, -1,0,0,1,0,2,3,1,2,2,2,2,2,2,1,1,0,0,0,1,0,1,0,2,1,1,1,0,0,0,0,1, -1,1,0,1,1,0,1,1,1,1,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1,0,0,0, -2,0,2,0,0,1,0,3,2,1,2,1,2,2,0,1,0,0,0,2,1,0,0,2,1,1,1,1,0,2,0,2, -2,1,1,1,1,1,1,1,1,1,1,1,1,2,1,0,1,1,1,1,0,0,0,1,1,1,1,0,1,0,0,1, -1,2,2,2,2,1,0,0,1,0,0,0,0,0,2,0,1,1,1,1,0,0,0,0,1,0,1,2,0,0,2,0, -1,0,1,1,1,2,1,0,1,0,1,1,0,0,1,0,1,1,1,0,1,0,0,0,1,0,0,1,0,1,1,0, -2,1,2,2,2,0,3,0,1,1,0,0,0,0,2,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1, -0,0,0,1,1,1,0,0,1,0,1,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0, -1,2,2,3,2,2,0,0,1,1,2,0,1,2,1,0,1,0,1,0,0,1,0,0,0,0,0,0,0,0,0,1, -0,1,1,0,0,1,1,0,0,1,1,0,0,1,1,0,1,1,0,0,1,0,0,0,0,0,0,0,0,1,1,0, -2,2,1,1,2,1,2,2,2,2,2,1,2,2,0,1,0,0,0,1,2,2,2,1,2,1,1,1,1,1,2,1, -1,1,1,1,1,1,1,1,1,1,0,0,1,1,1,0,1,1,1,0,0,0,0,1,1,1,0,1,1,0,0,1, -1,2,2,2,2,0,1,0,2,2,0,0,0,0,2,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,2,0, -0,0,1,0,0,1,0,0,0,0,1,0,1,1,0,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0, -0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -1,2,2,2,2,0,0,0,2,2,2,0,1,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1, -0,1,1,0,0,1,1,0,0,0,1,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -1,2,2,2,2,0,0,0,0,1,0,0,1,1,2,0,0,0,0,1,0,1,0,0,1,0,0,2,0,0,0,1, -0,0,1,0,0,1,0,0,0,1,1,0,0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0, -1,2,2,2,1,1,2,0,2,1,1,1,1,0,2,2,0,0,0,0,0,0,0,0,0,1,1,0,0,0,1,1, -0,0,1,0,1,1,0,0,0,0,1,0,0,0,0,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0, -1,0,2,1,2,0,0,0,0,0,1,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,0, -0,0,1,0,1,1,0,0,0,0,1,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0, -1,0,0,0,0,2,0,1,2,1,0,1,1,1,0,1,0,0,0,1,0,1,0,0,1,0,1,0,0,0,0,1, -0,0,0,0,0,1,0,0,1,1,0,0,1,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1, -2,2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1, -1,0,0,0,1,0,0,0,1,1,0,0,0,0,0,0,0,1,0,0,0,0,0,1,0,0,1,0,0,0,0,0, -2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1, -1,1,1,0,1,0,1,0,0,1,1,1,1,0,0,0,1,0,0,0,0,1,0,0,0,1,0,1,0,0,0,0, -1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1, -1,1,0,1,1,0,1,0,1,0,0,0,0,1,1,0,1,1,0,0,0,0,0,1,0,1,1,0,1,0,0,0, -0,1,1,1,1,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0, -) - -Koi8rModel = { - 'char_to_order_map': KOI8R_char_to_order_map, - 'precedence_matrix': RussianLangModel, - 'typical_positive_ratio': 0.976601, - 'keep_english_letter': False, - 'charset_name': "KOI8-R", - 'language': 'Russian', -} - -Win1251CyrillicModel = { - 'char_to_order_map': win1251_char_to_order_map, - 'precedence_matrix': RussianLangModel, - 'typical_positive_ratio': 0.976601, - 'keep_english_letter': False, - 'charset_name': "windows-1251", - 'language': 'Russian', -} - -Latin5CyrillicModel = { - 'char_to_order_map': latin5_char_to_order_map, - 'precedence_matrix': RussianLangModel, - 'typical_positive_ratio': 0.976601, - 'keep_english_letter': False, - 'charset_name': "ISO-8859-5", - 'language': 'Russian', -} - -MacCyrillicModel = { - 'char_to_order_map': macCyrillic_char_to_order_map, - 'precedence_matrix': RussianLangModel, - 'typical_positive_ratio': 0.976601, - 'keep_english_letter': False, - 'charset_name': "MacCyrillic", - 'language': 'Russian', -} - -Ibm866Model = { - 'char_to_order_map': IBM866_char_to_order_map, - 'precedence_matrix': RussianLangModel, - 'typical_positive_ratio': 0.976601, - 'keep_english_letter': False, - 'charset_name': "IBM866", - 'language': 'Russian', -} - -Ibm855Model = { - 'char_to_order_map': IBM855_char_to_order_map, - 'precedence_matrix': RussianLangModel, - 'typical_positive_ratio': 0.976601, - 'keep_english_letter': False, - 'charset_name': "IBM855", - 'language': 'Russian', -} diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/langgreekmodel.py b/src/fetchcode/vcs/pip/_vendor/chardet/langgreekmodel.py deleted file mode 100644 index 53322216..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/langgreekmodel.py +++ /dev/null @@ -1,225 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Communicator client code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -# 255: Control characters that usually does not exist in any text -# 254: Carriage/Return -# 253: symbol (punctuation) that does not belong to word -# 252: 0 - 9 - -# Character Mapping Table: -Latin7_char_to_order_map = ( -255,255,255,255,255,255,255,255,255,255,254,255,255,254,255,255, # 00 -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 10 -253,253,253,253,253,253,253,253,253,253,253,253,253,253,253,253, # 20 -252,252,252,252,252,252,252,252,252,252,253,253,253,253,253,253, # 30 -253, 82,100,104, 94, 98,101,116,102,111,187,117, 92, 88,113, 85, # 40 - 79,118,105, 83, 67,114,119, 95, 99,109,188,253,253,253,253,253, # 50 -253, 72, 70, 80, 81, 60, 96, 93, 89, 68,120, 97, 77, 86, 69, 55, # 60 - 78,115, 65, 66, 58, 76,106,103, 87,107,112,253,253,253,253,253, # 70 -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 80 -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 90 -253,233, 90,253,253,253,253,253,253,253,253,253,253, 74,253,253, # a0 -253,253,253,253,247,248, 61, 36, 46, 71, 73,253, 54,253,108,123, # b0 -110, 31, 51, 43, 41, 34, 91, 40, 52, 47, 44, 53, 38, 49, 59, 39, # c0 - 35, 48,250, 37, 33, 45, 56, 50, 84, 57,120,121, 17, 18, 22, 15, # d0 -124, 1, 29, 20, 21, 3, 32, 13, 25, 5, 11, 16, 10, 6, 30, 4, # e0 - 9, 8, 14, 7, 2, 12, 28, 23, 42, 24, 64, 75, 19, 26, 27,253, # f0 -) - -win1253_char_to_order_map = ( -255,255,255,255,255,255,255,255,255,255,254,255,255,254,255,255, # 00 -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 10 -253,253,253,253,253,253,253,253,253,253,253,253,253,253,253,253, # 20 -252,252,252,252,252,252,252,252,252,252,253,253,253,253,253,253, # 30 -253, 82,100,104, 94, 98,101,116,102,111,187,117, 92, 88,113, 85, # 40 - 79,118,105, 83, 67,114,119, 95, 99,109,188,253,253,253,253,253, # 50 -253, 72, 70, 80, 81, 60, 96, 93, 89, 68,120, 97, 77, 86, 69, 55, # 60 - 78,115, 65, 66, 58, 76,106,103, 87,107,112,253,253,253,253,253, # 70 -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 80 -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 90 -253,233, 61,253,253,253,253,253,253,253,253,253,253, 74,253,253, # a0 -253,253,253,253,247,253,253, 36, 46, 71, 73,253, 54,253,108,123, # b0 -110, 31, 51, 43, 41, 34, 91, 40, 52, 47, 44, 53, 38, 49, 59, 39, # c0 - 35, 48,250, 37, 33, 45, 56, 50, 84, 57,120,121, 17, 18, 22, 15, # d0 -124, 1, 29, 20, 21, 3, 32, 13, 25, 5, 11, 16, 10, 6, 30, 4, # e0 - 9, 8, 14, 7, 2, 12, 28, 23, 42, 24, 64, 75, 19, 26, 27,253, # f0 -) - -# Model Table: -# total sequences: 100% -# first 512 sequences: 98.2851% -# first 1024 sequences:1.7001% -# rest sequences: 0.0359% -# negative sequences: 0.0148% -GreekLangModel = ( -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,3,2,2,3,3,3,3,3,3,3,3,1,3,3,3,0,2,2,3,3,0,3,0,3,2,0,3,3,3,0, -3,0,0,0,2,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,3,3,3,3,3,0,3,3,0,3,2,3,3,0,3,2,3,3,3,0,0,3,0,3,0,3,3,2,0,0,0, -2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0, -0,2,3,2,2,3,3,3,3,3,3,3,3,0,3,3,3,3,0,2,3,3,0,3,3,3,3,2,3,3,3,0, -2,0,0,0,2,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,2,3,3,2,3,3,3,3,3,3,3,3,3,3,3,3,0,2,1,3,3,3,3,2,3,3,2,3,3,2,0, -0,0,0,0,2,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,3,3,3,3,0,3,3,3,3,3,3,0,3,3,0,3,3,3,3,3,3,3,3,3,3,0,3,2,3,3,0, -2,0,1,0,2,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0, -0,3,3,3,3,3,2,3,0,0,0,0,3,3,0,3,1,3,3,3,0,3,3,0,3,3,3,3,0,0,0,0, -2,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,3,3,3,3,3,0,3,0,3,3,3,3,3,0,3,2,2,2,3,0,2,3,3,3,3,3,2,3,3,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,3,3,3,3,3,3,2,2,2,3,3,3,3,0,3,1,3,3,3,3,2,3,3,3,3,3,3,3,2,2,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,3,3,3,3,3,2,0,3,0,0,0,3,3,2,3,3,3,3,3,0,0,3,2,3,0,2,3,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,3,0,3,3,3,3,0,0,3,3,0,2,3,0,3,0,3,3,3,0,0,3,0,3,0,2,2,3,3,0,0, -0,0,1,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,3,3,3,3,3,2,0,3,2,3,3,3,3,0,3,3,3,3,3,0,3,3,2,3,2,3,3,2,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,3,3,2,3,2,3,3,3,3,3,3,0,2,3,2,3,2,2,2,3,2,3,3,2,3,0,2,2,2,3,0, -2,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,3,0,0,0,3,3,3,2,3,3,0,0,3,0,3,0,0,0,3,2,0,3,0,3,0,0,2,0,2,0, -0,0,0,0,2,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,3,3,3,3,0,3,3,3,3,3,3,0,3,3,0,3,0,0,0,3,3,0,3,3,3,0,0,1,2,3,0, -3,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,3,3,3,3,3,2,0,0,3,2,2,3,3,0,3,3,3,3,3,2,1,3,0,3,2,3,3,2,1,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,3,3,0,2,3,3,3,3,3,3,0,0,3,0,3,0,0,0,3,3,0,3,2,3,0,0,3,3,3,0, -3,0,0,0,2,0,0,0,0,0,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,3,3,3,3,0,3,3,3,3,3,3,0,0,3,0,3,0,0,0,3,2,0,3,2,3,0,0,3,2,3,0, -2,0,0,0,0,0,0,0,0,0,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,3,1,2,2,3,3,3,3,3,3,0,2,3,0,3,0,0,0,3,3,0,3,0,2,0,0,2,3,1,0, -2,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,3,0,3,3,3,3,0,3,0,3,3,2,3,0,3,3,3,3,3,3,0,3,3,3,0,2,3,0,0,3,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,3,0,3,3,3,0,0,3,0,0,0,3,3,0,3,0,2,3,3,0,0,3,0,3,0,3,3,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,3,0,0,0,3,3,3,3,3,3,0,0,3,0,2,0,0,0,3,3,0,3,0,3,0,0,2,0,2,0, -0,0,0,0,1,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,3,3,3,3,3,3,0,3,0,2,0,3,2,0,3,2,3,2,3,0,0,3,2,3,2,3,3,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,3,0,0,2,3,3,3,3,3,0,0,0,3,0,2,1,0,0,3,2,2,2,0,3,0,0,2,2,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,3,0,3,3,3,2,0,3,0,3,0,3,3,0,2,1,2,3,3,0,0,3,0,3,0,3,3,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,2,3,3,3,0,3,3,3,3,3,3,0,2,3,0,3,0,0,0,2,1,0,2,2,3,0,0,2,2,2,0, -0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,3,0,0,2,3,3,3,2,3,0,0,1,3,0,2,0,0,0,0,3,0,1,0,2,0,0,1,1,1,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,3,3,3,3,3,1,0,3,0,0,0,3,2,0,3,2,3,3,3,0,0,3,0,3,2,2,2,1,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,3,0,3,3,3,0,0,3,0,0,0,0,2,0,2,3,3,2,2,2,2,3,0,2,0,2,2,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,3,3,3,3,2,0,0,0,0,0,0,2,3,0,2,0,2,3,2,0,0,3,0,3,0,3,1,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,3,2,3,3,2,2,3,0,2,0,3,0,0,0,2,0,0,0,0,1,2,0,2,0,2,0, -0,2,0,2,0,2,2,0,0,1,0,2,2,2,0,2,2,2,0,2,2,2,0,0,2,0,0,1,0,0,0,0, -0,2,0,3,3,2,0,0,0,0,0,0,1,3,0,2,0,2,2,2,0,0,2,0,3,0,0,2,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,3,0,2,3,2,0,2,2,0,2,0,2,2,0,2,0,2,2,2,0,0,0,0,0,0,2,3,0,0,0,2, -0,1,2,0,0,0,0,2,2,0,0,0,2,1,0,2,2,0,0,0,0,0,0,1,0,2,0,0,0,0,0,0, -0,0,2,1,0,2,3,2,2,3,2,3,2,0,0,3,3,3,0,0,3,2,0,0,0,1,1,0,2,0,2,2, -0,2,0,2,0,2,2,0,0,2,0,2,2,2,0,2,2,2,2,0,0,2,0,0,0,2,0,1,0,0,0,0, -0,3,0,3,3,2,2,0,3,0,0,0,2,2,0,2,2,2,1,2,0,0,1,2,2,0,0,3,0,0,0,2, -0,1,2,0,0,0,1,2,0,0,0,0,0,0,0,2,2,0,1,0,0,2,0,0,0,2,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,2,3,3,2,2,0,0,0,2,0,2,3,3,0,2,0,0,0,0,0,0,2,2,2,0,2,2,0,2,0,2, -0,2,2,0,0,2,2,2,2,1,0,0,2,2,0,2,0,0,2,0,0,0,0,0,0,2,0,0,0,0,0,0, -0,2,0,3,2,3,0,0,0,3,0,0,2,2,0,2,0,2,2,2,0,0,2,0,0,0,0,0,0,0,0,2, -0,0,2,2,0,0,2,2,2,0,0,0,0,0,0,2,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,2,0,0,3,2,0,2,2,2,2,2,0,0,0,2,0,0,0,0,2,0,1,0,0,2,0,1,0,0,0, -0,2,2,2,0,2,2,0,1,2,0,2,2,2,0,2,2,2,2,1,2,2,0,0,2,0,0,0,0,0,0,0, -0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0, -0,2,0,2,0,2,2,0,0,0,0,1,2,1,0,0,2,2,0,0,2,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,3,2,3,0,0,2,0,0,0,2,2,0,2,0,0,0,1,0,0,2,0,2,0,2,2,0,0,0,0, -0,0,2,0,0,0,0,2,2,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0, -0,2,2,3,2,2,0,0,0,0,0,0,1,3,0,2,0,2,2,0,0,0,1,0,2,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,2,0,2,0,3,2,0,2,0,0,0,0,0,0,2,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1, -0,0,2,0,0,0,0,1,1,0,0,2,1,2,0,2,2,0,1,0,0,1,0,0,0,2,0,0,0,0,0,0, -0,3,0,2,2,2,0,0,2,0,0,0,2,0,0,0,2,3,0,2,0,0,0,0,0,0,2,2,0,0,0,2, -0,1,2,0,0,0,1,2,2,1,0,0,0,2,0,0,2,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,3,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,2,1,2,0,2,2,0,2,0,0,2,0,0,0,0,1,2,1,0,2,1,0,0,0,0,0,0,0,0,0,0, -0,0,2,0,0,0,3,1,2,2,0,2,0,0,0,0,2,0,0,0,2,0,0,3,0,0,0,0,2,2,2,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,2,1,0,2,0,1,2,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,1,0,0,0,0,0,0,2, -0,2,2,0,0,2,2,2,2,2,0,1,2,0,0,0,2,2,0,1,0,2,0,0,2,2,0,0,0,0,0,0, -0,0,0,0,1,0,0,0,0,0,0,0,3,0,0,2,0,0,0,0,0,0,0,0,2,0,2,0,0,0,0,2, -0,1,2,0,0,0,0,2,2,1,0,1,0,1,0,2,2,2,1,0,0,0,0,0,0,1,0,0,0,0,0,0, -0,2,0,1,2,0,0,0,0,0,0,0,0,0,0,2,0,0,2,2,0,0,0,0,1,0,0,0,0,0,0,2, -0,2,2,0,0,0,0,2,2,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,2,0,0,2,0,0,0, -0,2,2,2,2,0,0,0,3,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,2,0,0,0,0,0,0,1, -0,0,2,0,0,0,0,1,2,0,0,0,0,0,0,2,2,1,1,0,0,0,0,0,0,1,0,0,0,0,0,0, -0,2,0,2,2,2,0,0,2,0,0,0,0,0,0,0,2,2,2,0,0,0,2,0,0,0,0,0,0,0,0,2, -0,0,1,0,0,0,0,2,1,0,0,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0, -0,3,0,2,0,0,0,0,0,0,0,0,2,0,0,0,0,0,2,0,0,0,0,0,0,0,2,0,0,0,0,2, -0,0,2,0,0,0,0,2,2,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,2,0,2,2,1,0,0,0,0,0,0,2,0,0,2,0,2,2,2,0,0,0,0,0,0,2,0,0,0,0,2, -0,0,2,0,0,2,0,2,2,0,0,0,0,2,0,2,0,0,0,0,0,2,0,0,0,2,0,0,0,0,0,0, -0,0,3,0,0,0,2,2,0,2,2,0,0,0,0,0,2,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,2,0,0,0,0,0, -0,2,2,2,2,2,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,1, -0,0,0,0,0,0,0,2,1,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,2,2,0,0,0,0,0,2,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0, -0,2,0,0,0,2,0,0,0,0,0,1,0,0,0,0,2,2,0,0,0,1,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,2,0,0,0, -0,2,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,1,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,2,0,2,0,0,0, -0,0,0,0,0,0,0,0,2,1,0,0,0,0,0,0,2,0,0,0,1,2,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -) - -Latin7GreekModel = { - 'char_to_order_map': Latin7_char_to_order_map, - 'precedence_matrix': GreekLangModel, - 'typical_positive_ratio': 0.982851, - 'keep_english_letter': False, - 'charset_name': "ISO-8859-7", - 'language': 'Greek', -} - -Win1253GreekModel = { - 'char_to_order_map': win1253_char_to_order_map, - 'precedence_matrix': GreekLangModel, - 'typical_positive_ratio': 0.982851, - 'keep_english_letter': False, - 'charset_name': "windows-1253", - 'language': 'Greek', -} diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/langhebrewmodel.py b/src/fetchcode/vcs/pip/_vendor/chardet/langhebrewmodel.py deleted file mode 100644 index 58f4c875..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/langhebrewmodel.py +++ /dev/null @@ -1,200 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Universal charset detector code. -# -# The Initial Developer of the Original Code is -# Simon Montagu -# Portions created by the Initial Developer are Copyright (C) 2005 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# Shy Shalom - original C code -# Shoshannah Forbes - original C code (?) -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -# 255: Control characters that usually does not exist in any text -# 254: Carriage/Return -# 253: symbol (punctuation) that does not belong to word -# 252: 0 - 9 - -# Windows-1255 language model -# Character Mapping Table: -WIN1255_CHAR_TO_ORDER_MAP = ( -255,255,255,255,255,255,255,255,255,255,254,255,255,254,255,255, # 00 -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 10 -253,253,253,253,253,253,253,253,253,253,253,253,253,253,253,253, # 20 -252,252,252,252,252,252,252,252,252,252,253,253,253,253,253,253, # 30 -253, 69, 91, 79, 80, 92, 89, 97, 90, 68,111,112, 82, 73, 95, 85, # 40 - 78,121, 86, 71, 67,102,107, 84,114,103,115,253,253,253,253,253, # 50 -253, 50, 74, 60, 61, 42, 76, 70, 64, 53,105, 93, 56, 65, 54, 49, # 60 - 66,110, 51, 43, 44, 63, 81, 77, 98, 75,108,253,253,253,253,253, # 70 -124,202,203,204,205, 40, 58,206,207,208,209,210,211,212,213,214, -215, 83, 52, 47, 46, 72, 32, 94,216,113,217,109,218,219,220,221, - 34,116,222,118,100,223,224,117,119,104,125,225,226, 87, 99,227, -106,122,123,228, 55,229,230,101,231,232,120,233, 48, 39, 57,234, - 30, 59, 41, 88, 33, 37, 36, 31, 29, 35,235, 62, 28,236,126,237, -238, 38, 45,239,240,241,242,243,127,244,245,246,247,248,249,250, - 9, 8, 20, 16, 3, 2, 24, 14, 22, 1, 25, 15, 4, 11, 6, 23, - 12, 19, 13, 26, 18, 27, 21, 17, 7, 10, 5,251,252,128, 96,253, -) - -# Model Table: -# total sequences: 100% -# first 512 sequences: 98.4004% -# first 1024 sequences: 1.5981% -# rest sequences: 0.087% -# negative sequences: 0.0015% -HEBREW_LANG_MODEL = ( -0,3,3,3,3,3,3,3,3,3,3,2,3,3,3,3,3,3,3,3,3,3,3,2,3,2,1,2,0,1,0,0, -3,0,3,1,0,0,1,3,2,0,1,1,2,0,2,2,2,1,1,1,1,2,1,1,1,2,0,0,2,2,0,1, -3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,2,2,2,2, -1,2,1,2,1,2,0,0,2,0,0,0,0,0,1,0,1,0,0,0,0,0,0,1,0,0,0,0,0,0,1,0, -3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,2,2,2, -1,2,1,3,1,1,0,0,2,0,0,0,1,0,1,0,1,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0, -3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,1,0,1,2,2,1,3, -1,2,1,1,2,2,0,0,2,2,0,0,0,0,1,0,1,0,0,0,1,0,0,0,0,0,0,1,0,1,1,0, -3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,2,3,3,2,2,2,2,3,2, -1,2,1,2,2,2,0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,1,0, -3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,2,3,3,2,3,2,2,3,2,2,2,1,2,2,2,2, -1,2,1,1,2,2,0,1,2,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0, -3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,2,0,2,2,2,2,2, -0,2,0,2,2,2,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,1,0, -3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,2,3,0,2,2,2, -0,2,1,2,2,2,0,0,2,1,0,0,0,0,1,0,1,0,0,0,0,0,0,2,0,0,0,0,0,0,1,0, -3,3,3,3,3,3,3,3,3,3,3,2,3,3,3,3,3,3,3,3,3,3,3,3,3,2,1,2,3,2,2,2, -1,2,1,2,2,2,0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,0,0,1,1,0, -3,3,3,3,3,3,3,3,3,2,3,3,3,2,3,3,3,3,3,3,3,3,3,3,3,3,3,1,0,2,0,2, -0,2,1,2,2,2,0,0,1,2,0,0,0,0,1,0,1,0,0,0,0,0,0,1,0,0,0,2,0,0,1,0, -3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,2,3,2,3,2,2,3,2,1,2,1,1,1, -0,1,1,1,1,1,3,0,1,0,0,0,0,2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0, -3,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,0,1,1,0,1,1,0,0,1,0,0,1,0,0,0,0, -0,0,1,0,0,0,0,0,2,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,2,2,2,2,2,2,2, -0,2,0,1,2,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0, -3,3,3,3,3,3,3,3,3,2,3,3,3,2,1,2,3,3,2,3,3,3,3,2,3,2,1,2,0,2,1,2, -0,2,0,2,2,2,0,0,1,2,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0, -3,3,3,3,3,3,3,3,3,2,3,3,3,1,2,2,3,3,2,3,2,3,2,2,3,1,2,2,0,2,2,2, -0,2,1,2,2,2,0,0,1,2,0,0,0,0,1,0,0,0,0,0,1,0,0,1,0,0,0,1,0,0,1,0, -3,3,3,3,3,3,3,3,3,3,3,3,3,2,3,3,3,2,3,3,2,2,2,3,3,3,3,1,3,2,2,2, -0,2,0,1,2,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0, -3,3,3,3,3,3,3,3,3,3,3,3,3,3,2,2,3,3,3,2,3,2,2,2,1,2,2,0,2,2,2,2, -0,2,0,2,2,2,0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0, -3,3,3,3,3,3,3,3,3,3,3,2,3,3,3,1,3,2,3,3,2,3,3,2,2,1,2,2,2,2,2,2, -0,2,1,2,1,2,0,0,1,0,0,0,0,0,1,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,1,0, -3,3,3,3,3,3,2,3,2,3,3,2,3,3,3,3,2,3,2,3,3,3,3,3,2,2,2,2,2,2,2,1, -0,2,0,1,2,1,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,1,0, -3,3,3,3,3,3,3,3,3,2,1,2,3,3,3,3,3,3,3,2,3,2,3,2,1,2,3,0,2,1,2,2, -0,2,1,1,2,1,0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,2,0, -3,3,3,3,3,3,3,3,3,2,3,3,3,3,2,1,3,1,2,2,2,1,2,3,3,1,2,1,2,2,2,2, -0,1,1,1,1,1,0,0,0,0,0,0,0,0,1,0,0,0,0,0,1,0,0,2,0,0,0,0,0,0,0,0, -3,3,3,3,3,3,3,3,3,3,0,2,3,3,3,1,3,3,3,1,2,2,2,2,1,1,2,2,2,2,2,2, -0,2,0,1,1,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,1,0, -3,3,3,3,3,3,2,3,3,3,2,2,3,3,3,2,1,2,3,2,3,2,2,2,2,1,2,1,1,1,2,2, -0,2,1,1,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0, -3,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,0,1,0,0,0,1,0,0,0,0,0, -1,0,1,0,0,0,0,0,2,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,2,3,3,2,3,1,2,2,2,2,3,2,3,1,1,2,2,1,2,2,1,1,0,2,2,2,2, -0,1,0,1,2,2,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,1,0, -3,0,0,1,1,0,1,0,0,1,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,2,2,0, -0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,0,1,0,1,0,1,1,0,1,1,0,0,0,1,1,0,1,1,1,0,0,0,0,0,0,1,0,0,0,0,0, -0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,0,0,0,1,1,0,1,0,1,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0, -3,2,2,1,2,2,2,2,2,2,2,1,2,2,1,2,2,1,1,1,1,1,1,1,1,2,1,1,0,3,3,3, -0,3,0,2,2,2,2,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0, -2,2,2,3,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,1,2,2,1,2,2,2,1,1,1,2,0,1, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,2,2,2,2,2,2,2,2,2,2,1,2,2,2,2,2,2,2,2,2,2,2,0,2,2,0,0,0,0,0,0, -0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,3,1,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,1,2,1,0,2,1,0, -0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,1,1,1,1,1,1,1,1,1,1,0,0,1,1,1,1,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0, -0,3,1,1,2,2,2,2,2,1,2,2,2,1,1,2,2,2,2,2,2,2,1,2,2,1,0,1,1,1,1,0, -0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,2,1,1,1,1,2,1,1,2,1,0,1,1,1,1,1,1,1,1,1,1,1,0,1,0,0,0,0,0,0,0, -0,0,2,0,0,0,0,0,0,0,0,1,1,0,0,0,0,1,1,0,0,1,1,0,0,0,0,0,0,1,0,0, -2,1,1,2,2,2,2,2,2,2,2,2,2,2,1,2,2,2,2,2,1,2,1,2,1,1,1,1,0,0,0,0, -0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -1,2,1,2,2,2,2,2,2,2,2,2,2,1,2,1,2,1,1,2,1,1,1,2,1,2,1,2,0,1,0,1, -0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,3,1,2,2,2,1,2,2,2,2,2,2,2,2,1,2,1,1,1,1,1,1,2,1,2,1,1,0,1,0,1, -0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,1,2,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,2,2, -0,2,0,1,2,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0, -3,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,1,1,1,1,1,1,1,0,1,1,0,1,0,0,1,0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,2,0,1,1,1,0,1,0,0,0,1,1,0,1,1,0,0,0,0,0,1,1,0,0, -0,1,1,1,2,1,2,2,2,0,2,0,2,0,1,1,2,1,1,1,1,2,1,0,1,1,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0, -1,0,1,0,0,0,0,0,1,0,1,2,2,0,1,0,0,1,1,2,2,1,2,0,2,0,0,0,1,2,0,1, -2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,2,0,2,1,2,0,2,0,0,1,1,1,1,1,1,0,1,0,0,0,1,0,0,1, -2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,1,0,0,0,0,0,1,0,2,1,1,0,1,0,0,1,1,1,2,2,0,0,1,0,0,0,1,0,0,1, -1,1,2,1,0,1,1,1,0,1,0,1,1,1,1,0,0,0,1,0,1,0,0,0,0,0,0,0,0,2,2,1, -0,2,0,1,2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,1,0,0,1,0,1,1,1,1,0,0,0,0,0,1,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -1,1,1,1,1,1,1,1,1,2,1,0,1,1,1,1,1,1,1,1,1,1,1,0,1,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,1,1,1,0,1,1,0,1,0,0,0,1,1,0,1, -2,0,1,0,1,0,1,0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,1,0,1,1,1,0,1,0,0,1,1,2,1,1,2,0,1,0,0,0,1,1,0,1, -1,0,0,1,0,0,1,0,0,0,1,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,1,0,1,1,2,0,1,0,0,0,0,2,1,1,2,0,2,0,0,0,1,1,0,1, -1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,1,0,2,1,1,0,1,0,0,2,2,1,2,1,1,0,1,0,0,0,1,1,0,1, -2,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,1,2,2,0,0,0,0,0,1,1,0,1,0,0,1,0,0,0,0,1,0,1, -1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,1,2,2,0,0,0,0,2,1,1,1,0,2,1,1,0,0,0,2,1,0,1, -1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,1,0,1,1,2,0,1,0,0,1,1,0,2,1,1,0,1,0,0,0,1,1,0,1, -2,2,1,1,1,0,1,1,0,1,1,0,1,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,1,0,2,1,1,0,1,0,0,1,1,0,1,2,1,0,2,0,0,0,1,1,0,1, -2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0, -0,1,0,0,2,0,2,1,1,0,1,0,1,0,0,1,0,0,0,0,1,0,0,0,1,0,0,0,0,0,1,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -1,0,0,1,0,0,1,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,1,0,1,1,2,0,1,0,0,1,1,1,0,1,0,0,1,0,0,0,1,0,0,1, -1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -1,0,0,0,0,0,0,0,1,0,1,1,0,0,1,0,0,2,1,1,1,1,1,0,1,0,0,0,0,1,0,1, -0,1,1,1,2,1,1,1,1,0,1,1,1,1,1,1,1,1,1,1,1,1,0,1,1,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,1,2,1,0,0,0,0,0,1,1,1,1,1,0,1,0,0,0,1,1,0,0, -) - -Win1255HebrewModel = { - 'char_to_order_map': WIN1255_CHAR_TO_ORDER_MAP, - 'precedence_matrix': HEBREW_LANG_MODEL, - 'typical_positive_ratio': 0.984004, - 'keep_english_letter': False, - 'charset_name': "windows-1255", - 'language': 'Hebrew', -} diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/langhungarianmodel.py b/src/fetchcode/vcs/pip/_vendor/chardet/langhungarianmodel.py deleted file mode 100644 index bb7c095e..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/langhungarianmodel.py +++ /dev/null @@ -1,225 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Communicator client code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -# 255: Control characters that usually does not exist in any text -# 254: Carriage/Return -# 253: symbol (punctuation) that does not belong to word -# 252: 0 - 9 - -# Character Mapping Table: -Latin2_HungarianCharToOrderMap = ( -255,255,255,255,255,255,255,255,255,255,254,255,255,254,255,255, # 00 -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 10 -253,253,253,253,253,253,253,253,253,253,253,253,253,253,253,253, # 20 -252,252,252,252,252,252,252,252,252,252,253,253,253,253,253,253, # 30 -253, 28, 40, 54, 45, 32, 50, 49, 38, 39, 53, 36, 41, 34, 35, 47, - 46, 71, 43, 33, 37, 57, 48, 64, 68, 55, 52,253,253,253,253,253, -253, 2, 18, 26, 17, 1, 27, 12, 20, 9, 22, 7, 6, 13, 4, 8, - 23, 67, 10, 5, 3, 21, 19, 65, 62, 16, 11,253,253,253,253,253, -159,160,161,162,163,164,165,166,167,168,169,170,171,172,173,174, -175,176,177,178,179,180,181,182,183,184,185,186,187,188,189,190, -191,192,193,194,195,196,197, 75,198,199,200,201,202,203,204,205, - 79,206,207,208,209,210,211,212,213,214,215,216,217,218,219,220, -221, 51, 81,222, 78,223,224,225,226, 44,227,228,229, 61,230,231, -232,233,234, 58,235, 66, 59,236,237,238, 60, 69, 63,239,240,241, - 82, 14, 74,242, 70, 80,243, 72,244, 15, 83, 77, 84, 30, 76, 85, -245,246,247, 25, 73, 42, 24,248,249,250, 31, 56, 29,251,252,253, -) - -win1250HungarianCharToOrderMap = ( -255,255,255,255,255,255,255,255,255,255,254,255,255,254,255,255, # 00 -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 10 -253,253,253,253,253,253,253,253,253,253,253,253,253,253,253,253, # 20 -252,252,252,252,252,252,252,252,252,252,253,253,253,253,253,253, # 30 -253, 28, 40, 54, 45, 32, 50, 49, 38, 39, 53, 36, 41, 34, 35, 47, - 46, 72, 43, 33, 37, 57, 48, 64, 68, 55, 52,253,253,253,253,253, -253, 2, 18, 26, 17, 1, 27, 12, 20, 9, 22, 7, 6, 13, 4, 8, - 23, 67, 10, 5, 3, 21, 19, 65, 62, 16, 11,253,253,253,253,253, -161,162,163,164,165,166,167,168,169,170,171,172,173,174,175,176, -177,178,179,180, 78,181, 69,182,183,184,185,186,187,188,189,190, -191,192,193,194,195,196,197, 76,198,199,200,201,202,203,204,205, - 81,206,207,208,209,210,211,212,213,214,215,216,217,218,219,220, -221, 51, 83,222, 80,223,224,225,226, 44,227,228,229, 61,230,231, -232,233,234, 58,235, 66, 59,236,237,238, 60, 70, 63,239,240,241, - 84, 14, 75,242, 71, 82,243, 73,244, 15, 85, 79, 86, 30, 77, 87, -245,246,247, 25, 74, 42, 24,248,249,250, 31, 56, 29,251,252,253, -) - -# Model Table: -# total sequences: 100% -# first 512 sequences: 94.7368% -# first 1024 sequences:5.2623% -# rest sequences: 0.8894% -# negative sequences: 0.0009% -HungarianLangModel = ( -0,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,1,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3, -3,3,3,3,3,3,3,3,3,3,2,3,3,3,3,3,3,3,3,2,2,3,3,1,1,2,2,2,2,2,1,2, -3,2,2,3,3,3,3,3,2,3,3,3,3,3,3,1,2,3,3,3,3,2,3,3,1,1,3,3,0,1,1,1, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0, -3,2,1,3,3,3,3,3,2,3,3,3,3,3,1,1,2,3,3,3,3,3,3,3,1,1,3,2,0,1,1,1, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0, -3,3,3,3,3,3,3,3,3,3,3,1,1,2,3,3,3,1,3,3,3,3,3,1,3,3,2,2,0,3,2,3, -0,0,0,0,0,0,0,0,0,0,3,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0, -3,3,3,3,3,3,2,3,3,3,2,3,3,2,3,3,3,3,3,2,3,3,2,2,3,2,3,2,0,3,2,2, -0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,1,0, -3,3,3,3,3,3,2,3,3,3,3,3,2,3,3,3,1,2,3,2,2,3,1,2,3,3,2,2,0,3,3,3, -0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0, -3,3,3,3,3,3,3,3,3,3,2,2,3,3,3,3,3,3,2,3,3,3,3,2,3,3,3,3,0,2,3,2, -0,0,0,1,1,0,0,0,0,0,3,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0, -3,3,3,3,3,3,3,3,3,3,3,1,1,1,3,3,2,1,3,2,2,3,2,1,3,2,2,1,0,3,3,1, -0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0, -3,2,2,3,3,3,3,3,1,2,3,3,3,3,1,2,1,3,3,3,3,2,2,3,1,1,3,2,0,1,1,1, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0, -3,3,3,3,3,3,3,3,2,2,3,3,3,3,3,2,1,3,3,3,3,3,2,2,1,3,3,3,0,1,1,2, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,1,0, -3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,2,3,3,3,2,3,3,2,3,3,3,2,0,3,2,3, -0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,1,0, -3,3,3,3,3,3,2,3,3,3,2,3,2,3,3,3,1,3,2,2,2,3,1,1,3,3,1,1,0,3,3,2, -0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0, -3,3,3,3,3,3,3,2,3,3,3,2,3,2,3,3,3,2,3,3,3,3,3,1,2,3,2,2,0,2,2,2, -0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0, -3,3,3,2,2,2,3,1,3,3,2,2,1,3,3,3,1,1,3,1,2,3,2,3,2,2,2,1,0,2,2,2, -0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0, -3,1,1,3,3,3,3,3,1,2,3,3,3,3,1,2,1,3,3,3,2,2,3,2,1,0,3,2,0,1,1,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,1,1,3,3,3,3,3,1,2,3,3,3,3,1,1,0,3,3,3,3,0,2,3,0,0,2,1,0,1,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,3,2,2,3,3,2,2,2,2,3,3,0,1,2,3,2,3,2,2,3,2,1,2,0,2,2,2, -0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0, -3,3,3,3,3,3,1,2,3,3,3,2,1,2,3,3,2,2,2,3,2,3,3,1,3,3,1,1,0,2,3,2, -0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0, -3,3,3,1,2,2,2,2,3,3,3,1,1,1,3,3,1,1,3,1,1,3,2,1,2,3,1,1,0,2,2,2, -0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0, -3,3,3,2,1,2,1,1,3,3,1,1,1,1,3,3,1,1,2,2,1,2,1,1,2,2,1,1,0,2,2,1, -0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0, -3,3,3,1,1,2,1,1,3,3,1,0,1,1,3,3,2,0,1,1,2,3,1,0,2,2,1,0,0,1,3,2, -0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0, -3,2,1,3,3,3,3,3,1,2,3,2,3,3,2,1,1,3,2,3,2,1,2,2,0,1,2,1,0,0,1,1, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0, -3,3,3,3,2,2,2,2,3,1,2,2,1,1,3,3,0,3,2,1,2,3,2,1,3,3,1,1,0,2,1,3, -0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0, -3,3,3,2,2,2,3,2,3,3,3,2,1,1,3,3,1,1,1,2,2,3,2,3,2,2,2,1,0,2,2,1, -0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0, -1,0,0,3,3,3,3,3,0,0,3,3,2,3,0,0,0,2,3,3,1,0,1,2,0,0,1,1,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,1,2,3,3,3,3,3,1,2,3,3,2,2,1,1,0,3,3,2,2,1,2,2,1,0,2,2,0,1,1,1, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,2,2,1,3,1,2,3,3,2,2,1,1,2,2,1,1,1,1,3,2,1,1,1,1,2,1,0,1,2,1, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0, -2,3,3,1,1,1,1,1,3,3,3,0,1,1,3,3,1,1,1,1,1,2,2,0,3,1,1,2,0,2,1,1, -0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0, -3,1,0,1,2,1,2,2,0,1,2,3,1,2,0,0,0,2,1,1,1,1,1,2,0,0,1,1,0,0,0,0, -1,2,1,2,2,2,1,2,1,2,0,2,0,2,2,1,1,2,1,1,2,1,1,1,0,1,0,0,0,1,1,0, -1,1,1,2,3,2,3,3,0,1,2,2,3,1,0,1,0,2,1,2,2,0,1,1,0,0,1,1,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -1,0,0,3,3,2,2,1,0,0,3,2,3,2,0,0,0,1,1,3,0,0,1,1,0,0,2,1,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,1,1,2,2,3,3,1,0,1,3,2,3,1,1,1,0,1,1,1,1,1,3,1,0,0,2,2,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,1,1,1,2,2,2,1,0,1,2,3,3,2,0,0,0,2,1,1,1,2,1,1,1,0,1,1,1,0,0,0, -1,2,2,2,2,2,1,1,1,2,0,2,1,1,1,1,1,2,1,1,1,1,1,1,0,1,1,1,0,0,1,1, -3,2,2,1,0,0,1,1,2,2,0,3,0,1,2,1,1,0,0,1,1,1,0,1,1,1,1,0,2,1,1,1, -2,2,1,1,1,2,1,2,1,1,1,1,1,1,1,2,1,1,1,2,3,1,1,1,1,1,1,1,1,1,0,1, -2,3,3,0,1,0,0,0,3,3,1,0,0,1,2,2,1,0,0,0,0,2,0,0,1,1,1,0,2,1,1,1, -2,1,1,1,1,1,1,2,1,1,0,1,1,0,1,1,1,0,1,2,1,1,0,1,1,1,1,1,1,1,0,1, -2,3,3,0,1,0,0,0,2,2,0,0,0,0,1,2,2,0,0,0,0,1,0,0,1,1,0,0,2,0,1,0, -2,1,1,1,1,2,1,1,1,1,1,1,1,2,1,1,1,1,1,1,1,1,1,2,0,1,1,1,1,1,0,1, -3,2,2,0,1,0,1,0,2,3,2,0,0,1,2,2,1,0,0,1,1,1,0,0,2,1,0,1,2,2,1,1, -2,1,1,1,1,1,1,2,1,1,1,1,1,1,0,2,1,0,1,1,0,1,1,1,0,1,1,2,1,1,0,1, -2,2,2,0,0,1,0,0,2,2,1,1,0,0,2,1,1,0,0,0,1,2,0,0,2,1,0,0,2,1,1,1, -2,1,1,1,1,2,1,2,1,1,1,2,2,1,1,2,1,1,1,2,1,1,1,1,1,1,1,1,1,1,0,1, -1,2,3,0,0,0,1,0,3,2,1,0,0,1,2,1,1,0,0,0,0,2,1,0,1,1,0,0,2,1,2,1, -1,1,0,0,0,1,0,1,1,1,1,1,2,0,0,1,0,0,0,2,0,0,1,1,1,1,1,1,1,1,0,1, -3,0,0,2,1,2,2,1,0,0,2,1,2,2,0,0,0,2,1,1,1,0,1,1,0,0,1,1,2,0,0,0, -1,2,1,2,2,1,1,2,1,2,0,1,1,1,1,1,1,1,1,1,2,1,1,0,0,1,1,1,1,0,0,1, -1,3,2,0,0,0,1,0,2,2,2,0,0,0,2,2,1,0,0,0,0,3,1,1,1,1,0,0,2,1,1,1, -2,1,0,1,1,1,0,1,1,1,1,1,1,1,0,2,1,0,0,1,0,1,1,0,1,1,1,1,1,1,0,1, -2,3,2,0,0,0,1,0,2,2,0,0,0,0,2,1,1,0,0,0,0,2,1,0,1,1,0,0,2,1,1,0, -2,1,1,1,1,2,1,2,1,2,0,1,1,1,0,2,1,1,1,2,1,1,1,1,0,1,1,1,1,1,0,1, -3,1,1,2,2,2,3,2,1,1,2,2,1,1,0,1,0,2,2,1,1,1,1,1,0,0,1,1,0,1,1,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,2,2,0,0,0,0,0,2,2,0,0,0,0,2,2,1,0,0,0,1,1,0,0,1,2,0,0,2,1,1,1, -2,2,1,1,1,2,1,2,1,1,0,1,1,1,1,2,1,1,1,2,1,1,1,1,0,1,2,1,1,1,0,1, -1,0,0,1,2,3,2,1,0,0,2,0,1,1,0,0,0,1,1,1,1,0,1,1,0,0,1,0,0,0,0,0, -1,2,1,2,1,2,1,1,1,2,0,2,1,1,1,0,1,2,0,0,1,1,1,0,0,0,0,0,0,0,0,0, -2,3,2,0,0,0,0,0,1,1,2,1,0,0,1,1,1,0,0,0,0,2,0,0,1,1,0,0,2,1,1,1, -2,1,1,1,1,1,1,2,1,0,1,1,1,1,0,2,1,1,1,1,1,1,0,1,0,1,1,1,1,1,0,1, -1,2,2,0,1,1,1,0,2,2,2,0,0,0,3,2,1,0,0,0,1,1,0,0,1,1,0,1,1,1,0,0, -1,1,0,1,1,1,1,1,1,1,1,2,1,1,1,1,1,1,1,2,1,1,1,0,0,1,1,1,0,1,0,1, -2,1,0,2,1,1,2,2,1,1,2,1,1,1,0,0,0,1,1,0,1,1,1,1,0,0,1,1,1,0,0,0, -1,2,2,2,2,2,1,1,1,2,0,2,1,1,1,1,1,1,1,1,1,1,1,1,0,1,1,0,0,0,1,0, -1,2,3,0,0,0,1,0,2,2,0,0,0,0,2,2,0,0,0,0,0,1,0,0,1,0,0,0,2,0,1,0, -2,1,1,1,1,1,0,2,0,0,0,1,2,1,1,1,1,0,1,2,0,1,0,1,0,1,1,1,0,1,0,1, -2,2,2,0,0,0,1,0,2,1,2,0,0,0,1,1,2,0,0,0,0,1,0,0,1,1,0,0,2,1,0,1, -2,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,2,0,1,1,1,1,1,0,1, -1,2,2,0,0,0,1,0,2,2,2,0,0,0,1,1,0,0,0,0,0,1,1,0,2,0,0,1,1,1,0,1, -1,0,1,1,1,1,1,1,0,1,1,1,1,0,0,1,0,0,1,1,0,1,0,1,1,1,1,1,0,0,0,1, -1,0,0,1,0,1,2,1,0,0,1,1,1,2,0,0,0,1,1,0,1,0,1,1,0,0,1,0,0,0,0,0, -0,2,1,2,1,1,1,1,1,2,0,2,0,1,1,0,1,2,1,0,1,1,1,0,0,0,0,0,0,1,0,0, -2,1,1,0,1,2,0,0,1,1,1,0,0,0,1,1,0,0,0,0,0,1,0,0,1,0,0,0,2,1,0,1, -2,2,1,1,1,1,1,2,1,1,0,1,1,1,1,2,1,1,1,2,1,1,0,1,0,1,1,1,1,1,0,1, -1,2,2,0,0,0,0,0,1,1,0,0,0,0,2,1,0,0,0,0,0,2,0,0,2,2,0,0,2,0,0,1, -2,1,1,1,1,1,1,1,0,1,1,0,1,1,0,1,0,0,0,1,1,1,1,0,0,1,1,1,1,0,0,1, -1,1,2,0,0,3,1,0,2,1,1,1,0,0,1,1,1,0,0,0,1,1,0,0,0,1,0,0,1,0,1,0, -1,2,1,0,1,1,1,2,1,1,0,1,1,1,1,1,0,0,0,1,1,1,1,1,0,1,0,0,0,1,0,0, -2,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,0,1,0,0,0,1,0,0,0,0,2,0,0,0, -2,1,1,1,1,1,1,1,1,1,0,1,1,1,1,1,1,1,1,1,2,1,1,0,0,1,1,1,1,1,0,1, -2,1,1,1,2,1,1,1,0,1,1,2,1,0,0,0,0,1,1,1,1,0,1,0,0,0,0,1,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -1,1,0,1,1,1,1,1,0,0,1,1,2,1,0,0,0,1,1,0,0,0,1,1,0,0,1,0,1,0,0,0, -1,2,1,1,1,1,1,1,1,1,0,1,0,1,1,1,1,1,1,0,1,1,1,0,0,0,0,0,0,1,0,0, -2,0,0,0,1,1,1,1,0,0,1,1,0,0,0,0,0,1,1,1,2,0,0,1,0,0,1,0,1,0,0,0, -0,1,1,1,1,1,1,1,1,2,0,1,1,1,1,0,1,1,1,0,1,1,1,0,0,0,0,0,0,0,0,0, -1,0,0,1,1,1,1,1,0,0,2,1,0,1,0,0,0,1,0,1,0,0,0,0,0,0,1,0,0,0,0,0, -0,1,1,1,1,1,1,0,1,1,0,1,0,1,1,0,1,1,0,0,1,1,1,0,0,0,0,0,0,0,0,0, -1,0,0,1,1,1,0,0,0,0,1,0,2,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0, -0,1,1,1,1,1,0,0,1,1,0,1,0,1,0,0,1,1,1,0,1,1,1,0,0,0,0,0,0,0,0,0, -0,0,0,1,0,0,0,0,0,0,1,1,2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,1,1,1,0,1,0,0,1,1,0,1,0,1,1,0,1,1,1,0,1,1,1,0,0,0,0,0,0,0,0,0, -2,1,1,1,1,1,1,1,1,1,1,0,0,1,1,1,0,0,1,0,0,1,0,1,0,1,1,1,0,0,1,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -1,0,0,1,1,1,1,0,0,0,1,1,1,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0, -0,1,1,1,1,1,1,0,1,1,0,1,0,1,0,0,1,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0, -) - -Latin2HungarianModel = { - 'char_to_order_map': Latin2_HungarianCharToOrderMap, - 'precedence_matrix': HungarianLangModel, - 'typical_positive_ratio': 0.947368, - 'keep_english_letter': True, - 'charset_name': "ISO-8859-2", - 'language': 'Hungarian', -} - -Win1250HungarianModel = { - 'char_to_order_map': win1250HungarianCharToOrderMap, - 'precedence_matrix': HungarianLangModel, - 'typical_positive_ratio': 0.947368, - 'keep_english_letter': True, - 'charset_name': "windows-1250", - 'language': 'Hungarian', -} diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/langthaimodel.py b/src/fetchcode/vcs/pip/_vendor/chardet/langthaimodel.py deleted file mode 100644 index 15f94c2d..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/langthaimodel.py +++ /dev/null @@ -1,199 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Communicator client code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -# 255: Control characters that usually does not exist in any text -# 254: Carriage/Return -# 253: symbol (punctuation) that does not belong to word -# 252: 0 - 9 - -# The following result for thai was collected from a limited sample (1M). - -# Character Mapping Table: -TIS620CharToOrderMap = ( -255,255,255,255,255,255,255,255,255,255,254,255,255,254,255,255, # 00 -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 10 -253,253,253,253,253,253,253,253,253,253,253,253,253,253,253,253, # 20 -252,252,252,252,252,252,252,252,252,252,253,253,253,253,253,253, # 30 -253,182,106,107,100,183,184,185,101, 94,186,187,108,109,110,111, # 40 -188,189,190, 89, 95,112,113,191,192,193,194,253,253,253,253,253, # 50 -253, 64, 72, 73,114, 74,115,116,102, 81,201,117, 90,103, 78, 82, # 60 - 96,202, 91, 79, 84,104,105, 97, 98, 92,203,253,253,253,253,253, # 70 -209,210,211,212,213, 88,214,215,216,217,218,219,220,118,221,222, -223,224, 99, 85, 83,225,226,227,228,229,230,231,232,233,234,235, -236, 5, 30,237, 24,238, 75, 8, 26, 52, 34, 51,119, 47, 58, 57, - 49, 53, 55, 43, 20, 19, 44, 14, 48, 3, 17, 25, 39, 62, 31, 54, - 45, 9, 16, 2, 61, 15,239, 12, 42, 46, 18, 21, 76, 4, 66, 63, - 22, 10, 1, 36, 23, 13, 40, 27, 32, 35, 86,240,241,242,243,244, - 11, 28, 41, 29, 33,245, 50, 37, 6, 7, 67, 77, 38, 93,246,247, - 68, 56, 59, 65, 69, 60, 70, 80, 71, 87,248,249,250,251,252,253, -) - -# Model Table: -# total sequences: 100% -# first 512 sequences: 92.6386% -# first 1024 sequences:7.3177% -# rest sequences: 1.0230% -# negative sequences: 0.0436% -ThaiLangModel = ( -0,1,3,3,3,3,0,0,3,3,0,3,3,0,3,3,3,3,3,3,3,3,0,0,3,3,3,0,3,3,3,3, -0,3,3,0,0,0,1,3,0,3,3,2,3,3,0,1,2,3,3,3,3,0,2,0,2,0,0,3,2,1,2,2, -3,0,3,3,2,3,0,0,3,3,0,3,3,0,3,3,3,3,3,3,3,3,3,0,3,2,3,0,2,2,2,3, -0,2,3,0,0,0,0,1,0,1,2,3,1,1,3,2,2,0,1,1,0,0,1,0,0,0,0,0,0,0,1,1, -3,3,3,2,3,3,3,3,3,3,3,3,3,3,3,2,2,2,2,2,2,2,3,3,2,3,2,3,3,2,2,2, -3,1,2,3,0,3,3,2,2,1,2,3,3,1,2,0,1,3,0,1,0,0,1,0,0,0,0,0,0,0,1,1, -3,3,2,2,3,3,3,3,1,2,3,3,3,3,3,2,2,2,2,3,3,2,2,3,3,2,2,3,2,3,2,2, -3,3,1,2,3,1,2,2,3,3,1,0,2,1,0,0,3,1,2,1,0,0,1,0,0,0,0,0,0,1,0,1, -3,3,3,3,3,3,2,2,3,3,3,3,2,3,2,2,3,3,2,2,3,2,2,2,2,1,1,3,1,2,1,1, -3,2,1,0,2,1,0,1,0,1,1,0,1,1,0,0,1,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0, -3,3,3,2,3,2,3,3,2,2,3,2,3,3,2,3,1,1,2,3,2,2,2,3,2,2,2,2,2,1,2,1, -2,2,1,1,3,3,2,1,0,1,2,2,0,1,3,0,0,0,1,1,0,0,0,0,0,2,3,0,0,2,1,1, -3,3,2,3,3,2,0,0,3,3,0,3,3,0,2,2,3,1,2,2,1,1,1,0,2,2,2,0,2,2,1,1, -0,2,1,0,2,0,0,2,0,1,0,0,1,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,1,0, -3,3,2,3,3,2,0,0,3,3,0,2,3,0,2,1,2,2,2,2,1,2,0,0,2,2,2,0,2,2,1,1, -0,2,1,0,2,0,0,2,0,1,1,0,1,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0, -3,3,2,3,2,3,2,0,2,2,1,3,2,1,3,2,1,2,3,2,2,3,0,2,3,2,2,1,2,2,2,2, -1,2,2,0,0,0,0,2,0,1,2,0,1,1,1,0,1,0,3,1,1,0,0,0,0,0,0,0,0,0,1,0, -3,3,2,3,3,2,3,2,2,2,3,2,2,3,2,2,1,2,3,2,2,3,1,3,2,2,2,3,2,2,2,3, -3,2,1,3,0,1,1,1,0,2,1,1,1,1,1,0,1,0,1,1,0,0,0,0,0,0,0,0,0,2,0,0, -1,0,0,3,0,3,3,3,3,3,0,0,3,0,2,2,3,3,3,3,3,0,0,0,1,1,3,0,0,0,0,2, -0,0,1,0,0,0,0,0,0,0,2,3,0,0,0,3,0,2,0,0,0,0,0,3,0,0,0,0,0,0,0,0, -2,0,3,3,3,3,0,0,2,3,0,0,3,0,3,3,2,3,3,3,3,3,0,0,3,3,3,0,0,0,3,3, -0,0,3,0,0,0,0,2,0,0,2,1,1,3,0,0,1,0,0,2,3,0,1,0,0,0,0,0,0,0,1,0, -3,3,3,3,2,3,3,3,3,3,3,3,1,2,1,3,3,2,2,1,2,2,2,3,1,1,2,0,2,1,2,1, -2,2,1,0,0,0,1,1,0,1,0,1,1,0,0,0,0,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0, -3,0,2,1,2,3,3,3,0,2,0,2,2,0,2,1,3,2,2,1,2,1,0,0,2,2,1,0,2,1,2,2, -0,1,1,0,0,0,0,1,0,1,1,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,2,1,3,3,1,1,3,0,2,3,1,1,3,2,1,1,2,0,2,2,3,2,1,1,1,1,1,2, -3,0,0,1,3,1,2,1,2,0,3,0,0,0,1,0,3,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0, -3,3,1,1,3,2,3,3,3,1,3,2,1,3,2,1,3,2,2,2,2,1,3,3,1,2,1,3,1,2,3,0, -2,1,1,3,2,2,2,1,2,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2, -3,3,2,3,2,3,3,2,3,2,3,2,3,3,2,1,0,3,2,2,2,1,2,2,2,1,2,2,1,2,1,1, -2,2,2,3,0,1,3,1,1,1,1,0,1,1,0,2,1,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,2,3,2,2,1,1,3,2,3,2,3,2,0,3,2,2,1,2,0,2,2,2,1,2,2,2,2,1, -3,2,1,2,2,1,0,2,0,1,0,0,1,1,0,0,0,0,0,1,1,0,1,0,0,0,0,0,0,0,0,1, -3,3,3,3,3,2,3,1,2,3,3,2,2,3,0,1,1,2,0,3,3,2,2,3,0,1,1,3,0,0,0,0, -3,1,0,3,3,0,2,0,2,1,0,0,3,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,2,3,2,3,3,0,1,3,1,1,2,1,2,1,1,3,1,1,0,2,3,1,1,1,1,1,1,1,1, -3,1,1,2,2,2,2,1,1,1,0,0,2,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1, -3,2,2,1,1,2,1,3,3,2,3,2,2,3,2,2,3,1,2,2,1,2,0,3,2,1,2,2,2,2,2,1, -3,2,1,2,2,2,1,1,1,1,0,0,1,1,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,3,3,3,1,3,3,0,2,1,0,3,2,0,0,3,1,0,1,1,0,1,0,0,0,0,0,1, -1,0,0,1,0,3,2,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,0,2,2,2,3,0,0,1,3,0,3,2,0,3,2,2,3,3,3,3,3,1,0,2,2,2,0,2,2,1,2, -0,2,3,0,0,0,0,1,0,1,0,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1, -3,0,2,3,1,3,3,2,3,3,0,3,3,0,3,2,2,3,2,3,3,3,0,0,2,2,3,0,1,1,1,3, -0,0,3,0,0,0,2,2,0,1,3,0,1,2,2,2,3,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1, -3,2,3,3,2,0,3,3,2,2,3,1,3,2,1,3,2,0,1,2,2,0,2,3,2,1,0,3,0,0,0,0, -3,0,0,2,3,1,3,0,0,3,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,1,3,2,2,2,1,2,0,1,3,1,1,3,1,3,0,0,2,1,1,1,1,2,1,1,1,0,2,1,0,1, -1,2,0,0,0,3,1,1,0,0,0,0,1,0,1,0,0,1,0,1,0,0,0,0,0,3,1,0,0,0,1,0, -3,3,3,3,2,2,2,2,2,1,3,1,1,1,2,0,1,1,2,1,2,1,3,2,0,0,3,1,1,1,1,1, -3,1,0,2,3,0,0,0,3,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,2,3,0,3,3,0,2,0,0,0,0,0,0,0,3,0,0,1,0,0,0,0,0,0,0,0,0,0,0, -0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,2,3,1,3,0,0,1,2,0,0,2,0,3,3,2,3,3,3,2,3,0,0,2,2,2,0,0,0,2,2, -0,0,1,0,0,0,0,3,0,0,0,0,2,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0, -0,0,0,3,0,2,0,0,0,0,0,0,0,0,0,0,1,2,3,1,3,3,0,0,1,0,3,0,0,0,0,0, -0,0,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,1,2,3,1,2,3,1,0,3,0,2,2,1,0,2,1,1,2,0,1,0,0,1,1,1,1,0,1,0,0, -1,0,0,0,0,1,1,0,3,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,2,1,0,1,1,1,3,1,2,2,2,2,2,2,1,1,1,1,0,3,1,0,1,3,1,1,1,1, -1,1,0,2,0,1,3,1,1,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,2,0,1, -3,0,2,2,1,3,3,2,3,3,0,1,1,0,2,2,1,2,1,3,3,1,0,0,3,2,0,0,0,0,2,1, -0,1,0,0,0,0,1,2,0,1,1,3,1,1,2,2,1,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0, -0,0,3,0,0,1,0,0,0,3,0,0,3,0,3,1,0,1,1,1,3,2,0,0,0,3,0,0,0,0,2,0, -0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,2,0,0,0,0,0,0,0,0,0, -3,3,1,3,2,1,3,3,1,2,2,0,1,2,1,0,1,2,0,0,0,0,0,3,0,0,0,3,0,0,0,0, -3,0,0,1,1,1,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,0,1,2,0,3,3,3,2,2,0,1,1,0,1,3,0,0,0,2,2,0,0,0,0,3,1,0,1,0,0,0, -0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,0,2,3,1,2,0,0,2,1,0,3,1,0,1,2,0,1,1,1,1,3,0,0,3,1,1,0,2,2,1,1, -0,2,0,0,0,0,0,1,0,1,0,0,1,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,0,0,3,1,2,0,0,2,2,0,1,2,0,1,0,1,3,1,2,1,0,0,0,2,0,3,0,0,0,1,0, -0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,0,1,1,2,2,0,0,0,2,0,2,1,0,1,1,0,1,1,1,2,1,0,0,1,1,1,0,2,1,1,1, -0,1,1,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,1,0,1, -0,0,0,2,0,1,3,1,1,1,1,0,0,0,0,3,2,0,1,0,0,0,1,2,0,0,0,1,0,0,0,0, -0,0,0,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,3,3,3,3,1,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -1,0,2,3,2,2,0,0,0,1,0,0,0,0,2,3,2,1,2,2,3,0,0,0,2,3,1,0,0,0,1,1, -0,0,1,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,1,1,0,1,0,0,0,0,0,0,0,0,0, -3,3,2,2,0,1,0,0,0,0,2,0,2,0,1,0,0,0,1,1,0,0,0,2,1,0,1,0,1,1,0,0, -0,1,0,2,0,0,1,0,3,0,1,0,0,0,2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,1,0,0,1,0,0,0,0,0,1,1,2,0,0,0,0,1,0,0,1,3,1,0,0,0,0,1,1,0,0, -0,1,0,0,0,0,3,0,0,0,0,0,0,3,0,0,0,0,0,0,0,3,0,0,0,0,0,0,0,0,0,0, -3,3,1,1,1,1,2,3,0,0,2,1,1,1,1,1,0,2,1,1,0,0,0,2,1,0,1,2,1,1,0,1, -2,1,0,3,0,0,0,0,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -1,3,1,0,0,0,0,0,0,0,3,0,0,0,3,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,1, -0,0,0,2,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,2,0,0,0,0,0,0,1,2,1,0,1,1,0,2,0,0,1,0,0,2,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,2,0,0,0,1,3,0,1,0,0,0,2,0,0,0,0,0,0,0,1,2,0,0,0,0,0, -3,3,0,0,1,1,2,0,0,1,2,1,0,1,1,1,0,1,1,0,0,2,1,1,0,1,0,0,1,1,1,0, -0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,3,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,2,2,1,0,0,0,0,1,0,0,0,0,3,0,0,0,0,0,0,0,0,0,3,0,0,0,0,0,0,0,0, -2,0,0,0,0,0,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,3,0,0,1,1,0,0,0,2,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -1,1,0,1,2,0,1,2,0,0,1,1,0,2,0,1,0,0,1,0,0,0,0,1,0,0,0,2,0,0,0,0, -1,0,0,1,0,1,1,0,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,1,0,0,0,0,0,0,0,1,1,0,1,1,0,2,1,3,0,0,0,0,1,1,0,0,0,0,0,0,0,3, -1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,0,1,0,1,0,0,2,0,0,2,0,0,1,1,2,0,0,1,1,0,0,0,1,0,0,0,1,1,0,0,0, -1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0, -1,0,0,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,1, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,1,1,0,0,0, -2,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,3,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,0,0,0,0,2,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,3,0,0,0, -2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,1,0,0,0,0, -1,0,0,0,0,0,0,0,0,1,0,0,0,0,2,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,1,1,0,0,2,1,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -) - -TIS620ThaiModel = { - 'char_to_order_map': TIS620CharToOrderMap, - 'precedence_matrix': ThaiLangModel, - 'typical_positive_ratio': 0.926386, - 'keep_english_letter': False, - 'charset_name': "TIS-620", - 'language': 'Thai', -} diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/langturkishmodel.py b/src/fetchcode/vcs/pip/_vendor/chardet/langturkishmodel.py deleted file mode 100644 index a427a457..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/langturkishmodel.py +++ /dev/null @@ -1,193 +0,0 @@ -# -*- coding: utf-8 -*- -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Communicator client code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# Özgür Baskın - Turkish Language Model -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -# 255: Control characters that usually does not exist in any text -# 254: Carriage/Return -# 253: symbol (punctuation) that does not belong to word -# 252: 0 - 9 - -# Character Mapping Table: -Latin5_TurkishCharToOrderMap = ( -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, -255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, -255, 23, 37, 47, 39, 29, 52, 36, 45, 53, 60, 16, 49, 20, 46, 42, - 48, 69, 44, 35, 31, 51, 38, 62, 65, 43, 56,255,255,255,255,255, -255, 1, 21, 28, 12, 2, 18, 27, 25, 3, 24, 10, 5, 13, 4, 15, - 26, 64, 7, 8, 9, 14, 32, 57, 58, 11, 22,255,255,255,255,255, -180,179,178,177,176,175,174,173,172,171,170,169,168,167,166,165, -164,163,162,161,160,159,101,158,157,156,155,154,153,152,151,106, -150,149,148,147,146,145,144,100,143,142,141,140,139,138,137,136, - 94, 80, 93,135,105,134,133, 63,132,131,130,129,128,127,126,125, -124,104, 73, 99, 79, 85,123, 54,122, 98, 92,121,120, 91,103,119, - 68,118,117, 97,116,115, 50, 90,114,113,112,111, 55, 41, 40, 86, - 89, 70, 59, 78, 71, 82, 88, 33, 77, 66, 84, 83,110, 75, 61, 96, - 30, 67,109, 74, 87,102, 34, 95, 81,108, 76, 72, 17, 6, 19,107, -) - -TurkishLangModel = ( -3,2,3,3,3,1,3,3,3,3,3,3,3,3,2,1,1,3,3,1,3,3,0,3,3,3,3,3,0,3,1,3, -3,2,1,0,0,1,1,0,0,0,1,0,0,1,1,1,1,0,0,0,0,0,0,0,2,2,0,0,1,0,0,1, -3,2,2,3,3,0,3,3,3,3,3,3,3,2,3,1,0,3,3,1,3,3,0,3,3,3,3,3,0,3,0,3, -3,1,1,0,1,0,1,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,2,2,0,0,0,1,0,1, -3,3,2,3,3,0,3,3,3,3,3,3,3,2,3,1,1,3,3,0,3,3,1,2,3,3,3,3,0,3,0,3, -3,1,1,0,0,0,1,0,0,0,0,1,1,0,1,2,1,0,0,0,1,0,0,0,0,2,0,0,0,0,0,1, -3,3,3,3,3,3,2,3,3,3,3,3,3,3,3,1,3,3,2,0,3,2,1,2,2,1,3,3,0,0,0,2, -2,2,0,1,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,1,0,1,0,0,1, -3,3,3,2,3,3,1,2,3,3,3,3,3,3,3,1,3,2,1,0,3,2,0,1,2,3,3,2,1,0,0,2, -2,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,2,0,2,0,0,0, -1,0,1,3,3,1,3,3,3,3,3,3,3,1,2,0,0,2,3,0,2,3,0,0,2,2,2,3,0,3,0,1, -2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,0,3,3,3,0,3,2,0,2,3,2,3,3,1,0,0,2, -3,2,0,0,1,0,0,0,0,0,0,2,0,0,1,0,0,0,0,0,0,0,0,0,1,1,1,0,2,0,0,1, -3,3,3,2,3,3,2,3,3,3,3,2,3,3,3,0,3,3,0,0,2,1,0,0,2,3,2,2,0,0,0,2, -2,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,1,0,1,0,2,0,0,1, -3,3,3,2,3,3,3,3,3,3,3,2,3,3,3,0,3,2,0,1,3,2,1,1,3,2,3,2,1,0,0,2, -2,2,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0, -3,3,3,2,3,3,3,3,3,3,3,2,3,3,3,0,3,2,2,0,2,3,0,0,2,2,2,2,0,0,0,2, -3,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,2,0,1,0,0,0, -3,3,3,3,3,3,3,2,2,2,2,3,2,3,3,0,3,3,1,1,2,2,0,0,2,2,3,2,0,0,1,3, -0,3,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,1, -3,3,3,2,3,3,3,2,1,2,2,3,2,3,3,0,3,2,0,0,1,1,0,1,1,2,1,2,0,0,0,1, -0,3,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,1,0,0,0, -3,3,3,2,3,3,2,3,2,2,2,3,3,3,3,1,3,1,1,0,3,2,1,1,3,3,2,3,1,0,0,1, -1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,2,0,0,1, -3,2,2,3,3,0,3,3,3,3,3,3,3,2,2,1,0,3,3,1,3,3,0,1,3,3,2,3,0,3,0,3, -2,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0, -2,2,2,3,3,0,3,3,3,3,3,3,3,3,3,0,0,3,2,0,3,3,0,3,2,3,3,3,0,3,1,3, -2,0,0,0,0,0,0,0,0,0,0,1,0,1,2,0,1,0,0,0,0,0,0,0,2,2,0,0,1,0,0,1, -3,3,3,1,2,3,3,1,0,0,1,0,0,3,3,2,3,0,0,2,0,0,2,0,2,0,0,0,2,0,2,0, -0,3,1,0,1,0,0,0,2,2,1,0,1,1,2,1,2,2,2,0,2,1,1,0,0,0,2,0,0,0,0,0, -1,2,1,3,3,0,3,3,3,3,3,2,3,0,0,0,0,2,3,0,2,3,1,0,2,3,1,3,0,3,0,2, -3,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,1,3,3,2,2,3,2,2,0,1,2,3,0,1,2,1,0,1,0,0,0,1,0,2,2,0,0,0,1, -1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,1,0,0,0, -3,3,3,1,3,3,1,1,3,3,1,1,3,3,1,0,2,1,2,0,2,1,0,0,1,1,2,1,0,0,0,2, -2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,1,0,2,1,3,0,0,2,0,0,3,3,0,3,0,0,1,0,1,2,0,0,1,1,2,2,0,1,0, -0,1,2,1,1,0,1,0,1,1,1,1,1,0,1,1,1,2,2,1,2,0,1,0,0,0,0,0,0,1,0,0, -3,3,3,2,3,2,3,3,0,2,2,2,3,3,3,0,3,0,0,0,2,2,0,1,2,1,1,1,0,0,0,1, -0,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0, -3,3,3,3,3,3,2,1,2,2,3,3,3,3,2,0,2,0,0,0,2,2,0,0,2,1,3,3,0,0,1,1, -1,1,0,0,1,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0, -1,1,2,3,3,0,3,3,3,3,3,3,2,2,0,2,0,2,3,2,3,2,2,2,2,2,2,2,1,3,2,3, -2,0,2,1,2,2,2,2,1,1,2,2,1,2,2,1,2,0,0,2,1,1,0,2,1,0,0,1,0,0,0,1, -2,3,3,1,1,1,0,1,1,1,2,3,2,1,1,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0, -0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,2,2,2,3,2,3,2,2,1,3,3,3,0,2,1,2,0,2,1,0,0,1,1,1,1,1,0,0,1, -2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,1,0,2,0,1,0,0,0, -3,3,3,2,3,3,3,3,3,2,3,1,2,3,3,1,2,0,0,0,0,0,0,0,3,2,1,1,0,0,0,0, -2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0, -3,3,3,2,2,3,3,2,1,1,1,1,1,3,3,0,3,1,0,0,1,1,0,0,3,1,2,1,0,0,0,0, -0,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0, -3,3,3,2,2,3,2,2,2,3,2,1,1,3,3,0,3,0,0,0,0,1,0,0,3,1,1,2,0,0,0,1, -1,0,0,1,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1, -1,1,1,3,3,0,3,3,3,3,3,2,2,2,1,2,0,2,1,2,2,1,1,0,1,2,2,2,2,2,2,2, -0,0,2,1,2,1,2,1,0,1,1,3,1,2,1,1,2,0,0,2,0,1,0,1,0,1,0,0,0,1,0,1, -3,3,3,1,3,3,3,0,1,1,0,2,2,3,1,0,3,0,0,0,1,0,0,0,1,0,0,1,0,1,0,0, -1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,2,0,0,2,2,1,0,0,1,0,0,3,3,1,3,0,0,1,1,0,2,0,3,0,0,0,2,0,1,1, -0,1,2,0,1,2,2,0,2,2,2,2,1,0,2,1,1,0,2,0,2,1,2,0,0,0,0,0,0,0,0,0, -3,3,3,1,3,2,3,2,0,2,2,2,1,3,2,0,2,1,2,0,1,2,0,0,1,0,2,2,0,0,0,2, -1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,0,0,0, -3,3,3,0,3,3,1,1,2,3,1,0,3,2,3,0,3,0,0,0,1,0,0,0,1,0,1,0,0,0,0,0, -1,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,3,3,0,3,3,2,3,3,2,2,0,0,0,0,1,2,0,1,3,0,0,0,3,1,1,0,3,0,2, -2,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,1,2,2,1,0,3,1,1,1,1,3,3,2,3,0,0,1,0,1,2,0,2,2,0,2,2,0,2,1, -0,2,2,1,1,1,1,0,2,1,1,0,1,1,1,1,2,1,2,1,2,0,1,0,1,0,0,0,0,0,0,0, -3,3,3,0,1,1,3,0,0,1,1,0,0,2,2,0,3,0,0,1,1,0,1,0,0,0,0,0,2,0,0,0, -0,3,1,0,1,0,1,0,2,0,0,1,0,1,0,1,1,1,2,1,1,0,2,0,0,0,0,0,0,0,0,0, -3,3,3,0,2,0,2,0,1,1,1,0,0,3,3,0,2,0,0,1,0,0,2,1,1,0,1,0,1,0,1,0, -0,2,0,1,2,0,2,0,2,1,1,0,1,0,2,1,1,0,2,1,1,0,1,0,0,0,1,1,0,0,0,0, -3,2,3,0,1,0,0,0,0,0,0,0,0,1,2,0,1,0,0,1,0,0,1,0,0,0,0,0,2,0,0,0, -0,0,1,1,0,0,1,0,1,0,0,1,0,0,0,2,1,0,1,0,2,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,0,0,2,3,0,0,1,0,1,0,2,3,2,3,0,0,1,3,0,2,1,0,0,0,0,2,0,1,0, -0,2,1,0,0,1,1,0,2,1,0,0,1,0,0,1,1,0,1,1,2,0,1,0,0,0,0,1,0,0,0,0, -3,2,2,0,0,1,1,0,0,0,0,0,0,3,1,1,1,0,0,0,0,0,1,0,0,0,0,0,2,0,1,0, -0,1,0,0,0,0,0,0,1,0,0,0,0,0,0,0,1,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0, -0,0,0,3,3,0,2,3,2,2,1,2,2,1,1,2,0,1,3,2,2,2,0,0,2,2,0,0,0,1,2,1, -3,0,2,1,1,0,1,1,1,0,1,2,2,2,1,1,2,0,0,0,0,1,0,1,1,0,0,0,0,0,0,0, -0,1,1,2,3,0,3,3,3,2,2,2,2,1,0,1,0,1,0,1,2,2,0,0,2,2,1,3,1,1,2,1, -0,0,1,1,2,0,1,1,0,0,1,2,0,2,1,1,2,0,0,1,0,0,0,1,0,1,0,1,0,0,0,0, -3,3,2,0,0,3,1,0,0,0,0,0,0,3,2,1,2,0,0,1,0,0,2,0,0,0,0,0,2,0,1,0, -0,2,1,1,0,0,1,0,1,2,0,0,1,1,0,0,2,1,1,1,1,0,2,0,0,0,0,0,0,0,0,0, -3,3,2,0,0,1,0,0,0,0,1,0,0,3,3,2,2,0,0,1,0,0,2,0,1,0,0,0,2,0,1,0, -0,0,1,1,0,0,2,0,2,1,0,0,1,1,2,1,2,0,2,1,2,1,1,1,0,0,1,1,0,0,0,0, -3,3,2,0,0,2,2,0,0,0,1,1,0,2,2,1,3,1,0,1,0,1,2,0,0,0,0,0,1,0,1,0, -0,1,1,0,0,0,0,0,1,0,0,1,0,0,0,1,1,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,2,0,0,0,1,0,0,1,0,0,2,3,1,2,0,0,1,0,0,2,0,0,0,1,0,2,0,2,0, -0,1,1,2,2,1,2,0,2,1,1,0,0,1,1,0,1,1,1,1,2,1,1,0,0,0,0,0,0,0,0,0, -3,3,3,0,2,1,2,1,0,0,1,1,0,3,3,1,2,0,0,1,0,0,2,0,2,0,1,1,2,0,0,0, -0,0,1,1,1,1,2,0,1,1,0,1,1,1,1,0,0,0,1,1,1,0,1,0,0,0,1,0,0,0,0,0, -3,3,3,0,2,2,3,2,0,0,1,0,0,2,3,1,0,0,0,0,0,0,2,0,2,0,0,0,2,0,0,0, -0,1,1,0,0,0,1,0,0,1,0,1,1,0,1,0,1,1,1,0,1,0,0,0,0,0,0,0,0,0,0,0, -3,2,3,0,0,0,0,0,0,0,1,0,0,2,2,2,2,0,0,1,0,0,2,0,0,0,0,0,2,0,1,0, -0,0,2,1,1,0,1,0,2,1,1,0,0,1,1,2,1,0,2,0,2,0,1,0,0,0,2,0,0,0,0,0, -0,0,0,2,2,0,2,1,1,1,1,2,2,0,0,1,0,1,0,0,1,3,0,0,0,0,1,0,0,2,1,0, -0,0,1,0,1,0,0,0,0,0,2,1,0,1,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0, -2,0,0,2,3,0,2,3,1,2,2,0,2,0,0,2,0,2,1,1,1,2,1,0,0,1,2,1,1,2,1,0, -1,0,2,0,1,0,1,1,0,0,2,2,1,2,1,1,2,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0, -3,3,3,0,2,1,2,0,0,0,1,0,0,3,2,0,1,0,0,1,0,0,2,0,0,0,1,2,1,0,1,0, -0,0,0,0,1,0,1,0,0,1,0,0,0,0,1,0,1,0,1,1,1,0,1,0,0,0,0,0,0,0,0,0, -0,0,0,2,2,0,2,2,1,1,0,1,1,1,1,1,0,0,1,2,1,1,1,0,1,0,0,0,1,1,1,1, -0,0,2,1,0,1,1,1,0,1,1,2,1,2,1,1,2,0,1,1,2,1,0,2,0,0,0,0,0,0,0,0, -3,2,2,0,0,2,0,0,0,0,0,0,0,2,2,0,2,0,0,1,0,0,2,0,0,0,0,0,2,0,0,0, -0,2,1,0,0,0,0,0,1,0,0,0,0,0,0,0,1,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0, -0,0,0,3,2,0,2,2,0,1,1,0,1,0,0,1,0,0,0,1,0,1,0,0,0,0,0,1,0,0,0,0, -2,0,1,0,1,0,1,1,0,0,1,2,0,1,0,1,1,0,0,1,0,1,0,2,0,0,0,0,0,0,0,0, -2,2,2,0,1,1,0,0,0,1,0,0,0,1,2,0,1,0,0,1,0,0,1,0,0,0,0,1,2,0,1,0, -0,0,1,0,0,0,1,0,0,1,0,0,0,0,0,0,1,0,1,0,2,0,0,0,0,0,0,0,0,0,0,0, -2,2,2,2,1,0,1,1,1,0,0,0,0,1,2,0,0,1,0,0,0,1,0,0,1,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0, -1,1,2,0,1,0,0,0,1,0,1,0,0,0,1,0,0,1,0,0,0,0,0,0,0,1,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,2,0,0,0,0,0,1, -0,0,1,2,2,0,2,1,2,1,1,2,2,0,0,0,0,1,0,0,1,1,0,0,2,0,0,0,0,1,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0, -2,2,2,0,0,0,1,0,0,0,0,0,0,2,2,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0, -0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,1,1,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -2,2,2,0,1,0,1,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,1,0,0,0,0,0,0,0,0,0,0,2,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, -) - -Latin5TurkishModel = { - 'char_to_order_map': Latin5_TurkishCharToOrderMap, - 'precedence_matrix': TurkishLangModel, - 'typical_positive_ratio': 0.970290, - 'keep_english_letter': True, - 'charset_name': "ISO-8859-9", - 'language': 'Turkish', -} diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/latin1prober.py b/src/fetchcode/vcs/pip/_vendor/chardet/latin1prober.py deleted file mode 100644 index 7d1e8c20..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/latin1prober.py +++ /dev/null @@ -1,145 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Universal charset detector code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 2001 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# Shy Shalom - original C code -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -from .charsetprober import CharSetProber -from .enums import ProbingState - -FREQ_CAT_NUM = 4 - -UDF = 0 # undefined -OTH = 1 # other -ASC = 2 # ascii capital letter -ASS = 3 # ascii small letter -ACV = 4 # accent capital vowel -ACO = 5 # accent capital other -ASV = 6 # accent small vowel -ASO = 7 # accent small other -CLASS_NUM = 8 # total classes - -Latin1_CharToClass = ( - OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH, # 00 - 07 - OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH, # 08 - 0F - OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH, # 10 - 17 - OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH, # 18 - 1F - OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH, # 20 - 27 - OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH, # 28 - 2F - OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH, # 30 - 37 - OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH, # 38 - 3F - OTH, ASC, ASC, ASC, ASC, ASC, ASC, ASC, # 40 - 47 - ASC, ASC, ASC, ASC, ASC, ASC, ASC, ASC, # 48 - 4F - ASC, ASC, ASC, ASC, ASC, ASC, ASC, ASC, # 50 - 57 - ASC, ASC, ASC, OTH, OTH, OTH, OTH, OTH, # 58 - 5F - OTH, ASS, ASS, ASS, ASS, ASS, ASS, ASS, # 60 - 67 - ASS, ASS, ASS, ASS, ASS, ASS, ASS, ASS, # 68 - 6F - ASS, ASS, ASS, ASS, ASS, ASS, ASS, ASS, # 70 - 77 - ASS, ASS, ASS, OTH, OTH, OTH, OTH, OTH, # 78 - 7F - OTH, UDF, OTH, ASO, OTH, OTH, OTH, OTH, # 80 - 87 - OTH, OTH, ACO, OTH, ACO, UDF, ACO, UDF, # 88 - 8F - UDF, OTH, OTH, OTH, OTH, OTH, OTH, OTH, # 90 - 97 - OTH, OTH, ASO, OTH, ASO, UDF, ASO, ACO, # 98 - 9F - OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH, # A0 - A7 - OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH, # A8 - AF - OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH, # B0 - B7 - OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH, # B8 - BF - ACV, ACV, ACV, ACV, ACV, ACV, ACO, ACO, # C0 - C7 - ACV, ACV, ACV, ACV, ACV, ACV, ACV, ACV, # C8 - CF - ACO, ACO, ACV, ACV, ACV, ACV, ACV, OTH, # D0 - D7 - ACV, ACV, ACV, ACV, ACV, ACO, ACO, ACO, # D8 - DF - ASV, ASV, ASV, ASV, ASV, ASV, ASO, ASO, # E0 - E7 - ASV, ASV, ASV, ASV, ASV, ASV, ASV, ASV, # E8 - EF - ASO, ASO, ASV, ASV, ASV, ASV, ASV, OTH, # F0 - F7 - ASV, ASV, ASV, ASV, ASV, ASO, ASO, ASO, # F8 - FF -) - -# 0 : illegal -# 1 : very unlikely -# 2 : normal -# 3 : very likely -Latin1ClassModel = ( -# UDF OTH ASC ASS ACV ACO ASV ASO - 0, 0, 0, 0, 0, 0, 0, 0, # UDF - 0, 3, 3, 3, 3, 3, 3, 3, # OTH - 0, 3, 3, 3, 3, 3, 3, 3, # ASC - 0, 3, 3, 3, 1, 1, 3, 3, # ASS - 0, 3, 3, 3, 1, 2, 1, 2, # ACV - 0, 3, 3, 3, 3, 3, 3, 3, # ACO - 0, 3, 1, 3, 1, 1, 1, 3, # ASV - 0, 3, 1, 3, 1, 1, 3, 3, # ASO -) - - -class Latin1Prober(CharSetProber): - def __init__(self): - super(Latin1Prober, self).__init__() - self._last_char_class = None - self._freq_counter = None - self.reset() - - def reset(self): - self._last_char_class = OTH - self._freq_counter = [0] * FREQ_CAT_NUM - CharSetProber.reset(self) - - @property - def charset_name(self): - return "ISO-8859-1" - - @property - def language(self): - return "" - - def feed(self, byte_str): - byte_str = self.filter_with_english_letters(byte_str) - for c in byte_str: - char_class = Latin1_CharToClass[c] - freq = Latin1ClassModel[(self._last_char_class * CLASS_NUM) - + char_class] - if freq == 0: - self._state = ProbingState.NOT_ME - break - self._freq_counter[freq] += 1 - self._last_char_class = char_class - - return self.state - - def get_confidence(self): - if self.state == ProbingState.NOT_ME: - return 0.01 - - total = sum(self._freq_counter) - if total < 0.01: - confidence = 0.0 - else: - confidence = ((self._freq_counter[3] - self._freq_counter[1] * 20.0) - / total) - if confidence < 0.0: - confidence = 0.0 - # lower the confidence of latin1 so that other more accurate - # detector can take priority. - confidence = confidence * 0.73 - return confidence diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/mbcharsetprober.py b/src/fetchcode/vcs/pip/_vendor/chardet/mbcharsetprober.py deleted file mode 100644 index 6256ecfd..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/mbcharsetprober.py +++ /dev/null @@ -1,91 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Universal charset detector code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 2001 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# Shy Shalom - original C code -# Proofpoint, Inc. -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -from .charsetprober import CharSetProber -from .enums import ProbingState, MachineState - - -class MultiByteCharSetProber(CharSetProber): - """ - MultiByteCharSetProber - """ - - def __init__(self, lang_filter=None): - super(MultiByteCharSetProber, self).__init__(lang_filter=lang_filter) - self.distribution_analyzer = None - self.coding_sm = None - self._last_char = [0, 0] - - def reset(self): - super(MultiByteCharSetProber, self).reset() - if self.coding_sm: - self.coding_sm.reset() - if self.distribution_analyzer: - self.distribution_analyzer.reset() - self._last_char = [0, 0] - - @property - def charset_name(self): - raise NotImplementedError - - @property - def language(self): - raise NotImplementedError - - def feed(self, byte_str): - for i in range(len(byte_str)): - coding_state = self.coding_sm.next_state(byte_str[i]) - if coding_state == MachineState.ERROR: - self.logger.debug('%s %s prober hit error at byte %s', - self.charset_name, self.language, i) - self._state = ProbingState.NOT_ME - break - elif coding_state == MachineState.ITS_ME: - self._state = ProbingState.FOUND_IT - break - elif coding_state == MachineState.START: - char_len = self.coding_sm.get_current_charlen() - if i == 0: - self._last_char[1] = byte_str[0] - self.distribution_analyzer.feed(self._last_char, char_len) - else: - self.distribution_analyzer.feed(byte_str[i - 1:i + 1], - char_len) - - self._last_char[0] = byte_str[-1] - - if self.state == ProbingState.DETECTING: - if (self.distribution_analyzer.got_enough_data() and - (self.get_confidence() > self.SHORTCUT_THRESHOLD)): - self._state = ProbingState.FOUND_IT - - return self.state - - def get_confidence(self): - return self.distribution_analyzer.get_confidence() diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/mbcsgroupprober.py b/src/fetchcode/vcs/pip/_vendor/chardet/mbcsgroupprober.py deleted file mode 100644 index 530abe75..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/mbcsgroupprober.py +++ /dev/null @@ -1,54 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Universal charset detector code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 2001 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# Shy Shalom - original C code -# Proofpoint, Inc. -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -from .charsetgroupprober import CharSetGroupProber -from .utf8prober import UTF8Prober -from .sjisprober import SJISProber -from .eucjpprober import EUCJPProber -from .gb2312prober import GB2312Prober -from .euckrprober import EUCKRProber -from .cp949prober import CP949Prober -from .big5prober import Big5Prober -from .euctwprober import EUCTWProber - - -class MBCSGroupProber(CharSetGroupProber): - def __init__(self, lang_filter=None): - super(MBCSGroupProber, self).__init__(lang_filter=lang_filter) - self.probers = [ - UTF8Prober(), - SJISProber(), - EUCJPProber(), - GB2312Prober(), - EUCKRProber(), - CP949Prober(), - Big5Prober(), - EUCTWProber() - ] - self.reset() diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/mbcssm.py b/src/fetchcode/vcs/pip/_vendor/chardet/mbcssm.py deleted file mode 100644 index 8360d0f2..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/mbcssm.py +++ /dev/null @@ -1,572 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is mozilla.org code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -from .enums import MachineState - -# BIG5 - -BIG5_CLS = ( - 1,1,1,1,1,1,1,1, # 00 - 07 #allow 0x00 as legal value - 1,1,1,1,1,1,0,0, # 08 - 0f - 1,1,1,1,1,1,1,1, # 10 - 17 - 1,1,1,0,1,1,1,1, # 18 - 1f - 1,1,1,1,1,1,1,1, # 20 - 27 - 1,1,1,1,1,1,1,1, # 28 - 2f - 1,1,1,1,1,1,1,1, # 30 - 37 - 1,1,1,1,1,1,1,1, # 38 - 3f - 2,2,2,2,2,2,2,2, # 40 - 47 - 2,2,2,2,2,2,2,2, # 48 - 4f - 2,2,2,2,2,2,2,2, # 50 - 57 - 2,2,2,2,2,2,2,2, # 58 - 5f - 2,2,2,2,2,2,2,2, # 60 - 67 - 2,2,2,2,2,2,2,2, # 68 - 6f - 2,2,2,2,2,2,2,2, # 70 - 77 - 2,2,2,2,2,2,2,1, # 78 - 7f - 4,4,4,4,4,4,4,4, # 80 - 87 - 4,4,4,4,4,4,4,4, # 88 - 8f - 4,4,4,4,4,4,4,4, # 90 - 97 - 4,4,4,4,4,4,4,4, # 98 - 9f - 4,3,3,3,3,3,3,3, # a0 - a7 - 3,3,3,3,3,3,3,3, # a8 - af - 3,3,3,3,3,3,3,3, # b0 - b7 - 3,3,3,3,3,3,3,3, # b8 - bf - 3,3,3,3,3,3,3,3, # c0 - c7 - 3,3,3,3,3,3,3,3, # c8 - cf - 3,3,3,3,3,3,3,3, # d0 - d7 - 3,3,3,3,3,3,3,3, # d8 - df - 3,3,3,3,3,3,3,3, # e0 - e7 - 3,3,3,3,3,3,3,3, # e8 - ef - 3,3,3,3,3,3,3,3, # f0 - f7 - 3,3,3,3,3,3,3,0 # f8 - ff -) - -BIG5_ST = ( - MachineState.ERROR,MachineState.START,MachineState.START, 3,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#00-07 - MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ERROR,#08-0f - MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START#10-17 -) - -BIG5_CHAR_LEN_TABLE = (0, 1, 1, 2, 0) - -BIG5_SM_MODEL = {'class_table': BIG5_CLS, - 'class_factor': 5, - 'state_table': BIG5_ST, - 'char_len_table': BIG5_CHAR_LEN_TABLE, - 'name': 'Big5'} - -# CP949 - -CP949_CLS = ( - 1,1,1,1,1,1,1,1, 1,1,1,1,1,1,0,0, # 00 - 0f - 1,1,1,1,1,1,1,1, 1,1,1,0,1,1,1,1, # 10 - 1f - 1,1,1,1,1,1,1,1, 1,1,1,1,1,1,1,1, # 20 - 2f - 1,1,1,1,1,1,1,1, 1,1,1,1,1,1,1,1, # 30 - 3f - 1,4,4,4,4,4,4,4, 4,4,4,4,4,4,4,4, # 40 - 4f - 4,4,5,5,5,5,5,5, 5,5,5,1,1,1,1,1, # 50 - 5f - 1,5,5,5,5,5,5,5, 5,5,5,5,5,5,5,5, # 60 - 6f - 5,5,5,5,5,5,5,5, 5,5,5,1,1,1,1,1, # 70 - 7f - 0,6,6,6,6,6,6,6, 6,6,6,6,6,6,6,6, # 80 - 8f - 6,6,6,6,6,6,6,6, 6,6,6,6,6,6,6,6, # 90 - 9f - 6,7,7,7,7,7,7,7, 7,7,7,7,7,8,8,8, # a0 - af - 7,7,7,7,7,7,7,7, 7,7,7,7,7,7,7,7, # b0 - bf - 7,7,7,7,7,7,9,2, 2,3,2,2,2,2,2,2, # c0 - cf - 2,2,2,2,2,2,2,2, 2,2,2,2,2,2,2,2, # d0 - df - 2,2,2,2,2,2,2,2, 2,2,2,2,2,2,2,2, # e0 - ef - 2,2,2,2,2,2,2,2, 2,2,2,2,2,2,2,0, # f0 - ff -) - -CP949_ST = ( -#cls= 0 1 2 3 4 5 6 7 8 9 # previous state = - MachineState.ERROR,MachineState.START, 3,MachineState.ERROR,MachineState.START,MachineState.START, 4, 5,MachineState.ERROR, 6, # MachineState.START - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR, # MachineState.ERROR - MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME, # MachineState.ITS_ME - MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START, # 3 - MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START, # 4 - MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START, # 5 - MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START, # 6 -) - -CP949_CHAR_LEN_TABLE = (0, 1, 2, 0, 1, 1, 2, 2, 0, 2) - -CP949_SM_MODEL = {'class_table': CP949_CLS, - 'class_factor': 10, - 'state_table': CP949_ST, - 'char_len_table': CP949_CHAR_LEN_TABLE, - 'name': 'CP949'} - -# EUC-JP - -EUCJP_CLS = ( - 4,4,4,4,4,4,4,4, # 00 - 07 - 4,4,4,4,4,4,5,5, # 08 - 0f - 4,4,4,4,4,4,4,4, # 10 - 17 - 4,4,4,5,4,4,4,4, # 18 - 1f - 4,4,4,4,4,4,4,4, # 20 - 27 - 4,4,4,4,4,4,4,4, # 28 - 2f - 4,4,4,4,4,4,4,4, # 30 - 37 - 4,4,4,4,4,4,4,4, # 38 - 3f - 4,4,4,4,4,4,4,4, # 40 - 47 - 4,4,4,4,4,4,4,4, # 48 - 4f - 4,4,4,4,4,4,4,4, # 50 - 57 - 4,4,4,4,4,4,4,4, # 58 - 5f - 4,4,4,4,4,4,4,4, # 60 - 67 - 4,4,4,4,4,4,4,4, # 68 - 6f - 4,4,4,4,4,4,4,4, # 70 - 77 - 4,4,4,4,4,4,4,4, # 78 - 7f - 5,5,5,5,5,5,5,5, # 80 - 87 - 5,5,5,5,5,5,1,3, # 88 - 8f - 5,5,5,5,5,5,5,5, # 90 - 97 - 5,5,5,5,5,5,5,5, # 98 - 9f - 5,2,2,2,2,2,2,2, # a0 - a7 - 2,2,2,2,2,2,2,2, # a8 - af - 2,2,2,2,2,2,2,2, # b0 - b7 - 2,2,2,2,2,2,2,2, # b8 - bf - 2,2,2,2,2,2,2,2, # c0 - c7 - 2,2,2,2,2,2,2,2, # c8 - cf - 2,2,2,2,2,2,2,2, # d0 - d7 - 2,2,2,2,2,2,2,2, # d8 - df - 0,0,0,0,0,0,0,0, # e0 - e7 - 0,0,0,0,0,0,0,0, # e8 - ef - 0,0,0,0,0,0,0,0, # f0 - f7 - 0,0,0,0,0,0,0,5 # f8 - ff -) - -EUCJP_ST = ( - 3, 4, 3, 5,MachineState.START,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#00-07 - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,#08-0f - MachineState.ITS_ME,MachineState.ITS_ME,MachineState.START,MachineState.ERROR,MachineState.START,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#10-17 - MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR, 3,MachineState.ERROR,#18-1f - 3,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START#20-27 -) - -EUCJP_CHAR_LEN_TABLE = (2, 2, 2, 3, 1, 0) - -EUCJP_SM_MODEL = {'class_table': EUCJP_CLS, - 'class_factor': 6, - 'state_table': EUCJP_ST, - 'char_len_table': EUCJP_CHAR_LEN_TABLE, - 'name': 'EUC-JP'} - -# EUC-KR - -EUCKR_CLS = ( - 1,1,1,1,1,1,1,1, # 00 - 07 - 1,1,1,1,1,1,0,0, # 08 - 0f - 1,1,1,1,1,1,1,1, # 10 - 17 - 1,1,1,0,1,1,1,1, # 18 - 1f - 1,1,1,1,1,1,1,1, # 20 - 27 - 1,1,1,1,1,1,1,1, # 28 - 2f - 1,1,1,1,1,1,1,1, # 30 - 37 - 1,1,1,1,1,1,1,1, # 38 - 3f - 1,1,1,1,1,1,1,1, # 40 - 47 - 1,1,1,1,1,1,1,1, # 48 - 4f - 1,1,1,1,1,1,1,1, # 50 - 57 - 1,1,1,1,1,1,1,1, # 58 - 5f - 1,1,1,1,1,1,1,1, # 60 - 67 - 1,1,1,1,1,1,1,1, # 68 - 6f - 1,1,1,1,1,1,1,1, # 70 - 77 - 1,1,1,1,1,1,1,1, # 78 - 7f - 0,0,0,0,0,0,0,0, # 80 - 87 - 0,0,0,0,0,0,0,0, # 88 - 8f - 0,0,0,0,0,0,0,0, # 90 - 97 - 0,0,0,0,0,0,0,0, # 98 - 9f - 0,2,2,2,2,2,2,2, # a0 - a7 - 2,2,2,2,2,3,3,3, # a8 - af - 2,2,2,2,2,2,2,2, # b0 - b7 - 2,2,2,2,2,2,2,2, # b8 - bf - 2,2,2,2,2,2,2,2, # c0 - c7 - 2,3,2,2,2,2,2,2, # c8 - cf - 2,2,2,2,2,2,2,2, # d0 - d7 - 2,2,2,2,2,2,2,2, # d8 - df - 2,2,2,2,2,2,2,2, # e0 - e7 - 2,2,2,2,2,2,2,2, # e8 - ef - 2,2,2,2,2,2,2,2, # f0 - f7 - 2,2,2,2,2,2,2,0 # f8 - ff -) - -EUCKR_ST = ( - MachineState.ERROR,MachineState.START, 3,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#00-07 - MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START #08-0f -) - -EUCKR_CHAR_LEN_TABLE = (0, 1, 2, 0) - -EUCKR_SM_MODEL = {'class_table': EUCKR_CLS, - 'class_factor': 4, - 'state_table': EUCKR_ST, - 'char_len_table': EUCKR_CHAR_LEN_TABLE, - 'name': 'EUC-KR'} - -# EUC-TW - -EUCTW_CLS = ( - 2,2,2,2,2,2,2,2, # 00 - 07 - 2,2,2,2,2,2,0,0, # 08 - 0f - 2,2,2,2,2,2,2,2, # 10 - 17 - 2,2,2,0,2,2,2,2, # 18 - 1f - 2,2,2,2,2,2,2,2, # 20 - 27 - 2,2,2,2,2,2,2,2, # 28 - 2f - 2,2,2,2,2,2,2,2, # 30 - 37 - 2,2,2,2,2,2,2,2, # 38 - 3f - 2,2,2,2,2,2,2,2, # 40 - 47 - 2,2,2,2,2,2,2,2, # 48 - 4f - 2,2,2,2,2,2,2,2, # 50 - 57 - 2,2,2,2,2,2,2,2, # 58 - 5f - 2,2,2,2,2,2,2,2, # 60 - 67 - 2,2,2,2,2,2,2,2, # 68 - 6f - 2,2,2,2,2,2,2,2, # 70 - 77 - 2,2,2,2,2,2,2,2, # 78 - 7f - 0,0,0,0,0,0,0,0, # 80 - 87 - 0,0,0,0,0,0,6,0, # 88 - 8f - 0,0,0,0,0,0,0,0, # 90 - 97 - 0,0,0,0,0,0,0,0, # 98 - 9f - 0,3,4,4,4,4,4,4, # a0 - a7 - 5,5,1,1,1,1,1,1, # a8 - af - 1,1,1,1,1,1,1,1, # b0 - b7 - 1,1,1,1,1,1,1,1, # b8 - bf - 1,1,3,1,3,3,3,3, # c0 - c7 - 3,3,3,3,3,3,3,3, # c8 - cf - 3,3,3,3,3,3,3,3, # d0 - d7 - 3,3,3,3,3,3,3,3, # d8 - df - 3,3,3,3,3,3,3,3, # e0 - e7 - 3,3,3,3,3,3,3,3, # e8 - ef - 3,3,3,3,3,3,3,3, # f0 - f7 - 3,3,3,3,3,3,3,0 # f8 - ff -) - -EUCTW_ST = ( - MachineState.ERROR,MachineState.ERROR,MachineState.START, 3, 3, 3, 4,MachineState.ERROR,#00-07 - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,#08-0f - MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ERROR,MachineState.START,MachineState.ERROR,#10-17 - MachineState.START,MachineState.START,MachineState.START,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#18-1f - 5,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.ERROR,MachineState.START,MachineState.START,#20-27 - MachineState.START,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START #28-2f -) - -EUCTW_CHAR_LEN_TABLE = (0, 0, 1, 2, 2, 2, 3) - -EUCTW_SM_MODEL = {'class_table': EUCTW_CLS, - 'class_factor': 7, - 'state_table': EUCTW_ST, - 'char_len_table': EUCTW_CHAR_LEN_TABLE, - 'name': 'x-euc-tw'} - -# GB2312 - -GB2312_CLS = ( - 1,1,1,1,1,1,1,1, # 00 - 07 - 1,1,1,1,1,1,0,0, # 08 - 0f - 1,1,1,1,1,1,1,1, # 10 - 17 - 1,1,1,0,1,1,1,1, # 18 - 1f - 1,1,1,1,1,1,1,1, # 20 - 27 - 1,1,1,1,1,1,1,1, # 28 - 2f - 3,3,3,3,3,3,3,3, # 30 - 37 - 3,3,1,1,1,1,1,1, # 38 - 3f - 2,2,2,2,2,2,2,2, # 40 - 47 - 2,2,2,2,2,2,2,2, # 48 - 4f - 2,2,2,2,2,2,2,2, # 50 - 57 - 2,2,2,2,2,2,2,2, # 58 - 5f - 2,2,2,2,2,2,2,2, # 60 - 67 - 2,2,2,2,2,2,2,2, # 68 - 6f - 2,2,2,2,2,2,2,2, # 70 - 77 - 2,2,2,2,2,2,2,4, # 78 - 7f - 5,6,6,6,6,6,6,6, # 80 - 87 - 6,6,6,6,6,6,6,6, # 88 - 8f - 6,6,6,6,6,6,6,6, # 90 - 97 - 6,6,6,6,6,6,6,6, # 98 - 9f - 6,6,6,6,6,6,6,6, # a0 - a7 - 6,6,6,6,6,6,6,6, # a8 - af - 6,6,6,6,6,6,6,6, # b0 - b7 - 6,6,6,6,6,6,6,6, # b8 - bf - 6,6,6,6,6,6,6,6, # c0 - c7 - 6,6,6,6,6,6,6,6, # c8 - cf - 6,6,6,6,6,6,6,6, # d0 - d7 - 6,6,6,6,6,6,6,6, # d8 - df - 6,6,6,6,6,6,6,6, # e0 - e7 - 6,6,6,6,6,6,6,6, # e8 - ef - 6,6,6,6,6,6,6,6, # f0 - f7 - 6,6,6,6,6,6,6,0 # f8 - ff -) - -GB2312_ST = ( - MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START, 3,MachineState.ERROR,#00-07 - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,#08-0f - MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ERROR,MachineState.ERROR,MachineState.START,#10-17 - 4,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#18-1f - MachineState.ERROR,MachineState.ERROR, 5,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ERROR,#20-27 - MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START #28-2f -) - -# To be accurate, the length of class 6 can be either 2 or 4. -# But it is not necessary to discriminate between the two since -# it is used for frequency analysis only, and we are validating -# each code range there as well. So it is safe to set it to be -# 2 here. -GB2312_CHAR_LEN_TABLE = (0, 1, 1, 1, 1, 1, 2) - -GB2312_SM_MODEL = {'class_table': GB2312_CLS, - 'class_factor': 7, - 'state_table': GB2312_ST, - 'char_len_table': GB2312_CHAR_LEN_TABLE, - 'name': 'GB2312'} - -# Shift_JIS - -SJIS_CLS = ( - 1,1,1,1,1,1,1,1, # 00 - 07 - 1,1,1,1,1,1,0,0, # 08 - 0f - 1,1,1,1,1,1,1,1, # 10 - 17 - 1,1,1,0,1,1,1,1, # 18 - 1f - 1,1,1,1,1,1,1,1, # 20 - 27 - 1,1,1,1,1,1,1,1, # 28 - 2f - 1,1,1,1,1,1,1,1, # 30 - 37 - 1,1,1,1,1,1,1,1, # 38 - 3f - 2,2,2,2,2,2,2,2, # 40 - 47 - 2,2,2,2,2,2,2,2, # 48 - 4f - 2,2,2,2,2,2,2,2, # 50 - 57 - 2,2,2,2,2,2,2,2, # 58 - 5f - 2,2,2,2,2,2,2,2, # 60 - 67 - 2,2,2,2,2,2,2,2, # 68 - 6f - 2,2,2,2,2,2,2,2, # 70 - 77 - 2,2,2,2,2,2,2,1, # 78 - 7f - 3,3,3,3,3,2,2,3, # 80 - 87 - 3,3,3,3,3,3,3,3, # 88 - 8f - 3,3,3,3,3,3,3,3, # 90 - 97 - 3,3,3,3,3,3,3,3, # 98 - 9f - #0xa0 is illegal in sjis encoding, but some pages does - #contain such byte. We need to be more error forgiven. - 2,2,2,2,2,2,2,2, # a0 - a7 - 2,2,2,2,2,2,2,2, # a8 - af - 2,2,2,2,2,2,2,2, # b0 - b7 - 2,2,2,2,2,2,2,2, # b8 - bf - 2,2,2,2,2,2,2,2, # c0 - c7 - 2,2,2,2,2,2,2,2, # c8 - cf - 2,2,2,2,2,2,2,2, # d0 - d7 - 2,2,2,2,2,2,2,2, # d8 - df - 3,3,3,3,3,3,3,3, # e0 - e7 - 3,3,3,3,3,4,4,4, # e8 - ef - 3,3,3,3,3,3,3,3, # f0 - f7 - 3,3,3,3,3,0,0,0) # f8 - ff - - -SJIS_ST = ( - MachineState.ERROR,MachineState.START,MachineState.START, 3,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#00-07 - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,#08-0f - MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START #10-17 -) - -SJIS_CHAR_LEN_TABLE = (0, 1, 1, 2, 0, 0) - -SJIS_SM_MODEL = {'class_table': SJIS_CLS, - 'class_factor': 6, - 'state_table': SJIS_ST, - 'char_len_table': SJIS_CHAR_LEN_TABLE, - 'name': 'Shift_JIS'} - -# UCS2-BE - -UCS2BE_CLS = ( - 0,0,0,0,0,0,0,0, # 00 - 07 - 0,0,1,0,0,2,0,0, # 08 - 0f - 0,0,0,0,0,0,0,0, # 10 - 17 - 0,0,0,3,0,0,0,0, # 18 - 1f - 0,0,0,0,0,0,0,0, # 20 - 27 - 0,3,3,3,3,3,0,0, # 28 - 2f - 0,0,0,0,0,0,0,0, # 30 - 37 - 0,0,0,0,0,0,0,0, # 38 - 3f - 0,0,0,0,0,0,0,0, # 40 - 47 - 0,0,0,0,0,0,0,0, # 48 - 4f - 0,0,0,0,0,0,0,0, # 50 - 57 - 0,0,0,0,0,0,0,0, # 58 - 5f - 0,0,0,0,0,0,0,0, # 60 - 67 - 0,0,0,0,0,0,0,0, # 68 - 6f - 0,0,0,0,0,0,0,0, # 70 - 77 - 0,0,0,0,0,0,0,0, # 78 - 7f - 0,0,0,0,0,0,0,0, # 80 - 87 - 0,0,0,0,0,0,0,0, # 88 - 8f - 0,0,0,0,0,0,0,0, # 90 - 97 - 0,0,0,0,0,0,0,0, # 98 - 9f - 0,0,0,0,0,0,0,0, # a0 - a7 - 0,0,0,0,0,0,0,0, # a8 - af - 0,0,0,0,0,0,0,0, # b0 - b7 - 0,0,0,0,0,0,0,0, # b8 - bf - 0,0,0,0,0,0,0,0, # c0 - c7 - 0,0,0,0,0,0,0,0, # c8 - cf - 0,0,0,0,0,0,0,0, # d0 - d7 - 0,0,0,0,0,0,0,0, # d8 - df - 0,0,0,0,0,0,0,0, # e0 - e7 - 0,0,0,0,0,0,0,0, # e8 - ef - 0,0,0,0,0,0,0,0, # f0 - f7 - 0,0,0,0,0,0,4,5 # f8 - ff -) - -UCS2BE_ST = ( - 5, 7, 7,MachineState.ERROR, 4, 3,MachineState.ERROR,MachineState.ERROR,#00-07 - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,#08-0f - MachineState.ITS_ME,MachineState.ITS_ME, 6, 6, 6, 6,MachineState.ERROR,MachineState.ERROR,#10-17 - 6, 6, 6, 6, 6,MachineState.ITS_ME, 6, 6,#18-1f - 6, 6, 6, 6, 5, 7, 7,MachineState.ERROR,#20-27 - 5, 8, 6, 6,MachineState.ERROR, 6, 6, 6,#28-2f - 6, 6, 6, 6,MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START #30-37 -) - -UCS2BE_CHAR_LEN_TABLE = (2, 2, 2, 0, 2, 2) - -UCS2BE_SM_MODEL = {'class_table': UCS2BE_CLS, - 'class_factor': 6, - 'state_table': UCS2BE_ST, - 'char_len_table': UCS2BE_CHAR_LEN_TABLE, - 'name': 'UTF-16BE'} - -# UCS2-LE - -UCS2LE_CLS = ( - 0,0,0,0,0,0,0,0, # 00 - 07 - 0,0,1,0,0,2,0,0, # 08 - 0f - 0,0,0,0,0,0,0,0, # 10 - 17 - 0,0,0,3,0,0,0,0, # 18 - 1f - 0,0,0,0,0,0,0,0, # 20 - 27 - 0,3,3,3,3,3,0,0, # 28 - 2f - 0,0,0,0,0,0,0,0, # 30 - 37 - 0,0,0,0,0,0,0,0, # 38 - 3f - 0,0,0,0,0,0,0,0, # 40 - 47 - 0,0,0,0,0,0,0,0, # 48 - 4f - 0,0,0,0,0,0,0,0, # 50 - 57 - 0,0,0,0,0,0,0,0, # 58 - 5f - 0,0,0,0,0,0,0,0, # 60 - 67 - 0,0,0,0,0,0,0,0, # 68 - 6f - 0,0,0,0,0,0,0,0, # 70 - 77 - 0,0,0,0,0,0,0,0, # 78 - 7f - 0,0,0,0,0,0,0,0, # 80 - 87 - 0,0,0,0,0,0,0,0, # 88 - 8f - 0,0,0,0,0,0,0,0, # 90 - 97 - 0,0,0,0,0,0,0,0, # 98 - 9f - 0,0,0,0,0,0,0,0, # a0 - a7 - 0,0,0,0,0,0,0,0, # a8 - af - 0,0,0,0,0,0,0,0, # b0 - b7 - 0,0,0,0,0,0,0,0, # b8 - bf - 0,0,0,0,0,0,0,0, # c0 - c7 - 0,0,0,0,0,0,0,0, # c8 - cf - 0,0,0,0,0,0,0,0, # d0 - d7 - 0,0,0,0,0,0,0,0, # d8 - df - 0,0,0,0,0,0,0,0, # e0 - e7 - 0,0,0,0,0,0,0,0, # e8 - ef - 0,0,0,0,0,0,0,0, # f0 - f7 - 0,0,0,0,0,0,4,5 # f8 - ff -) - -UCS2LE_ST = ( - 6, 6, 7, 6, 4, 3,MachineState.ERROR,MachineState.ERROR,#00-07 - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,#08-0f - MachineState.ITS_ME,MachineState.ITS_ME, 5, 5, 5,MachineState.ERROR,MachineState.ITS_ME,MachineState.ERROR,#10-17 - 5, 5, 5,MachineState.ERROR, 5,MachineState.ERROR, 6, 6,#18-1f - 7, 6, 8, 8, 5, 5, 5,MachineState.ERROR,#20-27 - 5, 5, 5,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR, 5, 5,#28-2f - 5, 5, 5,MachineState.ERROR, 5,MachineState.ERROR,MachineState.START,MachineState.START #30-37 -) - -UCS2LE_CHAR_LEN_TABLE = (2, 2, 2, 2, 2, 2) - -UCS2LE_SM_MODEL = {'class_table': UCS2LE_CLS, - 'class_factor': 6, - 'state_table': UCS2LE_ST, - 'char_len_table': UCS2LE_CHAR_LEN_TABLE, - 'name': 'UTF-16LE'} - -# UTF-8 - -UTF8_CLS = ( - 1,1,1,1,1,1,1,1, # 00 - 07 #allow 0x00 as a legal value - 1,1,1,1,1,1,0,0, # 08 - 0f - 1,1,1,1,1,1,1,1, # 10 - 17 - 1,1,1,0,1,1,1,1, # 18 - 1f - 1,1,1,1,1,1,1,1, # 20 - 27 - 1,1,1,1,1,1,1,1, # 28 - 2f - 1,1,1,1,1,1,1,1, # 30 - 37 - 1,1,1,1,1,1,1,1, # 38 - 3f - 1,1,1,1,1,1,1,1, # 40 - 47 - 1,1,1,1,1,1,1,1, # 48 - 4f - 1,1,1,1,1,1,1,1, # 50 - 57 - 1,1,1,1,1,1,1,1, # 58 - 5f - 1,1,1,1,1,1,1,1, # 60 - 67 - 1,1,1,1,1,1,1,1, # 68 - 6f - 1,1,1,1,1,1,1,1, # 70 - 77 - 1,1,1,1,1,1,1,1, # 78 - 7f - 2,2,2,2,3,3,3,3, # 80 - 87 - 4,4,4,4,4,4,4,4, # 88 - 8f - 4,4,4,4,4,4,4,4, # 90 - 97 - 4,4,4,4,4,4,4,4, # 98 - 9f - 5,5,5,5,5,5,5,5, # a0 - a7 - 5,5,5,5,5,5,5,5, # a8 - af - 5,5,5,5,5,5,5,5, # b0 - b7 - 5,5,5,5,5,5,5,5, # b8 - bf - 0,0,6,6,6,6,6,6, # c0 - c7 - 6,6,6,6,6,6,6,6, # c8 - cf - 6,6,6,6,6,6,6,6, # d0 - d7 - 6,6,6,6,6,6,6,6, # d8 - df - 7,8,8,8,8,8,8,8, # e0 - e7 - 8,8,8,8,8,9,8,8, # e8 - ef - 10,11,11,11,11,11,11,11, # f0 - f7 - 12,13,13,13,14,15,0,0 # f8 - ff -) - -UTF8_ST = ( - MachineState.ERROR,MachineState.START,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR, 12, 10,#00-07 - 9, 11, 8, 7, 6, 5, 4, 3,#08-0f - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#10-17 - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#18-1f - MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,#20-27 - MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,#28-2f - MachineState.ERROR,MachineState.ERROR, 5, 5, 5, 5,MachineState.ERROR,MachineState.ERROR,#30-37 - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#38-3f - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR, 5, 5, 5,MachineState.ERROR,MachineState.ERROR,#40-47 - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#48-4f - MachineState.ERROR,MachineState.ERROR, 7, 7, 7, 7,MachineState.ERROR,MachineState.ERROR,#50-57 - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#58-5f - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR, 7, 7,MachineState.ERROR,MachineState.ERROR,#60-67 - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#68-6f - MachineState.ERROR,MachineState.ERROR, 9, 9, 9, 9,MachineState.ERROR,MachineState.ERROR,#70-77 - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#78-7f - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR, 9,MachineState.ERROR,MachineState.ERROR,#80-87 - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#88-8f - MachineState.ERROR,MachineState.ERROR, 12, 12, 12, 12,MachineState.ERROR,MachineState.ERROR,#90-97 - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#98-9f - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR, 12,MachineState.ERROR,MachineState.ERROR,#a0-a7 - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#a8-af - MachineState.ERROR,MachineState.ERROR, 12, 12, 12,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#b0-b7 - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#b8-bf - MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.ERROR,MachineState.ERROR,#c0-c7 - MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR #c8-cf -) - -UTF8_CHAR_LEN_TABLE = (0, 1, 0, 0, 0, 0, 2, 3, 3, 3, 4, 4, 5, 5, 6, 6) - -UTF8_SM_MODEL = {'class_table': UTF8_CLS, - 'class_factor': 16, - 'state_table': UTF8_ST, - 'char_len_table': UTF8_CHAR_LEN_TABLE, - 'name': 'UTF-8'} diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/sbcharsetprober.py b/src/fetchcode/vcs/pip/_vendor/chardet/sbcharsetprober.py deleted file mode 100644 index 0adb51de..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/sbcharsetprober.py +++ /dev/null @@ -1,132 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Universal charset detector code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 2001 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# Shy Shalom - original C code -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -from .charsetprober import CharSetProber -from .enums import CharacterCategory, ProbingState, SequenceLikelihood - - -class SingleByteCharSetProber(CharSetProber): - SAMPLE_SIZE = 64 - SB_ENOUGH_REL_THRESHOLD = 1024 # 0.25 * SAMPLE_SIZE^2 - POSITIVE_SHORTCUT_THRESHOLD = 0.95 - NEGATIVE_SHORTCUT_THRESHOLD = 0.05 - - def __init__(self, model, reversed=False, name_prober=None): - super(SingleByteCharSetProber, self).__init__() - self._model = model - # TRUE if we need to reverse every pair in the model lookup - self._reversed = reversed - # Optional auxiliary prober for name decision - self._name_prober = name_prober - self._last_order = None - self._seq_counters = None - self._total_seqs = None - self._total_char = None - self._freq_char = None - self.reset() - - def reset(self): - super(SingleByteCharSetProber, self).reset() - # char order of last character - self._last_order = 255 - self._seq_counters = [0] * SequenceLikelihood.get_num_categories() - self._total_seqs = 0 - self._total_char = 0 - # characters that fall in our sampling range - self._freq_char = 0 - - @property - def charset_name(self): - if self._name_prober: - return self._name_prober.charset_name - else: - return self._model['charset_name'] - - @property - def language(self): - if self._name_prober: - return self._name_prober.language - else: - return self._model.get('language') - - def feed(self, byte_str): - if not self._model['keep_english_letter']: - byte_str = self.filter_international_words(byte_str) - if not byte_str: - return self.state - char_to_order_map = self._model['char_to_order_map'] - for i, c in enumerate(byte_str): - # XXX: Order is in range 1-64, so one would think we want 0-63 here, - # but that leads to 27 more test failures than before. - order = char_to_order_map[c] - # XXX: This was SYMBOL_CAT_ORDER before, with a value of 250, but - # CharacterCategory.SYMBOL is actually 253, so we use CONTROL - # to make it closer to the original intent. The only difference - # is whether or not we count digits and control characters for - # _total_char purposes. - if order < CharacterCategory.CONTROL: - self._total_char += 1 - if order < self.SAMPLE_SIZE: - self._freq_char += 1 - if self._last_order < self.SAMPLE_SIZE: - self._total_seqs += 1 - if not self._reversed: - i = (self._last_order * self.SAMPLE_SIZE) + order - model = self._model['precedence_matrix'][i] - else: # reverse the order of the letters in the lookup - i = (order * self.SAMPLE_SIZE) + self._last_order - model = self._model['precedence_matrix'][i] - self._seq_counters[model] += 1 - self._last_order = order - - charset_name = self._model['charset_name'] - if self.state == ProbingState.DETECTING: - if self._total_seqs > self.SB_ENOUGH_REL_THRESHOLD: - confidence = self.get_confidence() - if confidence > self.POSITIVE_SHORTCUT_THRESHOLD: - self.logger.debug('%s confidence = %s, we have a winner', - charset_name, confidence) - self._state = ProbingState.FOUND_IT - elif confidence < self.NEGATIVE_SHORTCUT_THRESHOLD: - self.logger.debug('%s confidence = %s, below negative ' - 'shortcut threshhold %s', charset_name, - confidence, - self.NEGATIVE_SHORTCUT_THRESHOLD) - self._state = ProbingState.NOT_ME - - return self.state - - def get_confidence(self): - r = 0.01 - if self._total_seqs > 0: - r = ((1.0 * self._seq_counters[SequenceLikelihood.POSITIVE]) / - self._total_seqs / self._model['typical_positive_ratio']) - r = r * self._freq_char / self._total_char - if r >= 1.0: - r = 0.99 - return r diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/sbcsgroupprober.py b/src/fetchcode/vcs/pip/_vendor/chardet/sbcsgroupprober.py deleted file mode 100644 index 98e95dc1..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/sbcsgroupprober.py +++ /dev/null @@ -1,73 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Universal charset detector code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 2001 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# Shy Shalom - original C code -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -from .charsetgroupprober import CharSetGroupProber -from .sbcharsetprober import SingleByteCharSetProber -from .langcyrillicmodel import (Win1251CyrillicModel, Koi8rModel, - Latin5CyrillicModel, MacCyrillicModel, - Ibm866Model, Ibm855Model) -from .langgreekmodel import Latin7GreekModel, Win1253GreekModel -from .langbulgarianmodel import Latin5BulgarianModel, Win1251BulgarianModel -# from .langhungarianmodel import Latin2HungarianModel, Win1250HungarianModel -from .langthaimodel import TIS620ThaiModel -from .langhebrewmodel import Win1255HebrewModel -from .hebrewprober import HebrewProber -from .langturkishmodel import Latin5TurkishModel - - -class SBCSGroupProber(CharSetGroupProber): - def __init__(self): - super(SBCSGroupProber, self).__init__() - self.probers = [ - SingleByteCharSetProber(Win1251CyrillicModel), - SingleByteCharSetProber(Koi8rModel), - SingleByteCharSetProber(Latin5CyrillicModel), - SingleByteCharSetProber(MacCyrillicModel), - SingleByteCharSetProber(Ibm866Model), - SingleByteCharSetProber(Ibm855Model), - SingleByteCharSetProber(Latin7GreekModel), - SingleByteCharSetProber(Win1253GreekModel), - SingleByteCharSetProber(Latin5BulgarianModel), - SingleByteCharSetProber(Win1251BulgarianModel), - # TODO: Restore Hungarian encodings (iso-8859-2 and windows-1250) - # after we retrain model. - # SingleByteCharSetProber(Latin2HungarianModel), - # SingleByteCharSetProber(Win1250HungarianModel), - SingleByteCharSetProber(TIS620ThaiModel), - SingleByteCharSetProber(Latin5TurkishModel), - ] - hebrew_prober = HebrewProber() - logical_hebrew_prober = SingleByteCharSetProber(Win1255HebrewModel, - False, hebrew_prober) - visual_hebrew_prober = SingleByteCharSetProber(Win1255HebrewModel, True, - hebrew_prober) - hebrew_prober.set_model_probers(logical_hebrew_prober, visual_hebrew_prober) - self.probers.extend([hebrew_prober, logical_hebrew_prober, - visual_hebrew_prober]) - - self.reset() diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/sjisprober.py b/src/fetchcode/vcs/pip/_vendor/chardet/sjisprober.py deleted file mode 100644 index 9e29623b..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/sjisprober.py +++ /dev/null @@ -1,92 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is mozilla.org code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -from .mbcharsetprober import MultiByteCharSetProber -from .codingstatemachine import CodingStateMachine -from .chardistribution import SJISDistributionAnalysis -from .jpcntx import SJISContextAnalysis -from .mbcssm import SJIS_SM_MODEL -from .enums import ProbingState, MachineState - - -class SJISProber(MultiByteCharSetProber): - def __init__(self): - super(SJISProber, self).__init__() - self.coding_sm = CodingStateMachine(SJIS_SM_MODEL) - self.distribution_analyzer = SJISDistributionAnalysis() - self.context_analyzer = SJISContextAnalysis() - self.reset() - - def reset(self): - super(SJISProber, self).reset() - self.context_analyzer.reset() - - @property - def charset_name(self): - return self.context_analyzer.charset_name - - @property - def language(self): - return "Japanese" - - def feed(self, byte_str): - for i in range(len(byte_str)): - coding_state = self.coding_sm.next_state(byte_str[i]) - if coding_state == MachineState.ERROR: - self.logger.debug('%s %s prober hit error at byte %s', - self.charset_name, self.language, i) - self._state = ProbingState.NOT_ME - break - elif coding_state == MachineState.ITS_ME: - self._state = ProbingState.FOUND_IT - break - elif coding_state == MachineState.START: - char_len = self.coding_sm.get_current_charlen() - if i == 0: - self._last_char[1] = byte_str[0] - self.context_analyzer.feed(self._last_char[2 - char_len:], - char_len) - self.distribution_analyzer.feed(self._last_char, char_len) - else: - self.context_analyzer.feed(byte_str[i + 1 - char_len:i + 3 - - char_len], char_len) - self.distribution_analyzer.feed(byte_str[i - 1:i + 1], - char_len) - - self._last_char[0] = byte_str[-1] - - if self.state == ProbingState.DETECTING: - if (self.context_analyzer.got_enough_data() and - (self.get_confidence() > self.SHORTCUT_THRESHOLD)): - self._state = ProbingState.FOUND_IT - - return self.state - - def get_confidence(self): - context_conf = self.context_analyzer.get_confidence() - distrib_conf = self.distribution_analyzer.get_confidence() - return max(context_conf, distrib_conf) diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/universaldetector.py b/src/fetchcode/vcs/pip/_vendor/chardet/universaldetector.py deleted file mode 100644 index 7b4e92d6..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/universaldetector.py +++ /dev/null @@ -1,286 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is Mozilla Universal charset detector code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 2001 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# Shy Shalom - original C code -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### -""" -Module containing the UniversalDetector detector class, which is the primary -class a user of ``chardet`` should use. - -:author: Mark Pilgrim (initial port to Python) -:author: Shy Shalom (original C code) -:author: Dan Blanchard (major refactoring for 3.0) -:author: Ian Cordasco -""" - - -import codecs -import logging -import re - -from .charsetgroupprober import CharSetGroupProber -from .enums import InputState, LanguageFilter, ProbingState -from .escprober import EscCharSetProber -from .latin1prober import Latin1Prober -from .mbcsgroupprober import MBCSGroupProber -from .sbcsgroupprober import SBCSGroupProber - - -class UniversalDetector(object): - """ - The ``UniversalDetector`` class underlies the ``chardet.detect`` function - and coordinates all of the different charset probers. - - To get a ``dict`` containing an encoding and its confidence, you can simply - run: - - .. code:: - - u = UniversalDetector() - u.feed(some_bytes) - u.close() - detected = u.result - - """ - - MINIMUM_THRESHOLD = 0.20 - HIGH_BYTE_DETECTOR = re.compile(b'[\x80-\xFF]') - ESC_DETECTOR = re.compile(b'(\033|~{)') - WIN_BYTE_DETECTOR = re.compile(b'[\x80-\x9F]') - ISO_WIN_MAP = {'iso-8859-1': 'Windows-1252', - 'iso-8859-2': 'Windows-1250', - 'iso-8859-5': 'Windows-1251', - 'iso-8859-6': 'Windows-1256', - 'iso-8859-7': 'Windows-1253', - 'iso-8859-8': 'Windows-1255', - 'iso-8859-9': 'Windows-1254', - 'iso-8859-13': 'Windows-1257'} - - def __init__(self, lang_filter=LanguageFilter.ALL): - self._esc_charset_prober = None - self._charset_probers = [] - self.result = None - self.done = None - self._got_data = None - self._input_state = None - self._last_char = None - self.lang_filter = lang_filter - self.logger = logging.getLogger(__name__) - self._has_win_bytes = None - self.reset() - - def reset(self): - """ - Reset the UniversalDetector and all of its probers back to their - initial states. This is called by ``__init__``, so you only need to - call this directly in between analyses of different documents. - """ - self.result = {'encoding': None, 'confidence': 0.0, 'language': None} - self.done = False - self._got_data = False - self._has_win_bytes = False - self._input_state = InputState.PURE_ASCII - self._last_char = b'' - if self._esc_charset_prober: - self._esc_charset_prober.reset() - for prober in self._charset_probers: - prober.reset() - - def feed(self, byte_str): - """ - Takes a chunk of a document and feeds it through all of the relevant - charset probers. - - After calling ``feed``, you can check the value of the ``done`` - attribute to see if you need to continue feeding the - ``UniversalDetector`` more data, or if it has made a prediction - (in the ``result`` attribute). - - .. note:: - You should always call ``close`` when you're done feeding in your - document if ``done`` is not already ``True``. - """ - if self.done: - return - - if not len(byte_str): - return - - if not isinstance(byte_str, bytearray): - byte_str = bytearray(byte_str) - - # First check for known BOMs, since these are guaranteed to be correct - if not self._got_data: - # If the data starts with BOM, we know it is UTF - if byte_str.startswith(codecs.BOM_UTF8): - # EF BB BF UTF-8 with BOM - self.result = {'encoding': "UTF-8-SIG", - 'confidence': 1.0, - 'language': ''} - elif byte_str.startswith((codecs.BOM_UTF32_LE, - codecs.BOM_UTF32_BE)): - # FF FE 00 00 UTF-32, little-endian BOM - # 00 00 FE FF UTF-32, big-endian BOM - self.result = {'encoding': "UTF-32", - 'confidence': 1.0, - 'language': ''} - elif byte_str.startswith(b'\xFE\xFF\x00\x00'): - # FE FF 00 00 UCS-4, unusual octet order BOM (3412) - self.result = {'encoding': "X-ISO-10646-UCS-4-3412", - 'confidence': 1.0, - 'language': ''} - elif byte_str.startswith(b'\x00\x00\xFF\xFE'): - # 00 00 FF FE UCS-4, unusual octet order BOM (2143) - self.result = {'encoding': "X-ISO-10646-UCS-4-2143", - 'confidence': 1.0, - 'language': ''} - elif byte_str.startswith((codecs.BOM_LE, codecs.BOM_BE)): - # FF FE UTF-16, little endian BOM - # FE FF UTF-16, big endian BOM - self.result = {'encoding': "UTF-16", - 'confidence': 1.0, - 'language': ''} - - self._got_data = True - if self.result['encoding'] is not None: - self.done = True - return - - # If none of those matched and we've only see ASCII so far, check - # for high bytes and escape sequences - if self._input_state == InputState.PURE_ASCII: - if self.HIGH_BYTE_DETECTOR.search(byte_str): - self._input_state = InputState.HIGH_BYTE - elif self._input_state == InputState.PURE_ASCII and \ - self.ESC_DETECTOR.search(self._last_char + byte_str): - self._input_state = InputState.ESC_ASCII - - self._last_char = byte_str[-1:] - - # If we've seen escape sequences, use the EscCharSetProber, which - # uses a simple state machine to check for known escape sequences in - # HZ and ISO-2022 encodings, since those are the only encodings that - # use such sequences. - if self._input_state == InputState.ESC_ASCII: - if not self._esc_charset_prober: - self._esc_charset_prober = EscCharSetProber(self.lang_filter) - if self._esc_charset_prober.feed(byte_str) == ProbingState.FOUND_IT: - self.result = {'encoding': - self._esc_charset_prober.charset_name, - 'confidence': - self._esc_charset_prober.get_confidence(), - 'language': - self._esc_charset_prober.language} - self.done = True - # If we've seen high bytes (i.e., those with values greater than 127), - # we need to do more complicated checks using all our multi-byte and - # single-byte probers that are left. The single-byte probers - # use character bigram distributions to determine the encoding, whereas - # the multi-byte probers use a combination of character unigram and - # bigram distributions. - elif self._input_state == InputState.HIGH_BYTE: - if not self._charset_probers: - self._charset_probers = [MBCSGroupProber(self.lang_filter)] - # If we're checking non-CJK encodings, use single-byte prober - if self.lang_filter & LanguageFilter.NON_CJK: - self._charset_probers.append(SBCSGroupProber()) - self._charset_probers.append(Latin1Prober()) - for prober in self._charset_probers: - if prober.feed(byte_str) == ProbingState.FOUND_IT: - self.result = {'encoding': prober.charset_name, - 'confidence': prober.get_confidence(), - 'language': prober.language} - self.done = True - break - if self.WIN_BYTE_DETECTOR.search(byte_str): - self._has_win_bytes = True - - def close(self): - """ - Stop analyzing the current document and come up with a final - prediction. - - :returns: The ``result`` attribute, a ``dict`` with the keys - `encoding`, `confidence`, and `language`. - """ - # Don't bother with checks if we're already done - if self.done: - return self.result - self.done = True - - if not self._got_data: - self.logger.debug('no data received!') - - # Default to ASCII if it is all we've seen so far - elif self._input_state == InputState.PURE_ASCII: - self.result = {'encoding': 'ascii', - 'confidence': 1.0, - 'language': ''} - - # If we have seen non-ASCII, return the best that met MINIMUM_THRESHOLD - elif self._input_state == InputState.HIGH_BYTE: - prober_confidence = None - max_prober_confidence = 0.0 - max_prober = None - for prober in self._charset_probers: - if not prober: - continue - prober_confidence = prober.get_confidence() - if prober_confidence > max_prober_confidence: - max_prober_confidence = prober_confidence - max_prober = prober - if max_prober and (max_prober_confidence > self.MINIMUM_THRESHOLD): - charset_name = max_prober.charset_name - lower_charset_name = max_prober.charset_name.lower() - confidence = max_prober.get_confidence() - # Use Windows encoding name instead of ISO-8859 if we saw any - # extra Windows-specific bytes - if lower_charset_name.startswith('iso-8859'): - if self._has_win_bytes: - charset_name = self.ISO_WIN_MAP.get(lower_charset_name, - charset_name) - self.result = {'encoding': charset_name, - 'confidence': confidence, - 'language': max_prober.language} - - # Log all prober confidences if none met MINIMUM_THRESHOLD - if self.logger.getEffectiveLevel() == logging.DEBUG: - if self.result['encoding'] is None: - self.logger.debug('no probers hit minimum threshold') - for group_prober in self._charset_probers: - if not group_prober: - continue - if isinstance(group_prober, CharSetGroupProber): - for prober in group_prober.probers: - self.logger.debug('%s %s confidence = %s', - prober.charset_name, - prober.language, - prober.get_confidence()) - else: - self.logger.debug('%s %s confidence = %s', - prober.charset_name, - prober.language, - prober.get_confidence()) - return self.result diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/utf8prober.py b/src/fetchcode/vcs/pip/_vendor/chardet/utf8prober.py deleted file mode 100644 index 6c3196cc..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/utf8prober.py +++ /dev/null @@ -1,82 +0,0 @@ -######################## BEGIN LICENSE BLOCK ######################## -# The Original Code is mozilla.org code. -# -# The Initial Developer of the Original Code is -# Netscape Communications Corporation. -# Portions created by the Initial Developer are Copyright (C) 1998 -# the Initial Developer. All Rights Reserved. -# -# Contributor(s): -# Mark Pilgrim - port to Python -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2.1 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, write to the Free Software -# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA -# 02110-1301 USA -######################### END LICENSE BLOCK ######################### - -from .charsetprober import CharSetProber -from .enums import ProbingState, MachineState -from .codingstatemachine import CodingStateMachine -from .mbcssm import UTF8_SM_MODEL - - - -class UTF8Prober(CharSetProber): - ONE_CHAR_PROB = 0.5 - - def __init__(self): - super(UTF8Prober, self).__init__() - self.coding_sm = CodingStateMachine(UTF8_SM_MODEL) - self._num_mb_chars = None - self.reset() - - def reset(self): - super(UTF8Prober, self).reset() - self.coding_sm.reset() - self._num_mb_chars = 0 - - @property - def charset_name(self): - return "utf-8" - - @property - def language(self): - return "" - - def feed(self, byte_str): - for c in byte_str: - coding_state = self.coding_sm.next_state(c) - if coding_state == MachineState.ERROR: - self._state = ProbingState.NOT_ME - break - elif coding_state == MachineState.ITS_ME: - self._state = ProbingState.FOUND_IT - break - elif coding_state == MachineState.START: - if self.coding_sm.get_current_charlen() >= 2: - self._num_mb_chars += 1 - - if self.state == ProbingState.DETECTING: - if self.get_confidence() > self.SHORTCUT_THRESHOLD: - self._state = ProbingState.FOUND_IT - - return self.state - - def get_confidence(self): - unlike = 0.99 - if self._num_mb_chars < 6: - unlike *= self.ONE_CHAR_PROB ** self._num_mb_chars - return 1.0 - unlike - else: - return unlike diff --git a/src/fetchcode/vcs/pip/_vendor/chardet/version.py b/src/fetchcode/vcs/pip/_vendor/chardet/version.py deleted file mode 100644 index bb2a34a7..00000000 --- a/src/fetchcode/vcs/pip/_vendor/chardet/version.py +++ /dev/null @@ -1,9 +0,0 @@ -""" -This module exists only to simplify retrieving the version number of chardet -from within setup.py and from chardet subpackages. - -:author: Dan Blanchard (dan.blanchard@gmail.com) -""" - -__version__ = "3.0.4" -VERSION = __version__.split('.') diff --git a/src/fetchcode/vcs/pip/_vendor/colorama.pyi b/src/fetchcode/vcs/pip/_vendor/colorama.pyi deleted file mode 100644 index 60a6c254..00000000 --- a/src/fetchcode/vcs/pip/_vendor/colorama.pyi +++ /dev/null @@ -1 +0,0 @@ -from colorama import * \ No newline at end of file diff --git a/src/fetchcode/vcs/pip/_vendor/colorama/LICENSE.txt b/src/fetchcode/vcs/pip/_vendor/colorama/LICENSE.txt deleted file mode 100644 index 3105888e..00000000 --- a/src/fetchcode/vcs/pip/_vendor/colorama/LICENSE.txt +++ /dev/null @@ -1,27 +0,0 @@ -Copyright (c) 2010 Jonathan Hartley -All rights reserved. - -Redistribution and use in source and binary forms, with or without -modification, are permitted provided that the following conditions are met: - -* Redistributions of source code must retain the above copyright notice, this - list of conditions and the following disclaimer. - -* Redistributions in binary form must reproduce the above copyright notice, - this list of conditions and the following disclaimer in the documentation - and/or other materials provided with the distribution. - -* Neither the name of the copyright holders, nor those of its contributors - may be used to endorse or promote products derived from this software without - specific prior written permission. - -THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND -ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED -WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE -DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE -FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL -DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR -SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER -CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, -OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. diff --git a/src/fetchcode/vcs/pip/_vendor/colorama/__init__.py b/src/fetchcode/vcs/pip/_vendor/colorama/__init__.py deleted file mode 100644 index 34c263cc..00000000 --- a/src/fetchcode/vcs/pip/_vendor/colorama/__init__.py +++ /dev/null @@ -1,6 +0,0 @@ -# Copyright Jonathan Hartley 2013. BSD 3-Clause license, see LICENSE file. -from .initialise import init, deinit, reinit, colorama_text -from .ansi import Fore, Back, Style, Cursor -from .ansitowin32 import AnsiToWin32 - -__version__ = '0.4.3' diff --git a/src/fetchcode/vcs/pip/_vendor/colorama/ansi.py b/src/fetchcode/vcs/pip/_vendor/colorama/ansi.py deleted file mode 100644 index 78776588..00000000 --- a/src/fetchcode/vcs/pip/_vendor/colorama/ansi.py +++ /dev/null @@ -1,102 +0,0 @@ -# Copyright Jonathan Hartley 2013. BSD 3-Clause license, see LICENSE file. -''' -This module generates ANSI character codes to printing colors to terminals. -See: http://en.wikipedia.org/wiki/ANSI_escape_code -''' - -CSI = '\033[' -OSC = '\033]' -BEL = '\007' - - -def code_to_chars(code): - return CSI + str(code) + 'm' - -def set_title(title): - return OSC + '2;' + title + BEL - -def clear_screen(mode=2): - return CSI + str(mode) + 'J' - -def clear_line(mode=2): - return CSI + str(mode) + 'K' - - -class AnsiCodes(object): - def __init__(self): - # the subclasses declare class attributes which are numbers. - # Upon instantiation we define instance attributes, which are the same - # as the class attributes but wrapped with the ANSI escape sequence - for name in dir(self): - if not name.startswith('_'): - value = getattr(self, name) - setattr(self, name, code_to_chars(value)) - - -class AnsiCursor(object): - def UP(self, n=1): - return CSI + str(n) + 'A' - def DOWN(self, n=1): - return CSI + str(n) + 'B' - def FORWARD(self, n=1): - return CSI + str(n) + 'C' - def BACK(self, n=1): - return CSI + str(n) + 'D' - def POS(self, x=1, y=1): - return CSI + str(y) + ';' + str(x) + 'H' - - -class AnsiFore(AnsiCodes): - BLACK = 30 - RED = 31 - GREEN = 32 - YELLOW = 33 - BLUE = 34 - MAGENTA = 35 - CYAN = 36 - WHITE = 37 - RESET = 39 - - # These are fairly well supported, but not part of the standard. - LIGHTBLACK_EX = 90 - LIGHTRED_EX = 91 - LIGHTGREEN_EX = 92 - LIGHTYELLOW_EX = 93 - LIGHTBLUE_EX = 94 - LIGHTMAGENTA_EX = 95 - LIGHTCYAN_EX = 96 - LIGHTWHITE_EX = 97 - - -class AnsiBack(AnsiCodes): - BLACK = 40 - RED = 41 - GREEN = 42 - YELLOW = 43 - BLUE = 44 - MAGENTA = 45 - CYAN = 46 - WHITE = 47 - RESET = 49 - - # These are fairly well supported, but not part of the standard. - LIGHTBLACK_EX = 100 - LIGHTRED_EX = 101 - LIGHTGREEN_EX = 102 - LIGHTYELLOW_EX = 103 - LIGHTBLUE_EX = 104 - LIGHTMAGENTA_EX = 105 - LIGHTCYAN_EX = 106 - LIGHTWHITE_EX = 107 - - -class AnsiStyle(AnsiCodes): - BRIGHT = 1 - DIM = 2 - NORMAL = 22 - RESET_ALL = 0 - -Fore = AnsiFore() -Back = AnsiBack() -Style = AnsiStyle() -Cursor = AnsiCursor() diff --git a/src/fetchcode/vcs/pip/_vendor/colorama/ansitowin32.py b/src/fetchcode/vcs/pip/_vendor/colorama/ansitowin32.py deleted file mode 100644 index 359c92be..00000000 --- a/src/fetchcode/vcs/pip/_vendor/colorama/ansitowin32.py +++ /dev/null @@ -1,257 +0,0 @@ -# Copyright Jonathan Hartley 2013. BSD 3-Clause license, see LICENSE file. -import re -import sys -import os - -from .ansi import AnsiFore, AnsiBack, AnsiStyle, Style -from .winterm import WinTerm, WinColor, WinStyle -from .win32 import windll, winapi_test - - -winterm = None -if windll is not None: - winterm = WinTerm() - - -class StreamWrapper(object): - ''' - Wraps a stream (such as stdout), acting as a transparent proxy for all - attribute access apart from method 'write()', which is delegated to our - Converter instance. - ''' - def __init__(self, wrapped, converter): - # double-underscore everything to prevent clashes with names of - # attributes on the wrapped stream object. - self.__wrapped = wrapped - self.__convertor = converter - - def __getattr__(self, name): - return getattr(self.__wrapped, name) - - def __enter__(self, *args, **kwargs): - # special method lookup bypasses __getattr__/__getattribute__, see - # https://stackoverflow.com/questions/12632894/why-doesnt-getattr-work-with-exit - # thus, contextlib magic methods are not proxied via __getattr__ - return self.__wrapped.__enter__(*args, **kwargs) - - def __exit__(self, *args, **kwargs): - return self.__wrapped.__exit__(*args, **kwargs) - - def write(self, text): - self.__convertor.write(text) - - def isatty(self): - stream = self.__wrapped - if 'PYCHARM_HOSTED' in os.environ: - if stream is not None and (stream is sys.__stdout__ or stream is sys.__stderr__): - return True - try: - stream_isatty = stream.isatty - except AttributeError: - return False - else: - return stream_isatty() - - @property - def closed(self): - stream = self.__wrapped - try: - return stream.closed - except AttributeError: - return True - - -class AnsiToWin32(object): - ''' - Implements a 'write()' method which, on Windows, will strip ANSI character - sequences from the text, and if outputting to a tty, will convert them into - win32 function calls. - ''' - ANSI_CSI_RE = re.compile('\001?\033\\[((?:\\d|;)*)([a-zA-Z])\002?') # Control Sequence Introducer - ANSI_OSC_RE = re.compile('\001?\033\\]((?:.|;)*?)(\x07)\002?') # Operating System Command - - def __init__(self, wrapped, convert=None, strip=None, autoreset=False): - # The wrapped stream (normally sys.stdout or sys.stderr) - self.wrapped = wrapped - - # should we reset colors to defaults after every .write() - self.autoreset = autoreset - - # create the proxy wrapping our output stream - self.stream = StreamWrapper(wrapped, self) - - on_windows = os.name == 'nt' - # We test if the WinAPI works, because even if we are on Windows - # we may be using a terminal that doesn't support the WinAPI - # (e.g. Cygwin Terminal). In this case it's up to the terminal - # to support the ANSI codes. - conversion_supported = on_windows and winapi_test() - - # should we strip ANSI sequences from our output? - if strip is None: - strip = conversion_supported or (not self.stream.closed and not self.stream.isatty()) - self.strip = strip - - # should we should convert ANSI sequences into win32 calls? - if convert is None: - convert = conversion_supported and not self.stream.closed and self.stream.isatty() - self.convert = convert - - # dict of ansi codes to win32 functions and parameters - self.win32_calls = self.get_win32_calls() - - # are we wrapping stderr? - self.on_stderr = self.wrapped is sys.stderr - - def should_wrap(self): - ''' - True if this class is actually needed. If false, then the output - stream will not be affected, nor will win32 calls be issued, so - wrapping stdout is not actually required. This will generally be - False on non-Windows platforms, unless optional functionality like - autoreset has been requested using kwargs to init() - ''' - return self.convert or self.strip or self.autoreset - - def get_win32_calls(self): - if self.convert and winterm: - return { - AnsiStyle.RESET_ALL: (winterm.reset_all, ), - AnsiStyle.BRIGHT: (winterm.style, WinStyle.BRIGHT), - AnsiStyle.DIM: (winterm.style, WinStyle.NORMAL), - AnsiStyle.NORMAL: (winterm.style, WinStyle.NORMAL), - AnsiFore.BLACK: (winterm.fore, WinColor.BLACK), - AnsiFore.RED: (winterm.fore, WinColor.RED), - AnsiFore.GREEN: (winterm.fore, WinColor.GREEN), - AnsiFore.YELLOW: (winterm.fore, WinColor.YELLOW), - AnsiFore.BLUE: (winterm.fore, WinColor.BLUE), - AnsiFore.MAGENTA: (winterm.fore, WinColor.MAGENTA), - AnsiFore.CYAN: (winterm.fore, WinColor.CYAN), - AnsiFore.WHITE: (winterm.fore, WinColor.GREY), - AnsiFore.RESET: (winterm.fore, ), - AnsiFore.LIGHTBLACK_EX: (winterm.fore, WinColor.BLACK, True), - AnsiFore.LIGHTRED_EX: (winterm.fore, WinColor.RED, True), - AnsiFore.LIGHTGREEN_EX: (winterm.fore, WinColor.GREEN, True), - AnsiFore.LIGHTYELLOW_EX: (winterm.fore, WinColor.YELLOW, True), - AnsiFore.LIGHTBLUE_EX: (winterm.fore, WinColor.BLUE, True), - AnsiFore.LIGHTMAGENTA_EX: (winterm.fore, WinColor.MAGENTA, True), - AnsiFore.LIGHTCYAN_EX: (winterm.fore, WinColor.CYAN, True), - AnsiFore.LIGHTWHITE_EX: (winterm.fore, WinColor.GREY, True), - AnsiBack.BLACK: (winterm.back, WinColor.BLACK), - AnsiBack.RED: (winterm.back, WinColor.RED), - AnsiBack.GREEN: (winterm.back, WinColor.GREEN), - AnsiBack.YELLOW: (winterm.back, WinColor.YELLOW), - AnsiBack.BLUE: (winterm.back, WinColor.BLUE), - AnsiBack.MAGENTA: (winterm.back, WinColor.MAGENTA), - AnsiBack.CYAN: (winterm.back, WinColor.CYAN), - AnsiBack.WHITE: (winterm.back, WinColor.GREY), - AnsiBack.RESET: (winterm.back, ), - AnsiBack.LIGHTBLACK_EX: (winterm.back, WinColor.BLACK, True), - AnsiBack.LIGHTRED_EX: (winterm.back, WinColor.RED, True), - AnsiBack.LIGHTGREEN_EX: (winterm.back, WinColor.GREEN, True), - AnsiBack.LIGHTYELLOW_EX: (winterm.back, WinColor.YELLOW, True), - AnsiBack.LIGHTBLUE_EX: (winterm.back, WinColor.BLUE, True), - AnsiBack.LIGHTMAGENTA_EX: (winterm.back, WinColor.MAGENTA, True), - AnsiBack.LIGHTCYAN_EX: (winterm.back, WinColor.CYAN, True), - AnsiBack.LIGHTWHITE_EX: (winterm.back, WinColor.GREY, True), - } - return dict() - - def write(self, text): - if self.strip or self.convert: - self.write_and_convert(text) - else: - self.wrapped.write(text) - self.wrapped.flush() - if self.autoreset: - self.reset_all() - - - def reset_all(self): - if self.convert: - self.call_win32('m', (0,)) - elif not self.strip and not self.stream.closed: - self.wrapped.write(Style.RESET_ALL) - - - def write_and_convert(self, text): - ''' - Write the given text to our wrapped stream, stripping any ANSI - sequences from the text, and optionally converting them into win32 - calls. - ''' - cursor = 0 - text = self.convert_osc(text) - for match in self.ANSI_CSI_RE.finditer(text): - start, end = match.span() - self.write_plain_text(text, cursor, start) - self.convert_ansi(*match.groups()) - cursor = end - self.write_plain_text(text, cursor, len(text)) - - - def write_plain_text(self, text, start, end): - if start < end: - self.wrapped.write(text[start:end]) - self.wrapped.flush() - - - def convert_ansi(self, paramstring, command): - if self.convert: - params = self.extract_params(command, paramstring) - self.call_win32(command, params) - - - def extract_params(self, command, paramstring): - if command in 'Hf': - params = tuple(int(p) if len(p) != 0 else 1 for p in paramstring.split(';')) - while len(params) < 2: - # defaults: - params = params + (1,) - else: - params = tuple(int(p) for p in paramstring.split(';') if len(p) != 0) - if len(params) == 0: - # defaults: - if command in 'JKm': - params = (0,) - elif command in 'ABCD': - params = (1,) - - return params - - - def call_win32(self, command, params): - if command == 'm': - for param in params: - if param in self.win32_calls: - func_args = self.win32_calls[param] - func = func_args[0] - args = func_args[1:] - kwargs = dict(on_stderr=self.on_stderr) - func(*args, **kwargs) - elif command in 'J': - winterm.erase_screen(params[0], on_stderr=self.on_stderr) - elif command in 'K': - winterm.erase_line(params[0], on_stderr=self.on_stderr) - elif command in 'Hf': # cursor position - absolute - winterm.set_cursor_position(params, on_stderr=self.on_stderr) - elif command in 'ABCD': # cursor position - relative - n = params[0] - # A - up, B - down, C - forward, D - back - x, y = {'A': (0, -n), 'B': (0, n), 'C': (n, 0), 'D': (-n, 0)}[command] - winterm.cursor_adjust(x, y, on_stderr=self.on_stderr) - - - def convert_osc(self, text): - for match in self.ANSI_OSC_RE.finditer(text): - start, end = match.span() - text = text[:start] + text[end:] - paramstring, command = match.groups() - if command in '\x07': # \x07 = BEL - params = paramstring.split(";") - # 0 - change title and icon (we will only change title) - # 1 - change icon (we don't support this) - # 2 - change title - if params[0] in '02': - winterm.set_title(params[1]) - return text diff --git a/src/fetchcode/vcs/pip/_vendor/colorama/initialise.py b/src/fetchcode/vcs/pip/_vendor/colorama/initialise.py deleted file mode 100644 index 430d0668..00000000 --- a/src/fetchcode/vcs/pip/_vendor/colorama/initialise.py +++ /dev/null @@ -1,80 +0,0 @@ -# Copyright Jonathan Hartley 2013. BSD 3-Clause license, see LICENSE file. -import atexit -import contextlib -import sys - -from .ansitowin32 import AnsiToWin32 - - -orig_stdout = None -orig_stderr = None - -wrapped_stdout = None -wrapped_stderr = None - -atexit_done = False - - -def reset_all(): - if AnsiToWin32 is not None: # Issue #74: objects might become None at exit - AnsiToWin32(orig_stdout).reset_all() - - -def init(autoreset=False, convert=None, strip=None, wrap=True): - - if not wrap and any([autoreset, convert, strip]): - raise ValueError('wrap=False conflicts with any other arg=True') - - global wrapped_stdout, wrapped_stderr - global orig_stdout, orig_stderr - - orig_stdout = sys.stdout - orig_stderr = sys.stderr - - if sys.stdout is None: - wrapped_stdout = None - else: - sys.stdout = wrapped_stdout = \ - wrap_stream(orig_stdout, convert, strip, autoreset, wrap) - if sys.stderr is None: - wrapped_stderr = None - else: - sys.stderr = wrapped_stderr = \ - wrap_stream(orig_stderr, convert, strip, autoreset, wrap) - - global atexit_done - if not atexit_done: - atexit.register(reset_all) - atexit_done = True - - -def deinit(): - if orig_stdout is not None: - sys.stdout = orig_stdout - if orig_stderr is not None: - sys.stderr = orig_stderr - - -@contextlib.contextmanager -def colorama_text(*args, **kwargs): - init(*args, **kwargs) - try: - yield - finally: - deinit() - - -def reinit(): - if wrapped_stdout is not None: - sys.stdout = wrapped_stdout - if wrapped_stderr is not None: - sys.stderr = wrapped_stderr - - -def wrap_stream(stream, convert, strip, autoreset, wrap): - if wrap: - wrapper = AnsiToWin32(stream, - convert=convert, strip=strip, autoreset=autoreset) - if wrapper.should_wrap(): - stream = wrapper.stream - return stream diff --git a/src/fetchcode/vcs/pip/_vendor/colorama/win32.py b/src/fetchcode/vcs/pip/_vendor/colorama/win32.py deleted file mode 100644 index c2d83603..00000000 --- a/src/fetchcode/vcs/pip/_vendor/colorama/win32.py +++ /dev/null @@ -1,152 +0,0 @@ -# Copyright Jonathan Hartley 2013. BSD 3-Clause license, see LICENSE file. - -# from winbase.h -STDOUT = -11 -STDERR = -12 - -try: - import ctypes - from ctypes import LibraryLoader - windll = LibraryLoader(ctypes.WinDLL) - from ctypes import wintypes -except (AttributeError, ImportError): - windll = None - SetConsoleTextAttribute = lambda *_: None - winapi_test = lambda *_: None -else: - from ctypes import byref, Structure, c_char, POINTER - - COORD = wintypes._COORD - - class CONSOLE_SCREEN_BUFFER_INFO(Structure): - """struct in wincon.h.""" - _fields_ = [ - ("dwSize", COORD), - ("dwCursorPosition", COORD), - ("wAttributes", wintypes.WORD), - ("srWindow", wintypes.SMALL_RECT), - ("dwMaximumWindowSize", COORD), - ] - def __str__(self): - return '(%d,%d,%d,%d,%d,%d,%d,%d,%d,%d,%d)' % ( - self.dwSize.Y, self.dwSize.X - , self.dwCursorPosition.Y, self.dwCursorPosition.X - , self.wAttributes - , self.srWindow.Top, self.srWindow.Left, self.srWindow.Bottom, self.srWindow.Right - , self.dwMaximumWindowSize.Y, self.dwMaximumWindowSize.X - ) - - _GetStdHandle = windll.kernel32.GetStdHandle - _GetStdHandle.argtypes = [ - wintypes.DWORD, - ] - _GetStdHandle.restype = wintypes.HANDLE - - _GetConsoleScreenBufferInfo = windll.kernel32.GetConsoleScreenBufferInfo - _GetConsoleScreenBufferInfo.argtypes = [ - wintypes.HANDLE, - POINTER(CONSOLE_SCREEN_BUFFER_INFO), - ] - _GetConsoleScreenBufferInfo.restype = wintypes.BOOL - - _SetConsoleTextAttribute = windll.kernel32.SetConsoleTextAttribute - _SetConsoleTextAttribute.argtypes = [ - wintypes.HANDLE, - wintypes.WORD, - ] - _SetConsoleTextAttribute.restype = wintypes.BOOL - - _SetConsoleCursorPosition = windll.kernel32.SetConsoleCursorPosition - _SetConsoleCursorPosition.argtypes = [ - wintypes.HANDLE, - COORD, - ] - _SetConsoleCursorPosition.restype = wintypes.BOOL - - _FillConsoleOutputCharacterA = windll.kernel32.FillConsoleOutputCharacterA - _FillConsoleOutputCharacterA.argtypes = [ - wintypes.HANDLE, - c_char, - wintypes.DWORD, - COORD, - POINTER(wintypes.DWORD), - ] - _FillConsoleOutputCharacterA.restype = wintypes.BOOL - - _FillConsoleOutputAttribute = windll.kernel32.FillConsoleOutputAttribute - _FillConsoleOutputAttribute.argtypes = [ - wintypes.HANDLE, - wintypes.WORD, - wintypes.DWORD, - COORD, - POINTER(wintypes.DWORD), - ] - _FillConsoleOutputAttribute.restype = wintypes.BOOL - - _SetConsoleTitleW = windll.kernel32.SetConsoleTitleW - _SetConsoleTitleW.argtypes = [ - wintypes.LPCWSTR - ] - _SetConsoleTitleW.restype = wintypes.BOOL - - def _winapi_test(handle): - csbi = CONSOLE_SCREEN_BUFFER_INFO() - success = _GetConsoleScreenBufferInfo( - handle, byref(csbi)) - return bool(success) - - def winapi_test(): - return any(_winapi_test(h) for h in - (_GetStdHandle(STDOUT), _GetStdHandle(STDERR))) - - def GetConsoleScreenBufferInfo(stream_id=STDOUT): - handle = _GetStdHandle(stream_id) - csbi = CONSOLE_SCREEN_BUFFER_INFO() - success = _GetConsoleScreenBufferInfo( - handle, byref(csbi)) - return csbi - - def SetConsoleTextAttribute(stream_id, attrs): - handle = _GetStdHandle(stream_id) - return _SetConsoleTextAttribute(handle, attrs) - - def SetConsoleCursorPosition(stream_id, position, adjust=True): - position = COORD(*position) - # If the position is out of range, do nothing. - if position.Y <= 0 or position.X <= 0: - return - # Adjust for Windows' SetConsoleCursorPosition: - # 1. being 0-based, while ANSI is 1-based. - # 2. expecting (x,y), while ANSI uses (y,x). - adjusted_position = COORD(position.Y - 1, position.X - 1) - if adjust: - # Adjust for viewport's scroll position - sr = GetConsoleScreenBufferInfo(STDOUT).srWindow - adjusted_position.Y += sr.Top - adjusted_position.X += sr.Left - # Resume normal processing - handle = _GetStdHandle(stream_id) - return _SetConsoleCursorPosition(handle, adjusted_position) - - def FillConsoleOutputCharacter(stream_id, char, length, start): - handle = _GetStdHandle(stream_id) - char = c_char(char.encode()) - length = wintypes.DWORD(length) - num_written = wintypes.DWORD(0) - # Note that this is hard-coded for ANSI (vs wide) bytes. - success = _FillConsoleOutputCharacterA( - handle, char, length, start, byref(num_written)) - return num_written.value - - def FillConsoleOutputAttribute(stream_id, attr, length, start): - ''' FillConsoleOutputAttribute( hConsole, csbi.wAttributes, dwConSize, coordScreen, &cCharsWritten )''' - handle = _GetStdHandle(stream_id) - attribute = wintypes.WORD(attr) - length = wintypes.DWORD(length) - num_written = wintypes.DWORD(0) - # Note that this is hard-coded for ANSI (vs wide) bytes. - return _FillConsoleOutputAttribute( - handle, attribute, length, start, byref(num_written)) - - def SetConsoleTitle(title): - return _SetConsoleTitleW(title) diff --git a/src/fetchcode/vcs/pip/_vendor/colorama/winterm.py b/src/fetchcode/vcs/pip/_vendor/colorama/winterm.py deleted file mode 100644 index 0fdb4ec4..00000000 --- a/src/fetchcode/vcs/pip/_vendor/colorama/winterm.py +++ /dev/null @@ -1,169 +0,0 @@ -# Copyright Jonathan Hartley 2013. BSD 3-Clause license, see LICENSE file. -from . import win32 - - -# from wincon.h -class WinColor(object): - BLACK = 0 - BLUE = 1 - GREEN = 2 - CYAN = 3 - RED = 4 - MAGENTA = 5 - YELLOW = 6 - GREY = 7 - -# from wincon.h -class WinStyle(object): - NORMAL = 0x00 # dim text, dim background - BRIGHT = 0x08 # bright text, dim background - BRIGHT_BACKGROUND = 0x80 # dim text, bright background - -class WinTerm(object): - - def __init__(self): - self._default = win32.GetConsoleScreenBufferInfo(win32.STDOUT).wAttributes - self.set_attrs(self._default) - self._default_fore = self._fore - self._default_back = self._back - self._default_style = self._style - # In order to emulate LIGHT_EX in windows, we borrow the BRIGHT style. - # So that LIGHT_EX colors and BRIGHT style do not clobber each other, - # we track them separately, since LIGHT_EX is overwritten by Fore/Back - # and BRIGHT is overwritten by Style codes. - self._light = 0 - - def get_attrs(self): - return self._fore + self._back * 16 + (self._style | self._light) - - def set_attrs(self, value): - self._fore = value & 7 - self._back = (value >> 4) & 7 - self._style = value & (WinStyle.BRIGHT | WinStyle.BRIGHT_BACKGROUND) - - def reset_all(self, on_stderr=None): - self.set_attrs(self._default) - self.set_console(attrs=self._default) - self._light = 0 - - def fore(self, fore=None, light=False, on_stderr=False): - if fore is None: - fore = self._default_fore - self._fore = fore - # Emulate LIGHT_EX with BRIGHT Style - if light: - self._light |= WinStyle.BRIGHT - else: - self._light &= ~WinStyle.BRIGHT - self.set_console(on_stderr=on_stderr) - - def back(self, back=None, light=False, on_stderr=False): - if back is None: - back = self._default_back - self._back = back - # Emulate LIGHT_EX with BRIGHT_BACKGROUND Style - if light: - self._light |= WinStyle.BRIGHT_BACKGROUND - else: - self._light &= ~WinStyle.BRIGHT_BACKGROUND - self.set_console(on_stderr=on_stderr) - - def style(self, style=None, on_stderr=False): - if style is None: - style = self._default_style - self._style = style - self.set_console(on_stderr=on_stderr) - - def set_console(self, attrs=None, on_stderr=False): - if attrs is None: - attrs = self.get_attrs() - handle = win32.STDOUT - if on_stderr: - handle = win32.STDERR - win32.SetConsoleTextAttribute(handle, attrs) - - def get_position(self, handle): - position = win32.GetConsoleScreenBufferInfo(handle).dwCursorPosition - # Because Windows coordinates are 0-based, - # and win32.SetConsoleCursorPosition expects 1-based. - position.X += 1 - position.Y += 1 - return position - - def set_cursor_position(self, position=None, on_stderr=False): - if position is None: - # I'm not currently tracking the position, so there is no default. - # position = self.get_position() - return - handle = win32.STDOUT - if on_stderr: - handle = win32.STDERR - win32.SetConsoleCursorPosition(handle, position) - - def cursor_adjust(self, x, y, on_stderr=False): - handle = win32.STDOUT - if on_stderr: - handle = win32.STDERR - position = self.get_position(handle) - adjusted_position = (position.Y + y, position.X + x) - win32.SetConsoleCursorPosition(handle, adjusted_position, adjust=False) - - def erase_screen(self, mode=0, on_stderr=False): - # 0 should clear from the cursor to the end of the screen. - # 1 should clear from the cursor to the beginning of the screen. - # 2 should clear the entire screen, and move cursor to (1,1) - handle = win32.STDOUT - if on_stderr: - handle = win32.STDERR - csbi = win32.GetConsoleScreenBufferInfo(handle) - # get the number of character cells in the current buffer - cells_in_screen = csbi.dwSize.X * csbi.dwSize.Y - # get number of character cells before current cursor position - cells_before_cursor = csbi.dwSize.X * csbi.dwCursorPosition.Y + csbi.dwCursorPosition.X - if mode == 0: - from_coord = csbi.dwCursorPosition - cells_to_erase = cells_in_screen - cells_before_cursor - elif mode == 1: - from_coord = win32.COORD(0, 0) - cells_to_erase = cells_before_cursor - elif mode == 2: - from_coord = win32.COORD(0, 0) - cells_to_erase = cells_in_screen - else: - # invalid mode - return - # fill the entire screen with blanks - win32.FillConsoleOutputCharacter(handle, ' ', cells_to_erase, from_coord) - # now set the buffer's attributes accordingly - win32.FillConsoleOutputAttribute(handle, self.get_attrs(), cells_to_erase, from_coord) - if mode == 2: - # put the cursor where needed - win32.SetConsoleCursorPosition(handle, (1, 1)) - - def erase_line(self, mode=0, on_stderr=False): - # 0 should clear from the cursor to the end of the line. - # 1 should clear from the cursor to the beginning of the line. - # 2 should clear the entire line. - handle = win32.STDOUT - if on_stderr: - handle = win32.STDERR - csbi = win32.GetConsoleScreenBufferInfo(handle) - if mode == 0: - from_coord = csbi.dwCursorPosition - cells_to_erase = csbi.dwSize.X - csbi.dwCursorPosition.X - elif mode == 1: - from_coord = win32.COORD(0, csbi.dwCursorPosition.Y) - cells_to_erase = csbi.dwCursorPosition.X - elif mode == 2: - from_coord = win32.COORD(0, csbi.dwCursorPosition.Y) - cells_to_erase = csbi.dwSize.X - else: - # invalid mode - return - # fill the entire screen with blanks - win32.FillConsoleOutputCharacter(handle, ' ', cells_to_erase, from_coord) - # now set the buffer's attributes accordingly - win32.FillConsoleOutputAttribute(handle, self.get_attrs(), cells_to_erase, from_coord) - - def set_title(self, title): - win32.SetConsoleTitle(title) diff --git a/src/fetchcode/vcs/pip/_vendor/contextlib2.LICENSE.txt b/src/fetchcode/vcs/pip/_vendor/contextlib2.LICENSE.txt deleted file mode 100644 index 5de20277..00000000 --- a/src/fetchcode/vcs/pip/_vendor/contextlib2.LICENSE.txt +++ /dev/null @@ -1,122 +0,0 @@ - - -A. HISTORY OF THE SOFTWARE -========================== - -contextlib2 is a derivative of the contextlib module distributed by the PSF -as part of the Python standard library. According, it is itself redistributed -under the PSF license (reproduced in full below). As the contextlib module -was added only in Python 2.5, the licenses for earlier Python versions are -not applicable and have not been included. - -Python was created in the early 1990s by Guido van Rossum at Stichting -Mathematisch Centrum (CWI, see http://www.cwi.nl) in the Netherlands -as a successor of a language called ABC. Guido remains Python's -principal author, although it includes many contributions from others. - -In 1995, Guido continued his work on Python at the Corporation for -National Research Initiatives (CNRI, see http://www.cnri.reston.va.us) -in Reston, Virginia where he released several versions of the -software. - -In May 2000, Guido and the Python core development team moved to -BeOpen.com to form the BeOpen PythonLabs team. In October of the same -year, the PythonLabs team moved to Digital Creations (now Zope -Corporation, see http://www.zope.com). In 2001, the Python Software -Foundation (PSF, see http://www.python.org/psf/) was formed, a -non-profit organization created specifically to own Python-related -Intellectual Property. Zope Corporation is a sponsoring member of -the PSF. - -All Python releases are Open Source (see http://www.opensource.org for -the Open Source Definition). Historically, most, but not all, Python -releases have also been GPL-compatible; the table below summarizes -the various releases that included the contextlib module. - - Release Derived Year Owner GPL- - from compatible? (1) - - 2.5 2.4 2006 PSF yes - 2.5.1 2.5 2007 PSF yes - 2.5.2 2.5.1 2008 PSF yes - 2.5.3 2.5.2 2008 PSF yes - 2.6 2.5 2008 PSF yes - 2.6.1 2.6 2008 PSF yes - 2.6.2 2.6.1 2009 PSF yes - 2.6.3 2.6.2 2009 PSF yes - 2.6.4 2.6.3 2009 PSF yes - 2.6.5 2.6.4 2010 PSF yes - 3.0 2.6 2008 PSF yes - 3.0.1 3.0 2009 PSF yes - 3.1 3.0.1 2009 PSF yes - 3.1.1 3.1 2009 PSF yes - 3.1.2 3.1.1 2010 PSF yes - 3.1.3 3.1.2 2010 PSF yes - 3.1.4 3.1.3 2011 PSF yes - 3.2 3.1 2011 PSF yes - 3.2.1 3.2 2011 PSF yes - 3.2.2 3.2.1 2011 PSF yes - 3.3 3.2 2012 PSF yes - -Footnotes: - -(1) GPL-compatible doesn't mean that we're distributing Python under - the GPL. All Python licenses, unlike the GPL, let you distribute - a modified version without making your changes open source. The - GPL-compatible licenses make it possible to combine Python with - other software that is released under the GPL; the others don't. - -Thanks to the many outside volunteers who have worked under Guido's -direction to make these releases possible. - - -B. TERMS AND CONDITIONS FOR ACCESSING OR OTHERWISE USING PYTHON -=============================================================== - -PYTHON SOFTWARE FOUNDATION LICENSE VERSION 2 --------------------------------------------- - -1. This LICENSE AGREEMENT is between the Python Software Foundation -("PSF"), and the Individual or Organization ("Licensee") accessing and -otherwise using this software ("Python") in source or binary form and -its associated documentation. - -2. Subject to the terms and conditions of this License Agreement, PSF hereby -grants Licensee a nonexclusive, royalty-free, world-wide license to reproduce, -analyze, test, perform and/or display publicly, prepare derivative works, -distribute, and otherwise use Python alone or in any derivative version, -provided, however, that PSF's License Agreement and PSF's notice of copyright, -i.e., "Copyright (c) 2001, 2002, 2003, 2004, 2005, 2006, 2007, 2008, 2009, 2010, -2011 Python Software Foundation; All Rights Reserved" are retained in Python -alone or in any derivative version prepared by Licensee. - -3. In the event Licensee prepares a derivative work that is based on -or incorporates Python or any part thereof, and wants to make -the derivative work available to others as provided herein, then -Licensee hereby agrees to include in any such work a brief summary of -the changes made to Python. - -4. PSF is making Python available to Licensee on an "AS IS" -basis. PSF MAKES NO REPRESENTATIONS OR WARRANTIES, EXPRESS OR -IMPLIED. BY WAY OF EXAMPLE, BUT NOT LIMITATION, PSF MAKES NO AND -DISCLAIMS ANY REPRESENTATION OR WARRANTY OF MERCHANTABILITY OR FITNESS -FOR ANY PARTICULAR PURPOSE OR THAT THE USE OF PYTHON WILL NOT -INFRINGE ANY THIRD PARTY RIGHTS. - -5. PSF SHALL NOT BE LIABLE TO LICENSEE OR ANY OTHER USERS OF PYTHON -FOR ANY INCIDENTAL, SPECIAL, OR CONSEQUENTIAL DAMAGES OR LOSS AS -A RESULT OF MODIFYING, DISTRIBUTING, OR OTHERWISE USING PYTHON, -OR ANY DERIVATIVE THEREOF, EVEN IF ADVISED OF THE POSSIBILITY THEREOF. - -6. This License Agreement will automatically terminate upon a material -breach of its terms and conditions. - -7. Nothing in this License Agreement shall be deemed to create any -relationship of agency, partnership, or joint venture between PSF and -Licensee. This License Agreement does not grant permission to use PSF -trademarks or trade name in a trademark sense to endorse or promote -products or services of Licensee, or any third party. - -8. By copying, installing or otherwise using Python, Licensee -agrees to be bound by the terms and conditions of this License -Agreement. diff --git a/src/fetchcode/vcs/pip/_vendor/contextlib2.py b/src/fetchcode/vcs/pip/_vendor/contextlib2.py deleted file mode 100644 index 3aae8f41..00000000 --- a/src/fetchcode/vcs/pip/_vendor/contextlib2.py +++ /dev/null @@ -1,518 +0,0 @@ -"""contextlib2 - backports and enhancements to the contextlib module""" - -import abc -import sys -import warnings -from collections import deque -from functools import wraps - -__all__ = ["contextmanager", "closing", "nullcontext", - "AbstractContextManager", - "ContextDecorator", "ExitStack", - "redirect_stdout", "redirect_stderr", "suppress"] - -# Backwards compatibility -__all__ += ["ContextStack"] - - -# Backport abc.ABC -if sys.version_info[:2] >= (3, 4): - _abc_ABC = abc.ABC -else: - _abc_ABC = abc.ABCMeta('ABC', (object,), {'__slots__': ()}) - - -# Backport classic class MRO -def _classic_mro(C, result): - if C in result: - return - result.append(C) - for B in C.__bases__: - _classic_mro(B, result) - return result - - -# Backport _collections_abc._check_methods -def _check_methods(C, *methods): - try: - mro = C.__mro__ - except AttributeError: - mro = tuple(_classic_mro(C, [])) - - for method in methods: - for B in mro: - if method in B.__dict__: - if B.__dict__[method] is None: - return NotImplemented - break - else: - return NotImplemented - return True - - -class AbstractContextManager(_abc_ABC): - """An abstract base class for context managers.""" - - def __enter__(self): - """Return `self` upon entering the runtime context.""" - return self - - @abc.abstractmethod - def __exit__(self, exc_type, exc_value, traceback): - """Raise any exception triggered within the runtime context.""" - return None - - @classmethod - def __subclasshook__(cls, C): - """Check whether subclass is considered a subclass of this ABC.""" - if cls is AbstractContextManager: - return _check_methods(C, "__enter__", "__exit__") - return NotImplemented - - -class ContextDecorator(object): - """A base class or mixin that enables context managers to work as decorators.""" - - def refresh_cm(self): - """Returns the context manager used to actually wrap the call to the - decorated function. - - The default implementation just returns *self*. - - Overriding this method allows otherwise one-shot context managers - like _GeneratorContextManager to support use as decorators via - implicit recreation. - - DEPRECATED: refresh_cm was never added to the standard library's - ContextDecorator API - """ - warnings.warn("refresh_cm was never added to the standard library", - DeprecationWarning) - return self._recreate_cm() - - def _recreate_cm(self): - """Return a recreated instance of self. - - Allows an otherwise one-shot context manager like - _GeneratorContextManager to support use as - a decorator via implicit recreation. - - This is a private interface just for _GeneratorContextManager. - See issue #11647 for details. - """ - return self - - def __call__(self, func): - @wraps(func) - def inner(*args, **kwds): - with self._recreate_cm(): - return func(*args, **kwds) - return inner - - -class _GeneratorContextManager(ContextDecorator): - """Helper for @contextmanager decorator.""" - - def __init__(self, func, args, kwds): - self.gen = func(*args, **kwds) - self.func, self.args, self.kwds = func, args, kwds - # Issue 19330: ensure context manager instances have good docstrings - doc = getattr(func, "__doc__", None) - if doc is None: - doc = type(self).__doc__ - self.__doc__ = doc - # Unfortunately, this still doesn't provide good help output when - # inspecting the created context manager instances, since pydoc - # currently bypasses the instance docstring and shows the docstring - # for the class instead. - # See http://bugs.python.org/issue19404 for more details. - - def _recreate_cm(self): - # _GCM instances are one-shot context managers, so the - # CM must be recreated each time a decorated function is - # called - return self.__class__(self.func, self.args, self.kwds) - - def __enter__(self): - try: - return next(self.gen) - except StopIteration: - raise RuntimeError("generator didn't yield") - - def __exit__(self, type, value, traceback): - if type is None: - try: - next(self.gen) - except StopIteration: - return - else: - raise RuntimeError("generator didn't stop") - else: - if value is None: - # Need to force instantiation so we can reliably - # tell if we get the same exception back - value = type() - try: - self.gen.throw(type, value, traceback) - raise RuntimeError("generator didn't stop after throw()") - except StopIteration as exc: - # Suppress StopIteration *unless* it's the same exception that - # was passed to throw(). This prevents a StopIteration - # raised inside the "with" statement from being suppressed. - return exc is not value - except RuntimeError as exc: - # Don't re-raise the passed in exception - if exc is value: - return False - # Likewise, avoid suppressing if a StopIteration exception - # was passed to throw() and later wrapped into a RuntimeError - # (see PEP 479). - if _HAVE_EXCEPTION_CHAINING and exc.__cause__ is value: - return False - raise - except: - # only re-raise if it's *not* the exception that was - # passed to throw(), because __exit__() must not raise - # an exception unless __exit__() itself failed. But throw() - # has to raise the exception to signal propagation, so this - # fixes the impedance mismatch between the throw() protocol - # and the __exit__() protocol. - # - if sys.exc_info()[1] is not value: - raise - - -def contextmanager(func): - """@contextmanager decorator. - - Typical usage: - - @contextmanager - def some_generator(): - - try: - yield - finally: - - - This makes this: - - with some_generator() as : - - - equivalent to this: - - - try: - = - - finally: - - - """ - @wraps(func) - def helper(*args, **kwds): - return _GeneratorContextManager(func, args, kwds) - return helper - - -class closing(object): - """Context to automatically close something at the end of a block. - - Code like this: - - with closing(.open()) as f: - - - is equivalent to this: - - f = .open() - try: - - finally: - f.close() - - """ - def __init__(self, thing): - self.thing = thing - - def __enter__(self): - return self.thing - - def __exit__(self, *exc_info): - self.thing.close() - - -class _RedirectStream(object): - - _stream = None - - def __init__(self, new_target): - self._new_target = new_target - # We use a list of old targets to make this CM re-entrant - self._old_targets = [] - - def __enter__(self): - self._old_targets.append(getattr(sys, self._stream)) - setattr(sys, self._stream, self._new_target) - return self._new_target - - def __exit__(self, exctype, excinst, exctb): - setattr(sys, self._stream, self._old_targets.pop()) - - -class redirect_stdout(_RedirectStream): - """Context manager for temporarily redirecting stdout to another file. - - # How to send help() to stderr - with redirect_stdout(sys.stderr): - help(dir) - - # How to write help() to a file - with open('help.txt', 'w') as f: - with redirect_stdout(f): - help(pow) - """ - - _stream = "stdout" - - -class redirect_stderr(_RedirectStream): - """Context manager for temporarily redirecting stderr to another file.""" - - _stream = "stderr" - - -class suppress(object): - """Context manager to suppress specified exceptions - - After the exception is suppressed, execution proceeds with the next - statement following the with statement. - - with suppress(FileNotFoundError): - os.remove(somefile) - # Execution still resumes here if the file was already removed - """ - - def __init__(self, *exceptions): - self._exceptions = exceptions - - def __enter__(self): - pass - - def __exit__(self, exctype, excinst, exctb): - # Unlike isinstance and issubclass, CPython exception handling - # currently only looks at the concrete type hierarchy (ignoring - # the instance and subclass checking hooks). While Guido considers - # that a bug rather than a feature, it's a fairly hard one to fix - # due to various internal implementation details. suppress provides - # the simpler issubclass based semantics, rather than trying to - # exactly reproduce the limitations of the CPython interpreter. - # - # See http://bugs.python.org/issue12029 for more details - return exctype is not None and issubclass(exctype, self._exceptions) - - -# Context manipulation is Python 3 only -_HAVE_EXCEPTION_CHAINING = sys.version_info[0] >= 3 -if _HAVE_EXCEPTION_CHAINING: - def _make_context_fixer(frame_exc): - def _fix_exception_context(new_exc, old_exc): - # Context may not be correct, so find the end of the chain - while 1: - exc_context = new_exc.__context__ - if exc_context is old_exc: - # Context is already set correctly (see issue 20317) - return - if exc_context is None or exc_context is frame_exc: - break - new_exc = exc_context - # Change the end of the chain to point to the exception - # we expect it to reference - new_exc.__context__ = old_exc - return _fix_exception_context - - def _reraise_with_existing_context(exc_details): - try: - # bare "raise exc_details[1]" replaces our carefully - # set-up context - fixed_ctx = exc_details[1].__context__ - raise exc_details[1] - except BaseException: - exc_details[1].__context__ = fixed_ctx - raise -else: - # No exception context in Python 2 - def _make_context_fixer(frame_exc): - return lambda new_exc, old_exc: None - - # Use 3 argument raise in Python 2, - # but use exec to avoid SyntaxError in Python 3 - def _reraise_with_existing_context(exc_details): - exc_type, exc_value, exc_tb = exc_details - exec("raise exc_type, exc_value, exc_tb") - -# Handle old-style classes if they exist -try: - from types import InstanceType -except ImportError: - # Python 3 doesn't have old-style classes - _get_type = type -else: - # Need to handle old-style context managers on Python 2 - def _get_type(obj): - obj_type = type(obj) - if obj_type is InstanceType: - return obj.__class__ # Old-style class - return obj_type # New-style class - - -# Inspired by discussions on http://bugs.python.org/issue13585 -class ExitStack(object): - """Context manager for dynamic management of a stack of exit callbacks - - For example: - - with ExitStack() as stack: - files = [stack.enter_context(open(fname)) for fname in filenames] - # All opened files will automatically be closed at the end of - # the with statement, even if attempts to open files later - # in the list raise an exception - - """ - def __init__(self): - self._exit_callbacks = deque() - - def pop_all(self): - """Preserve the context stack by transferring it to a new instance""" - new_stack = type(self)() - new_stack._exit_callbacks = self._exit_callbacks - self._exit_callbacks = deque() - return new_stack - - def _push_cm_exit(self, cm, cm_exit): - """Helper to correctly register callbacks to __exit__ methods""" - def _exit_wrapper(*exc_details): - return cm_exit(cm, *exc_details) - _exit_wrapper.__self__ = cm - self.push(_exit_wrapper) - - def push(self, exit): - """Registers a callback with the standard __exit__ method signature - - Can suppress exceptions the same way __exit__ methods can. - - Also accepts any object with an __exit__ method (registering a call - to the method instead of the object itself) - """ - # We use an unbound method rather than a bound method to follow - # the standard lookup behaviour for special methods - _cb_type = _get_type(exit) - try: - exit_method = _cb_type.__exit__ - except AttributeError: - # Not a context manager, so assume its a callable - self._exit_callbacks.append(exit) - else: - self._push_cm_exit(exit, exit_method) - return exit # Allow use as a decorator - - def callback(self, callback, *args, **kwds): - """Registers an arbitrary callback and arguments. - - Cannot suppress exceptions. - """ - def _exit_wrapper(exc_type, exc, tb): - callback(*args, **kwds) - # We changed the signature, so using @wraps is not appropriate, but - # setting __wrapped__ may still help with introspection - _exit_wrapper.__wrapped__ = callback - self.push(_exit_wrapper) - return callback # Allow use as a decorator - - def enter_context(self, cm): - """Enters the supplied context manager - - If successful, also pushes its __exit__ method as a callback and - returns the result of the __enter__ method. - """ - # We look up the special methods on the type to match the with statement - _cm_type = _get_type(cm) - _exit = _cm_type.__exit__ - result = _cm_type.__enter__(cm) - self._push_cm_exit(cm, _exit) - return result - - def close(self): - """Immediately unwind the context stack""" - self.__exit__(None, None, None) - - def __enter__(self): - return self - - def __exit__(self, *exc_details): - received_exc = exc_details[0] is not None - - # We manipulate the exception state so it behaves as though - # we were actually nesting multiple with statements - frame_exc = sys.exc_info()[1] - _fix_exception_context = _make_context_fixer(frame_exc) - - # Callbacks are invoked in LIFO order to match the behaviour of - # nested context managers - suppressed_exc = False - pending_raise = False - while self._exit_callbacks: - cb = self._exit_callbacks.pop() - try: - if cb(*exc_details): - suppressed_exc = True - pending_raise = False - exc_details = (None, None, None) - except: - new_exc_details = sys.exc_info() - # simulate the stack of exceptions by setting the context - _fix_exception_context(new_exc_details[1], exc_details[1]) - pending_raise = True - exc_details = new_exc_details - if pending_raise: - _reraise_with_existing_context(exc_details) - return received_exc and suppressed_exc - - -# Preserve backwards compatibility -class ContextStack(ExitStack): - """Backwards compatibility alias for ExitStack""" - - def __init__(self): - warnings.warn("ContextStack has been renamed to ExitStack", - DeprecationWarning) - super(ContextStack, self).__init__() - - def register_exit(self, callback): - return self.push(callback) - - def register(self, callback, *args, **kwds): - return self.callback(callback, *args, **kwds) - - def preserve(self): - return self.pop_all() - - -class nullcontext(AbstractContextManager): - """Context manager that does no additional processing. - Used as a stand-in for a normal context manager, when a particular - block of code is only sometimes used with a normal context manager: - cm = optional_cm if condition else nullcontext() - with cm: - # Perform operation, using optional_cm if condition is True - """ - - def __init__(self, enter_result=None): - self.enter_result = enter_result - - def __enter__(self): - return self.enter_result - - def __exit__(self, *excinfo): - pass diff --git a/src/fetchcode/vcs/pip/_vendor/distlib.pyi b/src/fetchcode/vcs/pip/_vendor/distlib.pyi deleted file mode 100644 index ea94b159..00000000 --- a/src/fetchcode/vcs/pip/_vendor/distlib.pyi +++ /dev/null @@ -1 +0,0 @@ -from distlib import * \ No newline at end of file diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/__init__.py b/src/fetchcode/vcs/pip/_vendor/distlib/__init__.py index e19aebdc..e999438f 100644 --- a/src/fetchcode/vcs/pip/_vendor/distlib/__init__.py +++ b/src/fetchcode/vcs/pip/_vendor/distlib/__init__.py @@ -1,23 +1,33 @@ # -*- coding: utf-8 -*- # -# Copyright (C) 2012-2019 Vinay Sajip. +# Copyright (C) 2012-2023 Vinay Sajip. # Licensed to the Python Software Foundation under a contributor agreement. # See LICENSE.txt and CONTRIBUTORS.txt. # import logging -__version__ = '0.3.0' +__version__ = '0.3.8' + class DistlibException(Exception): pass + try: from logging import NullHandler -except ImportError: # pragma: no cover +except ImportError: # pragma: no cover + class NullHandler(logging.Handler): - def handle(self, record): pass - def emit(self, record): pass - def createLock(self): self.lock = None + + def handle(self, record): + pass + + def emit(self, record): + pass + + def createLock(self): + self.lock = None + logger = logging.getLogger(__name__) logger.addHandler(NullHandler()) diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/_backport/__init__.py b/src/fetchcode/vcs/pip/_vendor/distlib/_backport/__init__.py deleted file mode 100644 index f7dbf4c9..00000000 --- a/src/fetchcode/vcs/pip/_vendor/distlib/_backport/__init__.py +++ /dev/null @@ -1,6 +0,0 @@ -"""Modules copied from Python 3 standard libraries, for internal use only. - -Individual classes and functions are found in d2._backport.misc. Intended -usage is to always import things missing from 3.1 from that module: the -built-in/stdlib objects will be used if found. -""" diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/_backport/misc.py b/src/fetchcode/vcs/pip/_vendor/distlib/_backport/misc.py deleted file mode 100644 index cfb318d3..00000000 --- a/src/fetchcode/vcs/pip/_vendor/distlib/_backport/misc.py +++ /dev/null @@ -1,41 +0,0 @@ -# -*- coding: utf-8 -*- -# -# Copyright (C) 2012 The Python Software Foundation. -# See LICENSE.txt and CONTRIBUTORS.txt. -# -"""Backports for individual classes and functions.""" - -import os -import sys - -__all__ = ['cache_from_source', 'callable', 'fsencode'] - - -try: - from imp import cache_from_source -except ImportError: - def cache_from_source(py_file, debug=__debug__): - ext = debug and 'c' or 'o' - return py_file + ext - - -try: - callable = callable -except NameError: - from collections import Callable - - def callable(obj): - return isinstance(obj, Callable) - - -try: - fsencode = os.fsencode -except AttributeError: - def fsencode(filename): - if isinstance(filename, bytes): - return filename - elif isinstance(filename, str): - return filename.encode(sys.getfilesystemencoding()) - else: - raise TypeError("expect bytes or str, not %s" % - type(filename).__name__) diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/_backport/shutil.py b/src/fetchcode/vcs/pip/_vendor/distlib/_backport/shutil.py deleted file mode 100644 index 159e49ee..00000000 --- a/src/fetchcode/vcs/pip/_vendor/distlib/_backport/shutil.py +++ /dev/null @@ -1,761 +0,0 @@ -# -*- coding: utf-8 -*- -# -# Copyright (C) 2012 The Python Software Foundation. -# See LICENSE.txt and CONTRIBUTORS.txt. -# -"""Utility functions for copying and archiving files and directory trees. - -XXX The functions here don't copy the resource fork or other metadata on Mac. - -""" - -import os -import sys -import stat -from os.path import abspath -import fnmatch -import collections -import errno -from . import tarfile - -try: - import bz2 - _BZ2_SUPPORTED = True -except ImportError: - _BZ2_SUPPORTED = False - -try: - from pwd import getpwnam -except ImportError: - getpwnam = None - -try: - from grp import getgrnam -except ImportError: - getgrnam = None - -__all__ = ["copyfileobj", "copyfile", "copymode", "copystat", "copy", "copy2", - "copytree", "move", "rmtree", "Error", "SpecialFileError", - "ExecError", "make_archive", "get_archive_formats", - "register_archive_format", "unregister_archive_format", - "get_unpack_formats", "register_unpack_format", - "unregister_unpack_format", "unpack_archive", "ignore_patterns"] - -class Error(EnvironmentError): - pass - -class SpecialFileError(EnvironmentError): - """Raised when trying to do a kind of operation (e.g. copying) which is - not supported on a special file (e.g. a named pipe)""" - -class ExecError(EnvironmentError): - """Raised when a command could not be executed""" - -class ReadError(EnvironmentError): - """Raised when an archive cannot be read""" - -class RegistryError(Exception): - """Raised when a registry operation with the archiving - and unpacking registries fails""" - - -try: - WindowsError -except NameError: - WindowsError = None - -def copyfileobj(fsrc, fdst, length=16*1024): - """copy data from file-like object fsrc to file-like object fdst""" - while 1: - buf = fsrc.read(length) - if not buf: - break - fdst.write(buf) - -def _samefile(src, dst): - # Macintosh, Unix. - if hasattr(os.path, 'samefile'): - try: - return os.path.samefile(src, dst) - except OSError: - return False - - # All other platforms: check for same pathname. - return (os.path.normcase(os.path.abspath(src)) == - os.path.normcase(os.path.abspath(dst))) - -def copyfile(src, dst): - """Copy data from src to dst""" - if _samefile(src, dst): - raise Error("`%s` and `%s` are the same file" % (src, dst)) - - for fn in [src, dst]: - try: - st = os.stat(fn) - except OSError: - # File most likely does not exist - pass - else: - # XXX What about other special files? (sockets, devices...) - if stat.S_ISFIFO(st.st_mode): - raise SpecialFileError("`%s` is a named pipe" % fn) - - with open(src, 'rb') as fsrc: - with open(dst, 'wb') as fdst: - copyfileobj(fsrc, fdst) - -def copymode(src, dst): - """Copy mode bits from src to dst""" - if hasattr(os, 'chmod'): - st = os.stat(src) - mode = stat.S_IMODE(st.st_mode) - os.chmod(dst, mode) - -def copystat(src, dst): - """Copy all stat info (mode bits, atime, mtime, flags) from src to dst""" - st = os.stat(src) - mode = stat.S_IMODE(st.st_mode) - if hasattr(os, 'utime'): - os.utime(dst, (st.st_atime, st.st_mtime)) - if hasattr(os, 'chmod'): - os.chmod(dst, mode) - if hasattr(os, 'chflags') and hasattr(st, 'st_flags'): - try: - os.chflags(dst, st.st_flags) - except OSError as why: - if (not hasattr(errno, 'EOPNOTSUPP') or - why.errno != errno.EOPNOTSUPP): - raise - -def copy(src, dst): - """Copy data and mode bits ("cp src dst"). - - The destination may be a directory. - - """ - if os.path.isdir(dst): - dst = os.path.join(dst, os.path.basename(src)) - copyfile(src, dst) - copymode(src, dst) - -def copy2(src, dst): - """Copy data and all stat info ("cp -p src dst"). - - The destination may be a directory. - - """ - if os.path.isdir(dst): - dst = os.path.join(dst, os.path.basename(src)) - copyfile(src, dst) - copystat(src, dst) - -def ignore_patterns(*patterns): - """Function that can be used as copytree() ignore parameter. - - Patterns is a sequence of glob-style patterns - that are used to exclude files""" - def _ignore_patterns(path, names): - ignored_names = [] - for pattern in patterns: - ignored_names.extend(fnmatch.filter(names, pattern)) - return set(ignored_names) - return _ignore_patterns - -def copytree(src, dst, symlinks=False, ignore=None, copy_function=copy2, - ignore_dangling_symlinks=False): - """Recursively copy a directory tree. - - The destination directory must not already exist. - If exception(s) occur, an Error is raised with a list of reasons. - - If the optional symlinks flag is true, symbolic links in the - source tree result in symbolic links in the destination tree; if - it is false, the contents of the files pointed to by symbolic - links are copied. If the file pointed by the symlink doesn't - exist, an exception will be added in the list of errors raised in - an Error exception at the end of the copy process. - - You can set the optional ignore_dangling_symlinks flag to true if you - want to silence this exception. Notice that this has no effect on - platforms that don't support os.symlink. - - The optional ignore argument is a callable. If given, it - is called with the `src` parameter, which is the directory - being visited by copytree(), and `names` which is the list of - `src` contents, as returned by os.listdir(): - - callable(src, names) -> ignored_names - - Since copytree() is called recursively, the callable will be - called once for each directory that is copied. It returns a - list of names relative to the `src` directory that should - not be copied. - - The optional copy_function argument is a callable that will be used - to copy each file. It will be called with the source path and the - destination path as arguments. By default, copy2() is used, but any - function that supports the same signature (like copy()) can be used. - - """ - names = os.listdir(src) - if ignore is not None: - ignored_names = ignore(src, names) - else: - ignored_names = set() - - os.makedirs(dst) - errors = [] - for name in names: - if name in ignored_names: - continue - srcname = os.path.join(src, name) - dstname = os.path.join(dst, name) - try: - if os.path.islink(srcname): - linkto = os.readlink(srcname) - if symlinks: - os.symlink(linkto, dstname) - else: - # ignore dangling symlink if the flag is on - if not os.path.exists(linkto) and ignore_dangling_symlinks: - continue - # otherwise let the copy occurs. copy2 will raise an error - copy_function(srcname, dstname) - elif os.path.isdir(srcname): - copytree(srcname, dstname, symlinks, ignore, copy_function) - else: - # Will raise a SpecialFileError for unsupported file types - copy_function(srcname, dstname) - # catch the Error from the recursive copytree so that we can - # continue with other files - except Error as err: - errors.extend(err.args[0]) - except EnvironmentError as why: - errors.append((srcname, dstname, str(why))) - try: - copystat(src, dst) - except OSError as why: - if WindowsError is not None and isinstance(why, WindowsError): - # Copying file access times may fail on Windows - pass - else: - errors.extend((src, dst, str(why))) - if errors: - raise Error(errors) - -def rmtree(path, ignore_errors=False, onerror=None): - """Recursively delete a directory tree. - - If ignore_errors is set, errors are ignored; otherwise, if onerror - is set, it is called to handle the error with arguments (func, - path, exc_info) where func is os.listdir, os.remove, or os.rmdir; - path is the argument to that function that caused it to fail; and - exc_info is a tuple returned by sys.exc_info(). If ignore_errors - is false and onerror is None, an exception is raised. - - """ - if ignore_errors: - def onerror(*args): - pass - elif onerror is None: - def onerror(*args): - raise - try: - if os.path.islink(path): - # symlinks to directories are forbidden, see bug #1669 - raise OSError("Cannot call rmtree on a symbolic link") - except OSError: - onerror(os.path.islink, path, sys.exc_info()) - # can't continue even if onerror hook returns - return - names = [] - try: - names = os.listdir(path) - except os.error: - onerror(os.listdir, path, sys.exc_info()) - for name in names: - fullname = os.path.join(path, name) - try: - mode = os.lstat(fullname).st_mode - except os.error: - mode = 0 - if stat.S_ISDIR(mode): - rmtree(fullname, ignore_errors, onerror) - else: - try: - os.remove(fullname) - except os.error: - onerror(os.remove, fullname, sys.exc_info()) - try: - os.rmdir(path) - except os.error: - onerror(os.rmdir, path, sys.exc_info()) - - -def _basename(path): - # A basename() variant which first strips the trailing slash, if present. - # Thus we always get the last component of the path, even for directories. - return os.path.basename(path.rstrip(os.path.sep)) - -def move(src, dst): - """Recursively move a file or directory to another location. This is - similar to the Unix "mv" command. - - If the destination is a directory or a symlink to a directory, the source - is moved inside the directory. The destination path must not already - exist. - - If the destination already exists but is not a directory, it may be - overwritten depending on os.rename() semantics. - - If the destination is on our current filesystem, then rename() is used. - Otherwise, src is copied to the destination and then removed. - A lot more could be done here... A look at a mv.c shows a lot of - the issues this implementation glosses over. - - """ - real_dst = dst - if os.path.isdir(dst): - if _samefile(src, dst): - # We might be on a case insensitive filesystem, - # perform the rename anyway. - os.rename(src, dst) - return - - real_dst = os.path.join(dst, _basename(src)) - if os.path.exists(real_dst): - raise Error("Destination path '%s' already exists" % real_dst) - try: - os.rename(src, real_dst) - except OSError: - if os.path.isdir(src): - if _destinsrc(src, dst): - raise Error("Cannot move a directory '%s' into itself '%s'." % (src, dst)) - copytree(src, real_dst, symlinks=True) - rmtree(src) - else: - copy2(src, real_dst) - os.unlink(src) - -def _destinsrc(src, dst): - src = abspath(src) - dst = abspath(dst) - if not src.endswith(os.path.sep): - src += os.path.sep - if not dst.endswith(os.path.sep): - dst += os.path.sep - return dst.startswith(src) - -def _get_gid(name): - """Returns a gid, given a group name.""" - if getgrnam is None or name is None: - return None - try: - result = getgrnam(name) - except KeyError: - result = None - if result is not None: - return result[2] - return None - -def _get_uid(name): - """Returns an uid, given a user name.""" - if getpwnam is None or name is None: - return None - try: - result = getpwnam(name) - except KeyError: - result = None - if result is not None: - return result[2] - return None - -def _make_tarball(base_name, base_dir, compress="gzip", verbose=0, dry_run=0, - owner=None, group=None, logger=None): - """Create a (possibly compressed) tar file from all the files under - 'base_dir'. - - 'compress' must be "gzip" (the default), "bzip2", or None. - - 'owner' and 'group' can be used to define an owner and a group for the - archive that is being built. If not provided, the current owner and group - will be used. - - The output tar file will be named 'base_name' + ".tar", possibly plus - the appropriate compression extension (".gz", or ".bz2"). - - Returns the output filename. - """ - tar_compression = {'gzip': 'gz', None: ''} - compress_ext = {'gzip': '.gz'} - - if _BZ2_SUPPORTED: - tar_compression['bzip2'] = 'bz2' - compress_ext['bzip2'] = '.bz2' - - # flags for compression program, each element of list will be an argument - if compress is not None and compress not in compress_ext: - raise ValueError("bad value for 'compress', or compression format not " - "supported : {0}".format(compress)) - - archive_name = base_name + '.tar' + compress_ext.get(compress, '') - archive_dir = os.path.dirname(archive_name) - - if not os.path.exists(archive_dir): - if logger is not None: - logger.info("creating %s", archive_dir) - if not dry_run: - os.makedirs(archive_dir) - - # creating the tarball - if logger is not None: - logger.info('Creating tar archive') - - uid = _get_uid(owner) - gid = _get_gid(group) - - def _set_uid_gid(tarinfo): - if gid is not None: - tarinfo.gid = gid - tarinfo.gname = group - if uid is not None: - tarinfo.uid = uid - tarinfo.uname = owner - return tarinfo - - if not dry_run: - tar = tarfile.open(archive_name, 'w|%s' % tar_compression[compress]) - try: - tar.add(base_dir, filter=_set_uid_gid) - finally: - tar.close() - - return archive_name - -def _call_external_zip(base_dir, zip_filename, verbose=False, dry_run=False): - # XXX see if we want to keep an external call here - if verbose: - zipoptions = "-r" - else: - zipoptions = "-rq" - from distutils.errors import DistutilsExecError - from distutils.spawn import spawn - try: - spawn(["zip", zipoptions, zip_filename, base_dir], dry_run=dry_run) - except DistutilsExecError: - # XXX really should distinguish between "couldn't find - # external 'zip' command" and "zip failed". - raise ExecError("unable to create zip file '%s': " - "could neither import the 'zipfile' module nor " - "find a standalone zip utility") % zip_filename - -def _make_zipfile(base_name, base_dir, verbose=0, dry_run=0, logger=None): - """Create a zip file from all the files under 'base_dir'. - - The output zip file will be named 'base_name' + ".zip". Uses either the - "zipfile" Python module (if available) or the InfoZIP "zip" utility - (if installed and found on the default search path). If neither tool is - available, raises ExecError. Returns the name of the output zip - file. - """ - zip_filename = base_name + ".zip" - archive_dir = os.path.dirname(base_name) - - if not os.path.exists(archive_dir): - if logger is not None: - logger.info("creating %s", archive_dir) - if not dry_run: - os.makedirs(archive_dir) - - # If zipfile module is not available, try spawning an external 'zip' - # command. - try: - import zipfile - except ImportError: - zipfile = None - - if zipfile is None: - _call_external_zip(base_dir, zip_filename, verbose, dry_run) - else: - if logger is not None: - logger.info("creating '%s' and adding '%s' to it", - zip_filename, base_dir) - - if not dry_run: - zip = zipfile.ZipFile(zip_filename, "w", - compression=zipfile.ZIP_DEFLATED) - - for dirpath, dirnames, filenames in os.walk(base_dir): - for name in filenames: - path = os.path.normpath(os.path.join(dirpath, name)) - if os.path.isfile(path): - zip.write(path, path) - if logger is not None: - logger.info("adding '%s'", path) - zip.close() - - return zip_filename - -_ARCHIVE_FORMATS = { - 'gztar': (_make_tarball, [('compress', 'gzip')], "gzip'ed tar-file"), - 'bztar': (_make_tarball, [('compress', 'bzip2')], "bzip2'ed tar-file"), - 'tar': (_make_tarball, [('compress', None)], "uncompressed tar file"), - 'zip': (_make_zipfile, [], "ZIP file"), - } - -if _BZ2_SUPPORTED: - _ARCHIVE_FORMATS['bztar'] = (_make_tarball, [('compress', 'bzip2')], - "bzip2'ed tar-file") - -def get_archive_formats(): - """Returns a list of supported formats for archiving and unarchiving. - - Each element of the returned sequence is a tuple (name, description) - """ - formats = [(name, registry[2]) for name, registry in - _ARCHIVE_FORMATS.items()] - formats.sort() - return formats - -def register_archive_format(name, function, extra_args=None, description=''): - """Registers an archive format. - - name is the name of the format. function is the callable that will be - used to create archives. If provided, extra_args is a sequence of - (name, value) tuples that will be passed as arguments to the callable. - description can be provided to describe the format, and will be returned - by the get_archive_formats() function. - """ - if extra_args is None: - extra_args = [] - if not isinstance(function, collections.Callable): - raise TypeError('The %s object is not callable' % function) - if not isinstance(extra_args, (tuple, list)): - raise TypeError('extra_args needs to be a sequence') - for element in extra_args: - if not isinstance(element, (tuple, list)) or len(element) !=2: - raise TypeError('extra_args elements are : (arg_name, value)') - - _ARCHIVE_FORMATS[name] = (function, extra_args, description) - -def unregister_archive_format(name): - del _ARCHIVE_FORMATS[name] - -def make_archive(base_name, format, root_dir=None, base_dir=None, verbose=0, - dry_run=0, owner=None, group=None, logger=None): - """Create an archive file (eg. zip or tar). - - 'base_name' is the name of the file to create, minus any format-specific - extension; 'format' is the archive format: one of "zip", "tar", "bztar" - or "gztar". - - 'root_dir' is a directory that will be the root directory of the - archive; ie. we typically chdir into 'root_dir' before creating the - archive. 'base_dir' is the directory where we start archiving from; - ie. 'base_dir' will be the common prefix of all files and - directories in the archive. 'root_dir' and 'base_dir' both default - to the current directory. Returns the name of the archive file. - - 'owner' and 'group' are used when creating a tar archive. By default, - uses the current owner and group. - """ - save_cwd = os.getcwd() - if root_dir is not None: - if logger is not None: - logger.debug("changing into '%s'", root_dir) - base_name = os.path.abspath(base_name) - if not dry_run: - os.chdir(root_dir) - - if base_dir is None: - base_dir = os.curdir - - kwargs = {'dry_run': dry_run, 'logger': logger} - - try: - format_info = _ARCHIVE_FORMATS[format] - except KeyError: - raise ValueError("unknown archive format '%s'" % format) - - func = format_info[0] - for arg, val in format_info[1]: - kwargs[arg] = val - - if format != 'zip': - kwargs['owner'] = owner - kwargs['group'] = group - - try: - filename = func(base_name, base_dir, **kwargs) - finally: - if root_dir is not None: - if logger is not None: - logger.debug("changing back to '%s'", save_cwd) - os.chdir(save_cwd) - - return filename - - -def get_unpack_formats(): - """Returns a list of supported formats for unpacking. - - Each element of the returned sequence is a tuple - (name, extensions, description) - """ - formats = [(name, info[0], info[3]) for name, info in - _UNPACK_FORMATS.items()] - formats.sort() - return formats - -def _check_unpack_options(extensions, function, extra_args): - """Checks what gets registered as an unpacker.""" - # first make sure no other unpacker is registered for this extension - existing_extensions = {} - for name, info in _UNPACK_FORMATS.items(): - for ext in info[0]: - existing_extensions[ext] = name - - for extension in extensions: - if extension in existing_extensions: - msg = '%s is already registered for "%s"' - raise RegistryError(msg % (extension, - existing_extensions[extension])) - - if not isinstance(function, collections.Callable): - raise TypeError('The registered function must be a callable') - - -def register_unpack_format(name, extensions, function, extra_args=None, - description=''): - """Registers an unpack format. - - `name` is the name of the format. `extensions` is a list of extensions - corresponding to the format. - - `function` is the callable that will be - used to unpack archives. The callable will receive archives to unpack. - If it's unable to handle an archive, it needs to raise a ReadError - exception. - - If provided, `extra_args` is a sequence of - (name, value) tuples that will be passed as arguments to the callable. - description can be provided to describe the format, and will be returned - by the get_unpack_formats() function. - """ - if extra_args is None: - extra_args = [] - _check_unpack_options(extensions, function, extra_args) - _UNPACK_FORMATS[name] = extensions, function, extra_args, description - -def unregister_unpack_format(name): - """Removes the pack format from the registry.""" - del _UNPACK_FORMATS[name] - -def _ensure_directory(path): - """Ensure that the parent directory of `path` exists""" - dirname = os.path.dirname(path) - if not os.path.isdir(dirname): - os.makedirs(dirname) - -def _unpack_zipfile(filename, extract_dir): - """Unpack zip `filename` to `extract_dir` - """ - try: - import zipfile - except ImportError: - raise ReadError('zlib not supported, cannot unpack this archive.') - - if not zipfile.is_zipfile(filename): - raise ReadError("%s is not a zip file" % filename) - - zip = zipfile.ZipFile(filename) - try: - for info in zip.infolist(): - name = info.filename - - # don't extract absolute paths or ones with .. in them - if name.startswith('/') or '..' in name: - continue - - target = os.path.join(extract_dir, *name.split('/')) - if not target: - continue - - _ensure_directory(target) - if not name.endswith('/'): - # file - data = zip.read(info.filename) - f = open(target, 'wb') - try: - f.write(data) - finally: - f.close() - del data - finally: - zip.close() - -def _unpack_tarfile(filename, extract_dir): - """Unpack tar/tar.gz/tar.bz2 `filename` to `extract_dir` - """ - try: - tarobj = tarfile.open(filename) - except tarfile.TarError: - raise ReadError( - "%s is not a compressed or uncompressed tar file" % filename) - try: - tarobj.extractall(extract_dir) - finally: - tarobj.close() - -_UNPACK_FORMATS = { - 'gztar': (['.tar.gz', '.tgz'], _unpack_tarfile, [], "gzip'ed tar-file"), - 'tar': (['.tar'], _unpack_tarfile, [], "uncompressed tar file"), - 'zip': (['.zip'], _unpack_zipfile, [], "ZIP file") - } - -if _BZ2_SUPPORTED: - _UNPACK_FORMATS['bztar'] = (['.bz2'], _unpack_tarfile, [], - "bzip2'ed tar-file") - -def _find_unpack_format(filename): - for name, info in _UNPACK_FORMATS.items(): - for extension in info[0]: - if filename.endswith(extension): - return name - return None - -def unpack_archive(filename, extract_dir=None, format=None): - """Unpack an archive. - - `filename` is the name of the archive. - - `extract_dir` is the name of the target directory, where the archive - is unpacked. If not provided, the current working directory is used. - - `format` is the archive format: one of "zip", "tar", or "gztar". Or any - other registered format. If not provided, unpack_archive will use the - filename extension and see if an unpacker was registered for that - extension. - - In case none is found, a ValueError is raised. - """ - if extract_dir is None: - extract_dir = os.getcwd() - - if format is not None: - try: - format_info = _UNPACK_FORMATS[format] - except KeyError: - raise ValueError("Unknown unpack format '{0}'".format(format)) - - func = format_info[1] - func(filename, extract_dir, **dict(format_info[2])) - else: - # we need to look at the registered unpackers supported extensions - format = _find_unpack_format(filename) - if format is None: - raise ReadError("Unknown archive format '{0}'".format(filename)) - - func = _UNPACK_FORMATS[format][1] - kwargs = dict(_UNPACK_FORMATS[format][2]) - func(filename, extract_dir, **kwargs) diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/_backport/sysconfig.cfg b/src/fetchcode/vcs/pip/_vendor/distlib/_backport/sysconfig.cfg deleted file mode 100644 index 1746bd01..00000000 --- a/src/fetchcode/vcs/pip/_vendor/distlib/_backport/sysconfig.cfg +++ /dev/null @@ -1,84 +0,0 @@ -[posix_prefix] -# Configuration directories. Some of these come straight out of the -# configure script. They are for implementing the other variables, not to -# be used directly in [resource_locations]. -confdir = /etc -datadir = /usr/share -libdir = /usr/lib -statedir = /var -# User resource directory -local = ~/.local/{distribution.name} - -stdlib = {base}/lib/python{py_version_short} -platstdlib = {platbase}/lib/python{py_version_short} -purelib = {base}/lib/python{py_version_short}/site-packages -platlib = {platbase}/lib/python{py_version_short}/site-packages -include = {base}/include/python{py_version_short}{abiflags} -platinclude = {platbase}/include/python{py_version_short}{abiflags} -data = {base} - -[posix_home] -stdlib = {base}/lib/python -platstdlib = {base}/lib/python -purelib = {base}/lib/python -platlib = {base}/lib/python -include = {base}/include/python -platinclude = {base}/include/python -scripts = {base}/bin -data = {base} - -[nt] -stdlib = {base}/Lib -platstdlib = {base}/Lib -purelib = {base}/Lib/site-packages -platlib = {base}/Lib/site-packages -include = {base}/Include -platinclude = {base}/Include -scripts = {base}/Scripts -data = {base} - -[os2] -stdlib = {base}/Lib -platstdlib = {base}/Lib -purelib = {base}/Lib/site-packages -platlib = {base}/Lib/site-packages -include = {base}/Include -platinclude = {base}/Include -scripts = {base}/Scripts -data = {base} - -[os2_home] -stdlib = {userbase}/lib/python{py_version_short} -platstdlib = {userbase}/lib/python{py_version_short} -purelib = {userbase}/lib/python{py_version_short}/site-packages -platlib = {userbase}/lib/python{py_version_short}/site-packages -include = {userbase}/include/python{py_version_short} -scripts = {userbase}/bin -data = {userbase} - -[nt_user] -stdlib = {userbase}/Python{py_version_nodot} -platstdlib = {userbase}/Python{py_version_nodot} -purelib = {userbase}/Python{py_version_nodot}/site-packages -platlib = {userbase}/Python{py_version_nodot}/site-packages -include = {userbase}/Python{py_version_nodot}/Include -scripts = {userbase}/Scripts -data = {userbase} - -[posix_user] -stdlib = {userbase}/lib/python{py_version_short} -platstdlib = {userbase}/lib/python{py_version_short} -purelib = {userbase}/lib/python{py_version_short}/site-packages -platlib = {userbase}/lib/python{py_version_short}/site-packages -include = {userbase}/include/python{py_version_short} -scripts = {userbase}/bin -data = {userbase} - -[osx_framework_user] -stdlib = {userbase}/lib/python -platstdlib = {userbase}/lib/python -purelib = {userbase}/lib/python/site-packages -platlib = {userbase}/lib/python/site-packages -include = {userbase}/include -scripts = {userbase}/bin -data = {userbase} diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/_backport/sysconfig.py b/src/fetchcode/vcs/pip/_vendor/distlib/_backport/sysconfig.py deleted file mode 100644 index b470a373..00000000 --- a/src/fetchcode/vcs/pip/_vendor/distlib/_backport/sysconfig.py +++ /dev/null @@ -1,786 +0,0 @@ -# -*- coding: utf-8 -*- -# -# Copyright (C) 2012 The Python Software Foundation. -# See LICENSE.txt and CONTRIBUTORS.txt. -# -"""Access to Python's configuration information.""" - -import codecs -import os -import re -import sys -from os.path import pardir, realpath -try: - import configparser -except ImportError: - import ConfigParser as configparser - - -__all__ = [ - 'get_config_h_filename', - 'get_config_var', - 'get_config_vars', - 'get_makefile_filename', - 'get_path', - 'get_path_names', - 'get_paths', - 'get_platform', - 'get_python_version', - 'get_scheme_names', - 'parse_config_h', -] - - -def _safe_realpath(path): - try: - return realpath(path) - except OSError: - return path - - -if sys.executable: - _PROJECT_BASE = os.path.dirname(_safe_realpath(sys.executable)) -else: - # sys.executable can be empty if argv[0] has been changed and Python is - # unable to retrieve the real program name - _PROJECT_BASE = _safe_realpath(os.getcwd()) - -if os.name == "nt" and "pcbuild" in _PROJECT_BASE[-8:].lower(): - _PROJECT_BASE = _safe_realpath(os.path.join(_PROJECT_BASE, pardir)) -# PC/VS7.1 -if os.name == "nt" and "\\pc\\v" in _PROJECT_BASE[-10:].lower(): - _PROJECT_BASE = _safe_realpath(os.path.join(_PROJECT_BASE, pardir, pardir)) -# PC/AMD64 -if os.name == "nt" and "\\pcbuild\\amd64" in _PROJECT_BASE[-14:].lower(): - _PROJECT_BASE = _safe_realpath(os.path.join(_PROJECT_BASE, pardir, pardir)) - - -def is_python_build(): - for fn in ("Setup.dist", "Setup.local"): - if os.path.isfile(os.path.join(_PROJECT_BASE, "Modules", fn)): - return True - return False - -_PYTHON_BUILD = is_python_build() - -_cfg_read = False - -def _ensure_cfg_read(): - global _cfg_read - if not _cfg_read: - from ..resources import finder - backport_package = __name__.rsplit('.', 1)[0] - _finder = finder(backport_package) - _cfgfile = _finder.find('sysconfig.cfg') - assert _cfgfile, 'sysconfig.cfg exists' - with _cfgfile.as_stream() as s: - _SCHEMES.readfp(s) - if _PYTHON_BUILD: - for scheme in ('posix_prefix', 'posix_home'): - _SCHEMES.set(scheme, 'include', '{srcdir}/Include') - _SCHEMES.set(scheme, 'platinclude', '{projectbase}/.') - - _cfg_read = True - - -_SCHEMES = configparser.RawConfigParser() -_VAR_REPL = re.compile(r'\{([^{]*?)\}') - -def _expand_globals(config): - _ensure_cfg_read() - if config.has_section('globals'): - globals = config.items('globals') - else: - globals = tuple() - - sections = config.sections() - for section in sections: - if section == 'globals': - continue - for option, value in globals: - if config.has_option(section, option): - continue - config.set(section, option, value) - config.remove_section('globals') - - # now expanding local variables defined in the cfg file - # - for section in config.sections(): - variables = dict(config.items(section)) - - def _replacer(matchobj): - name = matchobj.group(1) - if name in variables: - return variables[name] - return matchobj.group(0) - - for option, value in config.items(section): - config.set(section, option, _VAR_REPL.sub(_replacer, value)) - -#_expand_globals(_SCHEMES) - -_PY_VERSION = '%s.%s.%s' % sys.version_info[:3] -_PY_VERSION_SHORT = '%s.%s' % sys.version_info[:2] -_PY_VERSION_SHORT_NO_DOT = '%s%s' % sys.version_info[:2] -_PREFIX = os.path.normpath(sys.prefix) -_EXEC_PREFIX = os.path.normpath(sys.exec_prefix) -_CONFIG_VARS = None -_USER_BASE = None - - -def _subst_vars(path, local_vars): - """In the string `path`, replace tokens like {some.thing} with the - corresponding value from the map `local_vars`. - - If there is no corresponding value, leave the token unchanged. - """ - def _replacer(matchobj): - name = matchobj.group(1) - if name in local_vars: - return local_vars[name] - elif name in os.environ: - return os.environ[name] - return matchobj.group(0) - return _VAR_REPL.sub(_replacer, path) - - -def _extend_dict(target_dict, other_dict): - target_keys = target_dict.keys() - for key, value in other_dict.items(): - if key in target_keys: - continue - target_dict[key] = value - - -def _expand_vars(scheme, vars): - res = {} - if vars is None: - vars = {} - _extend_dict(vars, get_config_vars()) - - for key, value in _SCHEMES.items(scheme): - if os.name in ('posix', 'nt'): - value = os.path.expanduser(value) - res[key] = os.path.normpath(_subst_vars(value, vars)) - return res - - -def format_value(value, vars): - def _replacer(matchobj): - name = matchobj.group(1) - if name in vars: - return vars[name] - return matchobj.group(0) - return _VAR_REPL.sub(_replacer, value) - - -def _get_default_scheme(): - if os.name == 'posix': - # the default scheme for posix is posix_prefix - return 'posix_prefix' - return os.name - - -def _getuserbase(): - env_base = os.environ.get("PYTHONUSERBASE", None) - - def joinuser(*args): - return os.path.expanduser(os.path.join(*args)) - - # what about 'os2emx', 'riscos' ? - if os.name == "nt": - base = os.environ.get("APPDATA") or "~" - if env_base: - return env_base - else: - return joinuser(base, "Python") - - if sys.platform == "darwin": - framework = get_config_var("PYTHONFRAMEWORK") - if framework: - if env_base: - return env_base - else: - return joinuser("~", "Library", framework, "%d.%d" % - sys.version_info[:2]) - - if env_base: - return env_base - else: - return joinuser("~", ".local") - - -def _parse_makefile(filename, vars=None): - """Parse a Makefile-style file. - - A dictionary containing name/value pairs is returned. If an - optional dictionary is passed in as the second argument, it is - used instead of a new dictionary. - """ - # Regexes needed for parsing Makefile (and similar syntaxes, - # like old-style Setup files). - _variable_rx = re.compile(r"([a-zA-Z][a-zA-Z0-9_]+)\s*=\s*(.*)") - _findvar1_rx = re.compile(r"\$\(([A-Za-z][A-Za-z0-9_]*)\)") - _findvar2_rx = re.compile(r"\${([A-Za-z][A-Za-z0-9_]*)}") - - if vars is None: - vars = {} - done = {} - notdone = {} - - with codecs.open(filename, encoding='utf-8', errors="surrogateescape") as f: - lines = f.readlines() - - for line in lines: - if line.startswith('#') or line.strip() == '': - continue - m = _variable_rx.match(line) - if m: - n, v = m.group(1, 2) - v = v.strip() - # `$$' is a literal `$' in make - tmpv = v.replace('$$', '') - - if "$" in tmpv: - notdone[n] = v - else: - try: - v = int(v) - except ValueError: - # insert literal `$' - done[n] = v.replace('$$', '$') - else: - done[n] = v - - # do variable interpolation here - variables = list(notdone.keys()) - - # Variables with a 'PY_' prefix in the makefile. These need to - # be made available without that prefix through sysconfig. - # Special care is needed to ensure that variable expansion works, even - # if the expansion uses the name without a prefix. - renamed_variables = ('CFLAGS', 'LDFLAGS', 'CPPFLAGS') - - while len(variables) > 0: - for name in tuple(variables): - value = notdone[name] - m = _findvar1_rx.search(value) or _findvar2_rx.search(value) - if m is not None: - n = m.group(1) - found = True - if n in done: - item = str(done[n]) - elif n in notdone: - # get it on a subsequent round - found = False - elif n in os.environ: - # do it like make: fall back to environment - item = os.environ[n] - - elif n in renamed_variables: - if (name.startswith('PY_') and - name[3:] in renamed_variables): - item = "" - - elif 'PY_' + n in notdone: - found = False - - else: - item = str(done['PY_' + n]) - - else: - done[n] = item = "" - - if found: - after = value[m.end():] - value = value[:m.start()] + item + after - if "$" in after: - notdone[name] = value - else: - try: - value = int(value) - except ValueError: - done[name] = value.strip() - else: - done[name] = value - variables.remove(name) - - if (name.startswith('PY_') and - name[3:] in renamed_variables): - - name = name[3:] - if name not in done: - done[name] = value - - else: - # bogus variable reference (e.g. "prefix=$/opt/python"); - # just drop it since we can't deal - done[name] = value - variables.remove(name) - - # strip spurious spaces - for k, v in done.items(): - if isinstance(v, str): - done[k] = v.strip() - - # save the results in the global dictionary - vars.update(done) - return vars - - -def get_makefile_filename(): - """Return the path of the Makefile.""" - if _PYTHON_BUILD: - return os.path.join(_PROJECT_BASE, "Makefile") - if hasattr(sys, 'abiflags'): - config_dir_name = 'config-%s%s' % (_PY_VERSION_SHORT, sys.abiflags) - else: - config_dir_name = 'config' - return os.path.join(get_path('stdlib'), config_dir_name, 'Makefile') - - -def _init_posix(vars): - """Initialize the module as appropriate for POSIX systems.""" - # load the installed Makefile: - makefile = get_makefile_filename() - try: - _parse_makefile(makefile, vars) - except IOError as e: - msg = "invalid Python installation: unable to open %s" % makefile - if hasattr(e, "strerror"): - msg = msg + " (%s)" % e.strerror - raise IOError(msg) - # load the installed pyconfig.h: - config_h = get_config_h_filename() - try: - with open(config_h) as f: - parse_config_h(f, vars) - except IOError as e: - msg = "invalid Python installation: unable to open %s" % config_h - if hasattr(e, "strerror"): - msg = msg + " (%s)" % e.strerror - raise IOError(msg) - # On AIX, there are wrong paths to the linker scripts in the Makefile - # -- these paths are relative to the Python source, but when installed - # the scripts are in another directory. - if _PYTHON_BUILD: - vars['LDSHARED'] = vars['BLDSHARED'] - - -def _init_non_posix(vars): - """Initialize the module as appropriate for NT""" - # set basic install directories - vars['LIBDEST'] = get_path('stdlib') - vars['BINLIBDEST'] = get_path('platstdlib') - vars['INCLUDEPY'] = get_path('include') - vars['SO'] = '.pyd' - vars['EXE'] = '.exe' - vars['VERSION'] = _PY_VERSION_SHORT_NO_DOT - vars['BINDIR'] = os.path.dirname(_safe_realpath(sys.executable)) - -# -# public APIs -# - - -def parse_config_h(fp, vars=None): - """Parse a config.h-style file. - - A dictionary containing name/value pairs is returned. If an - optional dictionary is passed in as the second argument, it is - used instead of a new dictionary. - """ - if vars is None: - vars = {} - define_rx = re.compile("#define ([A-Z][A-Za-z0-9_]+) (.*)\n") - undef_rx = re.compile("/[*] #undef ([A-Z][A-Za-z0-9_]+) [*]/\n") - - while True: - line = fp.readline() - if not line: - break - m = define_rx.match(line) - if m: - n, v = m.group(1, 2) - try: - v = int(v) - except ValueError: - pass - vars[n] = v - else: - m = undef_rx.match(line) - if m: - vars[m.group(1)] = 0 - return vars - - -def get_config_h_filename(): - """Return the path of pyconfig.h.""" - if _PYTHON_BUILD: - if os.name == "nt": - inc_dir = os.path.join(_PROJECT_BASE, "PC") - else: - inc_dir = _PROJECT_BASE - else: - inc_dir = get_path('platinclude') - return os.path.join(inc_dir, 'pyconfig.h') - - -def get_scheme_names(): - """Return a tuple containing the schemes names.""" - return tuple(sorted(_SCHEMES.sections())) - - -def get_path_names(): - """Return a tuple containing the paths names.""" - # xxx see if we want a static list - return _SCHEMES.options('posix_prefix') - - -def get_paths(scheme=_get_default_scheme(), vars=None, expand=True): - """Return a mapping containing an install scheme. - - ``scheme`` is the install scheme name. If not provided, it will - return the default scheme for the current platform. - """ - _ensure_cfg_read() - if expand: - return _expand_vars(scheme, vars) - else: - return dict(_SCHEMES.items(scheme)) - - -def get_path(name, scheme=_get_default_scheme(), vars=None, expand=True): - """Return a path corresponding to the scheme. - - ``scheme`` is the install scheme name. - """ - return get_paths(scheme, vars, expand)[name] - - -def get_config_vars(*args): - """With no arguments, return a dictionary of all configuration - variables relevant for the current platform. - - On Unix, this means every variable defined in Python's installed Makefile; - On Windows and Mac OS it's a much smaller set. - - With arguments, return a list of values that result from looking up - each argument in the configuration variable dictionary. - """ - global _CONFIG_VARS - if _CONFIG_VARS is None: - _CONFIG_VARS = {} - # Normalized versions of prefix and exec_prefix are handy to have; - # in fact, these are the standard versions used most places in the - # distutils2 module. - _CONFIG_VARS['prefix'] = _PREFIX - _CONFIG_VARS['exec_prefix'] = _EXEC_PREFIX - _CONFIG_VARS['py_version'] = _PY_VERSION - _CONFIG_VARS['py_version_short'] = _PY_VERSION_SHORT - _CONFIG_VARS['py_version_nodot'] = _PY_VERSION[0] + _PY_VERSION[2] - _CONFIG_VARS['base'] = _PREFIX - _CONFIG_VARS['platbase'] = _EXEC_PREFIX - _CONFIG_VARS['projectbase'] = _PROJECT_BASE - try: - _CONFIG_VARS['abiflags'] = sys.abiflags - except AttributeError: - # sys.abiflags may not be defined on all platforms. - _CONFIG_VARS['abiflags'] = '' - - if os.name in ('nt', 'os2'): - _init_non_posix(_CONFIG_VARS) - if os.name == 'posix': - _init_posix(_CONFIG_VARS) - # Setting 'userbase' is done below the call to the - # init function to enable using 'get_config_var' in - # the init-function. - if sys.version >= '2.6': - _CONFIG_VARS['userbase'] = _getuserbase() - - if 'srcdir' not in _CONFIG_VARS: - _CONFIG_VARS['srcdir'] = _PROJECT_BASE - else: - _CONFIG_VARS['srcdir'] = _safe_realpath(_CONFIG_VARS['srcdir']) - - # Convert srcdir into an absolute path if it appears necessary. - # Normally it is relative to the build directory. However, during - # testing, for example, we might be running a non-installed python - # from a different directory. - if _PYTHON_BUILD and os.name == "posix": - base = _PROJECT_BASE - try: - cwd = os.getcwd() - except OSError: - cwd = None - if (not os.path.isabs(_CONFIG_VARS['srcdir']) and - base != cwd): - # srcdir is relative and we are not in the same directory - # as the executable. Assume executable is in the build - # directory and make srcdir absolute. - srcdir = os.path.join(base, _CONFIG_VARS['srcdir']) - _CONFIG_VARS['srcdir'] = os.path.normpath(srcdir) - - if sys.platform == 'darwin': - kernel_version = os.uname()[2] # Kernel version (8.4.3) - major_version = int(kernel_version.split('.')[0]) - - if major_version < 8: - # On Mac OS X before 10.4, check if -arch and -isysroot - # are in CFLAGS or LDFLAGS and remove them if they are. - # This is needed when building extensions on a 10.3 system - # using a universal build of python. - for key in ('LDFLAGS', 'BASECFLAGS', - # a number of derived variables. These need to be - # patched up as well. - 'CFLAGS', 'PY_CFLAGS', 'BLDSHARED'): - flags = _CONFIG_VARS[key] - flags = re.sub(r'-arch\s+\w+\s', ' ', flags) - flags = re.sub('-isysroot [^ \t]*', ' ', flags) - _CONFIG_VARS[key] = flags - else: - # Allow the user to override the architecture flags using - # an environment variable. - # NOTE: This name was introduced by Apple in OSX 10.5 and - # is used by several scripting languages distributed with - # that OS release. - if 'ARCHFLAGS' in os.environ: - arch = os.environ['ARCHFLAGS'] - for key in ('LDFLAGS', 'BASECFLAGS', - # a number of derived variables. These need to be - # patched up as well. - 'CFLAGS', 'PY_CFLAGS', 'BLDSHARED'): - - flags = _CONFIG_VARS[key] - flags = re.sub(r'-arch\s+\w+\s', ' ', flags) - flags = flags + ' ' + arch - _CONFIG_VARS[key] = flags - - # If we're on OSX 10.5 or later and the user tries to - # compiles an extension using an SDK that is not present - # on the current machine it is better to not use an SDK - # than to fail. - # - # The major usecase for this is users using a Python.org - # binary installer on OSX 10.6: that installer uses - # the 10.4u SDK, but that SDK is not installed by default - # when you install Xcode. - # - CFLAGS = _CONFIG_VARS.get('CFLAGS', '') - m = re.search(r'-isysroot\s+(\S+)', CFLAGS) - if m is not None: - sdk = m.group(1) - if not os.path.exists(sdk): - for key in ('LDFLAGS', 'BASECFLAGS', - # a number of derived variables. These need to be - # patched up as well. - 'CFLAGS', 'PY_CFLAGS', 'BLDSHARED'): - - flags = _CONFIG_VARS[key] - flags = re.sub(r'-isysroot\s+\S+(\s|$)', ' ', flags) - _CONFIG_VARS[key] = flags - - if args: - vals = [] - for name in args: - vals.append(_CONFIG_VARS.get(name)) - return vals - else: - return _CONFIG_VARS - - -def get_config_var(name): - """Return the value of a single variable using the dictionary returned by - 'get_config_vars()'. - - Equivalent to get_config_vars().get(name) - """ - return get_config_vars().get(name) - - -def get_platform(): - """Return a string that identifies the current platform. - - This is used mainly to distinguish platform-specific build directories and - platform-specific built distributions. Typically includes the OS name - and version and the architecture (as supplied by 'os.uname()'), - although the exact information included depends on the OS; eg. for IRIX - the architecture isn't particularly important (IRIX only runs on SGI - hardware), but for Linux the kernel version isn't particularly - important. - - Examples of returned values: - linux-i586 - linux-alpha (?) - solaris-2.6-sun4u - irix-5.3 - irix64-6.2 - - Windows will return one of: - win-amd64 (64bit Windows on AMD64 (aka x86_64, Intel64, EM64T, etc) - win-ia64 (64bit Windows on Itanium) - win32 (all others - specifically, sys.platform is returned) - - For other non-POSIX platforms, currently just returns 'sys.platform'. - """ - if os.name == 'nt': - # sniff sys.version for architecture. - prefix = " bit (" - i = sys.version.find(prefix) - if i == -1: - return sys.platform - j = sys.version.find(")", i) - look = sys.version[i+len(prefix):j].lower() - if look == 'amd64': - return 'win-amd64' - if look == 'itanium': - return 'win-ia64' - return sys.platform - - if os.name != "posix" or not hasattr(os, 'uname'): - # XXX what about the architecture? NT is Intel or Alpha, - # Mac OS is M68k or PPC, etc. - return sys.platform - - # Try to distinguish various flavours of Unix - osname, host, release, version, machine = os.uname() - - # Convert the OS name to lowercase, remove '/' characters - # (to accommodate BSD/OS), and translate spaces (for "Power Macintosh") - osname = osname.lower().replace('/', '') - machine = machine.replace(' ', '_') - machine = machine.replace('/', '-') - - if osname[:5] == "linux": - # At least on Linux/Intel, 'machine' is the processor -- - # i386, etc. - # XXX what about Alpha, SPARC, etc? - return "%s-%s" % (osname, machine) - elif osname[:5] == "sunos": - if release[0] >= "5": # SunOS 5 == Solaris 2 - osname = "solaris" - release = "%d.%s" % (int(release[0]) - 3, release[2:]) - # fall through to standard osname-release-machine representation - elif osname[:4] == "irix": # could be "irix64"! - return "%s-%s" % (osname, release) - elif osname[:3] == "aix": - return "%s-%s.%s" % (osname, version, release) - elif osname[:6] == "cygwin": - osname = "cygwin" - rel_re = re.compile(r'[\d.]+') - m = rel_re.match(release) - if m: - release = m.group() - elif osname[:6] == "darwin": - # - # For our purposes, we'll assume that the system version from - # distutils' perspective is what MACOSX_DEPLOYMENT_TARGET is set - # to. This makes the compatibility story a bit more sane because the - # machine is going to compile and link as if it were - # MACOSX_DEPLOYMENT_TARGET. - cfgvars = get_config_vars() - macver = cfgvars.get('MACOSX_DEPLOYMENT_TARGET') - - if True: - # Always calculate the release of the running machine, - # needed to determine if we can build fat binaries or not. - - macrelease = macver - # Get the system version. Reading this plist is a documented - # way to get the system version (see the documentation for - # the Gestalt Manager) - try: - f = open('/System/Library/CoreServices/SystemVersion.plist') - except IOError: - # We're on a plain darwin box, fall back to the default - # behaviour. - pass - else: - try: - m = re.search(r'ProductUserVisibleVersion\s*' - r'(.*?)', f.read()) - finally: - f.close() - if m is not None: - macrelease = '.'.join(m.group(1).split('.')[:2]) - # else: fall back to the default behaviour - - if not macver: - macver = macrelease - - if macver: - release = macver - osname = "macosx" - - if ((macrelease + '.') >= '10.4.' and - '-arch' in get_config_vars().get('CFLAGS', '').strip()): - # The universal build will build fat binaries, but not on - # systems before 10.4 - # - # Try to detect 4-way universal builds, those have machine-type - # 'universal' instead of 'fat'. - - machine = 'fat' - cflags = get_config_vars().get('CFLAGS') - - archs = re.findall(r'-arch\s+(\S+)', cflags) - archs = tuple(sorted(set(archs))) - - if len(archs) == 1: - machine = archs[0] - elif archs == ('i386', 'ppc'): - machine = 'fat' - elif archs == ('i386', 'x86_64'): - machine = 'intel' - elif archs == ('i386', 'ppc', 'x86_64'): - machine = 'fat3' - elif archs == ('ppc64', 'x86_64'): - machine = 'fat64' - elif archs == ('i386', 'ppc', 'ppc64', 'x86_64'): - machine = 'universal' - else: - raise ValueError( - "Don't know machine value for archs=%r" % (archs,)) - - elif machine == 'i386': - # On OSX the machine type returned by uname is always the - # 32-bit variant, even if the executable architecture is - # the 64-bit variant - if sys.maxsize >= 2**32: - machine = 'x86_64' - - elif machine in ('PowerPC', 'Power_Macintosh'): - # Pick a sane name for the PPC architecture. - # See 'i386' case - if sys.maxsize >= 2**32: - machine = 'ppc64' - else: - machine = 'ppc' - - return "%s-%s-%s" % (osname, release, machine) - - -def get_python_version(): - return _PY_VERSION_SHORT - - -def _print_dict(title, data): - for index, (key, value) in enumerate(sorted(data.items())): - if index == 0: - print('%s: ' % (title)) - print('\t%s = "%s"' % (key, value)) - - -def _main(): - """Display all information sysconfig detains.""" - print('Platform: "%s"' % get_platform()) - print('Python version: "%s"' % get_python_version()) - print('Current installation scheme: "%s"' % _get_default_scheme()) - print() - _print_dict('Paths', get_paths()) - print() - _print_dict('Variables', get_config_vars()) - - -if __name__ == '__main__': - _main() diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/_backport/tarfile.py b/src/fetchcode/vcs/pip/_vendor/distlib/_backport/tarfile.py deleted file mode 100644 index d66d8566..00000000 --- a/src/fetchcode/vcs/pip/_vendor/distlib/_backport/tarfile.py +++ /dev/null @@ -1,2607 +0,0 @@ -#------------------------------------------------------------------- -# tarfile.py -#------------------------------------------------------------------- -# Copyright (C) 2002 Lars Gustaebel -# All rights reserved. -# -# Permission is hereby granted, free of charge, to any person -# obtaining a copy of this software and associated documentation -# files (the "Software"), to deal in the Software without -# restriction, including without limitation the rights to use, -# copy, modify, merge, publish, distribute, sublicense, and/or sell -# copies of the Software, and to permit persons to whom the -# Software is furnished to do so, subject to the following -# conditions: -# -# The above copyright notice and this permission notice shall be -# included in all copies or substantial portions of the Software. -# -# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, -# EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES -# OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND -# NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT -# HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, -# WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING -# FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR -# OTHER DEALINGS IN THE SOFTWARE. -# -from __future__ import print_function - -"""Read from and write to tar format archives. -""" - -__version__ = "$Revision$" - -version = "0.9.0" -__author__ = "Lars Gust\u00e4bel (lars@gustaebel.de)" -__date__ = "$Date: 2011-02-25 17:42:01 +0200 (Fri, 25 Feb 2011) $" -__cvsid__ = "$Id: tarfile.py 88586 2011-02-25 15:42:01Z marc-andre.lemburg $" -__credits__ = "Gustavo Niemeyer, Niels Gust\u00e4bel, Richard Townsend." - -#--------- -# Imports -#--------- -import sys -import os -import stat -import errno -import time -import struct -import copy -import re - -try: - import grp, pwd -except ImportError: - grp = pwd = None - -# os.symlink on Windows prior to 6.0 raises NotImplementedError -symlink_exception = (AttributeError, NotImplementedError) -try: - # WindowsError (1314) will be raised if the caller does not hold the - # SeCreateSymbolicLinkPrivilege privilege - symlink_exception += (WindowsError,) -except NameError: - pass - -# from tarfile import * -__all__ = ["TarFile", "TarInfo", "is_tarfile", "TarError"] - -if sys.version_info[0] < 3: - import __builtin__ as builtins -else: - import builtins - -_open = builtins.open # Since 'open' is TarFile.open - -#--------------------------------------------------------- -# tar constants -#--------------------------------------------------------- -NUL = b"\0" # the null character -BLOCKSIZE = 512 # length of processing blocks -RECORDSIZE = BLOCKSIZE * 20 # length of records -GNU_MAGIC = b"ustar \0" # magic gnu tar string -POSIX_MAGIC = b"ustar\x0000" # magic posix tar string - -LENGTH_NAME = 100 # maximum length of a filename -LENGTH_LINK = 100 # maximum length of a linkname -LENGTH_PREFIX = 155 # maximum length of the prefix field - -REGTYPE = b"0" # regular file -AREGTYPE = b"\0" # regular file -LNKTYPE = b"1" # link (inside tarfile) -SYMTYPE = b"2" # symbolic link -CHRTYPE = b"3" # character special device -BLKTYPE = b"4" # block special device -DIRTYPE = b"5" # directory -FIFOTYPE = b"6" # fifo special device -CONTTYPE = b"7" # contiguous file - -GNUTYPE_LONGNAME = b"L" # GNU tar longname -GNUTYPE_LONGLINK = b"K" # GNU tar longlink -GNUTYPE_SPARSE = b"S" # GNU tar sparse file - -XHDTYPE = b"x" # POSIX.1-2001 extended header -XGLTYPE = b"g" # POSIX.1-2001 global header -SOLARIS_XHDTYPE = b"X" # Solaris extended header - -USTAR_FORMAT = 0 # POSIX.1-1988 (ustar) format -GNU_FORMAT = 1 # GNU tar format -PAX_FORMAT = 2 # POSIX.1-2001 (pax) format -DEFAULT_FORMAT = GNU_FORMAT - -#--------------------------------------------------------- -# tarfile constants -#--------------------------------------------------------- -# File types that tarfile supports: -SUPPORTED_TYPES = (REGTYPE, AREGTYPE, LNKTYPE, - SYMTYPE, DIRTYPE, FIFOTYPE, - CONTTYPE, CHRTYPE, BLKTYPE, - GNUTYPE_LONGNAME, GNUTYPE_LONGLINK, - GNUTYPE_SPARSE) - -# File types that will be treated as a regular file. -REGULAR_TYPES = (REGTYPE, AREGTYPE, - CONTTYPE, GNUTYPE_SPARSE) - -# File types that are part of the GNU tar format. -GNU_TYPES = (GNUTYPE_LONGNAME, GNUTYPE_LONGLINK, - GNUTYPE_SPARSE) - -# Fields from a pax header that override a TarInfo attribute. -PAX_FIELDS = ("path", "linkpath", "size", "mtime", - "uid", "gid", "uname", "gname") - -# Fields from a pax header that are affected by hdrcharset. -PAX_NAME_FIELDS = set(("path", "linkpath", "uname", "gname")) - -# Fields in a pax header that are numbers, all other fields -# are treated as strings. -PAX_NUMBER_FIELDS = { - "atime": float, - "ctime": float, - "mtime": float, - "uid": int, - "gid": int, - "size": int -} - -#--------------------------------------------------------- -# Bits used in the mode field, values in octal. -#--------------------------------------------------------- -S_IFLNK = 0o120000 # symbolic link -S_IFREG = 0o100000 # regular file -S_IFBLK = 0o060000 # block device -S_IFDIR = 0o040000 # directory -S_IFCHR = 0o020000 # character device -S_IFIFO = 0o010000 # fifo - -TSUID = 0o4000 # set UID on execution -TSGID = 0o2000 # set GID on execution -TSVTX = 0o1000 # reserved - -TUREAD = 0o400 # read by owner -TUWRITE = 0o200 # write by owner -TUEXEC = 0o100 # execute/search by owner -TGREAD = 0o040 # read by group -TGWRITE = 0o020 # write by group -TGEXEC = 0o010 # execute/search by group -TOREAD = 0o004 # read by other -TOWRITE = 0o002 # write by other -TOEXEC = 0o001 # execute/search by other - -#--------------------------------------------------------- -# initialization -#--------------------------------------------------------- -if os.name in ("nt", "ce"): - ENCODING = "utf-8" -else: - ENCODING = sys.getfilesystemencoding() - -#--------------------------------------------------------- -# Some useful functions -#--------------------------------------------------------- - -def stn(s, length, encoding, errors): - """Convert a string to a null-terminated bytes object. - """ - s = s.encode(encoding, errors) - return s[:length] + (length - len(s)) * NUL - -def nts(s, encoding, errors): - """Convert a null-terminated bytes object to a string. - """ - p = s.find(b"\0") - if p != -1: - s = s[:p] - return s.decode(encoding, errors) - -def nti(s): - """Convert a number field to a python number. - """ - # There are two possible encodings for a number field, see - # itn() below. - if s[0] != chr(0o200): - try: - n = int(nts(s, "ascii", "strict") or "0", 8) - except ValueError: - raise InvalidHeaderError("invalid header") - else: - n = 0 - for i in range(len(s) - 1): - n <<= 8 - n += ord(s[i + 1]) - return n - -def itn(n, digits=8, format=DEFAULT_FORMAT): - """Convert a python number to a number field. - """ - # POSIX 1003.1-1988 requires numbers to be encoded as a string of - # octal digits followed by a null-byte, this allows values up to - # (8**(digits-1))-1. GNU tar allows storing numbers greater than - # that if necessary. A leading 0o200 byte indicates this particular - # encoding, the following digits-1 bytes are a big-endian - # representation. This allows values up to (256**(digits-1))-1. - if 0 <= n < 8 ** (digits - 1): - s = ("%0*o" % (digits - 1, n)).encode("ascii") + NUL - else: - if format != GNU_FORMAT or n >= 256 ** (digits - 1): - raise ValueError("overflow in number field") - - if n < 0: - # XXX We mimic GNU tar's behaviour with negative numbers, - # this could raise OverflowError. - n = struct.unpack("L", struct.pack("l", n))[0] - - s = bytearray() - for i in range(digits - 1): - s.insert(0, n & 0o377) - n >>= 8 - s.insert(0, 0o200) - return s - -def calc_chksums(buf): - """Calculate the checksum for a member's header by summing up all - characters except for the chksum field which is treated as if - it was filled with spaces. According to the GNU tar sources, - some tars (Sun and NeXT) calculate chksum with signed char, - which will be different if there are chars in the buffer with - the high bit set. So we calculate two checksums, unsigned and - signed. - """ - unsigned_chksum = 256 + sum(struct.unpack("148B", buf[:148]) + struct.unpack("356B", buf[156:512])) - signed_chksum = 256 + sum(struct.unpack("148b", buf[:148]) + struct.unpack("356b", buf[156:512])) - return unsigned_chksum, signed_chksum - -def copyfileobj(src, dst, length=None): - """Copy length bytes from fileobj src to fileobj dst. - If length is None, copy the entire content. - """ - if length == 0: - return - if length is None: - while True: - buf = src.read(16*1024) - if not buf: - break - dst.write(buf) - return - - BUFSIZE = 16 * 1024 - blocks, remainder = divmod(length, BUFSIZE) - for b in range(blocks): - buf = src.read(BUFSIZE) - if len(buf) < BUFSIZE: - raise IOError("end of file reached") - dst.write(buf) - - if remainder != 0: - buf = src.read(remainder) - if len(buf) < remainder: - raise IOError("end of file reached") - dst.write(buf) - return - -filemode_table = ( - ((S_IFLNK, "l"), - (S_IFREG, "-"), - (S_IFBLK, "b"), - (S_IFDIR, "d"), - (S_IFCHR, "c"), - (S_IFIFO, "p")), - - ((TUREAD, "r"),), - ((TUWRITE, "w"),), - ((TUEXEC|TSUID, "s"), - (TSUID, "S"), - (TUEXEC, "x")), - - ((TGREAD, "r"),), - ((TGWRITE, "w"),), - ((TGEXEC|TSGID, "s"), - (TSGID, "S"), - (TGEXEC, "x")), - - ((TOREAD, "r"),), - ((TOWRITE, "w"),), - ((TOEXEC|TSVTX, "t"), - (TSVTX, "T"), - (TOEXEC, "x")) -) - -def filemode(mode): - """Convert a file's mode to a string of the form - -rwxrwxrwx. - Used by TarFile.list() - """ - perm = [] - for table in filemode_table: - for bit, char in table: - if mode & bit == bit: - perm.append(char) - break - else: - perm.append("-") - return "".join(perm) - -class TarError(Exception): - """Base exception.""" - pass -class ExtractError(TarError): - """General exception for extract errors.""" - pass -class ReadError(TarError): - """Exception for unreadable tar archives.""" - pass -class CompressionError(TarError): - """Exception for unavailable compression methods.""" - pass -class StreamError(TarError): - """Exception for unsupported operations on stream-like TarFiles.""" - pass -class HeaderError(TarError): - """Base exception for header errors.""" - pass -class EmptyHeaderError(HeaderError): - """Exception for empty headers.""" - pass -class TruncatedHeaderError(HeaderError): - """Exception for truncated headers.""" - pass -class EOFHeaderError(HeaderError): - """Exception for end of file headers.""" - pass -class InvalidHeaderError(HeaderError): - """Exception for invalid headers.""" - pass -class SubsequentHeaderError(HeaderError): - """Exception for missing and invalid extended headers.""" - pass - -#--------------------------- -# internal stream interface -#--------------------------- -class _LowLevelFile(object): - """Low-level file object. Supports reading and writing. - It is used instead of a regular file object for streaming - access. - """ - - def __init__(self, name, mode): - mode = { - "r": os.O_RDONLY, - "w": os.O_WRONLY | os.O_CREAT | os.O_TRUNC, - }[mode] - if hasattr(os, "O_BINARY"): - mode |= os.O_BINARY - self.fd = os.open(name, mode, 0o666) - - def close(self): - os.close(self.fd) - - def read(self, size): - return os.read(self.fd, size) - - def write(self, s): - os.write(self.fd, s) - -class _Stream(object): - """Class that serves as an adapter between TarFile and - a stream-like object. The stream-like object only - needs to have a read() or write() method and is accessed - blockwise. Use of gzip or bzip2 compression is possible. - A stream-like object could be for example: sys.stdin, - sys.stdout, a socket, a tape device etc. - - _Stream is intended to be used only internally. - """ - - def __init__(self, name, mode, comptype, fileobj, bufsize): - """Construct a _Stream object. - """ - self._extfileobj = True - if fileobj is None: - fileobj = _LowLevelFile(name, mode) - self._extfileobj = False - - if comptype == '*': - # Enable transparent compression detection for the - # stream interface - fileobj = _StreamProxy(fileobj) - comptype = fileobj.getcomptype() - - self.name = name or "" - self.mode = mode - self.comptype = comptype - self.fileobj = fileobj - self.bufsize = bufsize - self.buf = b"" - self.pos = 0 - self.closed = False - - try: - if comptype == "gz": - try: - import zlib - except ImportError: - raise CompressionError("zlib module is not available") - self.zlib = zlib - self.crc = zlib.crc32(b"") - if mode == "r": - self._init_read_gz() - else: - self._init_write_gz() - - if comptype == "bz2": - try: - import bz2 - except ImportError: - raise CompressionError("bz2 module is not available") - if mode == "r": - self.dbuf = b"" - self.cmp = bz2.BZ2Decompressor() - else: - self.cmp = bz2.BZ2Compressor() - except: - if not self._extfileobj: - self.fileobj.close() - self.closed = True - raise - - def __del__(self): - if hasattr(self, "closed") and not self.closed: - self.close() - - def _init_write_gz(self): - """Initialize for writing with gzip compression. - """ - self.cmp = self.zlib.compressobj(9, self.zlib.DEFLATED, - -self.zlib.MAX_WBITS, - self.zlib.DEF_MEM_LEVEL, - 0) - timestamp = struct.pack(" self.bufsize: - self.fileobj.write(self.buf[:self.bufsize]) - self.buf = self.buf[self.bufsize:] - - def close(self): - """Close the _Stream object. No operation should be - done on it afterwards. - """ - if self.closed: - return - - if self.mode == "w" and self.comptype != "tar": - self.buf += self.cmp.flush() - - if self.mode == "w" and self.buf: - self.fileobj.write(self.buf) - self.buf = b"" - if self.comptype == "gz": - # The native zlib crc is an unsigned 32-bit integer, but - # the Python wrapper implicitly casts that to a signed C - # long. So, on a 32-bit box self.crc may "look negative", - # while the same crc on a 64-bit box may "look positive". - # To avoid irksome warnings from the `struct` module, force - # it to look positive on all boxes. - self.fileobj.write(struct.pack("= 0: - blocks, remainder = divmod(pos - self.pos, self.bufsize) - for i in range(blocks): - self.read(self.bufsize) - self.read(remainder) - else: - raise StreamError("seeking backwards is not allowed") - return self.pos - - def read(self, size=None): - """Return the next size number of bytes from the stream. - If size is not defined, return all bytes of the stream - up to EOF. - """ - if size is None: - t = [] - while True: - buf = self._read(self.bufsize) - if not buf: - break - t.append(buf) - buf = "".join(t) - else: - buf = self._read(size) - self.pos += len(buf) - return buf - - def _read(self, size): - """Return size bytes from the stream. - """ - if self.comptype == "tar": - return self.__read(size) - - c = len(self.dbuf) - while c < size: - buf = self.__read(self.bufsize) - if not buf: - break - try: - buf = self.cmp.decompress(buf) - except IOError: - raise ReadError("invalid compressed data") - self.dbuf += buf - c += len(buf) - buf = self.dbuf[:size] - self.dbuf = self.dbuf[size:] - return buf - - def __read(self, size): - """Return size bytes from stream. If internal buffer is empty, - read another block from the stream. - """ - c = len(self.buf) - while c < size: - buf = self.fileobj.read(self.bufsize) - if not buf: - break - self.buf += buf - c += len(buf) - buf = self.buf[:size] - self.buf = self.buf[size:] - return buf -# class _Stream - -class _StreamProxy(object): - """Small proxy class that enables transparent compression - detection for the Stream interface (mode 'r|*'). - """ - - def __init__(self, fileobj): - self.fileobj = fileobj - self.buf = self.fileobj.read(BLOCKSIZE) - - def read(self, size): - self.read = self.fileobj.read - return self.buf - - def getcomptype(self): - if self.buf.startswith(b"\037\213\010"): - return "gz" - if self.buf.startswith(b"BZh91"): - return "bz2" - return "tar" - - def close(self): - self.fileobj.close() -# class StreamProxy - -class _BZ2Proxy(object): - """Small proxy class that enables external file object - support for "r:bz2" and "w:bz2" modes. This is actually - a workaround for a limitation in bz2 module's BZ2File - class which (unlike gzip.GzipFile) has no support for - a file object argument. - """ - - blocksize = 16 * 1024 - - def __init__(self, fileobj, mode): - self.fileobj = fileobj - self.mode = mode - self.name = getattr(self.fileobj, "name", None) - self.init() - - def init(self): - import bz2 - self.pos = 0 - if self.mode == "r": - self.bz2obj = bz2.BZ2Decompressor() - self.fileobj.seek(0) - self.buf = b"" - else: - self.bz2obj = bz2.BZ2Compressor() - - def read(self, size): - x = len(self.buf) - while x < size: - raw = self.fileobj.read(self.blocksize) - if not raw: - break - data = self.bz2obj.decompress(raw) - self.buf += data - x += len(data) - - buf = self.buf[:size] - self.buf = self.buf[size:] - self.pos += len(buf) - return buf - - def seek(self, pos): - if pos < self.pos: - self.init() - self.read(pos - self.pos) - - def tell(self): - return self.pos - - def write(self, data): - self.pos += len(data) - raw = self.bz2obj.compress(data) - self.fileobj.write(raw) - - def close(self): - if self.mode == "w": - raw = self.bz2obj.flush() - self.fileobj.write(raw) -# class _BZ2Proxy - -#------------------------ -# Extraction file object -#------------------------ -class _FileInFile(object): - """A thin wrapper around an existing file object that - provides a part of its data as an individual file - object. - """ - - def __init__(self, fileobj, offset, size, blockinfo=None): - self.fileobj = fileobj - self.offset = offset - self.size = size - self.position = 0 - - if blockinfo is None: - blockinfo = [(0, size)] - - # Construct a map with data and zero blocks. - self.map_index = 0 - self.map = [] - lastpos = 0 - realpos = self.offset - for offset, size in blockinfo: - if offset > lastpos: - self.map.append((False, lastpos, offset, None)) - self.map.append((True, offset, offset + size, realpos)) - realpos += size - lastpos = offset + size - if lastpos < self.size: - self.map.append((False, lastpos, self.size, None)) - - def seekable(self): - if not hasattr(self.fileobj, "seekable"): - # XXX gzip.GzipFile and bz2.BZ2File - return True - return self.fileobj.seekable() - - def tell(self): - """Return the current file position. - """ - return self.position - - def seek(self, position): - """Seek to a position in the file. - """ - self.position = position - - def read(self, size=None): - """Read data from the file. - """ - if size is None: - size = self.size - self.position - else: - size = min(size, self.size - self.position) - - buf = b"" - while size > 0: - while True: - data, start, stop, offset = self.map[self.map_index] - if start <= self.position < stop: - break - else: - self.map_index += 1 - if self.map_index == len(self.map): - self.map_index = 0 - length = min(size, stop - self.position) - if data: - self.fileobj.seek(offset + (self.position - start)) - buf += self.fileobj.read(length) - else: - buf += NUL * length - size -= length - self.position += length - return buf -#class _FileInFile - - -class ExFileObject(object): - """File-like object for reading an archive member. - Is returned by TarFile.extractfile(). - """ - blocksize = 1024 - - def __init__(self, tarfile, tarinfo): - self.fileobj = _FileInFile(tarfile.fileobj, - tarinfo.offset_data, - tarinfo.size, - tarinfo.sparse) - self.name = tarinfo.name - self.mode = "r" - self.closed = False - self.size = tarinfo.size - - self.position = 0 - self.buffer = b"" - - def readable(self): - return True - - def writable(self): - return False - - def seekable(self): - return self.fileobj.seekable() - - def read(self, size=None): - """Read at most size bytes from the file. If size is not - present or None, read all data until EOF is reached. - """ - if self.closed: - raise ValueError("I/O operation on closed file") - - buf = b"" - if self.buffer: - if size is None: - buf = self.buffer - self.buffer = b"" - else: - buf = self.buffer[:size] - self.buffer = self.buffer[size:] - - if size is None: - buf += self.fileobj.read() - else: - buf += self.fileobj.read(size - len(buf)) - - self.position += len(buf) - return buf - - # XXX TextIOWrapper uses the read1() method. - read1 = read - - def readline(self, size=-1): - """Read one entire line from the file. If size is present - and non-negative, return a string with at most that - size, which may be an incomplete line. - """ - if self.closed: - raise ValueError("I/O operation on closed file") - - pos = self.buffer.find(b"\n") + 1 - if pos == 0: - # no newline found. - while True: - buf = self.fileobj.read(self.blocksize) - self.buffer += buf - if not buf or b"\n" in buf: - pos = self.buffer.find(b"\n") + 1 - if pos == 0: - # no newline found. - pos = len(self.buffer) - break - - if size != -1: - pos = min(size, pos) - - buf = self.buffer[:pos] - self.buffer = self.buffer[pos:] - self.position += len(buf) - return buf - - def readlines(self): - """Return a list with all remaining lines. - """ - result = [] - while True: - line = self.readline() - if not line: break - result.append(line) - return result - - def tell(self): - """Return the current file position. - """ - if self.closed: - raise ValueError("I/O operation on closed file") - - return self.position - - def seek(self, pos, whence=os.SEEK_SET): - """Seek to a position in the file. - """ - if self.closed: - raise ValueError("I/O operation on closed file") - - if whence == os.SEEK_SET: - self.position = min(max(pos, 0), self.size) - elif whence == os.SEEK_CUR: - if pos < 0: - self.position = max(self.position + pos, 0) - else: - self.position = min(self.position + pos, self.size) - elif whence == os.SEEK_END: - self.position = max(min(self.size + pos, self.size), 0) - else: - raise ValueError("Invalid argument") - - self.buffer = b"" - self.fileobj.seek(self.position) - - def close(self): - """Close the file object. - """ - self.closed = True - - def __iter__(self): - """Get an iterator over the file's lines. - """ - while True: - line = self.readline() - if not line: - break - yield line -#class ExFileObject - -#------------------ -# Exported Classes -#------------------ -class TarInfo(object): - """Informational class which holds the details about an - archive member given by a tar header block. - TarInfo objects are returned by TarFile.getmember(), - TarFile.getmembers() and TarFile.gettarinfo() and are - usually created internally. - """ - - __slots__ = ("name", "mode", "uid", "gid", "size", "mtime", - "chksum", "type", "linkname", "uname", "gname", - "devmajor", "devminor", - "offset", "offset_data", "pax_headers", "sparse", - "tarfile", "_sparse_structs", "_link_target") - - def __init__(self, name=""): - """Construct a TarInfo object. name is the optional name - of the member. - """ - self.name = name # member name - self.mode = 0o644 # file permissions - self.uid = 0 # user id - self.gid = 0 # group id - self.size = 0 # file size - self.mtime = 0 # modification time - self.chksum = 0 # header checksum - self.type = REGTYPE # member type - self.linkname = "" # link name - self.uname = "" # user name - self.gname = "" # group name - self.devmajor = 0 # device major number - self.devminor = 0 # device minor number - - self.offset = 0 # the tar header starts here - self.offset_data = 0 # the file's data starts here - - self.sparse = None # sparse member information - self.pax_headers = {} # pax header information - - # In pax headers the "name" and "linkname" field are called - # "path" and "linkpath". - def _getpath(self): - return self.name - def _setpath(self, name): - self.name = name - path = property(_getpath, _setpath) - - def _getlinkpath(self): - return self.linkname - def _setlinkpath(self, linkname): - self.linkname = linkname - linkpath = property(_getlinkpath, _setlinkpath) - - def __repr__(self): - return "<%s %r at %#x>" % (self.__class__.__name__,self.name,id(self)) - - def get_info(self): - """Return the TarInfo's attributes as a dictionary. - """ - info = { - "name": self.name, - "mode": self.mode & 0o7777, - "uid": self.uid, - "gid": self.gid, - "size": self.size, - "mtime": self.mtime, - "chksum": self.chksum, - "type": self.type, - "linkname": self.linkname, - "uname": self.uname, - "gname": self.gname, - "devmajor": self.devmajor, - "devminor": self.devminor - } - - if info["type"] == DIRTYPE and not info["name"].endswith("/"): - info["name"] += "/" - - return info - - def tobuf(self, format=DEFAULT_FORMAT, encoding=ENCODING, errors="surrogateescape"): - """Return a tar header as a string of 512 byte blocks. - """ - info = self.get_info() - - if format == USTAR_FORMAT: - return self.create_ustar_header(info, encoding, errors) - elif format == GNU_FORMAT: - return self.create_gnu_header(info, encoding, errors) - elif format == PAX_FORMAT: - return self.create_pax_header(info, encoding) - else: - raise ValueError("invalid format") - - def create_ustar_header(self, info, encoding, errors): - """Return the object as a ustar header block. - """ - info["magic"] = POSIX_MAGIC - - if len(info["linkname"]) > LENGTH_LINK: - raise ValueError("linkname is too long") - - if len(info["name"]) > LENGTH_NAME: - info["prefix"], info["name"] = self._posix_split_name(info["name"]) - - return self._create_header(info, USTAR_FORMAT, encoding, errors) - - def create_gnu_header(self, info, encoding, errors): - """Return the object as a GNU header block sequence. - """ - info["magic"] = GNU_MAGIC - - buf = b"" - if len(info["linkname"]) > LENGTH_LINK: - buf += self._create_gnu_long_header(info["linkname"], GNUTYPE_LONGLINK, encoding, errors) - - if len(info["name"]) > LENGTH_NAME: - buf += self._create_gnu_long_header(info["name"], GNUTYPE_LONGNAME, encoding, errors) - - return buf + self._create_header(info, GNU_FORMAT, encoding, errors) - - def create_pax_header(self, info, encoding): - """Return the object as a ustar header block. If it cannot be - represented this way, prepend a pax extended header sequence - with supplement information. - """ - info["magic"] = POSIX_MAGIC - pax_headers = self.pax_headers.copy() - - # Test string fields for values that exceed the field length or cannot - # be represented in ASCII encoding. - for name, hname, length in ( - ("name", "path", LENGTH_NAME), ("linkname", "linkpath", LENGTH_LINK), - ("uname", "uname", 32), ("gname", "gname", 32)): - - if hname in pax_headers: - # The pax header has priority. - continue - - # Try to encode the string as ASCII. - try: - info[name].encode("ascii", "strict") - except UnicodeEncodeError: - pax_headers[hname] = info[name] - continue - - if len(info[name]) > length: - pax_headers[hname] = info[name] - - # Test number fields for values that exceed the field limit or values - # that like to be stored as float. - for name, digits in (("uid", 8), ("gid", 8), ("size", 12), ("mtime", 12)): - if name in pax_headers: - # The pax header has priority. Avoid overflow. - info[name] = 0 - continue - - val = info[name] - if not 0 <= val < 8 ** (digits - 1) or isinstance(val, float): - pax_headers[name] = str(val) - info[name] = 0 - - # Create a pax extended header if necessary. - if pax_headers: - buf = self._create_pax_generic_header(pax_headers, XHDTYPE, encoding) - else: - buf = b"" - - return buf + self._create_header(info, USTAR_FORMAT, "ascii", "replace") - - @classmethod - def create_pax_global_header(cls, pax_headers): - """Return the object as a pax global header block sequence. - """ - return cls._create_pax_generic_header(pax_headers, XGLTYPE, "utf8") - - def _posix_split_name(self, name): - """Split a name longer than 100 chars into a prefix - and a name part. - """ - prefix = name[:LENGTH_PREFIX + 1] - while prefix and prefix[-1] != "/": - prefix = prefix[:-1] - - name = name[len(prefix):] - prefix = prefix[:-1] - - if not prefix or len(name) > LENGTH_NAME: - raise ValueError("name is too long") - return prefix, name - - @staticmethod - def _create_header(info, format, encoding, errors): - """Return a header block. info is a dictionary with file - information, format must be one of the *_FORMAT constants. - """ - parts = [ - stn(info.get("name", ""), 100, encoding, errors), - itn(info.get("mode", 0) & 0o7777, 8, format), - itn(info.get("uid", 0), 8, format), - itn(info.get("gid", 0), 8, format), - itn(info.get("size", 0), 12, format), - itn(info.get("mtime", 0), 12, format), - b" ", # checksum field - info.get("type", REGTYPE), - stn(info.get("linkname", ""), 100, encoding, errors), - info.get("magic", POSIX_MAGIC), - stn(info.get("uname", ""), 32, encoding, errors), - stn(info.get("gname", ""), 32, encoding, errors), - itn(info.get("devmajor", 0), 8, format), - itn(info.get("devminor", 0), 8, format), - stn(info.get("prefix", ""), 155, encoding, errors) - ] - - buf = struct.pack("%ds" % BLOCKSIZE, b"".join(parts)) - chksum = calc_chksums(buf[-BLOCKSIZE:])[0] - buf = buf[:-364] + ("%06o\0" % chksum).encode("ascii") + buf[-357:] - return buf - - @staticmethod - def _create_payload(payload): - """Return the string payload filled with zero bytes - up to the next 512 byte border. - """ - blocks, remainder = divmod(len(payload), BLOCKSIZE) - if remainder > 0: - payload += (BLOCKSIZE - remainder) * NUL - return payload - - @classmethod - def _create_gnu_long_header(cls, name, type, encoding, errors): - """Return a GNUTYPE_LONGNAME or GNUTYPE_LONGLINK sequence - for name. - """ - name = name.encode(encoding, errors) + NUL - - info = {} - info["name"] = "././@LongLink" - info["type"] = type - info["size"] = len(name) - info["magic"] = GNU_MAGIC - - # create extended header + name blocks. - return cls._create_header(info, USTAR_FORMAT, encoding, errors) + \ - cls._create_payload(name) - - @classmethod - def _create_pax_generic_header(cls, pax_headers, type, encoding): - """Return a POSIX.1-2008 extended or global header sequence - that contains a list of keyword, value pairs. The values - must be strings. - """ - # Check if one of the fields contains surrogate characters and thereby - # forces hdrcharset=BINARY, see _proc_pax() for more information. - binary = False - for keyword, value in pax_headers.items(): - try: - value.encode("utf8", "strict") - except UnicodeEncodeError: - binary = True - break - - records = b"" - if binary: - # Put the hdrcharset field at the beginning of the header. - records += b"21 hdrcharset=BINARY\n" - - for keyword, value in pax_headers.items(): - keyword = keyword.encode("utf8") - if binary: - # Try to restore the original byte representation of `value'. - # Needless to say, that the encoding must match the string. - value = value.encode(encoding, "surrogateescape") - else: - value = value.encode("utf8") - - l = len(keyword) + len(value) + 3 # ' ' + '=' + '\n' - n = p = 0 - while True: - n = l + len(str(p)) - if n == p: - break - p = n - records += bytes(str(p), "ascii") + b" " + keyword + b"=" + value + b"\n" - - # We use a hardcoded "././@PaxHeader" name like star does - # instead of the one that POSIX recommends. - info = {} - info["name"] = "././@PaxHeader" - info["type"] = type - info["size"] = len(records) - info["magic"] = POSIX_MAGIC - - # Create pax header + record blocks. - return cls._create_header(info, USTAR_FORMAT, "ascii", "replace") + \ - cls._create_payload(records) - - @classmethod - def frombuf(cls, buf, encoding, errors): - """Construct a TarInfo object from a 512 byte bytes object. - """ - if len(buf) == 0: - raise EmptyHeaderError("empty header") - if len(buf) != BLOCKSIZE: - raise TruncatedHeaderError("truncated header") - if buf.count(NUL) == BLOCKSIZE: - raise EOFHeaderError("end of file header") - - chksum = nti(buf[148:156]) - if chksum not in calc_chksums(buf): - raise InvalidHeaderError("bad checksum") - - obj = cls() - obj.name = nts(buf[0:100], encoding, errors) - obj.mode = nti(buf[100:108]) - obj.uid = nti(buf[108:116]) - obj.gid = nti(buf[116:124]) - obj.size = nti(buf[124:136]) - obj.mtime = nti(buf[136:148]) - obj.chksum = chksum - obj.type = buf[156:157] - obj.linkname = nts(buf[157:257], encoding, errors) - obj.uname = nts(buf[265:297], encoding, errors) - obj.gname = nts(buf[297:329], encoding, errors) - obj.devmajor = nti(buf[329:337]) - obj.devminor = nti(buf[337:345]) - prefix = nts(buf[345:500], encoding, errors) - - # Old V7 tar format represents a directory as a regular - # file with a trailing slash. - if obj.type == AREGTYPE and obj.name.endswith("/"): - obj.type = DIRTYPE - - # The old GNU sparse format occupies some of the unused - # space in the buffer for up to 4 sparse structures. - # Save the them for later processing in _proc_sparse(). - if obj.type == GNUTYPE_SPARSE: - pos = 386 - structs = [] - for i in range(4): - try: - offset = nti(buf[pos:pos + 12]) - numbytes = nti(buf[pos + 12:pos + 24]) - except ValueError: - break - structs.append((offset, numbytes)) - pos += 24 - isextended = bool(buf[482]) - origsize = nti(buf[483:495]) - obj._sparse_structs = (structs, isextended, origsize) - - # Remove redundant slashes from directories. - if obj.isdir(): - obj.name = obj.name.rstrip("/") - - # Reconstruct a ustar longname. - if prefix and obj.type not in GNU_TYPES: - obj.name = prefix + "/" + obj.name - return obj - - @classmethod - def fromtarfile(cls, tarfile): - """Return the next TarInfo object from TarFile object - tarfile. - """ - buf = tarfile.fileobj.read(BLOCKSIZE) - obj = cls.frombuf(buf, tarfile.encoding, tarfile.errors) - obj.offset = tarfile.fileobj.tell() - BLOCKSIZE - return obj._proc_member(tarfile) - - #-------------------------------------------------------------------------- - # The following are methods that are called depending on the type of a - # member. The entry point is _proc_member() which can be overridden in a - # subclass to add custom _proc_*() methods. A _proc_*() method MUST - # implement the following - # operations: - # 1. Set self.offset_data to the position where the data blocks begin, - # if there is data that follows. - # 2. Set tarfile.offset to the position where the next member's header will - # begin. - # 3. Return self or another valid TarInfo object. - def _proc_member(self, tarfile): - """Choose the right processing method depending on - the type and call it. - """ - if self.type in (GNUTYPE_LONGNAME, GNUTYPE_LONGLINK): - return self._proc_gnulong(tarfile) - elif self.type == GNUTYPE_SPARSE: - return self._proc_sparse(tarfile) - elif self.type in (XHDTYPE, XGLTYPE, SOLARIS_XHDTYPE): - return self._proc_pax(tarfile) - else: - return self._proc_builtin(tarfile) - - def _proc_builtin(self, tarfile): - """Process a builtin type or an unknown type which - will be treated as a regular file. - """ - self.offset_data = tarfile.fileobj.tell() - offset = self.offset_data - if self.isreg() or self.type not in SUPPORTED_TYPES: - # Skip the following data blocks. - offset += self._block(self.size) - tarfile.offset = offset - - # Patch the TarInfo object with saved global - # header information. - self._apply_pax_info(tarfile.pax_headers, tarfile.encoding, tarfile.errors) - - return self - - def _proc_gnulong(self, tarfile): - """Process the blocks that hold a GNU longname - or longlink member. - """ - buf = tarfile.fileobj.read(self._block(self.size)) - - # Fetch the next header and process it. - try: - next = self.fromtarfile(tarfile) - except HeaderError: - raise SubsequentHeaderError("missing or bad subsequent header") - - # Patch the TarInfo object from the next header with - # the longname information. - next.offset = self.offset - if self.type == GNUTYPE_LONGNAME: - next.name = nts(buf, tarfile.encoding, tarfile.errors) - elif self.type == GNUTYPE_LONGLINK: - next.linkname = nts(buf, tarfile.encoding, tarfile.errors) - - return next - - def _proc_sparse(self, tarfile): - """Process a GNU sparse header plus extra headers. - """ - # We already collected some sparse structures in frombuf(). - structs, isextended, origsize = self._sparse_structs - del self._sparse_structs - - # Collect sparse structures from extended header blocks. - while isextended: - buf = tarfile.fileobj.read(BLOCKSIZE) - pos = 0 - for i in range(21): - try: - offset = nti(buf[pos:pos + 12]) - numbytes = nti(buf[pos + 12:pos + 24]) - except ValueError: - break - if offset and numbytes: - structs.append((offset, numbytes)) - pos += 24 - isextended = bool(buf[504]) - self.sparse = structs - - self.offset_data = tarfile.fileobj.tell() - tarfile.offset = self.offset_data + self._block(self.size) - self.size = origsize - return self - - def _proc_pax(self, tarfile): - """Process an extended or global header as described in - POSIX.1-2008. - """ - # Read the header information. - buf = tarfile.fileobj.read(self._block(self.size)) - - # A pax header stores supplemental information for either - # the following file (extended) or all following files - # (global). - if self.type == XGLTYPE: - pax_headers = tarfile.pax_headers - else: - pax_headers = tarfile.pax_headers.copy() - - # Check if the pax header contains a hdrcharset field. This tells us - # the encoding of the path, linkpath, uname and gname fields. Normally, - # these fields are UTF-8 encoded but since POSIX.1-2008 tar - # implementations are allowed to store them as raw binary strings if - # the translation to UTF-8 fails. - match = re.search(br"\d+ hdrcharset=([^\n]+)\n", buf) - if match is not None: - pax_headers["hdrcharset"] = match.group(1).decode("utf8") - - # For the time being, we don't care about anything other than "BINARY". - # The only other value that is currently allowed by the standard is - # "ISO-IR 10646 2000 UTF-8" in other words UTF-8. - hdrcharset = pax_headers.get("hdrcharset") - if hdrcharset == "BINARY": - encoding = tarfile.encoding - else: - encoding = "utf8" - - # Parse pax header information. A record looks like that: - # "%d %s=%s\n" % (length, keyword, value). length is the size - # of the complete record including the length field itself and - # the newline. keyword and value are both UTF-8 encoded strings. - regex = re.compile(br"(\d+) ([^=]+)=") - pos = 0 - while True: - match = regex.match(buf, pos) - if not match: - break - - length, keyword = match.groups() - length = int(length) - value = buf[match.end(2) + 1:match.start(1) + length - 1] - - # Normally, we could just use "utf8" as the encoding and "strict" - # as the error handler, but we better not take the risk. For - # example, GNU tar <= 1.23 is known to store filenames it cannot - # translate to UTF-8 as raw strings (unfortunately without a - # hdrcharset=BINARY header). - # We first try the strict standard encoding, and if that fails we - # fall back on the user's encoding and error handler. - keyword = self._decode_pax_field(keyword, "utf8", "utf8", - tarfile.errors) - if keyword in PAX_NAME_FIELDS: - value = self._decode_pax_field(value, encoding, tarfile.encoding, - tarfile.errors) - else: - value = self._decode_pax_field(value, "utf8", "utf8", - tarfile.errors) - - pax_headers[keyword] = value - pos += length - - # Fetch the next header. - try: - next = self.fromtarfile(tarfile) - except HeaderError: - raise SubsequentHeaderError("missing or bad subsequent header") - - # Process GNU sparse information. - if "GNU.sparse.map" in pax_headers: - # GNU extended sparse format version 0.1. - self._proc_gnusparse_01(next, pax_headers) - - elif "GNU.sparse.size" in pax_headers: - # GNU extended sparse format version 0.0. - self._proc_gnusparse_00(next, pax_headers, buf) - - elif pax_headers.get("GNU.sparse.major") == "1" and pax_headers.get("GNU.sparse.minor") == "0": - # GNU extended sparse format version 1.0. - self._proc_gnusparse_10(next, pax_headers, tarfile) - - if self.type in (XHDTYPE, SOLARIS_XHDTYPE): - # Patch the TarInfo object with the extended header info. - next._apply_pax_info(pax_headers, tarfile.encoding, tarfile.errors) - next.offset = self.offset - - if "size" in pax_headers: - # If the extended header replaces the size field, - # we need to recalculate the offset where the next - # header starts. - offset = next.offset_data - if next.isreg() or next.type not in SUPPORTED_TYPES: - offset += next._block(next.size) - tarfile.offset = offset - - return next - - def _proc_gnusparse_00(self, next, pax_headers, buf): - """Process a GNU tar extended sparse header, version 0.0. - """ - offsets = [] - for match in re.finditer(br"\d+ GNU.sparse.offset=(\d+)\n", buf): - offsets.append(int(match.group(1))) - numbytes = [] - for match in re.finditer(br"\d+ GNU.sparse.numbytes=(\d+)\n", buf): - numbytes.append(int(match.group(1))) - next.sparse = list(zip(offsets, numbytes)) - - def _proc_gnusparse_01(self, next, pax_headers): - """Process a GNU tar extended sparse header, version 0.1. - """ - sparse = [int(x) for x in pax_headers["GNU.sparse.map"].split(",")] - next.sparse = list(zip(sparse[::2], sparse[1::2])) - - def _proc_gnusparse_10(self, next, pax_headers, tarfile): - """Process a GNU tar extended sparse header, version 1.0. - """ - fields = None - sparse = [] - buf = tarfile.fileobj.read(BLOCKSIZE) - fields, buf = buf.split(b"\n", 1) - fields = int(fields) - while len(sparse) < fields * 2: - if b"\n" not in buf: - buf += tarfile.fileobj.read(BLOCKSIZE) - number, buf = buf.split(b"\n", 1) - sparse.append(int(number)) - next.offset_data = tarfile.fileobj.tell() - next.sparse = list(zip(sparse[::2], sparse[1::2])) - - def _apply_pax_info(self, pax_headers, encoding, errors): - """Replace fields with supplemental information from a previous - pax extended or global header. - """ - for keyword, value in pax_headers.items(): - if keyword == "GNU.sparse.name": - setattr(self, "path", value) - elif keyword == "GNU.sparse.size": - setattr(self, "size", int(value)) - elif keyword == "GNU.sparse.realsize": - setattr(self, "size", int(value)) - elif keyword in PAX_FIELDS: - if keyword in PAX_NUMBER_FIELDS: - try: - value = PAX_NUMBER_FIELDS[keyword](value) - except ValueError: - value = 0 - if keyword == "path": - value = value.rstrip("/") - setattr(self, keyword, value) - - self.pax_headers = pax_headers.copy() - - def _decode_pax_field(self, value, encoding, fallback_encoding, fallback_errors): - """Decode a single field from a pax record. - """ - try: - return value.decode(encoding, "strict") - except UnicodeDecodeError: - return value.decode(fallback_encoding, fallback_errors) - - def _block(self, count): - """Round up a byte count by BLOCKSIZE and return it, - e.g. _block(834) => 1024. - """ - blocks, remainder = divmod(count, BLOCKSIZE) - if remainder: - blocks += 1 - return blocks * BLOCKSIZE - - def isreg(self): - return self.type in REGULAR_TYPES - def isfile(self): - return self.isreg() - def isdir(self): - return self.type == DIRTYPE - def issym(self): - return self.type == SYMTYPE - def islnk(self): - return self.type == LNKTYPE - def ischr(self): - return self.type == CHRTYPE - def isblk(self): - return self.type == BLKTYPE - def isfifo(self): - return self.type == FIFOTYPE - def issparse(self): - return self.sparse is not None - def isdev(self): - return self.type in (CHRTYPE, BLKTYPE, FIFOTYPE) -# class TarInfo - -class TarFile(object): - """The TarFile Class provides an interface to tar archives. - """ - - debug = 0 # May be set from 0 (no msgs) to 3 (all msgs) - - dereference = False # If true, add content of linked file to the - # tar file, else the link. - - ignore_zeros = False # If true, skips empty or invalid blocks and - # continues processing. - - errorlevel = 1 # If 0, fatal errors only appear in debug - # messages (if debug >= 0). If > 0, errors - # are passed to the caller as exceptions. - - format = DEFAULT_FORMAT # The format to use when creating an archive. - - encoding = ENCODING # Encoding for 8-bit character strings. - - errors = None # Error handler for unicode conversion. - - tarinfo = TarInfo # The default TarInfo class to use. - - fileobject = ExFileObject # The default ExFileObject class to use. - - def __init__(self, name=None, mode="r", fileobj=None, format=None, - tarinfo=None, dereference=None, ignore_zeros=None, encoding=None, - errors="surrogateescape", pax_headers=None, debug=None, errorlevel=None): - """Open an (uncompressed) tar archive `name'. `mode' is either 'r' to - read from an existing archive, 'a' to append data to an existing - file or 'w' to create a new file overwriting an existing one. `mode' - defaults to 'r'. - If `fileobj' is given, it is used for reading or writing data. If it - can be determined, `mode' is overridden by `fileobj's mode. - `fileobj' is not closed, when TarFile is closed. - """ - if len(mode) > 1 or mode not in "raw": - raise ValueError("mode must be 'r', 'a' or 'w'") - self.mode = mode - self._mode = {"r": "rb", "a": "r+b", "w": "wb"}[mode] - - if not fileobj: - if self.mode == "a" and not os.path.exists(name): - # Create nonexistent files in append mode. - self.mode = "w" - self._mode = "wb" - fileobj = bltn_open(name, self._mode) - self._extfileobj = False - else: - if name is None and hasattr(fileobj, "name"): - name = fileobj.name - if hasattr(fileobj, "mode"): - self._mode = fileobj.mode - self._extfileobj = True - self.name = os.path.abspath(name) if name else None - self.fileobj = fileobj - - # Init attributes. - if format is not None: - self.format = format - if tarinfo is not None: - self.tarinfo = tarinfo - if dereference is not None: - self.dereference = dereference - if ignore_zeros is not None: - self.ignore_zeros = ignore_zeros - if encoding is not None: - self.encoding = encoding - self.errors = errors - - if pax_headers is not None and self.format == PAX_FORMAT: - self.pax_headers = pax_headers - else: - self.pax_headers = {} - - if debug is not None: - self.debug = debug - if errorlevel is not None: - self.errorlevel = errorlevel - - # Init datastructures. - self.closed = False - self.members = [] # list of members as TarInfo objects - self._loaded = False # flag if all members have been read - self.offset = self.fileobj.tell() - # current position in the archive file - self.inodes = {} # dictionary caching the inodes of - # archive members already added - - try: - if self.mode == "r": - self.firstmember = None - self.firstmember = self.next() - - if self.mode == "a": - # Move to the end of the archive, - # before the first empty block. - while True: - self.fileobj.seek(self.offset) - try: - tarinfo = self.tarinfo.fromtarfile(self) - self.members.append(tarinfo) - except EOFHeaderError: - self.fileobj.seek(self.offset) - break - except HeaderError as e: - raise ReadError(str(e)) - - if self.mode in "aw": - self._loaded = True - - if self.pax_headers: - buf = self.tarinfo.create_pax_global_header(self.pax_headers.copy()) - self.fileobj.write(buf) - self.offset += len(buf) - except: - if not self._extfileobj: - self.fileobj.close() - self.closed = True - raise - - #-------------------------------------------------------------------------- - # Below are the classmethods which act as alternate constructors to the - # TarFile class. The open() method is the only one that is needed for - # public use; it is the "super"-constructor and is able to select an - # adequate "sub"-constructor for a particular compression using the mapping - # from OPEN_METH. - # - # This concept allows one to subclass TarFile without losing the comfort of - # the super-constructor. A sub-constructor is registered and made available - # by adding it to the mapping in OPEN_METH. - - @classmethod - def open(cls, name=None, mode="r", fileobj=None, bufsize=RECORDSIZE, **kwargs): - """Open a tar archive for reading, writing or appending. Return - an appropriate TarFile class. - - mode: - 'r' or 'r:*' open for reading with transparent compression - 'r:' open for reading exclusively uncompressed - 'r:gz' open for reading with gzip compression - 'r:bz2' open for reading with bzip2 compression - 'a' or 'a:' open for appending, creating the file if necessary - 'w' or 'w:' open for writing without compression - 'w:gz' open for writing with gzip compression - 'w:bz2' open for writing with bzip2 compression - - 'r|*' open a stream of tar blocks with transparent compression - 'r|' open an uncompressed stream of tar blocks for reading - 'r|gz' open a gzip compressed stream of tar blocks - 'r|bz2' open a bzip2 compressed stream of tar blocks - 'w|' open an uncompressed stream for writing - 'w|gz' open a gzip compressed stream for writing - 'w|bz2' open a bzip2 compressed stream for writing - """ - - if not name and not fileobj: - raise ValueError("nothing to open") - - if mode in ("r", "r:*"): - # Find out which *open() is appropriate for opening the file. - for comptype in cls.OPEN_METH: - func = getattr(cls, cls.OPEN_METH[comptype]) - if fileobj is not None: - saved_pos = fileobj.tell() - try: - return func(name, "r", fileobj, **kwargs) - except (ReadError, CompressionError) as e: - if fileobj is not None: - fileobj.seek(saved_pos) - continue - raise ReadError("file could not be opened successfully") - - elif ":" in mode: - filemode, comptype = mode.split(":", 1) - filemode = filemode or "r" - comptype = comptype or "tar" - - # Select the *open() function according to - # given compression. - if comptype in cls.OPEN_METH: - func = getattr(cls, cls.OPEN_METH[comptype]) - else: - raise CompressionError("unknown compression type %r" % comptype) - return func(name, filemode, fileobj, **kwargs) - - elif "|" in mode: - filemode, comptype = mode.split("|", 1) - filemode = filemode or "r" - comptype = comptype or "tar" - - if filemode not in "rw": - raise ValueError("mode must be 'r' or 'w'") - - stream = _Stream(name, filemode, comptype, fileobj, bufsize) - try: - t = cls(name, filemode, stream, **kwargs) - except: - stream.close() - raise - t._extfileobj = False - return t - - elif mode in "aw": - return cls.taropen(name, mode, fileobj, **kwargs) - - raise ValueError("undiscernible mode") - - @classmethod - def taropen(cls, name, mode="r", fileobj=None, **kwargs): - """Open uncompressed tar archive name for reading or writing. - """ - if len(mode) > 1 or mode not in "raw": - raise ValueError("mode must be 'r', 'a' or 'w'") - return cls(name, mode, fileobj, **kwargs) - - @classmethod - def gzopen(cls, name, mode="r", fileobj=None, compresslevel=9, **kwargs): - """Open gzip compressed tar archive name for reading or writing. - Appending is not allowed. - """ - if len(mode) > 1 or mode not in "rw": - raise ValueError("mode must be 'r' or 'w'") - - try: - import gzip - gzip.GzipFile - except (ImportError, AttributeError): - raise CompressionError("gzip module is not available") - - extfileobj = fileobj is not None - try: - fileobj = gzip.GzipFile(name, mode + "b", compresslevel, fileobj) - t = cls.taropen(name, mode, fileobj, **kwargs) - except IOError: - if not extfileobj and fileobj is not None: - fileobj.close() - if fileobj is None: - raise - raise ReadError("not a gzip file") - except: - if not extfileobj and fileobj is not None: - fileobj.close() - raise - t._extfileobj = extfileobj - return t - - @classmethod - def bz2open(cls, name, mode="r", fileobj=None, compresslevel=9, **kwargs): - """Open bzip2 compressed tar archive name for reading or writing. - Appending is not allowed. - """ - if len(mode) > 1 or mode not in "rw": - raise ValueError("mode must be 'r' or 'w'.") - - try: - import bz2 - except ImportError: - raise CompressionError("bz2 module is not available") - - if fileobj is not None: - fileobj = _BZ2Proxy(fileobj, mode) - else: - fileobj = bz2.BZ2File(name, mode, compresslevel=compresslevel) - - try: - t = cls.taropen(name, mode, fileobj, **kwargs) - except (IOError, EOFError): - fileobj.close() - raise ReadError("not a bzip2 file") - t._extfileobj = False - return t - - # All *open() methods are registered here. - OPEN_METH = { - "tar": "taropen", # uncompressed tar - "gz": "gzopen", # gzip compressed tar - "bz2": "bz2open" # bzip2 compressed tar - } - - #-------------------------------------------------------------------------- - # The public methods which TarFile provides: - - def close(self): - """Close the TarFile. In write-mode, two finishing zero blocks are - appended to the archive. - """ - if self.closed: - return - - if self.mode in "aw": - self.fileobj.write(NUL * (BLOCKSIZE * 2)) - self.offset += (BLOCKSIZE * 2) - # fill up the end with zero-blocks - # (like option -b20 for tar does) - blocks, remainder = divmod(self.offset, RECORDSIZE) - if remainder > 0: - self.fileobj.write(NUL * (RECORDSIZE - remainder)) - - if not self._extfileobj: - self.fileobj.close() - self.closed = True - - def getmember(self, name): - """Return a TarInfo object for member `name'. If `name' can not be - found in the archive, KeyError is raised. If a member occurs more - than once in the archive, its last occurrence is assumed to be the - most up-to-date version. - """ - tarinfo = self._getmember(name) - if tarinfo is None: - raise KeyError("filename %r not found" % name) - return tarinfo - - def getmembers(self): - """Return the members of the archive as a list of TarInfo objects. The - list has the same order as the members in the archive. - """ - self._check() - if not self._loaded: # if we want to obtain a list of - self._load() # all members, we first have to - # scan the whole archive. - return self.members - - def getnames(self): - """Return the members of the archive as a list of their names. It has - the same order as the list returned by getmembers(). - """ - return [tarinfo.name for tarinfo in self.getmembers()] - - def gettarinfo(self, name=None, arcname=None, fileobj=None): - """Create a TarInfo object for either the file `name' or the file - object `fileobj' (using os.fstat on its file descriptor). You can - modify some of the TarInfo's attributes before you add it using - addfile(). If given, `arcname' specifies an alternative name for the - file in the archive. - """ - self._check("aw") - - # When fileobj is given, replace name by - # fileobj's real name. - if fileobj is not None: - name = fileobj.name - - # Building the name of the member in the archive. - # Backward slashes are converted to forward slashes, - # Absolute paths are turned to relative paths. - if arcname is None: - arcname = name - drv, arcname = os.path.splitdrive(arcname) - arcname = arcname.replace(os.sep, "/") - arcname = arcname.lstrip("/") - - # Now, fill the TarInfo object with - # information specific for the file. - tarinfo = self.tarinfo() - tarinfo.tarfile = self - - # Use os.stat or os.lstat, depending on platform - # and if symlinks shall be resolved. - if fileobj is None: - if hasattr(os, "lstat") and not self.dereference: - statres = os.lstat(name) - else: - statres = os.stat(name) - else: - statres = os.fstat(fileobj.fileno()) - linkname = "" - - stmd = statres.st_mode - if stat.S_ISREG(stmd): - inode = (statres.st_ino, statres.st_dev) - if not self.dereference and statres.st_nlink > 1 and \ - inode in self.inodes and arcname != self.inodes[inode]: - # Is it a hardlink to an already - # archived file? - type = LNKTYPE - linkname = self.inodes[inode] - else: - # The inode is added only if its valid. - # For win32 it is always 0. - type = REGTYPE - if inode[0]: - self.inodes[inode] = arcname - elif stat.S_ISDIR(stmd): - type = DIRTYPE - elif stat.S_ISFIFO(stmd): - type = FIFOTYPE - elif stat.S_ISLNK(stmd): - type = SYMTYPE - linkname = os.readlink(name) - elif stat.S_ISCHR(stmd): - type = CHRTYPE - elif stat.S_ISBLK(stmd): - type = BLKTYPE - else: - return None - - # Fill the TarInfo object with all - # information we can get. - tarinfo.name = arcname - tarinfo.mode = stmd - tarinfo.uid = statres.st_uid - tarinfo.gid = statres.st_gid - if type == REGTYPE: - tarinfo.size = statres.st_size - else: - tarinfo.size = 0 - tarinfo.mtime = statres.st_mtime - tarinfo.type = type - tarinfo.linkname = linkname - if pwd: - try: - tarinfo.uname = pwd.getpwuid(tarinfo.uid)[0] - except KeyError: - pass - if grp: - try: - tarinfo.gname = grp.getgrgid(tarinfo.gid)[0] - except KeyError: - pass - - if type in (CHRTYPE, BLKTYPE): - if hasattr(os, "major") and hasattr(os, "minor"): - tarinfo.devmajor = os.major(statres.st_rdev) - tarinfo.devminor = os.minor(statres.st_rdev) - return tarinfo - - def list(self, verbose=True): - """Print a table of contents to sys.stdout. If `verbose' is False, only - the names of the members are printed. If it is True, an `ls -l'-like - output is produced. - """ - self._check() - - for tarinfo in self: - if verbose: - print(filemode(tarinfo.mode), end=' ') - print("%s/%s" % (tarinfo.uname or tarinfo.uid, - tarinfo.gname or tarinfo.gid), end=' ') - if tarinfo.ischr() or tarinfo.isblk(): - print("%10s" % ("%d,%d" \ - % (tarinfo.devmajor, tarinfo.devminor)), end=' ') - else: - print("%10d" % tarinfo.size, end=' ') - print("%d-%02d-%02d %02d:%02d:%02d" \ - % time.localtime(tarinfo.mtime)[:6], end=' ') - - print(tarinfo.name + ("/" if tarinfo.isdir() else ""), end=' ') - - if verbose: - if tarinfo.issym(): - print("->", tarinfo.linkname, end=' ') - if tarinfo.islnk(): - print("link to", tarinfo.linkname, end=' ') - print() - - def add(self, name, arcname=None, recursive=True, exclude=None, filter=None): - """Add the file `name' to the archive. `name' may be any type of file - (directory, fifo, symbolic link, etc.). If given, `arcname' - specifies an alternative name for the file in the archive. - Directories are added recursively by default. This can be avoided by - setting `recursive' to False. `exclude' is a function that should - return True for each filename to be excluded. `filter' is a function - that expects a TarInfo object argument and returns the changed - TarInfo object, if it returns None the TarInfo object will be - excluded from the archive. - """ - self._check("aw") - - if arcname is None: - arcname = name - - # Exclude pathnames. - if exclude is not None: - import warnings - warnings.warn("use the filter argument instead", - DeprecationWarning, 2) - if exclude(name): - self._dbg(2, "tarfile: Excluded %r" % name) - return - - # Skip if somebody tries to archive the archive... - if self.name is not None and os.path.abspath(name) == self.name: - self._dbg(2, "tarfile: Skipped %r" % name) - return - - self._dbg(1, name) - - # Create a TarInfo object from the file. - tarinfo = self.gettarinfo(name, arcname) - - if tarinfo is None: - self._dbg(1, "tarfile: Unsupported type %r" % name) - return - - # Change or exclude the TarInfo object. - if filter is not None: - tarinfo = filter(tarinfo) - if tarinfo is None: - self._dbg(2, "tarfile: Excluded %r" % name) - return - - # Append the tar header and data to the archive. - if tarinfo.isreg(): - f = bltn_open(name, "rb") - self.addfile(tarinfo, f) - f.close() - - elif tarinfo.isdir(): - self.addfile(tarinfo) - if recursive: - for f in os.listdir(name): - self.add(os.path.join(name, f), os.path.join(arcname, f), - recursive, exclude, filter=filter) - - else: - self.addfile(tarinfo) - - def addfile(self, tarinfo, fileobj=None): - """Add the TarInfo object `tarinfo' to the archive. If `fileobj' is - given, tarinfo.size bytes are read from it and added to the archive. - You can create TarInfo objects using gettarinfo(). - On Windows platforms, `fileobj' should always be opened with mode - 'rb' to avoid irritation about the file size. - """ - self._check("aw") - - tarinfo = copy.copy(tarinfo) - - buf = tarinfo.tobuf(self.format, self.encoding, self.errors) - self.fileobj.write(buf) - self.offset += len(buf) - - # If there's data to follow, append it. - if fileobj is not None: - copyfileobj(fileobj, self.fileobj, tarinfo.size) - blocks, remainder = divmod(tarinfo.size, BLOCKSIZE) - if remainder > 0: - self.fileobj.write(NUL * (BLOCKSIZE - remainder)) - blocks += 1 - self.offset += blocks * BLOCKSIZE - - self.members.append(tarinfo) - - def extractall(self, path=".", members=None): - """Extract all members from the archive to the current working - directory and set owner, modification time and permissions on - directories afterwards. `path' specifies a different directory - to extract to. `members' is optional and must be a subset of the - list returned by getmembers(). - """ - directories = [] - - if members is None: - members = self - - for tarinfo in members: - if tarinfo.isdir(): - # Extract directories with a safe mode. - directories.append(tarinfo) - tarinfo = copy.copy(tarinfo) - tarinfo.mode = 0o700 - # Do not set_attrs directories, as we will do that further down - self.extract(tarinfo, path, set_attrs=not tarinfo.isdir()) - - # Reverse sort directories. - directories.sort(key=lambda a: a.name) - directories.reverse() - - # Set correct owner, mtime and filemode on directories. - for tarinfo in directories: - dirpath = os.path.join(path, tarinfo.name) - try: - self.chown(tarinfo, dirpath) - self.utime(tarinfo, dirpath) - self.chmod(tarinfo, dirpath) - except ExtractError as e: - if self.errorlevel > 1: - raise - else: - self._dbg(1, "tarfile: %s" % e) - - def extract(self, member, path="", set_attrs=True): - """Extract a member from the archive to the current working directory, - using its full name. Its file information is extracted as accurately - as possible. `member' may be a filename or a TarInfo object. You can - specify a different directory using `path'. File attributes (owner, - mtime, mode) are set unless `set_attrs' is False. - """ - self._check("r") - - if isinstance(member, str): - tarinfo = self.getmember(member) - else: - tarinfo = member - - # Prepare the link target for makelink(). - if tarinfo.islnk(): - tarinfo._link_target = os.path.join(path, tarinfo.linkname) - - try: - self._extract_member(tarinfo, os.path.join(path, tarinfo.name), - set_attrs=set_attrs) - except EnvironmentError as e: - if self.errorlevel > 0: - raise - else: - if e.filename is None: - self._dbg(1, "tarfile: %s" % e.strerror) - else: - self._dbg(1, "tarfile: %s %r" % (e.strerror, e.filename)) - except ExtractError as e: - if self.errorlevel > 1: - raise - else: - self._dbg(1, "tarfile: %s" % e) - - def extractfile(self, member): - """Extract a member from the archive as a file object. `member' may be - a filename or a TarInfo object. If `member' is a regular file, a - file-like object is returned. If `member' is a link, a file-like - object is constructed from the link's target. If `member' is none of - the above, None is returned. - The file-like object is read-only and provides the following - methods: read(), readline(), readlines(), seek() and tell() - """ - self._check("r") - - if isinstance(member, str): - tarinfo = self.getmember(member) - else: - tarinfo = member - - if tarinfo.isreg(): - return self.fileobject(self, tarinfo) - - elif tarinfo.type not in SUPPORTED_TYPES: - # If a member's type is unknown, it is treated as a - # regular file. - return self.fileobject(self, tarinfo) - - elif tarinfo.islnk() or tarinfo.issym(): - if isinstance(self.fileobj, _Stream): - # A small but ugly workaround for the case that someone tries - # to extract a (sym)link as a file-object from a non-seekable - # stream of tar blocks. - raise StreamError("cannot extract (sym)link as file object") - else: - # A (sym)link's file object is its target's file object. - return self.extractfile(self._find_link_target(tarinfo)) - else: - # If there's no data associated with the member (directory, chrdev, - # blkdev, etc.), return None instead of a file object. - return None - - def _extract_member(self, tarinfo, targetpath, set_attrs=True): - """Extract the TarInfo object tarinfo to a physical - file called targetpath. - """ - # Fetch the TarInfo object for the given name - # and build the destination pathname, replacing - # forward slashes to platform specific separators. - targetpath = targetpath.rstrip("/") - targetpath = targetpath.replace("/", os.sep) - - # Create all upper directories. - upperdirs = os.path.dirname(targetpath) - if upperdirs and not os.path.exists(upperdirs): - # Create directories that are not part of the archive with - # default permissions. - os.makedirs(upperdirs) - - if tarinfo.islnk() or tarinfo.issym(): - self._dbg(1, "%s -> %s" % (tarinfo.name, tarinfo.linkname)) - else: - self._dbg(1, tarinfo.name) - - if tarinfo.isreg(): - self.makefile(tarinfo, targetpath) - elif tarinfo.isdir(): - self.makedir(tarinfo, targetpath) - elif tarinfo.isfifo(): - self.makefifo(tarinfo, targetpath) - elif tarinfo.ischr() or tarinfo.isblk(): - self.makedev(tarinfo, targetpath) - elif tarinfo.islnk() or tarinfo.issym(): - self.makelink(tarinfo, targetpath) - elif tarinfo.type not in SUPPORTED_TYPES: - self.makeunknown(tarinfo, targetpath) - else: - self.makefile(tarinfo, targetpath) - - if set_attrs: - self.chown(tarinfo, targetpath) - if not tarinfo.issym(): - self.chmod(tarinfo, targetpath) - self.utime(tarinfo, targetpath) - - #-------------------------------------------------------------------------- - # Below are the different file methods. They are called via - # _extract_member() when extract() is called. They can be replaced in a - # subclass to implement other functionality. - - def makedir(self, tarinfo, targetpath): - """Make a directory called targetpath. - """ - try: - # Use a safe mode for the directory, the real mode is set - # later in _extract_member(). - os.mkdir(targetpath, 0o700) - except EnvironmentError as e: - if e.errno != errno.EEXIST: - raise - - def makefile(self, tarinfo, targetpath): - """Make a file called targetpath. - """ - source = self.fileobj - source.seek(tarinfo.offset_data) - target = bltn_open(targetpath, "wb") - if tarinfo.sparse is not None: - for offset, size in tarinfo.sparse: - target.seek(offset) - copyfileobj(source, target, size) - else: - copyfileobj(source, target, tarinfo.size) - target.seek(tarinfo.size) - target.truncate() - target.close() - - def makeunknown(self, tarinfo, targetpath): - """Make a file from a TarInfo object with an unknown type - at targetpath. - """ - self.makefile(tarinfo, targetpath) - self._dbg(1, "tarfile: Unknown file type %r, " \ - "extracted as regular file." % tarinfo.type) - - def makefifo(self, tarinfo, targetpath): - """Make a fifo called targetpath. - """ - if hasattr(os, "mkfifo"): - os.mkfifo(targetpath) - else: - raise ExtractError("fifo not supported by system") - - def makedev(self, tarinfo, targetpath): - """Make a character or block device called targetpath. - """ - if not hasattr(os, "mknod") or not hasattr(os, "makedev"): - raise ExtractError("special devices not supported by system") - - mode = tarinfo.mode - if tarinfo.isblk(): - mode |= stat.S_IFBLK - else: - mode |= stat.S_IFCHR - - os.mknod(targetpath, mode, - os.makedev(tarinfo.devmajor, tarinfo.devminor)) - - def makelink(self, tarinfo, targetpath): - """Make a (symbolic) link called targetpath. If it cannot be created - (platform limitation), we try to make a copy of the referenced file - instead of a link. - """ - try: - # For systems that support symbolic and hard links. - if tarinfo.issym(): - os.symlink(tarinfo.linkname, targetpath) - else: - # See extract(). - if os.path.exists(tarinfo._link_target): - os.link(tarinfo._link_target, targetpath) - else: - self._extract_member(self._find_link_target(tarinfo), - targetpath) - except symlink_exception: - if tarinfo.issym(): - linkpath = os.path.join(os.path.dirname(tarinfo.name), - tarinfo.linkname) - else: - linkpath = tarinfo.linkname - else: - try: - self._extract_member(self._find_link_target(tarinfo), - targetpath) - except KeyError: - raise ExtractError("unable to resolve link inside archive") - - def chown(self, tarinfo, targetpath): - """Set owner of targetpath according to tarinfo. - """ - if pwd and hasattr(os, "geteuid") and os.geteuid() == 0: - # We have to be root to do so. - try: - g = grp.getgrnam(tarinfo.gname)[2] - except KeyError: - g = tarinfo.gid - try: - u = pwd.getpwnam(tarinfo.uname)[2] - except KeyError: - u = tarinfo.uid - try: - if tarinfo.issym() and hasattr(os, "lchown"): - os.lchown(targetpath, u, g) - else: - if sys.platform != "os2emx": - os.chown(targetpath, u, g) - except EnvironmentError as e: - raise ExtractError("could not change owner") - - def chmod(self, tarinfo, targetpath): - """Set file permissions of targetpath according to tarinfo. - """ - if hasattr(os, 'chmod'): - try: - os.chmod(targetpath, tarinfo.mode) - except EnvironmentError as e: - raise ExtractError("could not change mode") - - def utime(self, tarinfo, targetpath): - """Set modification time of targetpath according to tarinfo. - """ - if not hasattr(os, 'utime'): - return - try: - os.utime(targetpath, (tarinfo.mtime, tarinfo.mtime)) - except EnvironmentError as e: - raise ExtractError("could not change modification time") - - #-------------------------------------------------------------------------- - def next(self): - """Return the next member of the archive as a TarInfo object, when - TarFile is opened for reading. Return None if there is no more - available. - """ - self._check("ra") - if self.firstmember is not None: - m = self.firstmember - self.firstmember = None - return m - - # Read the next block. - self.fileobj.seek(self.offset) - tarinfo = None - while True: - try: - tarinfo = self.tarinfo.fromtarfile(self) - except EOFHeaderError as e: - if self.ignore_zeros: - self._dbg(2, "0x%X: %s" % (self.offset, e)) - self.offset += BLOCKSIZE - continue - except InvalidHeaderError as e: - if self.ignore_zeros: - self._dbg(2, "0x%X: %s" % (self.offset, e)) - self.offset += BLOCKSIZE - continue - elif self.offset == 0: - raise ReadError(str(e)) - except EmptyHeaderError: - if self.offset == 0: - raise ReadError("empty file") - except TruncatedHeaderError as e: - if self.offset == 0: - raise ReadError(str(e)) - except SubsequentHeaderError as e: - raise ReadError(str(e)) - break - - if tarinfo is not None: - self.members.append(tarinfo) - else: - self._loaded = True - - return tarinfo - - #-------------------------------------------------------------------------- - # Little helper methods: - - def _getmember(self, name, tarinfo=None, normalize=False): - """Find an archive member by name from bottom to top. - If tarinfo is given, it is used as the starting point. - """ - # Ensure that all members have been loaded. - members = self.getmembers() - - # Limit the member search list up to tarinfo. - if tarinfo is not None: - members = members[:members.index(tarinfo)] - - if normalize: - name = os.path.normpath(name) - - for member in reversed(members): - if normalize: - member_name = os.path.normpath(member.name) - else: - member_name = member.name - - if name == member_name: - return member - - def _load(self): - """Read through the entire archive file and look for readable - members. - """ - while True: - tarinfo = self.next() - if tarinfo is None: - break - self._loaded = True - - def _check(self, mode=None): - """Check if TarFile is still open, and if the operation's mode - corresponds to TarFile's mode. - """ - if self.closed: - raise IOError("%s is closed" % self.__class__.__name__) - if mode is not None and self.mode not in mode: - raise IOError("bad operation for mode %r" % self.mode) - - def _find_link_target(self, tarinfo): - """Find the target member of a symlink or hardlink member in the - archive. - """ - if tarinfo.issym(): - # Always search the entire archive. - linkname = os.path.dirname(tarinfo.name) + "/" + tarinfo.linkname - limit = None - else: - # Search the archive before the link, because a hard link is - # just a reference to an already archived file. - linkname = tarinfo.linkname - limit = tarinfo - - member = self._getmember(linkname, tarinfo=limit, normalize=True) - if member is None: - raise KeyError("linkname %r not found" % linkname) - return member - - def __iter__(self): - """Provide an iterator object. - """ - if self._loaded: - return iter(self.members) - else: - return TarIter(self) - - def _dbg(self, level, msg): - """Write debugging output to sys.stderr. - """ - if level <= self.debug: - print(msg, file=sys.stderr) - - def __enter__(self): - self._check() - return self - - def __exit__(self, type, value, traceback): - if type is None: - self.close() - else: - # An exception occurred. We must not call close() because - # it would try to write end-of-archive blocks and padding. - if not self._extfileobj: - self.fileobj.close() - self.closed = True -# class TarFile - -class TarIter(object): - """Iterator Class. - - for tarinfo in TarFile(...): - suite... - """ - - def __init__(self, tarfile): - """Construct a TarIter object. - """ - self.tarfile = tarfile - self.index = 0 - def __iter__(self): - """Return iterator object. - """ - return self - - def __next__(self): - """Return the next item using TarFile's next() method. - When all members have been read, set TarFile as _loaded. - """ - # Fix for SF #1100429: Under rare circumstances it can - # happen that getmembers() is called during iteration, - # which will cause TarIter to stop prematurely. - if not self.tarfile._loaded: - tarinfo = self.tarfile.next() - if not tarinfo: - self.tarfile._loaded = True - raise StopIteration - else: - try: - tarinfo = self.tarfile.members[self.index] - except IndexError: - raise StopIteration - self.index += 1 - return tarinfo - - next = __next__ # for Python 2.x - -#-------------------- -# exported functions -#-------------------- -def is_tarfile(name): - """Return True if name points to a tar archive that we - are able to handle, else return False. - """ - try: - t = open(name) - t.close() - return True - except TarError: - return False - -bltn_open = open -open = TarFile.open diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/compat.py b/src/fetchcode/vcs/pip/_vendor/distlib/compat.py index ff328c8e..e93dc27a 100644 --- a/src/fetchcode/vcs/pip/_vendor/distlib/compat.py +++ b/src/fetchcode/vcs/pip/_vendor/distlib/compat.py @@ -8,6 +8,7 @@ import os import re +import shutil import sys try: @@ -22,7 +23,6 @@ from types import FileType as file_type import __builtin__ as builtins import ConfigParser as configparser - from ._backport import shutil from urlparse import urlparse, urlunparse, urljoin, urlsplit, urlunsplit from urllib import (urlretrieve, quote as _quote, unquote, url2pathname, pathname2url, ContentTooShortError, splittype) @@ -34,9 +34,8 @@ def quote(s): import urllib2 from urllib2 import (Request, urlopen, URLError, HTTPError, - HTTPBasicAuthHandler, HTTPPasswordMgr, - HTTPHandler, HTTPRedirectHandler, - build_opener) + HTTPBasicAuthHandler, HTTPPasswordMgr, HTTPHandler, + HTTPRedirectHandler, build_opener) if ssl: from urllib2 import HTTPSHandler import httplib @@ -48,17 +47,18 @@ def quote(s): from itertools import ifilter as filter from itertools import ifilterfalse as filterfalse - _userprog = None - def splituser(host): - """splituser('user[:passwd]@host[:port]') --> 'user[:passwd]', 'host[:port]'.""" - global _userprog - if _userprog is None: - import re - _userprog = re.compile('^(.*)@(.*)$') + # Leaving this around for now, in case it needs resurrecting in some way + # _userprog = None + # def splituser(host): + # """splituser('user[:passwd]@host[:port]') --> 'user[:passwd]', 'host[:port]'.""" + # global _userprog + # if _userprog is None: + # import re + # _userprog = re.compile('^(.*)@(.*)$') - match = _userprog.match(host) - if match: return match.group(1, 2) - return None, host + # match = _userprog.match(host) + # if match: return match.group(1, 2) + # return None, host else: # pragma: no cover from io import StringIO @@ -67,14 +67,12 @@ def splituser(host): from io import TextIOWrapper as file_type import builtins import configparser - import shutil - from urllib.parse import (urlparse, urlunparse, urljoin, splituser, quote, - unquote, urlsplit, urlunsplit, splittype) + from urllib.parse import (urlparse, urlunparse, urljoin, quote, unquote, + urlsplit, urlunsplit, splittype) from urllib.request import (urlopen, urlretrieve, Request, url2pathname, - pathname2url, - HTTPBasicAuthHandler, HTTPPasswordMgr, - HTTPHandler, HTTPRedirectHandler, - build_opener) + pathname2url, HTTPBasicAuthHandler, + HTTPPasswordMgr, HTTPHandler, + HTTPRedirectHandler, build_opener) if ssl: from urllib.request import HTTPSHandler from urllib.error import HTTPError, URLError, ContentTooShortError @@ -90,11 +88,11 @@ def splituser(host): try: from ssl import match_hostname, CertificateError -except ImportError: # pragma: no cover +except ImportError: # pragma: no cover + class CertificateError(ValueError): pass - def _dnsname_match(dn, hostname, max_wildcards=1): """Matching according to RFC 6125, section 6.4.3 @@ -144,7 +142,6 @@ def _dnsname_match(dn, hostname, max_wildcards=1): pat = re.compile(r'\A' + r'\.'.join(pats) + r'\Z', re.IGNORECASE) return pat.match(hostname) - def match_hostname(cert, hostname): """Verify that *cert* (in decoded format as returned by SSLSocket.getpeercert()) matches the *hostname*. RFC 2818 and RFC 6125 @@ -177,24 +174,26 @@ def match_hostname(cert, hostname): dnsnames.append(value) if len(dnsnames) > 1: raise CertificateError("hostname %r " - "doesn't match either of %s" - % (hostname, ', '.join(map(repr, dnsnames)))) + "doesn't match either of %s" % + (hostname, ', '.join(map(repr, dnsnames)))) elif len(dnsnames) == 1: raise CertificateError("hostname %r " - "doesn't match %r" - % (hostname, dnsnames[0])) + "doesn't match %r" % + (hostname, dnsnames[0])) else: raise CertificateError("no appropriate commonName or " - "subjectAltName fields were found") + "subjectAltName fields were found") try: from types import SimpleNamespace as Container except ImportError: # pragma: no cover + class Container(object): """ A generic container for when multiple values need to be returned """ + def __init__(self, **kwargs): self.__dict__.update(kwargs) @@ -213,6 +212,7 @@ def which(cmd, mode=os.F_OK | os.X_OK, path=None): path. """ + # Check that a given file can be accessed with the correct mode. # Additionally check that `file` is not a directory, as on Windows # directories pass the os.access check. @@ -236,7 +236,7 @@ def _access_check(fn, mode): if sys.platform == "win32": # The current directory takes precedence on Windows. - if not os.curdir in path: + if os.curdir not in path: path.insert(0, os.curdir) # PATHEXT is necessary to check on Windows. @@ -257,7 +257,7 @@ def _access_check(fn, mode): seen = set() for dir in path: normdir = os.path.normcase(dir) - if not normdir in seen: + if normdir not in seen: seen.add(normdir) for thefile in files: name = os.path.join(dir, thefile) @@ -276,6 +276,7 @@ def _access_check(fn, mode): from zipfile import ZipExtFile as BaseZipExtFile class ZipExtFile(BaseZipExtFile): + def __init__(self, base): self.__dict__.update(base.__dict__) @@ -287,6 +288,7 @@ def __exit__(self, *exc_info): # return None, so if an exception occurred, it will propagate class ZipFile(BaseZipFile): + def __enter__(self): return self @@ -298,9 +300,11 @@ def open(self, *args, **kwargs): base = BaseZipFile.open(self, *args, **kwargs) return ZipExtFile(base) + try: from platform import python_implementation -except ImportError: # pragma: no cover +except ImportError: # pragma: no cover + def python_implementation(): """Return a string identifying the Python implementation.""" if 'PyPy' in sys.version: @@ -311,15 +315,13 @@ def python_implementation(): return 'IronPython' return 'CPython' -try: - import sysconfig -except ImportError: # pragma: no cover - from ._backport import sysconfig + +import sysconfig try: callable = callable -except NameError: # pragma: no cover - from collections import Callable +except NameError: # pragma: no cover + from collections.abc import Callable def callable(obj): return isinstance(obj, Callable) @@ -359,11 +361,11 @@ def fsdecode(filename): raise TypeError("expect bytes or str, not %s" % type(filename).__name__) + try: from tokenize import detect_encoding -except ImportError: # pragma: no cover +except ImportError: # pragma: no cover from codecs import BOM_UTF8, lookup - import re cookie_re = re.compile(r"coding[:=]\s*([-\w.]+)") @@ -402,6 +404,7 @@ def detect_encoding(readline): bom_found = False encoding = None default = 'utf-8' + def read_or_stop(): try: return readline() @@ -431,8 +434,8 @@ def find_cookie(line): if filename is None: msg = "unknown encoding: " + encoding else: - msg = "unknown encoding for {!r}: {}".format(filename, - encoding) + msg = "unknown encoding for {!r}: {}".format( + filename, encoding) raise SyntaxError(msg) if bom_found: @@ -441,7 +444,8 @@ def find_cookie(line): if filename is None: msg = 'encoding problem: utf-8' else: - msg = 'encoding problem for {!r}: utf-8'.format(filename) + msg = 'encoding problem for {!r}: utf-8'.format( + filename) raise SyntaxError(msg) encoding += '-sig' return encoding @@ -468,6 +472,7 @@ def find_cookie(line): return default, [first, second] + # For converting & <-> & etc. try: from html import escape @@ -480,12 +485,13 @@ def find_cookie(line): try: from collections import ChainMap -except ImportError: # pragma: no cover +except ImportError: # pragma: no cover from collections import MutableMapping try: from reprlib import recursive_repr as _recursive_repr except ImportError: + def _recursive_repr(fillvalue='...'): ''' Decorator to make a repr function return fillvalue for a recursive @@ -510,13 +516,15 @@ def wrapper(self): wrapper.__module__ = getattr(user_function, '__module__') wrapper.__doc__ = getattr(user_function, '__doc__') wrapper.__name__ = getattr(user_function, '__name__') - wrapper.__annotations__ = getattr(user_function, '__annotations__', {}) + wrapper.__annotations__ = getattr(user_function, + '__annotations__', {}) return wrapper return decorating_function class ChainMap(MutableMapping): - ''' A ChainMap groups multiple dicts (or other mappings) together + ''' + A ChainMap groups multiple dicts (or other mappings) together to create a single, updateable view. The underlying mappings are stored in a list. That list is public and can @@ -525,7 +533,6 @@ class ChainMap(MutableMapping): Lookups search the underlying mappings successively until a key is found. In contrast, writes, updates, and deletions only operate on the first mapping. - ''' def __init__(self, *maps): @@ -533,7 +540,7 @@ def __init__(self, *maps): If no mappings are provided, a single empty dictionary is used. ''' - self.maps = list(maps) or [{}] # always at least one map + self.maps = list(maps) or [{}] # always at least one map def __missing__(self, key): raise KeyError(key) @@ -541,16 +548,19 @@ def __missing__(self, key): def __getitem__(self, key): for mapping in self.maps: try: - return mapping[key] # can't use 'key in mapping' with defaultdict + return mapping[ + key] # can't use 'key in mapping' with defaultdict except KeyError: pass - return self.__missing__(key) # support subclasses that define __missing__ + return self.__missing__( + key) # support subclasses that define __missing__ def get(self, key, default=None): return self[key] if key in self else default def __len__(self): - return len(set().union(*self.maps)) # reuses stored hash values if possible + return len(set().union( + *self.maps)) # reuses stored hash values if possible def __iter__(self): return iter(set().union(*self.maps)) @@ -577,12 +587,12 @@ def copy(self): __copy__ = copy - def new_child(self): # like Django's Context.push() + def new_child(self): # like Django's Context.push() 'New ChainMap with a new dict followed by all previous maps.' return self.__class__({}, *self.maps) @property - def parents(self): # like Django's Context.pop() + def parents(self): # like Django's Context.pop() 'New ChainMap from maps[1:].' return self.__class__(*self.maps[1:]) @@ -593,7 +603,8 @@ def __delitem__(self, key): try: del self.maps[0][key] except KeyError: - raise KeyError('Key not found in the first mapping: {!r}'.format(key)) + raise KeyError( + 'Key not found in the first mapping: {!r}'.format(key)) def popitem(self): 'Remove and return an item pair from maps[0]. Raise KeyError is maps[0] is empty.' @@ -607,34 +618,35 @@ def pop(self, key, *args): try: return self.maps[0].pop(key, *args) except KeyError: - raise KeyError('Key not found in the first mapping: {!r}'.format(key)) + raise KeyError( + 'Key not found in the first mapping: {!r}'.format(key)) def clear(self): 'Clear maps[0], leaving maps[1:] intact.' self.maps[0].clear() + try: from importlib.util import cache_from_source # Python >= 3.4 except ImportError: # pragma: no cover - try: - from imp import cache_from_source - except ImportError: # pragma: no cover - def cache_from_source(path, debug_override=None): - assert path.endswith('.py') - if debug_override is None: - debug_override = __debug__ - if debug_override: - suffix = 'c' - else: - suffix = 'o' - return path + suffix + + def cache_from_source(path, debug_override=None): + assert path.endswith('.py') + if debug_override is None: + debug_override = __debug__ + if debug_override: + suffix = 'c' + else: + suffix = 'o' + return path + suffix + try: from collections import OrderedDict -except ImportError: # pragma: no cover -## {{{ http://code.activestate.com/recipes/576693/ (r9) -# Backport of OrderedDict() class that runs on Python 2.4, 2.5, 2.6, 2.7 and pypy. -# Passes Python2.7's test suite and incorporates all the latest updates. +except ImportError: # pragma: no cover + # {{{ http://code.activestate.com/recipes/576693/ (r9) + # Backport of OrderedDict() class that runs on Python 2.4, 2.5, 2.6, 2.7 and pypy. + # Passes Python2.7's test suite and incorporates all the latest updates. try: from thread import get_ident as _get_ident except ImportError: @@ -645,9 +657,9 @@ def cache_from_source(path, debug_override=None): except ImportError: pass - class OrderedDict(dict): 'Dictionary that remembers insertion order' + # An inherited dict maps keys to values. # The inherited dict provides __getitem__, __len__, __contains__, and get. # The remaining methods are order-aware. @@ -665,11 +677,12 @@ def __init__(self, *args, **kwds): ''' if len(args) > 1: - raise TypeError('expected at most 1 arguments, got %d' % len(args)) + raise TypeError('expected at most 1 arguments, got %d' % + len(args)) try: self.__root except AttributeError: - self.__root = root = [] # sentinel node + self.__root = root = [] # sentinel node root[:] = [root, root, None] self.__map = {} self.__update(*args, **kwds) @@ -783,7 +796,7 @@ def update(*args, **kwds): ''' if len(args) > 2: raise TypeError('update() takes at most 2 positional ' - 'arguments (%d given)' % (len(args),)) + 'arguments (%d given)' % (len(args), )) elif not args: raise TypeError('update() takes at least 1 argument (0 given)') self = args[0] @@ -829,14 +842,15 @@ def setdefault(self, key, default=None): def __repr__(self, _repr_running=None): 'od.__repr__() <==> repr(od)' - if not _repr_running: _repr_running = {} + if not _repr_running: + _repr_running = {} call_key = id(self), _get_ident() if call_key in _repr_running: return '...' _repr_running[call_key] = 1 try: if not self: - return '%s()' % (self.__class__.__name__,) + return '%s()' % (self.__class__.__name__, ) return '%s(%r)' % (self.__class__.__name__, self.items()) finally: del _repr_running[call_key] @@ -848,8 +862,8 @@ def __reduce__(self): for k in vars(OrderedDict()): inst_dict.pop(k, None) if inst_dict: - return (self.__class__, (items,), inst_dict) - return self.__class__, (items,) + return (self.__class__, (items, ), inst_dict) + return self.__class__, (items, ) def copy(self): 'od.copy() -> a shallow copy of od' @@ -872,7 +886,8 @@ def __eq__(self, other): ''' if isinstance(other, OrderedDict): - return len(self)==len(other) and self.items() == other.items() + return len(self) == len( + other) and self.items() == other.items() return dict.__eq__(self, other) def __ne__(self, other): @@ -892,19 +907,18 @@ def viewitems(self): "od.viewitems() -> a set-like object providing a view on od's items" return ItemsView(self) + try: from logging.config import BaseConfigurator, valid_ident -except ImportError: # pragma: no cover +except ImportError: # pragma: no cover IDENTIFIER = re.compile('^[a-z_][a-z0-9_]*$', re.I) - def valid_ident(s): m = IDENTIFIER.match(s) if not m: raise ValueError('Not a valid Python identifier: %r' % s) return True - # The ConvertingXXX classes are wrappers around standard Python containers, # and they serve to convert any suitable values in the container. The # conversion converts base dicts, lists and tuples to their wrapped @@ -920,7 +934,7 @@ class ConvertingDict(dict): def __getitem__(self, key): value = dict.__getitem__(self, key) result = self.configurator.convert(value) - #If the converted value is different, save for next time + # If the converted value is different, save for next time if value is not result: self[key] = result if type(result) in (ConvertingDict, ConvertingList, @@ -932,7 +946,7 @@ def __getitem__(self, key): def get(self, key, default=None): value = dict.get(self, key, default) result = self.configurator.convert(value) - #If the converted value is different, save for next time + # If the converted value is different, save for next time if value is not result: self[key] = result if type(result) in (ConvertingDict, ConvertingList, @@ -953,10 +967,11 @@ def pop(self, key, default=None): class ConvertingList(list): """A converting list wrapper.""" + def __getitem__(self, key): value = list.__getitem__(self, key) result = self.configurator.convert(value) - #If the converted value is different, save for next time + # If the converted value is different, save for next time if value is not result: self[key] = result if type(result) in (ConvertingDict, ConvertingList, @@ -976,6 +991,7 @@ def pop(self, idx=-1): class ConvertingTuple(tuple): """A converting tuple wrapper.""" + def __getitem__(self, key): value = tuple.__getitem__(self, key) result = self.configurator.convert(value) @@ -999,8 +1015,8 @@ class BaseConfigurator(object): DIGIT_PATTERN = re.compile(r'^\d+$') value_converters = { - 'ext' : 'ext_convert', - 'cfg' : 'cfg_convert', + 'ext': 'ext_convert', + 'cfg': 'cfg_convert', } # We might want to use a different one, e.g. importlib @@ -1046,7 +1062,6 @@ def cfg_convert(self, value): else: rest = rest[m.end():] d = self.config[m.groups()[0]] - #print d, rest while rest: m = self.DOT_PATTERN.match(rest) if m: @@ -1059,7 +1074,9 @@ def cfg_convert(self, value): d = d[idx] else: try: - n = int(idx) # try as number first (most likely) + n = int( + idx + ) # try as number first (most likely) d = d[n] except TypeError: d = d[idx] @@ -1068,7 +1085,7 @@ def cfg_convert(self, value): else: raise ValueError('Unable to convert ' '%r at %r' % (value, rest)) - #rest should be empty + # rest should be empty return d def convert(self, value): @@ -1077,14 +1094,15 @@ def convert(self, value): replaced by their converting alternatives. Strings are checked to see if they have a conversion format and are converted if they do. """ - if not isinstance(value, ConvertingDict) and isinstance(value, dict): + if not isinstance(value, ConvertingDict) and isinstance( + value, dict): value = ConvertingDict(value) value.configurator = self - elif not isinstance(value, ConvertingList) and isinstance(value, list): + elif not isinstance(value, ConvertingList) and isinstance( + value, list): value = ConvertingList(value) value.configurator = self - elif not isinstance(value, ConvertingTuple) and\ - isinstance(value, tuple): + elif not isinstance(value, ConvertingTuple) and isinstance(value, tuple): value = ConvertingTuple(value) value.configurator = self elif isinstance(value, string_types): diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/database.py b/src/fetchcode/vcs/pip/_vendor/distlib/database.py index c16c0c8d..eb3765f1 100644 --- a/src/fetchcode/vcs/pip/_vendor/distlib/database.py +++ b/src/fetchcode/vcs/pip/_vendor/distlib/database.py @@ -1,6 +1,6 @@ # -*- coding: utf-8 -*- # -# Copyright (C) 2012-2017 The Python Software Foundation. +# Copyright (C) 2012-2023 The Python Software Foundation. # See LICENSE.txt and CONTRIBUTORS.txt. # """PEP 376 implementation.""" @@ -25,11 +25,10 @@ from .util import (parse_requirement, cached_property, parse_name_and_version, read_exports, write_exports, CSVReader, CSVWriter) - -__all__ = ['Distribution', 'BaseInstalledDistribution', - 'InstalledDistribution', 'EggInfoDistribution', - 'DistributionPath'] - +__all__ = [ + 'Distribution', 'BaseInstalledDistribution', 'InstalledDistribution', + 'EggInfoDistribution', 'DistributionPath' +] logger = logging.getLogger(__name__) @@ -46,6 +45,7 @@ class _Cache(object): """ A simple cache mapping names and .dist-info paths to distributions """ + def __init__(self): """ Initialise an instance. There is normally one for each DistributionPath. @@ -76,6 +76,7 @@ class DistributionPath(object): """ Represents a set of distributions installed on a path (typically sys.path). """ + def __init__(self, path=None, include_egg=False): """ Create an instance from a path, optionally including legacy (distutils/ @@ -111,7 +112,6 @@ def clear_cache(self): self._cache.clear() self._cache_egg.clear() - def _yield_distributions(self): """ Yield .dist-info and/or .egg(-info) distributions. @@ -132,29 +132,39 @@ def _yield_distributions(self): r = finder.find(entry) if not r or r.path in seen: continue - if self._include_dist and entry.endswith(DISTINFO_EXT): - possible_filenames = [METADATA_FILENAME, - WHEEL_METADATA_FILENAME, - LEGACY_METADATA_FILENAME] - for metadata_filename in possible_filenames: - metadata_path = posixpath.join(entry, metadata_filename) - pydist = finder.find(metadata_path) - if pydist: - break - else: - continue + try: + if self._include_dist and entry.endswith(DISTINFO_EXT): + possible_filenames = [ + METADATA_FILENAME, WHEEL_METADATA_FILENAME, + LEGACY_METADATA_FILENAME + ] + for metadata_filename in possible_filenames: + metadata_path = posixpath.join( + entry, metadata_filename) + pydist = finder.find(metadata_path) + if pydist: + break + else: + continue - with contextlib.closing(pydist.as_stream()) as stream: - metadata = Metadata(fileobj=stream, scheme='legacy') - logger.debug('Found %s', r.path) - seen.add(r.path) - yield new_dist_class(r.path, metadata=metadata, - env=self) - elif self._include_egg and entry.endswith(('.egg-info', - '.egg')): - logger.debug('Found %s', r.path) - seen.add(r.path) - yield old_dist_class(r.path, self) + with contextlib.closing(pydist.as_stream()) as stream: + metadata = Metadata(fileobj=stream, + scheme='legacy') + logger.debug('Found %s', r.path) + seen.add(r.path) + yield new_dist_class(r.path, + metadata=metadata, + env=self) + elif self._include_egg and entry.endswith( + ('.egg-info', '.egg')): + logger.debug('Found %s', r.path) + seen.add(r.path) + yield old_dist_class(r.path, self) + except Exception as e: + msg = 'Unable to read distribution at %s, perhaps due to bad metadata: %s' + logger.warning(msg, r.path, e) + import warnings + warnings.warn(msg % (r.path, e), stacklevel=2) def _generate_cache(self): """ @@ -265,7 +275,7 @@ def provides_distribution(self, name, version=None): matcher = self._scheme.matcher('%s (%s)' % (name, version)) except ValueError: raise DistlibException('invalid name or version: %r, %r' % - (name, version)) + (name, version)) for dist in self.get_distributions(): # We hit a problem on Travis where enum34 was installed and doesn't @@ -340,12 +350,12 @@ def __init__(self, metadata): """ self.metadata = metadata self.name = metadata.name - self.key = self.name.lower() # for case-insensitive comparisons + self.key = self.name.lower() # for case-insensitive comparisons self.version = metadata.version self.locator = None self.digest = None - self.extras = None # additional features requested - self.context = None # environment marker overrides + self.extras = None # additional features requested + self.context = None # environment marker overrides self.download_urls = set() self.digests = {} @@ -356,7 +366,7 @@ def source_url(self): """ return self.metadata.source_url - download_url = source_url # Backward compatibility + download_url = source_url # Backward compatibility @property def name_and_version(self): @@ -379,10 +389,11 @@ def provides(self): def _get_requirements(self, req_attr): md = self.metadata - logger.debug('Getting requirements from metadata %r', md.todict()) reqts = getattr(md, req_attr) - return set(md.get_requirements(reqts, extras=self.extras, - env=self.context)) + logger.debug('%s: got requirements %r from metadata: %r', self.name, + req_attr, reqts) + return set( + md.get_requirements(reqts, extras=self.extras, env=self.context)) @property def run_requires(self): @@ -419,12 +430,11 @@ def matches_requirement(self, req): matcher = scheme.matcher(r.requirement) except UnsupportedVersionError: # XXX compat-mode if cannot read the version - logger.warning('could not read version %r - using name only', - req) + logger.warning('could not read version %r - using name only', req) name = req.split()[0] matcher = scheme.matcher(name) - name = matcher.key # case-insensitive + name = matcher.key # case-insensitive result = False for p in self.provides: @@ -459,9 +469,8 @@ def __eq__(self, other): if type(other) is not type(self): result = False else: - result = (self.name == other.name and - self.version == other.version and - self.source_url == other.source_url) + result = (self.name == other.name and self.version == other.version + and self.source_url == other.source_url) return result def __hash__(self): @@ -550,10 +559,10 @@ def __init__(self, path, metadata=None, env=None): r = finder.find(WHEEL_METADATA_FILENAME) # Temporary - for legacy support if r is None: - r = finder.find('METADATA') + r = finder.find(LEGACY_METADATA_FILENAME) if r is None: - raise ValueError('no %s found in %s' % (METADATA_FILENAME, - path)) + raise ValueError('no %s found in %s' % + (METADATA_FILENAME, path)) with contextlib.closing(r.as_stream()) as stream: metadata = Metadata(fileobj=stream, scheme='legacy') @@ -564,7 +573,7 @@ def __init__(self, path, metadata=None, env=None): r = finder.find('REQUESTED') self.requested = r is not None - p = os.path.join(path, 'top_level.txt') + p = os.path.join(path, 'top_level.txt') if os.path.exists(p): with open(p, 'rb') as f: data = f.read().decode('utf-8') @@ -589,14 +598,14 @@ def _get_records(self): with contextlib.closing(r.as_stream()) as stream: with CSVReader(stream=stream) as record_reader: # Base location is parent dir of .dist-info dir - #base_location = os.path.dirname(self.path) - #base_location = os.path.abspath(base_location) + # base_location = os.path.dirname(self.path) + # base_location = os.path.abspath(base_location) for row in record_reader: missing = [None for i in range(len(row), 3)] path, checksum, size = row + missing - #if not os.path.isabs(path): - # path = path.replace('/', os.sep) - # path = os.path.join(base_location, path) + # if not os.path.isabs(path): + # path = path.replace('/', os.sep) + # path = os.path.join(base_location, path) results.append((path, checksum, size)) return results @@ -694,8 +703,8 @@ def write_installed_files(self, paths, prefix, dry_run=False): size = '%d' % os.path.getsize(path) with open(path, 'rb') as fp: hash_value = self.get_hash(fp.read()) - if path.startswith(base) or (base_under_prefix and - path.startswith(prefix)): + if path.startswith(base) or (base_under_prefix + and path.startswith(prefix)): path = os.path.relpath(path, base) writer.writerow((path, hash_value, size)) @@ -737,7 +746,8 @@ def check_installed_files(self): with open(path, 'rb') as f: actual_hash = self.get_hash(f.read(), hasher) if actual_hash != hash_value: - mismatches.append((path, 'hash', hash_value, actual_hash)) + mismatches.append( + (path, 'hash', hash_value, actual_hash)) return mismatches @cached_property @@ -784,7 +794,7 @@ def write_shared_locations(self, paths, dry_run=False): for key in ('prefix', 'lib', 'headers', 'scripts', 'data'): path = paths[key] if os.path.isdir(paths[key]): - lines.append('%s=%s' % (key, path)) + lines.append('%s=%s' % (key, path)) for ns in paths.get('namespace', ()): lines.append('namespace=%s' % ns) @@ -847,8 +857,8 @@ def list_distinfo_files(self): yield path def __eq__(self, other): - return (isinstance(other, InstalledDistribution) and - self.path == other.path) + return (isinstance(other, InstalledDistribution) + and self.path == other.path) # See http://docs.python.org/reference/datamodel#object.__hash__ __hash__ = object.__hash__ @@ -860,13 +870,14 @@ class EggInfoDistribution(BaseInstalledDistribution): if the given path happens to be a directory, the metadata is read from the file ``PKG-INFO`` under that directory.""" - requested = True # as we have no way of knowing, assume it was + requested = True # as we have no way of knowing, assume it was shared_locations = {} def __init__(self, path, env=None): + def set_name_and_version(s, n, v): s.name = n - s.key = n.lower() # for case-insensitive comparisons + s.key = n.lower() # for case-insensitive comparisons s.version = v self.path = path @@ -896,15 +907,18 @@ def parse_requires_data(data): lines = data.splitlines() for line in lines: line = line.strip() - if line.startswith('['): - logger.warning('Unexpected line: quitting requirement scan: %r', - line) + # sectioned files have bare newlines (separating sections) + if not line: # pragma: no cover + continue + if line.startswith('['): # pragma: no cover + logger.warning( + 'Unexpected line: quitting requirement scan: %r', line) break r = parse_requirement(line) - if not r: + if not r: # pragma: no cover logger.warning('Not recognised as a requirement: %r', line) continue - if r.extras: + if r.extras: # pragma: no cover logger.warning('extra requirements in requires.txt are ' 'not supported') if not r.constraints: @@ -945,7 +959,8 @@ def parse_requires_path(req_path): metadata = Metadata(fileobj=fileobj, scheme='legacy') try: data = zipf.get_data('EGG-INFO/requires.txt') - tl_data = zipf.get_data('EGG-INFO/top_level.txt').decode('utf-8') + tl_data = zipf.get_data('EGG-INFO/top_level.txt').decode( + 'utf-8') requires = parse_requires_data(data.decode('utf-8')) except IOError: requires = None @@ -975,8 +990,8 @@ def parse_requires_path(req_path): return metadata def __repr__(self): - return '' % ( - self.name, self.version, self.path) + return '' % (self.name, self.version, + self.path) def __str__(self): return "%s %s" % (self.name, self.version) @@ -1032,7 +1047,7 @@ def _size(path): logger.warning('Non-existent file: %s', p) if p.endswith(('.pyc', '.pyo')): continue - #otherwise fall through and fail + # otherwise fall through and fail if not os.path.isdir(p): result.append((p, _md5(p), _size(p))) result.append((record_path, None, None)) @@ -1068,12 +1083,13 @@ def list_distinfo_files(self, absolute=False): yield line def __eq__(self, other): - return (isinstance(other, EggInfoDistribution) and - self.path == other.path) + return (isinstance(other, EggInfoDistribution) + and self.path == other.path) # See http://docs.python.org/reference/datamodel#object.__hash__ __hash__ = object.__hash__ + new_dist_class = InstalledDistribution old_dist_class = EggInfoDistribution @@ -1107,7 +1123,7 @@ def add_distribution(self, distribution): """ self.adjacency_list[distribution] = [] self.reverse_list[distribution] = [] - #self.missing[distribution] = [] + # self.missing[distribution] = [] def add_edge(self, x, y, label=None): """Add an edge from distribution *x* to distribution *y* with the given @@ -1167,7 +1183,7 @@ def to_dot(self, f, skip_disconnected=True): if len(adjs) == 0 and not skip_disconnected: disconnected.append(dist) for other, label in adjs: - if not label is None: + if label is not None: f.write('"%s" -> "%s" [label="%s"]\n' % (dist.name, other.name, label)) else: @@ -1245,8 +1261,8 @@ def make_graph(dists, scheme='default'): # now make the edges for dist in dists: - requires = (dist.run_requires | dist.meta_requires | - dist.build_requires | dist.dev_requires) + requires = (dist.run_requires | dist.meta_requires + | dist.build_requires | dist.dev_requires) for req in requires: try: matcher = scheme.matcher(req) @@ -1257,7 +1273,7 @@ def make_graph(dists, scheme='default'): name = req.split()[0] matcher = scheme.matcher(name) - name = matcher.key # case-insensitive + name = matcher.key # case-insensitive matched = False if name in provided: @@ -1308,22 +1324,26 @@ def get_required_dists(dists, dist): :param dists: a list of distributions :param dist: a distribution, member of *dists* for which we are interested + in finding the dependencies. """ if dist not in dists: raise DistlibException('given distribution %r is not a member ' 'of the list' % dist.name) graph = make_graph(dists) - req = [] # required distributions + req = set() # required distributions todo = graph.adjacency_list[dist] # list of nodes we should inspect + seen = set(t[0] for t in todo) # already added to todo while todo: d = todo.pop()[0] - req.append(d) - for pred in graph.adjacency_list[d]: - if pred not in req: + req.add(d) + pred_list = graph.adjacency_list[d] + for pred in pred_list: + d = pred[0] + if d not in req and d not in seen: + seen.add(d) todo.append(pred) - return req diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/index.py b/src/fetchcode/vcs/pip/_vendor/distlib/index.py index 7a87cdcf..56cd2867 100644 --- a/src/fetchcode/vcs/pip/_vendor/distlib/index.py +++ b/src/fetchcode/vcs/pip/_vendor/distlib/index.py @@ -1,6 +1,6 @@ # -*- coding: utf-8 -*- # -# Copyright (C) 2013 Vinay Sajip. +# Copyright (C) 2013-2023 Vinay Sajip. # Licensed to the Python Software Foundation under a contributor agreement. # See LICENSE.txt and CONTRIBUTORS.txt. # @@ -12,19 +12,20 @@ import tempfile try: from threading import Thread -except ImportError: +except ImportError: # pragma: no cover from dummy_threading import Thread from . import DistlibException from .compat import (HTTPBasicAuthHandler, Request, HTTPPasswordMgr, urlparse, build_opener, string_types) -from .util import cached_property, zip_dir, ServerProxy +from .util import zip_dir, ServerProxy logger = logging.getLogger(__name__) DEFAULT_INDEX = 'https://pypi.org/pypi' DEFAULT_REALM = 'pypi' + class PackageIndex(object): """ This class represents a package index compatible with PyPI, the Python @@ -67,21 +68,17 @@ def _get_pypirc_command(self): Get the distutils command for interacting with PyPI configurations. :return: the command. """ - from distutils.core import Distribution - from distutils.config import PyPIRCCommand - d = Distribution() - return PyPIRCCommand(d) + from .util import _get_pypirc_command as cmd + return cmd() def read_configuration(self): """ - Read the PyPI access configuration as supported by distutils, getting - PyPI to do the actual work. This populates ``username``, ``password``, - ``realm`` and ``url`` attributes from the configuration. + Read the PyPI access configuration as supported by distutils. This populates + ``username``, ``password``, ``realm`` and ``url`` attributes from the + configuration. """ - # get distutils to do the work - c = self._get_pypirc_command() - c.repository = self.url - cfg = c._read_pypirc() + from .util import _load_pypirc + cfg = _load_pypirc(self) self.username = cfg.get('username') self.password = cfg.get('password') self.realm = cfg.get('realm', 'pypi') @@ -91,13 +88,10 @@ def save_configuration(self): """ Save the PyPI access configuration. You must have set ``username`` and ``password`` attributes before calling this method. - - Again, distutils is used to do the actual work. """ self.check_credentials() - # get distutils to do the work - c = self._get_pypirc_command() - c._store_pypirc(self.username, self.password) + from .util import _store_pypirc + _store_pypirc(self) def check_credentials(self): """ @@ -111,7 +105,7 @@ def check_credentials(self): pm.add_password(self.realm, netloc, self.username, self.password) self.password_handler = HTTPBasicAuthHandler(pm) - def register(self, metadata): + def register(self, metadata): # pragma: no cover """ Register a distribution on PyPI, using the provided metadata. @@ -126,7 +120,7 @@ def register(self, metadata): d = metadata.todict() d[':action'] = 'verify' request = self.encode_request(d.items(), []) - response = self.send_request(request) + self.send_request(request) d[':action'] = 'submit' request = self.encode_request(d.items(), []) return self.send_request(request) @@ -149,8 +143,7 @@ def _reader(self, name, stream, outbuf): logger.debug('%s: %s' % (name, s)) stream.close() - def get_sign_command(self, filename, signer, sign_password, - keystore=None): + def get_sign_command(self, filename, signer, sign_password, keystore=None): # pragma: no cover """ Return a suitable command for signing a file. @@ -213,7 +206,7 @@ def run_command(self, cmd, input_data=None): t2.join() return p.returncode, stdout, stderr - def sign_file(self, filename, signer, sign_password, keystore=None): + def sign_file(self, filename, signer, sign_password, keystore=None): # pragma: no cover """ Sign a file. @@ -293,7 +286,7 @@ def upload_file(self, metadata, filename, signer=None, sign_password=None, request = self.encode_request(d.items(), files) return self.send_request(request) - def upload_documentation(self, metadata, doc_dir): + def upload_documentation(self, metadata, doc_dir): # pragma: no cover """ Upload documentation to the index. @@ -366,8 +359,7 @@ def verify_signature(self, signature_filename, data_filename, keystore) rc, stdout, stderr = self.run_command(cmd) if rc not in (0, 1): - raise DistlibException('verify command failed with error ' - 'code %s' % rc) + raise DistlibException('verify command failed with error code %s' % rc) return rc == 0 def download_file(self, url, destfile, digest=None, reporthook=None): @@ -506,7 +498,7 @@ def encode_request(self, fields, files): } return Request(self.url, body, headers) - def search(self, terms, operator=None): + def search(self, terms, operator=None): # pragma: no cover if isinstance(terms, string_types): terms = {'name': terms} rpc_proxy = ServerProxy(self.url, timeout=3.0) diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/locators.py b/src/fetchcode/vcs/pip/_vendor/distlib/locators.py index 12a1d063..f9f0788f 100644 --- a/src/fetchcode/vcs/pip/_vendor/distlib/locators.py +++ b/src/fetchcode/vcs/pip/_vendor/distlib/locators.py @@ -1,6 +1,6 @@ # -*- coding: utf-8 -*- # -# Copyright (C) 2012-2015 Vinay Sajip. +# Copyright (C) 2012-2023 Vinay Sajip. # Licensed to the Python Software Foundation under a contributor agreement. # See LICENSE.txt and CONTRIBUTORS.txt. # @@ -20,14 +20,14 @@ from . import DistlibException from .compat import (urljoin, urlparse, urlunparse, url2pathname, pathname2url, - queue, quote, unescape, string_types, build_opener, + queue, quote, unescape, build_opener, HTTPRedirectHandler as BaseRedirectHandler, text_type, Request, HTTPError, URLError) from .database import Distribution, DistributionPath, make_dist from .metadata import Metadata, MetadataInvalidError -from .util import (cached_property, parse_credentials, ensure_slash, - split_filename, get_project_data, parse_requirement, - parse_name_and_version, ServerProxy, normalize_name) +from .util import (cached_property, ensure_slash, split_filename, get_project_data, + parse_requirement, parse_name_and_version, ServerProxy, + normalize_name) from .version import get_scheme, UnsupportedVersionError from .wheel import Wheel, is_compatible @@ -38,6 +38,7 @@ HTML_CONTENT_TYPE = re.compile('text/html|application/x(ht)?ml') DEFAULT_INDEX = 'https://pypi.org/pypi' + def get_all_distribution_names(url=None): """ Return all distribution names known by an index. @@ -52,6 +53,7 @@ def get_all_distribution_names(url=None): finally: client('close')() + class RedirectHandler(BaseRedirectHandler): """ A class to work around a bug in some Python 3.2.x releases. @@ -83,6 +85,7 @@ def http_error_302(self, req, fp, code, msg, headers): http_error_301 = http_error_303 = http_error_307 = http_error_302 + class Locator(object): """ A base class for locators - things that locate distributions. @@ -272,7 +275,7 @@ def same_project(name1, name2): 'python-version': ', '.join( ['.'.join(list(v[2:])) for v in wheel.pyver]), } - except Exception as e: # pragma: no cover + except Exception: # pragma: no cover logger.warning('invalid path for wheel: %s', path) elif not path.endswith(self.downloadable_extensions): # pragma: no cover logger.debug('Not downloadable: %s', path) @@ -293,7 +296,6 @@ def same_project(name1, name2): 'filename': filename, 'url': urlunparse((scheme, netloc, origpath, params, query, '')), - #'packagetype': 'sdist', } if pyver: # pragma: no cover result['python-version'] = pyver @@ -378,16 +380,13 @@ def locate(self, requirement, prereleases=False): continue try: if not matcher.match(k): - logger.debug('%s did not match %r', matcher, k) + pass # logger.debug('%s did not match %r', matcher, k) else: if prereleases or not vcls(k).is_prerelease: slist.append(k) - else: - logger.debug('skipping pre-release ' - 'version %s of %s', k, matcher.name) except Exception: # pragma: no cover logger.warning('error matching %s with %r', matcher, k) - pass # slist.append(k) + pass # slist.append(k) if len(slist) > 1: slist = sorted(slist, key=scheme.key) if slist: @@ -456,6 +455,7 @@ def _get_project(self, name): result['digests'][url] = digest return result + class PyPIJSONLocator(Locator): """ This locator uses PyPI's JSON interface. It's very limited in functionality @@ -476,7 +476,7 @@ def _get_project(self, name): url = urljoin(self.base_url, '%s/json' % quote(name)) try: resp = self.opener.open(url) - data = resp.read().decode() # for now + data = resp.read().decode() # for now d = json.loads(data) md = Metadata(scheme=self.scheme) data = d['info'] @@ -487,7 +487,7 @@ def _get_project(self, name): md.summary = data.get('summary') dist = Distribution(md) dist.locator = self - urls = d['urls'] + # urls = d['urls'] result[md.version] = dist for info in d['urls']: url = info['url'] @@ -593,7 +593,7 @@ class SimpleScrapingLocator(Locator): # These are used to deal with various Content-Encoding schemes. decoders = { 'deflate': zlib.decompress, - 'gzip': lambda b: gzip.GzipFile(fileobj=BytesIO(d)).read(), + 'gzip': lambda b: gzip.GzipFile(fileobj=BytesIO(b)).read(), 'none': lambda b: b, } @@ -633,7 +633,7 @@ def _prepare_threads(self): self._threads = [] for i in range(self.num_workers): t = threading.Thread(target=self._fetch) - t.setDaemon(True) + t.daemon = True t.start() self._threads.append(t) @@ -745,7 +745,7 @@ def _fetch(self): try: self._seen.add(link) if (not self._process_download(link) and - self._should_queue(link, url, rel)): + self._should_queue(link, url, rel)): logger.debug('Queueing %s from %s', link, url) self._to_fetch.put(link) except MetadataInvalidError: # e.g. invalid versions @@ -756,7 +756,7 @@ def _fetch(self): # always do this, to avoid hangs :-) self._to_fetch.task_done() if not url: - #logger.debug('Sentinel seen, quitting.') + # logger.debug('Sentinel seen, quitting.') break def get_page(self, url): @@ -832,6 +832,7 @@ def get_distribution_names(self): result.add(match.group(1)) return result + class DirectoryLocator(Locator): """ This class locates distributions in a directory tree. @@ -897,6 +898,7 @@ def get_distribution_names(self): break return result + class JSONLocator(Locator): """ This locator uses special extended metadata (not available on PyPI) and is @@ -935,6 +937,7 @@ def _get_project(self, name): result['urls'].setdefault(dist.version, set()).add(info['url']) return result + class DistPathLocator(Locator): """ This locator finds installed distributions in a path. It can be useful for @@ -1053,17 +1056,15 @@ def get_distribution_names(self): # We use a legacy scheme simply because most of the dists on PyPI use legacy -# versions which don't conform to PEP 426 / PEP 440. +# versions which don't conform to PEP 440. default_locator = AggregatingLocator( - JSONLocator(), + # JSONLocator(), # don't use as PEP 426 is withdrawn SimpleScrapingLocator('https://pypi.org/simple/', timeout=3.0), scheme='legacy') locate = default_locator.locate -NAME_VERSION_RE = re.compile(r'(?P[\w-]+)\s*' - r'\(\s*(==\s*)?(?P[^)]+)\)$') class DependencyFinder(object): """ @@ -1247,7 +1248,7 @@ def find(self, requirement, meta_extras=None, prereleases=False): if name not in self.dists_by_name: self.add_distribution(dist) else: - #import pdb; pdb.set_trace() + # import pdb; pdb.set_trace() other = self.dists_by_name[name] if other != dist: self.try_to_replace(dist, other, problems) diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/manifest.py b/src/fetchcode/vcs/pip/_vendor/distlib/manifest.py index ca0fe442..420dcf12 100644 --- a/src/fetchcode/vcs/pip/_vendor/distlib/manifest.py +++ b/src/fetchcode/vcs/pip/_vendor/distlib/manifest.py @@ -1,6 +1,6 @@ # -*- coding: utf-8 -*- # -# Copyright (C) 2012-2013 Python Software Foundation. +# Copyright (C) 2012-2023 Python Software Foundation. # See LICENSE.txt and CONTRIBUTORS.txt. # """ @@ -34,9 +34,11 @@ # _PYTHON_VERSION = sys.version_info[:2] + class Manifest(object): - """A list of files built by on exploring the filesystem and filtered by - applying various patterns to what we find there. + """ + A list of files built by exploring the filesystem and filtered by applying various + patterns to what we find there. """ def __init__(self, base=None): @@ -154,10 +156,7 @@ def process_directive(self, directive): elif action == 'exclude': for pattern in patterns: - found = self._exclude_pattern(pattern, anchor=True) - #if not found: - # logger.warning('no previously-included files ' - # 'found matching %r', pattern) + self._exclude_pattern(pattern, anchor=True) elif action == 'global-include': for pattern in patterns: @@ -167,11 +166,7 @@ def process_directive(self, directive): elif action == 'global-exclude': for pattern in patterns: - found = self._exclude_pattern(pattern, anchor=False) - #if not found: - # logger.warning('no previously-included files ' - # 'matching %r found anywhere in ' - # 'distribution', pattern) + self._exclude_pattern(pattern, anchor=False) elif action == 'recursive-include': for pattern in patterns: @@ -181,11 +176,7 @@ def process_directive(self, directive): elif action == 'recursive-exclude': for pattern in patterns: - found = self._exclude_pattern(pattern, prefix=thedir) - #if not found: - # logger.warning('no previously-included files ' - # 'matching %r found under directory %r', - # pattern, thedir) + self._exclude_pattern(pattern, prefix=thedir) elif action == 'graft': if not self._include_pattern(None, prefix=dirpattern): diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/markers.py b/src/fetchcode/vcs/pip/_vendor/distlib/markers.py index ee1f3e23..1514d460 100644 --- a/src/fetchcode/vcs/pip/_vendor/distlib/markers.py +++ b/src/fetchcode/vcs/pip/_vendor/distlib/markers.py @@ -1,6 +1,6 @@ # -*- coding: utf-8 -*- # -# Copyright (C) 2012-2017 Vinay Sajip. +# Copyright (C) 2012-2023 Vinay Sajip. # Licensed to the Python Software Foundation under a contributor agreement. # See LICENSE.txt and CONTRIBUTORS.txt. # @@ -13,23 +13,38 @@ # as ~= and === which aren't in Python, necessitating a different approach. import os +import re import sys import platform -import re -from .compat import python_implementation, urlparse, string_types +from .compat import string_types from .util import in_venv, parse_marker +from .version import LegacyVersion as LV __all__ = ['interpret'] +_VERSION_PATTERN = re.compile( + r'((\d+(\.\d+)*\w*)|\'(\d+(\.\d+)*\w*)\'|\"(\d+(\.\d+)*\w*)\")') +_VERSION_MARKERS = {'python_version', 'python_full_version'} + + +def _is_version_marker(s): + return isinstance(s, string_types) and s in _VERSION_MARKERS + + def _is_literal(o): if not isinstance(o, string_types) or not o: return False return o[0] in '\'"' + +def _get_versions(s): + return {LV(m.groups()[0]) for m in _VERSION_PATTERN.finditer(s)} + + class Evaluator(object): """ - This class is used to evaluate marker expessions. + This class is used to evaluate marker expressions. """ operations = { @@ -37,10 +52,10 @@ class Evaluator(object): '===': lambda x, y: x == y, '~=': lambda x, y: x == y or x > y, '!=': lambda x, y: x != y, - '<': lambda x, y: x < y, - '<=': lambda x, y: x == y or x < y, - '>': lambda x, y: x > y, - '>=': lambda x, y: x == y or x > y, + '<': lambda x, y: x < y, + '<=': lambda x, y: x == y or x < y, + '>': lambda x, y: x > y, + '>=': lambda x, y: x == y or x > y, 'and': lambda x, y: x and y, 'or': lambda x, y: x or y, 'in': lambda x, y: x in y, @@ -67,14 +82,27 @@ def evaluate(self, expr, context): elhs = expr['lhs'] erhs = expr['rhs'] if _is_literal(expr['lhs']) and _is_literal(expr['rhs']): - raise SyntaxError('invalid comparison: %s %s %s' % (elhs, op, erhs)) + raise SyntaxError('invalid comparison: %s %s %s' % + (elhs, op, erhs)) lhs = self.evaluate(elhs, context) rhs = self.evaluate(erhs, context) + if ((_is_version_marker(elhs) or _is_version_marker(erhs)) + and op in ('<', '<=', '>', '>=', '===', '==', '!=', '~=')): + lhs = LV(lhs) + rhs = LV(rhs) + elif _is_version_marker(elhs) and op in ('in', 'not in'): + lhs = LV(lhs) + rhs = _get_versions(rhs) result = self.operations[op](lhs, rhs) return result + +_DIGITS = re.compile(r'\d+\.\d+') + + def default_context(): + def format_full_version(info): version = '%s.%s.%s' % (info.major, info.minor, info.micro) kind = info.releaselevel @@ -83,12 +111,16 @@ def format_full_version(info): return version if hasattr(sys, 'implementation'): - implementation_version = format_full_version(sys.implementation.version) + implementation_version = format_full_version( + sys.implementation.version) implementation_name = sys.implementation.name else: implementation_version = '0' implementation_name = '' + ppv = platform.python_version() + m = _DIGITS.match(ppv) + pv = m.group(0) result = { 'implementation_name': implementation_name, 'implementation_version': implementation_version, @@ -99,17 +131,19 @@ def format_full_version(info): 'platform_system': platform.system(), 'platform_version': platform.version(), 'platform_in_venv': str(in_venv()), - 'python_full_version': platform.python_version(), - 'python_version': platform.python_version()[:3], + 'python_full_version': ppv, + 'python_version': pv, 'sys_platform': sys.platform, } return result + DEFAULT_CONTEXT = default_context() del default_context evaluator = Evaluator() + def interpret(marker, execution_context=None): """ Interpret a marker and return a result depending on environment. @@ -122,9 +156,11 @@ def interpret(marker, execution_context=None): try: expr, rest = parse_marker(marker) except Exception as e: - raise SyntaxError('Unable to interpret marker syntax: %s: %s' % (marker, e)) + raise SyntaxError('Unable to interpret marker syntax: %s: %s' % + (marker, e)) if rest and rest[0] != '#': - raise SyntaxError('unexpected trailing data in marker: %s: %s' % (marker, rest)) + raise SyntaxError('unexpected trailing data in marker: %s: %s' % + (marker, rest)) context = dict(DEFAULT_CONTEXT) if execution_context: context.update(execution_context) diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/metadata.py b/src/fetchcode/vcs/pip/_vendor/distlib/metadata.py index 2d61378e..7189aeef 100644 --- a/src/fetchcode/vcs/pip/_vendor/distlib/metadata.py +++ b/src/fetchcode/vcs/pip/_vendor/distlib/metadata.py @@ -5,7 +5,7 @@ # """Implementation of the Metadata for Python packages PEPs. -Supports all metadata formats (1.0, 1.1, 1.2, and 2.0 experimental). +Supports all metadata formats (1.0, 1.1, 1.2, 1.3/2.1 and 2.2). """ from __future__ import unicode_literals @@ -94,17 +94,23 @@ class MetadataInvalidError(DistlibException): # See issue #106: Sometimes 'Requires' and 'Provides' occur wrongly in # the metadata. Include them in the tuple literal below to allow them # (for now). +# Ditto for Obsoletes - see issue #140. _566_FIELDS = _426_FIELDS + ('Description-Content-Type', - 'Requires', 'Provides') + 'Requires', 'Provides', 'Obsoletes') _566_MARKERS = ('Description-Content-Type',) +_643_MARKERS = ('Dynamic', 'License-File') + +_643_FIELDS = _566_FIELDS + _643_MARKERS + _ALL_FIELDS = set() _ALL_FIELDS.update(_241_FIELDS) _ALL_FIELDS.update(_314_FIELDS) _ALL_FIELDS.update(_345_FIELDS) _ALL_FIELDS.update(_426_FIELDS) _ALL_FIELDS.update(_566_FIELDS) +_ALL_FIELDS.update(_643_FIELDS) EXTRA_RE = re.compile(r'''extra\s*==\s*("([^"]+)"|'([^']+)')''') @@ -117,27 +123,23 @@ def _version2fieldlist(version): elif version == '1.2': return _345_FIELDS elif version in ('1.3', '2.1'): - return _345_FIELDS + _566_FIELDS + # avoid adding field names if already there + return _345_FIELDS + tuple(f for f in _566_FIELDS if f not in _345_FIELDS) elif version == '2.0': - return _426_FIELDS + raise ValueError('Metadata 2.0 is withdrawn and not supported') + # return _426_FIELDS + elif version == '2.2': + return _643_FIELDS raise MetadataUnrecognizedVersionError(version) def _best_version(fields): """Detect the best version depending on the fields used.""" def _has_marker(keys, markers): - for marker in markers: - if marker in keys: - return True - return False - - keys = [] - for key, value in fields.items(): - if value in ([], 'UNKNOWN', None): - continue - keys.append(key) + return any(marker in keys for marker in markers) - possible_versions = ['1.0', '1.1', '1.2', '1.3', '2.0', '2.1'] + keys = [key for key, value in fields.items() if value not in ([], 'UNKNOWN', None)] + possible_versions = ['1.0', '1.1', '1.2', '1.3', '2.1', '2.2'] # 2.0 removed # first let's try to see if a field is not part of one of the version for key in keys: @@ -157,9 +159,12 @@ def _has_marker(keys, markers): if key != 'Description': # In 2.1, description allowed after headers possible_versions.remove('2.1') logger.debug('Removed 2.1 due to %s', key) - if key not in _426_FIELDS and '2.0' in possible_versions: - possible_versions.remove('2.0') - logger.debug('Removed 2.0 due to %s', key) + if key not in _643_FIELDS and '2.2' in possible_versions: + possible_versions.remove('2.2') + logger.debug('Removed 2.2 due to %s', key) + # if key not in _426_FIELDS and '2.0' in possible_versions: + # possible_versions.remove('2.0') + # logger.debug('Removed 2.0 due to %s', key) # possible_version contains qualified versions if len(possible_versions) == 1: @@ -172,16 +177,18 @@ def _has_marker(keys, markers): is_1_1 = '1.1' in possible_versions and _has_marker(keys, _314_MARKERS) is_1_2 = '1.2' in possible_versions and _has_marker(keys, _345_MARKERS) is_2_1 = '2.1' in possible_versions and _has_marker(keys, _566_MARKERS) - is_2_0 = '2.0' in possible_versions and _has_marker(keys, _426_MARKERS) - if int(is_1_1) + int(is_1_2) + int(is_2_1) + int(is_2_0) > 1: - raise MetadataConflictError('You used incompatible 1.1/1.2/2.0/2.1 fields') + # is_2_0 = '2.0' in possible_versions and _has_marker(keys, _426_MARKERS) + is_2_2 = '2.2' in possible_versions and _has_marker(keys, _643_MARKERS) + if int(is_1_1) + int(is_1_2) + int(is_2_1) + int(is_2_2) > 1: + raise MetadataConflictError('You used incompatible 1.1/1.2/2.1/2.2 fields') - # we have the choice, 1.0, or 1.2, or 2.0 + # we have the choice, 1.0, or 1.2, 2.1 or 2.2 # - 1.0 has a broken Summary field but works with all tools # - 1.1 is to avoid # - 1.2 fixes Summary but has little adoption - # - 2.0 adds more features and is very new - if not is_1_1 and not is_1_2 and not is_2_1 and not is_2_0: + # - 2.1 adds more features + # - 2.2 is the latest + if not is_1_1 and not is_1_2 and not is_2_1 and not is_2_2: # we couldn't find any specific marker if PKG_INFO_PREFERRED_VERSION in possible_versions: return PKG_INFO_PREFERRED_VERSION @@ -191,41 +198,17 @@ def _has_marker(keys, markers): return '1.2' if is_2_1: return '2.1' + # if is_2_2: + # return '2.2' - return '2.0' + return '2.2' +# This follows the rules about transforming keys as described in +# https://www.python.org/dev/peps/pep-0566/#id17 _ATTR2FIELD = { - 'metadata_version': 'Metadata-Version', - 'name': 'Name', - 'version': 'Version', - 'platform': 'Platform', - 'supported_platform': 'Supported-Platform', - 'summary': 'Summary', - 'description': 'Description', - 'keywords': 'Keywords', - 'home_page': 'Home-page', - 'author': 'Author', - 'author_email': 'Author-email', - 'maintainer': 'Maintainer', - 'maintainer_email': 'Maintainer-email', - 'license': 'License', - 'classifier': 'Classifier', - 'download_url': 'Download-URL', - 'obsoletes_dist': 'Obsoletes-Dist', - 'provides_dist': 'Provides-Dist', - 'requires_dist': 'Requires-Dist', - 'setup_requires_dist': 'Setup-Requires-Dist', - 'requires_python': 'Requires-Python', - 'requires_external': 'Requires-External', - 'requires': 'Requires', - 'provides': 'Provides', - 'obsoletes': 'Obsoletes', - 'project_url': 'Project-URL', - 'private_version': 'Private-Version', - 'obsoleted_by': 'Obsoleted-By', - 'extension': 'Extension', - 'provides_extra': 'Provides-Extra', + name.lower().replace("-", "_"): name for name in _ALL_FIELDS } +_FIELD2ATTR = {field: attr for attr, field in _ATTR2FIELD.items()} _PREDICATE_FIELDS = ('Requires-Dist', 'Obsoletes-Dist', 'Provides-Dist') _VERSIONS_FIELDS = ('Requires-Python',) @@ -234,7 +217,7 @@ def _has_marker(keys, markers): 'Requires', 'Provides', 'Obsoletes-Dist', 'Provides-Dist', 'Requires-Dist', 'Requires-External', 'Project-URL', 'Supported-Platform', 'Setup-Requires-Dist', - 'Provides-Extra', 'Extension') + 'Provides-Extra', 'Extension', 'License-File') _LISTTUPLEFIELDS = ('Project-URL',) _ELEMENTSFIELD = ('Keywords',) @@ -262,7 +245,7 @@ def _get_name_and_version(name, version, for_filename=False): class LegacyMetadata(object): """The legacy metadata of a release. - Supports versions 1.0, 1.1 and 1.2 (auto-detected). You can + Supports versions 1.0, 1.1, 1.2, 2.0 and 1.3/2.1 (auto-detected). You can instantiate the class with one of these arguments (or none): - *path*, the path to a metadata file - *fileobj* give a file-like object with metadata as content @@ -381,6 +364,11 @@ def read_file(self, fileob): value = msg[field] if value is not None and value != 'UNKNOWN': self.set(field, value) + + # PEP 566 specifies that the body be used for the description, if + # available + body = msg.get_payload() + self["Description"] = body if body else self["Description"] # logger.debug('Attempting to set metadata for %s', self) # self.set_metadata_version() @@ -567,57 +555,21 @@ def todict(self, skip_missing=False): Field names will be converted to use the underscore-lowercase style instead of hyphen-mixed case (i.e. home_page instead of Home-page). + This is as per https://www.python.org/dev/peps/pep-0566/#id17. """ self.set_metadata_version() - mapping_1_0 = ( - ('metadata_version', 'Metadata-Version'), - ('name', 'Name'), - ('version', 'Version'), - ('summary', 'Summary'), - ('home_page', 'Home-page'), - ('author', 'Author'), - ('author_email', 'Author-email'), - ('license', 'License'), - ('description', 'Description'), - ('keywords', 'Keywords'), - ('platform', 'Platform'), - ('classifiers', 'Classifier'), - ('download_url', 'Download-URL'), - ) + fields = _version2fieldlist(self['Metadata-Version']) data = {} - for key, field_name in mapping_1_0: + + for field_name in fields: if not skip_missing or field_name in self._fields: - data[key] = self[field_name] - - if self['Metadata-Version'] == '1.2': - mapping_1_2 = ( - ('requires_dist', 'Requires-Dist'), - ('requires_python', 'Requires-Python'), - ('requires_external', 'Requires-External'), - ('provides_dist', 'Provides-Dist'), - ('obsoletes_dist', 'Obsoletes-Dist'), - ('project_url', 'Project-URL'), - ('maintainer', 'Maintainer'), - ('maintainer_email', 'Maintainer-email'), - ) - for key, field_name in mapping_1_2: - if not skip_missing or field_name in self._fields: - if key != 'project_url': - data[key] = self[field_name] - else: - data[key] = [','.join(u) for u in self[field_name]] - - elif self['Metadata-Version'] == '1.1': - mapping_1_1 = ( - ('provides', 'Provides'), - ('requires', 'Requires'), - ('obsoletes', 'Obsoletes'), - ) - for key, field_name in mapping_1_1: - if not skip_missing or field_name in self._fields: + key = _FIELD2ATTR[field_name] + if key != 'project_url': data[key] = self[field_name] + else: + data[key] = [','.join(u) for u in self[field_name]] return data @@ -657,7 +609,7 @@ def __repr__(self): class Metadata(object): """ - The metadata of a release. This implementation uses 2.0 (JSON) + The metadata of a release. This implementation uses 2.1 metadata where possible. If not possible, it wraps a LegacyMetadata instance which handles the key-value metadata format. """ @@ -666,6 +618,8 @@ class Metadata(object): NAME_MATCHER = re.compile('^[0-9A-Z]([0-9A-Z_.-]*[0-9A-Z])?$', re.I) + FIELDNAME_MATCHER = re.compile('^[A-Z]([0-9A-Z-]*[0-9A-Z])?$', re.I) + VERSION_MATCHER = PEP440_VERSION_RE SUMMARY_MATCHER = re.compile('.{1,2047}') @@ -693,6 +647,7 @@ class Metadata(object): 'name': (NAME_MATCHER, ('legacy',)), 'version': (VERSION_MATCHER, ('legacy',)), 'summary': (SUMMARY_MATCHER, ('legacy',)), + 'dynamic': (FIELDNAME_MATCHER, ('legacy',)), } __slots__ = ('_legacy', '_data', 'scheme') @@ -1003,10 +958,14 @@ def _from_legacy(self): LEGACY_MAPPING = { 'name': 'Name', 'version': 'Version', - 'license': 'License', + ('extensions', 'python.details', 'license'): 'License', 'summary': 'Summary', 'description': 'Description', - 'classifiers': 'Classifier', + ('extensions', 'python.project', 'project_urls', 'Home'): 'Home-page', + ('extensions', 'python.project', 'contacts', 0, 'name'): 'Author', + ('extensions', 'python.project', 'contacts', 0, 'email'): 'Author-email', + 'source_url': 'Download-URL', + ('extensions', 'python.details', 'classifiers'): 'Classifier', } def _to_legacy(self): @@ -1034,16 +993,29 @@ def process_entries(entries): assert self._data and not self._legacy result = LegacyMetadata() nmd = self._data + # import pdb; pdb.set_trace() for nk, ok in self.LEGACY_MAPPING.items(): - if nk in nmd: - result[ok] = nmd[nk] + if not isinstance(nk, tuple): + if nk in nmd: + result[ok] = nmd[nk] + else: + d = nmd + found = True + for k in nk: + try: + d = d[k] + except (KeyError, IndexError): + found = False + break + if found: + result[ok] = d r1 = process_entries(self.run_requires + self.meta_requires) r2 = process_entries(self.build_requires + self.dev_requires) if self.extras: result['Provides-Extra'] = sorted(self.extras) result['Requires-Dist'] = sorted(r1) result['Setup-Requires-Dist'] = sorted(r2) - # TODO: other fields such as contacts + # TODO: any other fields wanted return result def write(self, path=None, fileobj=None, legacy=False, skip_unknown=True): diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/resources.py b/src/fetchcode/vcs/pip/_vendor/distlib/resources.py index 18840167..fef52aa1 100644 --- a/src/fetchcode/vcs/pip/_vendor/distlib/resources.py +++ b/src/fetchcode/vcs/pip/_vendor/distlib/resources.py @@ -11,13 +11,12 @@ import logging import os import pkgutil -import shutil import sys import types import zipimport from . import DistlibException -from .util import cached_property, get_cache_base, path_to_cache_dir, Cache +from .util import cached_property, get_cache_base, Cache logger = logging.getLogger(__name__) @@ -283,6 +282,7 @@ def _is_directory(self, path): result = False return result + _finder_registry = { type(None): ResourceFinder, zipimport.zipimporter: ZipResourceFinder @@ -296,6 +296,8 @@ def _is_directory(self, path): import _frozen_importlib as _fi _finder_registry[_fi.SourceFileLoader] = ResourceFinder _finder_registry[_fi.FileFinder] = ResourceFinder + # See issue #146 + _finder_registry[_fi.SourcelessFileLoader] = ResourceFinder del _fi except (ImportError, AttributeError): pass @@ -304,6 +306,7 @@ def _is_directory(self, path): def register_finder(loader, finder_maker): _finder_registry[type(loader)] = finder_maker + _finder_cache = {} diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/scripts.py b/src/fetchcode/vcs/pip/_vendor/distlib/scripts.py index 51859741..e16292b8 100644 --- a/src/fetchcode/vcs/pip/_vendor/distlib/scripts.py +++ b/src/fetchcode/vcs/pip/_vendor/distlib/scripts.py @@ -1,6 +1,6 @@ # -*- coding: utf-8 -*- # -# Copyright (C) 2013-2015 Vinay Sajip. +# Copyright (C) 2013-2023 Vinay Sajip. # Licensed to the Python Software Foundation under a contributor agreement. # See LICENSE.txt and CONTRIBUTORS.txt. # @@ -10,11 +10,13 @@ import re import struct import sys +import time +from zipfile import ZipInfo from .compat import sysconfig, detect_encoding, ZipFile from .resources import finder from .util import (FileOperator, get_export_entry, convert_path, - get_executable, in_venv) + get_executable, get_platform, in_venv) logger = logging.getLogger(__name__) @@ -47,8 +49,26 @@ sys.exit(%(func)s()) ''' +# Pre-fetch the contents of all executable wrapper stubs. +# This is to address https://github.com/pypa/pip/issues/12666. +# When updating pip, we rename the old pip in place before installing the +# new version. If we try to fetch a wrapper *after* that rename, the finder +# machinery will be confused as the package is no longer available at the +# location where it was imported from. So we load everything into memory in +# advance. -def _enquote_executable(executable): +# Issue 31: don't hardcode an absolute package name, but +# determine it relative to the current package +distlib_package = __name__.rsplit('.', 1)[0] + +WRAPPERS = { + r.name: r.bytes + for r in finder(distlib_package).iterator("") + if r.name.endswith(".exe") +} + + +def enquote_executable(executable): if ' ' in executable: # make sure we quote only the executable in case of env # for example /usr/bin/env "/dir with spaces/bin/jython" @@ -64,6 +84,10 @@ def _enquote_executable(executable): return executable +# Keep the old name around (for now), as there is at least one project using it! +_enquote_executable = enquote_executable + + class ScriptMaker(object): """ A class to copy or create scripts from source scripts or callable @@ -73,21 +97,26 @@ class ScriptMaker(object): executable = None # for shebangs - def __init__(self, source_dir, target_dir, add_launchers=True, - dry_run=False, fileop=None): + def __init__(self, + source_dir, + target_dir, + add_launchers=True, + dry_run=False, + fileop=None): self.source_dir = source_dir self.target_dir = target_dir self.add_launchers = add_launchers self.force = False self.clobber = False # It only makes sense to set mode bits on POSIX. - self.set_mode = (os.name == 'posix') or (os.name == 'java' and - os._name == 'posix') + self.set_mode = (os.name == 'posix') or (os.name == 'java' + and os._name == 'posix') self.variants = set(('', 'X.Y')) self._fileop = fileop or FileOperator(dry_run) - self._is_nt = os.name == 'nt' or ( - os.name == 'java' and os._name == 'nt') + self._is_nt = os.name == 'nt' or (os.name == 'java' + and os._name == 'nt') + self.version_info = sys.version_info def _get_alternate_executable(self, executable, options): if options.get('gui', False) and self._is_nt: # pragma: no cover @@ -97,6 +126,7 @@ def _get_alternate_executable(self, executable, options): return executable if sys.platform.startswith('java'): # pragma: no cover + def _is_shell(self, executable): """ Determine if the specified executable is a script @@ -141,8 +171,8 @@ def _build_shebang(self, executable, post_interp): max_shebang_length = 512 else: max_shebang_length = 127 - simple_shebang = ((b' ' not in executable) and - (shebang_length <= max_shebang_length)) + simple_shebang = ((b' ' not in executable) + and (shebang_length <= max_shebang_length)) if simple_shebang: result = b'#!' + executable + post_interp + b'\n' @@ -156,17 +186,25 @@ def _get_shebang(self, encoding, post_interp=b'', options=None): enquote = True if self.executable: executable = self.executable - enquote = False # assume this will be taken care of + enquote = False # assume this will be taken care of elif not sysconfig.is_python_build(): executable = get_executable() elif in_venv(): # pragma: no cover - executable = os.path.join(sysconfig.get_path('scripts'), - 'python%s' % sysconfig.get_config_var('EXE')) - else: # pragma: no cover executable = os.path.join( - sysconfig.get_config_var('BINDIR'), - 'python%s%s' % (sysconfig.get_config_var('VERSION'), - sysconfig.get_config_var('EXE'))) + sysconfig.get_path('scripts'), + 'python%s' % sysconfig.get_config_var('EXE')) + else: # pragma: no cover + if os.name == 'nt': + # for Python builds from source on Windows, no Python executables with + # a version suffix are created, so we use python.exe + executable = os.path.join( + sysconfig.get_config_var('BINDIR'), + 'python%s' % (sysconfig.get_config_var('EXE'))) + else: + executable = os.path.join( + sysconfig.get_config_var('BINDIR'), + 'python%s%s' % (sysconfig.get_config_var('VERSION'), + sysconfig.get_config_var('EXE'))) if options: executable = self._get_alternate_executable(executable, options) @@ -185,13 +223,13 @@ def _get_shebang(self, encoding, post_interp=b'', options=None): # If the user didn't specify an executable, it may be necessary to # cater for executable paths with spaces (not uncommon on Windows) if enquote: - executable = _enquote_executable(executable) + executable = enquote_executable(executable) # Issue #51: don't use fsencode, since we later try to # check that the shebang is decodable using utf-8. executable = executable.encode('utf-8') # in case of IronPython, play safe and enable frames support if (sys.platform == 'cli' and '-X:Frames' not in post_interp - and '-X:FullFrames' not in post_interp): # pragma: no cover + and '-X:FullFrames' not in post_interp): # pragma: no cover post_interp += b' -X:Frames' shebang = self._build_shebang(executable, post_interp) # Python parser starts to read a script using UTF-8 until @@ -202,8 +240,8 @@ def _get_shebang(self, encoding, post_interp=b'', options=None): try: shebang.decode('utf-8') except UnicodeDecodeError: # pragma: no cover - raise ValueError( - 'The shebang (%r) is not decodable from utf-8' % shebang) + raise ValueError('The shebang (%r) is not decodable from utf-8' % + shebang) # If the script is encoded to a custom encoding (use a # #coding:xxx cookie), the shebang has to be decodable from # the script encoding too. @@ -211,15 +249,16 @@ def _get_shebang(self, encoding, post_interp=b'', options=None): try: shebang.decode(encoding) except UnicodeDecodeError: # pragma: no cover - raise ValueError( - 'The shebang (%r) is not decodable ' - 'from the script encoding (%r)' % (shebang, encoding)) + raise ValueError('The shebang (%r) is not decodable ' + 'from the script encoding (%r)' % + (shebang, encoding)) return shebang def _get_script_text(self, entry): - return self.script_template % dict(module=entry.prefix, - import_name=entry.suffix.split('.')[0], - func=entry.suffix) + return self.script_template % dict( + module=entry.prefix, + import_name=entry.suffix.split('.')[0], + func=entry.suffix) manifest = _DEFAULT_MANIFEST @@ -241,7 +280,14 @@ def _write_script(self, names, shebang, script_bytes, filenames, ext): launcher = self._get_launcher('w') stream = BytesIO() with ZipFile(stream, 'w') as zf: - zf.writestr('__main__.py', script_bytes) + source_date_epoch = os.environ.get('SOURCE_DATE_EPOCH') + if source_date_epoch: + date_time = time.gmtime(int(source_date_epoch))[:6] + zinfo = ZipInfo(filename='__main__.py', + date_time=date_time) + zf.writestr(zinfo, script_bytes) + else: + zf.writestr('__main__.py', script_bytes) zip_data = stream.getvalue() script_bytes = launcher + shebang + zip_data for name in names: @@ -259,7 +305,7 @@ def _write_script(self, names, shebang, script_bytes, filenames, ext): 'use .deleteme logic') dfname = '%s.deleteme' % outname if os.path.exists(dfname): - os.remove(dfname) # Not allowed to fail here + os.remove(dfname) # Not allowed to fail here os.rename(outname, dfname) # nor here self._fileop.write_binary_file(outname, script_bytes) logger.debug('Able to replace executable using ' @@ -267,9 +313,10 @@ def _write_script(self, names, shebang, script_bytes, filenames, ext): try: os.remove(dfname) except Exception: - pass # still in use - ignore error + pass # still in use - ignore error else: - if self._is_nt and not outname.endswith('.' + ext): # pragma: no cover + if self._is_nt and not outname.endswith( + '.' + ext): # pragma: no cover outname = '%s.%s' % (outname, ext) if os.path.exists(outname) and not self.clobber: logger.warning('Skipping existing file %s', outname) @@ -279,6 +326,20 @@ def _write_script(self, names, shebang, script_bytes, filenames, ext): self._fileop.set_executable_mode([outname]) filenames.append(outname) + variant_separator = '-' + + def get_script_filenames(self, name): + result = set() + if '' in self.variants: + result.add(name) + if 'X' in self.variants: + result.add('%s%s' % (name, self.version_info[0])) + if 'X.Y' in self.variants: + result.add('%s%s%s.%s' % + (name, self.variant_separator, self.version_info[0], + self.version_info[1])) + return result + def _make_script(self, entry, filenames, options=None): post_interp = b'' if options: @@ -288,15 +349,7 @@ def _make_script(self, entry, filenames, options=None): post_interp = args.encode('utf-8') shebang = self._get_shebang('utf-8', post_interp, options=options) script = self._get_script_text(entry).encode('utf-8') - name = entry.name - scriptnames = set() - if '' in self.variants: - scriptnames.add(name) - if 'X' in self.variants: - scriptnames.add('%s%s' % (name, sys.version_info[0])) - if 'X.Y' in self.variants: - scriptnames.add('%s-%s.%s' % (name, sys.version_info[0], - sys.version_info[1])) + scriptnames = self.get_script_filenames(entry.name) if options and options.get('gui', False): ext = 'pyw' else: @@ -323,8 +376,7 @@ def _copy_script(self, script, filenames): else: first_line = f.readline() if not first_line: # pragma: no cover - logger.warning('%s: %s is an empty file (skipping)', - self.get_command_name(), script) + logger.warning('%s is an empty file (skipping)', script) return match = FIRST_LINE_RE.match(first_line.replace(b'\r\n', b'\n')) @@ -363,25 +415,23 @@ def dry_run(self): def dry_run(self, value): self._fileop.dry_run = value - if os.name == 'nt' or (os.name == 'java' and os._name == 'nt'): # pragma: no cover + if os.name == 'nt' or (os.name == 'java' + and os._name == 'nt'): # pragma: no cover # Executable launcher support. # Launchers are from https://bitbucket.org/vinay.sajip/simple_launcher/ def _get_launcher(self, kind): - if struct.calcsize('P') == 8: # 64-bit + if struct.calcsize('P') == 8: # 64-bit bits = '64' else: bits = '32' - name = '%s%s.exe' % (kind, bits) - # Issue 31: don't hardcode an absolute package name, but - # determine it relative to the current package - distlib_package = __name__.rsplit('.', 1)[0] - resource = finder(distlib_package).find(name) - if not resource: - msg = ('Unable to find resource %s in package %s' % (name, - distlib_package)) + platform_suffix = '-arm' if get_platform() == 'win-arm64' else '' + name = '%s%s%s.exe' % (kind, bits, platform_suffix) + if name not in WRAPPERS: + msg = ('Unable to find resource %s in package %s' % + (name, distlib_package)) raise ValueError(msg) - return resource.bytes + return WRAPPERS[name] # Public API follows diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/t32.exe b/src/fetchcode/vcs/pip/_vendor/distlib/t32.exe index 8932a18e4596952373a38c60b81b7116d4ef9ee8..52154f0be32cc2bdbf98af131d477900667d0abd 100644 GIT binary patch delta 24794 zcmeIaeL&R37C-#mFQAAZsDOx|h^Q#2%f7M8?y_rwq8DrxT_6QT1=rOSao4P?t-4U+ zN~vRM-Sldh*{hoQQmJ8>XsOwoQfg7xYh!o|%aoGb-{*XH#kw!g^ZV!d?^&IlGiT1c zojG&n%$Zr1-w0aX9<2>}%Q;H|fnQ5as1^>sAEeN@?AVCcKVa8otvmnZPbd`k& zLK9HW`_NbQELJvLRMfF?vO(SEMal#NWlWa|H}co@?-^Pz6ApF}gss;FA#Pp&KGB=P zWrEpP)c4&g2vcS2`p5Pd?kHR0050}Cz9JBs*l)7Fni;78K_m=ExQOtFk694R!)KK` zN`ae|j;{onuV@GxFUD$a z2d@uQ^!rzmAh?f3)>gOQ@_H>cL&!*h>v{lJZ?^C?-PnY{o}wX)l?Q72eOq#u^fr+$ zA<~b!`$^ZwNTi>Hva^AG#0fpwzXA=)X8lOPQnReRW?3uOh6#f5q`onH!%^E-6rI5P zdA(k{Sm1TG^M}w2ieHImQ-cQd80Smc0?}i;ET8|q7h4xJLXiTV)nwoBWb+^}>~>L2 zcALI&SwPLQHdN9^<-Z#s)nm2Gyv`duvlsg*D04(QT6P~x_2gX+$ZNI>L-M%SiT~~H zuLQo#SKwW0seqdLIGRoCl-VuShnG;*(d?B@=_xNjbUfLuMbk=7Hca+hb6*R!+J}3c z!OEL8&NlvLS3xK$s;+m$TWxJDtn-keFH2S2@JJn-(^KM|{*H5OZFZaG-H5Iu1sZuv z0+CPe+}9P=nWhOb zzo4PIoIedsazCQ0LT zUyHC7@co!_3*QR?L(!LL{_7JgypGbXHIR&_B@`rlEl}WkO}4Am?e#h$^moFmLLdwJ zw(u$uvHK1oZ-rO=f(4{+4X^qaVj55cB)O`YkjKNTNLzT+W-u_{e$#u-{)?vx-#5Sv z?>Pu5PnNnf6avBH?{+2*DUUh_Wr~XQw;iq^UGWWPXXOQ{22XaozReM&%f8_Zu?zNJ zyyx6y?U25ZB>aQ1?Bq~nkPFr z-&&yb78KM}^IyQnC<7XkmiSXgZ7~q=Iug$byytITZ&6Wg^)(P#>=j2yKCSjHUNvQV zqf)7h^)B@BFi})EgCK(~qO#tWiaIR(^IyH*6=5Xbf}CW)Uh#**B;vMQiS9Mf*|Tb^ zpF<@D79Kg1))N#PrLh=PpuAOm0tum$AcR*}cg&eIWv;t(R^1E|@45V%>dOH5iFU8I zF)(SK)Wl!Vgl8WSSotnw*w#pB70}*&68STP-7R7J60uZ?`&hz?TB&77RL$nkP`+kg z6Ma`-#J7zbM+z=~M+*v|a|#N)`-r6l-lzz4Ti@t3dZV%c6}%+`3c`fir&^Iaae}45 zMjG84m4Lk3YC>9kC2dXULjg}~Bk_*#s;P;)437h`ci$Nhd#cakn@6G3t|I#?8Y+cp6;fzQSkR{PhO@B^LQ$70H%dV1P}E&M)gQIcl9D_8lpi&CrF zTdXuBGPb8b-}zm++a!2yIeLd5=tKp7V^4;}%>D)pmWl@}>&t68xsQtld^)B!|5lSN zGHj&K$7>e~SA5roQHMWOZ{hVfC6<@F^@+|wDB{jc%n#pCZ_5F#g+Gs6Up>b{QI)$v zbT^9jaNAnet!v_l-4acyR6{6wL1P$1T`}G5M2u9pr5hoKvT0rC%idvUy7uo*JFV3Y zz9)ls^>>m1IV$wDD(j(Nex?etq%!}w!S;tHDPHe(mpK?rxvsDCJ88rN zY!#rws4e_&H>h&EIL+7RWh^u-HhSybT$_~Z%gSShFnQ5uQuPCYSJvyQn;=s*A>7Z} zaA57>Q}PRJYr!b~9hQRoSWV#$Fb@qLFBXl};|#98>Fmk3C!$l)B{g#XTNpbOo+LUF*|*_ot_td9H8Pd;AXR#? zyFFEpUPbZ2K>GW<9}inNTaQ z97F@Wj6MbdavzIp%x=E}hH~%@yVPYq8TB_9sK2wQ5?_B|WeAR8Ktg)eNea>(^B{7V z)fg*(8a(r~`gEU#kM7UTMD*?3ltxJK1jCHcWw$$e^2wmXS_$E~{aHx&1o5horFI{b zxJi#5b@Gf3)CVOej}#oyp3%Xcu!JEnnsO@8jay3Q=sr5^*@TX9|Jao^cCYNQ#e^;5 zKr77xw26g3VPb#jkqL=e-6J+^cfStupss9Bj}gO34+Z_Q(u63 zpgb-(0zD0|>R?mf8)XG8thQ>ebC5l>L>F{Oiw=h6b6FOOVDvAwbmkvmt}%ZNQsilz zR?-rg&zsRm5#I$|qY#Qll=Y!giW?3J2_uDuZPN=_eb1PY(+gk%23q)}(bRrj))+Dq zoqg?r)*3bCas8h%g6y${B@N0?-l9rcg0^LiVZ9>SEpl13yFS9g`B)N# z&=om8dXQv?l5K_}tUYpgx8rE5sJOw0TEetZ{d;Zv6ABdF-P=7RWN$$If1>QCa+EZp|clX3i?K2D9zMD zfF9xM>wtB6!Hc>}dX`V=!+wh%s0gEi$Mg-3xJKdBDJn3uipJ;VX$Zog4sP^>c>26HIQ4FuA%%?EUW3&H+l&6tdstga4iB$ zyf&|uPiBXEO%yF*tb2?+tVe8zxXlh?<6{Pin|kf0vYu_}lOkKn zKIqdY6;g>65yi`)qWvRy2jT*q2bnZS+>_n9%fnen%0Th$DJ(4|T0AnIjZL}WdjA11 z-ILcoO?i3>-*mlW4y|7G<>92m)S=*E3q!*fv7mSZ zP2Q;Zbn&MiY-zkMp5_qym?z-ql}j zDLqKEz<-F0#Bdo9{!UoYDs3{!U=)wL!K(WXP)$U(G>N5T8gyt7Ye9AWk|KSOfW{Y1 z@SHgMYY?`8EB$H>wdkF;QPr1DQl8E=ah;U)Q(B^V9+q08Ksyhg43MpEjE|BE=X;JG zZ42^TmkL*-@S#SHzpy*B6ReJwDBcKj%U}2-fGpxf%$U6Z+zA5T)R8C23c3~S>sDnf zf6`Cm=%C>zp(a{B_jNdw`nupmuW_x!na4nc#YAV~AkQsy2mB0Bp+_Ty#_Y?O7pyp4 zNw8CG58-aOoM=^1y0>jByg9X_Lm@5Cv9l*qlc-2-mX;8tcxkI0gUTC}7f?S+XX`PB zgp*NoiSTH1r`jwT(JldPaf3a^7SWxw11hD}r}Wk=yJ+vqd!Z2RL%}2_rRWZD?-Ne) z=QO#!=I~lQ&>q!csx!r{1O-kO+n*3gr0kr zMZk6n#aTQ~|1)&k)_Y{X!02P4{iey%*}Q&nGHC98^03R;eErqv+Sq~z64;e~5ffkR zz+Ux`@NGoe>x}a)T>b6ELp|D2@SQAOWfu54Y4OLvHR2r0=Y!C#O>`FXS%7M^Et5Pk z1w_bl65_jR{kenra8{8J9o7r!Mj??}n0Je5r}SrUCd7&#b!BH0`h*MUt_=HXS%gqz zdBs30u7~xMHtsUvbHVxBNOCqol?xsJjRU;7oA+EOn_M+SUu_r7z?CuTOKNakv3eXwgKUo%@axmq3#+~ zAaQ?+tsQRX3q+}85Nhd5qUXiD*CE3Dl1nL_X$9g)IyA4Z1jLz<` zI&lhmni&(5!V0PX%R?Zgo%!=@eqv0wUX+l`-v@ENVH*;M4gRnM>!w0x7p?q$F6rHx z&ai(QVF?_Crtq0ld_Uuf>Ym+Dya9*tf&%WQV@GG+naKzA={6CBzG`;@&58ynWIfoL z0e#DU#(C7ws~qHXZEVT9IXcws&|0}sK}>1mPhkvJo{yu3g=?{NV8(+5LGF*dEj>J= z;~>UxZvA)#Ju8ocgRDGMAe&E`pCvTDFM?TkQdG*r_gXshvrmJDp((Q?Oe%ua|4&df zKz)DD?oUeUclP^^8KT1hHqdey9Yy5DvTVazdfnaEr3c0?Lmw5C1x7pKtb7o`P!x2Q z#}tM5>~PG7sDuM#C7mmelky6bl$K9}6&=TxAXIxD*s+hm2rD854`+@vjtJW(8plcC zJ;lN9+jpFO_+v=8v!Ph$0`MtaT9H){?dl8At?>f13v}x=q61l02N{0`oNW2T(||tI z8c3C|t%?aeL4x{fRiD70Izj6#I2cJUOrEt0r`zBV=Kf^gh9J0MY0n4H$(1P z?bipP->^dMk+y@>-;1b8c~hr0L%N+^t=PY@UvFpiNqMf7Fds-s+T(aJw%H;+ILVKQ z;JFf_dgdIG4D#5pEl;uDW^nvUt_g5qZ^9@y$cO^$B@)v(9=~9z-(K3PWnTlG#m@@9kz6gctUUpd^`f#EjfiM3_&Xe8DPvfZBz%QZ4 zSD(7$f(X!|bH%k}n2>dX9Hv27`v`nTwb6I{4fKWVf)Uqd2iHYXy1E%>@cxZK7SPJ@ZyV*-1tQfL z)CN{|frZr=)XrWRlr%mESTHVd95E2aSrs^MI|RRyH)=URw~+`0%%Y4pj|Gj53jfI@ zyJwmnclzK*W#WO^tT{Q&Z1wRSa?#a&JY>a(0-Sa&BjHhL zLX%dW{>SVH$`*x|M0;o3H7PshTQ*_Hz@$bf2!8`{P_Zp>w&lRk0UshmAK zL_MO3MDz{hDKF68eHnu|Jq&}urb|~7&3fs?7hd}TRurA=?hF~^I++Rr_|B>}_>d48 z)ov?a{100EAOe2_ci%KaeT7=A{IrA~-LlTphWcA(+O@m_LQ&)iBg&eZh!g7_yaS3ath;2&E ziX0gyEvrrk+Xmp6i^Em}Hp13aReuY_Gv!I;CrYpP3YG|b znRa2K?ddEtJe7|=&nC%-4EXV%v=uA|vE*S~%J*Etimgx3w@RC5INx@OZIvg;;@DC7 zMo~74J&^Xuz|&vAr4jqxP8>*tZ6`Y5XZV^{o9q$1Hp{+AMS$V-{v2>_#|*5CDAFv1Y+8 zNIq&>cCTWVC710}N9&T<1YF$**G6_-tA!%PFj!l`VcX?uwMBy|HOOxt>MLFz$YCE+XX2uZLwI$QaX z4@l<7{Zli>=21(4^D-3po6Kz<7Iz)lHSx(%m+m`voP8UE2&75%M%A-Z=9s?s(uYuY z%LVtHc4rUVxp>ThAZFsoCsU6}jfPiQQLZ+d!yOj2+mA{I$|fWHs5q;Ie=~p`9+@qB zmId5*r9agiUOmZo5~H(Z2^}U&k4xi;w8ivbA)^*0_(#;;GumeKjR`hSBgRxyV$1g! z-(;Idr6y}p6{HXoMR91I8L_e+&Nz2l(6WefPwuFQ@|w`>RCayTs2G2ZJtd}e)OqPR z%gs`ac@fPQ@nV=qK3QtPMN%`ga5Fy-T!Y(~N=wJt!~Q$BqUS}8L4?2n(VlN-&4>N> zCTq-kHjo-(XR;na`yEJ-0@?g*x>87a^Ghr=6n3K&l-EZ{y~bq=P)=N9m^b`YE; zg5xC(+vb6u46^zbD2AcD2?4#xwu~L9p8#0A z=V;fJoT!a~A zTMB!|HVa>T8=1o7O_*dYu)VMum}M|I4^Bnik?AaCe4kVm)i|xh(Z@y*Z_4u38Hx*Q zaxO&?R#^|5Ji)e;O&H(XunLUx-MrUT0bHFO@jSVx-OECuRJTeTPevJ!nz-XV=V_`b zJ_|>@4VMgK*pBh(p_EH_32}9N(vaoo&5C80*C1Nd9ZnPv@kR8Wb!T4=Y`M~y-Mm&J zXx_jy_Yak=Vx{+A?|B`=z)4CLUey`XX%FDdQ(2iM#kCe)Dd69HA{pSA^6G==k#se6 z{-^yF$c+`XT42y?a z`BzBP24>S_drqXBu?KwotL+#nNckdPZsBE^HQaxL=Z9^FfISikixsZECg@h2|ABwO zyDtzb9Jpf-Fw=ynn30%bWVMv!)932Lm^3^9h%&Qz6ZC2)&9bGoIv7PP*giQr-mz&t zcR<>-0{CLIO8js{ffsY8u!|Gqkq2?J;rFL{)s%o5WC;_8$W(0H#DT%&J*rW|Y4gy; zWO6lbn>ea3o%xo{eK!e_-E2?br@&mvM4BMEcf|bcP$F6Sd`l+}4u~v2e~2uXEz9rK zk7~m^2@l=-WVf#tR*eq)lhbK${s3yEDSvoRGmW5)o_*X0g_8+|lgoNf8YX*=jhU3# zndsHAF?gM+y6jNL|E_54D%Zl@L|7UXrMgN1Y?i9g6*E-=VWus_aos>Kz5j~+cYCpp^;2ncg3Vf&WN$mB)OxZ2=ufn9kzO#t4R{;-w z1I_XG;3@Yxd~mDL(zPbDrE>wlzK3ax@?;M7SkXY0R-(Z99@QSfI)+nXEfV+KguUS= zjN;wzk(I+tEpmsg!0PoI;^z*-F(~hz)q%PCIMt_OyB~;7Ir1U&;9)E$+!2uva18GX zI`RP)aTy@^04NDl+hEvpG_}Da+88nPPo769B{PH$*xC>u6_orbabCUPERX$yp1ZeaefLd+PM-A zu?uIZwl3(cdwggD4SodH;7&N1cs)y+dSCJvd(ay?syIUVX$dyt*b;`T8op8sQGkMh z_5Fd3Ju!8-_*Ef0GF8!=7H7xxPhf&0(k|xbPvB*a_ zE!M{DMPG|rf^fF0>EvA;<6Vp^y?qcG_D=DjVIGnml47})yL|8+M-rS|l}NxdWR+Y) zkc}BsqFc!d*KEi=d6IGw-*Y%E3lfBUdgQP)SPyLkpHn|>1Vv2@AJ9R_mp~H`Gsc5j za()%uy+?rB=?oA4>Rfr?&7(5h$_T!~S5Dt;rNy=aM>OfP1_-BXLF|S>)MVr0Ctmqjtwa_ ziLM#UUbBMzGd5JM@qj%3Gp+2ega2@flQ8Vo^=7=X%Q< z_UJ+=NV>%qagqIY*`ScN?2$02g07yQ#=bAJup^5S*jopqV%sdR5pjQ>>2-EgPUOrG9vx~*dPNbdB zFguB&8M9Z3+aKF|ZFZCx*ogkJ*AMj`6q_dqs{=;WY_3O5Rb;5baP-ukMXGk+MSSzN zTtV2IWlxcb=BaGq+_hq83JaK5DY}NThvt1Pi)IB6Dr9R}T79p*+aA0u6E_Xn`}l%* zS>T4zXkur@FqwiaZRpK@tmqM#PN-KJlG(6@{luA8Hh$soKo9xUmoFSDD`pdp^=8Kw zMhBL_?qFXnj2GjdVZSe&FgjqP)GQd6nuP}uK1ArX%q%1#*bp`$&poi6SlThO zP#Mp@puVK;;yAj&xGGQwpt4b z_y82>5b$Qw5)knB5#Q4=8~#nellWVHx-A`a{AWz(fB3W$(Q{U|Xh8R@dxW`CS(4*! zuLATrxa~xEXpyeB;xKxU9V|j}ik#sjIuUl0r-8?^ZjML#i{K$QLGpbrBuQrENcv@_ zx{kf%*d@ko*!zGpR+R1C>wKucOdOiPHZ55%miJ|;4_Awm;@RGZXNjN3vEGktkR4#J zKZ1GR`^_T}qL>u7_vfYEWU@g_e6$!l_w+}{h*x9SOOGbXrm|Cy+QgstFpVogeA~kc zU9%?D*3g93jurNL%);Ipv+z#@^J=rO24NGzR|uaYPui~B*CO&g#SwUB(OFr4wUZ1B zjh1RGKly4$|NCxY5zC`SzIYGQjsatQUf0zIOhN$+7OzYH_2MY*zDLspnyXJD8#nCq z2xE|kl`o&!?KZ?zdc@(3~` zpO$pTG#s+o(H@i8z7_Hm8|t7J744o*1#mN1JVrSLFhAe!wdCWSEUR$$P;d{x>6R|^z+)V} zitrA&0geqc!do9FDgvcPVt5teJET0n0gvOz{}C0;M!AQXZor7}+09|V1=4bBoHnJT z0lrZFOh4jlXv-C8TIpg6w#QDlqwmgOM=Z8@KJ#%FUVVT6^NG}YF&rzl_)R#;NzVJb zz9uXG<=9?V^(dLk=avfMv6x%%hAva^HW`u~LwgI&lmR})F;@NvUIP@>Iy%W#cdDv) zes2#PQ@}%Da*%OeW29-pVckG>;FV&(;ojhmV&Hax47UYS13S0D^C;HM({qYmSN-x0Ohn)&P z54r8^&Gjh0LQ5wqhsCSm*P&jndg&Zd)&vu=HoJ+iPhjoQ0kYm1i(hgy#X5U>8lRa+ zIhTF*3iQi+k~yj2OE2%8_GA~?u(K*!lM!9tbDYA(H%Q+FV!(Sq=^MhUo}Zo*-7@l@L9;x%=t`+x8vwy=pajc6A%BO)!CH?h>&l*7r{P+hfpR zIJs8Po__Lgbeg2YWWC23bN8KWPi_O=CxAyDV@>O)i>v>}GFV~{Y6x#@{H-K-uour~ zGuTJ$na*obA6-=~*l;)`c_-8u?8a-EO!L$ragvjjJf#%nI~jXwB}KiS9UDphuKqV_C#dlgsf}HRPWnTMoIjK>{9B3CHdpgF!T6n!+nZc@1sjm`^Aq<8 znewDAD2dJ5bV@wBh$)_#DU-9OpV5f3H?T9$sF4VKRwEv$+-rC?MTV5^Z{wmLIEv~1 zw>O;XMeM}iv^Xcro}1fyYdr|s@%+cu*qI)Ax*LOAVl8}0C98h!FUAN+8FWAaPjG^j z^y{)-0N`mVsx7+=g$P%e@-%<@LEkuV-v>hllT3!hl?5zp^K^7${${hL(R=UEZ9iS_<`nF68s?7u{|7T1U&Z^ZrZAZ_Q^JTe4ig zqF1z|UYFufepQMi_%~7<$-j`|82&NE8vcQlOyKWHagsDB@8V?shLld_FH5nUKQF~9 z{;U+o@%2)y<*TLGz*i!6A2WMhi_QEY375qeNO2CIEycNfniN}jffP^T_e*ghA0x$6 zxk-wr^K>bm#Yaf7jSrUMdAy$#SMXj^?BL;2yo7g_;-%aRwz#Mh`?$4z{>=7 zmOu{B#oOo7u&)vNx&(ek;1vmMCGdg--XQR-1hx_Qz67=t__hR+Dadz8U?73pB`}!4 zX8=ll4yBYw%7`Fv`CJz@j-SoqGEPFr5NMacI0C0hU;=@YBru7~h^XmCq59 zF5G=3Tz{uznltF?!>NLBs}Y`d!5Mkob1*_ALh@Uz+m2abYu=`eTVJ3(!`sZYqj&ULWFkcwYqtVEhH&jI z_WF)~;+nVFM>~=!`gMnVAW@)oGol;>3qm2nbOakh1%mr+HvYw&NM9q*0kIw7O@tF~ zvzK3dCUiI1C@L5%{DtS){W}v>mxe$b^MGiu$=^Xgd@s-CI34)hAh3DsR7lln>B~T0 zW>&W|Gm1dp-89{=ouztLRkNErhsG7b)}~iKO}L1Nq>~h_i6d0WHWf3xG&u7t4l;hX z!N3YwBX1p{%C^y0OY{+HXd8XCgd8ywOke)j&9ML1vTZNLhN)h}E0*lbv}C^C%}%^D zON{2~%!{q`@=>OEhV zgfkXc>3h}aI|1owCArja z(xH2AG6~z+y1u=F+ktN2=bvY%Ux_6L;diejy3#ij)iu0TvtY!q1GsaU?aeMUZ^FMq(p6G=Qb*b*r3>$->!t&$ zu2(oBY!+XYC{#BU$ZhoH$>Tnr>yJ9d=)#EDVFCr-A+on$&FDk*>a)3H2yi$rCwkA; zWM2c&ds`t3f33I7!iKz-B$gJl`(KLO1i3<}&eh zxA>}yJ^uPY7ro#tZrDaqO_(pi7h<>Z%j0pVP>8HwSR_YHSGt~vo{o7c#!~?&{elcP z7Qx(A3>1#Xc?B(U;AErW+x^22(o%`T^w8k$@3hQTuH2r7G8js{`%d%X53PK_cBJ`F zk5c=gV7c$jEB^)v`V|QI7cG1pQbl-+@hw03C;Y>YvL(9)iW?qff8P}^9!+FNb`6N_ za)7oc{AwOno_`bvFUMkyN3h$wGE%}(LyeaIvaDmoq-|-DdOBgorKh=f@@wPhW$g6b zO4l9`l2`KSdj(o-I6B6I`ev3E_c7sg*G>-x9*UHXbymLmpVA=1tDX^2Wx&m>&VkFr ztJWaFyFH76ndiY0&%v*(6F`Wc;#xcbUR7t$bzInkXZSJr(eSG2pw?e^M1*QT;YTXCD~Q z{l&!{z1#)Roj}gtY2hgDbN-$NVG(snPF;EfB>pb-1|Hun^~RkYmKygo<%t5a4qH;8nvrW!tvycUW{FaU-U8_i12^V zAt~>2O)x+1WDg#U3)i42Uw3a%1x!~yK>X`=R#KkM2E3*3Uk|qO-yN8P>gRs<-)8)P zq<=||n$BEr4HUCW{_QrS^{uVqC+GLBe>+B$z06)Y6es?Eo1HwAAYQn=x8=|%4Dv+Dev`@iI4okavMg7G27X)hM{8o zFYKj;vEtYucB5gqxUh~T9ZL-yRfqX_ePJ}SAB!H4bpsP|n65^=cp&6;eeFKH6$!!7 z)4PuV9+1Q8j?L7Z&n9ufhX24vc#w1&2RVd)BBY(cJqCgsVIT77QmEs@M!Y}Tm0E}< zKEeU9(b=7Uj0NkHR8OPle9Jh=SKU4VKfc0L(0v9ujvMjXR?;_E6kRO50cDC>KKfs- zs^96k#ALiagg8gqEF?Q`q6E3piZ9j_yPC1r(G7DPw-FkGop-Mj4W3VZ;`GO#L7Dw@ z13T#WBbe8gkENeMNk7@b&Otxl!iA|PJp4#*TA=$Ind7*Hzf|xCNs~j;;@G0&@nH?K z?+%-PIGb%g-Zya&>Xv?g@=q$^GY`B$&DaX0=cYX3Np|VDhJE|NkxmA%rLK+b_;6$& zJ(h@XhgmlPf_c0JdOn8E=*?M7Je3@^XA4GnER?rGHvK;g{bw07o$4F1IL$Sfo)0&y52X!MydP`CPGK2Yh2ieBcgSrz6e=Fry6YdE;$5VnT*qPI*kwI(F z=~nojo$c7Ag6&;+r^i{hGnuY)Gdq?@%KaDu7Wr~4@_alWz!ii3D@U2#n9oO!L~*#e z^>-YT^A03R~=W*NQ z`wh=^T&ds*LILkuDnX9%eEdvy?o7h)W3cXv>d3u~B45_Xb-rH&^|S}_m!vFT-Hvx_ zi_Bklp-fL^QR?7ob+jDH>j-mWIgmXNPx|vQk#FinD!V z2)_Eo9b6*NhqGLep^l?$$Ju_eKJ56}#p3t3SmwvE;+b3QfsYjl(`jG#MpYofA4P#C z?8V(MF#JCzS~8uz`ms{{{Tuep$HV*iY3G4fdEx3~m?7R1v~F#Xvn#(jjSc@~VD6_k zy4>wx!aCf@*|a|qu!jT$ANGXFb3 z$>d;v|D>O2yUFT5NhP1@51;JGbeD8UO4C?c422M|uvswEYVZdx{8onlO-Sgf$z~IJ?tcKHlHiU9AXncqHC}A^I+NZ)#pgQ9KrTzz6M~?&7iGV-=d>C z{`^RkAeiyQY#H<_CKS*@SVOpbBF`IL$wl$)Q#wZa>u6AsCHN<+ZYtY)E=u+t+kb8V z*729;k}`Yl#znX92Vxs_B!8$H4#9pXh8wGAp)w@laRz!CTSBEra@>m!NH-$I4a|7{ z;dt^!*X>3(8@ACm<_3SV369s!n7l&v>G@FsJ6p}Hf75raz-XD!10e}Pfsl((g0L9j z351ss4k3Jkz!3s_$%I&hR0IQp1z|cu1;TQKClGcZ96=Zj$BGIe2_Xa_20oW01UW(u z!t@xKOQ;0!1j4Hb^$2GXnh|aw1os9p!hHyZ2(u6>5gtW&8sT+>g9yhE&LMn@(28&e zAvjhh#37_2+=nm;VLC!3!fJ%|2+zkBpb`M}2wx)HKnUq06A} zow;DXqjJH#1Y7C+S@X&&6Bd;@9MrV0p}peGCj#Q+xO5}8$i)1t{A~k@f?qlCudF?D zv;XNf7eujFe>H@h#bL2^6#jJ5&UO8xS;TKK0kiv>S=Zm<14{e(;^ZN|c*0O$JZpp$ zk64gqM#X{vUFfCwgg?HURw>PdmOd2GPKDpT3(d?{ZGuw5$ceh^C?%@VZV|Q-H ziU(&(wY%+9M|jvgv%fcj@WouIKjPX4eeFLiMIoz*P)yx{Mq)c%c+o8U@FIa-E%261 zLcoW4Z}$!`9`<$T?ITjT7eA09T#cgm(6wKfg`MQi^nXr4>gdn+-H5t}3&j%3y$+85 z|Ev7}qg8&`TRNg5Fu%-^Bdw0H^UD`hJ}9kJLHI4OpmdI7^n%KKESGs@_s_7G&2(fh zd1|Jtw*O3VOqnC!QCjJ6R#2e=p_(w0${wt+lsaq$!c#xW)On9n&H0X5zRCsRXBlWl zV|`D0xS~uLM!BOF%wM!%UfC!|@P9~Hiwfw)}shLe1 zH=|hKZonJ2#>gTA>pP&c9Z+)z)Y<_F=$x-aL5T)w^_Gp#a|H*eKwuT}g=zTC7A683E8H)PM|v#2M+2tM;lLm7 z2!c%dM{`IJ#U{$cClLQIpg(=eF=83&a|o5f9AUmN8@0_5<_Tq}c|N9O0j{Ei`y?W{ zpv52a(;=vY8Tb~2aY7Udj{?zyLWNL@0uQ4*rAU=YjpU;$%Ax8Kgk0n}K{FG>E<cJ9iaJ7T!N4c>=1C|zj4gRJk zB`D@&{^ucf2;J_@eV4%9fd_JC%WU`?z_heB2H9o9Tzi#=l_!)RE5B9VQI1fpQ5n?> z)o0cHH3m(NCQma{Gfz{g@n|+`c4}VN?A0`Be$Ygvr>5VRUYTB-zA=4o`upjhrGJ^; zoZdy-S36LfqLphiwOQIcZN7H8wp=?`>(bV0*K4n6f7K@I*6TLwj_5Axyt=}S@{HPy zBN>gZj58Vk$hecyLqAskpnjSDJ^e3wy~C`T&q zSI$z-S8h-?D$gmuRtBgPswnk~+KD>XaouU%kGh+>6&cTDe4TMU<752=J=aGY;tYcg zsfIkmBy?rAq0KliGt~5`X`ShLQ=^I2?!z;6WVB(E?-xybLzI} z&g$B9XX=-LH&y!8dM|EUyBfM1dKuykNrn`I z!k{yl4P(GDtD(eD4zB&xu-x#NVXfgA!*;`N!#>0Nh7Sy94d)Ek3_lx0jMEOj4h5RA;-yZL@my=Dsbmq+QqahX@6B5RD7f`xl|U_YSl*7Hq|cGLDhSzn<|;Qt2#+NTs=}< zrhZQSvHG%ll4haiZOxGMob=nT&euA$%d{)C>$C^8-)Mi* zcG6AMZPQ)V88Za^aQzN_z5Zi;h#|u;5|f#4m}V$5%r`g;%M3M!Ck-17uFZy*47&`6 z495+pz@#=qm{Dn5WPI3IXMD-H%lN+Wk})82S?1Qvmmz-NXDUrP)BUDNrm3bGra7hs zrp2bErYe)iwAQrAw9WLI=?&8X(_zz5$jB+vCt&auQ#)-S)1^Hf@)9B^p^cL#$%n}0 za+5sEB_AixmoJh(EMF(Lq*>FZrIn|xO8Y%6ND-qLph#9^D6$k|6=jMAifYAL#mkC) zio=T2icb}n74gc!$~@&HNVQYBRe7IkvTC;Kb=7gzSyhYbXVtH&&T73nM_sIbLj8*R zmU^HjS5u^!rm4~F!i0UTNlH&iH>W?I?pl|A2rKxf^l0s9ZK3ut?MvES+Ap-Q?FA*PYRQsry+MnGv6nl`$d1nlUTGq@RH$yH@|Tevo09A>A<6FkNEp zM#B~?+z$;O8@_}*bT>{gz5qUUg3x_vf<7m46Xa>>Y1xXYil1DH0m_}K617bor%BN~ zkKR?KKaoCMJ5Resdt57IoXD7~uhH+)ha37r(#M-<;pd=&EO}E}AH^7jO_8hG4(0T! zX1C@|%|Xo}O}*y0=66j<`q1>O^u_60(pPC0={&k!x`T#xW4h@P(+<-Knv%7GuoRJN zPg3~@%A(P4BOS3!neH-2WkWem)W$c)KM$V|J@seK1ZLguf|&MmRXS5nptlW zcEfnVQp}JS%WulRfoue+d#Q)0?^lW%7`)bFTItG`xXR|jaKG=nuSW0BoR zznvbS4c3NgBeXj0G%T?y?JDgW?ON>y?I!JW+O68}wXND4+BR*wc82aj$jFnr4Z2Ob z=X8g4-|2d1BxEFIBxj^%$TL(K+6+U+V;O5Q)@E#gdQ;_0oSRvZ`8bwjkSW3x zZ5m`sF&Ru*rg2!B513||>?Vh4nQ4`&*7US#tLYUiL$b9C!E!+!E8mj#LfR|XGde4J zDqQ0fGZfbq7Uj>%o~i(Ks=8YJv*vXAX6+g6WvqZK-8Z_P8TyQU*c-eVYQ0JSf&M{* z$FRo`WgKSA$GcMDthBs^x$+rlKPxoK8A`XZT>VMrteGjZq0tpJDQ`KPc+f#bdsP| zVoRHsR*~jNTLQ(qYU8DaGJVl!rY4gR0UsA-%U8*FV(;P*!Q?bmngzsb({`twN;{j@ znihdg)uO0SELCj6rdqFPRtU;aWwJ6yX;ZFJZddM99)sn;mBCm$IV#(Mg_W{to&OK` CQ=ZKL delta 23300 zcmd_SeO#2q_dk4Hmy4h-g7P5GAOfPI?jyU)F3XyLXke@80x60Q=(?Gj?wXbawo=kc zsbi^qtR$?gtW-?RQZYk`}faH&CZ!K zXP(ZSIdkUBTvm1ZRo(PkYv9}0WoB3II=%GxxmT0B!jzXYl20K0c|ot_HZe9NH;ZvV z@-DxAU=MsKKTU^FZI!DeQ}l;Pn2qO@_CUHVk`q6ZN3B0#> z0OG&=nK-U-#Ns83X92g(j34xbn}|4qED)mlUh4}GeG*%?HM}6q0sq& zcnex=yvK2qE(j&R`WZ)B(X#Vsnya9-M?s5?8&W_$4*cKgyvotrm|lQ~SxZH%DL#}; z?`iD&x)(2^-VY&f_tXrFkL5U5PJ7ApX(yT{yRJL0OD(n$9!G%kR*j>B{t&=%B_-93 z_E?M6N&@|d3{4XIa?>StY%}H?Y}Vg(jIGV-Fdq)?Gm2v)cPB9NGXJ>BX`zflsz2tq z;O?QkmCJ_Hvll9z?mFge(c)_NgoU=7IA;oH6KuSVvzXDo*%Ourq^rI9mP6_ZGckgW zo(?`2XkZIBn2;W_@P#}plq~8 zV1#BG(&_Q6T$8jKvd767hs2V0Z%HXiTD2mbyAm>G%bsRIy{aNJRg{YSs<9`fy3_{H z*=~8hG$V61y?NK;X_h1vi{bn;D`r0C&E&0L zgqXPI?8>C69Av-~mW{^HP6TT#&0JIj5hyD1*wHIbSTOSR&5lejmy14Nt9L@}h9rv7 zO`KpZva-eH2}@#CtV(JD){K%7)}3gvh`#cN$KyO2%rLI(q@==nHY1*}1WQH@qn?rypo2YO9MGpBjT10auFSk&Ox%a5gh*5Pm&gHW zTC@rU3h4|?rt(B}BbHvF)eO`Gn*Imj14E-n_47$>*R&N{H;8Tc|3PGdp;7&P`Hmmw zII}p{ZF^+UzMi%KGyR=R3yhfCg5ionAE0cUQ`6IVOemrwG2?}|FIi8co@kGaD_;4d zHFFdf^n9b4KKHxGfjLfnq9X-GoW{h$pbd>yE|87u5#)O7IVwd}&L+XxEZBmqdE`o9 z;z+AVBNJB1X0mjm zJ)q|OZ#q04cVKfhi)WJ7z6s%10J$zH8>=s)zt;U?ld&BnDrjWDIWckVHkl9<6QM#) zu6AXk^0NL0qwEwjci$$}K}m{tplQ8a!6@eF;v7GULmgyI?F&9^03U47wUAn2y0^0{ z$mO7@@ZzpqjhO4r$|rq-WhHBnWu`Z8G5+YPFQJpx{fwIN^eV8npecn#);x5A{s@tD z9!;ulbS$;9f!0=EhoFm^)J!YFSfphih~4tYARZljjqxF{kP~j)=?<^96Gg`3&%pCO93hy-H(I9LH^TkK;;Wi!8JJn zpbMm%MY;&kxldp}nWNJ?66mV#!nV7$IqhaEYnnF1i(_=xK{p1YlP{nfSamfyt<5=| zPyil^UPHIF(yJ)n1O-ZGU^^*ka(S?hEG|a?^k@hj+y}X~@M$$NdMbvv`X>puW5}-l z@=6&SPbe~3M+1}|{cZ1H!suJrN_~SVV7uyZkNbYiSA|q0uzy?Q2(VBs=BEThu+aVq z=*FemoM1CH0;z@a1KD_6(d7-uaUPBNXMbB!X_J9_T z2q)%0!a^faf^8aXPcYLH(d4C&xVX76D_rdm1B^35m(ytvp_i~#VLA1tXQRniAqm1V zEx8ktk~kRFMomvwu0-#bnmUTJhr4nET!9Hgu#e;|LN}foK}Lt>2F9YaZ`_Sh)_Iz~1b5)9p_SA~7f#%@re8^Ym%%&_cS95l{ti&72gCC>w*Ols4_>5=L=Nr%#`* zYqXFx;Sr-6Eu%P_#7tXEtPAzqa@mgRh_gv7HENbeoBx*KXNxMH)}%cD0jtODSGO&f z{1BcZROOJs0VASjjpn$j%oS{yc0W22kw?vHq`_qJfMFx+vN>+$cqvzfrrf=q#}X`_ zFe`?KJzbPk?9R@P(28(pW3ZXpv&nk{B6uq~K43)OiKw`ww8`0+fF(YkNFoM?4Eh@i zwUjmyL&V9RGuc*BO61!@NPJ|Us0yq^ugnhw)RxAhPfTc|b4YgNMAasA$#vX$G@_=U zLx0LKfU=d_%nb#@=x{lrEV{adZkhlwNVp=7yZe1QL?}nU|9J?OBYJwqEm#t0+JiOP?(xe z_C-I~>+Ac$E;GF{jzq;o33N0`kICviAYW{W4hN;_|jk(wL}i)i*Z@@8x-lbPeOgDb0GF}VHqFB>+ByJbqMvWL>wu|xYt zHx{6?pw*)*ZFLMPqJ!9sU=0SF!8~aZ?TdaxzZ6;MYvMNW-}K9GC@<>6WvmZgo=({B z!J?a^%L#Bul#OIT+%SF-c`+_})C5;fFl0ZJwqg7AbzluZW#fq^C{ZXEd!VZzp!$R( z*2R7Y(8ha&vjCDS2|JX2_i)?UN#eMttsFvUff`><+b}vG2w!i5 z05Tn&$~3C~En*lnDDVVuA`9}ds}lX-f$99c0^4x>U_JZ@GVZDfW~0>KvB&6dU?RUp zp0C~tfMosP>2VcMStTXT4v#(2Q_#B1-<8v{I>OzrHmAj0NXtl2sw^n(Ua=$4J7m|) zI5Iyqc;a_Rc&lr3vDNh%(jG^Q#}VOigzE1s9qQ7KF0@#at4thShfNKu#Myzf1>Fup zr#8`1N?!(4qjgVmMHDeY@@Zx0qABfku`a< z5Bd{5M3>e}#9^h2+?wlb`Po?!%-7^f9epkIELJg}BXyPsb0rq~0I3=xFKcJY#!UB2 z0Pjy=rGNuL&mXBB<$>ygL9mle%oWuBw~2Wa2pN&@0Sk^n!(GS}; zEX_|y)6gX4IX0+cIB44ZDNP3{I~8_%SNmPI+K*wWOknwp z-X>UR&!5;lN^AUKQ#50}+yRYuH=^xi!7#O95{ftBKvGmh`ydZ36{yjYccC)F1!J}Q zyDFsE|H9}RMj|#QV~u?~OkvJvsKk<*0M{?>kGjpBBIj8Tm=sEfBilCx=Wt@^)9lW^ zl`D1HwHE4BFt&EkB=FBd^T1>?om56jl;#g zi9S87%P!cM3T9y6vxQoN`=RwGQMdAzPQ4W};pk(5PXIgUGV=O}g344_l9=-VTMR9Q zMOZ@5B>ND-G+#taF30-73EYz=ib;Sinywp)`aNOxb`w{>h3(TX&W13xGGY2+mX9ic zKE$Z#c<>3*6=3T{A7ls{Pvxw8`eSyOLzQC~u8zrMy%j%Zk9sO1kGdGYbWodD4su#_ zj#fucGd(*5%kdU|+#_5$*IfawUNr^Zm?zM!Oe8V=<CK1_N!(#Nlj)ee&?(S z3zo4cP$EXLEh5tI8!<}0Bu2w8AlmYrUjRs4Ib3}wlCB&{J%@B{PC&hs;j3G4NEq1c zmxoLq|JAYF`c5#h*{>bHy(C;?v)@g!PL|Zm1R5~z`Pt-c54Vxu6|utjHWHzX6-KrZ zjWRiD2DsrXAOYx(u~zX+%7SZiq|5*@(+l%Rl~O%Y#^kFQB+AR;sRo@|G7_D_zRlKs zi(a$_gKAGgX)>#?Rhd$`a0Cc&24A(oi^P1Zc3SA#|3LmqKm_JUcHS~!y-Cd$`lE=> zb+2=E;MC-vWz*7dCN!=~u$?5icdPC4nbfRjL@L0O@d5v|=1itSg^NfrVKgsvzIRB}W;cElZrQvB~3X`L)%p z0N_WWjtr6ze>r!Is3!|Lrz_7)pI)Lt(L#}E@2!7v^zS`M| z9#qNP0?HH0^Gc8B8de!RGd6Cb)nYKg?{er3vOzm!(90^OLC=9$bQG+hiQi+gr1>3IfrEl=-7)sqh`Na0wKZq`4 zqYyck?5YT9_#q>+hmbRyBpKHUPtGL!jMc)T8N`@1SNN%nypi=ezoGtdlOT*LYpA$y z6we#T;n7P3ztZ(j4Jd+n zb%ir?5BV|24)b|x?h<}0IhQ-0ze?iAqzO^eNx_&n;Y=xcc+4PSdns8zW?S#epR#Vi zt}qi*UQyb1Z!JM9?{Z1l^8GDzx9Et;d z*j67(1C*6#_|Y*IGv$-WpJQ|2m>NCq+CWxwQ1v9XHjaB-SU%I(`Z<3}98jorQ9m;7 z{v`>%5%qKBS~I<4GE@IE7*owOYoRN18@Y0STB-_FK^p6|Vl1sUAXYZQ+;?{PRRqs* z<&O@YQzOkuCGz~y5xyEjM5c7tIsPRaT-BGjk1F zLXQ5nKdA*hFKPB;2n-PIc^7mc46tp)zu-j)Yl!qMSWe=`kIsGcq{rjg`w<9fS2DD-IoMMW;Apnc#V3gMfns58Kl1Sd%Pea#&L&UTencNjIG4xb zQ7^&~%@g(!Frs5V+M~|Klt9agm~MhbmOM7Csct^#*-;@J#z4N8lB_jHCLINVGf=Fp zZ70;dEx@{y?LKd}!mF46n+a7OCuhtiejDjKp`14n`vkRrp09$P6T%|Tu?ogZJT*Pt zy)4wP8N*QCf`FbTS0^M_USf64hVl^|F+r<3zzJ6_+=t9g*?$bguG}Md_d+}0HR#K#xpD{?ou-e@_gx~ z@%7f@tjnxFm zRbX7us!L#PeHFlGa)MuF&NOF*6iRig$nn)E<5Dy3c)oO9swq8Jec5jHs8{8ZYZi@^ zMLlwU>aihc>n z!T}Cvd_HKEKc>^xgN~R+UMn6}xf)$5qW?NC+VYt4>iy`E=!QRZ09ls8IrSf_m_lHe z5+B}y0S4y8TTgpT9NZkUy2{z5%0@H$qZ2Q)3ArHSI1}14EFEg0%}CTra+rj;jt~3P z*5k9^tw&M8u&)YbX4(KHiz`P|d(iq3u!llnu)?#FBI2dMzvg*g0u>Hi|EZ+1BrM{& z2U!uRds-p;Yg>*+@QA)GWb_y|M>~@U7&T2a^MsQ>#X8&msFBOl7{D?;ac-X9uq0 zp~I^X*-W+sdK%0XO{ED=KTzH`B}d9gR=n!&>52t{Oc}zrkb_eO#IxF5IUO$Mxz6c) zhFhI05$)vKTq*3%HRQ&WL9CI8sUt!RYy=%nT)5o_g=2)mX(1(3hx2R5(^C`uS^b`{ zZ^+K6F@x3~f`;rUfuTuA_YIQxt5~`|T$`PfZX1U$XS2$wPLQK)-Zg z3?4_C4!!^1E2=3?9WBL7plfi_i{!!oJ2~r24`x5KL)ADA`cB}%)&z31G+`q1N>@yB z-4;&-x|p5zoUT(wkeLR$`53_D0Xb^yfez%XZ;rPnL) ziowNGUPeBdRu-L(l$m~32HfpSCDD!S{+GD{LaioZnw}_Qod~LRVgZP?)k=2iwDxH( zxG-_rvEa0`8q!lzqC9_J&E!tj#QNz4eb_liJRz++Nd7b3z&Db?Wr>)Yab-z7B@4@< zh4K5znzEeO`RJwSbFlV@-l)BReZVt@T{T_S#U_xS%8b0544;uScn#*xCsH7#^ zpM)doEG*<^4oKJ$(genUkywq};jf{4$n4qorH%tX#lwqTN{u3H!a*hwUT1n<2cZDh zVTWhf$fvVM2x$+JyR#LMfAeHZ66~Qip|EfQ{lZGdmM01aOUT3J^1$V1yIB}>mb_iA zjTnT!mbm?J+N}0)EcT; zyR;uGF@Da<$|+a@&IH^~BK8dt!>fR|ppaedt_aXWFOA|B$5S~dY9i=>ZbEthHaKI( zcu>2Jq52}{wRm3(F8?FfdOnjcliWJO!?Si1q$vC~B5x~h>q;EjJY@XbiJ{MUGZ%@O zqMU`CA$#T~4>5`uCZIG|1mVzP-mZ+n+E+!c-`X_cfR< zRo9*2YFOGO&VuV69C-B5TCmrG-z6t)2maR5a}P}t_RJtZJTx|fT{YVLrorDifL=M? zHD3Ay$(^4vbv0X|cYwtuu=CDcM>O4m1e<_$-K1=Qdoj@J%gX}vQ0?G;6s`7=*}BPS zAVBwdR}&a7GM#LFJe*ve-#5zWqa7sDGKRYM71DzkJL&oGpunGC0edHb-fAY-ACDvt zJluPjAM+`Bk1ue0fCFLlZIMu%knsS~FYIBa;8HT};mFd*fbn`);{dsix%1!@hta5W zNEi3Bz-7c{rgxTiJ8cG|y>kU0AlqO?+7Cgyq%W1RntV>#-L~*zC{FVA!zsPzW8}CP zY`Zm^^j=^luEUAKn%~HKhc%>ifqvk9Z^0KJz;dps-T=!N{@Y>n#wToKVuxS&gqRj4 z3x0RWyv9Vf-PJZGv1r%A)xzW_8uW|81W5|42l7Z$q$C2yA$eqRoa7e#2@PcNFkX0e z25DTfR(L3lOtdc+t_~x|?BDXq`{p7cE#L8QwgB#mQG zU+VHyUQHf4IgG4W9xv>gLS9-vLK4Yx&MqG+al?cle=bjxOudOFhd&l8y!0~3e{4df zbBpLp_}O9Nep+JU;_UDtARI^d93c|;u}e)}EyXc5>O9I_M(jJe9-p-^n1KHhD>i)nh zrRT`>#|I4c$>dQV@=e4(fyxK?51EXR#G$9;qsIqHAGk+YMBI-@1vLTDghmSK95xzK zvS1)luNc6OB?b7Elkydh4SWWiq2q81=XLss6Rk(mH_=!3t*>}b_{7z)t0GF^Pd1#c z9LNjn29lqv)(dCiiF0MO@L?S3{p4)H5Jwh0xq<(b_&M1mYMsG?;EHV+SKXKAGfBx) zrTn*K-&11*bJY5$6NM$aNpW@>F+Z&rdhaH+PbUcXtt0O}J$rE2b8Mz+$8uAhCT{9V z6Zb4a8-n^t6E_OsX@n=7Bx7~LtP4B3`}rt9^x9L|b`48_?K2oj6y1G~VsbY{6DW!f zGh+R>OT*~1_h{q*+nz%AB{CfmajROq9YWWKCIzLulA4jJnHBW}!Qp8kVjZ%~yIoYJT)Hn86SocBO=8 zmslVXkd`(k!iG#cKgL2=_CpJ`_MZGRJy$h4{$rDjDWXSU8nAOkO(v63uSKIAJUa&DTH*za9puhyv8`2B|UCbLAZ{4#6Hlh8YWz5opwfJ6pUY4hsdWY=^&u6-68@rrCQYPneVpU$66bCNE+eG$~(i5-yD)Z>+yB zyd5jGCc@d*Lq7-PwGJNLN<8aNhxhHy`sJ0btmj`L-;p8wB+_R?zUqn|+&cv)9iD{@ zq;?QNLTln^5uj!c*OeU22Z$FoWgbL10j^BEO50!)UxawqOD$xyD&cOI<#zRDFoW*K zFOI@F9c^AIA)i#^;C6Si;@2IleBgECbc_Pa>&%_K>HQ7HF4d8Sd z`hCOr7kcr-e^#-a;65MQPveRlq*s9$^f6HSrl3{Nq6u4%+MHHa$>lR-)y4_ok93s? zS~U(_^_H4S+#3f6Eclgi@7;fqs7*2Yp}1la+YDOuF^sKaId=uOQEVdEo8x}5Ij8mK z4-dotGsV>ecXVxbrr6M2^2jF9{(62>0%OO%O&S)pZ%XF#N%-dYych@;ro7wUe&a>( zswW;^G&2j3c{jn^U zp{A}UmyU?8SBfFD5Aa^ud~Y*p-drZ=UL-?aOzh7Zy4T7cv_eW>oJksA+~gk+GKyo@ zZPWjCu(!4jdJMMhH}{d0Eh&OyIT^o2DVXYrbIX$~@_%W(Zt)9U`j%fXYnfX)Yf#N45i17v8%O6aJ4U zbe2cS{#Ug)+yCKp4sb`A>E@-xQGb7C0OSl>pop55gH!C| z2n_)E$x3R=TcH{;6y+Iu`4R6paBJygkLd!Kk>zhEf7X|w8`HL#()%9z@7;LNfhw^e z|MHR#&-m;IY9(2CnpsG{ZYCGDC8`!NhU9b<(%#FuTE5=h^1F-vhnBxyOoqOeT^Y3) zi*soJ?Jves+FOi+sUXIo^bhnJz5x0wi_@uFOeWCp#W+bElxJxw{aj3^(Nkh9qfKJ0 zq92NJ3~dl&Eqz;z4YVGy^Qg&Fxzt2AiMVXKR*duLYBA2IRbp(W%fxsRT_ncE)GEeP zX_*+8(PA;4P0eC#rFmjJpPIyY5!H&Zoyx>`8BG=A6*K{HCF{iADz*jmXXw4mprQUE zxSD~T;GJjh8V3F%g3LcfuZiFW23{7yO$A1{y`Mg@H;D zY-M1o2vUHRG**PKGjx9uyud(z5o~8*=R(GTn+*I#1UndbO$0j`cv%FQWlYbBpoD>+ zh#>oniXH$c_F2kOJH(7&2G%XCWOau!_yrLi!9bS?#xSr-1QQtO5Wyq{+C(swfzw4W zje(OyP{zQqBB)}ZQ3SONREnU1fvF;B;>!TXif}eV_ZPuD2KoaOJDtx`zb{}NH8b!h z5uC)p??tefftN&ZDg)1mU>O5HUOl zQG|~W_Ut0B?#K&elr9geIRYUGLA#sW-7!0GO9RWe{T15FZy*cbi45P0Or%)G+MR$e zAaw2~FTWEnY-u2e-$`Xr+dHykMuFDNh>8))5UdD`5SAfSA&>@Qd^a!D+sFYx}8yht*_xNPEIRZGS9t7 z_V1i69D9kx?HVB@yi3OKN;CfWBs(Co_jWMb>rJTP-V6Tv?0ZRk!sWfIEb^$mSCit! zFc=i%o(nw#i~WF#7*eUL?jr)Xx@CJZ(S|K0L(D4ts24 zWlUbX5|g*Rj*#7n0XqOU(WmRk!QD~htKEr}ZU2IDD!iy1scaS?CY1q6j)3>%7~t^pu5u!P0b71#3+T274VT z^l2dOeIlemD!c?wMgC0}{w*WEg4p&Z3mFw;-QHMXcM93LcTm*j_u2jHBWUJ8C6B8^teC`(3#xTF@#C=M25z9^BsNr6L zZy(3$Fc#FevbDG*3Sv(I*n5V}q2gK3Lf4Q3`=U&Zn=lB|d|1_a_>wsRPag59quJHN zqw;tCf-82g2cI4v30hSq;I8%;yHs17#fVm!k@ke8lFl?&g87@cZe9XF>b&M>4=?KMKPm zoF6_i0Zus>2RP^8;FwYz5PUx3a)325cb4b!NK@s zcM-FQZNb*nqCP`@x1*~K1~+5Wrw&r~czBmT`ePsV4aB~-#sU1NTT2qt4@bL_LxTbf z?h#IQu$z$SM}YDAvSx#*gtfk%wLTkoU+W=M=q|>lvJPGTi+8O*{odZ){8IqkU;pKb zhCdvUTO5p<#b-Y7s`Z3dO+r?=*hLi`hk1E zS!!P*y43&l-JO&p=@Zg`CFH)wWZ}`b|8aNH(706y`mAB{kqCjWAPbsegoSUCHBAXZ z!J7^5HVx;6ANrCjN7wO3$ihz+@>=rkC*KKAY#<*Wlf&J9^Vsh^Ke6GbEd4j<=V`@b(fRvDw5eKzIzH5#d+l5zndI$~wq6p7_HP>*z-( z&+VEWSF`J)dz|P<@0@_o=fQq@KrrHF@hO|DnRWWMcPe_Q#OrmR zfrR%L(*hi7TW|XES_lyCPFOnfZ;x??C8voV%-U=%yRKu;F4&j3Gf{by@-pg2>9Xfg zy4!&Io4=^bHO9*No`FPI147Zg%d@}&_UQ>eI{wDihbmYlP%QyWXU%}s0bC2i)g-fj zp5l_zV+yL(GiyY3ka@!JwaTgwpxW@5-z5Qt+)0j;Y(xPPn-YOnDCsFl8rIuex!_DR zoXo6w=P{Gp>8iNoagSU;J{=X^4o|Sx_}BmlSJ1Et zBACObe-7z!HZ^QLyn)!A>1)XKeWdwf88MxWlRm~uCbCbP+SjfblE#8|xPmckfeuf%4bOXAbuUaX*xE9dJ%KXu}u8v1{d#kB|gO&NmuSc9=koG|7yTa(6}6*R`n4f`*_-8 z8oBsc!ief>aH5{wZ=lH6H8P#|>#`7=gxbU`Z{7C8H8NN1e3E=##%B}D`P5Y7ZWQd% zWTqJ)^Y!&XSnSMbPxR##(Ps*t50PKEqHK7*VEH$Aj-YS)w%(Oo9gqH}cE`?7NN z!!W-zP8AriyqBmBjz>#N8kpT|Woxb3UExKv_)^o1!PLmfBB8=uGX z1*G}&r9$*gl64_U@Z2C%E+`Tvp%Yk2>?M9pPA9rEfND|5_Z8akDde3CN*wf0-+qvc{31DDa>L^hJ@9bN2Cd3Q^Gqz>`3)mNGsP%+ z`%-k{?Mr!^YjWBLeOxp7CI}rvXg(_PUNO)~07d(hZ2cl$*!(Lw@*-KEYngClked#@+-x=zSKN#+f(?*K1us}v^o~WLm$W7c= zax&Bthi`)?{24E7$Ok^I#~iar=Kr5U6Kj`L-i_=VhDY*Hh#rPWnoOR!8x?fXau4;r z#YE0}B86pBSarhanXICqGqX&LUi{lU%S0aNj10QaF)>;+ z7}m$V6}2Py(ml)I*_yqFoJ9W$_IihLa-UfE{U$Mb_NdtT#`E6H&@V{uKjYM2eJNJK zs^P)_&u0#4-``XwIPDy_2g|VQZwqjF&#&VL{QrHS7XE**P&fX`^D}Z9`7!18(T>H7 z%NN+qix;r#N^ z_QmsamZBWDu~y(~HyV2Is*TU~;3N7`F59;_%kyR|m_5IIhtN4%9$)If# z^#Avv()XtVeG=&Z5qf$N`>G@O8h$_zQLUsxQR107w+L|r;?*06NAUw@Y(u>>(Lj;j0+y}M)9GN)!ophZfJWqw7(lV-VI&qf+{JXuD`BE@EJ4L0erwro}(B^t2cvdpUrcr zz;~r1=3t8oxR-8Rz;jlNuq!{+j{M|%>2sTrhF7F3f7x4nJib4elP(bw_nn*EeHJ#(|8B89!#suGEJZ-Zkto95a|Q^D;{_t(i4OMb?b0xmioI z-p@Lebt&t|tebed%q4K#Bqp%3VA%j!icBHPlFgDiWKP*8*{ibSvOBWg@({U3K1ZIT zxTtG7Hr%`VLe%_!aby2%-_hFrtA%#)etGr!OLIn&=b!^r*>z)E&a93ZQb zy(Zf&J1jdTyC}ON`&o8dCdkLi$IFZ5>*X8eFUz;dBNg$AWW_LrT%l1I6*-E0MWNzB z#SDd2u}HB@QKeX;cwX_MVyogUMT6py;-uoN;#)yOf6#Vl z<+|~@S9L2gR)KN*GWzP(`ce9P@U78MkvYfcFtUGCiZkIE6QXzI?TQT5O4U26v#M?B zk-AvjKwYwKh;Fz}u2bu>b!OcQx-Gibb^45D8P8=@di24DafTwp48uahGQ*RGy@n%( zQ-)Tp!F=LxCAgdy)CTmO9yICi)KFj(lt38V|i~5S|B}T)~wO2hgiO@`AGAz z=A`DV=Az~sO{=C&(+(k(Xe)bbL$pe*PCG#xtP9u0>r!-ix`Vn;beDABLsI|HNiuq6 zBxj7w$jLBg*fJa$6&bY|Bx75~6-aNOK2$#flToG@3}J@Bh6RSF3~LQ<8yXBB7@7@t z3;~(3nd37b%G{UvL#EUiX^c0HH7>+-y<%)Nx{X}cn5+qvSu3(OWbMxSGK(wYxHXu+ zp0Ws8u56xcp=>iG`(4=&vY%vs$dqz}JWpONe^&mYyg`0Jenl=(^izZ@k`x|AqH?&> ztb9dzP5EES+sa<51jbl!beif>l}oiw^&({9GgUuzIkA~+uQQcnKaon*;BF>`Eo^!GDBI0e@@`6@;^#X`fr*V z-DzEiE;8d!{bXZ_u@uXx+&Itpu<;S&>&6D-=f?js#$$Y36>cX{k6acbS79~HP`##Z zNgt$nK=X*^LER_%6P5Zi`p@)V=)clm(SN6xf>F02(e=jN#^1o8-dWLEDOvYrJ(0B| zt1asW(*tWcZWW?F5Vkq8XJzx`3*{dxK2!8khA1Z~rz__vA5&HPJbl5B7F_k+V1py=~vP@O`s-HGf-1GM5ESB)67ACqILh$^~{LO&}NLwD9u=y z@m|K+jBT06GCi3o#$2P>IM=ww_?~ep>%>-Z^XIIxhh>$rmt?zS`($bIb@G?wVTzFo ztzx-ii(;?hqT+YOeM*Qa!Kkksha+tKqcgwBvLSqqScuGkX}rjDw9j z<8-jjX?))JhVcXADdQF6&qh8gBr7>f!y4SlaqRKP0oiq#MBWFQEv6&r9o*@ zW-IfQmHEmSlv|WrmD`ltm7`UUsj5`fDx%t?dP()FYPae;)sL#5Rq1NI+M#|-y;A*@ zx>mhjjdx+`z0-r!H>JM;W*$mEmY%NBWBCE%EVfumkV*Mh$M?VefXJBSO zV}_9!xkir5#^`z3W!c;EZ{)wqm5PaqpA`d@W0g&i`9RfZY+L^7x75egKdC1|L#$2T zu6rdzW>{i)1v;{xX~%Ooh$0FC?*_=DB=$6 z`N}u2E1p;Ws?1S642AS6RB0y~P^(9)^VMsx|9!3YsHN$u^f~DZ(s!h9)x4(JtG%F| zsjGqRIHP+ml-JW#e_@`x#kW z3yy#2*$LS>*+j)=%+@=K1B$N{JC%EtA1aTa8{sN;q+-i)w%jV8FJC0L%d6z8HU=)` z^(AYv)@E_RJXehDR9T+PDqAhvjt!B@0^~{ZR45Ird^H5)IFv@ayi*iU&-Kf=>Px# diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/t64-arm.exe b/src/fetchcode/vcs/pip/_vendor/distlib/t64-arm.exe new file mode 100644 index 0000000000000000000000000000000000000000..e1ab8f8f589eadabaf3efa068dce3ff620a01898 GIT binary patch literal 182784 zcmeFa3wTuJo%sKrnIxRag-b335zGYCP6n_7;gZ%ili(!*t80k1Yj>9b+D<@QBhV6U zn+fzX7~3)sOAFm5Xl;|(y7dCB>5YKxZjfzPz*^h>S`*N2Lad5_WK_)m^L@`b$;l9u zZSDS_-}Bo%&*Yr*p7-*7zxVh1Ue1SZ-mJVzsT6*LgGxQlRsKTy|J(mF4W-5w{AjG& z<@;sfa za7w@bRv^Lk%`@LB`Ucm^=(fNaeXS{y>zijxeXWfCvrez7T^W|Pl6&@dS1EN%JWU;$ z{L%{N-79LE%J7Vxsvf7K0e<`~Eau;MeH+p#a_>QCG;#!yYxMnrkc9EJ?8c#(UM3X-fGQdz8FB*I$J93oX#CEKeD# z(*kwWBA2bbb^Ti2%={sR1R&|V-4CaLU#XhQ?!0CB+T}`p>poJcQ+HnUKMlR z&nDmOX;!}3US z^hfQdK5Dd|Ug>F1toF9Qu_mSc%-yN&XBHW;HGdasU-yx+_D|Q8x7UXkwcmZmmF>#| zp+rcP0?$?Cy^=DEC|gb)W$h=H{#%>W{kN#of!C{y0}aa4|3tc~e!@_@3RNVr%&(5q z$M`fvZ-p8^-|rvHs`2m1TA|GOe#O0#do_QW3YBmzarfV;*SVHQ)NyaBs+Q}@6jfd5 zk0i>Kar_CdsxHk8y^4rcH*l>nRJGs#+_e|;D-1;vg*A~xjlbpETlkeL&+%!4gVifM zm(NFJbNwlyho)tgJ#=wy`9qgZS@ck$^2CP9q^O5hWEv0MlIwZsV^h2jt>m4QU-1^@ z)#dc%xof4~6)G^_-^zWf-23PI@8JFpxt}EWncQdowS5;H09#YxD=hqrZTt$J3t!nWAo%ry zYd3y7!Ed25XfJfFTrZaEE87MNU)egalII$~x>l~Ya20&r55C5MYv4a9I2K%k%Y%Y< z!M{^ha4fhMoY#PR2iNz5YiUR7ls2XQ`@wl(gg)}~x02Q>Y5w_zA=3Ey?;!0CNt-mk z&`%ma(x8RP{7Z2*_=Iur%-sgExsYS z^~R6zOBo#8*6q!S_jq%8&Wj)QPK+-_es4s6KaBkTJ@Wfeiaq$^XSNjhy}xc<``pmG`v(W-@$TSW zkNT$idd+oeO5t^??T$})C|7Q(s#c{Agi<0+!fQg4o-vIF?!P#+exFjsg(>Rz_mIIC z2M6;d?TFL`uSyy9X4zILGY0P<6=MpYD}&Am%<#m5QolMr9e#WO*o6-r*zYAx=zi4e zJ>c$x=Z!(y+j!tXEMEQgU~^F zEjm!vqQjM2-LMNThv;C_SM!e00o&rvd&Er#V7}kRu|o&xlfw_vr~An-G>8lg)8LOu zG#Ff@4r<$t%|5=%bv(cw|9Kp1AoA> zN_52;-o1KeaGS`eyDoWuLpr{}n)mj~caK)i zT~9?;=+#zmYL=;kyY~?HzIpXlp;huvw^3-fksI1l{>I=g!S9SygVoZu z*vGvmEn2wSaqgGF*SAel)r&4t@mDcF>0behmVrz?}j-ssI1Ob1pa>{rlg=_sVnO*~aHN>2<#N{%h%F)8SnG zKJWhimGsoIgS|y=4!U_CJzTvEo-f3vxzm^PEa})T?c=Y)Z=`HY(q6Mx=!Lxp75r`pR?=w6ZNE0-}CsEM;juC zJdwzu9_pOsSIwr1WEUv3>?4uM!}afT)eqb6}o)8h^U`b$UPH4?TjC4=Lx%={GoPobUWW2wVm;?|8%u; zV(CvU`V(Jp^VnW%8r+xf>Z`yi{fI`vm+b&njIt zdw9MKz_VDLITj}S4RTEUlTthVxr_EIxS`W2R*0-UxYLmyeMA{L# zm3E7Dd$~b@F3UgyJ9>a<6`^kNe^FY!Lcek)X@!n z1eU0VCC`Q>r|G@ZRCDiluPl9FYD7M(OVo+Rg*dJNMT`fH-K7z<-XNoE=rkv1N z^JZCItY@Q&2|r$12(O^$=vz==dz|N`UNukdb7*7vhP>Dj_+}z_33=46-yr|J*`Wmb zK;Y#$95m=VSymMtnU zRb52<;aqh?K`Q02p8-R?qBfN$YQTfsd(cxA@V7TRQdMX~s(kg?*&08-P@;0Y>JVJ` zX~cRB4gq=jBJXjA9O3nive54>L`^p?ERH>d6s zW6B1c`km{ErS!w@i|~=Du){rhv0ivV^pVu_@8Bt-1{QoAnU_3LC%S7(H+JfenrAE> zbYiIs{!#QGIIr4A9Y^6&;TeJJXFQADm9|1W|A^<`^SmcbJtA!hZufX{`@fyN^z}%l zs%{}&(po%u{hueTJ40|_r8SUtOLq0^A<{HHreY6_NVed?c^;5et0m4^sB-9r^vH*g^CH!w19^d^e4v~ zITZb3Xuw3z6--c%grP~zTZ^_{iO;cyc>>AjMZbwI7drj}Jf-_kq+^P)R=?#ZfYTca zvQp~W9GYGgN{Fu^I0Ek2vv0twA1$X(Wyp!83lV?($YJ|BxM-S=VM<&=w^12N2u}%b zz)M=!L9eGK?wlPnxf(n3Vx2doFyCR>AMrzY|4I&eLYL19B|2}qGV%OI6%(Byey!*Y zfvFY!-8((hA$(a-=BWz-xA3d3*O(alH|h#w!=#+l;gt2I=FZV|qjPGrEuEvwG)|0l zP^R|*6)X6xYVX!pqrKp>9)4baDg08~3udc?E^k=;*8A~LM?q@-osEn;bowZ=CiHT6 z%gG~h?(mPux6tPo_&65)`LVA7i|}Ztlh&9Yi|MqS_>Y~m!b!1v1!iD4b$CqsiEC8C zG$zgI{kZCoe)N7+wd?k72qngV|1^HsA?!?~O2!G69m*fF3lWc+r}NKq$xrc zI=ZnNyV`bhSwhOyAbU=^t=QJB#B@a-PeId#ww*T`FN|$4#P`XKpOi6z$b>OPe55)V z`-tC}3+(r;P_b4|sJ;`Pm$qK^VXwIt-WMA*#P@|$!GCUi3}qfm$5*~Aw4OfcGTmM? z*h9Gx@$fKpoq*qZUGIKp)j2OrnJ{^3$Rlk%ZMTJwSigokg-`cTzXj_>S{LoXf7I~e zL-wNQqLeMzr0N>MjkG22x!7XsO`}-~Xz0=>r9)-Tkv&IlZF2~)UMwBjO*GM!Nw9s|hONt{ zMS%Rdj8Sx3m*}xXmCW}*_d~*CQOf&16FMyI)zJ27-g&8e8#pcU5f4qrKfSAHeUNci z#e_gf_XFc&KLTGCJqm~YX7IS4^i|m@u?A0ie;)PsqI1i^yDzVEVd*?o2Y#zdi!5KS zz)0_3Wc$u}*dO5oU1#AmO_N)xOT#=Xlz70CvxN60U+|+tRx0>3CocTu|7pvx%=b(a zSc33HozV0~%7;JKu(hDjJicLqs_vs)L2c3c6I^@ej-+QWiJoD>33w~qIXzbJW$Htg z%BU|Dyhytu4^z)oX8Fajf-hKQij&HWPb$;1Na=B;_=o+DUM{<2NG_xw)2U+spIYO_ zxHMMawd~zg(j`sVJTE@_s~0hspMrl54IBOFhK*+I*Hct8cznvdu`D6aLi09cq7OL{ zn(u@@W$;+*1{M2}N9&P?jlCBQ(N*RJ1rCGuMCN>+$RStm~@~Iz3pN4nHJIAeR0e-7NxipeZR(esBIw5VrrEF??-_z!;69A+0iGWA_Pi(dU)1*} z+Ij&vBrddpa*g~1|0~HS*WaU`kgaCLU$&lyVO<3CFZeTw5IVDlwGz&}ilh4#Jb}#U zSec9o?ED?w-cW}-zk595r>>hA6MWxL&ELFOHMesW{;j=4HRmX$eG?p?W^;;K+U8Zo z7YJ>{=Isr}ZYFOFzhI$i?!@l4a{ohq5pb^a>@1AcQunja>^JAKlcd-4yOcIYjK7BBgw{F~>GK!t zJ_VRhL`PKZS)@vCAuR{EiiuxF%df*HLKm@CBTU}{BbSnHEYA>s;3*G~;A~qI zDybm87Ke6Dy3oJIgRkt( zj?aWP_Y{`IGVrx3&MfN+8*2BGGrq2Nk#AsYFwB1JZMEdl@3B?R`#taX{@NTI+zsxA z#^=(vGq!v=?O|g!S4^lc5g(?A_eDl@a^-}Qk|V(PL*ajL1b#KIN}Cs5YNbW(w4aeD zL7lp8cO1YqKG!_vxvNBQ873COxVH}Zs$NH5BSL%Pr2&bfmzBgm4?nk3w|?GP5{vLG zd88jKTcs{*IwFmj7`&0s=B{ctf-dEzl@j|{V`G{9G zRiQ5*#`eSl2L6DNzS>~D%ghgiI)Zy=Hf6-4RTTzf1|ws28DoZu+2omJ))5b^&QQJs zMdS$-h3ZYi*Oa~-`<=yoYUI$&B6CMPLnW3Onf)PXRGAg(XecY{e*k-2;R|*63}1#k zZ{&ILPW()8NM8=Ef`-yp@ii(gA-1Rn=F4*$&(a^36z0ma<3|eriT-ozIg#s&z-=dT z@tBQEv#=yqaHge8_`Hna_k3rb>*~h$@MC{_SvwBR{b=jTXSRe1DtS)E2@7@S)oji>r^Yk%BVZuJviyxZa=zFq{G8KmJNz=%9()ZKQ zmQrAg8tNmav1xl0zqQXBhzqWLsS@LUqRrvY0_LIT-2dFHZ_H3?u6-3+3w;IWR_rUP zuTox>y-(}jfRPfj;4OGYr#HTDXgFB@F@F=c z$Y6otTU|&T#k?ZaAu^ZV8EC3R?n;rn<;agCcZS@50k}$39lTjxkq{XfIks~5gMMyS z_$Na9K;bZ+-0?%P7hAC(MP4;2?2$hER_n3QWbTRm6w+qxv`|L{ONoh{FDdY;I`l<# zuc0QD8s_R&;1D`#KDJ|tk0O6L;B+hF;NA9pKQgd#&F$U()U6k&&^HW=pPAQV-p!$t ziQi-{zamEc56^47B9>FgIE!C_k=HMNJqrk)T1)(B7xu4fMHx1eaRs=HU>_xpI3539 z;IQV|-1B>*w6oK0XYae$j_`wd!BrOD;0ro@BXpj1g(Bm$qNBe|x~l zIK~`RwaL45>e-aFxJh(ufVL%n5ov+Xh7NDas%Y}FW~5Tuv~1p?tGRzH%X3{Oeb`JN ze84YuC|ltF%9>A9(YCaI9d*5zUmtyuepCX_V)`O*Wf*_4d71TssMKae+BpSSL06fc103iYQf74@UjY?1{t5em427IccXIMW791TY1>K=Gl`q3q2(cMzl&&|+XqW?4CHRAsh zgdV+%RmXnn5}Jtpo3^N^X(@P+{z+Ye1F5IQ<6C{q>kA%bJ)h(goBGCd<^@NrdAHUa zi2fD5>*(9x;SaifT$y`((3RH9{p+Lnw@!Jd><`HI2Yy!j!*Bl#e_l8f~8B@|Gfh%mK4;u7=FLMg>H+zekRBk9Skv3&L z^;mu;_A+Yt=le)6Cf<#Iv@42T%kd43t*|ZD*h*rU!jsAVhhv}o%37=D^fQ}w6zgvs z+v=oeN_y*XdM7%=WKO^|-cMZA><_^MU!AOS9XsHZL)Xe!a+gzI6=NgFVy$y=*~5v{ z)g^W~Rm~PZQsFx+q_17<@G*B9w%?=U$q$=*QzZTy>?FVAzey~A5A$py<_UvXfV>Zcl-6sU%yf#w4nZOXrb#hQio_@ z3LYj~aTM(fx#n#hx(iIo_<#lba;vPK7dY8irTY(D>Ex-Xy}m=>!bhG^Sv#&H89s-f zMw=H9o;keg`i{fsWN9y#_5{|v#?)9oV>8AJH?+U?Tvtt1{tjugi@daR!~Rv{I_5IA zmUadHc@r z@7%sgv()1THDp-5EGgLS3?^ zWR>{Hj0B=7CzfuF_O>Cw_kQ zvJ}gYmi&@8cq&v9XKzZ6cZ{B^2VpWE%pH2)(^+y4@Gkd++F$0~I5n>6$ z2tQ_+?y|2V!FsrVourzXGY$5FAHmK0!OdiF;{!M0yYCuE2RAaFoKcv*13hu5lHYo8 zR4_)>&Bvz^9EbH7J2x&g^Wmon9`Sb%)PnCYc({?+LL2KCZi3bk##ilQpV|IF@|F+L zF+X?*{V!0tdc9-~_-*JJ2iN8WX-j*N-=XJ1Q^$_R zg=Fqh<$iKnq+=K-@Dex)_Jm5(jq&r(35#JiIsNX-dS-iI+%wyWogDBRxzKFiIqm$J zyMH`1pMQF#mwhs)@LA9uOIl`NJ!qqtF7MG-%OdV!RTVYo~?U zCqqNQ1%6zR7{GzxnaU=!!J_wE=$!|>bA{e_tsD3-vGPsRChsU@jwu72Ng9hhSm)Nhpsj(1iUqmmbfcvAs=TRG~dZt@33k8iV8cc)+lY*xiKh~rS{IC~4 zaS)#=6W^hayfe!JO~<%qP{(N*Q={|7qL;tS9FN4Gh}l$E81NkPK{9^EU)R1{MJj2` z4edn^7N@G1`1vOMuz!E=!o~O(_-?TVBfnqX?ZtoVrHu;uukg?9_|ulbF`?K`PYiBL ztiBSTe-UF@XU&1-Z&gfDIleRA2|?<85j-q|*8(c@fC)_K+*lC#YcY)emp0zBUEpUN zc<3ekg6Jio^$+o0;9H&7w0(~;wqN$-)Fc`X<8$OnnZAwD?eO9u#$Jc|qzrRj0;`l! zxsmP54Au`*Mr_5SZ^<4T45xjq4p{p#hU|xH-J=1$mC0O3z=~I|LGHw^`vpIt4Tsix z%%%p`_C)Z_*6mZV4>K3R8k#LX*e7!q6LFo1Xc6n8wv)1uWHX1QwG|q{$b%B z&y@8)oHk|Q59jbJpR&-LFg3faYRbZ{eZi(mY@PVA1qoGBny2bk0(%>E78tqxh2UTM zS_Mp^*R_2b+Ry2p_Z&IS7yix>o^Zh}ev0@*!h;jBiCx%2@jDc@aQ8GLChL|2Z_>_u z+LN}<@<*I?BpPp)Uc}c5?nw{Tt)%Y%z@H0%7ZZES$mrwgPz!+jd~StrzmCzY85ON$kM1NjrkbJ2LQy0q=Yt z+l)@FGmr&*h+>J0w<3!T>8z2X&o!*=5}6Ef6?p@P*;;QH9_@39Zr1(_^}`!gQa=+s ztQ}Fgx4H1YjC*Wr2YK*G&R%niSg7;DOWC>bV$ok{(ET3Pj?xYS(*ZrJ_T;M$p=X1~ zQ+E&j%!Q6$``wy>`ChfV5E@2)2MzJ3^*jW46q;4;P3>6svB^u}$zU^i+Q4bL*i_2S zfL7Sis&EhUN?ec8p5z;iw}j_^gRGuJCU+4JZedMYD>RB8(Q`QSWy~bwAdORK?u;4a zJ#?Kv^p1Wz?{?=f_C`jCPv=KH%#A`%l<2QPW(pUIcW#IwgsN5+D}_So9})#ZCLL2 zazBFaMp>ufj0;#7I=ctn-({EiiL0JT!uy`#{^K{Vz90E`JD#0^EZ5TiTKvBn`i~yU z{b#%G2VHgF#2Al0Y!SJz;#WTzS_d!t>>c=Wu&LoA*K1$R3|aFqtWm716}mnWoy;1A zL*XOiIu!ofPnjc|BysW6W7xwRIrM_enVr5y<~(sx-IiN-`tUv)OJ^}I zvi9V7J;7?|a}m7ZjN|?j-&yD-6E@+sm^whyOi?W{o< zZG8D#SH1UGILQ{76hB|at9nc^FKa12U9rG=D|C0puixWd(r(F9%}u!y^BI^gWn>Ii zLyWTqdp;HVHcXWDTsiSWI)5(uoH@Y}<1tz9Jv{z$#@&y)U|VU!CiHU34vj@-XJSv7 z|7U#CK^w1u&pI2oPjGeMy;G;<245b5Z>wD&^B28?EzIwmtLJhLKC4TOl)8P$l9r3B z$`Zc%+Vy@TCtc(=_oFjr1kHnaqzcZ`wxRBeWyBMd}+Fck$@GZ@cj6 z{LEhtvW`n=>mybowlc!FSy{Bzeu2Xa?^Ms9?RtLf8LsCgEA6a$WM7uEUR=hPGPYdv z3S-JAvJc8ye4)!pkxS-02E3_SKF7Ft^>N?>C)uU&Ni{Msk9n6K=qK~cYiR#Bo`zQk zw;GRsJ3aLJ;9+BZFYDn38&pZ}N>$SL0CUZ>cWNo?p4p4n`(aN)_D1y$rnC>PRvm*+ zqzw!%8Z$8HHR=bq^1RejGI%|CmU{YvR@z>TM$Fe?9hRTHlwH~Gm3fXJxQbl{&PI+KEjzOoSg;w{;Fd| zrh?-I3&qbLjSdsufesJ)hxO!5HXTBv&|$*6MhEae=Iv>5=&z?m{aLiQE3jT@@eASx z=R=DFFaHI!NP`wV1?zALGWU5!T6fjm7T_t+*YH%O5$A$2J)dk=^npU2e zJJ4_#-MqWOGyeD{@Z;5_-c^Mdn=bHKS`6gdC!ZQxwSekqCNFUwSQ zS5CV2E*ZBkE4Jn=15?ye-daw(+9;vV5a`ul9 zi>c*4fGx{VpLe2Oq`v!wGIOv{p)$6;wJORCS`)Z6Bela>9pLJ$=!X&OE zYoaoVL4}C{WnQdA=T|Xr7dIC&hhWf$XG3?j0+UJX-CU?T%JHKlz8N(4O|JZc=fWQ5 zD+7Dggp8>wr{M)v_hsM-=VfIOvz#!4H1=SMZ7T6`EutMCZO9oVo0vQE{hjJaC%uAr zkV(3OhY|4l-SCzl4y2k*nXJElJ9xX#2i^~;r2=oe3tr-1IZH?vSi{^03{@}i`)E(# z{x9H;64#WqwgS8TE++2;_7QN8g6Aw7&spRjtMQx_zaAgmh3nTlU9evd%>>so%>Ue8 z`-O2kWR5ieT+4*7;vX>Y z8R5Oe2pT*SSI@u(iatw@qn%jl8n<3Jtbfl1>+Qgrz?VI5ct5C?e&aILTsqyq`Y(Yu zaUpxlhT@f$jkomh@P0JMX5cpkTfndQQG#DNb3pKWJGx5N##?*OC&uSxsb=u1=Z1%I zl)IXB{T7Z!i8l<<(-mKN$M)#E#C1#llJ;ElwlZF8SfNT9;ei1A$HU)I?K0miKC#(Y z*3pIj;)Kav#9YYE$L{(}wpEc+g_4+^Yd>dM7e zmodUFXdg}sCGxO2-Q+V!-;eLK>N{oa;J9n4L+gj2iK5x*eDq+#W!8#N%UUhN%TMFZ6Dm4((wvoIq~JY7nLXOW9;0! z5?_)$ZQ@IMNBWZ1993R$KQL;)a#42d>-6Ik^iSUZIJD2x?(Y}R(ckr>^!LkuO@BAi z&wqdWJ|8fK`}~pf?DM^@K7VwSJ_lrd=VvAU)mb9+xR!Y!c&xaO{WQJIHI%X*M`XvG z(z#I5!px;+q%gi{`o@2W>c|Uc?M@CWUZjTEw>^W9!f{4d~-?{MP7L_TetF_7Exh zcMRFYu69Hy3to1Ag>n(fm&-d;yt zI^Lwu=^SnS=IHR=aX#>3XP-=hcg5cf-jVzBYvz|H{Ap?zlYDx;RYe;71zw`5YFGHq zE4A&by)&hw)W;miSUoqjId?Dn+NxAX!)2=D7W5$YF}5lg>Zk>O$1(>N+?KYtB-&`t zVb>aUz2G)V{MFZ}4$ipXvxo1=Rc927G3dB9TjR3FZ=0_={(~`hcuuH;^?XZK@$54$no~PRbp+^}*pY_!sgB8Z z{qSP-@*~JG{>{1S4O6f97-7nZKOTM`YaEGX*Yey?If41}^j&;+q?n|t9)5zhVTpVuCw^X|Hh!^moV4b zBz_^*#X5nUIEyDVuDUpja5;Rs;(qM+{S#s){Qo}S8;wtjZ9IPzK6x+q$$WB{SX2?| z=c4EN@QLt;@QIAOrT+g8pIAJihImBgz@)FMT|DwNn@6Irp3Nh*ZXN-*QeVmi*W9)9 zmBH#{y6&`j#w}azk9GXy(9)vHPk^0{aKDoig zCx5j0q~e6bC!zYZ6JRiRJ&V|Ez<-;~# z+%*bcL`Kmo%aZtFF?!`M;fqZiq3Y-rH*bhu5#A8JV%ec#-Vi@tbjT{~31ju3=#sqj zch)7?%~5&eMfTjjo4VxP;*o_e9+{hJ$wKdtE*a*L{h~`89&zas1HF-~ON>!?=6kj- zF;n)I6f|OwiH*p^7@2xHT+vlD8AR?I?dT9|wBZIj2DZa4%6@mXrnsbTn8w~-Vt zaoXK)w|f_3TbX++D74}a=i7F2yV6JT$)pecqQ4o_T4M&rt-q~VA2ud1pRVn5RTq1M z-wrkhCr<#|JK;q8ON>V>e~CCHxZfyiz-=F=SkIeTV}kR+pZFBgkJ3^5v3x{7dKvFI z{ztFLI9`0WBv@)(uuOOxSezJ*54kD5z|iXn8S5mXnUfauK~vvM*0Es26gCW-ky~0~ z=(S9eC!8^9fsZkifu?DAz{Nj7H0gKG#mp)?m;h>@SAC>-rXJ9YwGwNg?Oc(inO};4Q z4s#XUvUXX+tNTiMcV9R0Uih|_s}GnQ9ul2X;P?L87{9@9U?9(jewUzw--&)tvyRB& z-7owv{i^@R>4zP0%U_#I{=Q_>Z#n$nmcOV={%FVChQ@rM!eXR>#!^&RA|XGs3A zYmWRO7mobNI`v9;l5g@1$)EVOB7Y0ys^yP8ghTR2J?E4^-nr$EbJiUByM(&TQ!VV( zxBYWl1}l(3lf9A2GAQT0pN|YaZOfqSXC5hoz~acD-QTy9!L#J=ZRmIEJm`0l_3DoN zJ$TyEpZ`n0vvugZO~2&zV>(WbEly80ZdmpQ^nMQUy&N?@@b~e@eVjk-#F3o1z%Ir; zVk>0a%zROEp?&|cG}U~OeZPZriRVbH$7f8@F`hEyMb`OAxeWYbnHQ+#eL(hd5H}1M z7t9G03u0|ym8@-&HAe=%#Yyg^oSggN<2lUx+G%5oWF4B+HJNtB7p%y_*YT82B%(hCu6N8#YWzez2sfV z7>cuCf(9_^cnrQyvX5udPsjhn$J?!aO(QQu{9NhtE#;gouIph9AFxS1Gx4i5J+iz@ zrF?*WNyG={4YddSlJ7-yt-}ky0QXXNfHgLn)|A2DEgmXk!}uI!#GVVDl+ZG1JtDLn z`~tR}b%**Y&(2TM&-ceZS6$wnf1F+Z^eFF~eD3ziuAH`=bV+maI@d|0nf&O`&^xDn z#!f?N+1$(DJIP&?^TvGH=a#!?kSRy*K8b!Uq-o**D6-_p-A_q>JGq;M?7mxacZEH+ z?cz)(k-LhF|F`6hZ+2`OO^#keCeBBW?sUn~2hLNDZaoh`zSp z5W2rzY!@(y?F!Tf*SFZVt0mcXWvH0EUp!>H(zWescn90H5x?)RvR%L?HtFBbQ|3NS zdyeheg8n)m+jTQ#+_vk{7tWrS7rx&JuL|GG)h&1Oe7-;7`{cYLcakP@m;BCocDGN~ zv)Y$5C$DpL%bk1 zoHt+BxNP7QYt?rVS`#Tz!y}6wFUdd5sXnS{xtIibGChsK8L)vocT)Dy*O9ggm zk2O&7=_wUW)5!C2_WsM-wv!pE<5v3gG4O#;kvK3Nc?Cy4d;l4*$-0Rs_I^2hR&y!q z;*irJTurrb)kFWCGMij^zfRs&lvzUjx`Fv7C!g$NUq%0X#$-MIm-pdI?D+D1%pX%< zpT~+Xhm~5;K%CtvyLxP-YBlRJeDzbZWqpX0KAZ+t4bmv^3;XPA2 zUW}|2YI*x_@CQ75O#GjB;t%qjFMqtq*#iGp@JISEe=K?j{wNdvaB%ew{4xG-#UISI zzbpKaw7%7?V-G?*kx#8(^_jNvEvS(?dS&CX^&+Q@E_p=fYgzOlgSUuGl^d}rXUqGk zL+fbPRmt9Mw?3GOKG18YlIjmPKDRK+T0l#;59<)Iqi#KN)_i5sIm)6J=TT-II)6y# zB(loFEC2~yU=f9*)Nc~=i;H-|K*{fb(`?+dGQc?_^VdINA>66p&Lix zq58Msq3JFj`Z;CK%|px~Ci4($gl!&TZN}L=bd!sR#<+NBwEZ0qH*S9T+x(sS_QT&! zUhgToVKn(gZw}l&4IPNC7ab+K30bPH^^Vj{tAddZ&J#)a;AiM$DqhpRg4xv$qY zvp~m7;7=| zRyVQ7=iYPlS?r#(&cV#J_F+XQ*@; zV`6e1P%USvWOBa7%)-naGVk&|!xxLb&{-mO+$npmGBfNl|3aBC=iK!YUu1q_Xuh_@ z<~gZXcv5&z)@cab>G%PaJ45Zn8T%W7sbYDkqz&6sp)6UfRKshnD!4bvugEi=y4-oB zo(dm!0(@2aIAd3y!KHz1$lzYu4f8DTYmMRkk^ckjGpFckUt*rF_LJk&--3sR#-&wd zB>~Pv6TO7rQ};{O*sR0$<6tET>s9P^)&cwZ=D-ooOgkcb zT4X(&Ypt0nc6E4-n?2X((Q|#GE1dX_)F-jhImFkbeK~tp&fn!5HPyo3roHw}d=IH_ z1$mYO+xfO7zMj*Tw8Qx^%|as?o81ZylJ_ztF2= z0-QLoIQ`yCywTlnIS)?8`%b@QUR3%l<)q&$nIDyYY@nRekGZ4tW2ApNe7;8*S~E8{ z&*{S~&M0>Jz_$pq2UrWJeP}sXL(Z`PPht2? zk6infbi4*!LGK}4x#@i_96ilCXa`5Pqi1@+kF!Ql&eKTl>-o}qp$mTZSv^B^cI~^Q zeK{i|C}(6e^G((L5|`7uFTH;+{%`=jW(psR4{YRS$oe@sf6$NK+J^lr;B3OY+1s|u zp3vw2Fu3hW%2$X^GxGR$fJ*3R(Sh=e4%E-o8Rd*4`QBxOHZ=d@JIMM4X{X&|X4LYm z{o?c7Kl*O<`r_; zKke1v4{YtpW-6`DqcU+lfqBD>pd z7zvXe6DAIKBlTH(?nXcFYV`B$t=O}ix03_xvJX4~u1>pfb+3)99QxA>uYStL z)q*i*Q>izUM< z!#fGi1$y0^7GE$%?b;1a#Ls^^&HDaN3AkX-|E_9&^37m$QuXPvs`&-)WWHG)>UiFJ z0cSVqvz+4hbB3FoCH8w@k#qF+V)I@kFTSDPgSrzxsh4Ng+H_2#-WTk8CMMP6d?RZ= z^@Qy+cBH;C?1P-4UdZ(sSG_$vAL1GRr$hEcM3^H$ACyQLgLembSHT)48H@4Fk0E;3 z+jtipJK7m@!Z&|QJU$34Z->Wz7aq5h*Nw-o@$BGnt6k5z@c3K7qnwWeKIDuXHy%Il zs#oyHexqbO);oBt=beMcPjTN*%|V#yjRJgRii@(4wQvA6WWQd=)u2bs1;T z_*q-*oXzI@y2N6At7NUelU9&gaHk>9*_N-*S-{nD_Pm^_A?H5!%6B#xt0kY|YMuMo zSyIH=tl<|rOM3B(JbcqmuZLuO&_-WF-hm%Vz4yM=aM;%wIMC$bJ6>lRIEUE!W<^-v z#xe;Y`_`*hHb*p1#3tGFHmt9B_HQn=|>sjFsd}c-;r!&}Wvg4pGKMLv5|& z%q~}3S-|1YsYsvO=1VU~wdbGgI!ycQZ|3rF|_hSDXm}G1YjJrdOFVF?`z2GLu^+kAD zzNwPVb3fNU)|@tQHedwUHq8pH@8^7lTIf?to94{W`Y7jFN_o~k#VZ(#7TI;kISfVA zk;Su=>DBdQ>3X8N9_w6&C_EwcgsDg1E4Ww@8&B^CwnS)x_5B?OPe*`vwF^H}=okGt zBxBk#Y-|nhE9g%zXIaWQ9!?D6ZO+;2cm)0l<2NY2Ga@o5ve#?NVG;2XIYU`uCqcyg17Fde%2n@E1Bn58^Xa&!RkZ z?T~X{oO)z#KzRoiXgD}OSk8vsfZV>Fj&R^x%$aw+OC?U=EfOE1m$Avwm8=imh&^4S zVm}o*qW*DqyZNqmyZRkHX~oP(i4z7Zk71AD@l)VvIq=IFHdCoDeuQtQMgFr zI}1d=ihZas%H@2H`F`@2P+oLqi2M26V_W;X8F!0*ox^if;#fVy-#}rp#XOzn`YsCP z(aX8PW8gW8AKLD*zTJ=xy#joTL1J8$#H&hsR0;k^vC!hjnfQ~?$1MZNaydK>we*a@Y1C9TQco+TN{UmicI_z$6b{H9*1`de(C6?lc(x=2pcu>~26j^p8UW~un$20q! z>chZ9P?b-q59Xs=_4!!yiCLAw%W@{qS9%AxwNghXZGE4%qBehRwXV6APct$83S-(D z;v3Q?errN#A#mOdoEHC$JmWlt`n}{~?flUBm6fcw8HLVq77qxI$LI2GpIm-Xk9?0e zH!iRT?C_$Yp!3nzTlLlzpZnR zzzOG1AiKr2UkMGo1Wv-_^D$2JgCm<>heVc>=_TKhmp)DB8guE>Zt%9;MWbw+uf=Cm zlPr7S=my2#Xy+~N{ypDI_Oea_960BZo}}I-u6o~uUqwIPM!lBbkQ*OMxdN5fzY;lU z@u+&)tI$P%)IWz3o#^I!Set{47Wbc)eLWf0zWZC*A1!jzs3t@ya}VF#4x(e>*Qy(- z+vz8?%NE)je)c zbW%;ZI=FDW$VG0DZ(ql*q7S2;(JB03$p_!bb$ypp_&{`loJ%ZvwG^N7$VzLBB69vo z^lI+|R$TiS=i<1}1$BRqO6r%`@8Dodw0H2isKi1UB!U8^{!AMyz>&~h#*J~l7ct}F z(E1A2o)o|%mGB2^d2hJ3ans~l?jT;EH1|nD1U2kx7c^bZuW{j<1DT) zx}ux<<-8h4e=UIzn6k-sZ0K|JL*xV+aC~R-cbMJhnpw6?bflawbOOAHZrtdqyV$N< zc#JbiIbU^n9N5EnQSkU@YCLu@Mk2DmxlO^FY{$G_*- zf!>$}UY|kN$hh>6*gbdzop|nV!aYVg;(%LB-?Q5ryk`FSB}PP5L}_Nfh384VZiI7a!t9My5``$qfg)G6bNrgXJS z;B?Ao=y6cSDnH}(PcxoCXBD4Lm-BZQ^p{iTY2v> zdK}r^b^o)2+k~!1h!Y6hp91DS&WR9Q>=6HrdYj7^99Tv9n-?u$Jx>0Tin0Ywtmn{U z0QBHtIp?>foNpBqM>xiJs#-kh{ps-ihZ~zG?}wh1$d8;I*_*z9vhNSR3r|C9fol

!I7S{BS4i1|@R7s@ zBh>dJ`X=v~Z#j7L7qym3O>1w@`}Ib;lHEs;SuB-d~dOAuxs6o zj(1yQZ!1o$-ceA_ip^cMNbJHAzA zF5#OchBvm%$mo|fg`L0}NG*Gu^P*QrkQ=_&+t1jl8hJX@o2ho?5x?BcbwBFM#siz|d;E%O<^m>3KE=JyzULbu z!9rpT0nW-V=64D0ba8!}-!$&ST-WlG@kRyLYJMUoQLfGWF5rHDnyWqgUbko8>-Ow> zX>TX(w@rvt?c(R7d>7ZJ`3VoUQRgmx(%%Zws`-VH(WuxbeV)7Zy&mWrRQR#2e9Q4u zyf?`&I8M$dwp{Xj8oYi-^R5+}NamrKQk`Gs_Z5C$<;T2V=N5ij`R($k&fWZAd)14b zk#V3iZ*4s*_w#0%zwaks7#?uWjt_FC`!YFG*#0)KE(g3$J??cPygRE7;bGQZ>oe#b z8!7rOI)oSNyQ{+AHTwIOBfldcb*2MX262l@&YrL3>{@f?{_UJS|D*-01nD#~1loCCNM{vKql!)eY^z7ZXSKF&3X z-^={Fk2=8buIQ`JEyNEj!CuDhs>Q6m4+0nKpKpM^yFP8>!`*lB_an6H$3EVNEh>m`j@YD1eU|hR z;wsYL8p=r?!DXQff70d%oRi1Fp}P&qM_eRdU}zWxhKG4~6Yrjb)`AlUA3CNz%2~JW zb7)NU<`(((;|lZ=dU6Z4U7y8l`u1%vCH;QzRrDFpj^1(pIC`p|*if}UOEne3Zn!zW8COJkrrg7Uc71-$!*e@A^1Twq51l0y&|xd@pWn#Yn1e?9 zQT$0~>?|;f{dw(6W$n{|TkMhet0Fh7Y2D2kE<@uS>DLAHOUCb)1Ivc76Pu(?sn4=y zIq?f9i_OsAW$(o|5V>9Zh)PIV6TUgdbH(2q9kLfu*4oOq5kwz+ls-s2(P}-`cZYhHj4t*Pua$G%o@CMJ8|n?uZrujTn^9aF&PCoVE_ zd~3$VN1)$TpYY@3;ZJ>vJY_bHcL0ZsIqqYOx`BFLz~(%UUm#;fr(6x?Ldcsu%Q#{V zYu5#jk5e{Ej3N5w;BJ9UY?8cpci9c&Yb$LE z55qgViAQH=>bf~+lRl%1befDw1UKp%e6!{&_ySfx4{f1c;it{|dam!pSh&^gZy9bs zmpXrTj<#eQy~^1T4T)o>gw zQ^x=aX2kAIO>c@C`a4V-cQvH_k+GlPZZ0w;xRZK&!JYX5&bH^=#?j((a_0U5$}Y3< zXvt50`~Y<&8?~I|HG=AsO85@zsJaoo7bcV${dGW6Ri|A#1k6kOY&d3Wo^*j%K z+`3~8<%e}gZoH6nyVAFI>IlO-Cc1-XeKv^b*#}7HIa})t^z4bH?(>b|1$&-Q`eJ^A zz9{CVh{@H!+ssqyez|${tJLSlZ4=Mh2A=iZs9{}e-uKS+vOsY}z1YpO>vipW(YO~K z4!*$0?lAf?j2;tQh~E6M#Gc{dt=O#^ELHHoGea){c4_de)>IH{tqVQe_v8x-SC`*E|Kkw@9Coh+qwGuIlIsDeI2LIk?rzb zf|36C+M#}*trtZ00xItSXR0yJyXf#9kJ0}c--#8u`4s#>VCSh_UknZGLTB}^C`-Hy zkI6SoBl!K|_wl`ieghv=+Dz6@$#z(DhRjXNcU^_wzDC{asQcH{-KFcc##=Imis0)= zKB-&yTk3bObyNIVcyrF`w!Q-;V|_m{+^1!&I8T@syAQ%1`-hXG6A1U7=lcyNF zv`xFJq&H2?E23R!u-WX9@4vT1B|Z^_uNZT&hE2`uVs1#rIZGJh;InNG?38ahrR%;` z>haTo?JtXd6j@_DDR{H}HSs47$=FHiS7m8>ZzN;xx<-6PowhSAb`N{eqjr1y!H3ji z^kl|1arH0HjNJ>}by{JTm1bncz9ebLhQKIg0`#G6T53s@v(u$NKH@qWmoD9K5qr!9 zAJjET=Jw=UG%{9Px#sq6jVCRKKhbN9pkeUHmpnx)uyfM3Sthu7u1oqO->hs!r~8f#1%@c?}}6;^t!QP zWHLAx`pH@);?gC(tZV9HF0mJXvv0b$y@LK!^6PuRTXdAU#3K*jziwna$2mpyBE$K5 z&uHQo|588ts9C>TZhb>m(&U|+UwQwOu`{6kW}#icoTMk~No+vWljQdwJ?iYEGak+2if$f|(_2fNvJB=Re+viem z*sjO1P2K4A`dDfy$P=0;BWJDYJ|+pG39+OXk>eSL^$zp7Uxzhip=08gt@E`kC*JI?u!U-Wd;j z<9uLF>*L3Ks(FvdFm2{Gz9)79-TVxEBj4kUVsjlIWdDZrSQworWAaw$wwGt*C;n0V zo*Mk>HRR>oh5S}@`p>B|dFoYb3(eOctK3`PI$*6qH}D^2 zOyKZw^HFO(Lpn4w`9`0N^W^(6W}8>9jW+Syo&+Zq$9)~$tWgo%JFy0*U+N2Gn)j{# zSUwkVd>w1(aaoTC9RKO`$WsC@~kko{2GtZo{FU$@4Vf<*#`A&>KP0UMt2C;p=Aih$K&$f;;>gB!U zlNihaSDv4cCzo}oUm2UzBrv{)419TP9xip{(97Pj@it@ZkRK*+_2P$hp}SWzZhG+x zYJhc5IuuR6XwcW{+;Zq4aZlT8>qwlNu z@rds}zcJnY4J{S#qRkB2L!XIXufN;*JLq{W{ir2f=vNCpzXA+5&ia)_Phzd5S?ZB$ zo-0_(m|!e>8vGA{+a2h;=4tAntnHO~wkY)%7#H;8D_8rd>l^5k@XE5pT*^9jQ*7pb z@=2RBXj9q>b4@3|#MnPX-aYJzkmn`b*BBppy_@*&X>S(uOr`pC?*ZC=&#<2+ZS&nQ^sDu4GF^9Ly2Lv-2ac!c zIV-nbbNcwK_;S);Y!~M>Z}Z#u8HU5sgZI2S@}1kileTi`hu9L&5$jvFw;-GH{BQR4 z2gDuZ`6ry)@3i68zm6Rh`uz6)pv#$;ciTtzI1Zm|^QyP5Oo+~vcxeTCx82se0es0C z-aQRW(em-o(CimJA(%k?cYG}TmW2zg|J~zkw>_f%nqyZczDt>gV<`z~ujZvI6A#Gq zOZM}ltbvo~NBJJDy#GNcu|uAJFl^gL>t7V#RFxIQ43!c)j8B~0PX`aO{uR0u3vSI* z*CZ;?Rbv=u3GT#iao2NvlzQCs*+@J(hjWU>mrWisiA);QHw_yh{OYt1?+>k0^)ZHV z?Cg)&uPgpyE@f80+vF=5xlXljVIWT}3Gr)Tvedbr(md+dmb3i7# zPV|NNIrH&v>nT5xc3&cp z`ZGH9Hu4A#TYjMGdOZbmPG?6a%e}-7!9jJ)zglD0nUpD}jO6d-oxrvZc=nVtABBGA z8)5Nix&1vU=EK62jqX!7hz}h7oM(bQ+x=EM{hy2*^t|I+!|BogkG(gKkNT?j|39A@ z0y7CvvI4<630&IA&_#BlUNeDK2jWIZ>vq2q%GFMidao(8mew{2xJ(+CPC=`nvbiLw z)mA`>y-J|!4aKEEOZ6@`0c$6u)gq-$T4{dI*ZF)t$s~l@yZrvUc|7v?eCB-4`abXT zKJWA1&hM?2%PAvXiLwh~=rG_x@;*?|xeZtfjuGH^6L3=7G3{&n!(veC#4K2cb)Gk?ba*0|6e z&5goOhVa40uDfpGXtJRN=n5}STR)$`<`!Jb(fp|X`|QA?*H{08M%Vv3^X5#Z{`axAUi24KsGZkS z4&Dqt^^w-GVk%s0{YiF}0M5Ga^A>mg_1V^zC%}t*@aAj4@}F%h@5{ijYmD`GJv{kW zDKDM%e1CKUvHXm!h3d@3wvN3gl3)(dQB=N#x}xkCt*kZZb^U&DF8)1hWt;Bju!FUd z6M5xS=Be9f+^l{y<#Es2A*=oo?$&8IZPi2ns^YA#j{~FN$?cU5zF^(SiG15Qy?!_M zGBvbMck*nUzVb%$G#r{ye@fp^&v47mSh+^&TW8l}%Sqj{o%F46fo$U68+|es`wQOWgF5mG3d>bLvlFYtnFNj+;J5>zVvl*3UBOSGws}u3To)=hm+! z{(Zxtxo-O0m9pP!*gC(Sy^i$xZuV8^o1*hs|`y_>t`vww$w>q zTDo$MNq2ApR73k)o%D@wT`68G7}#iF_z3)DEU0&E z;em00^tdljQJ_43b@J@ing6_#NAnOMUA8!ZSHGsbj;*a|Ahajo+f*rkp5>>UIWqZ~ zFWnI*S|geuTbB5E$M&-3Gxo|h>MVm+iXJGB^m*}()=H(tH$tQS(^{#ufd0H*bn*l^ zWbA3cH^oqe*WQqcO*3-Ff%qCH?rNv{LSB_o{ur@JgkPK|pdZNVklj_2=SuKiIRDgB zjx8MdlB@Wxa>^eB7f<9g8#;OY5S=^$oh+E~fB5--`03%N@OpCg^|zk{rv}}9PS4T% z+)FbH%swZYx5S~j~^fAM{8jJ#%_ zdq4Z!Oa7@B-RDM^)Bb$1eJ;WIIIH~1r0j~`d1v%^;>?}LgEe*gLsqa(JQMao*{k5%c0eC=Rt!AvTPc}+Cch7}-lNH8bBBF1^$=$_SyU6>Fa9>2&zBtS zFW6c#H~YFcX9Nz8Gxqp_dpNtY|K-|N$G^c(C?EFJ!}%U^^EFyYS(??xG{l_3|A{&*#=_QLmS8^#1=!NOB5K4rWF6U=!>5U_=@+9lt5rsk1+HJi{{# zJOaF50gU!whs!vh58yYYy<(>9ckzRmiY)ms=}V9+Pe-OaLw;R%e4@W()(j(4KJBv> z2_9E+&crU`-Vxuxw_l~T)9JRYGuKJ@8b>~r8{zz82mh0q%{J-JRZR=@+(-nP2w#S|rooj?VNPW0ws426&`4%9hn&Ujm=gicYm`p;a&1 zRfA33InGOOrJj&o(07B`n}^$34xYu39s1M`&wOHVsLY%At#S^`GmKncwjz=bNtYvg zE!lAk_EGFMq$dl=e#0G`#uW$>bC$lZwga_M+C9CD`==S7_WoAdjnj^UllUR_TBW(P znIey1Q0=RF8J@8!Mn2gcx%#7fyZwa`Y>DqvKEY6VJISa1>$g|0^vVZ-(Xy8YtFrhu z)DHc*+j`UD7&3FgQSgoVoxZ3{wJ)#2;O{K(w{BV6unh9&MxXQFop1T?IrBX+ZNA}! z=)d^zi_Z6WXTH6;*1T_oS8l^zM)Tbbzsos2eZDtn-mYa_!}C4dj^;bz&bQ{>o$n2$ zy;WJigj>bYg)nnj9F|-$C5ADua4+w!OJm&Z0V@JLY$_CrT&So1Yj|KhXE zx90B%dvV;(NR|+r?F#A)9`0RIf}F61wC&lz@80*nyc=ALzq)$K=h&CShf}g=TktXJ zQMoqCWl@HD7QUUn#a#b8`j_(->W@<s;qa&G);Q&pOu8 z_8{j+@~Qtd$PA=^Fk@hTw@-5J=DXkXVWpLM=z6{pyT zhzWsxzPp0hwe}Uh63|g(WLRM}ap4h97087#N8e+ah zp!2MQ%0Tw^LhLOfHRl}LhKtSd>)F?3W3K(o+t+*PgZA|^Jj390fcFKAaZiTX*Wq8> zeVsF8+P?nSMfY{?DU`hL0K5SBQgzg=|93nkzZ4It{rYNrQe?O3&_0(ZVBh~uT0aiC zWu1MWdy>#MsjR#2!?Wf_)>04qeuRGK{m{+-)3iGNS#oUJdF_3FEWM2*`^VoV(gT0= z6nxNipNt;wVN73h+xcc%I~c>}-C8^hv!@vU$s((&%Uw5b79YgAnZ~+Ve#eIXNvxa9 z%O`EC07vf+jfB~!-1(IPj0W6MF*9-6`lC_F&dE(Uxi(BuqSA|%E&EVRCxrsvYQC~-& z1CQDc?2hSv7Cq0RXL*Eo&LI0fLcIwWU*6AqfcJ-~&)o~98;ty(^q4!=;H*j8il(Ed z;kU-EbqTkEe_C$4)~VL$4sa~i>+G?gDvaJ%%_ps*=l|PeKp7i)d%A2y#1dVetzGge% zFH9ZGYj9Q&y4_itIgWVuTJt^Qlik?!%l<(2JK{@<=(8EuGOJYcb`bhhL*6iW;^j3w z&06f|nIFaGG4?XhOjBnX9YH^L-`T0rfqwx0y}Or|EXW9CW%DkZB2(wVjFEN9mWVau z)tL%mn`Bw_nc3If-i6Gm7k(+v4rXaQ!czlR#&dVDe^E2|to{}On^W-7-+ZLP~4_*yVSWca~ z4;4Dz2hWL|W|%iv!MrEHLGe;Cp4^YqQfqV9H~IV6W6CJA2_9OsLjBGJ7nXxpVR)$+ zyj0wtYWB$m?30D$>7gI$?;ddUacJ;pn1=G4{@f32J-YY}7tXEVgw~UxU(^-a>)cHs z`*J_OKS@4))0{w~+F#_HA{wPLCF@9K^IFIwy1O~8{%i1&<~y|NXMbZKPm8-|{ED3T zSY`~iW9K4$i0~!#9c-J$6J3oxDR#Yi*rx@pPgUk~e%+oKt>65EX4ZRdo+cA4|) zwH^;ZDSpR%wJeyMyNo$W^w>28Qp|c4Lls^e9i^Wf9NtpfhIKo$9QE z{O3%waM?{)KH-o!nh(@vb_QKLM_)jPSLSSR#@&fA^K) zJr0;0Q-9B;oww8FJ|~&)nb?ME{MMvvWxM@+pxo(?Yyj04*1-j|`+c+_Sx}N^F7#j| zT#-kZIGqkW3~bA%3bvIRe?dm$F?h1$*2)6n3uD9hLQ~~K?9a5bpw$=aU`=H=yqSFt z-k6w_9lMFWCAbS`8INKfY|+zlL9;Scu(Fdu2>d$CfYqfXL zewB~z*I4|VqhnSbQ_G8zJ=l@FywR7O4o`=^t6h9aKW87=KtIm9YC%R?i){?PosSk_ zFC;owjXsJ0eNPdq(&B8JK;{#)=Nr11X$O+MbL_S{#(D>KE04cDy8kSEz3STx9%DE6 zP$&AC1;BpchV}hEtEn;uEWr(UCC4sY{MH`7wNErF%>EuQur50h`UmbPhsM(0gLZK+ zdD1QOUCtGjV?X<#y=FN5+oVTMPv`lMAYh^zD?VVHJCYY>&-gvMx^|Bt~j3v-C9(X(akIHG@ zJF&gL8yI`Gw7}W;&@R7ZdJF#ek4D^nTtYv*I-i9GYF=uX7ujsdmP`2q^gF;DW!YJe z&T-3u*R0oz$7^u#mv!a(iwL&9xs#Fx#^sYt9GYcQtYjndmz$t}nw$CZAJKS>@3b?1 z$@q5AzHB%I&mGv9n>FmxwNR1&3GPDgH0`C)Thrc7;(>!h_Q7CMZD}p8blW;BIBOlI zwdJ+ptr5?*<+aZB@PJ1-?-^N{&(6dy)%YkxOuNOI$rt_p$g(1M1=?0T$$o!EMDZkF z@lR;KioPgLMyWH8ZeOOGc~8A1I`9lKwm%Xh;%RgTe}JDl1wWPYW1j^+%ctDTwZH6y zY#&4e^}7$6W5!h+Ftj?otXEF>LBEZ>`eJ+681r|JZ%w8H&!$Xg{uoy^?fw;>RO3dj zegRB%))8EdZc*@Va%i^tkUo!Nws+8NugAdEz_C%|ba-I-rKYv#_>};INF(iuPc?Ap zaoSV*=r+^$Rn3>r&T__AoTYg*?d_nwpZ!(zcFpDaoRGr0m9X$tfcUz zlXkyNyW(>eju_`qOab`oToW%qb|z-52c5QEAKanwrthf^oJa0^vPUp|IcWMK9~%dM zdz`*RIU{d!`*PhFeQ|UK4nFPx_Cx(yGt?gkSKT?k&<|&<{ltOMi1mff%82#VG;wHs zHPXk?>+5>@{J*`v;32jcSeO5AtgkWf;h-}{^h53(Omy&p@y&GcaKVTgTXOfTL5<8RG!DGYdZQ`gzq!@)F9A>2N*iAJBK#WQ%^r)v zUwQrwUOCBS^d9=%{QkYf#}PmHd-hQE?*YE)OlovUX}q5UKkbX24s3*7{J_2w+{c9; zInF*TJKqv~M2Xwz#Qdb+t&(eGSUEb=NETa`QFt@5fU2PH{fpxGXi=uLSF%vbbG{#r z%gvp+FhdKJZjlbpWa9TQH&tpcrgJNN>~eIY;$P#FDa%eQ?lM6686SPnq&k0nRj=-i-DKs zyPI=b{HkbMo}E=8TgC0z%qW)Am872~HsBgyAo){Hp|yo?b(~kKhT|m-$59d=YB9&s zvBkZZu*8*3aRwjCkDrMbcgf9f*6Ij6NaMTIZ*XfL^;T-U;x9#u;BQTA$Dwwv%v#k- zJNotu`b;2}E!E(y%8AcYzYKqn=6_qnx9!YP`ngc;4b`c24ZlA0n>(Zv;7~R2_!x2^ zjVUYK$asj&YiDJqkE`J>t4%OWZ_{bVTc;L&r`RqsxA#z&a0a@vC17c-nm7}Ck@0qf zkhw6ox>LwmKgbMJrj0Tht64+tnBc2(LZ8Ggei|`>-is__3gtWPeq>_VQWa zSm38JujYJaqTRJG?_C^^H2bZ^%6A|6bS|*7IH$5V{F!s~4rn*?b8H%8YDLb(`1cxG z!C3X%J1bkz&Zk*Rdq^9?C)!GfyV|&oHZ}u8%{%ytOfL^OjpeWF?U&A+$d5CKuc(1{ z5c<#l6KF2)kStnt1&PC3)`aYy{Y&=(mVMeb`scF4#Jv&i{7<{Sm^=06S3CCRI`N#Ilg|h8>&XZLk?g#CpU!mp5JJIX^1|Q*d zzBkOvp)c!v)8`qxMPK2~J2~gB^A*mUNBZToyMX6M@i)GV_xa-ki)#bceD!zpIBULi z`CnvRt@F*8cQxNNmhI4fh&isy`ieeYg5^)HZVHP!DmV!KVh_ z)SFqOv|TVy_4E=SN_%U=_?(+<%B}O|%+p!qPqepULtc*fkT2-n&YyP`ZL3cWw5_v% z%4q*u=gW5L!e>}~<2tK)5x8VzDC>NA=35EB%P7}6W&74i_N12H7qDwZjyQ-+eJ`@9 z1p08zcY-*h2FA?yci=~q9wRnV4)>c&&l5oY^9Z=s1KpKfh}oOdc%uaC$?lnyyn%TX z4t|$A$cl*L++1iS!_2GN2ovksYeR8Dijj#gpf6+f!PI9@PVRE+doOgZ5qQLXljkkR zzf1FK@{~_bI{sdH?e7@MbICkxAddyTJWZ2bxkO(3+sIQrFi1U9<{hRUVj_1WYHLit zb^_D8eZiJkT0aB*gA32CwMq}MjxGKSJu&>7BeRhYTDPmW)IBekL92UaSskfO(b8`A z<~V(pyaC(xypzCg+?yIJWrJhR8LYuB^fMN;<5A!jJM*57KJ*=impF4t$MCw(ilnc5 z;`DWh-Zmg_743}|S{)93BoScV?5p1j;APV z!QW_ZhxyfEev7+ou$%F>dw%|xN(KqPAYLnSE%$h6ZxA1p7E8Svo}iR`7xF*D^(tPd z;eF_bc%KR6k5g|h^;WNlDpu~^59hMS@!lo+edd}0`R!P&vsNR+$sF?I2~3IN17a=w zCHU8{g7`MHKmBjCKYCoxygcN`KOi|!G5v`0|8@2(jVI&3Sl@Bg=fbH7I5je!z+WBT zq`+T?eUlj5LNlIn$1f=gK7N6+YBR+%9fs%cc-Zib;-7?PDfmX|ep13c(s<%q%-+d= z@9uFW)O}9?-{0Ox_y!L-N9pi)%{;ThR>!GL@L^XIBNKcHmd{Qq?$pUs_>^s?{$=bH z?8}N%>GiLo&EL_>@9?!wdJQtK69IVEn*BQQSI6itBZ#rKqTGPlH zkbYci;;Xvn(OnNp|66Nf0{J(yCf0GL6nuY9Y;eP~jXswc`sVOUzzx6k;TY`jj1d^r z=Xcz@v_t2B2sP zx*4m>6Q$9{f8wsn(eS1J&o{c`5^Z@0ZHY&QCpR+kDEqy}q&ynahxopOGo?4CPtXVD z*Z1S@m@3Jm{(X&a+OwpKeivyPlUvWR<^0xsIAh9f54vOeB4zftbCo_%W%PY`AA0@s zluJ8*Yo0D6Px?G9r4OU$NpxRhx{+_*JWV4leV+38ojy;GBY#nu4^n2Vc^YeuG(SUf z67bwYCf&|DYGs~?zGv`Hd=T}*PZEa+xHY=xCuEpgUIs^s_$Js5?PqE92G~Ad?anvl zs=iO%0nQcidwm^VxnViB$3Hvp4mDa2#l`pG2N`Buxpw1LGp^zZNe+hhNOvchhtc=Z z2gy&X9ev+}()T_4XW-2FO7s3L)+P8hQr|}zi!zoUR&Si?d#nW?6z`nrpU4e zW>JUsc)|Tq?lw)|*Z)@jZ-8Uz`G@h%^M`ce2^qOX(AxV1{|W5j65#b7-Z}4Ws4)2M z#P)s>-ckES1e>$)B%R?iO&ppoVCAU;-}TJ%t(;cdfTC;kdP z{_7Z90=+!vrVW}S`JBbZU+7P8rPe9=de&OWPf#EH%>uPQD+upI%!bdiz7P2Pbrn2| z=nwrp^E5oUVhby#Trua)mBfEjnNG$OrrZJU#5qO&#Q4tAR_3dSv~F-rdtQ2Z!QG-x zuO7vr+wt`8Pjv$9{=ls)v7$?U!WUr)F6jIP;y+ z9DFk~`m$(G8+g1*^vdVNXITs!1T%xzq@{>=js06+FZpKBACI^Hi0Jbqc5w!!!gLA_m3dA=YD4oPEp{MI_3HmqNf`fOaEsh_}%sN z;T981(kj*2opo6Wo^(i#@-#fRY`UId4_3@L#d^~`on(DL%PV#NKsPcKV~A<9HedEni0s|5z3?xOL(i_VvIkjHDaLVI z8mvlOSUtqNNwygqSbscpK6GXc^ft9E+VKo_tlWc3%+AKjr;*jkrj@(q+d0Ru<`ffL z_Fnnz-xt5moT@+Hpst-0nM=N%ox|BQ7kh{|TDkDGm64hmt18Ov>@3SF4NbI4?^U~2 z{|{TiJ@BVgeUJ@Mm(Qv!ywu8F9-6Ug(WO>)e?#N34WzFwdFNkda6f$yiz< zx6rSme_6u4MLNre`L4Q%Po0!KHMqKvHEH4oX^jQ$bk0#>Y<|Ojr;gySu*c+nJfG^e zyN|JE>hTw^QvE^OzIzVeTBpn%r3}MJjmD-Yo%-qeZ+^< z7>fUT&2iZ$?m9oX=mzc;It4rht0|oSq>EsT3$Q~z`W(L8H_?{_w#~um_R2DQlEJq- z&V`m}Uu_@`hj4N`?RA~DvZJ(_;OvJFd2WsE_y@$#Jhdc1TKr6CiTeI4@<_I7Gak*u zXAhx=Sh}Dk#lGJjX3LPj7Q2-?$*1pEQ+DYMQ*b|$v! z8K+$9L-^-rSf#6XUUFs|W%}@a&!TU^I~$M1CLP`?-}KD3;I;sL3j(KR@-(2E(fV1P zw9hDS^F$|adOo!;`)lRBCb;5d<&*3|b&i!U=GL!r4fIE}<6^YS(IH>l24JN9iX*do z0sVI()5`9H&S%oEYCC#}YIcd$qD)=?(7 z{LCE1%Vq6R54!b;bnBW|*Z0Z9{S)snmYvK9J8*M>ocC&Qg!?h&xzI zTUcMJYZ7(EGpGx`APz2s-L)N|{@P2eC0)MgvEae^Eujh4L-T?4-M(C4KCOjy``FC6 z^OCkdu~qA1HasHt2rjIKUVMT+wQ|N54QWLtAEd6bgYZY3N%-9{2|J~aS^bNi^G~X{ zeLVCxY1h}$j{x!bTki~S3!bk&9?!&A0mmJU@XC;4l z03KquyMJ^3O5eY=J~l&3v=_qXpt~5pi|X5l(JL1Dp9p=rZ{>|2{pd%B*27vK`ach6 zoY)#Y;Dz+?x^v<_@Z&?wb2~7Z2A%l?d|C|H;L}rQ1CJ4Ne=yP#1IJg_%wIAM8s&Xc z|65(!FPbU%iZ)%r{?i5B_TV3;&zb|#z5ws{F-LL!*Hgymg`jJ_%%jqV@hSI+z6+koxno&tw-_j|&WckkI6!pA%pAM@H-tB@e2Py`sGt<8Q_Q!W63fBXZ_(l3wjeV_y@44;opze*>qI&dt#=+N~ORc|Q zzc26JxY72&uKs~_Z9%_e-^RL*h@Tbx>Sdjtf>$WJ(`pkPV{eN5)~x>tktd)-hK{Wr zrepAq5v@V-ZCcOsSFyNDaTvdb4da?t0jrCZVGi-v6nE;_?2KXTTUW!cu;zuS4; z?T>;RgT!aJo4Jg0r$GvtY{SXWlBX9(2mX$^70pyzj-N+f`#%g0-e+hb-=#ZnXyJ5d z;WTLBaK4{$?iDS(n7y!fzAa&lVeFnyL!TRvZFU2fUG#-BP-!D)5$_D!%NT^40n$0g zJQR8Z(kt?vJ+n>Eqx1ber9dfxX| z@mU3TN8nc~WTH+^kGObyVTAd3p4jz`#5HZ9Oq{&peIn$2CwaFQJ3LJh za0VWEyIF@}Xhop-$*s)c(TJUO_iA`88{Ng=A;+(=E2rXa;)j|#D&S$ngEhdX^>N0l zwyhlLT9B93?*w+(&Mi=1Q&#_v@FAY6d!J{zMi!m)Swr!D?2a_EH14;#Bn#=VB`f(`s@-Y?->gYb%q;i&XR(xo#>oL_hR zv_BlFVC>cS9VEB|)&?$rpl_$3Ly}YcKX7^#c~euaIX;m*Dfop!>yEz^;Nx}RQ+7RZ zd#YN;(XS1KH}9bC19|inpNs?3Y_9009DKnu{)v$n;lp*_Iz>D-oqZW6){@SrJD?k>O%9z1K_^bqj**|z<}X?2 zsafzk@Fa%UA)lcSmpk-9^2s7-zR%9jNuZej9(1=3kB1+0nhnc{`vv76;kS}sht3Lbj2#>L=KQx>t7I<_m^*P5 zx;k{udF{U>ehopbo?9Ub^=5FKE!Y201>?(;9Y)kik;f$!$*R+ zrZU!ft@}r{^_{e~zWI9F z>Y*)v|03zYq}%J6B^qxf-wIw>=xkH|Xoqyk+HXJ4I)9PxosXVfB3W=T=Q;JK6Wmii z_U3H)f@HDoC1>a+J`Z-|c?s@G3I?r}&FIJu!xK3+efg1>pfQ52>>IVN^YPJX<|)4l z_Dk>+8No8{-ef*r!YAt7hnBL|%b{t*^7t_NCG8s}*c&XT?v0nq{~Nsld|rv3(BKGr zJ2cJQ`O_-B18cU8xhr}48e$nne^6#;_IYuhN1oH2yHHpE%DD>_I%U@Nhuq&g8Mp9o zIk5f;_X}6MbXoJL^`9;SNY@)kF7!v%w$68PWM#_xcDK$Z>ZFg)=apZclngb>w=(ye z&H6Oo)?A)!XB^T`GWP7{(oK4H5!~q}KlPo=mGqSr??n&f*ochKSqk1MBb(iL>;dGg zjlkPt{!ei3*EjVkpxE*^V^_f)PRscXugksDR!7ZDtHao&lCJyalwWarDesIqU4M}# z-_>|UhXd($Ewh8hu0?P>i_eAnob}#|*tO8E^sDJ`_4eylV5WN%yU>+)Gp-oCdY7vw z9{_%ci*ka%vmBZ(ncHgZvG_RL&D?1(l3b6c(Z5JGJuKgIcw6nCigO{|DErVpjaB>P z3%p;M*4FRcwr26|I=3y>glWsWR}n{E!y{T>HL^_iUd4Q9T3-7{>5pP~IsUYfs}6zF7fl<;D2=vE-N@ykgYXlkJS(rk;z9+v~f=eP79Fn;fNd^>b|1oq)bWCjsuevYiK9fh>^CI(? zeqKzc!=vdNJlwt;)8=WKSr@JyoJ$`@*>Vlrx@n)qR%Pfr_MV~(DfmB!mW{OgI}Pn| z(ybBpSYCPWp3F5ay*|r2igRWc%@Ixe9Wf4Xm>KO@L)x3+2?NY&HFOTUxB6CK43Arn z{R;FDc|10b>dRO(RW>01N?*KoLm8o)n>aVBole@&URiDU1mdddH+(|3;S;!TgWu2? z@Fcf=AMXk9jJbTQ#CGq`hEFJQ@6V3Sp12KKmB;r7!*(zk=!upt$4@a}2Qx!KpE>&v z_i@HZm@v+e9jIR;-0}7dYzqb3q2I*#2d|Ka62}di8)yHng~n^275(S@o{ZW?=QF^) z3($S2q5BM3cHa%anG|C)rQhB6Id>>1-!OmftU1wD?Fo_>U<SKVdHo^jyaj5)zMIl#F@HoeTpXB3;QTu>>ugl)Ba#&bia_N zZc%)G;I?l${gOY}zcG%0;#~qa?Ny>V>3pw8=U+l+hiXV`kis{0~k=hnr#9c=KmaR=kELC!ZZ?p|7jyyDyhM;Co|nc@-z+RweiDjlpK zMh$zc%5-aw-jsf)UZA}O{zzxn9@ebNoaB2KX`)-}$Kf-<9Zd1_gZs|DZ)sZ_GSo0J zVes?|W`j^VaF~XCnU(uou;SI@*6RSuZ}-EyQ?9PUn9+ao~5t zv(2Pj4>~68!#Z!hNPjqknS0!d=%3ox^WX4E)csA0dHxc1DVV%i7K+PAQ0L>Z%am`~I#=-+~!zUrN7Pi<+!Zn6M) z%cq02b}R#Y(p_M!zJlQx_T)WB40~eZ$sc)dC-%&*u-CkVoN$nDO`MrT`(GrDap7xm zalihtWIk{EnP`vb>uC9&LuUn}E+bd{@1XqK zNG~G2W>>ULdR3iqs+pfgXsnHHr;&O;O}*9b`Q&@N*YG_keR2~%`@5_*#fjmpawHFY z3tG?LyB)gA@2W3DVb8Yja{=OZ7~v zEjY2DMc-=JKkiYx@CVTT0r3vOV%p~XqqDJWZpNCE-#sE(9;`Y-@pJ2tr1S2{2uvGpUr;v6%p_U$KtF;``j})Dj)6Z>^s&s##7nzh;3x`&mq&n2W3EegZ6Z94!Uy_ zTVrN_cI<3VZt@wM0@?iZ18?bhZ2CDGK68POWHleA?XJtBZQT2NWHC5p#@sNjP5Se} zhK#ns`PRVTZo720#_!r=j2QnZcl@L5BfRkkHGc98jem5$^zo1SX2y^0X7_x5@)G1< z=a7wQ?A^Iv+A6#Wb-22paf*e7%|5@y_XYx;x^HHI#(WamDH+T;%4@vy;ipo&%S+>Y z*IIpo^QD2`K?i<4z)x#j@Z0ecw$U@AZ9(*KbAZjDFU#OsKz2hrGn3ItQ(M*$KSgV< z96hJjhxUy0`hVip-{jQq)ty?q)ae6rx|clA!w736dGN>hKF*2`ytH;1>&TMNYeCS@cpSB_Iynv?>e+gYO{32spkj_^wp??YN>yb-1e5K2WdifS&qaJg0n^B&o zEPUe}%DH@Fsr;=x9$;)?ZK-(jD&igN zDiWVyCpGuNSCc1=@BJnHhDPE$V*Iv`on7TKwBqb4f7>9mZg7@e`k`Nf@Ba-tQ3lQZ zAJGZcYd3wrgLTb$c7QP^uVSo@4AH&UNOoLl`U0Kz@pEtgAvjdTTu6?T(0*e(HudHB zC$(}e%inor?`8t0)%%@0jRU}f-^auQ&ZE6=0i)xbACYBuT#p_#y}oyT zWL)xo_LCOsd6s(aw#OwCvk$h$XE(KOM=rS&zCiWGNpHBbskH%|I7r$()T#QGsy^*U z>L33XyJN%4Q$!y6e@`IIru}!E)77UezBN;q>WO0)AG5LBCSAWP_+8`r6V#;n6Lk7G z?*Jp|)Y8uvp?fAf{y;k>8$YGcwIg{@ryR{I3q_K(lbY7KOnmNGkD%M$|J+=P@KAhv=IjG*a;vHwjK)?2aO!HVX`$> zd(((-oSi*CSkXc4Ih-@+GM_{F#P8*@w_QxOCK@IA+OS8^j;wwhKRm z58vxB_Okb60slC-w0)`7zdgquDB=6|LsoyLoqJ?^0cY%*bxTS(Z^sL)C2itwPyzC{OF-!-&3xakeZwygGQbkJe;VR(eP zPW-ZBz$j4ez@Z12nstt^SAf3u?uK`#e&`7C635l=2Jf<Cl*x^-*m1KoYYqQ(@u=~0Js;w6Wgz);g)}buUk%AVcLq(mUN%-4ByJo zN38mAT0SSwjCnP-KU?{(qMX{6-(K7|v1Rf{#>pNkTjwLS{{VK9%yaS_bm%O!{sMbp zL+6|`r11UZCBC*r$OesU7rM8Kvz+jGHuO?5M4jsbtovX_v;&^HZVmG)T0f0GSl2cl z+m1X_{Qg;ZhI7y_mGNB*o>S%xKf>>d^aDQmLghv7b{h7wWskcj&=A;WYZTlX|u8Tj`7P=v%OK|%C+&Y`XMJIF5 za@&WxA>BTBSk5K=L38`E4zEr2_u3u6G=Z-Ov~KU_w?_vq1-3hAubn>YcPsQ)GW8vt zUH1TEVxF3@$6p=xQ%V6JQzv6q-fHR;Jl9gk9?EZiJ9-<+K~J)4UJWE#p%>C;s+}K_ z=QwREzmXB8@d&{`YOfroXM3TuiJZ7>z17}>x$J=+tg%@#a4#vpRQuySoZFM}%dED~ zp-(yqjzpQ`wM(6~F1@zqBass>E#jF@yFA*tlyYjr?zueq0Qr?i`@mTG+PUDnBfGcT zZ)Hw|w|P0(Wzdg;redx8%B=C+T}NE^*1NK{1#52IUuNgkMWKJJ-R#x*Zd`Zh;&$|x zQT$qCqKiEaf7vTN&C}?*B{(i*H+Q9zP(Ets{4-lM3?m%BY z&U~7>JR0lQevoH;VRjH}?-sK^RhEOx?0Mu#q9-UtfwZRWgk3EDWtp zL3iiaQS2rIbus!9z<<7$v1IYzx}tUobYZ39Znv(8XrIWgW6ni`!_eRo;2~UST~V{7 z4EyhZeJS}$maK(NS3`fZ>9_J!GdAx$$Qif;zxSz{lYqUVBY>RFVoout%=?5H!_=WM zoMY_^_7lpBvAb$K_66$t1<&6w|E=@JZ|j>BARZ&Jz@-OdEE*GQudbDGyo-Jv$QW0@ zl=sw%kD0zVi#8L}Hnrl)CCU%2%U1rw$bxDa(`o3gXyC6I^JLzITkCgOC4#f!ii+-Q zEH}-wl1KP9+245VS2^@0GI^WeB>LU9A%0u)Z0k0`DaklBf0KX%W2}_UW(Q*$O9$=A zQHO2d?1(v|QwBRw^hf^4Hh47Aom%db3Si?F{7AGE{HqlXAq!d{9z#0m;O}a0i`e;@ zg1KE1dHK6`|SO zQv%KCt!k_c&+aOWBWDP;a&Cm?h-cB6J#c)hDYt`i;U~74_Zr^gC!BZ9MZo1Ns(JT# zjB@HNwl2+yX+5p+{MJGzl-IN^9SeQx#okFgiEw#zzfRMy82w84oPJ4{<@KjKtv}1t z`okG(s6U=&}T@yg4`LiMGW{|4oO$pY<(!ZY~^3ilu9Eb9FRNB5dB?sUc|8G+~_ z>&xsJW7`#b{xQz}HyTFj<9g2HYF9esgn?Dwe*;#;GzV4-fz=M=eYZ1y@ww~Vy2X?3 zGIBu2X6VJeW=$Xan$GmX$)Pd1_b)K^(fc6F2D$23{ua*IBwIzM8aY9mk!Oba8s;Yl zo*-{0vUbhQS^i^1))pg!@>^%dIXIpz!hgY^7`wx}<$F#Ef^ySBw;2()> z26}SrYG6=JouU)(=UwHiE&*?;mvw05ugW)@Z#(XF^1YjPeGg1_+IxJG@&6i_>MuB= z*!}_V(DC^QAone3|0D4VI>n#BhXiY0T2gk7cn#B|w}W$4t>_D0(4OMUJCctqNoPL2 zpSvV+iz*Mzi$>cbOGzH_m&|? zJ%sI*>Juzyq}4T%x?UhgZi04IR{?f!>wzT!hZc0Xa`w2^x@*_Dz%Dq_Nd3j+J<8c# zG~yQUMLL6A`XG6D6W_JATF_}!qiZnw1MJX6A6EOV`qkJ(_xSI|7CpcHMeOPtGXkyIh+X(hSf2Z+vHLoMA71$Eqh+^<>F{<;*(*C3j7HB&}V_=Io&_S_@hiXE~z^Pqt8>V8VIf%r|pb7u?$WEV8f;2-Zl#UvMuXcZD-q-aPQT7JvASO?;R`&u!w`-u$ zZRm8B?>FQV9TYD)*7z=Ee4nJ8Ll-=`saCPO3g6-x2YYq8*9x*kl(H$iCmmW}cuVF0JIi&ILrTv4_h+&@K-%ER+ z(h5oA-%G1h+8asZ-%ER!(q@vzzn2zM+GV8i@1-qM+9J~U_tM^`v^SH+zn503w6~GQ zzn8W^X-i4t-%Gn%X~;wp{CjB;r6E6&48=>EO`6uyh`4TPats46Xx$LJe3a+kE1$1A z$CJjtmo`Od*xw}h_tLVIb`@#-duij9Hjgy^y)>J&3;Q#YKh_#-pO4-v`YRpw zxlM_!Pq6MAu@!heh0P`S+~C>@7+R4gs~K&#ycyo4hkCG~xlhmPli26;-UKgGeF*)b z_rBDh{7>GsPO1-$likpL)w`ktkJDcDx1$5c^lXm~JjT=1voRxypS<&a!}#Q{hw7P- zJmS3{3W!EIbuP_JqEo-Gy0U!Wfah+wILHK3XPwdzh!r|B};V_nIu`AG&yyI34?feAD+LJ5b@si+(&(|6Vt)?Ugx> zTzvN)$;GXs+;5f?|K7tKa0e57i}8_;yy72+&qzk(tbcr@{paw<(8CydtUUAq4ft=B zU^^gLLoqgi-r3CYH`09(d6C)J(8C8sZl#>|lD9CHC$Q^lU$?+OFi|;A6#<&n&V$mKi6G-1iVdh7TBBk>Q@avSBBojrK@{e zAKq3ON4~iZ8H>(s!lAR&tFww=)j}2Z|pEf zf4l89zO8?aZ;coKCR#D6cuF#}+mRu~?A86;mE8vqUiD7S9lxY6&8~i|IZZznYx=M` zue`Hqb^pcktZ35t-&$FR?8>?)I&dBE9ZSxjx#r%#ff|=~dGV9FWn*}=WW~glu%jK| z!NTg>jcxhf06I+izOb8fX*FeUMy6GL>&S1vbHDva->Bya_xEw0@r_O_Q`HN8E{uOB zI$#a>kN#5M?&jV09X+h~FF5bNP+7|SOwW6x1M}ds1-q%h?$%x7lIx*EPk=iKpWx^C ztT#e?JAt3q*L{@jByA7R#KsTzck!)}xQmT$o*3i#F5gbf8T%M`!K%!UXCz%+LU!!@usj{xHiP3Vg&ffki@(wIRqN2JPcrvz z)@%I3m7xvv% zuuW}CM#8-n<_w4*?wNnex$fm%)U}KAgZ7L@AAWg1aCm;hZ={_~Jl-<~TmwDBvQLfC znv?7w^QXbZ%Pl5UPQglJYFbA^~NZ>$|yPy$uQqX{kK93USuq(8_Ea%%$d>R zEZ0eFaQPEa?*i?4;)@OfQ$4GphlGv(PtM0xjYH!ZfnRBUS^6e95^X*vJCZrv znQ=Gs70bRB+mFDbtFAn_+Z0 z^+|?29E%&Di26?EFEou<3VObk`;&lYsLHC}0d9ug@BF4e$K)p^I0bfwlU6Y?Mbdw- zQCuPCyVVmm_9O{stdnRMzeR~e}dwLc!W}b>?yZtK8 zg-UyybN_APdTXWeM{(r3kss2xm%xK!@~(N-k&`Th-T{{!>4PqIZ*d!IASNC!$XsLx zS|k@lj*3r6UY+1uQZBwUvu_Q$K*=5s!-wC3KZ+;+xQuuk^2uHdyx}n`4=|=_ zRrC9=$Vy(M4dC_IblXZoTNi}k0kgoVM}2`uuV?Ow2XLmE^?iaftHzmPoPn85m4Vrh z7bdt540+$c{#N941#7m-0 zuRp(`KZ5Cv|1LaqF*`L{ENY{#m>ajk`k^x1q%vyTL4BHMs% zF8lKCG#HaVrx6B7YyIffj$BKDirw_vUOdRuvxGi#>K=>BPar!OadddM<&tEwOJrwcI8 zsqc($n}ZJe4sy)!S_78xLp3v(^| zg9F%%R6rxK%NRI-Y;GcUd%r?gq_HV)mz|M}3~;CL;nPd5BL4|{QpNUzi~3s7Yoylt zl5uRymG3m~ef~+2D+aKot2u4#He%S^-H(1)HULi{XSkBJEdQBjRW84C$#Q#AmTWp! z+XZ!r!zD}V=m+*_$JDm$z=#>ua{97MWZRHjJz$l-i@Mb2;BJ3gOm%X9YYERK$Sr&g z(c`C1DV`ECA-|P;^LXRa|4?;4^qI8@-saW#qNN&70e32Qu}-%i4ENPdVK1hh!QYH) zi%{2c{@JQZocY|1?NP%VVm7gsTcPi%`PK$QQ>Z(=Uiw{Wcox?#A0NgA zsw<%-=k0B$iN6tLuaPa`cH*{u9@v9#28Rv{ziQ4d z!JZ~Nc(A0U4_vARj~?Y7>(%zezCLJl!p_UWN2){oht`Se_!xNT)uA;a{GMZHwx}+Z zZ+FXQkxyqH;j3tD?@ab0zOUeSGyU$R-o8lIwm7&U8z#-U=4v!;?Izxb;Fwyrw6v9V zdAe+rtO;2iXB>P8e8~sD_u8_ivFysWQReD!Yb(C!5|vCNH)LjCerlfDc7aY5>{(Gu(& zJMaz8N=>jhALrLCqdm2|8(YV*;M$X=7=*H6iw5A7>s5AqigLB-U9qHr!w0}l2g`8XYWoHL{AA+ZN-x>Hp zr^E6Q@nm+^5kES$A)46URQZQlGmil$0Ztd(Yd<}$or!MS;>XNhg3Lx`+5aTBiRk?S>d`qPKpxA@ zA4sdC16?KjJiObXi=GQ#3okzaTnGKO;par(WY_pafwf>h=gmL-rvsa3d3tBWuewyo-E;+noX9PcQ`oL_f-!eZU^-fD#x zv4_id0vnuyNEPGK^Iy=d?7$vZb@ft*?of{e`WG!g_nfnRo+KIO*h`iza;C-^)(0B#Y<-p@Gu?AmY z4PtX&Z_&Qw4aLxf^m6yS>exM`r#(rUsd6%fu+Nv zoI39AMf6SkP$%;63&t~_J9nrjjJ)ZzAN^erUn}ZcPF-5lsR`&Y@JnP5>S$yPx?|mD z43*%Sw+64}y_NY0ur~h@n3eGR2A)H64*uM2y=hULy){(h$X1q;C&4|fDzlJxt!v`+ zaL=7x;hiJF_2f6d``WZbG39P>b(F~B^4jyGOXlQ7m&7t0e~;x&R(72CcQ!Q_*6=Rh zjyUg)=p`G0$4&4nOVMjCeRXi3bDna})CuTDdK{wy`sl5zf5m55I*B0dm5+*J^dFqr zWIyL;K6qXS+o?uCUr}VEHXAR?ouQ{5FZZ@^s^@~Msovsrc%YJ#-uG*u` zo; ziS<~e)xK!QJMgqn0<3K<-<+RJ8P~gY?qt#Cf>Kw56M^O+qscHGmdbze2@Z}p^S-!U;PT- zwSLs!hRi897jeewV%?=Cy=n0i_!MMXt(C|Vv32CW_)CasJKmQApH|lmz00&7M1P!F z*GW5>)_(I|JdyI5z&?w$j}HX2+?+Rq&|B`Mt1sq^KA*bUpy@a8)R^VJrkEi2{obtW z`szo^Z$InbT&TD^I~b>6A^CDQ@5B|ax8cb{_Vl_n{MPt%HXYt;rbarzvrgn&KI(V& zpQ(|(q*32P;?b|>-9o>*m~w4*g;rrlF%=u{oWsOxl0JMVc`i%CRoS}sKIMGVT&do2 zv#(8!e8a7`8X2KrDZV_t-7kCJkrl4uj*|qi6)ZJw$%O?+!KsI5iZ+nxbgbgL>O~~NnF_*;Ku~7htHca&21bQtjfc+jk7`BM9yfA4Dj{aw|yObz7?Hd z{rTYo*h{DDcf2(mfc6kimGLFUTQ|dJMdX+LB$a+FL%zZ5GW-$+BlLsy z!UgVvt>4La$t@&HXy%#2vxcYWIsO?&HeY5>Gkqzu-(a3UaQpO+^eGNrgVRP<*+x0_ zp@gS=WezN5Ea2*ajn4ge2M#&+;~ntXCcfqYpRc}?{(Qr&@4fCAB|{9F^$@I6j3V}_ zyf}4=|BLa|c*Aa60iHqPKE=Onm&oU6i?zh;b@E}v2B37WecR&L1Zx`l##-TpXv1k_ zubmkVT?iFN>l;argG+%m(@mK?JG#iwpyew%dWlaOu?sS5ttE@qCb5K&Rj{8DH>y@+ zx5v+`pP*Be_TYo+NJOYr+J<$W1eP5+jIt85kzi@FJ^qU)m9rM zt_N%3%)dG7OgvG7J?kZYYkYrYeDZ&4Bu)Jm?fT=WywB>l%UAtH`vdSe{SM;QVr$h! zIeqKd{PvPmg~fef>moto-Xgp0z=v4=z{gG@GZephEq-2Mo%PBcnXk@?;vrwDySC&H zS=W?E=QYTD9AvI~7e_lLLIc7X+Dj(Z*{tic$hN)mi#Av};_I*V&xojv7em~uCfOwX zvRAKHpZ0)e_IHE(?Cru6;fs;~gDdI#`yBdLgN$GMe3<&B2fCZ_?`7OCB7-yS??iTK zY>ypTOf|f$lkWLIm7S1o;FFXat6gu;mnBr|_Fca?h_xr4QfkDs7zaOK8H(9Ew(o3&KRZN1s=s*{~_k!ULzyA$lSv_9=nDA zFZ2I;$Gyw<<)Lv?|FOSv_m*jLOQs)n_pV)P}TdY>!-`H6BEccCs^Vm<|lbUOamK1+-{3?8o zkKV+&Hg4OJy_H6#Z=;Xm)%3jl0@$YAdpC4{m~-xDFVML>BH7w*#wPfNo`A3a(e}d2 zsBa>Av?z9+Uix?GTOYVn1H7(-uMgwzAzYBZS?Hj1Pjo`{{AKkLcyh?Pr<4 zQ~2u~hu6=w;0K{cI=5d#TeP_sc-8A%5bNaL19V4aoh5zrzv={KRuM}>Yv~2*IF?q& zCh8EcGoNn-tcgbKMsB5!TY!n?GeA77ID3V~y(fZAfIO|lQ&~FBTB-JQ#yLQKAGQ-s z(8yEpOedjfUuDmdZCMPswy~B3Tgg@y*=E0A5_%e#Q8qw7!=xD){oJbGNq-aKi%D;W z7ZxvHOJ8D?zt}xCYxvF>+TZ(a@melS)9ma#c(O+x4Zw@V-e@hjoaY3d6VdhdF_tWR z@gw7+9oV**c@6L$0>+}tluM@Dl|73MufDGUHyW+CJsu1`S{UA$IYTrmFb>{~SgpX0 zvcU_=qUSp%J@fVqCvNSZSQ80iDlpz;=500mE1I&}-_0DU-gxlWTUj3~vmThdjXnDK zq^3zNDks~zICtM0o};swBPHG4z$lcvGoD^@D%)AxcKp{@38~zeqht( z_94sd19y!Zo8~`LW|=!je+oXb^>jn@2VFn$Z+i;;Vvv5dssEd*f6!$IT7t9v+y2D& z6#tx?`&y~@{=1^bzfZo}J11{@PT!h>?Xs=XxNEFnOBput=#7uAKwnhC8kx*~DV)(A z^OvI+N>Ns0jWDj*mDB+rmjyng?2Ih%nepDV|If^i<}GGVs;HaYvNdqq75fM633aoW z!-n#|>(kySo%1umfO?O$QWo9vag~qpAEb^sx5f7xI}OT(%jdLY+LyFsQqOy6OFAa? z@mXkX7v*xBa$CLyynn&}N6rqi)kiy8=)-A$R)n$D`4~4iX8Ikvtz^HxgIn}5P`8XR zVo!{3m$g86R%6Ccv%i-0rF3jFj9vQZb;w#jc43)ws8q0j1QB7e=x>yMWpdu{l2^e@%IkvUP~*mN$6!+ZN4 z!rlUUC%JJM_W~Y-?~RjB?V7bj`tecspd@HVI)?bG!}p-Xx#Q$X>X9612XibrkfrA& z*9H_Dksf}VHj8zRlP+vVdl`2ZS{to}^QjYBpJ%??c2c~6dGCP+dhZ8i>*AD~OSx#H zxevbbFMi6s9K9F1ot?s-$g*$zM&|kT$9v(qB-2@468%fB-zuHx>pmvf2e=QX*Z(>0 zHwv_9jPK+>fq&kL5-X?54rJX6%%*hy;~SAl<=pgY_3;7rsewJgk`DsspzpRfhLMkl zHiVFYH5YD&wwz;4tDPACbEv0}bMp3MQ|i}Zhm)HBM(KAuB(Dwlc2xGaOz22$p3rdx zve8V^p5^;<^X+8Y2Ws~x_?PUl+>31k4AB)-B_5r!vYG!$cV?`*9J~n7kMZ_|3fept zF8;;V3~;xXI^sOTABgXVcY^1^?!`Yl5`+%M@|RaM+rf%V&Pl83Z_3~N!8SW9tL#^H z{Y3N|0qjZ!p=a7ZbZ^ljzCHD8yPow}^?mC6CALJ{q05c%ED6>%G+<>2o2!lB$Z6Vr zH*Jdc$sVQ5O;3>?_<%zb0~)(MF7vRRk(oLB7lq;Ce=S@A-R&b!?FX*duf0a~3pe{J zxRa9k9(WIS50rb3wW+a{eIUGFdv*xgvJ;r-`-{L?d%Mx=@%|O^zcj(GJF12pNC$6;`_ z)y3DIfZ4CJSG)My49;m?1;9Ju;b`0of^%i$J$?E1!s}gJYi91f{wOB!VQ{WYyb3fU z41RccC45`q;+6X2;h}IT?&6{FCbl7w>!g+JFHw2w^6=|f7r%r*U7vxcg(p~lSH`N1 zcinYdG_-rqCCNSC;@q2K*-z> zNUnay0O%W zgZXh_aSGWp`s0o-5rgHW8B1AXOFLrMS<}8$+`qW>uGVdZ?9Cr&@UQBa&OHyv*t(!m zW!vCOo4#78yJT6rD@zXkeR6&<(Bi>(eP*=umdVl5jW6DHJmJgF5-hKytr}lcc60c9 zO-~BuyZCMggDufuUdt`OwYP6s2lH8O3Ml><^R3Txz5Zmk@Yg?>HG+`|1N*6QXb=2W)A803Ub zpSc(H{Pa&9*HuF&W9JU6Y(V^jkss-?2F-U(9^;;31 z;u(bZyF_@{KlClzpU`<^Zf$HrIqF3@Qak@s86v)#$D5bo(@!>S?CeyA2e7`O+6vi` z)3m=6`y;4*jKp^&FMZQA{AcWyFr(dg6?UH)*5-PjXIgnrIfymF4BDsO1LuFD-FU;^ zd1G%&66!VC!%vW3N4@s`NA(*0KYP7?9OdC$uhBN!H=w?L+|wyu*wd+a-@H_`mo}n) zt+uz9j`sEv+5_?8nJwB&i0luD_5wE%FKjI4FaoJgP{LT$oCQ_ zd(pI+?8U}^%U;Al=U1~AR+Ldo_Ch^toY5MoeGB@) zX0{iDP}hE05EDAn$yR*qY%A<}akUkWyc`y`qN9_o=qPOk!b)2KjNcAh(J{04;0J}R zcn7wE@<+Df?Pj*3V`d%gTSj~N82Si-t@s6Pp7uaF+lo7y*@}+;maV8>_AG2g|25SE zd|@jF!B({UldbrLZN)UU6?A^&9_-m}$yS_&tX9|x8t>9`Vj=7WjXfz2vKKhJUDyi) z`aQ~Ctbo0sGJFet2)nIN<%js*+uGb#JPiBa)~WpV;9bpM&}OKf7@D#8^db^!S^qJDyd&-)VO@Fczt6@9HPL$O(B|KQy)Z9+k>wxR?&rt_r$WCy3?uoH; zOZE8|oR>nfUIfK=1<+5IdzYH<{62~22)V65-+}5c=0i8srRvw!VqIkP?^u(Cy$dm7 zz75aGP|WF_!*kV&u^H`o&&OFnhj0d$vTgKyi-Qf4ZLqD&GtSpxgOKOHdS)xzM*9M2 zf0bQE*drk$(pU-^DGwzW_d|~9^bQJ*xAj%n2L{V#{{6ts{@9x2jx3zXaiwol$o7YUL2^R1Tm0*0hmiX@fOK%tMFW=N2&K1fKWb zV$Z_2$Zy8-R~~M{ILjAxe)V$qnPxnjyxPS2WO{l{S-4kgZ^pOT*T9{|e>7kHt+pkY z#(x<&;}h*;`X2PP>+oDIcMCX3@!^ch3!x}S+ zzH!5CCNbwr^BlLfmggB`&SYH$Y~R)P2Dh}g;PiIPD;(I2V>>kM@S#J8h&uMA$$5L) zOGj%&^zLgu=GABpm)2WNm@Dab%rW16{6)1-t@)fe=ASQ=_|N##dw9k2^QiRY%{>-W(+jZw}WN`V(v8qb6Xi$GqU!i^PsgwVJN4C z*xNA`ZQO#ncgP!7eo?OXT7maqA5jC=V2zj~?@4U|d-Ul!3HzYlQ(H^#QP3K5DAsLh zjhWV0X-$>JWz7S*DDy6zQhUan4h!F~?X%k;m~3*I|urWbtc)KxDju6n`V5a^}&Kh(>Smh|HD zZ|G$l>t!_Qh4#e#m0rYq;!-F2K$AG%Lf^)HCdy_Q=~&3m?o?+=P>=P!P1`UBb=HJF zq822>JdSq)X)Oq2kQYm6JWhSny{*I^3DohLUWn@?%8B-Rl;HUx_!f7;uR~j=av)wA z&E@op7T*dYnaO5Qn{!^L(YK+Yv%q_OJOgdEvhip*qcK6z`K1L>05#1pVpbsu7+ixy-*zz=h+m8 z_&cBL;fOb+*|+@c?!{f3`KP#5n(Et!bto4~+k8^4ZK3?cymSTjy3!u2kRgm1{vdNaOv;(_OvXV`P7 zBP+09j?O?Fk7p3I89ZATk-PlvX~*qj;Er}2wgYva_LI{4K=e)=?XUV8&z^ehk-)dV zu#VaRdmVzX&#Ke*Q`n14&%`xu0ht-gLjt@|zr)&~4}f;7=~f5RItbMt$6gcpt`yay ztF4E;3|X4?;ruRy{|W21(U61Wh1}b*7J@x-?l_m%6VK;XVZXY!4m<3Wj|Qx#Q$FZ9 z1DiB>O(*o_5tjUH6*}m=9b>BvwBLyQ?17(5tU)e8xs_sXNT)U??61{^J%DeQVefUO z3G?*uv*-}!yAj`_0kC~WvT@P!Y?SC;a&I!(FaySjB%!nyI`##9k0O!>knU8*rWr*>&1)lHf&W zSO=kA(b?536J#Zs4nrmr`oRvFJ3!{ncA4J+pWe?Mfw5%uEzM>AJ9#5n-k%_EKIBzr zuARglO?r+UV|ih-o1MA#GW6g$V~youdAC9y?6nwX!WwMzcGD60Q)jZ#xjbFMI-<=` z-6K6v|7$2@a6W5|#+-E4+L;gKIo;$3`y+PXT&-s3v6aA29sJOqI~tFe;7>g>lVA_f ze%77da`nEkwCBywIZ@{m(6?gFl&(Saa7G?Jh z>;iq$88!lY$FS!RnQxo`gspfy`25@KIp!7>u&v< za>BPScFe^+R+#o+ZE_FB+36o|Mn8UC zQV;Ed8RO{?_u#XqJ8KEY+=4O23Oaic?}?VeHZ_PdjO06;ypMUi!F}w<*k`h+;H?dt zkKwE+@=xCZI19UR27BGM0BZ<)w_DIb+5>}m@}sM8#zCnyEHVb~G-GXiXAsT}r!-5^ zmp_YkQ30KsFGn8>Sl;nrm1Y5Cp*^-T-C%pV_aN?bNb4+YLy~)-NNb@k;GjGsp=u=7 z1dqa&j{xpLs97H9yOho2QgSEZ+z668*8Q*L{&QS7Kc4kLa;dofi4J~(96t+v z1)SZ6uu_+S;@#Zl>BuweRXT^EYuaO^bZ# z_Hm^5E7ER4+B6On&q#}F`kh4jRacsZwL$)b%py#%_$KL}<(2+x;h)m+at}Coc8+OV zgIn-H8iyQ2I!W%G#JiPC&>m0^*WKJpTldZ?MH9Av3w1 zgqY9m{6^Ta0q{@ZsgCc&nPX??Xxos+(ckgjX#bN`1Jp9lVvs$dIqpdNwix?X4Woo=|FoXbav*$L@}P47pp8Cib-)oeVzqhwZ$T z+79-SQatp$(QCRjA!tjQH@?ROnMk(p-*@!Cw?G~f#zJ@((LVn$`7zk8+i33^-Uo2U z-W=@J8oNTfFb=kc=A6dD7KE&jXQ_m0rjb_2cnb;PwJzvf~3flE`!?v{ zKj}MuMEPi;?|3Wr*Iupf_@R$$-?0Pr3q;>B3bIw15mq>b8pgytNp&0Z{Er8cHw;|e1Dnt`Ze9^h3)}t*i4)=igcQua2rYA zBc^du)2*$mPrTr&qK*yQhmw zIxNp_k_Ue5^0@kQmdAxBg8Ab6q>V;Z&pl>|swr@cEG_N-r`{H{e z&h-N0yFZW(&*#8fS86YsF&827-NSv|Vx(8q7U|*LG}vm4!Q>hPrH6Uho#d|r{28n9 zz8`D?#a)N*6&UcF!cmy=D<`T|IO>;|Va!2m5@+$9-%>mm5XUZhK9c;-d(SB!>fFs* z%&lXPzX9bs?Bq=GO^O**rz_x=O#1F+DyDHN&HK^Z?$ypAq%|0-XH+-IHcf%PX?~0B z%s99aZxY;^F!#LPO*>eHGqz~`poGra!urcptaJ2$9~vv3!MKZ_DWum|IY0EA*J*2E z3p<2Gy@$EmR;CMH7$4HyLeH)1gXOrSL)Z)7i#3M9&|5+f`1x-c4pJCOBOCon3P*n3 zo>(9JB>ScFU)LiYy@oc5_u8qy@(^%2>+wPS4ginp-w%+j2V|l57rR0ZawB=Ej?#WN zoRjm4Y=bmT=n&?GJX1bRNSpB9ZIlk`zfRUODsuzYp_e;Et$tYcMKb7JxaiPzx8i)@g?^=*g*;y2`0kQ+VIVM3Bg*1B z{@u=R%jN|r3j^Lwx&e7Kw=osNCJ#S|z7y8X@jbuDYo?*!dKUX*(O+Fu+j`69>F}4~ z)dlbHxG$pT679{U{%%H_uKpn_q+1r;#Qx*<-TXuLx-Y83xhQwTtpPXnYtS#7Vnja@ za|qM>>wLVw!#fAq&m9=s-ZZie_P|cj;jFbwd}a;i@5&yEUb+u+e|vPknJWctefG5B zp3U#+bO({gz;|(O-?!M$U+obLx!e|6(EqK*`$RsSeZrbFue0vxKhoZL8gG>ALZ?hY zeqz0}VtV%=rj1ryg=ZYSyOcruhVkx_mFiRUghf3;^>_!Io>N}h3cTA?HskaYT4<@4 zb`<-S*BOg5DsRQy;VH!HtfwR$omG9$W}1Ve-+vrLKi=s*FnW%qqRf;Hay|dN7IVkZ zMZW$R6L_C?(}M7PSxEK%Wt0Wl+YIXO;f16P$(SGe8gVQ8_aN3(Y0N^;5ZJ065$uQB zdN|nwgrhXo`+l{5oX6|OuzybTJk;m;5NpvCw;Sfq(Z42p(=43pdxFVs_^ze!i@j5z z;`xcO*%>^`O?WQQ8EF~tL(j>*(ADWqIM)?%X5f7uax=lrAkx4&olf&RuIV%(PKx_O z#0z~4Z@}}Du#6w)TejDR(Af;fiI0Bq8OTEEQ2twrzpXa^F|R8>)KrJY-VZs>>QLS% z4$D3GweD@VZ*bEBzi-p{N*|27JYX|u9nGz&sm4stm96d9>FTvXec@)o9@B@*r?`8EX$c`a z@cWe+i6I?;KN*)8(s-sYBpBgm_$=>UJmtoaxeIOzNyd%wexM%v>`!2Rnf8z9m)6$! zd1>3C5W*=MLN}HFZ z)BD^4x?Y#kegCZ`H4V_2F+BEVKg6^DQo*tQx^`2nm)lJ*#eR!O_-Uil9VOlAqjl>F zy=~q7i?ja_c!k%xz-tS;$9p2K+0#R2fiA-RP;h?t0?;l~5<;#)zJ3`uE2IO$9Gf&J z!uzPUQ?2tstYwC?5Gbpc9ZVI92T8P2@BUEv1_WAZzL{~Pt zY3G}6tR7Y}dh9m|vvBPkJN$t~#%s2fM-jj1;OAqvOux{&S4_&-58r+5p@$w?1+z)l z5qC*T#(s6vox(l(x=%-qgTFZx4dMa6cHF>k#*Qu?{=>-Ma-uYkAn}a@{e)2;dDpI= z(GKrEH!N>Ev$d|He}kvDtql4P(Z!3iFG6(5i^@=ULUehHypr)v?SNImVh-OR-_L7n zlJ%JOZ#C(HY+h*RXLOzXPvad2*qB9n)X%dNuX|}Q+3Gvt59f=Aozr!I+nwS)h$(Q3 zf!lYwK)9_5-UEN+=QQo@SndtG>xHuey7(^youX?UmaJ?s@dbwM(G!Eh2NmF}Ykp}73!=oeKod{yR zc(7h@rijps?;@|DvYM|0RtdZ6sV@`n0DJnC)lfaf7^+atU-$ZS#b@gwh zb@3sv6lsyCF?Kwq1olkKDnA8iHjn$0KZcL~lYfIHzP z!lsfM@+UCD`WPI1A8l>T3AibKDEx)#Jh%JQuCFocd=BU*boq$%*pxh0*xS~SvmNAY z3pv|B&JeAOjdbket|jVqKC8X8t}}7=NX-n zt2Z~1Y}m`AWG9(IIsN3OCa^UP&#vu8%o)Qr)>NZE7^3OUmO#fEo}v1T(~+f^&mno8 zp9}eze@nhH;S%;UyofpU1hm@(~6`7zWNO1|Q8 ziTX?ELx}GdINSLQ#I0RM+`ih2<&Z;9Hlj|Q!4UE>$wG2BKleuVeWYKvb=0rKJj&Wy z$VmNfKL2Jr?L9h$F^W8IO@4=z`d#bsZ95UxsXs>N;_rn`rZ`%bW$ef1e>2}oK9Zvr zI*?<7Kg(n9=V8n%`}K}>TKzJQnciqmW_(LJMC)Xmf_IoQa28C6*3q^Qzs-2JD@5zS z?izjvvU?l+ZZF*R9^N+UL$qXX=$V!286?h%$;6p5)SfbRoe)-UL!W3!GSZ+jz&v1B zLjavetH*K>`XP8;uGV#(Npfw`h1liVqU&y#Ym2U%U9K&CsQr!l*qfTUzVhE)07lHlB7LfU|To?{b~5 zcN}aD+Ha--{lseM%sZRp>onwqU5<|Ut>o)~-%9R4{MK0R_W0dK$X)Fj#CrR)J(7C! zX1#f{-n@p8&Q5shZKt6tDxU_Om$wOW>0u9$KVc8nKyKCl3vt(rXNzP1n5^40-g+7D zxX>8vBQ0pe4$U+I{g|33(XYv0+SNY*<@Ax}2Y2%b6WkvY?qpNQU4wi4h!VJO67C_e ziR7+_d*X;foC&1CPT+u1#MRw@C-4TGE2O!V)Z8uHy7*&kWuHHBei!}n&9B?frTsP4 zecPp8+NFkm+U3jW6F;>HW08GsM~n4ZBF+%LY&0cz$2_I@UMa0dYS00lk4JkY&F%^_J)?;1uS-uBwihTG8p?WL`& zM;}mrdl~0JLk9Gp*M*`FJ7bTW2Qp%BW+3hh`%Kv+!}%xNvTHH-brEMxHsXBt3n4hc z0W#29=}d%4M&A+Rzp%dK7Gk3%W<*V1GzC`ans(KH53H#{uJt}ejzuXT7-C@2&w;eVU=d_tF;hYSt|9W68 z%TvrLV_(Cr({R7oT?-o92K$;(erJF;0GmDO>?dtV4EDmlh4Lo%i`|qv_Bjc6O}Ho6 z-Mh29pK#yNRr&A6?m@ym(C&`CONb`}YcJm+ey_{sW968WKN}DlhG)^_Gw3f2K>I62 zpQjwMlE3eTznEjy z4JDb5qmGoJosfPo_Jz)nC+M<_btc`Af8i$eBHf@X;U;w=-JnXHPPC|dL2i%1W825U4YC~w0M>)**M% z*xK&Du)9CTT)WP}Pbr1(5F~WpKfH^-IBN$y;kO6Zi5_^L8h$f?XQNGcWq0wn!j1f# zcQd^i-(GuD@Ga86)xzNk!*2}Z7vXpMbnQ*DK{Tc}wh0-Ufp|z(#B({qOS>%6<#;F@ z#Y=uO;3oXTjr=37%Y#F;%h#hk#Pf11pLL3M-4p(4%uDT$`Vk^d)Sq(HchalZr6u=c zyne|n+=AF`v6Gu0yA?XQ>DjHo?pBI=AmhuIZaU=GQ?nURHe+eOKFW;ViL64Ip}j4l zK9KP5x$q^L|3JQK2pc^YzMKJhsP7&FH#$Q_`1?6S#QPxnTk+FPZ(=U)O&Ep1>7k~} zFS@~Y(wS`b^MK^DA9lP7dyuNu<6ejRUfgSOtLGwYgQ!E;`zyX*(*W7(Fn49f_>FL&at$qf5NiW!M ztIq3y7sAmPnQR%w=Z&>T%&Tng?-ZX(N7+05Xwxow*U!7kxF#mlR2+->CZo=Kjb_1n z1(>V!HsE`Nuw7$Kp=X~!Ud&VmIxio>r+~4(hj#&eGqp0TZEV%nJ%RZ>^h0V?{BnI! zK09e1V^~vDMTwiXE|BIkc%6)%jiqq2xakiP?{hq3z}ty<#GYpCuibU})(q4&&7ayI zxr;fInU3>iFrQewEoRLOn!B)|9*BHxFO~U%%=9koR+4`(zLkmaA>sN(CSZz-WL4#t z(G~R<{`;d18>=%L7ea^VuOAH4>DHKfXh%QWfVp8#pTc46SA#Q#(Qn0>&h4QuU(9ix zTZT2wplkAURb%wN-Ddh?F0Zst_2%}Q4RJ%f4RJA1`Z$aQ&LEEoG3~29qkT%|sGvA6 z{0`OW+Q+!z{6X}+lG~Wqp#L1`dw%)-Yjvi}X0L8DaSp`&U0%o;1gWns_STF=rq~*&oG0{;D;vDc}2=*PQp= zJDOzoM;;bI_8GYQBkUlAoeCNY`W5U1*^oBapJeQDU^8sb(J;t51#zvyoJ=U*5%7Tw zW}R+2=1!UKV=q^dPo=vb=~9{2B5lftYNu3hbf`ZK_#XRzyYASK&VN#O(h%;yRd=Q% zt_<1EYj9RhAk`geA5?!%y9Zq}6!l8A6RImz9#j_X?;V;$^@i#e)uVwZk2X5*GXqeE zsNGOKaBeqfpY5|yFIK^w^j^*7P4%Kb_F9G0{#3N(65rwF^G5`Kg$%AAvk2>2w#|p}V|^rsu4CwRJYHsxjx> zeYR&KuQ@BUbsg>GU)ugZ2f<$qK~P6nlg zcDn=dm9ow&pnvMeopW!q{XBkq;rCjV9`t$+&(5`Qr}}da=Rx`T7~=FU<)ueGBwccO ztuH}+qVlJDYIO0d?5Xl=grNZU)X$SLI3jkyfKTGeDDc_@*T0 zlpN0|x%bqlnQ&fW3Ox}C2F{GqEz$~c>RL%P5MgB&#pCiuJ7CfcTe0(XJzR3 zPR2W?f2Gq}Jooz^_-lSS_8*kunMLsw(OM+ZqTj=>5@$L1D9cL&{@%~7AMifxU&)q? zIpC)Kf6eD~cKvb{KU8O^UR~aOm3S~0MtFY{mfpUKteE4N@!?g%mt)@qrE%~o ze#%f5`>w(-0sr-@@HOyXz6zh(0LpS$GW>13nm^>#j4@j|*3X`y@<*T9s*~sD!_JHM z{X+4sADtgd@B4+}4Az)BoS&-B%|||quc9l}j<-V&oE2|tsV!^&AH2*-=q2I0o~e!L z-Av1wzQXizOt(^-Y!O1~ZLdI*w@q(+s9dnXYE~Ak)X09$@-0(|V>qGW~;T+ccT3foTNO zaZKZwrZBZJy`Aa3OgA%qn&}%%KW2KKsW+ETSEdn6CovV33IDotc?@JafvMsb_0dc^ z1wf02G{=fnK3KXZGynKBiL0N`iSXwqN~-)T8vm~JSAtF~rQVMfj6RgO${Bkzu5`iGjM4WI*KwEd^^7q`Bd+rMY@C^$C%mL1jgu-iz}Y7ia(LDD${w4W$o0oWEY&~;y=TM zpXGw{TyPO%RUWsy@JkshdG2=MuXVv2T>Nix!H>A$GREi&i|a`jez^;N)&*BER{GiD z!mo7Uzv05Ka^b)2!moD0hg|Rxm+&9D;Nvd%3m07Lg6myyg9|>#Sn2D$3%=;$ztIK% z;exd=_&ntFvd7XTz-s|e*80PX-P?`Y3Zq1pm~=3R7)0n z1I{l<16B@6Nd*~J^1u#BN%JlF1jufR597}B?;=NH(lshH@YD=Dck$C71TkdvM!u;Nli3_0nTK9-juEj=r}AYB^+ z7=>L`qep965BwejhaR|ZxKaAc%}cjgATObC_@nyMAC)pZ95?;OO>HtfDoh&$O7#~} zX`>MbZPtnc9RO+!)wEkcZ^FF{cV7frh96+l@E zb8u^c)H#XZNZ%g2GuzIW;{R1F6pz(bPzX(=S@ZLERb{swdCCF9IJuqP`;IFMY2OlvZXi|D#_={zv0GaK9aL7IhF86SF7X0IuGX6V_|fao1KH8;t! z#A0PLYIh$!hO8KDX9DbsVM2O-s?D0`Fb|qx9GQ*rS(f~KC$7O#V33avXMVKJW+`U9 zauBl>P2HB7QS200_%Mvm&CVkWZJ3s_5Kjz4EbN$5y2+ND+??X<+`@bV$}c@X4^5h* z7V^?R)pxNw@DEpa^-Og3*OHs+OSy(unELk4;b`ncH}$z(DV&7+-%(1F!mG5&KV9VZ zPcD*4k6Vp9C?3KpJW;}QwU)$=36g(uSLrKC;mBRZgP->B^jlpjT}8>=H9h6mm6AWD zTcWO*{5$gryIv$Cg;RNR#>(B9Qh362sW7fsavgkyl|NUi_=+mNYy67;ckasH|I({# zK-C%YrF2fz)t%a`;*($J%eg(Pw&NOJVb#_sJcXk+PnT*VL|yHG!jwMoT`8RG;pw-! zR6A5u$*B8KUL?R3>z(vS|rZT#8SAUE?PC=q7jdTgluKhBtA!#-rR7ReTkf!ixWQ z?#kc)(!bV#QOnT^wJiK$;DyVWQ<#+{e+rCxS*6DE#xjAFn z1Lurs51ccmy>ZU?eiuyp=bZUF+?wN^F4&iZ#HHH5mxNljmanC1xtdK&$KOCLP3S<= z#^>55qW>Ci%S}zs&)2ltgU#o=O^2G}pr0j9WSSQ#v7&}^5?3)@8!zV%s+h7k+FFjc zw3Xx^XIgSbhOcI7ey4f-MI)qt6_27SeTDrF%Ww);aC*w0k#QsESGkjY?ueW04Bezh zqErX4e?izS?8DQ@K9FvyO_I*3jEIsOQJP;T9@Qnqqj)JSQJP<)c>Qpb8_g%t91z{) z-UBzeW1oh=(TvA4jt8YMGuVA5(>b6N|6JUZb{1}mCmXi`cOLGpxNW#e))L$#>k6hT zLCO6#+!W94xG6m9sc12eG2RMFaz2Th()kx|^7krklIwlk#Qy|0@#jK;G)EEyrI3Eq zmHeG7&oZX9Oby>l_dKRYm>PbN?#WEcnAS2ioM%4M3Z{M+*qvzw(`u%*OwTio{89S1 zKx0laAxhf_v+#2BU}o?*+l%!!;+5v zN%$|1wsmTKNVPrpcLQPyFKL6)~8hL+Q(^c-y<2dEJ7 zP%Tz%Bn`P)(t=Da$46mP5eJn-0W=bh*hnu_PqLwzLZM@7V^lYrm$W0VREg5HB7|~n zugxV)Lp)UDbF~Gileq}dvhMzAe6kE=j!9PS(tmGuypBgv->1ldb%^dF1=v3C95#q z)!$5KcY3B$_!#@N+;|%%=rBP!6SG0q+#IA$7d^MhU+nm)mOQ&uPVO}QogRlNqJ-So z@i7x9j6d$^yuA3_ENf~pq~T6cLb@&6nuB@r+?=WD1sS<%IQh``2%H;K7ZX z9@_lyBac3|rEKftPdxe5)Bh^pw*8rBpL_m=iWgscdB@IIUaj2q+Usw;xqDC5-naI> z{m#4ZRqsD=@X+D+KR9x<=GccHef-JsPe1$oi!Z-AQCoNNRQ=cAeA{sP%-M6_egDJx z3qSsJ@#kNDZM^i`?|)pr(xmC!+&w&7wbpxi`}nqL+s?1Oe?VY|j-7%!2Y2b(t$T=} zN6%ip`}FN+3=Qi)AbjAUh`~dK4jVq=n#hsYj>4SxSd%3M3!d{AWLOtwW@YE(<}I@2 z7ZfgDQdGS3hIu#Ml$89}{F|38zvb5d>HPmsr~iMg|LAe!Crq4lojGRml-Q|p)8ePk zn3*tZ_MF7I*Gv8XHUIw=`qx~nA6YqN|BNm0{%i&R3l>nf!v3fGhnKQoL|5t@e{Q;e z_NT%-=jWg0*Et0J*0*5){dE+7gq;(Jm*4y}>FS5y>KFO?hZ$yXuKxd7ZXLe-G{-FjmyeSkYLQi>FrF{ zGTp?qoaqjxRZI^tJ;L-urpK9n!L*iXJ<|rJ=a`;ndXZ@(Q_O0Zr~vg$1DSSZs^ag> zcmUI3OifJVnIiS=H_S#7?;e!-l2qyLT!@Gs?D?%XyBz; ziZ$X3s^Nmad;;*BBY%-A+%pPof`Q)%Z}*RQ3T=!gqy0@vPtm4YY#4{wK#R4>NPlu+ zmKI&OK%1GKr%l6#rnub2+JtmDq(T_%-6^JUHt_@TBYbKGDB`fv{2$^FgDS)WY2jYD zpfH~o^bu!4dUi^>jTpEPe=Z@U;p|N5*>Ut#O^e4(*VL(l($WSMi+BmA;Tm9=ipxMh zigA&%W;S1&oqa9MFw1ME7@s1%7|}yNz$D}x{HFm{0%lYSK_`hh0Zp4JM%6@VTuZZV zGF}^GTG-A{ul7=uAkI2RaSoNGggr=wUJF1h`Yl%J>UKtQ5~AxThdyhYK8W z$#iM{NnSF2t4NDnWcqSUt#B6Wjn}kaEd+7dbCUzLcCahexwKN?w|sQ7*z8 z%ir8R8dA~~&y;>o#or+4G!6fYo#Y`G`Wv7bgdg%}u)B~?>D#QCwQG@jHvZF0pt{mg zD&ZJ)(|kBx3!%YqXdzxg%yH68-FWC*UYXE)xHd)1qmzr|C!ii`&~>KtpooDPNq1MQl=I+JiFXf@R zcDnS8)m*zDW16FN~Kv?YvbZG)~-##p}-*ILHi8E;~&_EDEHHZZ@O zaSz57jC(Sk$Ipjej4PSnn{gH6K8&jw_ho#9aX-e#8LN&}En`|Up{s#$e+jkoj0Z4o zWE{>|3y}4J_MXzEXFNhe&5!XljDr}fc`O6tk<2$Tmb<=C9~h5iek5ZPV-w?O#%9Lj z8OJl8z&MfdB*w{%uVb9SIEHZ^r+>3Du&SKFusOy5#vb4rHn^1-o*G?#^sDhF|K4B z#kiXBSjNX0n;17Rj%M7*csyf$Cz<~Vj17z@GLB$8iLr_Cb&TT~$1qN2JehGG<0*_w z7{@YR%Q%j48RKb;D;UQ!u3|ic@e#(e8P_tN!}vVo>ltf7GQa7J{TOF5HZaa&9Km=6 zV-sT?cR=D9doYgR_Ka`HiYuA7)LRVU_73&iE$?5M8-PqP-QUoU|htwHRDpoILAd?n;7Hl z8gZ2~_GMhjIGAxY<57%j8D}y+&lnS5@(Pyu^I+`9xHV$~V=u-LjJ+9~82d7gXB^Bp zneiybd5m>|GWin59*oyAZq2xiu@~bC#@>vp82d6l!Z?_5E#py)8yV|@WcvCpGXGwT zgBW`=HZt~Q9LYGCv6=BG#>tFz+@Z{4?8Uf*u{YzjjDs1MF&@RZlCjPp<*jDy#rQa5 zZ^jKOJmW?czPAjo@5=dQ9K_h0v61m8#*vJ5Mj76$!ZS`(;TdNr{{y7|BITcPsq)Ww zlky)S{g*5Mj4PFY#?{LIFzNre^3S+I`Dfgy{6|Xv`ff7+-i(77>qbkyQSli^D!xhb z&5F-BQQ--apP{f>;v$7(B`#GsUgAxRJ(fva&N!HHC1W{DpkdTZvjDV?Y!OtAOU*oT zm8;8`Phmc->(fQ6`;yY^1L1r@X&(_?^Tp^K<0tknXHhh8r9awVK^KjS>7wy3T{#?Y z4&yw=dF-Ed-Oxod5OmSL8@lp&jGV76V!T*T+L=Qa?XIDVW>Dy|B1d%5tN~s5_WVe> zvcxP1?OT(0KFhh7^OwPM1s06d=}KpLvRMubG(cA-hofCJbY*aO+Q&s#Ci~Ci`0}w^ zg|0%5w}8{n<9HTuI<&ikE}9{sD^tt@(f&H656RDF_XV6k(sL&K)4mq+NqbO8Q&b9GCzbL;deX3{SK*~SWH^M4?s&^!h@=x`T^r!Mq^)MXTRPpCS%fwgupn4e& zX}MV=b&4;+9v{_Hl264)^_B9+W&zSrdb6C$i}W>4?6)KpP<^KKlzwTi85u9=rKj>o^*$WF)i{>wKjn|7U*UVHz}Mq{EKmX0J8Bos^@iF zk-u=%B318-gkBtYsR-}D)UKTUQ~L@>EmirFyVoY!%Z=I_rI#JyQFmc{x?jshtv5?Sk4XVU<3$TPhba&+Aa5pL|}M?d?hGDcUZ#)Kj$GZb>~k?6}m^czgTH7xhHu(^(EF zzayVYu6Wk4lq=TWZlqi@?Q%-FVx8om`sPSi%5|MxZY7tpkJ8?bbCNR~eFsO&mFZ8g z*B_bwRJ$Ee={xKprSGV3GX3#(dneO($V+>MsrE+Wzq5VI#}k+2P$etnm}-|(>OaO_ z4`n*>cD_t!Dwi(VZK`K2r$ZJ%)livEg1!F8bR6wTrZd5wf9H6cmut4L1kL52hL$8R z8Si8#OnR7Y&kyCFTvW}Fa?Q2#W%%)SJEHWgo~sm|N}_pqWksl8Mci@ra#i|M_Ex4h z*`817Kfzv)r2lApxk~@BPV!Lv*Nb%)%CYpXTAONbQ=ItZKiQSQg1PRL8fAMKQX z;?J_T3&|hn5i;D(Ue5eH#;-C~>waoou$1|#-Kcdx5B9%_ z`Kn)2&iF3ot93z7#+A%h_F1hHs&&L_=0DEnEIXCG%}6Jo5)Lu4aA_;}Ygy%lJ6+S28wn`t2F3b>*>)8`%FU#*K_;s_-2D z5XSl;vV4mfALsb{Fvb_L#J;x$jE^wCH)A97<*XUj{TScNd_AYvpRt+w*^Cnz-^N(2 zqmN;n!TdWJmvVfg7#A`BcE%CRk5K-Z|0v^4jOA<{*5w%&GQX1XEXHPknSX!A)y#j8 z@o~lpj2jqlWE}4=!v`>KWd1#jlUd#ojP*mM|63R*a`-^TLCl}6!ZW`EVLPsR<*Ph;H3csFDHFe%?Y#zBlL7#kT^GLB??kg=KZ zdyGpt{b0t4%%8_t&Eub9oWcCJ7?&`9lkq0TPcyDy{1D@{oL(2k)y&___&DPm8Jh!S zeupw{VE$6Z70fp=Ze;#y#(B(-WUL=9<11lY#Qc$rgP6a9@g|nHD`O+`Z(rmm4Ud1oh zU)1Uxt;f*31HGRuFF7BTf*9zXVMjc2orXsJ@}l)SIwwH2OOhoVT2^|J>o#ea<(8M6 zC!_eBv14Dak6uZ07sl6V``4*~e%0KC! z-tKnBsVKAN*p=TrJiXKH>|gH9CKs8Wqr4?vU@t$3Gwk(7;%vNY?i^o^z5FC!tv0Io z^PKca`H_3e9qH%U>#Yn=v*OP27rO8lJH~t24@=xm?6pOs%I+WF}Pq}WWR+(v? znbzsl>Z)9KCVi{;Mki_&~rAPCxS@!xU`PpJGJo%98AP)KDI*p@U%5_V} zzIVC)=a{X+9U0-s&E7RUS<8ob5t)kOwolLUUFSn?KY76 z51i|dTqjby5{OUfJNu{j9ri%3<2mG!>&$8w1D$(A^_||Jke9@2cLP2=&Ff^!VkoRy zyuzwQ%5`eBD?;vzkf$!l_~kmELms(~<7ht;JMtqj{a4{>{Zs9(Q0vNcQi$Tq^#c_H z&F9lINyVVnN1XYzUgFFrjVg)MdV&MfIy#+_q0&?Bhn{M*he2}W`o^E1htfSCdT+_U zba#G!%Jn>zI^~z@tCE&5t@qL?8A=|Kfqcj(w_K-n=v$sJjh*_ zq=j){*@6^S;U&(n%Of$J2qG_u^X>LQ;sX2gQ?BDwl zm-2~jV$aQ;`u3YfV^)*f1FiS2n%%qWzkYjXL3VZi7unC9wdp9p-uKh`J)U|X=*b`Y z+O}xHOW6h%3li%9+S!_U-+A&Mw=xq0e(M{*-9&m!BnUIkd+z_lswqnfIRW1kaxr zbbRCJj2B-0J@lUYk~f4=`dSa))vtGY^4=F_Z`*QRs}V0vxv@6sixThcSI(P0y|?^G zNpNPbRIJ}b#twP#y*amiG;VUerRTO!$M}tY)ArNX>k2x|t!ndhw{uyahWQ=)2C@0w z(ftO?v>a>L%|TsH#(VT1_T`}O`^Bd|v}pg+9Zyfb*Emq$DK{r3IHxmd+cc!F={b*h zTlY&J9o-Q6`r#YqKR9sIKkoNU z`6ExiI0oS~)8r%PE}k##b?ZCX-if&$d&cj(F}%^NZ~uMQxn;i@&tE(|_NjrL%NBk7 z#$6Mv%XK-I{2rYhTwM6%eVSJH(yq~kr7PEW9D6wP+8wLDy6Z9hq={2Ae4qaO^1AiC zPXwIHi0C(A$3x#ouXttO0k^~p(=$GOywBq0_biG-4NCj{o7=AcvE6fXHniVaqRrnK zYWSe$olll$kA14^)K?FGmG|gpW1>608h7BxqD=wypRB#sH0|WrXD;O~ncDZ87rd#K z`3;`+j8{#f@6!kST4vvMTi1nqFWyzSxM9M853VbVe#}$9KdRMhWewFId^@Y^x4M@m z=zgfU#n3yykL~D+OO^G{dv<7TN*T3u+Ko@XKQ#K}wCwD5J$@WI`|H^Od;4tsJ zoaa-{ZC};%hcADB{FVA2>N@@S=KP^0r8};P8c?}s-B-_c)~8+kFm7Sf@U{zQF9_WA zx#_^Du%e8Ihh%>G&I>=kVNL6c{M27eD*63|qqDv%H%xxGan8nPystd3yZPw>ldTWz z=^onk?f16Vd~|&Adn1ne&iZWq+DG5pTz+s$pPg zxHXsdKf2B*|Al2uo_BE3x_toz9QQ#(Jay%Tb>Dzm)Af1?WdzIZOu(tGJV*ZcN^yR|0?mT(H|r> zP3qZU*_`sZh2Qp_xnN1B)K|Ybo8*1@lyzjo(BWmC<#+8G{LA=uAw51HfBo?Xf9~?k zL2-j+JZJ)I&AxR=gwc+J{Xz5?yZ=! zhc@<4dSGm1yI)>jX9&Ee=;(=YNB4f!GpfU!6ECbz`ugW3YYu$-#0@hBxsRFl?Kp45 ztNDF-afa8#8|%jHcz)opm!HUe;_%p>SCJ{=LO^XFfAV&F@G@!RitX5^W1!yeqY?ZgS)m195El(-d_T?5f|4?hxr z;khr~8}Qti$ZMv)n}1|M_L)}Y4^MCW(f#em>t>W)+sp5hZTWXhc-`?4`5$Vxb=dk%;i*GEh7P)Y1?i@~ z(~`8`e?GQo%;K`sg94Y&kG(i_;DySc?rqa9z`7wezr0=FWK)1{w0G&ELp{Quy5@)d zpRL|8`1Ntov6nv1@?2A0sq;L!?7NLK-nQ&Z*?#=9yS_Rzw9=Ase#4ua4+SjzXzxAY zqQdwEG>lmK?qi`HFM5xS`RRkFpIY)nuia0syy5XfgKU1oZ=LYy?#h6-%HxK_?tgCR zV{g26<=MqUdc+PKayTGr$GR(dBR-w?`W+hs$6XFSG^z2+$Zu-yYHjUw)GheRg`H8) zKeKCJ?)UnxyX8aFwEZp9icgm`4jA)E-J(r{zWvB6b-L}#&gCymx<2gV_rGupyYl6v z_wVT)(C@jLUaxk3-Q(^4hg!XQ;OvXneKE!B(F1N@ymRmLZw8ml-gj?#;-F#Ux_hh% z{CdxofIpf%+ZCrSxIE&c`HSQ0cBQm}-YgXhg1vP$Rdt?6M&4lI7$V zsbOz;^xg3AM|qY9`hZtRoVNy zR_o`V&57xp`}sZV3s)B0G-7kkJA+?cJK~*j#_D^EpYQ!uTPmcq=)IduKYD1u_)mks zEqQ$4m|rr-6vwO_zwO)N$=_V=xjFCS+8;hR;Q7gu_q>^oH7f=@_Un@c?_Jur@!PHoBkPl%ePqo~t$#aL z`q8s{uaA2F&CY+ z(|6gx&(;){tt}Z9@XF`mYt0{I-{Mi<`=NM|ruyfGqI5?;$iqw1d9=}Vp8lGyRgmV^ zs=Maax{v0j57*rF!!>uWD9zp5ta*6P&^&x@)I5Bxnx}7}=GkVY=Gk_w=GktG)~em} zS}VWZTC4USXsz4VYOVc$)LI94>hu9YI(>(5oxbCEoma=XI z{@`A8;s?>`bmRvpo%SK!xt|8EyvnQY6U$PYR&EhAvEW~VMtHv<=qJ%H3tDsj6+yG_ ze@)QV{dNm_%O_PtOJ+Rqmcad|zAb1((>sE$dgeXBzkbGkf%pG(K;U2RJ0x)He(wt! zwCe++RlT1$BKY$b92Gb^v_{~Q4T73tpFSqs*H}Ij?#~5(BtB%~O_9X~f_RV90nvxo%d&Jr z7PR(sm7s=CTAvc}8{-76s$DN=vfHPE*522xUc|e`BB-hHNkO&a-wJ975Bpm1FJ%g9 zGCe0~ncMe*){YzajYyZ$R@di06)!6Z84*>zz-`ms{EVo_-Ywm7!@D;{Jv8W0!T`6# zsEC>V*Dl?c619F=zV*3zmZ;Nry!7J2K2xI}U%GVH8>?@K`r@g%5B3_C8?{7t%zU|Z zPSh_KAG+?ukj$vYu0LL^56z02+vSA0`cA*7i0H*TW?zhr+BEc)KL(YhMlJTrEPeVL z3&)cgHDlHjtM8kW8TH4m7Zz2Iv_z%7uWNnmqx7h8J&wNg#<$r~X&vJ>-J6#kHDb_v zrzcuxMSXTdL%=U54N<@J?zf}%)zMM!O)`G%@oa9??s0EFANcT1QGFN1xi{RH8Fg&# zOLcShW=Eww^6^41&rwld#=iDj#~$;duDjtIWqvFl?#OOC)AGIWD(u>!9lO5IV z&R=it@~<1Ck~db|nXxb`D(#1gGlg1i)We&8U+Ld^XjIMa3;V{OO^&*K-pl7&y^s`@ zk@WJIZ!W||J+&zK_wgGiMLjqt&UB({R@6-?pLuWJV2V0->hgoJ;W<&m?%Q_yql+m~ zyXPIs*}oty>iTnIj+pXtqDFkRa@?NX3!=sroeh3*RZ3Lk&$dq8R$U*pZfD4DziFva z>(6#PR`*DD)VQ~|bq<@I8WpQI|NPv^gs7$atLJ@vcRcGMB5F@Rx7VX8vZFRtJW_GP ztCpyqCw4v4F*hbErEBykl3&n?y z1Q+Kn=FnG=h(9045g9Bw21_11rXdWShGJcuK2Z2{_!D11nQzO@Mu;pitHm?-u#s&(qgxhaNJ1ff$P4fBOCB0h4*`ANC9Y{^{~ynZOXEZ}}zlJr|%_`*_wEjP=MVaZ9$!l!av^XVwR1?dGSK72AK zH8(BYkUQUyr_LUymLw24J|IQLW53|aDsHyaDW=`Xo|!)JsYY9<8?A9Qve`^6!vWohk}x(sBz?dBYbK=Hk0H^pT4PO^iYt^Jf`{AWXM#jIyiZe`6m+!SEmK4kr~EhwY{&o>y;7sFCn=Nr;dEdmec6TzGBZ{v^ETftSZTIEgD#r+)pHX`Wmsca62GMxH0<>TbvXPsw9X z;Q#JN`7JHy>pzWk#>@#b$9}!}?#xL$;vVhNB_aRCBkz(5#$P*cR(`rIf8J~yU0gg5 zM;X)CvhwHUTj}#zN$NQ}FM7t*5yR%qz()`9y{>r$BZh@rY}o_z(o$HUW1w`?RfGGw z@d>dwL{ZaY+&2)PKBL)aJvLwHpVCR(J3EK``t8$cbk>IO(;KUvM|4^h>d-(OeO6%U zJ)HcYEOCw=+#(`U=P%H-YK6zoNa&q7KAQX-ad9KZh?Cf!6DLCxIaMH&XUK^+`6K9RykH20TC9wVE(iTF&-+!925 z=H_m~Up(9@pu&MTD6YVQ$j}ahFwT>>_tK?H(^8;5EohpX)TgZ=wRcRkl&_iF^&&o_ zvs;4n7Xdes4+@(g{h7!g8vno|nXY-B8_p^PmSs2uKear#3EfGzq2yKyH>a{5MsD<- z`q`M5;IJdeP0K@?=;aOMY~kD6nh_hJ2Bz>iab<{`($8ZL6>eHQ)gfEHEwy*VV2UFC zB{!{ly4_94MDG9e>Yyd}tXpyF)TtvY!|GSizf-5WX=P#LcGRPq93_F5V6CEa9q3*4 zRVBK+>QCjpq0=gxnwp&4P0sEkwFtku>f5W`?vQ?~;rDqI(@D2+RjVl0zV)ZR1Fkbw z73`PrR4dJ-_tZ>UeZBZ@hQ98$daOg4{ETp`Tk)X|;+@LS;{E0kzU|1$-f#%QG&|Iy zRU!8cC;@7FL=Dh<1a5`B5jU1cv=s<5Ov|IVMOYo&s?j2EN2x?2EKwgPzg_UVlSse* zR6qRgDSr2??}6VP`1e39UN2<4qh~$ILvb!MVgDfRG{i}C9Q==jJ7uFaetT*--Ll?| z=4|!Iu7CbT@JrQ>RQ!~?0%@hgw-4FQg zU%g*neP^}DxYqT}0PX#W2&&s(y1x~b;VM~%;YSJFvwkww^&`#e`luG% zzJzRLNYmBt=ob8@Xz@W1SF4jHfEIVk-sV!lgJ>=C+KamFD4XN6FpZ46mhK$ z-{`@|`wbI%t!~!7d;d$*-o|ZYf3^7C=!9<*^;Oi?X1em!40h5spgtT?!9emH|U9<=G<653&@1efqR&luU z8^F8h#BZbd+4G_F(|8z}dtcf#qUbB@>ri^4Alg}&z&-0b zxX5J06T(b4rx)PD$1&a7|7q__0HUh)|IZyT(JZm$g|;wSDrw4ST39L|U_hedPAi~* zP>Lg<;>%<#D=IU5uUA-_!#?a-YP3o_l_lCnK1)n3EK4diD)an5pELKuh+tOl{eRm# zdic)0_blK2e9tm>e2|Uj7qxqC#j}-e&wblHx1;SkxAxP3TNrV<&!vO@`Z$0+EHyq1 z(&0!Q4p+VrqJ{TJW>;dyY@C*JP{Br3Dd<*>hiT%y!Ko^(#Ya++eCO7~h5?Z9YF}66iq= z(CO*Ow&JH~*qqQrv|i?M@Vh^8xaUNhHbDmE(=>7ObWLRFWe?)Bn9+89UOWBUh<59d z4gJO0O~hMm_(jju#0SWR&+p8a?23a-W@)WHV(TgRFhc8`M-w~IhY3oLP8_#mPSM9z z(4QVkb|>nObq;ue?9lmxaR#xWYm8k7S!Sy`uCe1e$iG}&x5@uD+#9d%wdo0bhqpKH z>AE_{(`x4X?>f_iUqnzZ&S89!dGh3YWVkhS1(MYEa_BX-($0O&ba;j(Y< zVCdUGmW6NN9H0q0_=>QaF@$znKUG-Nm|>Kki1G**#hzW=+^pMW%ZxF5Q(u<1=}W`& zm`|fG@umwhKIhySDClzrcsxuRaNiqq2(eE)$SDMVuQ13PQ$-DPj(y{BgJT!(&(`5# z9dgcQB=0zU#(domb2cMcanG@XBW}lu^E;*+b(pHdl{$3laI+5I(_y0yn|0VD*40iw z9S+xFxDIdB;cYs+ONZGybm*`~hg)^HQ-=q2__GdICAj)qrbC+!<8?Sqhc}ZaNQ-p1 zQipjud_jjhbl9vz&snZ|=jgDX4u|S+v<@S6I8BFhb+|%@c{+Sfhg)^npu>GSY|&wl z*{**3=+L6WaXOr)!z3O4RfiAg@KGJ+>F{kGex$>FI{aRT$8>mhysKZc4lOzyr^AUl zoTbB59j?&fDjhzq!y+BNpu>-J*sR0ux_mCup+$$|b*O}fKbPq87^=fi9U9N;P58mD zsaz6YXp6vp&P;K@6F)o_GCw74{@jF+^whcGuAL1rvI}K~luC8%HAZp><^yG0%((%)0Gn3;JLXzTA zMTs_E5fp`AfsLEPdU32bAucUFjGJ>q*(v^j3Vk@j7H-YATk*Lz7dn28k7HwElqz;ci+&EBe8B2e{_`Q?yNb37t~ZnM92E z>vPkV*b>ATUa>+5_L1~LLR=p=u#CP)v1lqb zk7kNV8h#)S|07UhYWRei!PW>>J=7&nYeK3LcFOF8kogN!u(2Wfuy36zz3Yg0af#dA zX-V8)#7%L4cc(5%!_O>COP-eyoTeziCo#*T?N{ukPDoBm4@VV|LXJ&ONmA+-AC@s2 zztDh9N7Q1^8EnlZV;9GP^A67`>2s&2U}FQt#W2=)liOOY?Kk+IR}%A$e9Re-UGGdrpOAfLE2g z%SM9Z<5Sfy9yheJxnOZAIY>0D^OMO45Fw7cN-1vu_5;H?3R~Slpyf~xh|lKwvBPln zSus=G+BtWEQ{t!El2epci6^^EYyGW62S2yUDVJBRa$3(T)%1%V6XwUkPbhj7bx)r7 zJB+}p%ws~^FCgOjW76;XJw#d&?kv{ELvfT1i#yfl@yYm2Gt#fxT&3cl_?eN9sZ= z4LFr6Np|$;Jj%xH!Hl@nWNw znE8KG`mJ9xGqp<|diuRQpgu7f_3Wa1cdi{9>8YPw`?Bt3_^kS{>%O5E2IMpM&%8cW z@=7-aovQqQ_4D?)TbFc!ed95H3R-4Y(24Rd=kLrnyB?qPclmMGtsot4I?eLI zdil*Lzb$WHULUbPuPN`Q&f8y>t7=T{BE;mKx9&8Hn4P!n9DZ{7_ZvQI{0ZmZm$@)& z(*1Ui2VDO3&%#VS9n{T1<(!{0BG)m`hY%{00KW7;4NaWRu4Fh z_+qi1_Z@(|FU;5v#&r_l3cz~6`$eKD8V+c(DZ#^esSsyr2_7|KnH)dKH^jnkq5 z*Fg@;!eO?z@?}kehuFNvQMklA@^W&P0618M$H(M6G1}sXFb3y zly3qgz9QHK0|0wKjy!yJzP3-ZjT88$e4N$O zS8FnPPv&>%_h~hN=S@LBQ~2(lIIRKjACQ|3hdIJ=zX@=O*rVB|K@V_kn+{T^LXLpv z0rCaByKp~iCeA(p`rIHyA@rtJhfTK#Q38CjZxv#b_|g;})H3rq8veEEn{8P>2uFPb1*#Vz$Y-l6WSXnpQzSOkbKTt4YT!tqJgDwB!7n z5JvzVeEgGZ`p+FTBplSS^#@YGYRi9W$&gWz{_x54;Y2^%;jho{voyk z{ss5BDR%{*19pS`M?Xz_g?z5A<$ud z)5Qlnp!amxEuaO^yH1D~=@;LEAENx;QGP^o00xU#Qw!h_wClIk9}s;wa1qp89T5Ne#Vy`b)UHG-ft)L3H7Wx z%m7@9^4b5SzRotq?h@iV_#f+M=-+g}!`2A70B`H(;1|ke?S+1O9`Fok65>}e=Zt^@ zF1qYrfvTjQ(wBvR<54t3&S|npofg{`S=1 zulAdwy)_Y!a@pr-@P~;W;pb`MVr`B`()l=tf%}%f;2Z&<-vyc&W{NjO_r`ftTqj+q zDY?`F9s$2ej z4flga0Dr)0oTIS-X5cIh^`!~0FYxoa8s|80KWCJN@oujv>sn3p*Ah(Cqc!D6>|-?H z;PW8h16UUVd;qIMHPHa*4AY>`v6}aI@E`s(=sFF0H%qGroD8{`!`V)p77zisqkQxA zxR3VjQJRW7yl@uA^9JArNPpH0mapHb6U@rW5b~N~SBlPnYv;&xPt0sZ~>tb*| z2QWMibfSFhEKM{6HYT{_8IY*KF7}$50c|Wl2YdmnzD*MafPV8d@d^B+NP#>(51ipl zRr+A^N+Z2{G%H{o%C)3Z??B%|LX3lp$d`Ru%VOYl6ZHNLO&mnItcM_XK+7sk#Zlpa z=WDY~u@AdoMmF);qghsiUhu#AZ!TzgM8i0lXlenx0`1vzDBmwVeAZ|x&Zq$l$Nlgf}QPe#T{lk5pTN#CGo+ZBm@g~oSM+5R~cR3)>j^-f#<2gmb za=_7uFB$xT;Cqh3!KIge8Ui0iHA5u^Nb-O&m;0a&!zF2=eY=ZW{Z&L zv4{`P9pylNJRj2neE)AIPvHT=E%>hb*?<{<-hkPFK7a_F>;P`5D@F^T!kILEm5}+A9Zj)5%=vK+-JWq6+3`i;$6hSE^yNQ z+D`7*cXGe6ll#q`+!rUQjjNz{C-=>r+_!XapTEN#g!}xyN|W33fK9LoH+*6@!`FMY zk?yqxzJfx~lkGJ;?{8dT`u!EIDP%p>{*3Yzo}T<9u8j-8Y+QR$$a=~#dBC3;S3-Ti z$BfG$WQ-X2+j$x~bf~!Qy6eQunKMOla{nhCT>Ij>RPp*m^c^rQ6`h^AH#Ddlu_kbpK0!oxITvdx4ogR(Fxn%&Hh#T z5&cW~;5Oig))Etm-?42!?Z(wn7Et)#b?kwy$F?428@xU~<>m+!I5rpUQWr@t<7xoum{?4T+=Ok;IoFTRHGBf=Pc)5kL`4nGjU(q zg@){b+!eVevaj?Lm>(gVk$+3iM~oODqN1Y2)TvVizHFy(gMup^&(6*krKP1}iaM=qEo<)#?9?EZmUZkiq=GG+dj zP1D)lv#Z#Crj+0O_+zzDT6~8P%Y1#McUNC2JoF6O!McvKQHJH2d%E~U9t9~rxj80W zw@>~1_ZNc)4;B`SMT{9UMg#{3i-?E_#kl--n|R~l>qUM3?ZWRRDXw@$ieYa^5%`)EBj1){^m|g={H_!sA4qZiPAMku zmLd&#`sY&I^raND8>N_aP>SWtmy7%FzhA6gy;?l>*kfYNnl)nW+O@*raEQXfLM7ko z>T0oR(#x5q-g@gT@$S3titXFCiw{2dKKf4)vD$fyPY9_shAV*IyZZl8^>iB@AS zi~4h#5evv(;sMlOi~5zQUyJ%XP=B9W{Q;;SWW^a?8|LoWLaeIB+7s%((k#Sw)Nerj z&rp9a>K{P;L#Th$t-k*NjBmlv^_wuxV7;q)74~XM5Kp`cBizwT$cAA;emYsmeG7#A zYPFEf8-zT%y%m4|^H9GZ>RV8MEb2$0KH6$dM*StdgnVF_kn1K3S+zjOmsg|B4MOhU z-c}#u8h=0Cb_VKqL;dck5C7AQU`U$J5pwJ>A#a{6aZzX0{0M*Wvi|4r0KpR0GF{(jVNZmoX>-g0q0T1Z9<_o0PCwD1~Q*n<|D zdr9&AFewgCmg49FDSlin#ZMcg_+`6nt$!iv^Ec80QGW{R&qn=4sGo)UYkEmhG)#(( zlcjiNffU-UkW56*nVVW zL||mh#pm|+!E+Sw(+doULH)pBHV_gSbK$w?_I5p2I9}Gz3?w63+mAW_ymQaJevE2h zu%faro&*0-)(?)v^YhNV_<9|;34@3N>c<2IM@@{342z8E!v=c$TyVi(di3aV(FEei z=NO`pBJuo$0x@I*pNE7+MMOqMg+=+EPyo*d4pQxpiwcbliwcbze@UP7^)Apu%;kPV zCV+mnAELT2xE3?o2Pp6g{8Q}@yEY0ejtYy6QbZ^m4gLfFn9yNCL?JRNDyGl)@#mg< zE{JG-9&^GRa9O)URRI6Yl4G^ zd-UvL35JP}Ph)}pI{dvqNfA4N*Xj-puU`Q0B zA_JjFZO?6^O%@%!IQ2%yK9%L&_hTJTaO5B>pv!h zdN=9XvrJus_&g*K>WbYIJyE_f(13`bGkcvqDJCQ;@LK4a@s#{EQn+`oQBlF*V}#pd z@;_9S!2PSQ47nDj;{F)Vp+9Ip6p9!THu9QwkKNjbpd*5>Ik7hE8DjZ_}p{P3C0Q^))tE=5N}i?R(S2T*Tj48y{BS>eV^|ZpMU;2zU#kFeEs#;;^4u9 z;`{Foso3Dikt5>QUw;*=f0SYaVukI$xcOQCz8L7P#y|(V%$|*bZV3juhcVCsFeURQ7a?c>wm-`{@_KbiaV@#)id zkgxaIXJ3MNxOYFl{{08`_wVK1^YXKB-KX!R0|%mPFK=J}!Ty8AMg0c(`Cr`KzZVLe zeb!l*T-3L}pZ}TN{VvgOn9Y8J{Jp$-_@U0_=bU|3AD;^b`Jd_K)xBGnEneA(c^eFrQ0EIq-h`Vhi%5sRz|GyeXmWI}C3>;e&cYwm*2?D_vk>WunR^-X-wL_Xiw*S9Yq z^JT~bx^?S@8IXWyAg2&QW6KZuEM$Y@`1mhLn{Pg-Y1-h?qenyY6-^j4R6Cem zDY^X~{!!K(n9mt7V8Eq-0Jv(;nc=F|!U(OsiZrqhmWwp{e7sje#3?hcRusXGj}5F3%mR7yC>Of zw$N3pR#_2O6SvEe2VZc(1y^EihC!Z)hAs>Z1Wz*Q$DF6A-f2wQJWMsrV>N9li@OR&R*8;u! z1#LF{{PWN9`|rR1;lmF z6C)FU@`}*-%xh+Y4npE%WZ*1~44F$K8}|(ONAQ33)mK+Sj{Tq$D@e;Lue_pYpnhPz zNm2$_XHxn?9BC7ec5G~jYJH;3jR_;9*PF~9Lj)vB|Ve{apz~s|Fh3NQ<%0ZxEz>8A$drrz#yRbKK z?AY4VK^|ok{_yqE$i&~shAa&DN9=@lKz*6{nUvQxNEz}W_)i)jgR9<=ayV#6`55~h z$FSbLT}sPKO0BT6cv`79)=8<$E3drr%6`y=6{G`lQ1lpd5R#YFBl5o^I>>kOoHC-` zQ2*(7F=tZICDZK4MsNT=tw38E$(#u^2;xB z`?Kq0;nGO?_<}(B==>ms*o|LEIeo8`lN+UsA`PGD_6(gFZP+t?66~2iX#nhi_B;go zM4LX2eYOs0Q20aME3tOb4}N_GX`w$PG-x2C-czsJ(_!crb)CAjd+SD7ym++CUpPiS znKo9@Kmu+#Amt4@4c6UKP5@5fprO^CNrP_BL+O*?Q~Sdw^#u*SH7?(ZW|p_3LE+!8 zU%yuWk1?L({7Lel{c=1o`n0fRVgng4dB1|St%#MPa(b65&rP^O8;SNs~8W_<_tOv zS||gK2ktU4Xdxv2^pl&G2g}Mku9c;rp%66Wrn~S@J0xZ7*Px+E%9)^n_Dr7?37=%x zGifmFnKT&oOrPZUyuu%+3{ObCX=fZ?5NEY!(qQEFbQrecP6t1e2Bv2+N6Dv_UL!02 zdaW$U01b=A$Ro{P$&X>qq#*`0V2@4CbhBsrB-(Q@Y=J&006vN1v*D8lZjtg*7yj-2 zlENSH{mPLeNA^REvVwHbFA`E`NC#scLej!}9BYlZjL$gUGG<_U&OSI6i|0BgPo`lB@Cw+=u_CzcBRwnrp5ZKv@`>_DZ~o z8*w*iq3@@^=e&k~lVcUfHl~;E3Xso&hE2<^k=0ADmF1wJy*>W}K56y5V7c=4aWX48 zMBY6oL}tc^$_E$TC_nqSUXlj)7(>}Jd=gVf_`}~T{Rcj)h&%BjBu>Pe_CP)xGNF%m zkB=D-kw)tOt65{^i=d$fG;F*RG{6>U&!mC&OrNv{ew8$2lZM;I$@`K(Lt?0079XZ) zAfAjdI6fOO#z>5BNGIt(eFn#1#u3a;r&E@dl}W~>MmFR@cq%)=x~!~!GETmH_to-+ z<)8t!@HA+s((Rc(DR<#mxh{3A%&`T^M^b|1!*he>gP`GF&~O)MxDzxmwU04wg>4|6 zDA#%&z0rGj>p6QbedoF-aA`MUIH2e*EwCeU? zd4IBt1|q6r48+?Lkk`Rq9hdT*cDm2Y-jq4U-_YsAq?4taTba?+$p5@(C#Lk~S9v8N;- zdE^m=59jyfH)-Hlz;S|l#mqSjX`nx)9&#)sk6wE)RPHDy{zv881&i8Z4A=r=4B9h& z(&HIXME^JEUs(IQA6uczsXx&5)tL8>#5tI9*s#PinURqpVH1+q*I$3VihFP#M)^2W|M|{ldGyGU_aX1oUVp$@O1WaiiXU)BN#1qWT}N)b@kTj! z?p(!p%qx{Hkp|iY*L{c^>1a<2*C03!Q|}og(N?$~u?6!<-eZi!ae}mTkideUKBK&T z@A?nhg6%r6M!EZrJMK76JaKMRW!khs!=OQfB+jlWIdIIN&NKEg=peqtpE^%EIX57r zJvyB5j6#@jhcJ#@z5Ff5Y0-RXAt(Fy|0|1b^y1 zGj)mU&7_IAam^WL@09(9goMbTpdg9UDM}7}&N>DSgoa)jGBNCg_CeeYNIl}5fqKO_ zkg_w@H-JC4+j=0MPE!68bg#q~KtJ$s1?~C87hhB~kUv~EgncL);^N}ugb5RrZedSe z;ZI(Z=cI*r(+=o=c%O3w>JMY>EwBUH0BwTp(gy?g<7j7vkB?6fZz1re z-{d~w=_ZB0p{IoII!qdjOx#Epb%?fb-+lK<_$Z-#GkS; zVhHZ*Q17Wr8O~K-`V}5zLYBE#*IqyTYG+81r9@J*%jf9D_&)ZIFDT|Ddnqc;qe<@^dQ^|xh;Y_p| zuc%|JL)wThb!!WJCuuQoM|@iJ2ld^Bb=zOH{sa3`^IweLD@Y6ZPiW9VXwX4Op3=Y2 z7MLj;-lH7oJ87qs8GRA=61WFNTF7_w@kjhj8^$MoQurJCk62*<`NVuGnJ~U*{6o69 zFUL&XB7EbGHkg^wyk;57 zf;vw=lSb+~^yoNZgeJ&}<0{Ws{GPFi`2AKo2L*Tr;oq6r2hTq6oW`F{(vHDp%i_hy zL|1t{aPi_sd_Skz1@TE;|LdTK-kzpIeES9cA^k~cv}d&MhPR!~Uc^Q z)O6^f!>&5)sY72K2I??VhtdT#LOeg@ z_Sm=<%sqwS*xR`oYr`LXuTVM_4y5&>E?v4@jk(O@V5sr7950M+KKo) z59?&NAqEab`C-WY5N{lsgFO!JlW{FH22R}%;4XbkaJ`&TnDBoGS7(P57 zeYVhl(noT>L>!onJq$hOKMue81AOlL@SD%EVeXl7zmfY~++X3EIQNG>JS=5BXg`g; zUSIlYtSiJ}?P&&aVC=-XC&x$nb&f^!)r@5sr|=m+Gu?!}D(-o5uZ?>=Jiov_D(>%a z-=BMIOq~&r`@Wn5U=Ess`NC}+e>tvkp38n2N0TQy{VH~HwdnfG{jwXecFX-f?v-t9 zc7NL8XKrG->x1f_^FhW4%*272{+*dT;rR8<{++V;?ip%dg6rW-v%bN8*H^B6)2YCL z`@TE_!L@zv`5OAj{a)@fjt2bjM4j~In1cCh9OB#=#KXXa@@EW8TsZzbzh<6V)8?Kt z*W``K_o%qn$$i2bU?WI%x=egIj$_>=1>e8Bjbk&j zfdgX>@`SN<<+2d9*T?-Ct|KzBZ)Wbja{rine(w6nvn<@B;a(s2n5d6~UsP3(8vQd4 zGWHiat};$D;x_{q*5JA_*8=mgp2#18%Ng_qobNL}C*8z_JRsbWpRD!`xnIS- zTkaWh?|HqxUr3wa{s{L9?bxF-c)-rweKhxOky`biwuSHZrojK-Mq8xc;k!1ZbBBlqyRHf`7h_nElY$9=L)9S3%1-R;`PzZre+`jK~9eULBX zOX?u)kTE|Y@n9xDn287ZL^+%)z5hy;+B4-|-3s8q{nB>nB>%_$AAIn^xF?@{at8N? z7>^SV@|^Q(&Z)U>N<7F5`XfGL;{Gkq0T?#Xkv_7Vu4^3uU&h#2uSmgq%RKVo;fEhq zevf{a;~)J$b%Fe#F0yap!bF?+_VW+qH=ln5n`ozxunFoTQ%4;e`p@||*EzVROCEE6 z1zS|}E!rk=GUVef8_MGK)mGWO?*n;g??-L=2%F&kHTCg{WjC<<4#$7^R>s@D>>qm( zdjbO))h$+g(|D(XeBglx)clS((J#^thzEH<`H(k1H0_k{<=!rzy*pT)1u*n6Z^?N1 z)IV(U>wO4L+H~y%HjEuR_DZbxE+HM*$DBWF)~q-2{cU;s?YFCSV4RUuYZu1chdd`g z$bZi7s0-u=@o%5lCd%!AZn=-OC-lz??2Ao@-K%*Z$9nF|aBm55mc+SlMYl2kA|A}- z2k~ex8{)@ti@t;NQZ#=RGUK~bC11`#5r2Hey+rOy;CrT$>#xLN!GZ;f@AL;82idlv z1MG`_!LU_~6UVSGGY7hyh8XEMI^~(C-X5M*%o6v@|$b9)IqMLb6&Qm|Jc8c!}sWBuy3|WITHuGCqUx8 z0&?QSi3&I3z|1nV0hXiQFixVs;JHMsh3BGO&J|84DgR6VgLCw^vAx|`ljoREpToTc zyt6>e97 z)H%wFye8fB$yj?S`kj5N{@D+Fjq?8;;bFwU_|AdE_Z`(*G2=q6IZ+?~MBl1^(rb*r z)Fsl(Hks-3Xmhk}(n%fT`5m6U!1n@_fB1uRvu`F}(%+u%?);|=nCW}z`zU9;6N38p z4c5c%M;lx}`J;*7HSzv~nfCQt{;gBIj``v{yISk_b$#vN1+s=LP9w9NQO6fmm;-Uo z$eH-eEbSe5{A zF~795RNY4z&;0!S2&`wnR8di(K`t2Jy^8x?TT5fm}V`6&udh=Wm=p za{kJ>0_Q3g%$3#vFYdK+PEOpZ|LnJ9S%ivle!v{;3E<2&I2Y&Ki)#^__i^4CfO#j^ zs2NicZ~84J&S5x*dioy=B-?OLU&F<6O^53UoU@UCloQ7w#z9Qa-y5dlEzUnU_u$-z zu`cHuBPJi-}9E4nVIU02=~<}OX`cke^nTN>R+x_asQo|`{;?pZX*9ME-1gnJvp8qU|mD! zHY^(}ziND6@so3?dy-ZEhYkIq?$OuN?@*Tb?z-YBeKlqI*(-VSsk^RI=egHqTrb~# zu9)3#GV0^ID=Jo_|DxW|Z!_kiJm^o|ap7l<2-WhS%Zj6)cpW_kd32d8w z8DoE@dwu$8;>~u*3p`i$#4(%eNvz9xHpg!4#hg%=?JVnB9_)=QJoGF8JIh8HmS^tC zvKkgL&1P>vPxS^i4Zrm4y5jrS@CHUR%Xq1`uX*5^>x%DQ^TZpCcx6^^UhATl#1+b% zp_eqPx2|=?yVV%(%@9*X4E~1UJuFl4_LxaxBCaFwcRW5bz36!ev7x5^(IN2@VIUtO zdEy_?Iw9dL+qgGXq>EI1_c}$)L2JokF5bNtfvZF@AEn2sG8Pel<2KfX{PONpi*po{)+;pgJ98PQGVG zFdc#~R+=G&)8|)!!fle?HYG>g4B7F@C584l1|cNS@IIbY;1&-q%?4JdFEdVj$j1}w zx94U@yyW7x2JMYQ&jyxl8pcIO!hTkq6 zGjP~Y%RqBN%Ix{^`0>aw1E)_LKO|tFIV~MOau_#veoDfaflCt71_q8jtNZAbWj|F9HPaGiFT^hpF43=XCpBTgLiCalZ%s{J zgr9qxlaSW>(EV{(2CDKW8z&?zN|Z01B`l7}nyo9FMyVzW8DGn;O7RMIbitWYO#g5|Y;@aZ+;>P0U zVho&4v(w@Xa$23SPMg#2%yv4Q)y`UHy|dBT>=Y&5CFT-KNl=NkB(}s>VlT-qag4l}VMh%8W{TWmaW&Wlp7|(pgzuSyNeCSyx$K*-+V7*;LtF z*-|O0ysEsbe5%Y<`g;$ zYYOWM8w#5WTME63e2V;v0*bGK#W_a*CWqHAQtr4Mj~wEk$0%KE-~;0mb3P z(Zxx{8O2$}ImOQ6n&P_RhT^8;mSQibkJHZ?;0$+0JCmFl&MapRCS5hoI%k8k$=Txc zD)A}tD+wqGFNrQmD#<9xD#?U)(1Pf)q_T{%tg@UkXITy8+W^_NlzBm_evoK*d31SFc}96wc@AV-1C42b zwzQOcK~MakBjM1GBuFp|(sM#`b&y(9MN5SjRZ~?9WrG(FX90LQd>np`07tkZ+L7eQaAY}h98O1# zqt4ObXmYeTymEbV{c;0x!*io^lX5e1vvPBCow>EppGKuS-q0HhbjAvOu|Zd|p(oYQ zk$UJyGjzindSQW1SgT^IY*qHE>?%i9byaOueN|&sGx0D359o++cstAvizCQkb;LSs z4!a{8)8}eOt)t%2=xBC`T<=_St|d1p*P0ufYsvJ1(n{!2;cb+-V zk{6U`&5O;m<=OMH^Bj5AdA0CejqqI}-y5FGk{<-m75l&5Z`t6dvf-nu;i2l`otjHU znRl5PzA328S{7SoE3=nnmpRI+%WC0q8sTmD0%=S+L<7oJm*ImA?@Px# literal 0 HcmV?d00001 diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/t64.exe b/src/fetchcode/vcs/pip/_vendor/distlib/t64.exe index 325b8057c08cf7113d4fd889991fa5638d443793..e8bebdba6d8f242244bf397ab067965d47c5093e 100644 GIT binary patch delta 29243 zcmeIbd0bS-@;7|Wfq{{2Kp6IYQ3Mwh7Zg!Y(1Q+&JMJjpz7QN-63t)+qvJRR?I>nb zV@x#BBqo}OYj8)!CB_ApD8^hphzm)KVl>YCt#bzS-kaa^yr0kWzW+UZK5bQ3S6A=V z)zy8*qP-49l@29c`T7+}sbl}jo_*&7x2T4ozpy&$Hsa46K8-pm!;MiZWLOgQ0Pw-B zkD`77yz_x3s!XP@xADJggG`14C+AEenJdD4V=~h?F86H*j(2@9p0oYcap9b^;KXrt zAbhJ3=l5Vu`PRb9jjRXXQTuCap6f~p>v`_cAixryOJ)|nw@@yu2=ixkyt-r91JK03NZ{f`7-<6rjl$uRSO1bZuDo@=k~P^G zDJYtA0qnuvL{_;rs1-LDq<;UEHMw;`-n5xHNVY>dCnz=Bu>G_U&abrZ9e6p0<0Mhb zaXL$CWSnRbvqg)x+J)yNjR(uI&kEc(1q|m=uU8qi62qWCcFDf0u*#chRTG=W4dOY` zl97-W|EtcFmmuY9*m704b1LME=8OdCSs2@+3JeZHT!(m?Wqu^T-yZc%F<(kIpGq|! zm9zouPgS@s3L_xqSkq(w7VVzZEJm&OpkM9HC#5Y-6=5ls{pjUqi>#u#E=40gPm3=% zUy?p;p@>d3A4)f063wTjP5vy$AwK1^Z)Aa>h_NpqQyQ(|IJs?v%GWiuR;n{04JDVA1_gV;{TUY>0r zMl=tOoFtlcky+*{>4BEHtGfvg0$HZo%dno9&7&eSpWF}w%EgEyqVaIt*fFCIRaRDt z#(UAmr-I(e_|$%OTjNuQIW0xA>xBO3-$lpup_<(tOu$`r7F7XY%B~Xh(C#Mq$EqD( z%W-8i2Out*WF-%Rv{2D}*{a=#v_<9m2xGlaXiz;qWLM~?u$oV%n#;{+ktKAbXpL(2 zij2^aB;+8+Rf?ujk?LVKVWP!VvRbW@X8^V0q-d^}midwNxi01ksYNJJt7`2<<4KIr z8R?!s5e`v%kiSrq|2#W13;`1lW5_dD zwXfd={wHvqS+Z)UQS6Z%tF~&#Ach$x&7eABW` z&^`W~EYzy)h`s$VpV5RVyGV-3)Q}Gc>8l zdk+AN57ITMk24nC*9WE>|F-JAF(Iwm$2A<+SlV$!G8X+x49{hTN~`vQ!k`qo84+>O zV|}1#d;tb51FLq2%-}4y)v8^L7%MLgF^J~FqAAN=H0Aq>rn#X*QU<3C5iOhRAxKNpX)}2HihxXX%Kr^*mG^rNpcuuqGfb?5Ejjro# zM0FNS`aLvde?VqLRYjntnopW9=}d-Ek)c-YW=Wo?&Yi(osIv60uN7pub0JsSq=;&X zObcBH;$JRKmXg>Y#j1@)epF*J*fD6cks>{$PBP_95rM zafJ(L>$QK-+K zP`;CrPNAkK)Z3tHc2ECMEt(3qiKYc*qG@!wXi<&t%X?u$9fvX!O{tNh2!))TZa&Xu zI(avbKYlDVG8C=FB2T$P{ZrW2s$B<6H0DOep+%Nzq_VG_0u3>s>XbNYSgGPM84t%F zm(HPuqFMV>9}JJ#^)LWt(ru2LU|g`lzCbUUhin!XozqVhbGm4DrfoncS?$Hd0|g_n z+&_Cp&DHzZOO@W#e`9=gmdF=aGnn7A+KT39q*;xLoZeS77!zs588OFQu=>RVPwk{N z&8VV>)S~g28aWCskX|)HlC(hWScmgBpLhNSOpFt8WgCjGo+NO$tSVce_ z1VBlH{_Vfeh25M(or94gnx9&=Q4*W%+&1$UY*eD9uiC1WsLRb&nBRq!;-ag1=XAU4 zqWMY984L}2`#2a=%4l~c#g~IAo?@;9{YTIjmFvvMH18ge=Kjghzhn*kaeVLKe?k>){sf^aF?#z4bdYxV7-z_zTCGPlL&#{f5|c>M3!<@2uRdz0-Q) zMW^Nd&R)1hJFT@v+x^bsG#w|eCVuLwYRYo&H`3$tBa_XiGhiH8w3~<<2|poW4-_|s z$kkaOeFGEWyi}r+SJHfFq5EkNWU*wU*}JLjQg#S&%LJ?T*e!NO6XH38mrKKPf;u3% z<9=hjd!S&r#X{W2r`AHU+}8SJ&8nlYKTg-7KBj4r2|ANLGEQf4p@(?jfcu0e2dvgY;g3V&f#dEquZ1&Lj}ER5 zF&uY{!|LIrAKY1gkEkK-qiMobk>1Ee4GLYQgDr^17d*8;Oi1nAx-eY&S|(98^o%k^ zbtt;L|9pupa^@#uMx53YPSx<}B7TIOJIerB!H9K5oQKG#{;PDm_aF@p>aMzMReW)@sK2^V}iD;4`b<9m(GJ4l}$q5_@KJ2K9p~ zt*<;i3nML=YLsOP8cd1KN@BDv(XkB18dp18-f{o%Lk^P{FB%WRa66zDtta(O(4Ehq zYin=QC|k6(;7PNHcIoB{okjbx$is>$7Z22{j5kBlSr?x~pNW|qSCo_$k5#2tNl&Rf zv+G-Z*fJkqH<3z6yj;)?8(?K1`_?B$@Eypm`}q0&d>K|}gZX;J_|$8Gr}2QhbbA1E z_x1DLLa8v7E!thaB;iCs2S`W?WGTKaV-}-8R5uRBY%cgHud8Nvfv*MvYNk-4r4J?c zzRXH}dkR+q*frk(-vQw-XL^$+99J`w)WF!mOyk$a|1{2gx}0hZS9j?}7#Oi=Xi5AY z#E4LF}u5w}fk>S-Zg3yj-DY)67S8W_{$O zbn`_lmBdkOXJC-9aS*#0*xPX?ijs4v1*} z5ljv#tRlE|(!iI3ol#BWY3riLg)yW8l2HkBRff6JrWdk^_{pf3df|^C79Wx#aFJ|w zNQlt?94ie8YeCCND=ozl!`34`admeg}TU$eKS@qZ4S|C#TER zl)gR5>e?j;1t(cde1|S!V2VFZCZGlw$q}vE=VxX0<-W*T=t8FtM~rZO9+J(LYSclD zuaqu7WbefXPJ89-^b~Vn5A|-2E-Ixu{NJTeVO6Y363Xa*8{t;E|mPBhn`pufVgm76c8o9jjM?^f-|v&^%79@`P( zsZrs$X6x4d5XLTs9PZE#yM>KbixSY)H6v&Vyan^s_=cM1{{nksjvBffCJ2maUOJh> z&V+^wW4p3HLNkPAKCDOB06~?*R)w__ejLw^hs6k^6WO0(LlYKa53vdF^0zF!)?X_7 z-D)kNh2e3jv5R6f&`v74fwhs9xJd*Eka1 zLd^LQt0Ni01FnT>%fLn4C5_xt`ZG4*THEZ430PXUu(Ymb^)P4UnhtzicAjSZ|5%RVu;-Ysn~w-%q&vhLSo?H+XG+4SonVG=xK zWYfgamOo3(xrAJ0urJe1_GgW%72x{v`ULyw-7z#iH^P}VC*E&s8nR$T`buLw8fTmj zOlx)_d^W#;=^_I8Y&JRKJ(qBIIwtkvF8KoCV-u!G4U)iLG;Ms0!(? zDaPuD@%K$BzA46f9_GkwI)D6a2OH_6^cfo))!ywX)5@7%ijeAH zTa%ntIDR-8pYeK?@tIKgvy^=os|l7Z?KNkl2ehQB2&=x!VwyEK42AW9B@)yrh?LB4 znqng=j3Z28yx2LYDTw+)DPRgnCRHR{CDNryH01M|NDeYmI1G881tKqs?MwcIUgTms zHIam3={l2hDqh-0@x)`Z{Xm&u>^d*0P^0XBc9Lcy`>EN0$gPPq<*vdCUi4VAsvPst zQ@RH;r@_kH2o1o*=B;9_(Azo~;JXD>6lz&o`YR5hCI>Q_yrdP#Vl$84c46zAHw$vY zY%w3I)sm|!R5?hKXF{4lT832z1u|>%Fn9Y`irR}578l)8sBmUuq9b)1oEyeT8i^$5 zW3{~=r}!H6U8G4JshKC?p(;Rc&`4c>i2kjR;Ps?lQW?)!YduAC-)QOvC&~U6J00D% zga1o`bObRUt8E5>$$AIlIA5-|PU>~XrT{xY2`dWl+%g*MRUYs-0x0X$Fc7uunR z!}dyS!?VB@A%;&u+=aqXnJQ@waU>qn2ZMVV;pPe)TFT`$>txlg0tUN7K(C*X$}xR( z<{Py2V3SWF9c0m-gwt6xyPU{?I_hyW)lbb~lwHuw8PG5^*#t7_8~t~piS1tDYY3)9 zXt&4EvWhOOS%?}z9p)tlr9!+sj1hFCRo&hXPojPN)!GCxN2UD%on~=8x1Y^z(bljI%NLU@ z+?b?q%5ws2i0@j#mWP4ENqu7*toju=i%5@f6i`^xZLCX)6}p9!LK|5p5NpjeOnj*p z$(S%syI|Phj1WUgeen+%Mw2QTo=fS>RL($=dmf0K#CnvO)x^92%u*Q^m`p;8IcW+= z@+^=f6ag*ih&ed~M4ef+602BeQZWFyee?ik*D18J-}{*+8}w=@RzGF9*8-xf_@~y# zTyLBo9hr>d6X-WVs+6AK-x6g{h8l;P8odb@wPc>g&cp;ZyN`-mwFY?PF*~c!38b;J z$ZXmDElnQ>_HC3GAd}s%5~G<)Io5go+&vXS`~NEs}55 zF5L$euX+wfr3-%_p!htdfR4$q1*2ls4nf42bP-Hkp_9UQ0(@3&F%ouDS@L3=v5y6| z+|}u=mf+I|BlkRD(vGR}pt$W;Y692oc#zF~?G=+Q-FzY4?u<0Ufw{DbcGU01D9m)& zp4iHn7hVryV(Z=-EeuJU*>oD}ef!$hX}tT9J+N@gU}K~GKS>|F#{Ov2cIYyhsFH~% zr7_p6R!mDLxP+WEyB{{K?T1A^PiCR1L<%hNC5t;)JZfXXGDBu@w6XYJW3R_`cVCxJ ztm^22)jK*@0vgJHMC%2AfUE4;FvTq}2EZ!!| z)BcI$w`lhI9ZKjTrl8IobPE)$`9B^fL(9Tjcos>C*R9rC6|`FIeCx5Ac_?yM7V=A; zn1*EP=3r^7zG%+M+j=J-+(`u#2E*1}EDLveDov=8Mb9D8#>i(}Z6x?7E2;H#EVODP zkPW2|ZqB^hwF_3aCc*OcM6!sJRl_8QzvjqOj=kBw_r&f2U_n7mU?^Ss3o{jW8|Cg~ zbsv(KLdYxWdm7TcHc)3D@&;>)K_RjD322ps|xL5sd`F}FUyo{qRf$x3s#jA&NNSY`nu4} z&4KoP-xb}TQ=e|^`kX6_-9=ZG(gxP4MwNbMA6N$l=+gE=FO-lPyi0( z3D2bx!q&-J;wz_+(l0LK$}m8N%L>!bbFiox_lFRt7j&^oc+|Zt1x;p z^GIzabess_$p)o{3j9P?n7Tpu{4n!N>mdC8fMuq&?fk=i=%pTHy9Lx)oeRA&k>}Tt zTpg_I#0qS2QAf#cP^|`|Xj%+)j6~Iv`)qewuV5Y0j8Sd>2GRKJMPbx2RXA``YF2Ig zPg!uU;E?*~4Fus+ERgfYXU_}0a3YHgJ*K+$oMR(DyEkx!lZWtlDNa;t`qHFOZJw=x|PI#oG7YW!Lu^$Le|qx(yDcH8XfR{t1e!W&YwJ z;Z`}jAubW#%Vm?(BMruX;fC1S1YT70a%xL@VkNclAlhhJON2VB-Q%|_fC!pJSdH9c zl>t4@YZjsX+d35-jcB7I>Z_7Q?`P99e1)F-+2V{~A!0wuCU->+Th%nuc(AdC~U>s`!l^8dJ6wUnV&kNbrS)4yj&`n3@PtJ&jz zJ%ofB)}?^a2gP)Au4I963bFz^Oe{nB{oBegV|0MuO|0m7v& ziFych(N)c&^E5tYR|&U>R&C5`c6PA0`k)+fSowqJ_zsPBkfu=P^*dcg*T0vrW}FPcxK;70AjSuBL(WGuoTi?mh;iBY~y zoM7s(5B&q17yizbpyCD49SDtqguBrC%43^%*|=do5r?6tE9s87md@SQa-4p<&hAZw zDk04`G{(0sxJlyZQI*zkl!<&0ZFCbIJVLZ*JOs@N1aP0W-)`6cB)p#%qHylrtR49moBk(zc z!E;0xyODsIBedEOVzexw{Fs1t7ff1Lq`W?&r{^*3>~v3`_|yC)wGQT>oqxmeIU>em z%sX;lr-?ZaF%#E->O8WQQ1uS$GBU!hC!*OSJNbRS9GT6M?5)V4HHJGtxFWt-&bEzg zBeZ>o-5eRO&WEXB)h>UB1&-?8>OIWQ`1`PLQjC?*45y{`-;#!qZ`h;Z)r_PEF9|5y zoHXM;TQaK9{|Tnl>~NaA-=MTq%Q^|&n3@=L;~w)FJ-k&}4pl+U-v*vQeN1Q7Yf28h zdx@2EthMDBevf@LI=hwizjORrE0=*BYb{hpLypQ?7BePeoC!wtf;=(j512@!(Hha% zwT{zg!h=#LJB-|=o_KjcE*l5yF&vR1jnBO1(L{%F4#3dD=PC!K(<4SW<0AJT5!MbtX)lK4!09pn1=3V{c}U5PYlHmF!TB1GY)g;!+`+G*)es zrL4)gVSP(3p|^32ixu(iC;+a1Ua!Tv!26e_RNqrl?=Mg>V+5Lg$X&FaE<6tfn}bZ) z@#<;kayBONg$`D&<5IR|TyyteO0qkSc30q%;_MRk%eZhM_9k-}A0-^Q$XbmL6Hfog zGRL>*RJuf7q&~Dm^!ywK&&nU^Y~h0T#XqifO*f7Do2zw@qLGAq16S!HO(bCn+cv(X z;B}E*AKzN|a4~bs>Eh$7LYeRbYo!xVNw(vOaVHy_GhVO?>`=}s*TZMf+lm9Y^F)@H z+ah#1D6$85F|ErY#7w@mZ%T83V9M7`U|;1XICoT-3ge}g?6+L4zc=EI-rpv21nwhF zI`|z6&08n%8`*)pX#!is+Du4fvnR~;*zgTTu&=L}Q}_75TZKMS46{xM;N4ij#2zgj zH&f5bCp*2H^mLOPlTUTdSQ-ZsE2bGdOSL~jB6Zd9@Xj0 z^*DcT$8Mf}z^XFd4(V%N_OzO{m>Sr`4%v|DS_SJjb;^JexchXR%F4e_#N$FCd55nonZ;4X_S;_DeSt(B!r2*yB zw1V|bc4=Cq%a8NXJ9fW-y}p_`PG5=Q-k%LHPq#*H9)~8S`X2E4Y^rv zR_&vCglQJ7#wekYnquf$+T91k>!Va{VSZ>JRR?uxvNY|oTn{DMSz*|h-zt209z1bQ zO4f^W@fND{kSe`*txE&#;9TbNnm+hqM&t9ZxQm^5%V7OIVGnaN_Rbha% z9eca&#;`(0F>kn<6xdeA`ss!C<{P!&p_!+URMQefI9~jvb%9U(VY1VC)SPBye*k}VL*#pW4@-&GIi zEUfL6u_Bn-ma8?L&IfXOWAFTT+En7dU4ZnyL6^&ynPHDs4Fw4@dM~vOAg-wl#F@+xTUC@PZ zx=W96idyv8@FNRa7{O{!MF>aBnEUBS)>R)8kyZ-L@BTxBn`yn`X0q;Y=m)9nQ}(95 zMZf8Kyrf+8DSCkRi-=E9HH$X-5KOtmWBO=%%@Co>DaTq*1tZ?4q!_EAA)l%;rj}=A z*cBhAiB;=_hJhta&;Ha$`x5DKkAxz9OG$iL9#7?YZ|_?;iFf+(uu|N6actl0G-2ss z=K7;Q3!W1s_>E^tbB+oN4>7lMjsdvCHEKLrLjVAHyO05&Po zQZ)OYi#d-)_|fpDGoQJ>o~&MnG^_UMOg84cuX>Igxjd7-c|OqoP)nTZ)}Egu)IVWv zZv@!4AwO5EH~fWtPgwdJZG|hN+59(R?Y|)h*OoV8cq6-UAu`bkQ4Dl4`RmMS+sGi( zz#x;H-jO?3XBx5%&*f6d6F-R@@hN-g3juV;&1tN|{FcJ@Wo-V%aJFdv_kQzQ&>jo{ zvE;JPCl6S1%F;?{GMjlR#=aZo#rC%rIPjs;j>cj7_yemoIiYY(&05@W!hLiZBXx+a zf75^P{&lpm)msbQrB7sjn#Od^G%(eC2Inoh!i3qFI1P%Wp7_*O?Pt*2bPa_nG(i)# zAYW}I9U0-eg#};5+zkENwFH&!gX|ln*1&Y68>v}|NyDl*Zq`)Md_!k`n10}I+(>oE zu42m#{`RS;58G-8?znb)qYBs#6rz+>^(`2%s&L_?P^H3`jIG~aS=3(G9L_ox6|nkQ z?(U&DT;aL2{0PLXa@)#s7xs=YEybiB4BPhKe5XA@u{{)dBfD1D!jE9p){mE6tFOtCpT^6s z)v^%QZizmu2PPZlPnix6Uy|wd%P4#VfoFnrl_L!yh19m2H~V>sqhMdpZZBymd>;s} ztF}qDJWHgaC#;>Rv+(6sHr*5{412CMF^b?%(`r<^Gcn(k2IH7e8vhbqx@!m2PYGGPVhpn`Dcw{ zs+Y05mJP!B9c=mXNbS&(=*ygXOrp84SLCZ@={h^MJiKWLQb6uP?MsHL+tQakTi#J^ zz%Fe*QhSjFc?CGc;SN0dD>mYtZoV$3s88zQCO*9#B9U`163>0fw!IVOd!`bJCP$*q zmEU__pjG!cWw*>`@2v6*%mh1e7R>>VF)A5RoG+Tg!P4U>`o))BTjlNj?INXb zZhwWoX{&Hv{)(wr8-xq#%(VKdP%p9xrRGkpp{1#1J)sBSQ(uFr(eS1j-E~y8^n0Y^ zTseda4GJTjZc4hcl|`=^CPW=!3)jTCg=|G0>&f`?R8x>t@l0^GF_xPe7+MdUWZrgvH2_k((tybHxQo2DhMuIyxKxQMrE>+skQ)^&RP zRjYQ?EBIO9#GTNPxN{)e@_uZvKjfr8`AKIA@&!isEjngChcq!DB}%I!wA&+Q9!UFy-~J zw5gQc`ly32tv~Z$JILFoKT0;GRm;l1i#MQo5^6rM zr+Q)|a6%gTwy%;QRn8E62J7Ek0WaDWaEs=QYOD5FQVCZJKbID-W_IhMT)c{qA{TOH zH*3Fc+}P$fQN^?hs;{5qfP4!RjF<`f470Ojg@&$q2)hJtSu8`&D_67=ObgCSDOk=Z z-UnYwocAwu7juSGfS)Ck;b_$#HanAx@SJp_di$ewPP`}isAv^bHY6PH$UHvw6Mi|t zT7KNwu`d?56#Oom_Hn2XcY`heI8bPMgKhn|xsOW8s44zKu7x7Ct^>RKakbDiomFmV zsa@6wvZ?wVtlDj&+++JzGS^SS?Ds=av38$y5K`u{$)7~JJzMc|%(l;EAAB-Bz#g-n zPKk6vG}|96%2lghm^IQW9EDiK#>Ju6OK8}dphIeVV9RWs=-M6&XI3*Yhra7c{f?d3 z7~;CFkx9y$!ya!85pI~6&!@qi_vmS@wJOKt!foU69ZK(Xz_gdvBZ=&>Xo@i6!&%XG zNG#T~1D^&l^QX?SBV;zKUe1&8YM$hO&(oIjM8`VX_~BVOBmIRnM`hr;_l$HOkypyN zK8tyM)>??^z*0X;7B(MXZ+|vK)3z(kF1r7MYc4gH-Tf@W{p;1Z>Y$?po-e^FOJhNs zf;GF~4#4vl<8fZ>?3%lY_1QF8P)%UpZJH~L?aA6~P7(qwY{uqv;X)6#Yjc!vC6`ri z7KIW532gl5Rl@FK*7b|ILRB*R{tI2;mA$B3Y_$jnXfk=IwjOh=tVExCN_Dv`cuTyH zn9K%ki4u0Cu-CWr5~iiFgIl@^*Y~X48YtA|u;8ujg)7Ny@YZhaPhbjSBe3+XTiE&M zpJk2rPBx7~|CfX}K`Wvd&#^}%`en`3siy8SqgC4ChmPvk3sp^LNee2JtgIjLiyc2sw=!wE4L zVc;c6!IuzzO~FeDKdazt2tTag>j^JY@Xds8Rq$j+O)@Or}AD>!-jrDz4G7tKp*Fng69*yQNiaDzDB`| z2w$e)O9(H@lKbCWLJ_@^u!itS3O9o^Z8-SMdf!o)42-Urh;h3N8`;tAf`NUZvpmgjXs!9h9VU1y>WkL&4n%->l%i zgny{up@f$xcr@X~zzvvmaTJ-WBqR_%O~L8XK+0Bdk?_F^o=Lc<;8}!sRq$-W;}m=n z;gJfSPq?pw&n4VN!Ha+olQ<=^gc2SNm3yLua7n?}5PnU;*AsqL!8a3rSi!duUZ&t> zgl`?nmVO^Jpb%;km!EITDSg#qqU%(JqE{&Hpo~)};}uF4D5VM|Q=!21>1t6Z-G;LI z??VlNN=&Pyha$bHLa|dQ(V(am%5OucJ$>=M!S%)4a$RmHlmt*7DU?$RB^i{v3T2N% z5kaX|D4#2oNuWG1D5O$_ln>Ic3T2T(nG4Dlg)&Q~$VvsoQwnj6LL}`1V!1+0SBRu$ zK-{hn+Y^zlLt&rUzJq~loq3mO+;XsCH<)7qxh%dcR)|@!J-cjyV3@O-+Ps7m3VGt~ ztv}Xs()TS8lfRgt*Gg~;!>f*)+(B(YHmSC%2c)7_gp=-tRAVAepc!yNU;_cA_P)Ubc(Rw35Y=J29O;>;6HGBb`1~{Z>iL zkrS`Ob#hhOvWoq%zpvr$Ok|_CMfl7(#VqN}zw693(qR~Oc$m~V=%;s(j1L?aU7*gk zxPBs1H8ZFs@_=ZH84ZR-`#KWjdVEYp!INA3x;#p-*dMl(IYgP`GN%@)H1+14wuU=t$Y&FdVN-yrIW?Wt82O@|a!kFDFoE)wXQOx|FB8 zE~I$k8GU=aK8R%{Ah=bBwMu~#Ilp+;X?O5>|uFZ zw@I*B8`f^_h4A~dRzq5hM84s08Oqd zDds6Tg5BmRDId<3AB>(bGy{um3clekbd(DE03Rh9|EiwT6E!)xu;GUx7F`uJsrX0} zpCRh#9R|W|!)8vx`Ra?jfu89khakEkG_<^sOm-N_#|;F!tq%zyUqwtmwVxFy`k+ig-4^5q&KoSIt~7(nG&Y zsEPnuOlJE7a&IT=ol}YyaQ>R&c3_d-kc$YUVJw-#{7(i5Pr9>CCsW$RA;B05v#hZW zuQ+W-Vh#O=(>8ax?6eiJL9&XI!5ueDZmj!Gpb(`vZNCSD?6mz@p-S^!GPcWHS>(&w zSNaQGCbP^+AFauPau&cjjIa4^pDV1iu=$n#hS`Imt(7kmd~p>Xi{s_PLDXbuioZgs zOAB#XS07A+L|PW}L@WI~i8@jG9zZ!!D^)rNS*Vhy^ig91K17t~^gn#C;5%gKuoQN= zqw~tsZo<)VxFG2nk(P|EbVqK*kG$S?IWm#Tk(WqWKU`M$vy#)Ae$#{g?Se#yO5=iL z8eNba{r3wJF#hi@NP^h?)Bk=!^0G{x%Jkme^UOTnrF=R~fs@jb%NU*?W^38cXIl&N zrsJBVlc08Hy?#6@AF!w|B`FQLo$tKH>*No1GUiCsgVgp#1^p#3JNNfZZDQ2cP(B-*lsTl zEyV4GH%qzPshvAgXfa5m8fH!nY#6%1n4OrE2sJ)y#=qTQq%Oo_@Mh(g?d|aiC>F%! zkc5Sed8O4rA&!2E2lMt_sv_oJp!B9fmxjJ%)%0W`SHd*QaGQ!}N6P!~#A4R}N)Oig zM$h1PN50&ibUoy0oQvNKNye{qR7;!K`#0j7F2=Qz)D3*cIP5B%6hD&PxDn^VgJKyQ zj^(f?g#}cNUzDDVcSY8p*n3J{4FLe%HbD_{MhE(QQ=#D!5oKF z7jcau^v!_Ut5OD-j0vhjXDMPZdwAQ|Fx`o423(g^&@V8|L9BMjywPy7islQbT$jg6 z4 zz<^(wf)ohHdq>Fn`A=(z+;mOx6O3DYc?|*iN40c02k?CAw>j+7JE#2D4s4v>!#Z27 zWu(z-oy@B6X>6Gk9e21RYBorII&|gE3(0JiEPt{AHSL5Uci5tu$X2z5insVb-xM7h z@?UR?A|J7fH7$jUx0(IjnDD;}Fe8+2#1>yf*_dojyX3uT38Ifxsu3*nu6AnA4*0ak zj{46+8RVlWF7N!Lm6?s5(64?N#&rDxKP03q2@%hm>ZD`MQHHW6T8oyU zynHiNA${1B^{WjOj&@+vYU2zdbQ9i!?uNwVRJ!-HYDYUlx=Hm}7L|8cRKAKsajRqk zo}RlRW@UHRWgrSg@@Ls0cl(R>$9(oE?=EFBJU@W+qAHKyP> zRz|XPK9;%P^A&c-vXlp&YUShaiSya=2Y%tY`7d+PbzT5IrKx47W_ip>&_@^X!=6~`AJ9W=5Y03Z0G5wXK|EL(W3@2s0 z!6NR53oYJYz3&G(hwEv?yiQ1qV%W_4VJ)t|{trddkPalJu^F?+)>Z7r{dB+7*OVEH z`wiG@<;wkrlXM{gjSd{#>7QgtLtyRxS9eoqFG={vgiLAXD+xPa`+&c}r&qVdqQl<< z8|2z4aW@-`w!tnoIM4>C*x=hXxQ30Z^D$K0VvlW5O+TH2KVKVcX@gyDP-laqY;cMV zzGZ{&+h8=Q2K;&3po0zG>!i>hDSU8>(x>A1bFsll8|-R>{cLcG4KA_44{dOp4W6*U z>o)kn24C1uqqW4eqzWDjWRW2GzDM4z|Gr8x(DD zj1A5O97Z*J+eTPzgBxvdhYcRJ!7Db{-Zqw$m;NekWB$DjR>kw|rVtghK1%+bQO;|e z0<+I4uyPaNiZD-C^)rP~)C>?=LZTXeDV20)20}G{RixTm>A<={d@^Ar+lT-0lfNdX zAa9(0^2`~b6UI%RmX})^;l`&KrjMICdB()hS^C_`GeRfNn3PvASwC+4v^*OV_cDEE zq5hv3AfNtT^5qp2{4e;1a5K1Z+<0ypm&b*2`qCS2e0p%s%o+NEnbSfijhm4>Ew3PS zR-Rr@wNfg&vzdl})6RX>{MdazSM*MM@3HHi7dJ-4mF{TD>vOO5lz%ss+YrieM-Y;q zQEfQxL@38yM`(z9_}g>-?Y#4~*^sby=t_$Zk*qQTd zk9dK$ymY=NA1ZM3OV@hxeKH$zSHKT-6G7jQ8^26uKc3?}0##gK58ra7ERJj7U($-> zdbNI8Bi>2ioZ30Wm2TDYz78DxM%Aj)2VT5?(}o){51aQoj5KZ;d5J!>cF!i0quNNy5E-%DV^-aHxFqLp5B$?vJk8d)prm$ zhk-)r8BgBN&>(R%ZVVqG{Ayzox^bKjLPOj`W6x>YJ93(K&iM_3bGl}q_70p^JNJCQHchx-{dZh&?hY=v{dO*x z-^RITqBL}OKK6MJxll(Mh{iI%8p(00QBX4|lm2+>nF1VrL)4smdw0&gz~!t{xjNr5 z+aZ(WpocjRKc~{4JbA6$b2PWSvN^6Cp+SfT327(9HMHK4%WOon_sP1J0a7<)1LFhD0sHhbdXy?neXAK;)jz>L*`iE;}IG&H?`-Q zws+;4u5`{U9qhwLhBj2~>J%8NQ{h4X_hRBoOMLi1S_WVF@LdcZ8V633+XTa}<}?Fe z4*wT05`tlzc4}ytyr{6Ca%~H0ptjEQtVcPOM)jO9o8!JfxMSmQ1^f_#k8N7;S~aKD zBjh@AT8yf;okzKcR^8%1iDw(C|H)jeON4ed0i|dCcwcw^KdM_==gSXn(~xJ|5{~-? zLHz>mToJZ?aw!BTn8VOVX7e%{9XE$KkrW?bi$txwx@=a)y}TQw z?mNSA2M}n&9yn9_MZhb{XULBncMD<9-=!A=_~6urJV(!A68^+-mu=na7mV>=>0I69 ztYf(<-#*(eQ;4HgfyIf>@Mw-eJ)k#8MU+Md@_}kf4Glouiu`!5E)8hUg85#h(O!IDq+h%<7p~9a!*g@_@b)=;I6sD$r-Ds~ zRO9qnnjg%EcA~5$ICSu~kqgD|g*5zamiKv{D{rtsksyJ#9vsAdLH(B3nX2Bp>ZPW7lkDrp4 zqfeV#+9{m3l>Q#YH`Oc?dgti}=S|Pg9H*Z&j4M49&UeUom&6Uy=PD98?k^rpy4gOac{B8>lM7JN%z}9k85Y5RSNc4HFDX43 z$%i&QZ9h1#VEW`4>pJHTNY=6HFROBtQ$^G9zC$5p9bYUXD*bv$pzd+{BakmlN zflGk>J#cG@AI!-FT;U137W8)k7kEL30xtqQj&KclC7`PoPmZH`&I@on!sbYx%LR<~ zfej5Fg8luWM1f}ljtWEo4rt)2LO5;*A_RXyAfa3+$Autd;)fYS0slrA44ez&xIqY1 zm>y6S4xyl{0n-qQP%wFnOoW4H8DOgj1&;&l6^XLIF9KF0Tx)H>nF`42D5b?Xrg1od z$(admLZAw526S$&&|LtNY;)a0HGJS-?jD9z>wl90t6LK$Um|*gXz*ALz+|Zy-=t&jM^^6LQ$x?=p7 zfJxt#^ARXB!4GXX z!Cw%la$H{wHNsi&#{r56RlqX=2OwMno((t-p&EEDU{9OURLm*9B253K9kx#G= zLEQo}4Md!S34;hh$GOT_h5|aw!#IJi2Gk)?D+#uHU0G)B0q4Jg)_{lL`v|nmJ_NKP zP-Dn;4MLzsh5{}?SP%YffG-dX+YsRv;QHrH)D-x7z@_Mwv%r@DUN2HwPz`9egyYVG zZVwn@!a)>xDBvapYUO6Y2MAPf9bn2*xD?SVseqResE4irE?I>o+X}t43hTewY9(_t zpkpa~pkPu1c1IvV$$%>mNYJ~04>l+(q7Kk^Bgd@)o#0jkYGnnW~j!Z2V4jE#UZp9_*TG}!{`Cv$$$aP`Zd6$ zQ_4_g0&=G@^D$T40WY1#_`4%ghsdXAFrGAt08b(W0HY9_RxdCbj z@K>dxT*9YV6Ohvo&;AUHU;Q$!{u*l%jc!QK-+}a~mvP@Kh@(>*(#!9HAF>v*__r5u`rBv! z5aUe0Z_0BtstvHaoJwPE58A)~O+brz$kTHL++^I4OoX(_@*PP!ViUNTSc6;-nJEtb z<~Xi1LVH}J#6eCUPK(Sv!8D!A=f)w!JXB{KVtH~YgCL4hlq(k)*OX68B&ka`cQMw0n`zlXy>(M5?N4L-ht}L)`}TC+<5v-{5!;pf_scf z%$^uEamgaCfGchZi5k%u(~41Cf*6hSez$ud`M&@6KkuFM-g$q|DekRXxAt4NZr!RL zOHSC7d}FgRk(XB`r5gKWFZgX-yYSXv(7hRc1M#Ov-wQvgz<0w}DR5=@9l$%Go5HUF z{x(h@J15lMRzOPVPksJjcZYG_k&dPtVe1pvfS)14aP;<6knzRdp#)v-&wA0X1_@ zfQjriWK|l2MsdSI68}rqxI?HcaHW&`O{wu>L#4x~{{a~Fal(lC%%`~@y;XEf+ zWW=Y%Ue}lB#mm`F>@`h@qYKIxrHpuat{?ka;}>u<5G5g=R#6o|KtIsVPM(K23|Q zl75iqS&B}T4y8*!h|(#!(v#)b#HM`zfl?qSV$?H~DQ92^mAa+0CTbk=TczC=qgqhP zUjb?Z_0mz4f|kjlp6mykzV09YmxMDOtds2kVPPE7g}(6*-x8p8_|v*l1IK79|6katQSr!!l7}@}xs2P(wsnZIPSgY%`YU%4!XE zv1iS??#Q!;$_H1(A!?9|SY90FHA<}bB&_6-*(@%*U1CpN&{Se^h_Q#o^0ERuB0P>g z8(VAEMW6(IMOs73-j!60R4+olGWQ`5BFBwld1_dYS@!_TrmV`~x~w=XllY2k_foBm zLToL>aFZRR?E0T(^RkR4F6N)*wU%Sn%|HoYT@6;LM&1qP*t24VKCB?MqW^UHorxT` zx7lnqW6%nwqhzygEoCrPjH<%8$!ngWD;2sTHTPoV-unl*&ARzYv7>5;X|BptpfZgj zriTjC5S6KpmXE7Yf zA|@{$Bc2-;!{JW^2eHJo#@3?oK~#!aS9X)et0>K^y9i8W#iA=n*-J_%RjP_Tn_8|9 zixBZAXfG)wvu+VM_25B2n{9bWOm%GSMA`4L*({}!c#aIq!ipo}x%}KZT0Z6aA3PaV%x2hsS7$R*)#&0@EEZDKgx- zku7T1EprFOD+b${bxkyc5Ds+egjjab;FxX&F@1FNG`fq7>%f?@hZLSDotAU#Nn4x% z{cX_8s?ZhfngjBR7WQYmaEC8J14SO&!a8Zkj0k#SHml4pQEdu0)P?qq1M;sB;Hjp5 zAw`|0ni4`(rV22jFe>0EC8w5$il#$?XsSjyDxl7tw(OiX+<6xC2m6_ zTUWqJ8yaxq4OZ#WO_*-Yj=4lmv>QZSAVo*ZV#*DUo7y|+K{`gQ&S0;v=ph4bbOdr0 zh-kWnswxVgucxMz$BjO!S<##4GUZS-3EHbdx6elD3x!sqbWa(4}bZv$11>#1+FG%_Yo4K4&`cT#>7I z;DF22M+eO2BH_Uy@j$K1&6D2D-@S+P;!rG5&I8?%Z9~~4_wZ3Gv6GNgVpV36XgbS1 z6(00~YurCIwRe}I5IM9h8%9PIQ7k(76OdP~jIC-6)JrvK)HP)*n&8KNbWd!zI|REK zRaNdK4+vqQx-g+Egr)07Id?&dr4?TXvyXKlp;jnKs@F@W6g@B!+5&6fPa}`*L;J={M4IWZS4O&{J*$)9PvXc}0fVtcXI{&jytii+Gel5}M z<+N_>FOQkdc|mO*xF5(2p7Y%9!f4X&){?>q((bO8uOzZZo|6Kei$}H>GEnVe5nGRs zoQL7Ibp3t+Gk7uM>|QkWwM9YlZ~dXcv6pi4l4M|J!VaNaQF0c+MA{5p-}Ez4nrN<~ zdux76LhIW>smXsJH|lY2{8%nTA=QfQBu{HYMd@gxn|vBv^6)ldY!zGXt<%o)=ea}4 zBS?!$n(GtTN8Z840~4?#J%tr1mTTU~P{gw+tRmBn6sZ5Zlz2r=TxdzO-Q$B6H!e{2 zGb@xRmb>$!>0pd#I$$T7j~nbUoP)vKc!OrSLifav=h7-ft90z@6}o#+9_iTP5431Z zR|C`8AfE)!gmjK8Ny_fYbDTjVpL|Xki?e#R(Z{P@7IKRTKNj{CD>Rn^*+HL3;r1YQ z$H&|I!X=J7q4ov_h3T>TlJ=$pF7m!XEXdc}>m5pk%okRwR>2N$daKfD>1$1@J_HGlb-@VwPfB{04 zo_!bal5l-A>mNAJJ$*FCrAbHiu#2XrOW$Lu_>N-Lf&N0Np4|u>Xgd^D%IiH@Tu_nQ zxk11KJlhKV*n@o?^pNB16X}&S|1!QOfTBej4iP?<=8>nVAF(w#mmzsvT+?k zgs2o|?2sW8w`Zq13=`g-z=At=7iLdjqdG%wSm2~#wbJ?cO>zww% z($QBu$73M7vi@BnLrXz0#fNCiUL}RG0I_nfn$+11ZE0Op4DmlZVg zad|UiUE|t~%^-`cS`n)^#YNS~JP&^Xi+W__Pc zjOjS-SQ2JHl>Bo#l6>6LmYUEMSV~2)VsHPxAR3C~Co@1Q*JK7m|Llg9iyqj^H>MF% zGzNJ)QeGHWgT=CD!*g^n`V@)SS4GJYTGkLNZ(_G&yoI|{nQgaWVeckj2aCOkeYLDv z`$iQCX)l*zskQD~F2GrFx6V;lrqZZq^ZNv}DP1}tf7S@Ix7>!*u)Ev|Y!-zZ;LOf< z>)gYTX1gjVFONRydtRpI9O+kJvwtKM4M(m}nRYd@2x!t1$oF-Ry57YNS z+%f^x$mCev_*`>}m&wpa&hAM)Ixe-yXK6Gr3w^O*v7*iDfM`so&5i9vX>bJf*g=lC z#e~>I;|`Q!DXikGr4@I|bdiI?V&F|RC`UmzBVikxLGRnzo+gN{&K(ux)235{M zGQ`1QFO)1N2ZP0|JAOMC7NmAU&O8^f(zo zt0E_)(}Zk|A)-_orqorgv7_QEEKySxj_R5>D7&P&qm-hVM7uEfS5n)`HQAt6YkGoc zZ_QOzYhr+5QBwwI_))0Z2{?yo>@AT3v#w_?6qaTS7}eo!htAXS!2@*kcG7X0M;;MV z(qb@iMGh+8jX9LyRRgP=>cYX)tUFu7>U!<%wE)%76r%M;otxn4?*|Dnl2-~J=%$Uh;=iPWmUI^&Kp7tK7?QGMwa`tIBZs+hm zX}rrKFsAIWQaX1|HX7Lb3Ef5~D@5?1$H!&iSF;(??ts&kgLdCtx5go`Iyx&XBzDAd z<$4>73s_tmqreiPu-IByzPZl2B=&PL&8J>y%YOU|%Tx1*2V0UjGVtSWq`%BK4g@76 z4Ih9bOT}%a(yv>xe+^L1M$|$mu_HEJ99E?6h6%+-)PJK}+-089G za8)UK78Pv@d%`t_p@0c$Y76Bx>-Ha{#pEw`XUCGd3ng9Ivm_6VSnhn&husSa^-rWD zk>YG=zffo;!MwP97hI<9?b8TmvqupMk{QtV>f$e-Azi zNYTi!Waw-x5xIE}D-MkpzS_gS3hiNh?*Pe~)B$( z&Yec%P(dvbV^7GVE~9r98^f9~)9KQ87d}F`>T58fU)t@^Y1uO_Udu zrDWE8kxH1p^mBnYQ>#`m}79cDO_ zKp&SU7qIjJ{eAqE=z}uz(hjNd9)@7;SM9`#Q7h%Vx$M0GTZ93#*}&9e)^9oT9Q*3^ zKo*xaMfhpf&b4Xp@K)d2^Q`YcPa$*<`)gHiyApd+wAXgCzN`K1MktYQce8H?1`Ee# zv4(;9d?L#b{jBHr#|+LE{e@>K>{YR=aB@2PR`e5INMRSn9@ZyPkm{ZuA`DMq@#*2h zxpO-wrHAms>3%y+gC_FAz|-u*H~m-Zy}B>liMcJ;EJMraoZ88(n+Qsx&3qp$tXZHc zZtIl2kU`}P9KVwd(svh%@3W=)j^0lkcn+sCoEXfyp*vJ5>1;~c;L5(!PvRvOkP##d zYG%n9ll><>Bq>(hm?Y2AJt^JV;_15d1^X@|lV8uYLlTABnJi;SXJPAfRy-tDm_40+ zF=VT7;W!&Lw1?2;F?)4rx86@4LDBXngUh$c>{#RhvE1{3%GATbo=}6eI_xCGM$()G zqgej(9z_Rf4nJbeL!S#+h%{4p9=uGP>B+OA@M_Iy;D}+^Ge2j~X9fgDw-Ew9$KG__ z^yF!gyNEq2s9NJ<%d^VNc%j=pb|y1YIN8kp%1rG#_cK-9-Jq)II4+$w>-t-WM-<`^ zKUnp9cvK|!W(C9cS{XGwiyrRRE<1=;#K_tB_6~2DRo{jMB`HmGfmjQO=n#~I0RIXV^$PYO0Sbto)yC$<874PfO}HZWy|iJ-_Rv7pnttA_>!`6p z+ynN?sExv#_gTc~JYmD1?Df%+;tNMK`1x&QLg9 zh&$;Y^lnEPKPF_3vr2+%jN4yAgC2ESetbl!Ad?D3^p+Zpp%Q*p@hnQJ&`iGv0O|2D z?Io`oRXrmQBuOF$`3qvMcroicY|F&n`5s#}#v`$u$R=Y8XMpWH)K0N|soffRu98w6 zB$`hZeTPvZgDZw^2vn=~=FlVTmoX_q?gQ2#D@K^o#Imxw36mZ&DXV+X&TXo@0NPb@ z?tqm;TaYu{`JkJe-5;{cSs}uQRg53oRmi%R58t5;%E}!D z`hFa2IG6+eIMK4k=Y3pFGBw+>x@}~?gUn-GwBt5Rn862Ak<`R8#)S%12Uy{_D7yjg zl4+N5kG(f8C@_j9RYK42ze5Yu5CA0)fpsG3)HT1uu8&I-qysE!{6N9&Br}Yk@89i1 zQUq3DCw=G*!$=E?V(8u2%B&}Z3);6?)P!CgdtRY|IWCV@r1rS{oCW@*?Aph%JqC@t z`RO!PKA~jP9Fkb+M}rrv!yMr9=7CCc?4;vn-Ebl}k3c23K-fyaxf(!TiZ)``eID@# zX^0=A3M%|JC@e`Ex3Y;7eZp+1WJ4f$o@r4`H?>?Xo&0K3Dl!GL?uM3aoS2f)|1qvb zwWXwgsx@6Gfk&V%#BaRC)=cW=XL=jt3g=t^c{gSN{Vh>R z^WJ7xCj|)pEsW3h7koCeplqWs_A~Z&cADUGmC4y1j4#cC=(nH;MQC`%5G#2e2pAq` zQ_`shrEBAnu+01vE!CEiVSu~BUz=gXedBIL3Rbfgx?Ge_rby@I%XN_QS~6_i!Kkn3 ze1{m7ON6QAJ#My;fpx*GYqtqh+@t=ufx zN`YwW{J@v8$0dl2jg#_?sc`dQx#Hrg^=wbqm_c$m75nc*!#{~CU%Safv)rzEZ)EGH zBnziA*~KZ|Zr13lD7_k%EC9mV4BNwOa-w}FqAeJ9gXdlhA}0@3o91sy`s8PpkrU#5 z^-r$Vi7QG!V#e!0z;fHbR^$v2(l)Y_IUQYdVGxTN&JlxmvfR0^9nUnmUHrm{YC4z& z@8%;l6)K?ch~)ebTkO!>KAnHSHoPwj)l%Q5P`j`qr%SrGKo=`kPcG6t!oziJL0+iUejA=` z%V-C%R45|wE0XW_TtY^3Lx(XTVnE%vRyCxhE%({=(v8<{6Iv;_35&JuA znH19r=z>#n_FmEn^~hWlQD!!T(XlUg9lD*g{t-Jowa6zFITnPFdZpS@E8dbZtf>k9 zj*r-gX=6Kwn|HM>_795}P+gycHjaaTV>%!Ntv{_?>6Ia>hS)Sc<)=+a! z%JaX#1mTV!hhrDon9oC@zIBz-2Di{o14;T2WY({>tFy}0LHNR7@ovW2q>`6Pq-0d)Zx}T*JgcEuf+uZ1!UOxUW zuF*ywi6r=Aox@PDS=V7TyPY2`WHvCL*37*xPeD1$6;MamT)l3gsZg%Dozw%AP=gV(pgXi^+?)W*4wsLwiw3EHnm~wJ+#9|4? zc@GOku^>M!WP9gz6W+{Wx95d*PJ|A@bqw9gh_h<5L@TF*jkf5=`JXthImIr>;2C>X zItBl@Ag9e^Nre%DZ_dtXg-N`#c)V@ME*xWD%s8FD7Y{tq$Ol19mFiPWzXWE?x|pT6^82~b z7B>4U^~_+{?C@tT_7rn@ynKfFVWB_6D%6_R!dk>es+U@Dp8jQp*_?jBtTEjP94x*1 zIEalc^0Ti*He@QOWlM`@4l8Y5b+R!?gvii z(>~+ugrfCQ{a~r|anKy{6ONBbRwAtTf>=I?&0Y|MzppLu&{J}X)lZ`NU|L0XOX`Z$ znBFNRfAK|KTJyf5kLi=rr^euq^Og4shgPzh;W4sOo<3>^nilAUz;b5$LYNc(GDOPi z8rWNcSg#jWqq<`+1STY`P$t>ItlPU>!KU_Ja>+0mTVSWm`aTpRI1y!@CCk^ zJz5xw7Wgjm^P~=gaYbsk-ljjTidk$584V(wD?V~8Vh~2Bzq778$bNq@CFy6F<*^sxzWEY}TIx;M z>|E@{vRds5m=0V*(ls1YwWd|T_0n&8=~4QDXSkC~`sFU0RUGN_0x}|hrjno3O5+%; zw9Bw3*yqK8#(=hR4OKHWT1hE>iUZ5?8jOW63F*HCiLDnS_P|kbBHg09i3#Tl zot1YG-^+H1X!>!zbPxw9dcQ$eK1alaC-a@;qywycNg!)l($%^5SWH7d+)$KVH2%h} zcJ*ZYBQ$KS>jtVnQv1dO{WDSoli+;W~*y)a*fTe3X zYL}w@?x~p5EW*WR)^&T4*<4Dtdl_kF-Oa^p$R#hkNlK)4F`IMA&-x;YWv^X&A!%g` zx+|q^(ZfxeOukaw9UR4r54uW_I}XPnG#{}uc;T;gk(`JTOMisA|VHs2m5TLH$R53WODpTY1_%-)57ADuf55hE)NuHUts<(M+?## zmj81IoBi@veygHrGY4A~U0oEAXDm78l@s!pLs4SC6&H7j?zlHeh%(~Z# zAa+7*6>7If8_meqcmnbu;uaCo@C$oZKBRjIsHPxJ&T1?5=Iaofvf7376OyEB95;WK zC|%J@che92g{9eJXrdL>S%31c0kLMATl%|1=`2=O;V6ir72tbrn(DIpRLp?->Dz`fH2*$f$Gzb2;( zQ@&y`Dbeh@y)6~eIKoxNAS#Pb)} zH><*huKQWiKri8q*^1}&J`|c+H*&V(c|D>;y3STSuZh8I`D%l(?Gbyr+E=*zh zVfrK1heZp00@!rsCtPo4%UN%s$0K%tg$Z9bv)@>_kXXL6-I|^}AGR~?jVZjat`uI^ z0O#}%+WPeU`|SLi8P0d#Z=)2y&$?8G3NJ2aBPwS(ZXTg*E*<6c_gQV_c<-`=oj*_9j zb`NJm*Yy+*nb^zg!Ue+~_Sw3=USW#wv;|IKAM(n8{~KI!2kc?i>%+Z%JAvg`ZcEgK zibA|VE5>y#OIhy|a%8tU55K+yW28L%D@66>9ji5u&#B%9y!)9v*O#qb@8U839h^g< z9kRd0_QX{Fk>9fK)&~hM<*|nKUGe(BWrME$h5#xOFBWlP>PY>P4q3FRjgcPj28lV3 z$zU6VB(mkk;v8Xh2A8M;q|IdJi|mt2_v`^)V$iF!`8%J6BL$E9Nc{ppy8!2zX zAwvMlSxGppyL}kwm_o?Gp zMV$AbK4{NM+>$+PPK~$Y4UQ2FI)K(U5-jVPXcyYb%E0-gIDx2KKAm4eL zHO0BJ%nd%m#1fXjVHn@YzT4m{JetfdZU`0ne6wTYG2S%^dY?R(-(&4#%J=VJ7dQDD z2f3qn72~b=SFB2_ihipY$nPUNRLrT7jn%|RqLPO(CZVW8zIi)fvLz@-N_g6kyCfm0 zqJTGa5p(Xe-iFGllvNd^`F{)2;t*xn2h$l|%sDDr)t7`Oa7DXOsvktgfnaODYkZV7mx2YsUo!lVkiGmtZ==zn_b#$6ebU46W+Qgln!Biwn)ADW22#N zUW9HDE0V@T%3I%$;@RI(%@8ayD1xV?r6vC!;1vLMpS|dBhit31UZXnmwu>- z^fcne6FDS6z{ief%eF?_`z`^oBFTLTyR)?eyR_Bb{z0)qxm^rOSK-A?Ebtxw0lAy7 zBP^h$jBCy6aa0f-%x2v#pg5+a$6hq+T>gQtAr^4QwC?G89dYW_b{>R-@_rN|yQF zEg^C{OWZaxbUHMCxfShS!LmzbEB}o*X$g3tT>0XgZ0EKfLUb0pwQYpQ4`Wb!d0KNVwYjQ2zQ2cJBRfr|bk&r&Mz5AhZ5p@+8|UXkuCowbxtj^9CNp$HSK9 z3>lK6Jn$Z@Eq$Q0LM^qfC)@W?gK%sZ+xT&`PLIt(8N425-9i{S7_ob+*yWFd ztsUaA!3&@C5Ykt&-k*fE>x#X$b%c#{N)ST@su{sP6e4#2osN0IGZar-9DrUTbsTKi2#UO}>^-MjPuwLZ*{aX{*|^Ufqed%iW`mL^ z{hxUf|20o%$`cV~W#Nb8>a^U9U5o0#bAZ$GP1?esz%8fxpH8srpLG#-^kMFwCktW~ z8~ynx?cNj&rE=i_PpI`ww(s*$7vo#FRdRqt;8B-Z=RJh|{&|45d?if6D$!KSi@lx0 zwlkmYGlV0vSlRZ4LN7ftZ%-03UuAv2NEbdG%*wwA7uL^Y+rJQnOak$&`;L0yl?vvx zbD^+3oh{#~_gnlm8W+_d!s(n$4!}>UZBIY};of`uOm=5ytk6E4h3>*$J%|n3^_&nr zh^^k0C@kH_>UQ}F+h?#lyW)iP=`8HazHzT$uYxC`VsO*4-j^>c8gJ-Snuei&6Z(Q6 z9mN8%6Jt-x5xMNO-O=pemu`Y#D*OJ+wL*tiSix8E!mH1*kG>k{H*SxmadeS|E$)kI zD@)^6O=HfxBL&Sq)_b>apN2GAk_H_wILh3CAivpX)KO6L1F6xI#m;smb zNjYW;>hCKQ=k7eU`x`5Kvqlp&*hTK9B3^QYiUi4lDiR@ks7Q?LtRRlEMkOT6PbVq4 zMfrh>WXiWxBul=kBH41iicFVJt4O?jR7DEpZzma*{0rq>3P~y7T?8&wD=8AE zCafhqLdB;O?x*4(5bmPl+X=T;@jZmMOi;?HBK)?B*AU*I;&p_dSMe=`*Qj_sZ$#u9 zHPS!{+f`g9{9P4qB7Ciiw-ElSiqn^Xa*2xD5pGa%7s98jxEJB$RXm9BOcjqHJQcVR zb4K5x$?bN6d?Dc{ zRJ;WE7`aM~te}KlD!!8N4^(_D;Tu(a3*jqO`~$*EReU?)3srm%;rZj)^ZWgWrD56N zI{7&zWw4!?;GCdR;#7(YDBV;_h)T%GlyEHgo(vFV-744jE*pBL}^7G6$N%{^O;6Y?%NeH zddj6dA%m0sj2ZuQQGGxz*W{B)sHBpk%sG#u<<3)8Fii#10niRQ&W$39IlQ?yJzubV z)o2Lw8j~_IWmF1_KHSe34C~k8-ui4Id>F~F2QF_wF3I_OfSa$PaUlJqlFk7%;&Z3) zgKGS1is#Fvj5?xt$?f1bxmk#{CrT@~P{*WW`W#KKk#NC|wZ@wad_tjo-_c+#S{)P< z*sa4Iyi(P|k^uH2Cx=U}Y}}K_vxp;p?bfWrF#L0+=|Th#O# z96Zgsaxwzt`4 z%7kc38BIooZV(cbb_{E^To)^|1e3kGhr;2daGXLL`=U^M(MpU)I(@&dGBjjT5ya3z zVXz|xN`F~R-_n-ePDz)EOzG2r;5GyqHPl=TO$b^vR%JKqs#^K_5mK(Huf$}!z;kfA zlIzYj;=?A*QIPOVv>HH!mvxFh;feuAbd>F`WOn|PQluuc&bkM4s7-fHQ8-T<;>^0w z;#hXAd&dY1r9HkZj#FDfS6${hy5NpXV*xNl_HJQ8upcVai?5YicOye+yQ z=eH(=stN9_+lih&;h`iYjz@gsp)9KuG1vHlv-eo{L6TELy5ywQ zt~rDUkLU#zwgkCg4a3vatSzM-)vSv~QgCrtR=RW<>u7u$b2$|v?AyrVPJI{nCls!7 z*`zHUqO1t`wCO|RAZpR&4b1hlw=ktEi$2{;DBr;HP6rD%UD=A$6NNWJnT*){Q07<{ zA?#hxdent`-;YMe>TljgwK%!!Z#Du!IWMkf^Xrxi57x2EXF`Ozomt?SZv7|s8e{TW$$f$1 zL!Uj5X^%M5i7h$jH|#;AIu1LZ!k8}(6Y){TyBKRfu-;q{^pSLOuisuYlpv9#<7i;L> zT-t^=6_@rJ7o1zyo)75x{q(lBHv&abs!LmN#ek7Z`+`c9KmD7r&vdx7y_od{A7RdP z7IeW=x9JU3v;@9$d^&A$(QSQ$=`Z+9IjC*((0Rc_7li|60!WIg*w={$rU*{Ns*W^E zq-G%~I@uL<;WF#iwAQv+-0WyH+SFb?)0QAloThm0K3LMt{dMe2h~d5dP2Gr+)8ZJ}Y!y2D*JZ`Y*RnabW!4+&;O6 zv&-NA%k9&~yteu}q0l&;wE&fYL|Ewiz09NW?m=ImmH z!66}Vl`muR8v>)AV$ndy#1j!)&1_#~9n&4rHCT>~c)Mb`BYLm`cSIhn+tps(bCIGf zDnN@ndbPT!79=Fih5b2y&cECf8CGCXd9dwQt*zHUzpy=51LH5i1hpuITR>5qs7f(M zO@tWx$!Ar%{L$a6y`E>c)7EjgJWL@B(@Vb@B&xZ9iMwSbNByj%{nkL7=q0Ygj{9FWk3!Xm>70 zQ8UDxA5ql)n5oXm3MLBlr7;@A{De5mY+sTS&x(oWu#w9$K^2gOvSNIToGRT>0@0iAGgf+;+NFJ-S%WtBO|QbP}-_UU*Qu^_E}_9phS#S_!8{s&9AV*U@1*5Lp8BjxSEek zIo1-X_p&33=|$8V41=(>+R>hDe3VbZ_aaG;Jv88;FzUYXOK-ei>eIs#TZfvy-hcZe+jZq|z^Y;M`DF zwG_^K*mS35na2I1AqbuIL$4!JB|H;2DqW!8zo2iW%sN5G4qu<-e04ZA9;3*~bB42+ zU;XHtj$Z>^*3Kf!;I9vW?>Qinwf)9c7d zxo~KK8@9(vCE8pRwAYT>$g50?ElitLrdc!Dh8x|5gpb(i8!>`rCgXqWB)GwD`7PCV z%?!FyQ8K?lnNjqMjfE)r#TiWcZI}?Zi2eMVf3iK;QhVozEi@F-$E~YBG@HK@P$S;( z*+}2Zw_Z?tj@+Xty%6ac^m`9@nB{IGI}Tt8a>&rPu-ih~+d=ur8-}(@1I+I_x&!n5)v0^)PaXi`qD;%{tyrKEhs#e- zcIReCW9Vh*6kJw1iEG6uSi2XO`ZivaA7_9XuPJhrPXnvw=n=e#t35nsK(!1p{0sw# zdg%!2qaS5>AF;+!izeVQK$I?kY4tsf7)sWbVlm+NOyo}=(h%a*k`LBL^`fs?d~~I^ zXg)1si5G>-Q`>5iUj(XH?QUSjk5GU?TiOoApmaQjMqxefv6Hs~ghh+l@3&6+4AQp= z@ehN|<~=0NjSiCL`MK@%%X3nL{I8W4Q+iG zI82RY(O&xT_aC`=B|J@dceAn+M{z zh_Ih=^14}cyJ)%{L6`5|`1ZN2_s7$zFL;PVZW6t804J)NDYz9lO9VJW9@ukSDBL4c zl@Dr7x8uPekEgkP{H(}H-b>RA=f`;Fb*H0H;mA_%M7hPhgWok$mx72#i%qhbExqFr zv^`N77>$T~UN7~`@V6w(&xvf;ofzZFaOHjYK9rlB3Oxk9^M(WJDA(K>N3|XnEv_EP z`LaU5rzK9*7`*q@eT_w{bi^DUfp(k2BbC2#_=_&I{Y%D|n7Dh#Zx2Zs$#Ot6xh!Cg z;yc7-EhWXMN!hY>H2dOSdu)gq@V~R6cRd4R(b_-iUhlhqSKBrMK-kcG>D~|Ls}hNO zGu+cxsrP0M@_{?(h@aOi9O(Z`nR0n9dv?!LFx+8p-SrlN=PSTwKHGOUK)6x3)APO` zZ(Q3ATQ~k~v%ozTc+>(LEbzVs+Tbf!`sZhX@fJAJ0;gJFu?4QSz|So3um#qu(8$Ra zLW>2u#Htm9Sl|iz+!_D&S>Q(&xY7cPEqp61aE%3iVu6P(@OulCE%2!V;e7XsQ!D6c zfrBh?q6IFrz*jABy#*e$z_S*3%>r92(58o4uDb;WSzx3E_9TeErj6~m~Vm0EpVd+?y$fU z7I?`5?^~dgWlDT3Fwz3!Es%cahyFcpfpdECcz%a}OD%+Y3si{sGtx3weJwDXebL;} zZ)J$`D~nwHXcexkRbk6!F1a#g3h(Yz^+=@@gsK!T`jN!&)*p)`eV2i-C4}d;xT#HQ z7{MnKRC3-P7Ci#a1qo_3t~b7_*l4I{WP3kf3GUFO#GasrPPIy9M=;roYt1`{!W~~!IAUNb>RF* z+H?MKcAS5A?*`90U6rfWn$yPFa@y{W1+Dq>FyV6%hN$)MTxEebAK})Ttn1EkVF*K7 zDnIh(J1EVg>FN}V-_JytZfV{r;J^GM&O2Ch4skZr!Gix(ZPy-ft07pb#>?UT4SY3<$ZloDR~a>+V{` zyW4R__gp*9eWWet4x)Q^mjX*iGF|^`hPu{nO`F4U?;&h#E!@|E^EJ3|zPVb?H?AG$ zyV||Zt;!{n|F_Q3NGp=%2Iy(WdFI+z?)T@@B3cVBS;}!c5R88kE0~l5unf^ST8fp~ z0la&p(zG~7>O+IJ&bi95z#+S}gij@oyMmyWQ2Ay6-?5)EbQm_Qb8oD3ZwxhtugcZc zhQp5$a<1L~wTRHo95)bQU+biO6TpXiDup7BAz)qQjLLsKIB#;kN zB-aA^^Cy4iH(?(@A9M`^<;|Bj(IW>q6h-+;{Gicu& zS}0* z*}#m%VMYpd4gXvXM{eRrkP&uT`164~H*(y$KPz7e=DSwb2J-_4slw$5;d0=#g^mT@ z(E{hgmn%9j20BfR+w(L4F?NVY*>XOHINm2NmiOUf_*Q*cU|-oKgnzDbR|p@}lbBb+ zN>;?YKO~_G&5rsnMll6=ZlT2LoZ7>fn_e{9iKo(bn?RJwKI-MiUo#!Q}Jcy4ath#7OH<>w8ZGBYp7z|H0d z=FQ0~oNUM&Fn7-Ux%qhm42Ah=3-aa|xXKwF_=-xej(l9F1nUubhJP@Kljr2-=M~~t z%u_A)Yo1D~T-TBRGVE0p3ijEv!KR-vCohFNWj!*laQ2Kjs3)^9P}s5vjpn3J2furj<8->LHF2tH`{oKC#J2mVRrJO`tA%uQ}8x+y3+V21Nu=1y5 zKESvzN3H1p78))8t`AqsE4GyTzkv>18$;!?6#iu+2W5tBp1Xn2)EPge065f@h|15XCjdvjbCaDp{{P>jF{ zo<$)3CO;hbv^Y4pgd#KG?+EU|n*av{pn#4%mkam@LMG@ffIR|n9S=Mna1jFW6a)GP zsdx}z&tM!0@FTMEfOQB|8NuZtYR7P7<8T;5UMx^TDiMfjE#MO+f#;qAb`Dbu-2?be z2aYR7#qR>%MR=DQ1nAR|*2DXt9Kr!LNoX8!JHSbO(Fx$w0eAM}xKDuZ0{pcf>ZgKa zM8+gT4lrj~fC~_4xT*jf5Qv8$m!jeXpF^NVX9JcZ)PsK|;0Fi|z_$Z_fpDG50X%>p z13wJ-Mt|56z}NOSavc3i?^8tT01x9JZruS)fIdUi%KQMoLZDIE1DHBg9bSTQ!!S+Y zCwK=z1l|N#G8~nmGX(z_$#MChw*Y=Wn&VaghpzxVM)^kcEGCQNJ{+sg%_o3X&#OJN z23&-&2N_lXeu_XNu^liE2c3G*rvu(bpvDn&8L#FgIA8+)PX~-pD&P?W8m$_@O%vf` z0lo#$_#^&M!X?1_7TjqPbN~X;GXYHqBY~F!ev3erRRPY;R`CKr^JM&Q6yS-U!f_K3 zDDSI)H*;Y;0ly8{HBaq8JRtpm19*&FK9IiCIc^u2k^#R%r~_UFIADf~rvlbmaQaP= zkvK)k;34=7LHX}3W?{4tNWzN&YY?ar1b5-o6oC~{1(-S;f(e}9@!3XoF0UhUYz|t7 zgcE=+bJd;^Y(}6Z*8-S44-G;E1ScYpkY)oOLZHgXRJ@8n({vp$rI6z;A#VxbDTF5A zb$}BLaD)IaFamkL2o<8x@qnK#Pc!JhAnDA1YNcwU}+J&@{Xz{8vwt57g`ncdcYI!srsLw^)|I) zf{7NK;9UfLB=mm?4tEC-vXJ>OVE*SEw-7kNzS|+az>@(B5oo@*0Gbh~q1+e1cR;|v z!|lXSBhc!|1U#|}BLR9HpzcdmA9w*yvfu@PUdFGG0Zd+i*Ydjt7(wDD$g; zy?;}ME)nnsLKApo!11!$P=fqT^bS@~G+;jy7 z52J1bQvr}h1Tv?%`zZVY0&s#;A40f*PY0|-P(}!_V>8bDpvM6k5Qt9jQwzQw(EbrB z1CJBnG6YgYrGWbp>OqHFnloO*AIeOAS@O271QS8>sD1!RCOE`GCrB<&;vq<`N#Y?$ zPA#GnB(EOf1j&O(I6-o{5l+zgv4R^FcOCiCD1jh(y9g(^$byr{iyTu#CrEA_!U>+U z;N+hnKMc_c4o0BH5hTwD(Fu}2gK&at|BO&0w@-cfMnIL~74fA4RN zDaR&F?|nFAFKGu@0O$p{qO$W4-gEL6L= 3: # needs to be a text stream @@ -382,7 +399,7 @@ def read_stream(cp, stream): s = '%s = %s' % (name, value) entry = get_export_entry(s) assert entry is not None - #entry.dist = self + # entry.dist = self entries[name] = entry return result @@ -414,6 +431,7 @@ def tempdir(): finally: shutil.rmtree(td) + @contextlib.contextmanager def chdir(d): cwd = os.getcwd() @@ -435,19 +453,21 @@ def socket_timeout(seconds=15): class cached_property(object): + def __init__(self, func): self.func = func - #for attr in ('__name__', '__module__', '__doc__'): - # setattr(self, attr, getattr(func, attr, None)) + # for attr in ('__name__', '__module__', '__doc__'): + # setattr(self, attr, getattr(func, attr, None)) def __get__(self, obj, cls=None): if obj is None: return self value = self.func(obj) object.__setattr__(obj, self.func.__name__, value) - #obj.__dict__[self.func.__name__] = value = self.func(obj) + # obj.__dict__[self.func.__name__] = value = self.func(obj) return value + def convert_path(pathname): """Return 'pathname' as a name that will work on the native filesystem. @@ -476,6 +496,7 @@ def convert_path(pathname): class FileOperator(object): + def __init__(self, dry_run=False): self.dry_run = dry_run self.ensured = set() @@ -580,7 +601,12 @@ def ensure_dir(self, path): if self.record: self.dirs_created.add(path) - def byte_compile(self, path, optimize=False, force=False, prefix=None, hashed_invalidation=False): + def byte_compile(self, + path, + optimize=False, + force=False, + prefix=None, + hashed_invalidation=False): dpath = cache_from_source(path, not optimize) logger.info('Byte-compiling %s to %s', path, dpath) if not self.dry_run: @@ -591,9 +617,12 @@ def byte_compile(self, path, optimize=False, force=False, prefix=None, hashed_in assert path.startswith(prefix) diagpath = path[len(prefix):] compile_kwargs = {} - if hashed_invalidation and hasattr(py_compile, 'PycInvalidationMode'): - compile_kwargs['invalidation_mode'] = py_compile.PycInvalidationMode.CHECKED_HASH - py_compile.compile(path, dpath, diagpath, True, **compile_kwargs) # raise error + if hashed_invalidation and hasattr(py_compile, + 'PycInvalidationMode'): + compile_kwargs[ + 'invalidation_mode'] = py_compile.PycInvalidationMode.CHECKED_HASH + py_compile.compile(path, dpath, diagpath, True, + **compile_kwargs) # raise error self.record_as_written(dpath) return dpath @@ -655,9 +684,10 @@ def rollback(self): assert flist == ['__pycache__'] sd = os.path.join(d, flist[0]) os.rmdir(sd) - os.rmdir(d) # should fail if non-empty + os.rmdir(d) # should fail if non-empty self._init_record() + def resolve(module_name, dotted_path): if module_name in sys.modules: mod = sys.modules[module_name] @@ -674,6 +704,7 @@ def resolve(module_name, dotted_path): class ExportEntry(object): + def __init__(self, name, prefix, suffix, flags): self.name = name self.prefix = prefix @@ -692,20 +723,21 @@ def __eq__(self, other): if not isinstance(other, ExportEntry): result = False else: - result = (self.name == other.name and - self.prefix == other.prefix and - self.suffix == other.suffix and - self.flags == other.flags) + result = (self.name == other.name and self.prefix == other.prefix + and self.suffix == other.suffix + and self.flags == other.flags) return result __hash__ = object.__hash__ -ENTRY_RE = re.compile(r'''(?P(\w|[-.+])+) +ENTRY_RE = re.compile( + r'''(?P([^\[]\S*)) \s*=\s*(?P(\w+)([:\.]\w+)*) \s*(\[\s*(?P[\w-]+(=\w+)?(,\s*\w+(=\w+)?)*)\s*\])? ''', re.VERBOSE) + def get_export_entry(specification): m = ENTRY_RE.search(specification) if not m: @@ -821,6 +853,7 @@ def get_process_umask(): os.umask(result) return result + def is_string_sequence(seq): result = True i = None @@ -831,8 +864,10 @@ def is_string_sequence(seq): assert i is not None return result -PROJECT_NAME_AND_VERSION = re.compile('([a-z0-9_]+([.-][a-z_][a-z0-9_]*)*)-' - '([a-z0-9_.+-]+)', re.I) + +PROJECT_NAME_AND_VERSION = re.compile( + '([a-z0-9_]+([.-][a-z_][a-z0-9_]*)*)-' + '([a-z0-9_.+-]+)', re.I) PYTHON_VERSION = re.compile(r'-py(\d\.?\d?)') @@ -860,10 +895,12 @@ def split_filename(filename, project_name=None): result = m.group(1), m.group(3), pyver return result + # Allow spaces in name because of legacy dists like "Twisted Core" NAME_VERSION_RE = re.compile(r'(?P[\w .-]+)\s*' r'\(\s*(?P[^\s)]+)\)$') + def parse_name_and_version(p): """ A utility method used to get name and version from a string. @@ -879,6 +916,7 @@ def parse_name_and_version(p): d = m.groupdict() return d['name'].strip().lower(), d['ver'] + def get_extras(requested, available): result = set() requested = set(requested or []) @@ -900,10 +938,13 @@ def get_extras(requested, available): logger.warning('undeclared extra: %s' % r) result.add(r) return result + + # # Extended metadata functionality # + def _get_external_data(url): result = {} try: @@ -917,21 +958,24 @@ def _get_external_data(url): logger.debug('Unexpected response for JSON request: %s', ct) else: reader = codecs.getreader('utf-8')(resp) - #data = reader.read().decode('utf-8') - #result = json.loads(data) + # data = reader.read().decode('utf-8') + # result = json.loads(data) result = json.load(reader) except Exception as e: logger.exception('Failed to get external data for %s: %s', url, e) return result + _external_data_base_url = 'https://www.red-dove.com/pypi/projects/' + def get_project_data(name): url = '%s/%s/project.json' % (name[0].upper(), name) url = urljoin(_external_data_base_url, url) result = _get_external_data(url) return result + def get_package_data(name, version): url = '%s/%s/package-%s.json' % (name[0].upper(), name, version) url = urljoin(_external_data_base_url, url) @@ -986,6 +1030,7 @@ class EventMixin(object): """ A very simple publish/subscribe system. """ + def __init__(self): self._subscribers = {} @@ -1047,18 +1092,20 @@ def publish(self, event, *args, **kwargs): logger.exception('Exception during event publication') value = None result.append(value) - logger.debug('publish %s: args = %s, kwargs = %s, result = %s', - event, args, kwargs, result) + logger.debug('publish %s: args = %s, kwargs = %s, result = %s', event, + args, kwargs, result) return result + # # Simple sequencing # class Sequencer(object): + def __init__(self): self._preds = {} self._succs = {} - self._nodes = set() # nodes with no preds/succs + self._nodes = set() # nodes with no preds/succs def add_node(self, node): self._nodes.add(node) @@ -1098,8 +1145,8 @@ def remove(self, pred, succ): raise ValueError('%r not a successor of %r' % (succ, pred)) def is_step(self, step): - return (step in self._preds or step in self._succs or - step in self._nodes) + return (step in self._preds or step in self._succs + or step in self._nodes) def get_steps(self, final): if not self.is_step(final): @@ -1128,7 +1175,7 @@ def get_steps(self, final): @property def strong_connections(self): - #http://en.wikipedia.org/wiki/Tarjan%27s_strongly_connected_components_algorithm + # http://en.wikipedia.org/wiki/Tarjan%27s_strongly_connected_components_algorithm index_counter = [0] stack = [] lowlinks = {} @@ -1153,11 +1200,11 @@ def strongconnect(node): if successor not in lowlinks: # Successor has not yet been visited strongconnect(successor) - lowlinks[node] = min(lowlinks[node],lowlinks[successor]) + lowlinks[node] = min(lowlinks[node], lowlinks[successor]) elif successor in stack: # the successor is in the stack and hence in the current # strongly connected component (SCC) - lowlinks[node] = min(lowlinks[node],index[successor]) + lowlinks[node] = min(lowlinks[node], index[successor]) # If `node` is a root node, pop the stack and generate an SCC if lowlinks[node] == index[node]: @@ -1166,7 +1213,8 @@ def strongconnect(node): while True: successor = stack.pop() connected_component.append(successor) - if successor == node: break + if successor == node: + break component = tuple(connected_component) # storing the result result.append(component) @@ -1189,12 +1237,14 @@ def dot(self): result.append('}') return '\n'.join(result) + # # Unarchiving functionality for zip, tar, tgz, tbz, whl # -ARCHIVE_EXTENSIONS = ('.tar.gz', '.tar.bz2', '.tar', '.zip', - '.tgz', '.tbz', '.whl') +ARCHIVE_EXTENSIONS = ('.tar.gz', '.tar.bz2', '.tar', '.zip', '.tgz', '.tbz', + '.whl') + def unarchive(archive_filename, dest_dir, format=None, check=True): @@ -1243,6 +1293,20 @@ def check_path(path): for tarinfo in archive.getmembers(): if not isinstance(tarinfo.name, text_type): tarinfo.name = tarinfo.name.decode('utf-8') + + # Limit extraction of dangerous items, if this Python + # allows it easily. If not, just trust the input. + # See: https://docs.python.org/3/library/tarfile.html#extraction-filters + def extraction_filter(member, path): + """Run tarfile.tar_filter, but raise the expected ValueError""" + # This is only called if the current Python has tarfile filters + try: + return tarfile.tar_filter(member, path) + except tarfile.FilterError as exc: + raise ValueError(str(exc)) + + archive.extraction_filter = extraction_filter + archive.extractall(dest_dir) finally: @@ -1263,11 +1327,12 @@ def zip_dir(directory): zf.write(full, dest) return result + # # Simple progress bar # -UNITS = ('', 'K', 'M', 'G','T','P') +UNITS = ('', 'K', 'M', 'G', 'T', 'P') class Progress(object): @@ -1322,8 +1387,8 @@ def percentage(self): def format_duration(self, duration): if (duration <= 0) and self.max is None or self.cur == self.min: result = '??:??:??' - #elif duration < 1: - # result = '--:--:--' + # elif duration < 1: + # result = '--:--:--' else: result = time.strftime('%H:%M:%S', time.gmtime(duration)) return result @@ -1333,7 +1398,7 @@ def ETA(self): if self.done: prefix = 'Done' t = self.elapsed - #import pdb; pdb.set_trace() + # import pdb; pdb.set_trace() else: prefix = 'ETA ' if self.max is None: @@ -1341,7 +1406,7 @@ def ETA(self): elif self.elapsed == 0 or (self.cur == self.min): t = 0 else: - #import pdb; pdb.set_trace() + # import pdb; pdb.set_trace() t = float(self.max - self.min) t /= self.cur - self.min t = (t - 1) * self.elapsed @@ -1359,6 +1424,7 @@ def speed(self): result /= 1000.0 return '%d %sB/s' % (result, unit) + # # Glob functionality # @@ -1406,49 +1472,40 @@ def _iglob(path_glob): for fn in _iglob(os.path.join(path, radical)): yield fn + if ssl: from .compat import (HTTPSHandler as BaseHTTPSHandler, match_hostname, CertificateError) - -# -# HTTPSConnection which verifies certificates/matches domains -# + # + # HTTPSConnection which verifies certificates/matches domains + # class HTTPSConnection(httplib.HTTPSConnection): - ca_certs = None # set this to the path to the certs file (.pem) - check_domain = True # only used if ca_certs is not None + ca_certs = None # set this to the path to the certs file (.pem) + check_domain = True # only used if ca_certs is not None # noinspection PyPropertyAccess def connect(self): - sock = socket.create_connection((self.host, self.port), self.timeout) + sock = socket.create_connection((self.host, self.port), + self.timeout) if getattr(self, '_tunnel_host', False): self.sock = sock self._tunnel() - if not hasattr(ssl, 'SSLContext'): - # For 2.x - if self.ca_certs: - cert_reqs = ssl.CERT_REQUIRED - else: - cert_reqs = ssl.CERT_NONE - self.sock = ssl.wrap_socket(sock, self.key_file, self.cert_file, - cert_reqs=cert_reqs, - ssl_version=ssl.PROTOCOL_SSLv23, - ca_certs=self.ca_certs) - else: # pragma: no cover - context = ssl.SSLContext(ssl.PROTOCOL_SSLv23) - if hasattr(ssl, 'OP_NO_SSLv2'): - context.options |= ssl.OP_NO_SSLv2 - if self.cert_file: - context.load_cert_chain(self.cert_file, self.key_file) - kwargs = {} - if self.ca_certs: - context.verify_mode = ssl.CERT_REQUIRED - context.load_verify_locations(cafile=self.ca_certs) - if getattr(ssl, 'HAS_SNI', False): - kwargs['server_hostname'] = self.host - self.sock = context.wrap_socket(sock, **kwargs) + context = ssl.SSLContext(ssl.PROTOCOL_SSLv23) + if hasattr(ssl, 'OP_NO_SSLv2'): + context.options |= ssl.OP_NO_SSLv2 + if getattr(self, 'cert_file', None): + context.load_cert_chain(self.cert_file, self.key_file) + kwargs = {} + if self.ca_certs: + context.verify_mode = ssl.CERT_REQUIRED + context.load_verify_locations(cafile=self.ca_certs) + if getattr(ssl, 'HAS_SNI', False): + kwargs['server_hostname'] = self.host + + self.sock = context.wrap_socket(sock, **kwargs) if self.ca_certs and self.check_domain: try: match_hostname(self.sock.getpeercert(), self.host) @@ -1459,6 +1516,7 @@ def connect(self): raise class HTTPSHandler(BaseHTTPSHandler): + def __init__(self, ca_certs, check_domain=True): BaseHTTPSHandler.__init__(self) self.ca_certs = ca_certs @@ -1485,8 +1543,9 @@ def https_open(self, req): return self.do_open(self._conn_maker, req) except URLError as e: if 'certificate verify failed' in str(e.reason): - raise CertificateError('Unable to verify server certificate ' - 'for %s' % req.host) + raise CertificateError( + 'Unable to verify server certificate ' + 'for %s' % req.host) else: raise @@ -1500,50 +1559,34 @@ def https_open(self, req): # handler for HTTP itself. # class HTTPSOnlyHandler(HTTPSHandler, HTTPHandler): + def http_open(self, req): - raise URLError('Unexpected HTTP request on what should be a secure ' - 'connection: %s' % req) + raise URLError( + 'Unexpected HTTP request on what should be a secure ' + 'connection: %s' % req) + # # XML-RPC with timeouts # - -_ver_info = sys.version_info[:2] - -if _ver_info == (2, 6): - class HTTP(httplib.HTTP): - def __init__(self, host='', port=None, **kwargs): - if port == 0: # 0 means use port 0, not the default port - port = None - self._setup(self._connection_class(host, port, **kwargs)) - - - if ssl: - class HTTPS(httplib.HTTPS): - def __init__(self, host='', port=None, **kwargs): - if port == 0: # 0 means use port 0, not the default port - port = None - self._setup(self._connection_class(host, port, **kwargs)) - - class Transport(xmlrpclib.Transport): + def __init__(self, timeout, use_datetime=0): self.timeout = timeout xmlrpclib.Transport.__init__(self, use_datetime) def make_connection(self, host): h, eh, x509 = self.get_host_info(host) - if _ver_info == (2, 6): - result = HTTP(h, timeout=self.timeout) - else: - if not self._connection or host != self._connection[0]: - self._extra_headers = eh - self._connection = host, httplib.HTTPConnection(h) - result = self._connection[1] - return result + if not self._connection or host != self._connection[0]: + self._extra_headers = eh + self._connection = host, httplib.HTTPConnection(h) + return self._connection[1] + if ssl: + class SafeTransport(xmlrpclib.SafeTransport): + def __init__(self, timeout, use_datetime=0): self.timeout = timeout xmlrpclib.SafeTransport.__init__(self, use_datetime) @@ -1553,24 +1596,22 @@ def make_connection(self, host): if not kwargs: kwargs = {} kwargs['timeout'] = self.timeout - if _ver_info == (2, 6): - result = HTTPS(host, None, **kwargs) - else: - if not self._connection or host != self._connection[0]: - self._extra_headers = eh - self._connection = host, httplib.HTTPSConnection(h, None, - **kwargs) - result = self._connection[1] - return result + if not self._connection or host != self._connection[0]: + self._extra_headers = eh + self._connection = host, httplib.HTTPSConnection( + h, None, **kwargs) + return self._connection[1] class ServerProxy(xmlrpclib.ServerProxy): + def __init__(self, uri, **kwargs): self.timeout = timeout = kwargs.pop('timeout', None) # The above classes only come into play if a timeout # is specified if timeout is not None: - scheme, _ = splittype(uri) + # scheme = splittype(uri) # deprecated as of Python 3.8 + scheme = urlparse(uri)[0] use_datetime = kwargs.get('use_datetime', 0) if scheme == 'https': tcls = SafeTransport @@ -1580,11 +1621,13 @@ def __init__(self, uri, **kwargs): self.transport = t xmlrpclib.ServerProxy.__init__(self, uri, **kwargs) + # # CSV functionality. This is provided because on 2.x, the csv module can't # handle Unicode. However, we need to deal with Unicode in e.g. RECORD files. # + def _csv_open(fn, mode, **kwargs): if sys.version_info[0] < 3: mode += 'b' @@ -1598,9 +1641,9 @@ def _csv_open(fn, mode, **kwargs): class CSVBase(object): defaults = { - 'delimiter': str(','), # The strs are used because we need native - 'quotechar': str('"'), # str in the csv API (2.x won't take - 'lineterminator': str('\n') # Unicode) + 'delimiter': str(','), # The strs are used because we need native + 'quotechar': str('"'), # str in the csv API (2.x won't take + 'lineterminator': str('\n') # Unicode) } def __enter__(self): @@ -1611,6 +1654,7 @@ def __exit__(self, *exc_info): class CSVReader(CSVBase): + def __init__(self, **kwargs): if 'stream' in kwargs: stream = kwargs['stream'] @@ -1635,7 +1679,9 @@ def next(self): __next__ = next + class CSVWriter(CSVBase): + def __init__(self, fn, **kwargs): self.stream = _csv_open(fn, 'w') self.writer = csv.writer(self.stream, **self.defaults) @@ -1650,10 +1696,12 @@ def writerow(self, row): row = r self.writer.writerow(row) + # # Configurator functionality # + class Configurator(BaseConfigurator): value_converters = dict(BaseConfigurator.value_converters) @@ -1664,6 +1712,7 @@ def __init__(self, config, base=None): self.base = base or os.getcwd() def configure_custom(self, config): + def convert(o): if isinstance(o, (list, tuple)): result = type(o)([convert(i) for i in o]) @@ -1713,6 +1762,7 @@ class SubprocessMixin(object): """ Mixin for running subprocesses and capturing their output """ + def __init__(self, verbose=False, progress=None): self.verbose = verbose self.progress = progress @@ -1739,8 +1789,10 @@ def reader(self, stream, context): stream.close() def run_command(self, cmd, **kwargs): - p = subprocess.Popen(cmd, stdout=subprocess.PIPE, - stderr=subprocess.PIPE, **kwargs) + p = subprocess.Popen(cmd, + stdout=subprocess.PIPE, + stderr=subprocess.PIPE, + **kwargs) t1 = threading.Thread(target=self.reader, args=(p.stdout, 'stdout')) t1.start() t2 = threading.Thread(target=self.reader, args=(p.stderr, 'stderr')) @@ -1759,3 +1811,215 @@ def normalize_name(name): """Normalize a python package name a la PEP 503""" # https://www.python.org/dev/peps/pep-0503/#normalized-names return re.sub('[-_.]+', '-', name).lower() + + +# def _get_pypirc_command(): +# """ +# Get the distutils command for interacting with PyPI configurations. +# :return: the command. +# """ +# from distutils.core import Distribution +# from distutils.config import PyPIRCCommand +# d = Distribution() +# return PyPIRCCommand(d) + + +class PyPIRCFile(object): + + DEFAULT_REPOSITORY = 'https://upload.pypi.org/legacy/' + DEFAULT_REALM = 'pypi' + + def __init__(self, fn=None, url=None): + if fn is None: + fn = os.path.join(os.path.expanduser('~'), '.pypirc') + self.filename = fn + self.url = url + + def read(self): + result = {} + + if os.path.exists(self.filename): + repository = self.url or self.DEFAULT_REPOSITORY + + config = configparser.RawConfigParser() + config.read(self.filename) + sections = config.sections() + if 'distutils' in sections: + # let's get the list of servers + index_servers = config.get('distutils', 'index-servers') + _servers = [ + server.strip() for server in index_servers.split('\n') + if server.strip() != '' + ] + if _servers == []: + # nothing set, let's try to get the default pypi + if 'pypi' in sections: + _servers = ['pypi'] + else: + for server in _servers: + result = {'server': server} + result['username'] = config.get(server, 'username') + + # optional params + for key, default in (('repository', + self.DEFAULT_REPOSITORY), + ('realm', self.DEFAULT_REALM), + ('password', None)): + if config.has_option(server, key): + result[key] = config.get(server, key) + else: + result[key] = default + + # work around people having "repository" for the "pypi" + # section of their config set to the HTTP (rather than + # HTTPS) URL + if (server == 'pypi' and repository + in (self.DEFAULT_REPOSITORY, 'pypi')): + result['repository'] = self.DEFAULT_REPOSITORY + elif (result['server'] != repository + and result['repository'] != repository): + result = {} + elif 'server-login' in sections: + # old format + server = 'server-login' + if config.has_option(server, 'repository'): + repository = config.get(server, 'repository') + else: + repository = self.DEFAULT_REPOSITORY + result = { + 'username': config.get(server, 'username'), + 'password': config.get(server, 'password'), + 'repository': repository, + 'server': server, + 'realm': self.DEFAULT_REALM + } + return result + + def update(self, username, password): + # import pdb; pdb.set_trace() + config = configparser.RawConfigParser() + fn = self.filename + config.read(fn) + if not config.has_section('pypi'): + config.add_section('pypi') + config.set('pypi', 'username', username) + config.set('pypi', 'password', password) + with open(fn, 'w') as f: + config.write(f) + + +def _load_pypirc(index): + """ + Read the PyPI access configuration as supported by distutils. + """ + return PyPIRCFile(url=index.url).read() + + +def _store_pypirc(index): + PyPIRCFile().update(index.username, index.password) + + +# +# get_platform()/get_host_platform() copied from Python 3.10.a0 source, with some minor +# tweaks +# + + +def get_host_platform(): + """Return a string that identifies the current platform. This is used mainly to + distinguish platform-specific build directories and platform-specific built + distributions. Typically includes the OS name and version and the + architecture (as supplied by 'os.uname()'), although the exact information + included depends on the OS; eg. on Linux, the kernel version isn't + particularly important. + + Examples of returned values: + linux-i586 + linux-alpha (?) + solaris-2.6-sun4u + + Windows will return one of: + win-amd64 (64bit Windows on AMD64 (aka x86_64, Intel64, EM64T, etc) + win32 (all others - specifically, sys.platform is returned) + + For other non-POSIX platforms, currently just returns 'sys.platform'. + + """ + if os.name == 'nt': + if 'amd64' in sys.version.lower(): + return 'win-amd64' + if '(arm)' in sys.version.lower(): + return 'win-arm32' + if '(arm64)' in sys.version.lower(): + return 'win-arm64' + return sys.platform + + # Set for cross builds explicitly + if "_PYTHON_HOST_PLATFORM" in os.environ: + return os.environ["_PYTHON_HOST_PLATFORM"] + + if os.name != 'posix' or not hasattr(os, 'uname'): + # XXX what about the architecture? NT is Intel or Alpha, + # Mac OS is M68k or PPC, etc. + return sys.platform + + # Try to distinguish various flavours of Unix + + (osname, host, release, version, machine) = os.uname() + + # Convert the OS name to lowercase, remove '/' characters, and translate + # spaces (for "Power Macintosh") + osname = osname.lower().replace('/', '') + machine = machine.replace(' ', '_').replace('/', '-') + + if osname[:5] == 'linux': + # At least on Linux/Intel, 'machine' is the processor -- + # i386, etc. + # XXX what about Alpha, SPARC, etc? + return "%s-%s" % (osname, machine) + + elif osname[:5] == 'sunos': + if release[0] >= '5': # SunOS 5 == Solaris 2 + osname = 'solaris' + release = '%d.%s' % (int(release[0]) - 3, release[2:]) + # We can't use 'platform.architecture()[0]' because a + # bootstrap problem. We use a dict to get an error + # if some suspicious happens. + bitness = {2147483647: '32bit', 9223372036854775807: '64bit'} + machine += '.%s' % bitness[sys.maxsize] + # fall through to standard osname-release-machine representation + elif osname[:3] == 'aix': + from _aix_support import aix_platform + return aix_platform() + elif osname[:6] == 'cygwin': + osname = 'cygwin' + rel_re = re.compile(r'[\d.]+', re.ASCII) + m = rel_re.match(release) + if m: + release = m.group() + elif osname[:6] == 'darwin': + import _osx_support + try: + from distutils import sysconfig + except ImportError: + import sysconfig + osname, release, machine = _osx_support.get_platform_osx( + sysconfig.get_config_vars(), osname, release, machine) + + return '%s-%s-%s' % (osname, release, machine) + + +_TARGET_TO_PLAT = { + 'x86': 'win32', + 'x64': 'win-amd64', + 'arm': 'win-arm32', +} + + +def get_platform(): + if os.name != 'nt': + return get_host_platform() + cross_compilation_target = os.environ.get('VSCMD_ARG_TGT_ARCH') + if cross_compilation_target not in _TARGET_TO_PLAT: + return get_host_platform() + return _TARGET_TO_PLAT[cross_compilation_target] diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/version.py b/src/fetchcode/vcs/pip/_vendor/distlib/version.py index 3eebe18e..14171ac9 100644 --- a/src/fetchcode/vcs/pip/_vendor/distlib/version.py +++ b/src/fetchcode/vcs/pip/_vendor/distlib/version.py @@ -1,6 +1,6 @@ # -*- coding: utf-8 -*- # -# Copyright (C) 2012-2017 The Python Software Foundation. +# Copyright (C) 2012-2023 The Python Software Foundation. # See LICENSE.txt and CONTRIBUTORS.txt. # """ @@ -176,9 +176,9 @@ def __str__(self): return self._string -PEP440_VERSION_RE = re.compile(r'^v?(\d+!)?(\d+(\.\d+)*)((a|b|c|rc)(\d+))?' - r'(\.(post)(\d+))?(\.(dev)(\d+))?' - r'(\+([a-zA-Z\d]+(\.[a-zA-Z\d]+)?))?$') +PEP440_VERSION_RE = re.compile(r'^v?(\d+!)?(\d+(\.\d+)*)((a|alpha|b|beta|c|rc|pre|preview)(\d+)?)?' + r'(\.(post|r|rev)(\d+)?)?([._-]?(dev)(\d+)?)?' + r'(\+([a-zA-Z\d]+(\.[a-zA-Z\d]+)?))?$', re.I) def _pep_440_key(s): @@ -194,7 +194,7 @@ def _pep_440_key(s): if not groups[0]: epoch = 0 else: - epoch = int(groups[0]) + epoch = int(groups[0][:-1]) pre = groups[4:6] post = groups[7:9] dev = groups[10:12] @@ -202,15 +202,24 @@ def _pep_440_key(s): if pre == (None, None): pre = () else: - pre = pre[0], int(pre[1]) + if pre[1] is None: + pre = pre[0], 0 + else: + pre = pre[0], int(pre[1]) if post == (None, None): post = () else: - post = post[0], int(post[1]) + if post[1] is None: + post = post[0], 0 + else: + post = post[0], int(post[1]) if dev == (None, None): dev = () else: - dev = dev[0], int(dev[1]) + if dev[1] is None: + dev = dev[0], 0 + else: + dev = dev[0], int(dev[1]) if local is None: local = () else: @@ -238,7 +247,6 @@ def _pep_440_key(s): if not dev: dev = ('final',) - #print('%s -> %s' % (s, m.groups())) return epoch, nums, pre, post, dev, local @@ -378,6 +386,7 @@ def _match_compatible(self, version, constraint, prefix): pfx = '.'.join([str(i) for i in release_clause]) return _match_prefix(version, pfx) + _REPLACEMENTS = ( (re.compile('[.+-]$'), ''), # remove trailing puncts (re.compile(r'^[.](\d)'), r'0.\1'), # .N -> 0.N at start @@ -388,7 +397,7 @@ def _match_compatible(self, version, constraint, prefix): (re.compile('[.]{2,}'), '.'), # multiple runs of '.' (re.compile(r'\b(alfa|apha)\b'), 'alpha'), # misspelt alpha (re.compile(r'\b(pre-alpha|prealpha)\b'), - 'pre.alpha'), # standardise + 'pre.alpha'), # standardise (re.compile(r'\(beta\)$'), 'beta'), # remove parentheses ) @@ -416,7 +425,7 @@ def _suggest_semantic_version(s): # Now look for numeric prefix, and separate it out from # the rest. - #import pdb; pdb.set_trace() + # import pdb; pdb.set_trace() m = _NUMERIC_PREFIX.match(result) if not m: prefix = '0.0.0' @@ -434,7 +443,7 @@ def _suggest_semantic_version(s): prefix = '.'.join([str(i) for i in prefix]) suffix = suffix.strip() if suffix: - #import pdb; pdb.set_trace() + # import pdb; pdb.set_trace() # massage the suffix. for pat, repl in _SUFFIX_REPLACEMENTS: suffix = pat.sub(repl, suffix) @@ -504,7 +513,7 @@ def _suggest_normalized_version(s): rs = rs[1:] # Clean leading '0's on numbers. - #TODO: unintended side-effect on, e.g., "2003.05.09" + # TODO: unintended side-effect on, e.g., "2003.05.09" # PyPI stats: 77 (~2%) better rs = re.sub(r"\b0+(\d+)(?!\d)", r"\1", rs) @@ -563,6 +572,7 @@ def _suggest_normalized_version(s): # Legacy version processing (distribute-compatible) # + _VERSION_PART = re.compile(r'([a-z]+|\d+|[\.-])', re.I) _VERSION_REPLACE = { 'pre': 'c', @@ -610,7 +620,7 @@ def is_prerelease(self): result = False for x in self._parts: if (isinstance(x, string_types) and x.startswith('*') and - x < '*final'): + x < '*final'): result = True break return result @@ -641,6 +651,7 @@ def _match_compatible(self, version, constraint, prefix): # Semantic versioning # + _SEMVER_RE = re.compile(r'^(\d+)\.(\d+)\.(\d+)' r'(-[a-z0-9]+(\.[a-z0-9-]+)*)?' r'(\+[a-z0-9]+(\.[a-z0-9-]+)*)?$', re.I) @@ -710,6 +721,9 @@ def is_valid_constraint_list(self, s): """ Used for processing some metadata fields """ + # See issue #140. Be tolerant of a single trailing comma. + if s.endswith(','): + s = s[:-1] return self.is_valid_matcher('dummy_name (%s)' % s) def suggest(self, s): @@ -719,6 +733,7 @@ def suggest(self, s): result = self.suggester(s) return result + _SCHEMES = { 'normalized': VersionScheme(_normalized_key, NormalizedMatcher, _suggest_normalized_version), diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/w32.exe b/src/fetchcode/vcs/pip/_vendor/distlib/w32.exe index e6439e9e45897365d5ac6a85a46864c158a225fd..4ee2d3a31b59e8b50f433ecdf0be9e496e8cc3b8 100644 GIT binary patch delta 25715 zcmeHvd0dmn_V<&A08xWM1=%qID(+b!Aq!Cf(TWYCL8_uw!I+{_Osw^)!3J7mq*X_& z-L=|YyW6_dTbH7EMO)iy7i+6+v9)$Awo6PXg+F@B90_|Godb(UUoI=FEQP z%$YM2s_R3mPlT>9idxs2M$I@SsWxwz`}@2O2ph0y-g?BJyq-I6mk?IYdsYZ@=KX~5 z`5w#WtwXqB?v;7h1iB^2?+k_ug>b^$S$0nT(xLz-}s)xoMnrhpk*DzS0tdG{3S}xm^>6Hn7CnpBY?X;7L+w?_DttY(57ax z1W_RCRWhEB&$wO$3R=QVM93q&>tkVAtI8@C%;LCs84}P?p;p_@7x@Q7>%^hPslQq{ zfAqwY7S7kQrM2G(uD0G4S9`;mVYeg2=8qnSq^*GjXS{uHH;!{~_DHnBUQb4dlkP2` z?d}4W-c>*#;Xevkg8~w|thg+eaNL#}o*WeS=b!3NE{gl~d?te9jyFv5T}G<7BQ=H$ z3zhMOHj*9MhqpwK8KDV$-`-?dsLFH(b=Bpzsr~k*gb|$8yR6N-tcAXcQq;!{jk5KJ z>`mY(qzb>^@8FC5t~P4u&v7hV(1%UzUX*@k+1 z8+j&dz*sS++;e1zui!#RL6d_USU~+Q{BQGo&(RTqaz11YC}u6#5+}6&SVH|ld9#JP zdB$d?VJNiR*KMznbPZ4K-jXI%EpShYBZhGKNFzAQ4LzLU`G$stzc+_?b6Z#==MLaF zUv5ju^ip9|J(s06$1uN3qQ2pEwbB$R$CZ@S);p7I_Ez#jcw(jn#nk0qw_;4KbiUBC z>pr2o`rfPyQHFNcn7Z6n>xWiwvhi$|v%D?g$?k731fC;O!%^3l%^~U=d941IyQ7Bb zzN5kFmq~c1)U!+?UNjKjVUMJ4v18iMfcO>ts)6}`z%*6bD}g$?SpCsb2%);(GL1=u z=W?R0Sbf8Wk3V|$DF~^}R+oDLTuwwBez#vslTbvoJ&@-3o3YyLlhIInCBBfYQb_W( z+KK}?D-bXCM=wSY93hO2KYB8%WsUN*@~*HOxsni<%pqd)jI>y!O=}Pu#adNZ&bm{z zptj!ad`R~GBuTMP=lkrT9Jqly$3>Q-+CbrIau&z`Jae|;q3_mO! z?7Ka<-4}w|dV5;^&K{oDcBj<8eF{ktRc6}C1!kv+*XDl$uyhI_*rLn&uzc6OOgL;2ge$yKC7SB=8W_# zYx6t1V)XrA7qL3@(^?r3y#uk85mChycTj>>T%L_>1bS@*r~Ws(=F627L2&)W#ok)_ zC&n?_g31K-wBeAw8iKi%ErLS&;Vr+vq@=F)GMKE6%7bhb*c@H_T9)jOR+S1Ngq5yvBm0U7wu(3flL1HNU76WJvhNZYOTg#IsjGb+r4(D~@CPta zN1_68EG8W|Z`K||A{WMSvfA3tIlE`dYMa2}KV9Uly#RzBYs30$6sLVDRB^Kv0<@cP z*yuZLe!u;?KyCurzgyschEZcLvxa}fn1xb2N8~H&gqk5z-WR?^e(&xx?7Q|Hz7MkH zZ3{NExER{6xY)m&4Jnpz)JeC6Gc>wP{^)%4p!i=LwvWQP4J}BXP-rc-v;EH>orbi! zT1K=6vf3?ZL^1vCCX?^Tpb80&xE##>-JgQlS9=oQGzzUQp-1qIGX6#x_E5CcA8iC< zzS|Un$RiQKB!Bb_P#q%sP%M9c^i)Lb>xB-VMO`+kxh1U@gAAF=`%f<_#oCHVMX>Fuv_W5 z8|06!Q3;M5P-TAI35k_9XIdF>`bs`6cT5a&MUDC|^vkVUmk(p0O`Q zK?!~bH+jV`t>`y)NGqLvLm2yVk0Hg?A6Y!+lp@*sdixHc(mbRF$~huMQJx0g)5tqy z*mGX&mZFILoi!{}C_##5uuk-6ZF#y!uoJOD=H~8wx3`MH_47e~HR+=po?%<~fI`gHBybuu=Be@bxmI}#%v~4o-NyP9H zBhsw&_%$|y`r0#?1G}BEvr*p(szWxVsMy|$ai*ignHorg6}U(+r67C|P{As)QePrV zlm=+f5mU*e?x~W%{1jXx)!mat6Uj^6>x#As6+Uy7t@AI1)qYO+W>pv4&mh%iE20}< zOnQzeis-{2pP;W}J)JN_Foo&juQgnHxeqKcslC1h&Fa;UxMjVK?_m>YKa+qw{-=mEMeg0soxJQkm^0fi{Ssi_^#{S+De2odr1B>vCJHc2*mG z92+e9)x5>(cSWLMGC7e(#twWJ%njJ==>ZH%NrTUiectACNosGndeJZ>I(nCS71Syt z1(9j|=~S{XQWJ9-%g!IY8dD4rjJBUc6eBf`9E==jC`)JMI3&~qbBq;eqXP$HTWhh= zUk7oVjs7TQ-=7Ki2g3I|$YP_L(n(4WRl&qeD0OF?orQFP5@nzc>^(3?7^#GAhsL*W zMwRxJ_@Y}d{!Er!6Ja4_7J`xnhP~BDJ6J=u^hnT8W+hH9btc#uMni*TO0neH@4U;U?jXS~x+V(YZ~wfULE)H8nIiGdua z4|o1%xLsxTXEF}@FOa<9{xAGmmX(nN?&_Xtl5J?zie(pekT-kANp=a5roH5op4vVS z!~84}rU8q|vn-NxO3;KUjM_l=^d)_J6?Od$4QDJF3bMRc|M>O7@VjDd^cguD-@s;I zrDNpeV6WulbKv&1V8$_%vHILLXD>P#d~7i3{c`eGFFF6Ofg~}if6DU)G&szcCpLsG z?lOXN_V(pTd=ZshVXWp?qQ2W|GCL|S;suCmP%bZaCGSU7^&EzbviDnkbilIc^JI%HU4fnOy`d$4U(o%hBT0H8ED57<&vzW)% zn#1WQu;w6jM}@%GIK8wvs)$xF2KpXojT~eMb2epK!(NUX!8L5LWRj@^V&_^g;tsKu z?lOY4Wb2Rt!dWlbA+~w7ER8;QPeQ08esXDp`tyIWLYhOjFn+I;_Z*UY4vSaZs>Zrm z`m3Y86daM@DE1vb)Eep&*UI~o?*^n4Rf4wk)9LE^(gvEDjpMILB4ffEN<&PTDlIFH z_G0_!opmL&*U(axvZ8$kihm^D_vzqchi<`yjx3~a^d%G02lLDO5_kFlY+)lf!=b7^ zrWs^MdT)NJoE%FZ+G{8#v1D3<2ga*EdNYP{S_65d_bGGfU#MqVprXRUJ0@9?&G(1; zY_Z{*D~jI2sU@(UN5S5?Gz5D%1jVyYHZS z>LtE{R^OT0vrtsxc(OJ&u1^%Qp*DzQk0PNRGq5i}VqwNh>4`XUJT^I|5t{yZpjV#D zkx*TA1ub;KEz%xa$md8&VVokOCyp9H!JHx{Psa7<^GB0CaSM1`9vKk-K$oY`%@Qm9 zJBO@_kLM+3@@jl`*O)w^D*9>+xe}iq&i#RfUmrt~6Ve8JkQ+!EgX2Z8+;uz?H~Mc- zpe$o48*DN+K`nlkA)6C&P zEHZtVCbrO*8;Mp#(N$5M{r>`bnE&HkA?Y2QYPQtlk7;+vsiOd94{G3!jjee7bA!v zdAv%_W_nIfHlO-j3ARyTVj@5YPK=Fd;A8&gbf#Wp=AkbJU37sEWp?tYinW2)xKKi*7Dwqlfg;oihB z=&$D9w4?{CxeHrkGl7!*Zze?xr4{)OF>UjQP#RM?zrWd=M;T#nCA$MdK@dc z`Kv(Rq(ZmI*0-X&SlB?W0~KZTw}-e{mbx4rBRtQ>B=-U>LU9f1v%X~#DclB$(DkwO%{y#= z!t^ATKGXSYO0uO-0)JXb4)z%?Yr!TOlzBblXzMdB2bYiI23JM^W$<`RU1dJYCifbCJ1m@K>PxQqWGe)vho}|bLrVgN4EH$<{ zxwih%B=zxfZ+DhGlBR;($ieX9Xax>xN_g<~g3GD=>^i-aM(RyNz;a zu_#ySDU0MvY_w_!8Qxd%;B9s+u+p!yAcV)Eo8Xj@Z8%gn!VfM`YHu1iTS z+BTNsw#uPL>A^9~+IqAL9VQ^I`+V9~Y!WDU1l(wx)41Bo*8W+X; z!pc3y#U@t|xEyg-izg4Kq^Y;Fwy%aw;0UK9Amb&>tJ#Kw0X%E%>@aW?>NS<63-8-z@edUSR<;tZTGx~%{JH|TxGYS+|H4Z{J znsjLG+B@}piX$GIx(HNcaYhJ5K%x8rjs_cD-HZm-rls`x2FTt0vC(qoUtr!l>lUm^ z$Gl{>0ThP_ZB(%&Oo=nmM*A}y+D34m)w{X|tZmF3uf_@IF$j+DXoqn)f`iV6;x9Rx zqY!hoey6mUe#DMftby=i2|bInQ`WPuK=)0Pcy71566iWaJh^g?J_015(2?a*4VFdi zSt#iwwlP3SKg9UJX6>S$sxvv3js57$NV%+@L)MIE?iQpv+zk#NRoYErt6;>@& z&V^fVi?IUnUTpcO?_&bk>iw@hgeW05PAo`u=Ngf$J;i`~!GsNyHE!P<08 zwo9&}7ZfGc_x!ZFgGR^YBhF-#LX5@``Tw9y;T_5vZ2 zjuZ&b_F{^+jqZgaz|xU85~%!VaQdTej5J}ofq9t|y?dqm_6lf7SXtU1KdwWQE$v$M-A`f3yWZ@Eo+ z*9BLYl}-ge_28>{m@oITPvYz1EjVW_qyu3(mSB*|s4w>%*geZGaEyN><9`rejk#?Y zWWO)BiGC7_k)?LD`I6fUqOc(c47;t0NSrnzhOqDGboQN;#=f<3_H9bUclqk*d?AQ1 z3PDn=5QIhvLBB{L7%D~Z{8OP81ioBut40XLTXhIwg>KDZ@!F;~HyRdN&88Ipwy@N# z8pIky)A8K}MqOj*5Fs__(De#7<1@22?c&cZ zK$*4aF);=X(=4L>8IPTn4I{BP5+NYWNW5dvZ_$yS12wYCFm!_oLwm*V-#KaEqoQyr z%A_L;$yb90=S2aBx`5b0gF$Axd5(5n@i7O-XY&a7oa92X=I-Q3mdta_DUNWvMo5nS ziR2CLr)$J!AFM0#H%xA#y&4pDvdERPV~Ca7tyo@1qOd88VP467xl#VmC>hy0xbN`v zt7y}E;8&kz7RCj1at36d897WLHyMOurmXHGtP&h2$+f}#-BTe0FhOKB>jOmC*qR+S zx)735QX-r=ec~NAEYOcqtBrmnkn@_?`dU$b^DKvseu9&Ci7!Gpor>SZnQFXXW^`l? z90lt!uC?_7&D9GP(4T%l`H070{t|aIC!mr5Co{G{C+`f=MB-$Il`r(8QjA(i-LRbK z?ojL)NjMpA|*wRwaq!@b{5i~ zn7MF7j0uqWCP9!U_@m_m$digVR`zR(em&*?QAp%Bq)CyC)OE^!-K}|@ZDdVSE62&! z6>!|YFl(G|l!R_8AWtfX@v1ZApmKJ~m2u1-z;qs0f3EiXFT($YzBsrE_8TatkeZ($ zIjVe7J*iWrC7;C5*$#4I<_j=GvG?*UhgOX!_J2`KN1r91smAm9S4kiBz;x*uHoMrZ z>NI~cetoi2>}zbE=ZK|cI7(yBH{5o~=xkD{P8$%IJ6B9GU4%g^p~GQu2X%2DYbA=cJQ+U35lSj zNxK&u&txRq*paZoAAJVFVKIjzp;f1(5khH~7NH;L+IF_yzQ?9y&%qiF_#82)jd@cilZlxb zB1(LjWukoYN2Yvi8%`TI8kWFD$wH;97QgmHXG{DiAy2UmsF7B-1!7ZaVfBp{`@Lb! zU2OEoSD>nI;C|&YnWBqxKgQy$y=cZs){Kp8lNZd7npP%CdGH{!jpY%6F>gp9Ha8H{ z1Y%PHF=HTB5QtdQ`n~d)`#jWXu4or6)^wM9@#+aR}_`%k7ZB zc(nIA`hoQ-Tj95&&rZn7&tDRy-p1E`Nxb?2C3g)SbGvn+8XR$9hBVlG+HuHhug2IG z`%l}QEZlm84cTeCU5E%XdD=chhzNsu+GAoIaEdmOE?Is1vw7dj*tW3wYYb(H7571Y zgt6ZmnBX(Snw1*S?G$zqEo1O$caiEW;{Ya>_MrusT7m5U+JACuhj`@Jqj*@DB08iO z`6+9#sD;Q3DH@iexWl)~JgX=SZdu%_`dV<+@Yo7AHSL3M*Nql?TE$haBI+1T$_y#_ zpZCPFV@JW>b)zB;4OM11!rSQnAk{hike|2KY4!%49)j-{P7o85cmLPyYDaAIf4voc z-_H!ab?HJjCforZ31deZGw)rYj#zjVJ^i6pAUslg>s1I(80=G8_!&turt^x=$Y^77 z*p3#ssCSUH#y;VD_cDAh@Ij(oWWVvvxY%nL7(v@^L5Dc2=Q6ulW%_J|Y5Rc4vk;1| z7Ef#Ms^6LG+9&>Z6iA(pqc-R|#%nbX%sT;t2L`T4Ruc}YBZTf_1dKxjb4kD)kk$N- z8dkKo;agUd0VeynbErKEoAQTeFj>X)#f#)ab6@umjK#j&m=xshV^0JVTrcGpgPXEb zs;;-P5B&sM%tnU{gHeT=jf)1^XdguC#JTI)+#ehKsUzgnEqfFS82n9OP$2E_DDHLxl`KxMvdIoExb{vT58u_BUeYJ^XEPyiKB)L>0Hq5fdy@&AE1pT zTbO&^JKFBk3W)0hdP0eF)X=2gknBwo!;|i5Z+9g(hB6cr35_W^MGlVY%ZpEu^P{wE z|BTIX43VB<8belN15aHpwL*z-K~g50e-WdPO2lKFHL>LFoLqiPEcr9%cDMEhp}kS| z$=As3+)V!IaH7e}k_X${!(I3lZ>hb=XQJfEyp6Fzrd~oXI@!KCN=&0Q zqqiOfH=T+TU|?5j##@CZn6pjvG-wSTlND==bMInjH-<6-T8Tv#IIh6Yt)*n| z=x4>OBJ$zrN6BaTqw@HFLlE{b2MK)+5`r!H@m+|(h>Fq2x_VjZ`6HbSJp=~-o*j() z_#OJeBV_WJKE*G9;(PoA5(w8!_r!mrBAaOO?=-&+_ zZtP@!@inq=>`>m;KwchO#y`?qojL_I`f*YojfX;9W|Fd^kEtTRscPL!4~dp z3kCnw)ycV1#{Xf44u zHb5Vm4k=&M2df|o!dv$^h1M{DNFH(Z@81^(Vib+j!vDzH;OaLJl7%8!nM3rsQs@6r zaF0dl?vac;P%hXb{|9z7E(GUyIE3-4>_E_?5ozB9)lOq@)Ks5^d58V2@d9QUTbLtc zx_4QV!NhK=mQD2C6pVR&;tpqS#{ofRrP|+4Gns(-6yndl;PFQ-O&9n&&HIxWQ-Q(` z0k0hdGh*boMX}va|A3{0`YLcvDrq3y?TRi>90O{lHTh(;ea6^i#H_Rh$LNxqXG#+4 znX`hKm{InZCb87m=0rFp67U`(;MtUb81!rw zLM6hp*@sLlQ1P*+$jX9r_lY~7!n$#3lEKG3=Vc*>5{`E}RC4KoBKEE)NDc3x=71{r z5ek{Z*GBh3NqBQ{(2|yHe++)T!Dm^GIK#oXmCD(;4aeHu4zC!E6p~Vq)cjMC`s{WQ zNgJ<9?|%UOtCt&EoKhMu;1)xRD*{(W^vQ{k1?+<1Hkvh&EFV9NZyiP69 zZ@qK`cf%r?V`G$PN@|u<~1QE1FE76>{L{%h8tw6snnB5nh@vW@a#+D+pe= zjrXH{GH*`zbZju*iQR|zWFep47$MI0H8_{78bdxS^zw=HCtcM`w&x?+K`)Ee9>=w8_E+R$(YX9*9G2PMm3v|Ew3bs8W-{}O zwBSkfou@r`E{uXT9d8aUe9306;4*w(Ee5DiY+aXkk|h&AxTgu(lP2=J?jsu}jfrFL z)*PXw@LI=E{TCgRMMr!=ZcgeqwTWq)`#|pWj7kt;6Z_le9lUH{_F#bSag3fvGGja1 z$wp<+fcH2?7u(X}pAGW41wObnx#staXd=XlOrO%5oUlbijJ;FDgwM%SQ{vq3ffDeN z&St_HFgvgWqhRwhN6}2?*2k-vHk@(-e(^jkCg0KKeBnVyZ@o)e)x(B~XZu(s79lfN z8!@_pd4++@C<5<3@Gk65-(w}Elz1h+Kbqf@I(korG7jhq@_KRqt{g-euMIAh+)utO zw%R@Y#Ej6{E-bD5`oM~Uu-!g+*T^*&;BFz3)yrtCS5aG+V7VV=dfS&4_pO6Y|#^duw=l74=nON3cBzl4jE?0_b$h7joqJhL)KAK-@A)lA0 zhz^iH%XOliWZ0bHcyd-cr!Swqm8_UEd*IF&nH63)hVwmwiE~-F1Ay2?@Ol9%0eb*% zA&qRAn>@S_+9%kpXVL4u&{OV97cnht_dv(v>F_%oi{B={&y7jyc84+%c`iK#c3c3m zw-;}`Oi~=Px{rdS32!%CiPU|V5J!H57ONsy(ek$lbtL!hdWTZ zTnmE#yyFmKqhohwEf_0`VW%71yu1io&!=Vg9?Y+B@P6`KWhy_(Lh37%A_FI)P?`fC z3``>VtuoHNwFmasI%k+@Rai~E>sN<(bTM563xwIXUQ-LZfAl*)1~+ zO(k`cMa~rVi;M@s4c>5!UA7Hp^U-*36&SnGc$*(`tr{bD%r6u`P0zEDV`Xu8?T3*Y z{+~m~YNRm&8$;Q4)_kGEO;G<&vg_dzJ}jHas(#}8n#uL5zI?lh#4H@sKR5t~q4$M{ znFkfH^=(BflQhBu@gW%KEolSD;|s@%H^9muUoOlMrIR@40Fi=>a3$j4J;nTM;n2wTLKP67V3mJdl!d?uq4!#9?4t;H&1)DtccR-Q;w6Kk68FV zW-{@Sk-ZONN#e;ycRzgvSYYS4vX#8~$cf%f5Zpjg1F%3+#8&d;;zgSG{)w^w7`{t9 z^zTdan1stYZ!#SStdWbwf~Cno@n&lVRy=z^Vag%8M;*emi@^P49qt6HMgo z5L$z89NltTFO@@I`dv*OKM`@|_%iNn0v&gp_msgNc{*4rLblmh-Vocjs1l?`4mZhM z^JWj}yCkkx-ZoSc<~zxrbo>?Si`szmV8N1Pb}XE?WT^IO9EgN77JTG!bjNNy_Sdn4 zh12TKpbtQkT=FyfyzC6jqPpBOjJoq2`D{sSRGdzbR%nvLg1wYMGcWACy`+n1XxR{! z64~vr^91g!Q`2x z{Vfxq21;o4&fuirInt%+BAoR7?Kt5l6gzlSQ1Bz#h~EgI_``qu{YP`#`DU|FwfaZm zUz*InmPq<6>zl?*X*LkTi-Ka^?A2Kf&miV6R`&`WcIR5HaagX>*r?U%C>Gi zygcCufm~nqQ(+%UzT-g#sb z5EHMsJC5_0Cc(mGt~GTL6~6!_h*P1Q83AS#FPsY5E9KbWD`l8L-@V%*u>)Qp&gCSF1Gv^Ql*{&GU=Qb5Uk!FP-iRy|h%raW@;`N#I;jhc5A+A45F1Bvk{iz2Z;X zux0}>ujrk;Z9N*7R=0VBg%yRwVXY0}gE0Patpt zDgd(qQvnkJ;{Z7u2>)1q6yx;yS;^Z0y8y>FkkZFyM-17>5^(vGR(I}c^4Vi?y%kRj ztzZdjEWl?2RzFSJAM3*p*+}|$(pfOZqv*#tP7DWC<=3gAGN03rdY8_A2F{HQ=B zMi6oU69M*({DT5`J=ytT(bwFvYJRNC7(&Sny?}X z)&X-?{)IWgy#ctaxSp`EgsrR)|JN)&M2K4g@%uoC6bLyT3|q$mgNQ`hTXCa;B9Uop zK4@Fmm#=^Z9(fxMIY;Xw7_T!M9Jpv8-`8eoS-FN>>Kq6f3Q6&w^yXd$61Rm_z*1y^ z6bY82km%PMNx|c3>A%%5V@)tYVQ@+Ac|o)1wT>miZ6MD)p2yFAjhug6FCExPU$~z1 zf5O8*yPdr8L_alqkTI=c3k$pvfe4+xy>n`|li<_{wc)ueTO=VXM~gJ%zLjHpXP&_N z{*PC>)4k-4l{)5(_<5y+U-mI6^hWb9uOxQweI^Cw6#GCiE$|5{{W{G70##5_S9Sp# zBpm4KPiXCm&i)AxbS2D4jC`E5cumaPJbYDORA0Et(%S$Z^?z1B>k0DDReeNXkj7Oh znr)a_)Q1i<&BPc1_}A zGU#;{ERxg9LYPM3L_<8Co)f|$^h+UB&=W$ap+|%;k?t2l9o;2_M*1c~&k>8?y~skh z2(%pfj1cD2wL&Nqp+*5UGE^y` z7SRmg{RKRSkrM!oT0mzo^y6x> z@5#Pnw}XNUY5YjUh@F*rc%BYW@OIb|q|^i`3PY$pNU0Jii!=spg+Q^<8P%l!Q)%w3 z0BfWnG(X6i9-u@SLdOLuy#th3L#Q=K5eF!VhR}&Y%2kX$nk6@cP7YF<0+cjE=+q$P z(*Q+h2(1iK_D>0rjD}EWko0IP$tj$s{|hKqRNC=Qpagub_o ze7`$s)=IV;vg`VepU7Bj^y(|%#;%j(X>VHeF@8$L;xrxZfR10pu&0GI2827m2Wo_| zR)QxEKT7WbYKwh;!ci)duHlmzY@IXZZiW$YJP?8(L_QGzXQqf1dPhCo*H438p1ZsdfE zt7SE-5urU^C9*9^eDz5(Yzy83oh0M8%;USqkpo-O`E{}6mn|v${8-ZMWhIV?rkB%2 zt)%?r$@u>8<(d4aa+0!jFuz4k?%(Jg)s(R8nQe6XTRU@ zz5C;4Slam{`Be+}roWnuHz>Qyvn`*-Ic3%&I;IrM6p`z?VX zkumHBite)`vD8&(Wce9OIU=}8xMCe!_p`sQ!%daOG1-PpWl?9!c&Rm1SC zN?AnbGa3%jONHUbk`dgK?DZ<{hc9sK<)t%@#<5luT<}V0+j`=AH&IprLZG?HthwZ! zwW<8fe~~ZOE@LK1(Yi@}*Dl1AoPxas|KHz1|KDLmgJRgtO^&bY*FRzza(}??IM{t# z0xop$<2YQ<^zpSEY7aehJ}gAK8?qcl(%9K7i##YgNnRi^c&W0R^y5qS;ZPdO^WKP^ zH#Ss=cv%dYy>Tsn_(yVc<9z<}E@aN8CVt49WXv;4_yqs)%qvj&ct#pD1c{KfYy+)q0!T=csZ?qR?Iz;S>K^btGZ?_2)jXgBu=#OgD= zdTn&|pzl7>F_peX-`CA!1t(rxq2Lb@{JzZX)`$m#!oWhpyWmeEAWup2XaA?O{@?6w zcM{%CKs2N5)l3Z5BM0+EOgrbD=7y~={N5yy<}I5wOSEtIiRe@N0Qm@Cs@FXO6pPXGZp{=B{7H^rTGkgG<|jh1vZ5 z|Ne^<*+HCZJx!M2JDh+89{ke?c%R^svd8JgvRd}W#zs>Ype6e{q-&Us_FRz2dIKp- zvC%N3-hJnFV?OzAd!IpH%@I3`jb*N09f$*-IFowG>k z>$70xAAUVUG?m=Kw}vR*$lzalkl5eIC-1!RG>O`AP_r0|cq=<}cX~U8yOZT`oZ^0B z^(*KY{(^@u7x$;jyTO$aOBcUFet0*Lcb+7FzpLbz93+bOW~IK2os~WFgX<37C5fX8 zeH0BV@iC8O^x$?Y{kRL+^WN}uW<3O#^Yn)xh|o#Z&`CwOyTx6*;d^J9!&F53OeGOJ zbCRF>(C-iYY=hmH;{vys#tDdXJbltl7VMPsGhZZYcgFE~FOoNRj^OVvC0BM1NFR3q zRfjZKX>YdY!pUlAcRs-WdIoba2Gj5~)J4p@1`ezYXfc_m;JTkav79w;5L#4B_fJD9 zdRMCxKNM@DZ%-p@cBy&ki{#(CM)JpAAYFD3?b>Mf|E){%R-bt64>MCKkvuryb_9yB7zDjOKOQ2RaA%Kq* z9?=-uu#~k2GY%7fDVg!Unh*bhYbE%y(^IQ=9z^0n`iQ$!CzHLJ?x!y9gMDl^y`1HN||Sn{s6SY zIKjgaS4*nvI{ws(tmZ~L{7&pt&HjpzgdV$gcU9jYQC&|S-y6>~@XLGOjcHiYIqlMN zW`j+}YK)k@)A6q!qKHC|X$S1IcDU9*3L$^(OX)uDl(4b?SPrdqyn!tGz>tvoXou-G zZ#LR=v>Eal%TBfbo=i@BkRCnepXmJ&DLu2eV+QFbi%FM#$Q9eR6*JtK`tWp<1Q}Y;@X8Hb)(I_Hpzil=huFd$@t|g8>t?gm7!W^AS?E zKPUZY3F-}A(|;h~o!9ik{=i6M8gWhkZ3{WIe>lIfiSP$h@yywF_ceWV5woWX>B1{y z{DDDypC+>8fWdrh7ydRz;73BICNVi?Zp+)wJ{WIYV|-j8^cTg8@q$^(e1Jt%fer|6 z@@Wn04pyd)`;~ccnCEdT8{md5?3+m*euLGH!vcOdL+}nQ?9I0ptLzfT;i{U^U=Hzz2XcfXje3KqOpIX@GHn zQh*as4cG+O4%h`a3^)n+7H}27-$vsA3V`u8&dg(h*Z?yC6@cY{HGoZk?SNf?qktyB zUw}w7v_HTCumZ{f)qp1f`RE<{h|#!K^!i5t`)o(J9>5+fR{{0{jsiH9kbLLsVn6r@ ztrpHQ)xr%wY~eC%ES-r>pIEq!{Vi+zM)&UcD{?TmIlL0!@^QM(zd$}XG)VXrN%hW4 zhq`rf@Awv;MWjwf$V^iXVOIouzk#72V@KCzySCHZF+W%Fh}G9JZr&xXVroVxqasR*%f6~@`YtiC#yP8(aw}}t3u>#bKcy7 zH~T1cLQzi9un_&C`~i>VKXpd-(#)?i6Ud1la=I#VAVGN;mUV0Kr{xq)v3yw{GPWr> z$A@(@W9rO|5~3d!VDRJ<2U32>hmD)hn!}$MOI|;pEPHI6h3lwk z?Klg0=X{(jXZ#(Sb-Yj`f8ELo2!C#Zg?oEKpiJ##^1zSDA>mU4QQvf-4VKxAD*LGH zPOG0TLmB7e_}LE&96J^=4nAg`5Igmj5WN02Bbp))Sh&mASyuy}*{ujqVnm^}IzFdC zn|%sv^)6Zz*5*W_yb$yMb>IJ6-A7*e`Jv&J;-WHVzOa(U%r9S1HBVU8(6r*>nRA_a z3#y8+v?|K(f6!4j%bB~FOuAS{RJ5n8mLFZ_oLDxm(mK;=FXqVki=(GK!OALf&JGj@ z^&T>A!OYp?<~~?8v+B_i9CuBGJbBOt6CbTCx76w^Cp#dGJ+Uhqin;$lwCTg%EGDkxjHaORw{V(bhd z!FG?CU+Hp=fr7x^$6FUHbl%~a#@f*teU-~8TL^{nXrNGTQc>;%wQAVxiV8SZIM%%h zW%Cv+Dnr9b-7gQ3SE(X$F6(`fvkGIw`ch16)SB`umpgyfg4t!(1#{;+%c{6~@x-#K zd2{EZz1FG)v(QA+{bFx&nx>PkG<|3HOI<~A3+@d6s0FV1PHgDnqKdMzN)q#H5}Ekx zi2+XqyUfO|nER5?FPm9eGv3MKl({VG^dg^%~1*BkTDB`)h(#X!(h3p zP<`1#j3(LIJd}Lj{B2hY9#XK6Wx&q(-xu-gevjO_reza9VAnKk1ylkm0CvD^zzo1tz+}KgfE6$fkPpZK zSO7Y}5I`Cr5y1GPez3SXDG)KB^?O|C0WJW}0KNpA1RMeE0Bi%`_m`X~sF|Ked_7<_ zpa!rM-~?0vrUJ$RjDU1N6hI7U`Obo;Ubx-_oB^Etj+Ea@8C4Cm0x$zG5pWEU4=@6T z0Ac|Wz;*nfqX}>Za1yW|upUr-&O)BLI*sW5s!XY_=F1u)q~D#qB98D*4u)R zrQ_$quBc18M0%IJ8P&j_g3>L4;TEq}1QrZ9&{RQiSUPj=KHC z!LVuF&cFQQ3*3?r4bp9}#HQmrmzw}=40k^_9`P~w&I8Uq2gFy}*_5(>Y~DJ;Y#|kP zG5(=H_k7CHVG+u6!rGe)%Wn=!o6A+e`Wu5-Ikx~eOWa6-Nx|vxcP8}!4fi1Xi<8_~ zE*hCffoUFB3Da>tOvv#F=OH`$6rm`V!itmsZ^osZo5U^bG&4(K+=@CY&jEcQ%uv=S zcO_ak7nRKazbZ5r@>+(hF65ethO)}pM+GbVP6alh;)QpWtHgp}pb$bf;_T0AN;r{Q z7z(G5VLpnixT}wkaB37h3~8)VJW%FRQOP*qa|BK{HkB9^){i-uZK!x*yzZsRxpYtm zVvPRVh+*IBu*dHmG97Cj<~~q3n?2UJj`?G;z(o9KBiZ>wggl0Q?o?#gyt#D2F*6jYnmOJdQGF|Yt1!Hns$X& zl~JDY?~LA=%1l${$jr%^Gc)I8R%foxd?9md=Jw1_GJneKs_U!E)XmX(balGdbo+E4 z=}ziS>-@SXeS*HPzQ5j}AE6(sx9X?qXX)+wrTSI+r}RJRZ|M7FJ>|}NHfvwjH(702 zlMJ&As|@=LM+_eunhduMGGm_6VO(PT(0J9TGu>w@H7zi$G3_#aYqFakH*YiVGApxx z%D$9+HTxD@qjHYhh9E()MzKlphGLImpW+k6*NStBRt2w&Ri-LU%6z3&IYU{ld_=iJ z`GK-Y`HNDda!XWd)wil9)m2r6+Nt)c*Q*kv-@Qa%2s4&WRJ@(%6>4Lv%t*( zRoF*asH{{zrM#+)R%NP2surtOs@_vIs7|WBR{f&7rD|7AP|sJdRc}+DQva?VsY%qP zX^q-q?c3U~FtV33;xgr#gWMRtB0;29WvSOeY5GJL5hQ3@sU%y-bf&RMwmY&NB&5~yI$m)|-o#oAXDl5j&-=H;&G~^rR80H(6 z7^)4c4eJcg8MYW+H|#O&hje{pIBhs*xMa9yxMc`2y89Wkj5CazjjtL{8c!R~LD;&Q z6sFfrpP0Tj$;>h48Rl~HQu7M4*Zh?EY4h{ut>)Lv@0$0T519XL{?z=f`3Lg_^RMRL z&DSwnevDVI?19Yi%~elR z&rv_7UZq~6ei46g;0N_}bwABW&6Aq_nx8d)Xfm}X?P%>eZCr*TV{_)ana!E4nY=Di z7pF_r73&_sGqW?#W?J;4^^>wHvUr2qu-7=>wAeJnY%srP{?NQW z`^9W*6r2$aFe{cS-d3DeM53atsvQufvbQSoI|J!|J!xI*nPgQM*n1 zmbOLf*Jfsn$apNHE@NZH_Kckw3w1TR)w*@Mjk*_fUGy^jV11GPCH))v5A{Flujp^< zL$bPM#bjBs^0WRlxQtE4YsQ{#Q<^E$G{)pLy=FoK)(Cz8ZaJjk8wID7D+eh@Kx288 z8LIuNzUrasSJWS>kE#DskJT*Itk*oJ*{a#Dc@qlz8;xI+gawwa9imleHCi2(nnk-x zyGFZSy9r~+XGCPAXAIBKWau)C8PhVJ$oN-Aea4Xt_pyu<87DKo%s7*AE~9_ukW59U zCQ}E!{&wa>-6OikbStq8U(mg++oRhLkvRdqeol8$cSZLXmZb#CGD$yKU!`BIk2J&> zk_;oD!c)<&c~I2@na0};-x>lonmW~3)l}7Es`aW@ReMz*sm`f>S8;B2Pj#9)L!GaF zKs{KigzP%CtF;LkOEQwM-3-;)bnobrvTkNI8d?qAjrSYp8@riIrZUqHrY`1z=IQ3S zW?Ml2?_hf$$Hgd2iYJwujxRaa_w&`=J+o)6Jp1mV77Xfs3` z)kgOM<3?j&Q>|&c>9FZ1Q>b~od82s*G#yqb+=X!DA}Cd;(Rz>Sb=CK1^A0HI!|IRK z->Z`}12sBLj^>Ic4~lt}_8sk88M`tRnVFgEGC$0Wz*sfwTlFzn!wm(7EynG}lcv+A zv!?T=tELPy^Y`R)a9XZbu2HU6Zc;w4+^&36DOO2T8=o!~rMb5m-Zboj#zOrqaD*Z- zDlCdK*w12>I^|^LVr7kTv$7uh7^g~9$yFBB4AoLqwQ94fUUfo6v71Dyho}{5tGW`J zVu$*q`kcB|9g7A})-2X+5^OXP#~vMV5c+%NCP$mUQ9oI9T+Erp=J&%&L8flPTehX{m#al;cQV zwjC|4gK6gJSW%X@Od(9Htf)L@UXnVb+lCCw6qT6wv!5BT)A|11_n+VI{pUxm*=w)8 z_I>TO*Is)+R-O*5tP9+z6SZy9k8wQyY($yz${bY}?Ad!qr9gbdQ=h1!h0v?Y7Q#;WfkaO`1$Yo!ww;wykM?{mC60(&y>e;6GVaB6NCSlfuw#yax=$^ zLPb3}?kJ))0*MmdK>&j_Xw6-z4Sg0|a)A2g1ejIfg<4!&=%BfWs5okG>7{d@R_M4Kcb zo5g`T-A_i&8#AHF$hq5g+6Io|s_N{0t8Up;mI9=hy)gwy+T}~IC0GuHa-5a3L~RdbL5zK5Pz*7xgt)9D#aXp+%VPM zic}1fFEx&g2#nx6#*&=C6kgnu%nt0wOM8=*finFe)K#6=A@^D={YG&n*UApp$~J06 zDe{lCjS*YwEp_0SL?vFY*UA@r?H%-aOgRg;#go>+0Z|rT)FQ?xSw%(E5kvY1jU0D0 zg5z9yH{E$OkmK6!mALZSwT&wSTr1mANIT0P=*wufioEulG&zQqdehxT&vtsf?lyUy z_IAYcb(V|dg`mOXPla)u^F*e*uqB|d*~$$mq+UDz?{I#{(d`LBdycaP6tfn*9WAu} zRKL1ooX5!BK4&)3AQU>^9eOB{^b8)*YrHRuz+Dziw85#P6TvxO+uIggplw)k-4o!- zYh#UUNaQ$oUR%kmnZl?#Tf@xO5ng+U{Fck!POo4BN=mBgY)NKIIe95KF)J0tROj6^ zVN6Z*lu*r0x6s{jy;&EcwVn2H)p_lv;}USP@${s!yluhB4j%1xo(R*Pw4d<=$ZzGd z`sYJ#N@lrF`l>g(+q^OH5G>~)&dOUwR?cifV@%$dDS{wY-Li*yV{#codyrj}k9?&} z7@R9ZID6F2azV;$J**sZq=HLcXY{4Fw6AV;T=a?4C{BLDwI=->V(z?+q66evI zn9N3T`Z6KkN63akfA=L?=r_K^Ekfd6U*b();zD}Fmsl$#x_yZvOIfs$Ys__|&%o-$ zqBfy2)DzDt%QO`iLrjW`y*9MU8xxH*ZKGZ98|C7|Oczb6-jYtyyUCMG#TKS1yfGQ9 zh&AcWpqh{~-qM5`i)sB`uh)4Zl2P2b>FGtg*o=5%(!uSujb|e~4XD)%2H=UT5H{An zXeOeT^j0>eU!$Co5|AUkF&xOJF~^fIQ*OQKOC$Gw0ofQnh#%FP910)in1UQdv|d=z zRdra=MV4}qCeg7dr247&h=ak-6Y=vge|@YWCi)}VFNk#4G!|Lj73z!^+8*X?{tCMYr1L(k=M<41tGPVj)agD z-9Be&MOFR0R&LtrtK~_fI4A%Ut+^u%+I*)r%{~ZOoQAZbh%I%N)SuZX7JPSTFzeJ1LKfomPN| z*<3`w@4<3Qis&Fr*=hPZq~o+UT+EH)?EG&H*UrUas|mA6+Ke3idkU1j*R`@mUgs_i z(OyPNu|A_|7U+!}7&&@AnS2-3-*FfFjVmt%sp0gGWR@C_RL{rQ^X6-`_Ebs*BFkd! ztj=pQSy+WMs=v?^&rWFCNVM%Pa$ya-@>&}6I-pLx6pe+ZY@vEo*#LD+7lNgv!R^JW zG`sB~(8_(NI~1i`V`sV~bYwDd^-1TaCzH4ODB@0I#d>2FV)7xBF_xopjF;D_!+ci3{UPcau;o-F{Sb|}XknvJw$^T?&>ewt=h z;;fmreinw&P5v^&tZY9*KNwYBSK44rwV2rSNRrU&5c*?(B8?e2;|x;W%5wD8d)9r< zRoYZ_!DjI)n?MGk-L>+JtMsC4&~tu_DOPnDilQ(9+c_TQ(wC(i?N)W`oLmo1A$qV$YSvyyh(eL-RJ$BUZ`(kwT2I z(LJtW-bxzi=tNQ)Thw!TCKNa(@x!6yNbH~ln=mo=i`o1AO` zx4R9Jj!BN!H2}Bi_%IBq%X>o)23>B%2OXi;)L2Bai!_&az7tfFS zlHcMgqK+szF5+Mt6Fewb6MbDto`^3ASNH8^ybw<=#k+bm48f2l)9?(kxNrKzWoV7N zy~=AJWDT3C2~1BI#o5A(Xx~&}vhDxy1k=e7M~K}6Lf~$kHPaJaMB7=XX%=XWTo?pc zUN^n8;WT-^-u-vd`82Vvsq>p5-X{9OA7~FqJsP=7m$3NiLfnPz?sHWau#ok!WOY*Cl(!UY zQO0A~dlw1qcpdv4X2sp+38R6rmWXt}bk&LHEe@q)Gg9L1a|2 zB>VwrE&r@sk0z^<2l4HCvL|^lAEqa9DG&8nhXYKBiB@aLlPL-OHU;@xN>0y45K$Gq z+=pCF$p|iQgV=M?q*rSCV5>jr=|0`%4g@N%qk*vXpey2OJqDkYq{_vAFl1wDUf&bp z*!T{LM|qN%5or}w=Y`1b20U3Fad2Pux>T8#IVRT3kxS`v{xuzm z8aQ$0#x{(}bJ+K>spN$~MWY8XbT|n9Ml@j`Mk72KY~?M6LM?HYbqLFFC{!#C5C5fX ztB_exw(s^J)k6kEe25bKg_Qv%$A`p6Y`;%4Fxr(Nvr;r8PlWMY=O>B-hQ4<(}Ez>_(b- z8qKi=(<}&V2|IsvV9Q?_qU~f`{z*XFOfNanCkGdK&5jRqco%p(7G{6@t}Q@D0$3e1wb- z4<(O`NOD|9+&@^GfQB}Da&P5_Ic;h)?VXGD2CYTaki^s-l&hQyHaf}7? z6pp%$ws4^g*dtd#ZZOkV1JLGSBh!k11=2k|iI@_RJ&;O;Q+Nn{68ntB%OvTphZzBA zz%)zb4K%=jVyg4D!6;eVWaQ+xG|E%hcI}~NoZR8(SVP3pLdOMR)3b+I6KN4MBIpDt zIIQvzYa|^l5Za*jF=zB`nIHxs)_(M-pTX&ksYJ26PM}Qs#6$>7G!vG03jJIm^aSIg z!yfPxdeHrhU=`7ItfAVIHf)xT8i8r=c#gZ3cS2+fia1z!$`fGiSu)9FwlL2UEx+dV zy7HPe_7;1Pi4FokEyRy!ggdX*9pdicD!gc#L=&0(VUWtGJMSXcohw^7#vjJ`Gx6g$ z`k*%GfF2lMIvs7k=5S#Z3kKin)(i~uu8c^i#_)9Z8mlGM@eFqwza`eT-2E z!WBZ06e$FOVL~u4LyG;=i3N zbyp^0je)87?IGss8UxdXRKX>LCRClnCciTTTf3|3B2aJ4HZ}|8V8Y0zqfaHwoZAAT zrR7W5YcU&vI`xc!!>o!Z$9OEPY`7#GNQ8i_NAPjOp-2?lu(J*`YPLqtgcJlH;&|-k zLE4imgWV^0u@tM=L_=8-CA)SH61->0RMm^Kx@LZnGN!<4zqGSfjb$$ zD~xy6r#e3rufDU9AH~J3_{rKi6D+A4b+9^ zSL#DpFKmJIAvD0NWaZI$OqlN5>_Q*3Z53&&JIhi26AA9mhn=dw6X>juoJ7}#lFJFh z`01e}tl!`?)+%j%MT%Zvj9)!F%!}z5R*yC~>mp55)00f?H$15)mJwE!vn~|_RfB=@ zTUf9~b}GnRxgz43z;0zS1nB-f^5vr7Rt!F!I)w~U4$Hp@+)OP$35pEE8tOdRbM>dp zALBJm*QX23w+qi7Z7M=G`32$V1!)k4pm(m)4yKozwZbe$ zRDS|>g4rUuszHuF*xmsD!5;6zIeHubKozH!}+ajmKoX!bs+fVN*j`H06sWD;vV{ZNUI za|bv<-<+kGz6!G0;9l8+ZuGvtgTeQ;s~~Dr*EIRo^`DGkoHmP{_q?{?A{xt@=6tmo zjpnQqs;gRnz{f_*hZ0-=w;gO{Sz)C6mQ%H`^dJLlL17XjVY`R<3+g|X^|7#Hmi)8n zRQJ8lUWJ-H7KdeKw(|5N!Kw&@+KoqU|B9RG6I-v6HP;)%j#rqqfKqR>V409l(*~09 zS%#Px=qpqhbf`T5>l0>l1Lb>=m$Kx8Py7BXm0BY=&kG{~HXs~Nu7&B5)qr8m@GUV{ za$DA9<`&6S_m6txG8Ep#Fn4}Pz!&l`HGSzac|!eo)KEX!EReHl!ew$-ot9RECC1rO zEi9z}W@g0+G1?g6o(fem8S7yXF=+a-vP(1rqqcnKE4!Eaf-I3Ej}BvI!hVlaaN-5@BQ- zdGBVuRCI>O^opKr&e@rJ;|)@-uLyq;ZWI6N*54t&=p~|Zl4?lc_gx}7LvG|LYUHrp zxx-Hb2?#$$HX06rV_Z(Mc`qx-8*>Q3hhh#62d^?ckV(~1P~2@e<@WO`Pxa}NGk7Ap z80r|q$(ap1Q^C*d>sS+x7qGc7y< zV}7PDcF7l0_+nr9Vme=}-WN0aV*7ouTwm-}U#!3v+j`!>YR{jcS7QcPEzhuYYdHP< zE<$$paj&vK029#OGw27FZNzG?34OLfR7%d1??;(=(Rrd7J-FmgL&y98J*WcbWqI8! z90!%`63Q|gV_WP!Ynj5rT_@O(owXDQ5n(3JT8u(O7|gRyJ>!6zrj8sQ-G309_g#!_ zC!4>2Bp-A{nC_%%sw?G8Z!|kiFk{n|5_c zN3JCrY8NVZPc|WAhKZWU@iA!%mZP}Kuj6wJF+WiUoIGvtiax!X%}nR8d(~sb&USHy zy@;MKAThaV1rxHd=$Pj!wt7sYwxPlRZ(Rpf`Kh*%^+Gpx zmYc!vK0|KjCI_Xq!Dp6AM&+dhAIe~O2JoSxy<}nDzP_o*Oq)sb z_YQ7551Hs`bhgJ!*Ao=Zu)vX;O!HU~Fc1XUCnXArrQN)^H5Ul}p7$aa9 z!kt(yFb5;nTt@|~J3H_ju|@?ZOXfvX9*y1m_&Ll~G2Poj7LV=kkYFg@zlS+N?i5SM z1-Q@6w7kV{YuLi%btd*lQ*l%=(*>C@s!_9U=}eWTQlm6$}<#Bv%TCi5?|UMDxWY4R== z!_V17WaA6@OEF~i_+fn5IP&KBWd8VA^7;4y{EK7BZ{v4|K6P5CI?Q7HmAvvm25)L2 zryt17>|V~DzU6GDBbu=0cP{Gz6J@N$trQS@3wm;jq)!;0GzZD9BrzO&&dyGIa$_Jv zF_F-hw23U5(4TK?dWh`4=(zm_%HPa(&3m z!aV+P6nVAqUTDR~LVKevvwtQ3oS4N&jwCG;vs3-;?d{ID=zT55ZR%%ev1_KK$gMv` z3=eLN_cQeodeO~Rc$)nDpkizs2BL&cz1p=WdBo1LW*ldl=~>VkoceQETDYII{z1)5 zcQ*zy0$PbR<}+>Jc~qVv7SoGjRuNfjT2363#^f*k1cGqjC>UvIK4!GU{mi?Fz&4E2 z#M}Fr=+yfRt06+z6%N!gZvB0_>=P0)IiEj=hk0(##D}N;eiiY!_oFbEp%J_Sy$l)Rdzp9D+QZ$hN zT|{E1WRPt`;uVAb-8Gu@;~&r*IPf;2O83f0<^;}5YhCQVROYgumW&W%Z z`U^1lvrK|NEAM48YI?uSkNl|v>H91--UBDc@_r|SU==W@)gBQ2POEX>lgEbjC;yl( zG?qqfZ+4l|t|FYzgkU^9KRgnTz+1pml1(yLU~m$^bcfrn1P zcYDhyPPIi_`J$ppRB0k5rD=S7FC#-S!x0w-b#KI-%_$}I%|QXdC9{Nf zkVP{ecg+3{JxS$g9zy?UU?f)JaEXp}I6LVd1iIn+R_A zNmN(typrOt(H#tr@45*R@vV9h9B+5Pip-Z0_-gT3Rzp*I$;?@k9UFX!b|DcyAdbcZ z^{Ct*eo0-0kvKWytp#(((^CUb(>a_X+u+(yrelO0VJyc30*WIhb8}r=BxZ&KZkz-M z@$Cu772(S5mBDW1ru52mM+AI+Ym~Ua-C%oi-8eFGwu`^nM2^gUH}s~U2JJm@heMXn z$>AS-fE<}qkx&PO%i4`fv)qjE$(KcQ9wOkxf?K{!eo+(l2~$mUay2Pe%=g$mra!CU=j72 zA&jwQ1A71VO?RvA0d=O5jjzNg9|m&+T2(~H7DHHl+8QTk3`PvC6|x9ZUc|DT^1_5M zMwXIAWqnH>p!j^M1Aw_tdJ5q6=4`PVdsszpGsJ;;gRMi@(1&>%G3LJ6&R{Ztn|{Xo zK6sb(qDod$T8S&feZ%vPHFg>fP2{JtK|NPMvf<>prk+L;=9~0_ws5viFjwcji-Yr8 zHsVHW>P#~olZ2+1;Js5l-YQ@iBY13(8###EF5!j)YlPUQWU$MRm#2(LW@EeZ(c`-j zumb3=M3xGI$jH*Bc5B`SI}nq`DC9m!8#jzrfVu0Qz0H9MY6$VHF972D`8i} zOeony9&|0tM6-~qgk|U_Xam8$29xAU3VKrgD7*bh{lo9Mjs4K@nSQ&Lmqx)K=1MQ87B*QnOtTZ2I`Z`ms!7katZH} z(H7)fhP#FYtl|)QXaf2AkyX9?e%obGE6fCc*fO>Mq=PW%W50r0tJ;C1D(wDYG#F9; zUi$i%GPh!wD3ffc7|R!pA?GU6M0Mm|g<5ogD3**A6_a^O`tv>Zk!P388?y2>X2eyG z;y*loPRys^eXADNYSgxwZtR9|>6ziu zd-yo`OXOdTc1`U73wf}lFqAp-1UKNL$H{9Y$)msd2-AkLA7iEwLsGDd_-7A((js?h zB({cNdEK=j?B>F&wgeiEO#XSPJwbA8G2xyQ-IV7}kU6%vq{jOcE6{5j&`#Xhu^ZqY z{z0C%&FwW}EK@AFLAEDS?FkZOF9^TU&|S=y1~T2A9DedXWkCaZ#$Kam%DNOP= z;T(xKjt#a^R^CL9!t^M?y=4ao5$#5u~X&66@Iopzf;o^^%<8&E})c@8pB`2X}I4Hlc zVj6>X3fc%VD%_SuT0(wZ?y)X z|D|OL(GFsj^yA0X5!Wk;L+ilgQ!B5s_WRTd4A2nRoqL_O!b1lVx&_-^vu6Az}| zfT-aB4@a8ihbN8D&s)&Pt}~69zOlb>*?3VLJM7>hPQg>O>7opt;r$G7|TrUux=*>^|bu7iKlvLY-L;~keR`J+k z`WRFjX0gPjZ(}#EUfZW&hCmc~`tz>QWax@NRmOPcH@JP)rDR3SoQ1NT11lkXxM&cbg%!(hu^s4e4JKV!UV}A(5_EKF3YJ+q zN~?Z`!cU?Fk-T9ZH0U*UK5fTp!%F<$u--8b9KzNP<{FHd8rYuy8}6~i84e@k_LU3v z$o~yH8W#XXgx5}Z&-_96R-eI^-Ajic=@;P5hR!bhfDLnLhc#f9Yh|-m&u;WqPI2EB z_61F1tu3$XrYFKgA6oeaTgcEC0p7e8r#E^8n8Tm}xdNGR$7LK7%Z+v^tpC z=(mgn!>F_CV2}m&5*jLG!ftVGK9-u--V9sLrw4Q0s{ht|)qgI7z5ToKE=m!y3F`Bf z{SA5uq(%-8F*r|+lgwJ(w@SSh4{&V$k<(HTJ0)=~u zzBCKjSF73XZ8LdQ^l!gnkMwq;zh&p3yQ}lgG3us&k?^PEqkqy0>Hr(thr$_5MJq1v z*F4=rG<;VYONk8q&~_S+D3D__2m?L@N!t*y<~fvS4XDm*VP%axLl!$H#SQGr5wT`G zhR~NQGL3xW9AtbB4>n3@@H_rV!HcA8v#{{>-cC4L`W0Jw6tflwLkKS`{=+}K-jjKq zyvHC^z339ruSw=_4I!m#`m;_~u1RIVD{Hb?aAD0rJU|syr8qLM*f77{XE*L`3``In zuVLKo1iHhj*<|8mU7ag}iQ9=v-y8w`cx@-_J zwp18V1#F?5uJs#H>^Vm~K5-y_uL=&E2dYo?=?jZ3ed{hX28h_D6G zjW+z$08Uhsch(N(!FgeAItxUu#DOMo!gw8vKmo`EqyrKG(SR_3_&GA#RS?ZM-CkC* z1z-cLeU7~Bnit->fhFAj7WM7eKqA)ljcr3BVk}`Jho5x7gbl>7E`@L1KxVDWV1a9$ zWFX@}?KKGY0%`$=0Cj*S2z5lsVRY$Vz4qD;r@-8N$yQ0U82 z_B0CTD$$s30)c1l{TW=ey31_^v92AS& zXOdih!=HYjJcTy+*`^ zZk{GjLtVm7V8^*`R;r#H|C#Q3(iD!eFIVA^fwDP4V$fk`k z{GRn>@5Tr8%-n$PE~YP43u^c#-2wz_zNEUW1!e^99^{`8k=Qm%b!~1@w6#oItCE&_t_*a0*>1 zgwv>92xrj8gm5-36T*44R0u6}st_)ug+f?P^Muew^+LFeDur+b9gfh!8gZZ!Yk{UP z@_{uBjTX?g4DBJH>lxbl3}e{H(BA};fbM1J z+X8yj!SFo-UdPbw0(yd>8wKRDzv$Y#TE()hZs6XK+iGMET9(|`hb8o zGt?-cEeur(C3z{h3DE`` z{gkynibNY|@l%!xl%)!-a-l$h*MAL}|F?99!N(e@4J`1pX8I`6+Q12ZN`j9PuMIT$ zDIq>eqBd}fpK=S1LNqH?8#v8RfpN_!>Ds^fWxl4t&yhP4z9m@B7 znTWTwlJ{OohdpGgVwvvv$i}=00$-;j$^Y<>0iLXoHrK^1Xj)+%B?tg0E zNfGZ(CS%@wgMZ0RetmQ9h^WW?9=Pv!8oBTOZsa0sjNC%NA;1YhPtbL*;G;eEM^?Qx z*5T|0yZH;eWo@+grgt~6KEeeAPmB8-&v?O2)iH_rkbM4G=4EQc>o&n-f}OR8iMAq7 ziRX*|r=!Ck%*8wvH{!SnkG5Et9om8%>}Z3}ZAu+5IzrgpTq51X$f+D%dcP43?ys-y zL~E+*aIpXfx8Oio;OCpz%|~xM2M5ZRZuUsBS|=FII4KI=VRuBtu?bSK^Lvr)BrcuN zS>Fqn;9EG@h-`_Zp*AVJ71{l+n48a$pKFu*n^Cs#tX#)(__oqREqG^4i*b@*dw9}| zieV1qbE#>WkX^WI2m#8(!X1PmWEWTN&Z6o>_84@`MIw9z_P+$_*A*Myy$n2%Wrj^Y$cO@g+&wp3FV7%S*Qj#W4Y8qx@eU{Plje$Gg~~e4FlnIoAeY#EEm~ zfZlOkoW70PceOO&8Bl$`wnxF}pMUlQbrdIjWJ~iuBg5a$b*#f4=IcItEAuBQjiQUY zlV}uB&v*ZQs*5;=W1X|v_@~FX-4;1s9oWN|VkwTOVlQXRwDe7^v-jC0FY{Q8Fw+Td zCW3;sV@oqrBT^x5-Pf{`H%Zt#DMKSzCO?0G4;R>Pzww{!SKlC0!G03tF7_inTws6m zFYGVBL0*0*omWJW6Yr?_@ZF^2osnrnqcBv$3sT|rOK};SO)M2O_>@Tgu_kZ|Dco=1 z_bedQ`{%*{y|G_~7law_s({aWSHswgZ zU5h*d_diZuFRxuGJf>+MI3QUK3)vm)N({nBWu91?@^se>(ZN88W>C@srbm_(x!EIW z_YJ6tKVK~S)Pl;XlXz_s|2rmo(y&i$k>hM3Gb{+c zR!Vjt`0-TLpwIdTVzYME5_0C4tWI^W=B)&})?tnt%Da%=`c(d2rnL zy=-C30u#PH5pME?JH7kYd5w83kQF?Q*31mCE3G~0FIAXO_`_Jd?|UO+>IZ$xlCbI}?~ zoy;D9zgD=;oP*|oCKZ~&{(XmYPOmYdS_`sU@c}Eg>Q=;>!)P;}Zo9>xkOwo8#gtY2 zct$)f7}sTtiGPdj=nc-57diWtllfv1d}+g2Zt`PBYJ&G8B;vNC8Tx8Yb(m17OC(%} zrVJ6j8Y>+#sMmB@gbirP4G2L=1F^q%%%Ov%>|*W(c*UV~+E`~)tGbm=ANuXeuGwVD zW+xi?rwi^VdZasLzdVBA7|n=6jt?5$D@Zkk%~87-ZSD)3KGxI*ZdmFzyi%&Pb4sK2vPyYB;{ zt8T5reF^Sbk8cMZPYK<5cm%U56xWXN^q&F^?^qKsTzDmmk4uI7?l7DZQSq0g>VTBr zaEKf_Fn7Ss4^bsvy~1q`3n;|egMI{GV~LwN=VOL-n&_3D#BgwA#?`0TI0^gTI-qQ@ zDlu3^c$Wjw(|%_wv+9fJqow5egSpAO4|~17Z?cZTVT#8&#dL&#*e23hPm$XPQWSjeD1Q4a^2DLR8FS!IEeU8aQ44Y~zq7W}_7L7) zvgen1{fp1&@GZH~n8Gs&sL$^7n_gdU#lVE?d1D0`VRh7C)p{}V{|SSN(; zD+LKUp|IE^!QhGVG0cDR#LQ#MjlG!d`fS!{rpvI9*nLZqv$J{Wr})-_#p!FPy6Y=h zcG*P#27>S@a0<#W(R@Tp@XYuJx~eNdppRHaw!AQqAGeIW|3VVK{WEg@g#lqd)?j+@ zWhy@Mnof*IlK6vH$*dy@{PwHlsUxz~87!waW*Gw99zqu5=<+AJs;7&dAfFwP^G-_m z_eZ4od2Qg8U%2)RMw0dbuTvdp52dY3$%F3?oKSPc>lIwE*nuL}Ek-r7 zk@fbues#Cc6}*6d@npbNd-8tWP=5P0a$qLBQ;@j|#$keq>`g zyG?}D;5+R<|B>1C zy}xtZ0l+E1SwI`06A*ce7D1l5g4m9w$S4I5wU)ve!wR8X*(l z=5J8|!fgm65gHNpX7qD-`@!f4`!G7fc!W*g7`eU(D-k9jj6~QEVa-{%+7X%%CNnz1 z6odu;GIFU1o6i`z{v1?x+gCw5w z<1Kxm^X6RX7wj2_lyO39_^Jm<0!} zSUk6E$r7^TXA_z6vxKBy&xrVm%Ud*e@w_tA;suLrWfl9ETn`nIZNDU@EM`sgHDb(S zJ37PthK}0up;}mdWdARp1-~OMT3A+AP8@Ch$lkWJUeB^TMYegqY~*5Fa-U*R{zCf_ z3(GdzK7W2$#S)I(X&WB?6<<&`r+jkRsD%p`&m{wIT-ZPR*8&l#ycyZEZV*NYe|3ZQ z_xyD!zklP+ZTw){M%+dIpMli#v-ZDv`xCKa-(}o20(Jvx06PHN08p9SdcYdM3P3qv z9$*H*1egFQ0OSIU03ARLPyi%=Oh5)89gqr$2ZRB{fR68c{5QWdTDew4S^(z&O@I@C zLx6VywSe7#jes=(eC*AM{HuBe;x@oSz&yYVz!bm)fDVueNCbocZeBtia(44sZxi3)lg$0cKw`k`1>?iMq2qPVzm{ zf5b3cMfvzBqd9aGr}(!KV_i?`yv9J5E+ZdwKJa{`D09RF+(!PNz~TQ>^m^VPN={P4 zq+;D40q7bAHX8GAp*rQgS49g%aRI_6l!2`kK1;3~A=_`)ZXJ$ipmS;vWgGh;glw;Z znykc#VmWb@cx2BCt%R$K!$nCYJdPZ$J1P+fe~gbYl;Ib57?mu>{<`R_J+rrdfNI3c zfN!niMbYB5-PpEnY;QMqs2e-gjh*Yl928jBUsok;{@Xzh87<-{deY)c;M#D(k^y>G zJU(Bq4*&XAcZTNYPykxfRY1Xp zgQLZwiG>HZO%Yi{j@)q*$C^gvvS8XYKk|Pu?htDJzgXd4>MQnj{r}AdFf%f8YoJ*) z&Y@I{SC|!x6)O~*6>lj{DlREFWuS7fQl~6YE>%`3iSjMwVdYunRb{7Atm>soPz_Oy zP#IMdRYj`TR7X_TR7F`cvm9A3Wz}W%RS#BAR6Epb)f?2i)yLIet9xojXbc*s=76R{ z6Oo;kEzh2ky)66X>`d)=twp;``@2K?n(htV0o@0>%epE08T!TgC-n~fCjEAOjsB?q zr2eA*s$OIWGejHG4MPoDgUztYu-ouY!$*eGhTDdhb9Uz($l+2s;h&zBNYs)-$u!Aw zX_{=3tXMW*Rwa+i`Z4Q|EW75K#+5xu_oQx>zEdA$h%*d8S3Zqkq*tXSvRSe+*`u<@WlzXf$*N@QWkmL(Y^Q9m ztX6hNRwp|x`^E7btfxFeo+3||50y*g`SL>f6!}APo7^eiDBmI9C*Lps zQ2sgl#}(wg9f}l%LSazMQ*2f2!32D*;FZ0VpR2g6lq^Hm&a4wzSF&zo4OVBV=cp?% zw9lyjrrxW5L;Z(3Li3B}jwUEOH~WF?Y1uQgw`U*BKA-(_c7LrwyIH$XXVb0My{7Bb z4b)5ZI>Q1(x#3B}dc!Wm+lGUN^M>yYcMZLAqH`QcIg*?yIa_jQ4rj!vi!Failw_Rb zLCF-!!;(iN3nh=Rj&?4s-knMZa*c2^c850gjB6Xb*BnR1PMjC`EDNN$&}lh?}IAw{L)t|AG7H%s}5vQ~LOd0P3U@>}I)$21PjOPaSe2Q@>ob=ecMpU-{+ z!}=8lwmJKHcA9piHdkxM@P4Na(nS>34i-m|jEqNYenI?Tydfp+uBV8nWQdTYdyX=VULk!;c z7_?Xj&p`PIxm-S4zCgZNzDs^k9-x@4cu{d(@ixY!SoOH-Y1R9xdi780XEe>(SF?Xd zgI>{nt(%5fjn6rqb0LSDjV%N9PL|A)D5M7IchW@JL`+nZT%(w!C{t`uoKaj*INB6h z%2CRN$~DSYlqZ$9lp!jWszfzQRf8e?Om$V&H)~YZL+U(Dsb;lio$hmekDLiP%W__1 zJu2sfQ=(IPUH&WP@~*s7&MSfxJr(te7R4`$DCHddQ$jnl_GhK3XQ>yd-&P-0pHzRN z{z2WT&em8ouWH`XG&(eGnzZck*#ou1v}<&mbT8^I=~{K^`VoRASfhU)dSG_W<2lQ7 z964)p*5%yF>B!+KaaKkB@e-qCy=0qYpX3Y4MPcc-OT_+VJ6pOyYM1Vlz9+4h-jxQ( zB4iU}Q)P27(%WUbWuMB<$u2=a4=Nm4==XcdGZ3>o%23qg zmUA-^rW0U+F`6amsO8wu8{mS7g6BP8B>?`O(Pi>-hw04SiKGg0y?GEjm+B)r5T1Sia zmNrC}pqqr&Mdhf`kau&==WsP}C!hhp%C5X#g>TlJpYOy9=^MvN_nnRj18d3I$Y)$qf zsBl&GvFxVoFs&3iZoYP#wgwCFE8RD`?{%U2Vfs1xNA>sgX$G@lz9A!LL{4_jm>hG? z^Eo?nYOxT}nmQO)2p)kd`&zU*q0wr}vzKN+mAyK9bN1P6l+ffWXrZ)RYLhO*l&r;+v`DG+CPqZ84byJ@s#KKj z0B18N{PT7U7CT(wMf^XjJ=r3GtxF`2PRV?Q_H5w zmO-Z;g0QyB#PU?RQ9fH3sWUGD+VKS1+Hg_Q zT9eSxIW+2U07dCLZ$fJ|iFSrgL7N#J0(E9WQKx{`%cxDD)r3@q0wt#{&HMT8OL7u| zc4nOS|NNh4@;oPd@3q(Ew|@8aTbKQTJGU9XF(!rI;Gi+jaMfSL{{IL5S;mYn_=oXk zU-0`y&sfX8UsM&X*-&`@`j35d{i=Hl*Q~nlzK?AzTzz-p`c3y0*4$Tk^X)4N@BP?a zcVC~Cm41y|^~q)bvgd2FT3_{kPrv``XP@Bt>GwZ*c8-1h#@TOhJ^$e+&;F%-{l?jE z+Sjk2{agEe+u0VbUz^qT>JhFvA8tAORr~s(S9Sf`tWx`$^WkwV$?=!s_N^t+WeI}A-%-%^pqt}=Fn>g8L z7CN9iSz0VJ;D9=2k?S|!y?G;V?)fx@1d#gf_QNYsXiW9>>+f2%ag{MI#7U)2vz_ZB zT$BDH#zd}nvYCo&c@d%EOs-dq1b}><*Wd4CwEIY3?R=q2@sfq{vZIX%&5;uC$5M1h&uUuH~Ir+?)O7a#nm zwu^q!-S^pCXU}l*%_i^ZTdlSy17@c&@B5(;Iry+S>im39t3r`P#FPThE#$qKGK(l%P90@!XIK17tLpymsMCYjZ>%Hr#@GLBx~Y2BGW&{5 zLtf`sx&k5Dxy|Z&$ZezRfUBwF1vgRq7IDHJ97ELrq*YGPWtOmD*S&xC+q5x_93Jqzd(82MWD;KVQvtZ%na4fu@bnz=Z z7oC1|K=}28>ty_Pg5M%z(OzVkuGi>#de=bF=^X=Wc&;uq%XGbqtMK(0_?iH&f&Zv* zEL?-jqr$uJ@0Asfg=^uw8r*xhehgfz9o4BeRsUn)yr_Xb@+)j1twm{t^NS*+@hiNa zwELAdWqwg1Y5Yio7AEsQi?hLJt)mwow2&nqGUZ3MQjoD!WG$_&$V3u7ev>#0ub=Z< z?dScz_SgLW_7nV41_yU_`*YiS{CPa*x1aP+YF~l;eh~Tn5c2z{$nQsx-#WY_|UJWM!>Zw-*%pf7G(^#i4hP z4GzxZ-O+ETLv(QIYx9oifNe?6dm@<*!2FntV~-B%lgAJ0(_`cp z4H^c9Y4FQYG#FfDj@q^xn|)@@>?3E!nM5*u^zQVyNW!-1Yi>AF;LGlB#O6@WmOaTx zw|`>05B`8xh}n5o{n#@;a&Zw=WN@px)!Ur*?=sinJUZt#s7)nq6BMJFAM~OO~xvCAv`JP zimCb8S=SU@HGN9n#LVJ%jQSH0Nq>>r=FCXZfviZ8$x?em&sO1x=Q7K&)6VnX0z}ISp9pO{Uf?i||ug5*YSvptGTJr3v((LIxJUdCQLCRC6Y>r+L};z(x7Q z$o4G*gZsxSmt0RpR^-(da2hBxM|=H&BjqM=BvKn&w)oBW?+~p<|Liu3+&1ze+soe= z+$a3bJU>{awq+j=o^xoC+>Up@6287?im6(3m6;fgu3oq%5I9n83WoaT4;^d`U30Lt z!1DK7SD30LS)D&C<`=Po{oUi5r#Id^bm) z3Go}s7a(77jC}Bah~FjU|Jd!I4@3Ow1*hOn0iM+V@8P)|9G?FDkK%jH+FhuYCE4{5*HslV2=aOqo>c|_nq~pw4cZKV}zaf75-l;_WmV& zz?1b2$9xS9$9nJu3#OZ$0n?CEV9pxKN)IsvfbuM+CD+MJ@}n3(WY;VHZSBx63#gB zpOf0Xe7_D1wJ~y<(5a*%Q}4yMFYKhap3LfX}q znIgYjxzx??dQ!b-SA56e zbK)Hye@m8L1I8(qd46(kZ^GnhU*p{FB4o{`B_yK4NyBBi~N=VvC&)J#$0x zD;P69l4jzqzVstOYf_VBY(};@y2H0)_ZfUY@rLL<93y$wjgh!;Aa55R{Em$`(a%?u z8XrW?;i1z-jm2@vUN`>TWZH~pnw_fOrgK?pynAxtzM_!aV4;uy+M=I|?4X}V)@GZIIhJW6Zk%Vfv>dySvcWAam0NwmrD{hw zRl9TT_VU7l<=@;kedKrA%x}M8sLlNLFVm*ow}Yv1jZf=+=8Sxihe(gU!+~RMPNbt7 z`UsYo4NJZYOK#J7bWZQrZZ1|^YCB{FGB%Wlj;OzT$LMc1ZJ!6{dse5#!RwjsUCq;0 z)7ClC8jqzBhjH-oIQ&y>%)ATWbR2c54Yegc*-AStj9oXRnWOc-N&PeML*?gR;GNpL9r}0sC&N$Awhrsk%k$I0 z)P~Bz2__+0 z_N2vo9yW3D<16Ui9`reV3k$YSf%_GHGf($t!QZMGo&8sJb`}Nr+c536R&{3{W%SlMHV`|E+@yvPcV{LJkfMWKNZrx zEP1IYt!7ecMZe`gx^)* zw>~ZMss(QCJo%Bt4BA4*D?QsN8hSWPu79YV6|~{fO!5){ALiincnUl(y{mdwz;g{| zV8O%aHsw*B=;-a;*w)uGBpZ&dJ-Z@_PGjhM;pmstaS~aSYzVHO@%*A>pXUhAFYx^D zJRe9iPpK{O&;egw|5tKWyxx#$s+vhxTC*>||1U}F&JYfqw0hF+%BgxiLYj?_E3tEy zLs##4K<$i5yFgmCn^w61*c77z@lak&cZIMRvr-=QRbh83!;nILX)`)H$;WUv66=JeW)R zYViJX@@!mf;^LWR@UWWxs|en$EH=yBOB z(ecmVDZ3BFb}TpEQs~GlI9*9!ZFfI^F;Cx2+A}-eiw?1R#>G4DNMWAIv9q#s(ucX|L%V!d zB++@t&54)L3(}1zkT2;+)zgBV>b*MBA-*gq^VLRxTl{L*TNa3Kqpm2nTIE!SS2mcM zccWc5y0s?9(XDowgMs+Nl<9rk#0wrWZQb_OYAbli$Is6%gj^zwMi%Og4W_($?B`usD#oDEA^FkTHT;?Yho?O-te5j!on{U5!w zRq63N1T!$4KR&MQ?5!pdu%_JD`%%-Oe)N9CwAtTG%jK8S#_6JK@w;bQNG4P+d`?4L=kW)k6quY%y@y_-%0Ja&|kYQWV|yi3AXq0 zOx>dv@9W?`~rO1vn4AodfEB8 zZor@T5inQhMC$Nm<`%?EZ3J1g;LU5$k=^(at#_6sRIVBs^vdnPm)Jo}U2=II+Aehc z8tY(Ie2XPtFR%SmQ}H1s8)f6=6V_@>D?j_^Zw~JI!fF$5@kQ!7;eE9gq^)Z1B?GcW zL;SyL0(8i0|2N7!ozB?c`p9PbWS8mo2f}Xz@rjs^h*H;C_>T65-kswebLzih*c&%BqmW0Wn}YHAz7joQ*XJ=5N4%gVOL z4__spSLaUZ-OtZ{Uz8I+;o_tOoZu_XjlIvQ&x}>y!JN2!_3hLrm^%+KPSx|7f*+FlLF|>QLx2b*t%R=$XCl_OlG%rtvz^044?W%f*p z7kt_&^CtE$xlDLenVv=X;Thtc-1dG)KR<6ZP9BmA^`oCU1{g!wxG9?wFYr5d?`6`J zW?Y`vnB6>OXyj8r4^azLrz5VJ*3Gi=kDT&PvC}Ku^j`8Vv!)#BL;ok|?;?GsHJSXAM&z$x3?IB$wj-R{ z)Kmt4ElvGw)5g@#H8r3c7c)-m75|h?9fDnOKkDMiUo z|2MBY^U6O(-;Xz1lQKLW`|TvNvyb0+?2po!*E))ylZc>0t696kvmNvFY=MDCkfEL4 z*vrk|(e00PB>!p`76fwZCu6AU%txh%r(aLS#XGU2CgtGTK#73 zyP#D-ws6&r@jJ=e%x^DajZW-w3-{mR*8mOdJU#D>|Bkwk!NW)SZRg#s)T^uNt03Px z#{Hk=`QP}ppvw*L2!3ze)6Mf<-l_jyKeyS>!(&V@y_@tpes|D@hXa)xjuSfTf(ohg z({7(a%N$KXMYU5{1!N?*7;seBjq`mf#cQDCa3{O!;b-+N~+zI+a_ z5PJ?Qt{|TX8QF_{ir)o)iRSl{Pjm@DXB^ad3n@#!O5y7tsaG%u>4WmBUFGcs&elbd zk_z&N$geyXA3QW7W-b~M|JkYQdvztx0ypo~Ts`{zfHuAWEFSH=^namDwGSWKpVOWR zZ5}BqiD%%0Ra{)z6}8O%r5A%;ZIX9jtG5FE)#w)GvEO66y!Qv*iL$c1g(z;QPM#pMHa1n^)E5Z3RwR%uSO%OHij> zcXAvgY22Pi8!_JlCBkKtSP@qO@f3w7JidrIPac~&0v zBa8Nx9v~la(7naHPiLH;UToUJ{L+Ia0nrlT-+B_2%PxW9#T@gZV@d^z&bzPPDU zyg_~OkcEF>rLVJCe{|h@BOT#`*EMCd$0{o<#u8S>x-!NR6|>1RD^N>3v?{{{j}((9 zR2-=bSiz?BRoL?^?o%6%T~{1m&%(#q_QK%>Q3k&gPZ;{M06&lSN)N6-pp==ouu zm+Zms1c&tH*ji|)zRLHgxR%(e8JMqUt@}`aSn8RlXV0$`|0!0Itmh=xbHQyVa`Ci_ z%Ro^HetUD5_##>7B$u;s3oodAF~17RBir-D!#{%#x_%gX3y$PEgwN!9M4M3Yt#*tF zz4Fv5`WU9XU{{=}5c*Z{ES{_ODONYE13fulea;V#4QnFAJK5=8?7n=nVk_{R zU6%QoKHjUWUCw>q{V=rk^6VLzr;jlU3n=2f_@(KM!RPuYQ(*<43s@P?1-}#7UJ7h6 z%X~OsZQUKikL~k^+J)<2s$%@lwtD^5Su%pEf;tL-pIw2Zx_!dzJU%gU>i~7vRMnx7*jT$L$WyjN1oja(kqs z?{UeZSzjR^G&{!JWS7n7-yWsE?6Rwz{(~!GUBn<52hy*)He|cN3a%?6ZWFjE(ji$+ z?+i69MxINN=T*q1C(oAdKTUs2Of7s;RgsYFjU2;`z>oZbJ=Z(O+~a>Gb7Sb!4h)jB$gHgO)auh)yuqKsTEk)0vw{hEgx>| z^6!!BTxhj}@%DcAJ{4J7v*F(E!qgp8P2|g#BPYxkvVPa2e=mMpVAb{UKCbf`uaD;z znRtL-ftBB{@xwQfjg7>s_F-4MK2?S-X8Zy!8?d{ITVBmLLU1_qw8`_sW3&Ta>~;?R z@!Aoe1&VKR_z|CTh##Tzteeb=-=q>F8ILcqv5vNmY3+H`E`vG;TTbe8Y7O!}~mJ_LbZ_AN*7e}2P#m9(w) zZ=c?W>Swdd~SB91TNXSaJ`;<<5>Qk2s%Q|RVSquABMhVAU+e zHi>sOWX0^cyO7(bCG@Gnn()YCYy2bdWPjJ|gH?yXm)F0gcA2K8$+Xi1&hK||HW7Gp zu6pXW<={?mRe(FeU%@^7wQ(nVP%#QGHQ;3?cv*`rUTI~mtG1eVi`GNy5WMy1R}asR zUSRD(?_V$vNL=XrRgr{Z=>6#MbZhF7R^(1LG*}neKF^xk)OTZ9Nd@z;=axtZnA$t7 z*zUl%NXOw!;+0l#?ZJSl8fdcGRIb?h?XpGTbHqVX2NuZo-HSgFV7yX1 z-b|5hI7t5Jv9g4Ct9r}$_=FYMTKPw$wMu^Ri{&FrE{o6$-Hfs17pq;(tDME|6xwa& zwx`jK(oc}bPaD|o_LpZw5-&|6o=2YOpkx31FYo*XKd1d+`|Q{m#}C@W7^@-0?CVQ0Q?PqBjx6!k6z;Jzd%%65 z5T8@;t)OjZ0!o_#uCSFsXn?++8@Ph`qJzavCNGkhM4MVS&o~{QJJJ!e3g-t&pKIx7 z_QkMkxxt|^o6VczOT|jXlcW6_&ps6zXHBBl&m7t@th4fLtCyas^p@fDPILw~U|Ya? z5AoMPe*_*F%r$wQ9q`JbYc)pQ=hb%)<3PybjEj@(A@kkax@3n_&20In2H#;JeT8qH z5A3mqd=dM(CslFga3}dazff}rngf)Ke<8)}xAR$PaW4;J$?b*U16bxMPB)#pt%>+2 z++V3F*=0Tx-_3<@WIMLcqc$Yt|B&m}y~puM zphsA6c(Kz|ekM4rWSnqXU zZ*x#VY?0qou_mYqAgU3gc$`bd<(U?(u zth~3`W%9zmasBCJzg6v@WgOa+RDOoj_V#7eq4FQkF*~E!7u8pReXVECPBsX>*L+F% zP0Ie7dOrHr;J2caA{`oc7NkyF-w0eP*WDR9QbJu?tFsne5@ft_oi&cRp!e^{%8zW1 z)vX&CA8a_*Ry6+6D8BBDmpe;tw*p6M+`39r&3gR&s+B2@AFce#8$KT?X=e{okAIxK zHUNZaoS8TDzNfR~5#U|ri?scYcN5HnrjLT>ZvTY#eDZm)2735B;lYFTzT)i z*f(~9GG053*LT9_d4i{mKK|}v+3{{`T$9bOBi0?b>nTV0F~fA%y2AwPDX#**%+ZE> z?X^s1?|Z<_G;k9HH_-<8mG=IN`DkRacnWa&ETkDoT;6UPa_;h?Xi1ayJ!}~ zPZJ*TcaPM7??%#lJX51JjitewC3;6DznAGol_&weP@!}?8p z8~($D!iHn*Tn9J2>xg7X^OWDncH{(~H0;o0Qu|_Hu9*>On+6Sq3;eh+v4tbyi;J5A z^;$<~_RfLc`OrI0^nPH|z=w#*Z=Esi(IV!ut_5dGV^LB2Bcv@Rt<)dvFT?k0mHp~T zi$_>%BfZQTpJOM1&u6w&_FV16LJTx=XfO#HObKUGUvbRgWH9Bv+l^XvXb4iM8l_P%@7Cgtilg7{Z>$dM!kxCkKPzRBNC8;JZKR*CJ z96sE;a0&hezFWNBn%u8wQJe{+l8Dk+h!Fg5D<3aP00PCl0;=xszQl6n-LGj&1Y>n(A3g)PQfc>5z$k zh}Z{fcD8@>kmf=r;X0#Vv1cE;+U$I0OThr}RUY_E%0O%7pDjG#yJGXVW?ZrGTQ~A6 zzhYq^`^ucw$}1Lj9SS!s#@5M?El8M>(tK092H0Dvv%t#hF9QGS>pj3Ey>8p5p?$l_ z^Q$Ar`QqApwi3cZP6T7g5@;eN+aQ_S|u60tvo7(vk+EZJX_#@uB6&r7k zUc}c5A4reXuA%N<;m?J@O8|TLkYbVCOI}B>qs87;cJouegyRF~O4jl2oy@QA&@CMY z@a<<(-y}cl=N{NFPywAq|Lw%YIv6TH-OxRk=tVAHc+>AAOo5g7G*s8b#PKsmKJy7js?T(8WtW`02pGm@h9rxJQ z4)WlWT)LJSu`q87FRLzx7l;0$!Rv))Cq8Zt0@DFKD-TRI9inHw&sX~h{mg@oU;NpI zf%$&3zX%#O{0ti6Puuew;88SNd@!|R(_c(m0Z)c^lBX4%rpu;Mb|$pKj#fr{n6KiR zV$(d|p5GbGTjKd2Bdh0-$$i9`n^^l?XL_0v-hlc z*DPrDZ`1>SWGgoiYlnt;ZQDF}Esu65(Prxd8wV!S z?sGFHKe`y(y47#ySHN=x@LVB0w*`OYKf!Z*11`_WZxqj|9rW6Ec&hR+ZG~OF>t%hf zC-?u({RqArV?BsBE?^z%>>hak2X2`clIocv-uDgnAHR9sW5|cgvy3$_mF1bpat-~j z!T+nK|LCE-Z@G0JNUHk|%l6f`OD>%F)whTC7w9|%4?g_t)6o;R+rC;L;>_=`CbG6h zbbTr|jkOTRq9-PF82q;%Gxvr-o;DZ8u~)m{*r4X(F5IfQ)csNNdwHvQ_dVWu`Odrf zqUbcrX}^!XZacS;m^-vzAe}-Pt?S%KJ@QLR$miLcYHZOzAP;J*?XCCkXrC0>etd0c zp!b2AfyG7B9-RpuqCxCq(446N*L955_M;nLVBUYjYAa4WW%<5~gY!gF;bUa{#M#%; zOTPkwhq;*a4C&1DM;*M4v^(#|_La@B$GCen?vid7%&$?t>$O`G)oY^zrR33l4fpNX zpu&SD9)Y(u0FPk#;ah|AI(h$h;N`{F#@X{8qwP>D`+x$j?3zybO-|gxPIu)tIeqlJ zJL4kuMcI4t!&T5{A-v&@EB<>psoYuQxMh)v z7tz1RiBCNYJ!kMf;BTlL;N1aq5o<}p_F8WA&j+q?{EVkbm#@)6nJ{H_{Q>$$*moyuqo zR!xkv8hcKVb!Yt~XYJ1`cK$r{IdhF8#$#GvK0N;N#@*jbf^Cfpo9N}09U6q&94JYA&zIU3s8Swq!)uygzDf}tfg0|2>>u>7}d57r$+-}T7^_h#~ zhW3^Z*@YxOM7(<*`o9MKiLHkR{3-2!cCE8tBlqWJ3Hg2S=HBFZ^h?Bdb-eMAb z!+Iw8&xr$G>K6!xAaTTguF;FPCeZOYnp>SS>$%-4E)MP=tFBYbUmE^2o^kgq9FxtE z9Ld(H&J61O4cAucTuYtzP-h^3uV|$&4PM9m2=lS6^u2+0q+3+qXt@>diwD&IbADfV z1+u$1i5I$|BW>o4mfdE2t1jk!#0zIwXU6(4b4l+*#7g zdJG?He#A?{=QjEp@eh1U^?vrP`s2aQ(2*t|>l`lDGoEr{?@{|cuulHvIb!H*8OsK( z^dqH=C+u-MWt$H%j>MK$FrJR0A8JSwkA|$grBPxR zShs}^smv#yq31Q&yeKgb#kTA|xUr`Yu_1h>xkGJjx@tsQl2MON#Ww#1(+g6qTKOSf zdccig$mU09m=2G29-P-h2gOCyuL$M1+WIuvUWk8-7{bB%L zzmR$&+$YmdvILC#BU)R)d~7ec33Gh~Ue??YKKR^zu6_Ot=4k@!8h~x7DD}kP3Dpy&9>G^| zjoDma())oe5n)f5;Pmix0(e&>;pY@`365EdZ}Q5RtE}dI1^wy8CLgBHo*$>ZUEcl? z#eWq0$^IQNpJ#K9YHyCqr`Udf`+z^ywx3!j?eXdJPCGfJv;$H1&7-Xz^sCmFO!t|c zKlatXI=I7n<}2xu*9VVVn|oOoJ6LZ@de@kezQ|un&U0*1aF{CA7A+Z!o28 zaGmKGd^T-haM8GdLBCZuxP#{vzLLS)$+N**0}YajFhO&R_J;P-@Kx2QAe;28ENxt9x@EB;@|u2;I5Hz9mfyc2rT$?IT}-Y z^(#g(hqP8-UQ@xDf`uA~k41;Sqt4$)haaFOX$#n@BYW31Na~J`)P6R zzfOy~OK9;xXtQW>f${m&r{w92f!TZI^@d+^60aocCP@&Yfex`PaV( zPWb?uYg?IVYHyyh>;c8qR?c3+I3xQu2emCec-;{V(lHe=W_K1ZzxYdul>PlD~=7r zCnJV=U3nrv+*NBKGl>yLiP2_WV-i`GsbpTgJ+RR6RSJI+d7uTD0?dsC7MhN7^tR?7 z!hu867Ju3|y$4$nI%u*pt~9yzFPqxW0Z%kPD}y|hCLZ>gI)1>WK6ihs|`0-!7aRda0~CtT)fA)AB}hEt=8Xz_n~o7QoKy#foAr; z$j&c-w{_+TwyN@8WJ!BCoHHRNwQu0O%~l`0&f0_s7n|V~`e&C;lU6heL&H}DXVEWswkp$_?X+;;7e!;5kq>irxXD09{z2>LU z&258|7Yya;vFv%_%mDphPgyC?853i>Tj}Gy=qine1sXxFf$ZxJ&T`XS;|^r||FDNf^xOdrHo=4JC+aB2?#9NmuXvHO7s}ATSD+nw zp`(Ga;AQ_d$~90XVBp)W%yYrhDxhTU zJcKMn*WawQ9@#bPQ#wk6teGBf&x7sFV;xO>rRk`@-gK-cA9>cfP#Fs-=yp7_r`9qd7Wr>O;pL$uaS0&Dv*V*8|8*Fe{M3p2!a3B^hDfFRj`5 z%sT3>FTyu?W3a0JN>dxg%HLkecXO+}fS|xKySJ@vC$9Na_&27$n(yC+5zOIij zJ}M@iGguP6e$Kl|jcgBF=}k+EOnhCAN!-)8@|k<)n~v`>c8J~>>0sU9(zQGXt*dUV zxzTilfNdJ(>)&lUowkx6?2N|Ewe*e!~#|EP;R4eTMxeZ#Gv(;fEmh zt`r{X0(VjPCmsIS3jZvtsj>Md6FjSbTRBqG5={`tlb|43hp zSJuK0_zYq3%um@5_=kArE|+KONAb+-jKBVvJfr&m7(DYU;PrUsTE=FQmEIwqiPY76 z#?*;t4vS}?#b}gpa0uutwJ>=?nU|U_bTtN4G|-spJ+ z=^W{lQr7wr|E-FqNfu@MB8jCxU9~xCW#@RhPH`kpKV1c{*>a0*bmSJFAKP;xZ&;q{ z*k6i1w#PuSQ^`Jsr_+PryL75$uUEm=C1ROV7W9!fc%A9^F|;?(8$QY_EwSvma^;C; zOj!`bhAPj$B(?YFZhOQ}^0dDsgq<#Jvua-9zO;(6e;5Wskv1ms_ZoQ4BoqrOw4%-@~<4Qh_SkgsBuwcK8>1z!Z1|E)EF8psP(k_BagDk9sak zKi(zNk8@`{`dv$1f%DCkte)?@-c$e6}E3t2EH|qt#Y^f?Y&cZHM|dLts;JF$eMa%6d#1S^GeO#+4CUy zcjvfQIqjDX@*L%T&5UuyikYjf|3$mDAH|;PsE<7*8D`!+)Z2@XjsH-mHD}C+m&CGD zR*^S}S6_3lLYQe9278-V1I{F@9;)_yze?Tt;#e zbK$-o*d#aOx!QTaqw?c-rQe2LY-dkxAFu?{fW4-8;el0FZ3E>Gq*}Fs37mJqeISFi zJ-5ogEMB-ywrsxUSu(=WO6G4;uUe0fuyn4Kzw{t^*I@fKH(>#zZ2v;$O71G>%u_p^ zIcM5ZdQ~c8J>fUYZ_klvuA~cihROqv@`Whx`J}%Fmr;DG5HV+)c9b~`K8DJ;ZBD0* z^n>~>JwJL54q6S?p(~gtx37BkeoBA4KlWwV<&*PIaLZp9%)auGo2w7SuS@PSm)RFGUQeD)QrRXSGT9+hC(PUZjn>zf~ zWA}iYzBigHW5^Nvhio~Dx^gtj<>OOu#4Ht~JX;_`At9l)e= z>#4(&qg#Kk969Z(%-xhpmZKVYXn1|Cv*uGgyBU6yEa{re%X)shKk@SDypkoQNtQ;x z^PZF2C+j8cE6vO6U8j)d$mwJ%xn(j-g6_GoylHZ;x*jvT*kYFs&Rp1H4m_^SK3 zu6|#nLw;;EF|A4XfE(Xo5^X-lVd7zQeF$EP;^P)SWUY|z)6Vi>&Q*y=$+wtmh!_;( z+IdPt#~zVCTvIvh`y|g<{FmW`_XX!$_&Wc+a4K%%*!DrrZcz+Mx_mCUSo`i_d%VOV z29Q=}+qNL`CcAc+Jo1sk?DJmB_+2;)BYW}>3jW5iTj+qwf-CXKu|*rngIrfuY)Z|x zh?P{c#;ltD3I8FLyU~u>><>hoIDhndXI>(>DRsHx12yCirIs_-Y0iYSR|2|Rj`sIA zu6%P9^_|W&7ioKcN>ZIE%ysM`%}3fQ>Rhur7}s2!+9M{}esxMk(+u)_l=Tss-#wRM zI_{=Ve*r$wZHbo)v3uYsh%S=7i;^Bg@2rB)s^1YwybNy*;p$2US3UIKE3-8z?>_Rb zrOZ;s6ZMQ6y?k1?zLx$4t!cJDr1#NlU4Q=z_+8Z3=X3o1s4)xb8TWW)*NtzeTt|!~ z$XW`;X;hZ^{C4#1T#tv@YkVZ?@-B2|AK&mADcgTX){1O-`w#F3JbOxd_z&?1`Q9#n z{Dk#G|G(gm^kM$^^FP2J8h^^}+qn7z{BhlX7k>~B`J?d1sQHLw9s3<Ad*ENGI*C{s@#fyl{wc!>acY*@lCDOPSyw4p%2EYKic@> z!Wgl6N4F2_5ZTdWJ(3)AfbEm)%<#Bp9%VM6^M`cKsJ_^JclGv0`t{fJUofiwUFbLI zu5*kNT zXL8=pIpyem?MbwG3;U#cLg?Hw`Vu3L>|5&v&Y6U_g@<6>v>cU{-|*rLedBHl>n;Gcxp@hsVDe*?TM$JIJ+Wg{rmKh>-)yyp_>2U zp`rN`c=v7b5Nq};*TP42m*JtZF?gu%_wdk!Bp%vJ*~{|~F}cw^#GIYWLv>^E(8dYF zJd}~dLy75}O+2*!%vlq*w{hFvukd&ND|^2(dL6lR!&vf*-W-_9`tNRZz4%MI30bPD z@sHF^Yr_p4>>o-5;cLYpByZyJXk}*|_jR@nd91U9dq?-9Po*=JXHilge#SW(hV5)C`8Hzz$>4ZOdX>gbNHM82b!?0P~O)+y}^S$TJaGlk=#aiXi%yb{7XZn?}#T#Twg^4DQ8KQJ(dlwOK>! zmi`~KPpmSjeZ?@6+8>HHX&nr4i94Xh(D<&htfZCm6D8mH549g)z0M|VNFIH8{2lt< zB72nlrbE70&()eMHu3kfu9-Oy&QFbWtcFfr+(Yjb7rdExli(>N$Zca zHb`rXck(Tc!{~>w&7Ha*alieIapHeVp4hG*Lk7-K)) ze`h})d%ga+N$yALe`i0E{FtPDd^&SV`@Nux`z@>#&T&X>f8C$fzF?f$w;w$%{rkf- z=X}Q!@W#5ZeO3JQtq5>e^>^%Jdf7jXZ?Z%>Uh+?6?})vpru{MY(rHiEDr`&v`xy@+ zXI~*Ny2f6svxhmm7M_VAcg&#Pzj5oCG^!rr_pDK49)xJ=Ur@;nNUbsipP z^UlNL_1t@S{O4ES4jzZ&D*6tazWuh3JpqN#{v)SFXPteH9$=q{_NDh~-T|n zzp?t&gN``ke3z<&1#?67Xbfij0}E_8*mJBsxB8Z};PYS`y?@?=ZwY%9dRNGf_=}~J zd$CC;*Dx>jF!Hg%#J_|71=c(caxXdZ#J9)r$HHgnoA%tC2S=-bUu!$B zq`vkOoCU@{A*EFmEt36Tp#5H=cZF51b&2z3drnebb|u1nH}}X&e>e6^`@Me7b4+U} zEoYCt_Bs>P)}ECS=S@+baa>+_r}Cq#+kDQMIO)(U#2Gi4&}?y0=A)%OrUZR5SG1_7 zpOUxXd9h2^y?g5)92mw8?K|(KuO=IxQ*f|P*4vM6r?dB3{gNDOK3;UT(;1_WcrF`$ z9=5YiTwQDSJ-x`e1C;^JLjk}0Hv+TA=NZuEICHGhUHEZ{6|A=dcZqZCz0rI{F+3@M zb1r*O`gmp?L0y!-<0QwKCz#BhGJE~>e0--ed=9ONy^1|KE!5FTTYpMhF_({bIM+N! z7fx#bd)ltXA5ohhC%KkWu*m)}4JQb`{K9Cmcp8>yBlQ*z=)8>Nn&b~Eg z-Em&~uh^qo4Nhu49~t;9IEj)kh#f2hM=oC+6CIQ179b}}K{W^6aO`TXamsmo(hc4+ zl4$hU8AlgM*PAKMKEq_b7;8Vyksi*~_Y?O92i|_tbJWZJz9HIwmU^`>jrnlJDzJHZ z?blGQz~uL@LEf4fgJ~|li~g9ui6lDNllBO6E$D)|{THDYj;D;mO^$IgPk zCxF4@=f}90zktok>jFnF`zMEA!p3|36wcJwZ_rn0I#=Iy@bG*V_@PZ$?QT=Q-2PW_ zrJa3I-^yp&fPt~kI_bXZa&zpC{HhzK!VqG4yxu01{oY|j`a?{7?aQc?Fsf2y7>)l ztOiG-yT)McocA-6Z^>3LmrwwYEQUXrBVN9&aqF}i?v)nhzJaw44e$YcIJcbrkL7$5 zMmo-uhhFCJv=%;!uIQ%zo57c-zy2LvjT+5SOh9`FU!)(B6KHS@|JXj~#?zN0{b1+( zM(Iec-?qP_O5L-P>i(`fYe?P5xsKwqAJ50+0QiZykl!UA%W_th1f++c__J zSBShZVg!>9PL8)QE-Nr8wdL?-<`t3Sy?lctSa;9nt(rq;O?zI=JqtC)$|IgDzUYBI zW$=#b&%4R2===lagKqg?Qu*0#`5@&Rd_fz}8^JUF@6jD^4DSDZG#|ZJ;`f~ctMjm= zX9nP}H^7a)fja;n|2w}9^u~|*PR9%A8u?-O*?83%43|IC^J&U4&eT}`=~8P53#L9?YYqkj6MrzqdY znKq~X&T4y|I<=p=Dc$T7oL>11d;fRF+Cu#NPvj`xZsuM{*Z#}}{pHlj_hiB=7cF3G zENiCG0Vh5cIlcv*lFm=xC`ESnJ@d=KU83s=#utM7v%t(A0ox~?2`-4qckC=*aAYmz z?_9KiZC(q%Q?fI@y9978EoboH0kjDhZ=WII}AM+BR|?( z)SG^ITJRUa=@+22;98_JWrzVN&gAW-{tMRs_P|q$TOYvA6mlk6ggxmy4;3B>;0xaQ zY~hhU&cr!I9_^=A+1ueG{Il5&)b|hcP49>k9KCa#IT}LNBE(XJU*R*I_2;KJ_u}07 zLC*G@-u@YE$XK*n0v?uwhYDAxKR~{rech}f^Y)NSwujTN97@;T@?OTf0qnSePwt_- z;qN8*?<9P90)7PF+Z`LMaZmC-A7@U;8PD@gEAnekk@!|;5U{Xg-zI*Xx3C}o^Qi`# z@339a?+W0)8n`_j^bhzu=sftDv#>^uzlA&Px0PKu36JSpV9xn_liIICKS(bt&tdXN z59^GIoAMofIuc&Tr<2W)zWOn1Ddpd67;pB8Zl%<7 z&Rg$iPk9ycbgVbi?91ny7W=s#n8bJO$xD3G_BG+vyil1r%Gwao>kQ9T^eg8hq<;b) z`god2fFaq{jDIa zieD5NMZfQL&f{>-TZ6t~gCE<%S-hOpNnDcs@Ze{3KG|~R`2=|Vs?EEOzdxFX;>L8c z53}?0{QjEXHh$ar?clf1XFB;9DQhj^H^H1-)Ho{f~giw)t$IX7GUU2UIFJo4NL z)tL@l8JyR!7?^6<+Zed+@NUiydCq~=fn`M8B558=XQU*MUHqbibWw<_VDjwx$I-#Z z8DHp3B)+j(8JpC((Ec_Q{5?o4@dA5oKZp*(HsuBUP8>K$9pHCg?9~?+;s=&sFXLa} zey2}k;Qa9KnBOSG556QnF>o(G`Y}F_JiFc`cDKdZBXQsz>?goJsqXq$3l>T@Nsn>{ z6yHth)EQ-SwMU>@a(S&Y&WKUZIo7lUm@m}#UU~y&ZgQVOz`Tp~q>5h`1Gnm79V=_U z9bS!HWuKv#T~Nz`ddvoZTLxeevxJc!AHp?(Y)_VVry-n|H|g%b}Sb}Vd+xvk{=5drh&cAee38oh*`+>ULx z_v!|Mhjy2e{uuZw{-p2G-U)^6nvW>vxiHH#6~S-MV?(8jWdjs%kv?~9j%>p4_f>|+ z=R>|{coXX`TR9gC{ZQKn{y)clQqQhjgH6!bdkMK% zQ*K*`*s-p2viT+pIEaG73h1zd_b)xn-ortwP4OCMY~Y+%E&KD@UzN4Z0B()j<*!O^ zSfj9i9dLN#9QEs2`la#v4ZyNx{G=w;srnpSmfJp+ve*p!yBhl5i{y4A@oSY0z&EFO zt_YeAtryqY3Fo_TdF?;NuK;)9v+!XzczgHSN26e z7a1sqXSz5mMQznkx61T#pYFE%Map~a?BHJU1!zP3?AZ}#Uk=m7v9YQnwLG z<$O2nYelNr$J*|l5n}!udH$gtQ^4mZE;4d_8)%o0K))+rfo7BNr#?=eG8f0U0EfmL zI{#w}^}LMD(YM+(X7tMal5!E`P0t!f{B5BfV>-5)vN2)|u{Q_z3pUv#z4v(SPTte5 z6KhD?3!1FIe~7M)eUF2SVXjeS{Zu@$Uwx1UFyH(jRfJI8AKN!o57Zu?5N?dfB* zeHs6$UHb79wn#h`kiI3JQ=Gb$^*hWhdh0{ER-n^#t{?t!{Kt$7g`eZl{Wbc!l{&V< z3#z9kxg5UeWb&6%r~Hf}o9>g_|CREhyL?I~{x!M%FIh8O%^6TXLSKn)vQK^FJIDM0 zYgFwx(mC2{g;so%vHb+^ME_QJUikhc&nowGu8L3S{-@mQo%GIVKVq!-+1ZyBKTGzD zhx|>mkM0D}AF12a5&XW+XhT~|N?sp4K{IjX)O)<+pN6*GxHMocYlW;c& z84~VPZ!fqDyqB{N*f%*=Tux^)ETHU47mtqoOm0VCUV_J#-w;O}g~wgw+nF}G>E$u- zsCY{<9zV;oaQ6sT508Jv{eJ|Hdu%*{J9qzb7jhDKitjX-h<&C?HN3C_+DDQ3WSZVP z3YHq4ZJTk)Ie89U?0TR4!}We>RJ|NFa!I|1+;fwY^@Hk{?o;2thfc4CC%d`Y>GWr; zKHM~_t%W?>`GIq+{5Oosf88k9I3v;?$0ft{1h!6f2C;9?ajn>sO1#gUkq*k`+g>%k zNJ%MvQxqFojK8rFel2xfS`%_;Bubqnz=F#uN?Me&ti$@OGu@AAQcIgI{MHthrOo zX>06TNKEr0<6HUb`0+cdeT?C38V#>6HvEkhcL$c!ttW0N=}j~9ifLCg3p}OoSv1Eq zUXQ{5j6s;Q=9^K>d&uu!iqC(Te4!4Vft7CeZLvL`II>&c+<`WV|1p)J7 zZTq}3Xc#{6SH9xa=qt4ySmk`{L;cabR11DV!2EH(p9T+y=Rm(+{50aUCB2jJ5t+m5 z#sBKN+TT_|e-`uWd)!}qk~zE+kK@lhjJ&fitqwf&PPW&?&%OBIcZbffRGs)6c7K!o zXZ7nd_$|bUYP;c0eM2)a8$5o8Z#i<_9%Buszh%SoUrzs>XXeJk>c8s8ui0t8+vCjF zDNXN^`IYy3tvyq1KT`AzH9GYuK5~NoiRV?upbPB0*0i|%OQ+4@c|ge&zUMwWJ?R@y z8k>v1RHnsy-!oh;d98hyc(7x5yo^819zVwiBk_ z)DuGnym9b4#=(+@J@9n!x+8oqE&aJi+;%?dbH3%EdZTVVp6%(zUwe@GFK--?)9=jl z$X^t!f#4GRtjYkfSivNIPL8(+CPtPQY?`>f2s#$r5UHg;Tc!dpW}D&mPBI>fxes_T2Z(m-T&-sX3jDy-Eus zwS`=LJlcT(X z(R=5WSxK3XA=|J`B{;}iXWz%{IiPxU$X3%*8REy9Zd;Cr^{_W?^S5(Qo~nxj1?Z`Be`DBAou?*UeRxZHJc=&T*s%qC9^@ImZl6^s{mL1^EO`K(cveNZSlV54Kz0q0~cHQ^?uD)%sH)`qj>4} zN+q{vZ?amGj^tikv;#go6O0fWYuv%NR&u|d&i6<^ zh~IR;m)n0e{OMzy9Kz%4p2orU=wGV=)LkOR`54TdA?1aJl0u# zetd3|V0;Z3_}uvXCXE|@=O5qRYKj>r*DzG@@X@`xukZ?SY>;$y<6#5 zIpG51+5_2f)eX~%u>@fM%=1kgDTT$VF{88He z0C^9vHc`(@xv#c9{CYQW$_xH1=0&n-H^ZJEJAsZO&hu0Qa;ezEd!fDRR3AmB>nwN8 zBKMp_m{6Nx?w5f_(NZ|P+;{xgd)p3QJTHGvF@0}clk{3Q`quVOT;10DFn$dFjQp1h ze$sJQ*g9^QF4Zo+)2S;MV4jGvnDnpgDedeH09T;Uv~ebw4V!e2^o7<_T=Wy)QoG}P zZ6}@`)~RZ@?{V6G=dk~zwmDx8{pzflx9e_9j~}Klp))D=+)A=u^ZNJ`;I;4ZRXMj~ zSD}laVK^Kf`jWf%-qUHxewW(Hr5~~-z3e5{bw0AG>zCYXCi~Cy{KCcIHj?$P=QoHx z2Tr2PnLqa05B?>s6q4+g!a%402vI?|ukOvGR%bw*ni$^*+wG z(sT4J2N$Y)#JDPH|C#EqK6P{AYm})!m6EWZe|vM{aXtUmea07W)-z|O+3z_M=utg$ zhLnfHvHBOo;G42(9AS|Xe*#~1bU!^jXs=f?^EZw^rFAJ$ z^u&4AReu89)876g!21*0pMsCN8hF1zpZU#gO(#~)xITcclfICj^9}slI?D46jmqB= zd!9AU#OL}uOYTJvzJl)RLO!-$#d(!Kt=asXnKJ(y*qwdMrThk+dJlPo!{%?A+FoD5 zjTdra({!&G7dWVD{#$2Ec^zfuQbzfEc_-L50ndSQ=1tJgoE6v}EB8C!USr-SO4-;U zvs}Ju>{C7`{cbn?e_6}z`LVZ#(__Ch%NJ8dz7l2URuF5)2g&;!Z8p)i&KFrn+sS8= zXq=^angDt4Vc*z(U>AJ9A@&iYT!Ftq_)W)-{|%yKwngk%B%ek z;ty*rL3;2<)0XbZpZ>?{7jqo5Blwei_j(=Y-dDZoOJDaaebzPITKYf@_gm?U=2u^s zkv~1iz4&5*H94FKym>CXKZjm1`pX2 z@{L~CRD-WUn|4f2^Ct%WnN8b0#N_pTyb5%(tqYjbdS{aSe1bKo@LG=OqxyGQf%$K* z{=N2`^xMpvqy4cTLbhJ>Pg;7pbvmQ_fNzGM`fTf1d$}$re~Kx^z%$>R2yb%bub(qW zegt3S!#7*N<;N_O_eF5nHHQ3s51;(Yl-Jzq9Dk&aGeqn)0je{XHDK7Da10s{ORp@V zt_b#`63*9(ruJX2MF$kaZ|*?jePmhBYlpWK4Yb<`z$s;-hNY|Q? zz+3-Yd7U+((t+rnfbZc-eQ#s&Nry&ue(2S&{K$}Wg4SF_)0{PeHJ`DTw^3&)x>EW; zc{J}7-C~w2ExH9A^$W9HGQ>P@t@Pvz@Q}Uk1AfyPGWgnc%mKz9fq$Z_oO5S71q*pq zM){eGJSzTTJc0h;ECVZJT{U@T!S~|%fBu!TW{Z6B6}(qD7Z2kzichD*r$2G+oW^~Ui_vy&t83>}-$*;Bed|-gpU0Q;?A*Q7 zTjBbNo3V4~^dUR4;fg7KY5Gem$jfDG*%3p_AM_cF}3MD zA4c9I@vZh*{9CDq^S$DQHPL6}-zNL{ile>CoC&nq-$na@gDLj9@xV65kl4Rm%XIt< ze?s|K_dA^LDK{VAVOJh|Y~-HB?fK(BPs(rHdKt-F^4<1h$@$!RP15(>N8A553Un5L z7r!#$r&+t{?iC0#wzBsTX+1Ueq2t$FL*OI8{dM5zY1WI<&%NHa3ukDZDeLto6XQKh z`ajPdd&+0{1|#@joXOnX_7U&kx9=yl)9JRYF&E#q3On(s z+%V%GJ3gg)Zlxa0Qvp}rJ~zMCv3RhqU|tr#Zo6W4ZxlolWBWzkcaMf~A)iR&k?&K# z_9Y!e=2)XQRIlpPT9zo|)xaKS?t%54_S}Q_4bN`ugY$m)oAw*1;|#K`c?H$m$sVN| z@Mhx@oQ>XprEea#7I3xzd&G!OrALWrt1ss%@6A6wLVbomzAI;6$6lxH^zC){uY&I! zdo&WgL?_Wfd(zbJ;;Slc0%`-X^#IXodrk@RZJi{x-xEOvEhd4}(}}6ABeb?~ zPQ#^_B-T?cP)Th|(6%Rt_DJy3_FPOr>x5`qib#TDzVFX|=1C?YSkL)>fB*gRdd+Jl z&$BOUuf6u#Yp>h3bId(AMut!HvjjRTMR({?KYV8s3#&RG=6}^QIPD&5;-%!%AU%z2 zEb=Rozp?B-Vf+hZzw*h?ATc(L%jYN86|k?geN|z2!STi9p{HN%{mt|np&y4PiOcJ; ziZbalP8s2#!c%?*nX$Z-a^rcPu^ZKyL35P1%PC3D^|J`@k#TIER>js^Un8JXTIIJ*1T^-R&K$sL-XB*yvsQ~dA`?c-mYa_L-Re{cV`rFAgF+LF!Ix0<=I>ZngpPT&k0d#fezcXEYVRZU%bc>3?>SN8zfA)+w z#oV7)VeDab%U3f-#bAk7jJ%IleEpeeQ@1U7jr;8IVe~zn-cY*yrixQsdTAM9CJh@P;LXHDtg@1dK|>|(7%-dnlwtL)1`?kVToJ@*T=qk1jW zOQR0$%)1U9YiS}*3jUk%5!#PXCt#1S*1BA?=(0wg14>y>+~I4^lbY{OFrU?|qb5J+ zN6P7*+e&l>vOkzHFuzR`oE**PykoA{lw5X5dk}h`GUdF}x|jU%UEo>cnOt^^eTbN6 z{P?Ee-#^*-|K2mA$j>v?#!`U+T?)z-SeibLR1e7^EjxF`glXB`l) z*_4MLM5ywV<8yF6z4-^(*X0YX{mk9hdw@av`U$>+(6o==a~b1q?j$yQQQ}VGX`CTH z(7sN-YU%rk*Qn{~h1bUrJY^{W_btHTf4hye}aWuN@+r?o?Gg z)!@F5%$gZmLOZh219#R>66Jr9)W(g{W7E(3@B5?KZ5-V{cAHQ)^3f~EK{s+=PB&xv zr$j&BP3i}y&8)i$hH&-@<3E~jm3Ji8%~g_v*kdQN$1b^deV^`+PQ7%(wlZk+k-%`A z{X?StuktONp^(bD@nUQ4%!_>s3Wz_`ozLcAVp5Sexef-DabUU6W|z z-b5Rk*9>rRgtp#7K9R0fa?6YFja$%spS>c$8f<{p=FW=ZAC=Ym&)9R2QCq=X=dU>n z>-!{jmZ$m68D#J0Xg8Xmmz(+RmhtBiiM7vclHOl{6&O}~R)PVe-8 z=uKxIH}$2%pl%gyxb@w-!kMYhn(t=MAFO{OsXosm=c#Y%!V@&kF7lCVN4_v^Ft7gU ze)x8KQECc#HYFDx7#Hip?_6=+@~e?tk`K&gT#KzD&D#O^Qzd1C-9u%KOtadGZDf9w zGs5_>z%xyoJBDOQ>~jyF%nkSdE%@)*wWx6J7++dCzvb)Wv^jRLf79j~Cy%t-W;}pj zmu1zYrr&r^2Rf%7^1bXa+(p#82^m_vLU5--3%Vmch%8l#EETbH z%|1DoeNwqtxF^cMyBivP0UkUOr-3X7o=3o~%NM_sz;iP+q4i|=7i|UhI5|<}gXrb| zFHuhKG$-(=*0(vQh)3y6$vRTqtVYU+?`}?NKScf%-ovYYd0yM7vyUg`KQZwy`;48^ zjKOyD2gn{GdWru4A6&^qSK@1p-))xiP+0%4Je%`tQ);+o^N$-?@0p=ZuNWJoImhjJ zYHp+SemMs==e-1-YTws6PVMIsUnbr&1-hui=TK*teCCgQ6OZdm-GcvR3;d_SLg&uC z-o)uDjr3J=Z+WO-%C?4$EBls^+n|~9njfvpy?kC`UEf4bhO_r}jH;Upe{a}B2vtNZCb^pB4hZ-0zn$7t;SMc0Jc^`QT z1yeINL_c;(#<(ZA{;s~fDVCG7X(DTxad!7P{-o!X7n1C)UBI*vI%RH+f7-x*4Ln?j z%*(56lkLLb|G_NqFB+++TH6Pn_H@!#ZPnbqHO3F>f)k6; zx8&2-IriTX`5U%VH||9@;BVimdn`A>GkjmNw-wU%i}bU7;$9x42Jo3#GEUW|Z?#Fydy!%u1Aze_6?@anh z*82?kn+?+TgIOiH!tHX6KW9u`CoX$#x{!Bk}n?0p%tf_SFbE@$p z87JZv>CB=zo3_ki%KA2Sc$Y*U-0r!%Uv9Ta~m}LiEviKX# z51JPb9-;3ez@T;nug09cV{9y!wt#6_3%IUWp1Vyr>7Xp|EjtK&bFc>qKEX3GW&B1d%tlXa3NoGUSeIPV$Js}IV|!Rvjp#^qe>pLT&*bAT zBtBPxJ&DiWS3Hw@EzY)#>v_NZAv3@$CIN@r<{R)p%}XWoBA+e!a;dxzxP8n~8h6Xj zOw@y}S+D2ME8x&C>nahaAsl*UU9Gu%-Y&V?l7Fyt zeB0??J{-d5c6`6h8cxJ*1oFKva)(vB=`V@jn*MfN<@9&ps+juHT3nXs>!k2JqA#}( zca043-_hvZ;Tgz)hdJ*VU75%J2!5$1b|7TNS@4nA+g@*IaXzvFeV^vs&wbe;C@&w`#6`)uagTl|uI zAH)Ooe=j`8jLY^LUY%Umtta{bZll*e-`+Lq{H^+kgU|YpIP=H2D(LrJWKxYAz4|#g z)mcY)HMT`Vs}9c=49W92YJUe3{dF6>8a&o%oQ@2v_{pUHvJBq9L8y-YB&QlY9B}&6 z^O1cf@2i?GkNr_+d<7rXJevNt)88-uJN?zaUw^Lt+o2uP=PIW!=?C?{UH{RT=%t;0 zzfZrCbLI^j=V0tR@^z+(YnPvi8S4(G??kNV;CPewR0q$%Ve&*eaL6CQz;eLAqSzOQ ze!Cr5!km#eC1AO66j&UafkThm!TlgSC4=xdw3?XnbK_$^;QCtivEl2h{$qpds}2}P zuCJSb`TzC$LWbC4a9!~KVttK54+oqvVjoJ(!N(kWV0_aObU1g|95{53&BowZw68Og zc)@J)EE(A&Nv6{LA7qaX9heqtW*?TnY-veK?0Wo^(be~|M|Hlj>8o1{(b<*ZGputT zI(H-2xc-tOWs_ zC%ue*2YxrFZx4A~BoFHDF~RpJ?{p?LHl!rkPxyZyzxe)b|H(Z7*pVab!}9YjBo3B5 zhQ?<<2;9xmYmBinbf%Fmws=h59q0ne{hpiWNATCAPEn6^q15Mm-|u(pa%T|2&|E#Y zc#h0u^2{(dwdmrv%a%#o&CE?J-@vzL5MIVw3cMr7 z+6}x1t(&po48f~&kNjL(8Gj37X{N15(Ca)ho%8Rd+*>n~_OIkOK8wt4e=~ZvU7M`f zPOlx?vCE2WXD{5wUaXvQjJ54*;68irug{6&bN!V~enxaZ=)C%Ss2gAH#CiFc_kjUV zp`Y*3V!dQHmCi=`ALSB`PYt)_b03WB&z!s;X+Pz>?96*&z6-#Y=DUk?8h4zwWZ7wD z@>OiYXGS@PF6a4@HwGSF1rCUP4!z15jCa+XSIUR-`V8g!ksN9<$Fi|S5`Fo&mn_a1 ze6T!mCT@P9M0vAThtWZj+$Ff7EtVxYTo|w9OYtJ)Ta$xuu%FA*RzMr(tau?XujbC! z3g}k#B{U+w!Qcez@K0D@vU9bg+uaTCT_M@h#RvEaa?TJO_5U&Oa5eadfD7VOW>=m~ zjkYGvAUz8rp$4zDK;<@2PUiwUjdLn{!=E@uZ-;j?Kb?~qQ!{!d#=pn#3dXAc-LtX< z?|hB5w3}yx^h96Dcvl~vrH{?vQ1cGGqSMO)Pou@ry8ERwCyKB16T@lnO)4{sc znpM7;cFwiB`?gLxy~h)G;ilU)oF~fz+)E^W99V+B6TAMu5xcv_bK&d^U|HjtGTZnq zdh+hr!8v!0CvWy_o?lA8I`e*>IN(e8J!h=4hafeO4HJ*&wI*a^~{+6!K%8(rLHT|};XX`FU!Bk7%It!?d_OCUbbf+z1c_lZl zu`1?6OGbyX#*<~<3CG3MYrddq>jZm3W6xXowW3EHK&QS3-Bc8NxaK=bK1_pS=KBZ8 zBYIv+4xbF}b&#FMhyJGpTI+`I$}hz1%}KIRgni!bzB+a*^C%j;k2^r}-y4|MkY~k$ z%&YncUOv=^@>vw16Q2t#qrqU>vx{Rp6YYH;|M@!b5%El%y@dEK&8sO>QXF&Qy|P;W zd9ci++4w+q`rR`1#Rl>JN zYPD7L6zkaHZ%BS~bT*1X>q^)y6`Otuyt;e3)fV3rF6v@$2F7{PH{jczbrjsCe3(4I ztPyk0U=1!J-@OIzXoVk@p17f{7kh`1B~DCg8(Q~iq2zT>KDjpW+gkLk;=PePtIgq$ zq|8jU7UPTe5J&*5l{ag_>RY^Ya!ogZinPm_-4ZYYGnNV zMbbecFG$u3UCTWP+8ZPXo%KbV6a26OnV^Vr=gL1r?T(CZ97p*G?Pk($#Zr^gb+Pl*$k6lOeM zqptdl^PP;tv-dt}4?#F52T6-^7u#ck=1kl~PE%fA$gk+w(NF6+SWb=l`V&YN*5|85<4CC0aP0qu8T2eTH2 z`{(i-U#x!X@^Jqw_xG%D{}t}p{=|+M4(s zG1Qw`6KgnA3cm=L=8QW`9vxxlfF2%9wh16Jh7eGn-~P~|HqNE9@>zo$SPy~?UFaU{ zPmdp9KjggGDL)sLL%;UWgOsVjXU1Od$U;BlK1S8i`xlUTDbx8C%1Aak%sctxi$`wc zSv%_@(N5=m{IC5*zWn;W8hokFFRA0e>g1|md`5>u+l)0K6D9G+FPePs!|@CJ&(|f! zCEoIJ`jU){Om1}KVfK5CNo6#q&+vXbXG(WWcLRgU>-~|$n3hvU@ck3-v}ef{{Ry6F zOo?_nm+-&l!x>X%t3NTOuTy7tVy=vAozUAB1N+cE-iHm+7kh(z ztQDNEH1Bt_j-j{V_CD%Z)M-!H`yL~Q^ZVKR+&#K$oZV6a&E#8F|8&~W9G!`PzOrQwG}(TS-y+2t@@wtM|L7-!vpla-{Y+=<2xUCfcL~}$mGg_ ztK4S=oI7U_PpdlZj44RH{Z9pZk5N84w!NsC`O4>67c{0lFS)+(ZqcUOPA2VaKQl0| zm^my>@(nyp-DYSd(4B%CmFVe8R%`z4+CvFX-&&U_~{2j5K%pAqlj zTwA_E{EEErYI6a25Y7x;^DItYX8hlJdnh**cwE{3$G>C!BInOEbC%ND#eOt^ystg6 zkU7(M6q_Kse>eW3N5Sg=a);LK<;<7*gudH2hu3_8Ggb<1Tj;UUY0t;p`BJdsQ7>@T zTXT@}j^O@oUiggQ9mG<@e8rC?a=v4y8=mw1m&4@UwZL$v$$4lM z>FmzBTn?SINssaxGPiuXUS|*H4BGiBaBH59vOeJD%bz7Kp$nag@y8|?syiOxP;*== z`6&N?X4KiuHdYF+3if8{xZcwThi?m z?PsluW|X5^w(5TDbuSU8dB!^~v}b!$-tEv*^a?9|fHf3n9A8SpO=bc&k26owWtR4@ zJrXz@II#-;8s8RfdmX=3?g=B8WZm-Dq?^Nsl{+n4Ik&LJl&f0)UD>Tal6=k_3ZC!K z){gPa9q%?=!dW!a+S75dm5Cg?JXAS#MOlfRo@QA^f$>(+L+Y11H=Cg~=%q+7$mgfS zV=d3S$jV$2n7U&AMOJ!WZC&Sjp06yt{x4Ix2eX@T+)Mdp${*xi;JNT$7N$3C)!9AB zd$mR0!qN+@W%4QXkhf4{D0ugpBl1Vwd3Ip_ zt=y|~415Y#7jWK_?SV1Q#qap=o5XJ44lGf8nf+6Am-GZfZ}*-GEYv<)OWqC9WE1^$ z9JkWL^cm%>M^3rSO52GCh@5z3p^s;g6M=<-{W;1=mufQ}Bj0pjcUUyHG0r~mD7Z%V zR8v|#bwzc?x3~J4549uQ-;Pc|G~l-KkX2Aq%r5H8dH>c z_F*Y|re!TIDy*6m}{W=)UT-so1Xk3wX^{Bh(b zfnVGWOwF9B#Y38nE`m6v1IR<1LHNII0)9zf;4arUy%Wmr84Lf7*)`R`;UiCc^M>Fy z|JjNokyK(sHjK%K~u> z=onvFB%{5^XHQ6rWLQPT_}Gf~`7HdQfvKL~()-ydR_qe+72lQEuQ_ie@7G!%o8cw$ z^F^+~R-rp!oSd-tJdYhQ-}_?V%e~8P`~2rWKe!%Nd(c0*G~?vb=!Pz2ch_AA4?`a> z{(j)$R&X*IK65v6St+<7hNs#FAH(MUK&Y`48edsCXW?Xcl>1KbHz#<%c&6|x-gFuJ zPX~P4#eWc(H3#B-K7K#U97Xu7rH-)!!Pk12M?D*&r_4ioFPf69q`m4te>Lzh`7Vm2 z6J7gMu<;dW=sD&!YU(HMe;TC6tOPw~wX()zzoD+=&CUFF_Ylz;apa+~41Lh%NxMSP6CDfC;j{>Oz@z=sSUTRp_bkQGB(gOby< zp69ThUx9a#&v;pf$A9R6$5~5nIX1DmP2fdmc9lo3x~w?en)RU_H{SCMv@t+FhWnYz z2)Xv-=qqcF1{S`yAl&~U=2kpYeK~O*S*?G|+0yVr-pf|t@WLtZLfx%BRPL9YbHxis zi{Z5RbA7rB8DkJX=i~6_T6C9P;AJPUaONqhpiLjoIk!9>xDag>s9(co}!WMWKu1?2V$wQa5L`J^qmTi+=^a%X*$k z**JA{SCEG~KJvL&K=WFw^6MRKt$zLA9R0)FEA2raHSz5zd$(eEKJ`(_SvmKIkXOoP zyu5V{a>*vzteu98xS%O7#C&{_oa%Mt7j2|YgtC%-LX^FpvP}h!Op^}`;3I1n>o5qP z@D;qgl{q{dveWKgiL7N~s~C97iDT@@D7&9Lo~Dg5WEja{wa96`oariTD?_#w^kX$U zz#YDFa|LVM>U)Bi;auH!Bp$E^Jn4Dqy8};X{F;NeS1N`vSs(7&?H#(7Ec_;8*M2NL zaxZv1OQ_34+K^A&tOtDHEQnS4_C>`!s*Y0X*WfWR+7`)nz&h{itd zoxI?O=$)epd02jyQO3QB_reYGYu2;Kt$t)hBlRL0mVh%<2iNAxUS5P*7!8J$5QznkQFks#LmmFfc_Wy%jL!O@U<`m#spLfT0 z+TNc9ti)jKpKRCYOch5ic-=cb^fq$1&R56Czos*|p!9W!KcP_6F;EN2xo`srx2tN$1h+@QwH;hff6H6G!RC=u7EyDSkTf>Bu_BBu3Vu zoZ%0bI{ZQU$b5Ld$IecR!XN744{zdw3XPj{LMH9M#yFt!< z`;7BJY_*PzN8EFg&q}k-aLi!~=exq!t|6CQ_{YU|YOkAr*X0M4dmi7-whnA>UC_?l zB-%OjqR!HfW%z-m1e(zqex=qhKGV-~m$7JFdIG+UK3BS~A$@_fcWQkqZeF$)=o}Rl zF~-z;i{Dd{`g$zU*L2?9nCJ_?Ak&w7pIATgm-Zm*+eSB%xKGTvo39yoloQB_H}oV?vuCk~?oE!_Y z(7l-duh#$EW7|dhdVdxFv&Nv4%+>|`jzVY5Sv9C$~(S2>m$FZ<&bg2F07thuF*CoCKmfd?Rc;<=mLD2tQ4?DTS9XfIMKWEK} zuW0X+JdSVW-pEew0SA_f)NubwbUd{?kz4Mwrvm-$F%skK@OeEU6qbMJQYm_i8{w{xfXJJ>QtgR#YPE{qob zDE|u>U-SAy@M?3%sC3MN3w^~F<)H>|dzS#0;wyg7IDE<<2j0Z1C5t5QD=vROjSkk` zL)^JQ2Y1d~yvzAHc`lZlxd^X6-(37TQ9hr!IB=eEjy4y_2ku-HX)eT9#ZyPl3Fkh> zo*Vk&`^?EQ=HrgUoLu@obMmVMtm4t%JP)j+%}F?EPFy(MxpCpUZj`wpC+azKa~j^7 z?0YeKK7+xyZFHpEL&Ufm%-scW-qLV7D*jo}Lw) z2TqN1Y&$0xE6&%~dg|j=(LkA1G#}omI$g9adB#0st4t;Gfn>vO)~x>5z1ST*6W>~! zf)6tHwnWYj>^=FZMJ+ApD1zi<&VpA2xi_L0zUJrdh*~?l_svZWTc-fecs`G^7gc>Z z+;=`*3Nea}t4p;kd>1*8Y1U&d5oBMn6flvMG`}f57=pGW~ z5I#*TM;E;J-+^nRWOQ^7J?!lfc+JpR+jEFD6I-)Q*w zbIiRyaqnE>nDIwvk8sbDqn$_N+~G4Be34%Z`K+$@)Sm!0xnKly|{rTZgisH^24ghnnEK{9pb}c&yIf^6zRT=08lEYU3bn2qwY1 z86K)L%-769G}%#G0kf6kI!k;yGr(tpQ&Hu19*SGWCwo%eRHQFYs}>S8*NTr zJfySylbru-e6>B~wDGoJRP8_OwPF`OjV$Y!(B9a_Ixt-9~x zhqNC|YP*?o5!$AY`Gt%{G9R{6^aWXm25QGKCsxfst*@mEnBfoo@U_l?)t;jManY^g zIFrW5IrcrDvB~F`nh! zPuoTp_9ka0e7gI!H)v0H=b$Syy0ujMbG_rYc65`+*yvvp5pAOv3F&DW2@*U(3a4xrYN@zKCt{Rx!31w zFN%X}jrl0NQ@$~$sIT$PL7s~5Dk+NaUhBe!)(!r4IQZ)Ze_G?h-}ci3^PZU&Zt$W zC<8wXu|{IIJVh+g^l<<1yTBd)FPo07d;lN3Pw5ve=o8M7X(V6jtSVWIzUL;(G7Et( ziodFl{oKfv2|3g)w+bKm(uD0G%(tnF+&GhZ3AwRI@e8gDFuJUg6Ix83YsDQGjv_x^ zLLVRJVt0|Kt_x9lr3y;JX+r&Y3o?PKEyyD~v zZ_5C@ZeY4yboK@4{eQqGs-eUGD?WjI(gp1IvaUJL_A|!VuNmvf6(#5$6ZeD+#?iPo zGY`Dj+xin|D4)4l3mrza-`I|iY6<@S&Cq)Gj!R?Pd7h8_vV;9@8hARr-?_`u2QK(u zwu$A>(BENjbcFLGx}vto;oW zB0R6%P~Th&O&s9a&uCNam8(7NM}m*f`Eu5~W%4N#l`le)kLZ}h7}I#yKwD}jf}d%r zjUOaq)c)(9-LTXA#Vk?Pc(l2ffpv$w8Mu=P2X3? z0>7OI&x&sKGd4Z5?5pvsNc<)aPv=|noH#?Hk1~E&gLa&d>~rGFpNF(|V^OYlKcE`waXNo`2p%VcjmpKeJ!O8TWC zKQH91MBKIZ_RY6SK0>a8K12`KK@Y-z8u*VuOHGTczNQSjzmWG$Pg#AbcIKg`9M0I4 zYZexA-j3v03tO-)G<-fK7J*jMZ2zHD%0?6OQg;GvyiU7+*yW2IBoA#5a!(~W)ygVwTNqrWoG2I04z99_M3+6}Yjxgz zKJjjmGD>%!Eo@c->r9;%Y z&d0j<>ppaBU8|T^@%qWYU|n0+dFYfQ@1I0wI0gSw9nZDUId%5@f;a}A@AoLSAS?7x z;+%*+eO5hV{3`j>I}&jroSQPVMvifobzeFc;F+jF$5b>9puDx-68Hh$+l{MK5Ok5s5-aob9KqzT%hStSxZlCP$RolU76yNIzynpj` z;r@%j?RNTm37GYNGyGTj^X>4`-QbuUf@bWI_lErbTkZz z>aV08{3N~dy`k6&^>dUmN9bGSjgBZuMwmw)`((ZD9{6lDBO>2P$?XqjvIn}j#wWnw zy|Cn>n1RKf5{q17wd})>?+FBvLSVugQc&Ey&(mEYW3=O=y zmiQTR(X40xX{&{Y6ITPxqjvz7zh^#8TP}|UUeUObrO91&r`aE}2lt0xFqXI%xg*iG zYDBBJ;=w_9a3S~*Ei^B!Tv&`hrO&>Ia)k?5!>23Yzv;lOG8K%?Jr8mQZX=E~S99XC zm$v!P(^fmRE{m2{oSf|6gs<@xU zhYyMWbKxTBuSzt8E@-Y~4B7bnzpGjova?f#bJ1)S>stNm|9rmBCxb}-NY)*S-va~Y zxw_6xrat_HG3j~pKYq*X{Zaake)(IR^L=H38P46XJ>_-FgEKnvBIp?c&72$IIg(j) zX7?T0YU*vLUhu_j=65B(BQH6>H5a~wTv5SqSH>uT7Zq3+WtcUv%FXK+cu8eV-?Fg) zQxE8s)q~C~V zenq~4HGjzb$E}sj$H(D^1IX5|qyIB-mpJlKKJjmZc0*#>U0!9%)Ta#QD3A{9M*J|F zIaBts2ZHOlJ=mOm_*6z{f9Ou>w81;2tGQD;LY!lV>|Qo^?c&3@fcv4*l_8f7*MVbW zZWZjJjei*!cv$zZlb{D4p40jV>$eVILDbiX~<-`vqZj>M5*T+4Y}{mNDxHMq+9AK;1{ z!Qg5hxY~}s@B55ja_*W$+mcCl8a<$6lmFV!%$n}((3xI5esD}#i7}nK53+2A>s5^( zqH|u9UiCC~;+N>}Q`|k|$~4SP#%}sP$(-y!*M1UNT=R1~K2f3%?dvH=$dFr*wWAJm{W#!z=h5rQP5KPFu3aNe8QP zf5W>lZK&LOeygvyCR%%5;3GZv6!6a+;?GTW6YNU`JMBq^73~c$FWztMH6>IL$V>{0E_hF^}N?wEBFog!}bv5&aalS zzQ$7R!6hj%Y&`@lI*ECEkljkL^XMGlgO@onc6Mtj@TStf%8ui= z-m6colOIsWrC+CfX6t*v#+W(d;*&OtudJ(cb@j}`aV|cbL+1@={cBAkXYSQ|BWEW0 zxI{!3JUKHGTQizm;}~=Z_tHjxj=?mC#}X_A8ZZcgEey zorRBYQA78*%_Q51v*Vahjubxfg86WrAU3zv2&-l2{ zYV>SA&-l2{Zr8J`c*e(lcAK7E%`-mkvs?9S5zqLz&;Cx&uHzXW_u0*Q#<@vnC->P3 zo+Y0pEdG+uF9u(%{Xu+rsL#i(zgTUK& z_t}*^I~She`=V&NVWQm_{B!sr@ekRqPHl>AeUWughmG)J+K^9iZNf%q)(&gd)#DD2 z+kqa)gWF9z*v=o;cg0cs-uPV)@2YqTyMg$JHjHHIzi;fm!Y{cqAn#Z4vs zz54!iNq>*NKU>oOI^PCANqJ^&b$A3hKCG-I_^2Y&?Mb!5>r>h5f_zKv*S?6qSFC^! zeBv@DbU^Xr?oa;D!Cur4?em8|+v>~^-*dp%=|zw)IYPda+l-pvvjp%`&#N{-$tk6R}Ld${e_p6j03#KUVC!j;df^0o{wv=i7v)A z>Yk~7ZO-^wx2a+5@T25|Nd5b8|1brIC^KR|?ysJC^JMoGM1sB9pyPUIQslBp~ zJ12WB=InZ8jc0+SApvJYl3u*D{xH)*5y1;4E8l*YJ1a z67QbSJKFIk{l@9uY-4W)jD+Bldi3c^&0uCbFX_8`2Go83ZBOj@VIt?zXHcFUFK+T zbaD51>d=~4mYgGH#F`lXZupuQ{%-i182)bfni%$uaemM>vFrn^iD6|%u8AKbe~pxT zZo{wC?3<1*mAKa>ZLEi?3g%+-(s2I@&MkG&+LuK~$WylRTP*U#(x6GLCB6GD?{-61 znychJx{tGznfDdRa+ckCgf%99on`54mf5?CzEoGbkxWZxyUf-j%>7z?o^?jp!rq76 z)h3&V@TD9xI@4948`1l*yjSi63x1v4|46x6<4oLF)_Nnf7dJV>UnFOEPVZB|@Ka>a zDEpBQ8DDK>0T1Ux14DX-GhTOJ*n3~jHubG@(Z#$|+r+n=_+rM5XLi!oPULCr8Fik_ z#((;$k^3E;aFVQ^r1N;tS1_dW5T30GV~~BU6J5kiXiSRu6(JqaTp(0_TH@CdfL%*;Y;w2#!p6@AKA)9 z6Ls`^6L8;2zv{2Uz&U>pnXza);Yf)A>b6aFF6Lb7Yt(rk-3h+nS=GXCVTt6@)ZSIt z0CW~Vh#bBPKXg|gaMi(qhZXCz0DL23F5l0XCR1(=v{u3YA?EaL`T$>#O|h*Qymf96 z888i+ddA~><|gKzIJ*-StnZgNvud1i#_5|@zuY(Dg}f;DYM}q>-v?jVk6)eerFf+` z&_hloC)bnYUg+U>G#y??u152o`Wln}^2NqZmYVKwEaU&mS&REi`Q3^SC+CIUcE-1I zjMcxCG3u^%!4=1!(1pDK*ae$-I(U>$HCOFRKmRO!YR?e=3Bt>?-o%Gi!P_nPuK1Ds z21dSTEr#??01EKxrgW4HqM>ZiV2aAk$Gm1JLj&&Lhf1&b%%?>c3y_oNhqnk109aN z!{yx;e}nRsJ;gdS}% z{#@f6pKIyc;J3?9a%N4{8QnX%_xPpcAq`mOjxxcaI2s@G9RI=CmYLX|?!^~cz9lQU z6Z>n>)I#dp-0wJZeWY&!zb8Qhkpq`E_Dq|!E!0nZG;=Mx<9=+#W$;MsvHkne&5g%j z=Q;A#XlyFm!F_(Ae(vTwczod%lz+*dP}X!{es3ds{`hK7EP_v|${pu-uXjS|vVMGM zDvuj`Z7IGskD$Ml?e7)z441Q(SebtTtd99u~*wU)zO263#)+xKa@`OE&mU4 zF*F{(WRh%}v6cN+(I;q2eGcsMwv?((?m;W$dm(xYPi^?fv17_pfKJE@4G^bb@`Pu@ zpIMvGZC0fxT%_^ja5rHG>$K@Wu$P>1<>cC}8Thx9mJn?%;d66(gVWx0a`fSEr}nVL zn)bRfkj0*w-x#7jb3Yn*?Y$&*+wI%|7_~3XP+OZAXPjqM%*i6gBY#srIFE2wV3d0T zkAKBe6ybiro=u)21e&as=%!Sb{243!zm2bJv58s72jedEWUnA&d$`NQv;*!0%YJyQ za?RCElx&&PDt#7I?d189F79xjqu-x5K`LXlP!9Ifiw~D*wlM9)V|%8t7x8{6|2F`4 z5AF7b(zZpQ4f!r;&NWvfd21JOw8CS2@uH$;*5&cy5xORHb)0eV9rnbQKJos#rqTRK z=27R$U|r`8d`tf(-U_dfZ1cPiK7!4tZR8!sIs@qJw06<7M(rDt@fX0;b`sC<_G)-7 zpGTP26B{>n9$!6+93olfPT)%DSTt2kK8V`_9@p7M?N@l%Bha${{m<8!PgAt+r`!4b8rBw*yDLFsaqy1LnDw30i=R+i5%`;iol@hS3{N@(9-f5exQ8;- z%kSgfsj)bEZk3mR4g7r&pOSs2691Tt%cmui*=dKo*wO}hVr%{KSEo;n9i=_ddp>*V z1Esnk$rXK{jeWT>gTFN-;&3iJ6;qYKm2~!1GJ-a zhL18=B+AcEYU4RXf-Ohh6{QJ@6$-U5{*EaH;_?!HhUd*xP&f&b7k?1dvZ}*J& zc%r|50iJ~X_t;0^mwCutlK&$2-qSZ3zJEHGoWIj1Z!3R|d|~f-l>=-=G4{|>=KRL9 z;}+=7)MhIzjWoI&0;kR(yL4XV<~^ z^nnA6DX*Llssl}#drR|KFWOthGc9sJmSe9+A68TM%%XdKh%B&v(Y!|a2Z*L~_z*wI zd`}^kFtc?d=kZgQgx7J-32m5QZP`C9zq)|EyN~(6IXL|-0{!+FK$F`sB zo5wjN<4){#+sVuNIb@gGwYhhc!29+y&-#Vl24M)upwj-b64z?Kt^e|RaD^#x9xrk9~k-;9O54@ zgNI@KG0DFPT~nwB-sZNgJ7%;eu}+HAkIu=G<&?LF+#n(6%uL>(jK2AIW?7Tb%T^UI ze_g<@bJ~OCYUE5HU+p%(on2MJn$g__m7eU>;`DG0_n0ofnNJc|E_k0*B~$j-46$6XOV0$T!}X8W=O)$W!0qz2;Z&)}~%?M?PnO4(K92 z;lm4F#Md>|YF>`Mi`bmHC@%JUHUFt>O1(XjnY<6nNS} zKCxSo7lf}&c#k=^^$(P15psd-T0NdKgmiS;(>|y_(n;*!R(PECrziIh>9dUe!Ck{X z?EU1?VtmoD)*aYeLW&oBnf05mIbj?Yc+oiW`Tjd_mka+m#s2GKJ^LBU9&7UaCg`hz zar&SYjn4;diT1cNetARc(-?~RjFg*`U5$GD4HoCMH>IsHIzU9Nce-UGW zR`+jo?m^lAlyeWtevfVPD(&}pYTAM4Ly7i&mKdXSTz<12{ME{x0UynZ(5B>*Qoc3b zV4^P{-+uCuM80nq$`^i%wb1Zl#nxb>F4|+STTnXAn%v~^Rf#Ud8}28s4f&l6UkDV0 zjqW=FE%{b)M=^D>?C^ZUgO)69>me^y$j(WvvKB5-pX9jUEDgUWpGlR*ZjYT^Nx$Ua zU~jo#w(uU}JXtkSuvE;Cpt~hE=xpUO*?dGwz9vQIA-aIWkU~4@R+&X%l}&cIlw;+xQN4C@9?0^(7UYH_;lqD zV&46qzUKh=H+P93TYiQ-(^_v9`_a%E&kRkcKk=3@JiuLhuFi7wJQdEJ-Pna#M^2va ze&UNHzaC|O>-MIEBuiX0F!1nkzE2f1PcyL*?GNDN_{YgpHq+|w;Jw~; zZ@#WDUS@Ic!vXU{Pk9b@NOy)j=eFy% zr#+y7{oT+$d%NgF^kVeW&`R?DJ`?yVksY)DPIwI4*?ZWTq3M!Db z%=09fQ@U#p^+xO0-SeeuNbXyG7SeyvMt7HxsgL|+U|t68G$a3R2L|nN_0U^_6G=!0}aNn`mPE`#jcc;h=A|RpjD9Yew(KgP(=K zaz8jLGkB`zE_SV{@!N`Dvx?@}zWHkVLTC>=hVG)6tN*i}*ZCqlbQ}F2hYt0OxQ`-3 zmz!)um&X=MzRe7E^S?`tvP?<#?5B&?V5t^plbuw=#3kaV2z0`x)|5j%Bh6rEVW>#vgRv&No0Uvs2(M$*Mh z@KFi>2-;I;N5Dr1a*f}}Fzo-x7x)P}^q810S9fG+;ebOU2kx;lMB6T%ibh@f6h9Dc zj-;Kn|KZHn`@)1QDwu@hJD6Yg^xX4}L-U%SZOQMPJ-?&*e0S!?kw3fHv&jjH+?=9Z ze()P#>kR5;o^|i7b7OSKv&CF~it{RaL%L#CJ4Ba*aX4$%<=g2QqDA%lesMRhe~;rQ z}sCZy}%k2_JJM8LuZHf3@&kSRc(>fsa6?$6BvFWw4!5^QJrP5@*nQ61$an zs&Y6hPx^dpsc)gv{sw-Z>)|C^`Mg8>+B+khzu^C!KE@Fl>-@in|Fy40$nDEHu4bBd z>sIvJ=8PenTql^BTl6gD@CL19n|fo;ocph^aMM>zhqa#k)3GglGOf%;mA~MK0 zH|=-&65|dd$96F~%4qd}aOTG_vomVi9Rp6YxhbyZF za*Hdb+`<3lHvjkO|G&w;qn`gmlueuL#8nl4^nyFI#|fW7#bTWe)bxPkr7D9>RdZJy zOuM6mXYsT9*-z|h-62t>_ZQbKe}g>J!7TO@4u8J zvIha9WHo(1N?z;XxJ|m>WbmE{$JWbUfNX4sr2E;$*o5D}i^%m)H051Fd*jDiTf)kl zk$C{0W{WS!WLK za##~}_%+-`8+U>e&8Lq%EfMw#i+tBwCqBwF6X(Au#agERbjI0Fc@MVzdU)h9WTvC= zv~RO#$@i%gytc5Ggj>bS%(u;czcBC`IHRr)ID}4!z#E6Dc!fp6o z7#)s}-vMwezD&JXvftAi_?YVbQfQ;jy7~pb|Czktj?}5*QN9#pGja%mJL>w+sf!(| zQ+Cj%F-|_kKe8sG#QHMcSn9gUeP#7&t)FC$)NaK8o2{&mWoeI2+{PY#WJ3LfM%9z= zSA?ABY1^$r^%;bY%J@X!(@wt=&gqvt|JBhPtNZ~wt&H4c)#OSJ^@@Jk`_*?T-&fGT zeox~26?;i0Z;3nt|i zcqYC5)4uGbWkzobtmWyAc|M-!*jZ)L2en-CQeH>F3wc){v+kYtLf!y!*FoBkk9XRd zNn4Q_FXd(1*&*@W?*Ye*34IHm1|Q$I#-+VtjjwK|dlVEbCV;HD!1Ls|j`?vyk> zw|}C};=~;N(f{<;b|{AI{Ca&9wW- zec>ZNqFmL6iQC@PyLx}Ce1A0VO3U9^jL$Z9&BIHvXB4tVCbD0OW^|Y6rPwXv)YVu+ zjH~o=+CYv=gC64cm^A2_@!r1gPt1?zt<;`SRz0I}t8d+9`v&ZB)zg{7+L90TYHyTH z^mTARyPeI{#rAnb^-KBqX=CQP$UfsoK)qne%*IrEQe!IZ+(2Klfe6Mo;I$pp%dF3A z{2uuJHJ{I&9AK*tw>1L8ac^3Pu~mB*H#BD84y-HOr}xknF#4(&Ge-Pfi9NFBiq0y{ z7%KNwvA*;i-vQ${J#rnwKReHTro7|lw_o+ZBL+5Ird(zr{FKSjE?a@O_3f9vEnTbe zxk6TXiu`N3b9n*y@ugcC`u zK71m`-5p)Khz;{tZ4Xes^32*Jh3H;ue-r*og=l0ZK8RNP{8DUSp2ug!UV+~!e|s_a z!5l#DjZjYgnzh99BO~%(N9jj8@W^{Z`L847wtks*qz96YMtUGi-xCr(iuf>e^S|k{ zU^eGC*=nY?{^})o=SFMZ9NL7}XPNi*?3iQ$^SgUi?3DZaz^uV~Pf#yhXYT!6{ueK> zoC!Y!-%gL??_$}veJAzo+9N&4T+->REDZmp$7>ah_jGj%_dfEV_jtcbUVC4o#&|uS zDDh)U3$2WD+n4qoaCSla|GF5RRL1S^RUGMOpX%T3FZ?uk_IuV{93*}wus(netRb%n z-g1gHt$s@R%%q)O&dE)k7u2lAPb5C)V%ZxzHLS<-AJOKs__{Q~m+O#OqO5Cpz_I{7K^viw+UzakE+Xc;{|4;x z)O(Y)sj(H`65OXfI{lfoBAv50|b)Z%Y$& zC3svq6fH#(bSSzhT_4SKo)zvZRDIfV>FbRIeTjZL{t=lLnPBaGV^(au@4h4Ap5HZQDuQN!J=*Y~_c?ortiO zx+(YbA961a;}Cy{rhhw6>rm@bYw^4|fp`9Yijxf=3@6Rt?1NzNEw^`E+OsB~|t#X<84^D$dJjBF^Hbf&)+7x}vV;nx}xH$GR?hC+iDR|>~ z+Yf#|gy%W^19NL>y9;~!%M*siG&G*f^N#0P=CPRZ4LmufEzWw=?_T=XxK_`)f2;YQ z-@@~)iT^(|>)|~1teV#{8j$PB1;gjN*y=ui^B4DVr}ey7Y+v;O=)C7-POJl*6yI{i zzBujAVmy)+-}FvYK8@v{!&hKGc(i0oQ60t}r%vFO%lE~tiD~R3={?lD{^r~Eb@P1Y z&0pB}I(}UP)|l$(hVO6lmt5Hx{n*&4A!I=OHq(wz#pm_r%lG{W-Ld4u_>3#ydpUsUL{0k^cti zdMERdOxLW<^f>hT(zK*@;isP<9y)5)%LGGT_Yba@oWb={SU|h%*=D^29C^U37oM4R z;j!p2va~ki(Bc_yc3Ej{^R~(0KbLXe?eL4|JD~f4t>PCC|F`(XN5FZ4Uzqr@k^JH_ z%+YB4!rg1W2hECKRKPE88p$uf!RY*AX#Cj6hQ=>`K^=GeIp9@$Oe1|rUUB=>nu_ua zx!drtx8WBmHz&a_j!LH@enDH5GyI}(1g)Kgf21VQ+BR%yZ{N10a77ZYcqy4z431?a zuNWH3bB0&sB=L$IhgVS7;T8PP?;&21Q~jY?-!#19SMUmrPrTxn!+1qbb+`O#SuZbv zj{v;lZPvVebdq_+mxu9+oc|WD==jnP;T1C*J1!d!uecIkG3gv$@pgh&+>qcE%4@bC zpW@MYMK7>c!Yg#%)jm-LztGuJZHQlxQ`qo}eB>UNU(~}dG>1$+h-{PBm*U0J4-8cj6-$&U; z^sFA)LG+70bW8VYU#tte$VcD7CJVm{1kty#PZpxndxL$olCzn7ve%Hi<=K=W9;E&4 zlkgyC4Ne@f$BiH2L5%r)d$!BlqB;KJEcVjiHWnKX5EljD*hHzIH1e zKdU0ah5vZ}arFG;>Tlaat`eOMr9*h|p{Bf4Pik5my*vBa;o_1|uk;Otv!Yw)+Ea}W z^E|%s$1DFMI3s87(i>o_{*ve z5D(6soQ*u!`A_=l*R7^Jo&O@_o@0G1+mEc>&A#01$vmX?$enhgka-Lqh;GfaGYw9< zPZJ;F#dePlSF*-Q;$h>P!baySJx3#XPNg44XA;{9?|VOgv(bDDiua&bcxEeUB`oW? zXPo0 zc)9NSl1^iP3%G@k>P{ocxtoC<8|U6{C)$(Vq=`7#jo@bs&!ivEMBbZ89~!UDzOLPb z{bNfeA2-%)v^X>AmX$xR$z#>X9?qQ?vO|!cIfZeJ36U48ydFOa#hU#H9whw}IIq!j z=bomV))t=2p1p^)FFzysG0T@nJ)1F3#{Qn$dW`!ZWt(pFWL8Hu z1Ty{W0~w(S&<10+?ECYi9~0a|b|}Zk+ss2!d@Z`R=SxrA&?Z|$0sToPw5*MtOHNE` z`x1U|73jC5Z~Gp`?jy>EH;cYzwy&M3heCD$Qy9n#B?FnMTDHl6Q z`S+t;a&hjtg-@~1Iw9R=kbEw}{X49I$(1?VPEGg4PW{Phksfw|VpF|&|DSR~G&Fk%%VwE59SK8vD}U?dBT6CHQ!tO6843fwNGu?dW0Bu zl~-QEUgr2XHqIvgM_w!cs%P>c**-WQ9d01@Dq0y@xP({>bfSAFdvgw{&nR}lUUbrr zf%DR{B}XzhNvWCIsw)H`9a1uC98GL2-w$QJG%PbQ! zFuWfpr=E-V=fJgWOD^t%)ctGaVPX%SX6^089=zX^)r%}&9;sElfAZXtZ!j^pKctWS z@E9*=1!qjz#4}_Y{M_!zKJUB|W6--LWAj$g*)#B=cd!pOd2)Mq(%)(N9-wdJ1N6es zAu+zg`^Q&?{@dRhm^*>-8JJC-Jab3Nx$rvWyC|=I{NBt%y`KuV#65Y3(1G?IqMw!C zT;ub7H|qg9jNLokiv9E-tv;;>_N{)=u3$ypDnA4NEdpje)Bd!9!Hlymrkvv0uf_ipD?!n)UXu{tw={`_3=e!>~DEv&0X~i%kX_U2(m20^6&xn-;?h0`*ReXQ34? zsszS2D0eyKx$ojg&~If_R^Gm_6}=gG=0b%Rh8IREFI$LBX`Vk(ekNsAM`a7CQ$^Vz zb&#p5>Vex&9hDdF2bS0lU~>1tipOMcwJG)^l?>Cg|+K+ENav=N8Va==VZ5Mqsafe4Ua7Apz-HkW- zX_r7g5fAxE?eIfY9{(YH2oA@0CGItWrjv6!Der*m5BPrj%Sz#%J^c(Yo%E)Jt~GRS z)v#~3Dkc~oLH2j-M&7X@-9!E+Fgg2pcI&^8b20h8#k(0t6dKN*fR15J82?Umnc&^c z$0fO|gt%1QBdNVz{=_B62Oj4Bm(V)uincd_17B_JX1~dO;N;|LAjj5rbXxD1murc} zL0sTljAJc(V*|D*`Br=-;g|UWbgR5vQJ!xhM~UP|`KY{*rhKfEkvWuy)qxASfE-=9 zp*&z~C+~i;eBzG9i9c1|mMO?5Q(HG9FUa5QM)cN2p1+wtpYQ#ioa$$h!>?gYC%2)z zUnA>pO4Pp-`+O0wHQzTJ2Y+q+{>1oCGX6D;-_1+*amFuM=SX7w$RES|8B>edP4}0GNS5?vOJ!=EHLDcOX zuWav#qIk*33vNIS6cG`X%7P7S1?A=j6nw4FZm4CLW@Z=Bte4D+N(+0tU|L~WAzD(I zqGDlMQCVT7e*b69Gkb4#TAlAb=XcKc`_Aa%zyI^hJTvpmWvw+cYc2R)>78UTm`A`f z7jft_JMr$RX74O3!nb_OPASLNSKx-xER~9)&ME&zh-|wPvMvaSi;~f~Zw}kpl z9qyIXc%8C{`|ISu@d)ogXw&fTf86Sxo?@@r#O^GCay;)9)&eLv94Pqho$!!wMIHdfjJO72Sew%4&Y^M zM-zT`fQkr&$UT#kH@ALSQ&QQ~NQ49|VOjPqZd%Fi8$OXaj6ZOTXQm-YFk zMx#}>Gpchl=B<9gw{TQ?vK^}Tjmz)5oQ9n4N59e0vRkG>?`-T8xBzL;Ju21pmG0|Q zdwuSo@*GC(-554#?d^Eq)dExTCcRyH55;?f#rH$O2?J0U*YoeTrp=qMw|JKsdl_#+ z86#WcySf;&-p4r?^K5u8eaN-bF-D!mdOEIi7S*-fylDo)rFnP9o!Cb2K@B>3hjEf_uu2O=mehgyBB^9IOu$c^ZC>OTz}x+bH-q;mCrZWqlEn* z9i!V>u4|1o^Qju%9dLo4!Nj^{>BC`5_hN1QP0e@V3LRUXJ!8Ia)4Q7XKFZkf9lVF) zEA0QN^6Z9u85UV_-BgABhpoD{3cM`5&wAjxgZA3cb!wS5XzEmyC)!&S(q5&g)}j#m z|GLm#$28h!fW40C)Skj7F6sj`;+YW|-@L_g?1e0iJM)AXRO~GdVt>QhfWowj+c76~ z3VL1bl&I0W*Y4Y-?`_2I#i-jzmwlr&{zakA)EMOcK0kbWFEFgYw>_>Uea;w!AFj_A zP`iH#b%FjCN7o71z_>mM=dwiTR^#8#XoIv*kmd<6R_zStFx1zBXgokXN>kl?*Ztmz z#}zRCxy%pHx$-dPu1L25&-UTEjK-U0@!aRWYRD|oPuaOtZJC*-R`rC+kk9%MGuBK{G^aFGcq@li#<6HDe$WNU|o9*ij zqT{!%n_ljZ>vd0z8GT9D<;yjZbY9tl@AuV2eJ{%l7TX%db?m5(%_dY(J3FtHQa%WXkH*&Xm+Eua%hFBP6F+4)8G?Z{T=f@SPW*m3F2TPO_>u4g z|E90={ktK4T&r7q6;8d`|AzUu_$T4OePTx=_T?PcjAv=S*|@Z>#^f#LB469vLI3@D zFZ}Q^IlYD&<~>e!qcwx753&x8Fp9aYF#prel+mkdSLUHxTZp+C*t>HZaGjRiYya&< zH4U&?Kv48cCg|CJA@9&&t?ktGi*08VV}3RSVOne2LD;U`7&c;UzK^|E`@+oMb-sMe z+KyxLd&PP|*X$Yovw(|mJlrj}S01qY)Oi1EQLdlEXZd$RoI{i6`2XxF=H~e<>GjaO z8~lF>%BpXhu?Yi;fEhyiD8pmwTHx_3CC`^TlWvU z{?oDH2sek+Ko8_~;X}WiFs|^L@2>kbD^z&;>F?|`P8|EOPu;q>ws=;e;np@YTWX!# zH+cEjOJRS1EmnV4&Rc9pSgS+nWeS zVb0K8@U1==cfIkfQTO%>fK#=Wfk|4MK(b)~^8YQ?W{+$Bc5kHf9^R3o(k+J0_qCpm zbfS>T|Jv}>PuazI$#gpxVXApI}_%8DH zE3G-*VWmECR=%188=iDwmSrt0af+(db!Ji~^1-Xc*}$~Wq3JLPQ^c)A{T)uzjoH_ZL3 zyi=YAar#M@FGDN@&RyFBFxQB&v8D>wgY-T0BG_2q9BNEE6H<(M0LriHc_A0`Hc6Kz zUhp^Kc`~$H>bGJKo~uH~2-qX9dFZ}~?txNq(ArNjP{Z@4bS;#0$@2nTFX0+Ozm9hu z=hL8DT!e04@j@B$VWjIA$1_|0pHN;X-;Vis{W)1VH>8IR^H@vQmAL;dsY4#=`kSAZ z#kvxnm&5$0e18DuACVWF;{w;=+XZ@DmuJlAJ&$`ZCXRuOzSQo%8p`Wu7yA(4QjlO(cKO#Xc2#9@FsbsK4lJUx43{*bC|}I8o%4={zgw9J6%IuX}qCx)*dJTX04|qUz@cac)qx}TeJwlR;|_}RJ~H$ z7wDtV{Q$ndvnl}XmF`#7^{o-lTqok&ZwRZ#(mIc}fxB_8KjSe3&)y23GR@Z~ z24fBS??N-K6RTh|pG?YMml4Mu`RI(_Du12uTjjeWehbcbJN#~~=ex?wkL~uy@krXu zhwbLWcJm%VHaqTRw4Z^ksD2tWZyyWtWyE-Z^6BHjYUEp8|M`0u^>d5+dKuksy@Wk+ zbPx8i@Eg5TSVrSIre+(iYjT(NXx{<#^sz9(KXSAM{*UYaG^UWhfPd`hBKU9A{rxc} zlD`rD38VAz46(pCffwdLSFiTFz}MqBWML?(xmWk=-k#R2)P1+>vnv;`@crn9f#Tvt zTobR}hI^5{hJ%Ghk$`t%T@0`!^}_s?{yiF+8xXJo-pPRHcj!4>nqO!Un2hvt4D%ny z^IdpH7TsHR4s67};n!fJI$Y@72Z1pY?{%^O*Lz7AhZAwV zSBEy8i0i%H@YnNz`(AG^F@n~s0?{^T9G(KCdJcdr(o2lQ`qgf_w|0cj$F~gNcMf5S zDSjtE-S&fny0_P#&4o-4+nd`&Z#;j2uxa44(I>n!ySGn=ABAZ&SYD59Ctla(RvA9s z%JGRKY!ves5O&54@j8t`bWb19+J9sk^iW=*=VGw8xTvSgdMF<0rLbx6)5F7$!b8`^ z;e*7*8&Mzn`EpMS`R*R{>ploi_q^2q=-jRAMEfa2dndbkUs&=0?$<9w>VAIgx7fwc z#D4iMen$4obNCgb9Y}q-(oaKqy@cJ2x|u-xdr)Vzhq@AVhW@sQu7ULMUJG8N`45z< zhIoM2f|oGItUuQn1wVS`QV;i2nyz;PuD4=mSYF3m-0K(=I?fnmx%h$sV<)}C!SM{K zJojT9ue=|}Ivi_ptie%&L!B2fHt220h-aQ?PhA7@b`$0nBXR#px)bm_0KbixUsC59 z0jJKlKFD7nUmKFYI=8k#pFQKzJ1|IVvue;X%ON8jn)j+_a4qSL@jG4fe$N~6=pLEI zGScURH7m@kY#;2RPo<;AJL5RZZmg@v`BVn1jv8bsjK+KuzJ` znqV1p_6d|FlIlS7ZbkeQIM#)*ci4BKD8)M&wurS)V15tRAvH?BT$hs9g*1;5fBAAr zkwL8ONb?!IzC!25V)$7N#`j6~InFVV?SdZty&=YacqaPxG_*C*p86m8>vJY9-!aB~ zV&T(KtK(?y!isjFmuq{mEEn=jd*`=M{)gk+FNp6SWL#tcCtZ|RRexzc(0&nqF#2#n z)xxF)upzG3-w)KZ)t27kU}Y)hhB+8nKN zywSJkOkd3975A^&)NYeGW`vJ9CMwhzgL{FmQO5YFc9oyeKC8%3zZh@)9;9jQq6~On z39h}8T1T$N^>atx##$50KstYer>}K0A&fh(DclXnxmHxhZc`kk?3Hkmj?h6%Mrf zjsd;#4%$UMXHMzAAJ6->T@|hQuA;pi^nB*sZT06}b}dBxV=Z{sM_6MkraU4~RQKEQ z8?v40KZgfjtSO_tzD0G>(GudX^o-BTS>o0*_t-#6kw|2iMv;?>^5up0I_YZu~ugaNZc{Y8ISZnvL9Dfeqp3Fp>$yzSfc3Lw!D;Vu}F!;6jJq7bb z@jyQuyL-J7Q_`yWebvpzTdT`)NnIRQvNnILP~8 z{+IUd?8RC{4%+sW%2bN_q%tM7E?#{Nm49q`RvM*+`q&A5#rT~DyOpEv==^-nqxJSi z{PxE05|tk0=P)Lgz@O^l9G(j{wKB&TIWE~x-{(l}7QdDuikE3WSpb>eur*2zeNZZ#52utV2g}AqVdjCtJLLHM0qq zHWw3`FC8z|$AA9D1eKe@q~vDSnZ4F^?Tw=k4rQ}aT+fWceu=Buv<~BWAl~b8HF+7{ z_fd?oi1aMOehLalzy1G2XBp%Ic07q^TCS!y3G;Tof0i!m@c5H3)E+3Eez-@vS~_Kz zzo0PR;~QmH3sZ{qfLHz`uedLbJMky+xx5!X!!y2DOQQ_?hbbQ){fVAZ)Wy4hA}@mc z%|DR~$Y1#rIrRb5<)|ctd-l)5!5)#g7B9oP$TL*`xE4y+C(?U_=-onr zco$F93A{%~y+0Ap@kYip#sbC?#;uGmGwx&jgt4CS2gY}X%k*AmEMt6>aSdYu zV=7}DV+dm}Mo-3lSllCbjPX39&rs>#k8uoR0%Hc_QpQz`8yL4SzR36*<6*{gj6QfE zj9eeaQH&PGIL2ET?TmLbKFV0m_!eUg<1xms8QWuFlUy&xA&g@fCo;}vOk*ryT*bJ7 zaXaIyj0YG$VLZwBEu)?|{@l%W-oUtnQOTp3k6|3Z*qL!IV+Ny&*Em2}G(~{MQO&s$ zg^ZN`jf|BuByY;s^!UYnC93cWOLxd{4a=pxDelVpRd|J_DF$6{Iio4OdA#!NQlF6{ zR5Hi>neM9G=>sUO!qy7ObM;`d{Ybu--syU8tXJeIl2UjlQL zej0OJck8aeE&dATEEab%$2FerN|-DC8=0%ReU!Ou%|evA@ojG5%iQEU+<3VguVAj~ zW1pM6in+?q0XKQI8$aw8zSfN&b>nr+aV@O7dN+B48$aj98<{KnoOhEqxygTblZzd) zy{Pz}%vE`fZrs<6o801ebmM+*yoVb%yYYT*Jiv_)VXo{I?8ZmAg%5G#EG zIermyJQuIKQs!pnRm}aFH!vT--1MAGPsI;r-j?MS=F0vPnd3f3cd>5albEag-N{^y zKPAkS{Y#m*VSVMyReq|OtMWH8@6X}=o|pN@UPRqlm@9c2^T90N$Xw~KU~Xo49rHfS zjpb5*0P_&$fy~pGV?UhkikYkNcq8+6EH7uS#!vV0veGSlEpw&6iMeWDrWa)XlzoDk zcjffl^}CO6?&Fht|8^e_+{Y*P_8-gXcjELD6O*mEwp3+4k(ij0YD>ug=2&x6tQi~( zJU1^DTzMoW=B1@m0QN{soM+9=bLER|DS6p;sy2 zM>u#+o_+Kvq;I$7=GoIzFwsLUF)=^Onvp&~%a*EhC8dIxvuv0al+K)L%dq9y#CXt9 z?3NifPKe(4JsuvtaolvX440i_vs;m0VnGN;?Po9=Wl#_f`i+D7WKd|J7zU*F3sqtq zbkHW8P~Z??z#t)R1KxsT8IA#nvt*2*)rS{c(gM(l{s>fZxk-+=3%%@w_~1 za)xan;*v{Rob1NQq60-zUSW&+x5IAQCqOuuO6TEyGBpX zlQyR;IMa7%cNIJ2GW%E4LVD8edHJwJYI<%?hBd{OY0JtpXIgVqIc;g?zsQQTyEB~A zR%U0GlAQ8C%~DcIc242dj7H_oPt7yiZF%|jEWMhNCPj**xcK-eJr|J6dUdpl=KRWv zsr{>VEi1&Ps7T7r%t|*?8_G?mR*~d1lI$t2l}h<@m)~@Avk*y{)-3CMTk2mK)jix5 zjY{b`?fM7HIj!F;oU*^0on1pCd}?8qH4|MUJu5vg9V5q5o83Gw+iq6vLdmUpwu1CL zm*(#p3VIxk=TDP!g!@~WqjLYjj(?%QS%1$Gf5+;|$rFm}dcTrp3y<)+xvbDYD2Fo)By5s_id&2^EQt$Aj7>Ts2Z+3nUswkt=8 zOh;F@r=%6SMAk!?BeFAdXoNOTPhNmChB+GJm`l1z)~xKT!p!XaTr=v=mYahv&8qeM z(jU5ZF*0*ScYnHuR&oyR0Cjy$9&{~5;oRdZPS@s255-gWPU@PAko>NM6p#Gf5nj4K zp{rBZbqZCwDlNqc)o&Hwm8(1xw#1>zsPqt$ziS*&-I2d?6i>w`Kb0?6J>>72AJ_OK zBS-llr}CxJRqszY6!icy$W|h%AvWGVl)#(gQ1aH^kqODZx&u1FEt!Y*w1H&mqR zHV`5r+dc{BuULC_iY+%+h~O8S&v%u$@wCsSyo^zd;qZ((HIkbSNemb&?`IMiIXj|` z_12!_aNkI*u9oO`P@-jxM3s)8pVX)HI8fz-J(6DHcq)M69Ik-Vk5uJje;S`VSHeUR|py($hOoc$>vl?gycs^^J^rdwxndFm*NuA{W$40;UGV{pQL+0 zI>^5_4)Vu3tOBw4J&9AYXv6QiqG4^#SFJ)|GG`=DI&5Xf}1&mf$ zEFE`r)PI7+68z2rs-6^teq%GB2cW_x+%A0G1nA=9Vz@( zd2?VX%CQZ)a-kZ{ge&$N(+S$vhFzm8uBYo#C`Ds{4$iv8F1 zmn#a9b81!dP%c*;(&~SezHE_pltRU&miPD5FVL+h+c0U6_3hMf<@6~%XCKG{%ZDCn z#j1~_qBKjckf$r_qqr&1K{b&Fiv&R%*@fCkCM=V$+nD+owaw-=?JO%bBAX~cEcgD} zJcp^!LoGgA%txEdMvN=l?(gc8bs$Sjc_wQ%??ui!p_aNBA!t+>>C#gd>ekD`-;8Gd zlyh~iru9)*jx6RaNk0uoe4SwW5IU_eL1NWZzsXG_{$#$E~HQe>KyxG=_e4Bf? znXdkHP9?ojj@$AGJ8sZ1K`|3|gX!5>NShp;w<%n7#58M;BUdi|H2r0Z!7XBZc63D4 zM2gcMM_5iyY<5O^N+EK?7e(4uWPsNJ$m-?H}~$- zw_pDO0|N#H4jvLTbXf545hF)kGy2+)>&A@5oYw@4H5m&Y^X8|eFIbq7nU$Th$ex>* zzj#SO;nJJt-h4}9($(evSa$1exBqvS|Gzu^f7||H;Sm!jO};)dYRc5;X))7dXT;5n zpEY|S{@yZ?Y{}Otac?A7Tdmr};nV&~K`s2f2lR{`~x0UeJTiJX>b6%}xS1=+7pGG+ezY zJqHh)CPXX_a?_>_OHCbCsOu%33OB?&4bDtI3gO6GL`IIu%p5~A%+k%&?@tk5zoUoU zz$ue+@Rx3|;xVI=51g#e2?#M$zpEys`&ycHlX_)*CzpD19U-JXy03QnQ4XYjnssx= zC4ZTo(@&;LGZRif3O5Bg$kM~xFo!_pup$g>i~Ddw;*`TFTuwr&i!k7P;7sHw2RI$~ zkyI5iu)|{DMEFz9O8xn|R?;&W{>jkn^noXwOqb@Lq?75V>uHgXOkdtpE1to&BQEnp zvjMU}$b89p7dbCUp;SI;4v%JeRDNl8M)~N$IRDN4!;njIv5fS48vYG~O;ho=&?P_Q zL;r^0Oiy9xpV{FOx5Sn$w$iL(GCS{y-)^n`K!j=#xLY}L37;XnwZnPgYJ$=Ii|gJXJqch+>g0h ze>5|v`3-Ub%xUd_TrhKQNks^An!6`wVNP@HzS7^k78cVdk@EzDaok7bVMTXmPj9M2BwPF>HmVxGftU*<*3TQe_V z-iCQ8^R~>(nVXnbGH=H`iSyr{c@gst%&R#(K5nYJI_902H!|-UmU&O+Nz8jO&tZ=D59qFlxtVzh^WMx$nfGB{&b%-4O6L8TS2OR= zypH(*=8en;G8ZRh`3EsKF%M*JW%*QZqU_O?46Z25!##6HV6PWukw=fT29>zR`c?9!F z<`bDGF`vXdhxugYMa-{fUcx+zc`5TL%*&ZiWnRfVnt3(z80K}%r!#M49?M+R%ksuC zH!+{h+{}Cq^I+yTGPf|dF^^@wka-gG4CXn^moqP7uJHv(33E^8IozLHGB0JhH}i7l zKFlkb`!cU)Zem`?yaV$_<{g=f)6yQDnVXp7V{*DPGw;bfg!vHWk<7<3PhuX*JcoG% z^CIR8nQvsS@dZ^Gb5G_K%v&)V%r(Bit7PuUyqbAS=5@@y znKv@`VJ^Os<@IH5V&09pnfX}eAwq>9fxLB(gD zrs6X%Q1Js~{9+ZK`9>9=d6^17M24?W;h9&d@XTvf_+S~nL4{}Dq{1^do{{C(M#=Dg z%)OZhF!x~|qQZyB@R3T+JVD9FNqL%*GcQnbi9o2nlB;UEp)n&}5u$Gb8gZzNow^eQIYxU@APd zH{vRNYIl?$6`tB3*-w?9+F=l^sr2W<$|P6zpmrIA+;X=@>ZC8;p^w@ry9CKDF;4__dL(o^N5b{~Y$ z>OPj*Kb4QCUlDq#&TquuXs|qPchoOj+YR-PAjDGTqHsZIMXKEu=yq}PrFwiPr+(!c zp88i1TB$0R+`Ts0QE$}Ws2nQanR+>=Im$^pzFgad++R7(VMpqpG%u&xIrUTGs$Wol zC9cw^eoOV@ZU^eeuI(pTFSj#a)W2QxN&VcFQ-5#HW&cp+qW7PbiGAnX-zXE`24i5Z-}98V$~{YB2NI`c>8A4)BO!^`m`+)*!bd~n8>9#|$#&-K|FWHhJM8V+Zd~IjyG_DaOHSHllB4~} z^L3=7KS?`udM&@g# zBcC!~(JuL*_U24i=IeS#zE!@|_$bHQaF={$;@ZL4b7lGy9qmV^Kg}@?sPvuVA*JtZ zZ!-M|$9N~xcjlM&3{&e3!(Z3&Ef;58%7?01nU86Xd`kOAIohF2C)OdC=}hC=rE#0u z*_G3w5kR$2nNGZ;{m68j{Yj=X(NTU^z0K=2Qy&SM=YKkSl5|q<6cp`Qhch3=JC}iLf2I!4R_S5vY#4nWqMN_<&@#$9qmYl4|CM3 z3?J>1AJTuLzRp4=mf=-zQ~hnKi=4v8x{O0q`UtcM>7@QJm-3T*mZM)tdAM8rSU33$ zNBfZZQKP2H-%Q7OTgoF{+5x3Udp)I-@>oato!9sBOi%5CoSG+A<9#g8BM*^kgr!N| zC&{a8d|ZjwnOH1hSzgN?Nz6ZEp2PeF=0(g)nU^rXpLr?skC>M;|Co6t^BU&W%%5gn z$9y;QM&|pNiwm+mN12Nd zSe~NfY+sFe3Cq>>Un%ojSzgZk73PB1fz-NS70Xq>@ng9shp%P1x~^$pehr_OPD7Z##PUeyg+zUn$Lf#s`N9?a>tWS+)y zwJsCFawE$NSpEd_V&)$(-^hFi^D^f9nO87>fO!@3SDDu`f17y&^PS9_n190D_?xs> zHFH1a>N+`qc_qt3m>*(p;qrMik7W5l<_XNdV4lYO2=fBw?=dfCzKi)r=6jfzF+a?_ zg84S)Rm`7dUd#MA^9JVYm^U%6V{ZIi+T&l${g{8qJb?KL<{`{KXCBG?81n??pE6Hl z{tj~^x39j;3s`PrZeh7v2Q6lK2FuMXAHaMg%hQ-ga{c%)FJpN$^CXTxlz9cqbD76- z{8r4XSiYWlE%PUtH!y#Wc@uMWUN>Hp_I;V0qZj{Z(zB)-%jBC2C=+} zhlem=|&S?U<`|RH@_x+IST1MHP(J2sSRTpY2Qx2bc_#CX%T)}iTPsYMj`h>Jj%S1 z)9cFIkL5Qp7o1)n<^e2EWgf!(4d#)|_cBjlUd}v?c?I(V=I=8vX8tboD$ZXw<{Mc) zm${n9KgGO^>sWd1tyTINqNZ)E;3^J-47J9B&!%A(=-Cgy(3Z)RT1_8G}MfaOb> zH?rKqJcQ+|n3uCWgn1;(iqvGjCx1y_i?9Jc)S;>mS9uisg?puVelX<^ddjAoC`cr!zMi zWc_3__has%;xj+VJcM}(^GN1fm?toQk$D>P?aT|9Pi9`sJcqfQm!&7SrIYu$R@_g! z@?^~Jy7E+BMEyMPFJpGxmEgwF6E`!PB|Nx;xZ2?edON^$dv|KhYCXOm43PYqE_cS%{@zJ*$w3QzMf^mMl?PeGkE=kD_6 z@#&dv*YI+0Hu=c(ob@gFd`JCBp5|ybl4s&sb60&?j{1{wwc4ok=eXFD$|LudJJZi` zv|AaUX2o6g=ex-lyXd2Nc6t|#3Qy}Eq(wTp4wde(r(8EwtIV{{OzU)NbycoAlf9Kb zIlt;`KZ>hyNOJnC(xds;3`cvE@=UCEN+;JrocWXMG|ql0*Dam<-sSp}tDN*xzm-nT zYpYdkS|6tJtJQMi)L*TR_AA!`)#|lePp5G~*-Nf-IQy&Q^jA8$4}F2Y=bwD!`q_Mk zz2th=LP!6Q>q^f2t92jcLp%stl>OxT*h0rRF4q;+DmtyF)4HQtU8nVGTAy?FN4Y+z zR?$gL>rN^);xrD@U+Lt!maAO$H>aF}IyA9<21K0K=*NN1w1d>zw zuHi|)b3Bmic+ULDb!N4Tf!=#V?VX;WkWO;7y8#~*<#jSOVkoY9yyB`y%5`eBD?;vz zkZ)a(`sF&GGe2@2$Ju`*ca}$T`m5s8`ls4mq1Kh@O(9Ay*AJ8in$M?mlG32oM_lE! zUg9bzi>e%{^#muUb#!`5hDuNMAA0)@?O~7-xxVqo^P%+5h25_dUi!P9KjnI!N}b9} z?N#NLIIZ{6TQXFBC=V1up1I{Zt<&D}4KuEs*8SD46xzQ*rFXU;xetK;N+;Ki)ovI0 zEJzUAh#EiSK7ky^`CG0d%D=AmarJjCuUyw(j9f@3*WaD_k?RFCp2!i#$z=~xT*a3> z&5<9;>5U-LNuKK%A0*FnoImAyzcYP~CPaO!c!N$%DM7cqi_eqk&hqs)d6TOj9kSV% zsP(7a&%C*9?%wYQrO502)Lj=xdF`vYXWj!BI*rYkDC3wuEwQa15}tfSo787hmjMH( z(><%j@LlTd-rDEsPsf|ay>9>U^xC{mH&nKMs^__kPXkSdzJxZ@UA=CyPR~jY{D)tU zld+zIM}0Bu+kvqu4=>vPROd5O)&vYScFE3)>Xy}&tZf-F!1AnTti9KTj}NXN^sf(Y znzv!-*sY)6`m1fn%wtO)*=jDIdEw1>p6S1S>a>yPo&W-#gpvzkbZ>(cX_3 z({o;b!?xIbcHe+U9&4E2Da*UlV}pMfv8cn0$TvRRdU{ga?Q}P6`d7jFUksIb-R~7Q zH64J{`DVo4n}eDnjqSeeaYN~^0gdNBnDFG#uBD4UdF`Hw>9=ZG7fg@M z?pBz;?S3InytsQ@e({QRohN*-aLmq?$L@LDIC;{vG~cH_zqoc?zvCUwr3DY1xbxv} z!W`F+u}nWX;h78BOQsF@@_8R>Wv1b?p7E|p@O|q2 z0oK{~+|gsfp7ZzQFK(DP>#EN`I2LnvW!kJQH>bb2a$NgMrp!F$k)}BhJ>zrfIqe@$4VjYu;G4Y$_1O3B_L`4t z7r#6Dpzo~D)|EW=?xwQ$r}p<-+}PT?xa^lalmD3C+eMy!c1_9+rmx;-XqU0KfB5PP`yX4|D);$i zm%Z)^iYQOd9~?B~&N)w+zVpZ1{+7ObI=P}@L2<3=^u`aW-k!D6H0$LXst>>MTf*=a z6&_tDMSEPge_Gca;%)zvmC=P;iuR8^81%xKraqasY+4!ekJ_$B1O~ z_4WGDA;gw5hv#Nw-&+63SD%i(uq8Wr$&68}-)XpE@Uetr<338bJh@M&Wpm1I$p31< z%=t^Yq`dOw*+idQ#Ah*E?$7<)yZ+H{r_XP_xzp31?ArBnn++kkYu}1Gd*GqLi4RT)Y5Vg_ zYt0?6EjV~Q{NSEreL_3UIsW{r#M3`5S^eHuPuvtY%wzoYuflzxSD3yyALl*k<`dyN zpBp;rr6;nV_+UcT(tw@6EZH-)x~gV=T~pa}^T+Og_NCcX8Lb|B>_M?7@}}{t-k5TJ z^rvyt8a`k2cJCe+yq8a!vY>YRk!8^*jE_F~NWLf@S1;%EEDszueed~(J=M>5&**!8 zd#0e{pyy}&e%(W7`*#W+9Nnso$Ck0xaq9HYVrY-?n$x z&!2nwiJ>ocjNN|UGuM3`K5D~5Pai+7T{`q*O_8Cn^jc(Hd*spB@1H&T?vQ84hg>`D zo!sjAnP0akdt^q_4<7rro`@?Q)7Nz5>D;>}zWQnVk39VD$$4bV(YS`V7f;9h)}`gJ zr>4De;>a^T)4xf+_aEnWf7shH@36R|)0QvuPaXJS(6Bp~lWpp|EJ^+Cr$dXzFD^YZ ztmCcoqR)>U`hCTZYg)JMkiI@Tx2)}eBufWvoKNwh1HFTuy!N~OpRL+C{9oZ=(HB0+ z@LF9}p?RHL_RT|a`>cDDx7U7l&#|vZR#?*-*T25$K!*h%@3}8XZ!o3~4WpO7^Z207 z=Y1wb{rJ&SPcC_)?;G1z+_d$;FuUoR+b2HuMn#9W%3?-D?|*jWG(+o|frd4jQ^$`hHjFbIyJ+Le(XKN(YM#k_;PsB?7eHs5{8Wm@8!9=ljJa2LAiQUO9U^i>|{BAy4P2~x%>an@)Kc6sk^Ajf@d(nFGXiDH~o&(lD@^OxJ zL;q86?apu2vG`^3ZLtDtPNFCYS^Al#UDRBB;wO(A;Q709 zCip*gC^M+R*YDOY+3VA9IWc5SYUFiW9&A_Fc2vT+E<4x%969jyIRlmr{cLr9X-U!8 z4ljQmR1*17=53z!{T`0h(^P*9s7eo92YGr6&9k-8yxI$`g`Y6A=p_s-`wN3HNO%~p z5gy*5!ow$0c>2T%&sH}JPv3Om<(n_OTCWgZZAye!+s&dy+vh|J(;K2iyN^W6c6Fj< z`yWKh4qlqEgP&&X6r>qDM`+%iZ_vEEp(l|nMn2G1%vM9gl>)q+?Cv%Q|L0@T!h22fm@>ZAU5zi{c)9 zOXq{9?b9*%^4mJDeCAzUeq-Ezo$vqgJ)Qq@{{fx19QdJ*e!D*+tnBwhwJx7K|Devp z2G!{NWP^^D=%)_p{;RErb^m9(eXQf_2Xz!xcYmVGk6d?zP&9q5W6{V>wK`va{inLT z?~ghboqO~%;+9_~eok04?W~TKzubRR_rEma3qrBru#T4Sykolm{q2wI;S-+Mu{5rK zo$eoM*Rjt2hK@x)dY>Tw%DwSAmVWuTj+Vp*=^s4&B+11sg*w)qsnpSYq~$4He?W|m zm38ZMOfr0`W8M8d>vg@WtvXtow&^HpztYhh6nI*fUs$N4#qz9`8{Zd&N$?-Xyo>783bA0BofeuyCz5b>P{SPG03yqDuFUq*-#?U2+lV7<0%goT8cmMK_?*F_wH0h!8yVDkg zhNga3{&l{{4t-?fZ!6lj92r{k#`k+8&L)N4IrpV=EuK#dO-p=f{FmRygg&{b+iwvM zO%B~KC&qHTM@Hx^$)EXbUvCLLck1GX=%B37QTIQ6=Hv6pp>NDRkhOn)YUqvU##dW% zvO-56TM_=|8}maa6rAn$!ph{(ke}>bdak@NbnPzxH%!x0Lf4(`eCWiZnW5qPp6(hr zJ0&#Q82Qt)C*wnx?ys7A`rcT!LvZMu0}cNQEzbOy}&V(BvLr z5C8a*Ep*n*pKl0Vm=M~2$BD0pM8<`VE4=2r>we7&&B&fVoW>7&lOgSer1u>51JBR3 z*+-5TnyS9*G7WE$#`hh(zDFb*8HrLG`z3Gj=3FjldX6=WvV$Z+l((7 zWSMORwv>GRgDAQb?`O=SuceTD9$s5ywq}{FIS81FIP}((^u@NJdPrwD{Ywe+?Ae)! zk)cmN^zBXWvY@@bk;rp`zUNV`XUAebA{ymso@;vFN4y?`-UUf+N_Mh21*++-lID4O zrXBuxA5*qHQ;HVl+w6t(eH9Kv@m%G(w!AzSvB~>}2UbDB^IlFoR)*?dh$)X4GUh{-&TdYVb3-f%Ccv!3VM@=(qA1V>7Cs?D4|&zz&)Wloxg zA}0knlpCg}Z^v6llJzz=46-c8tho@A3UcrsQQ3lI3xVEDd~74tj8;OXgEJ)=S(1hijOd@EFk9#0LpJ7t!}4?O!;;gphS{_=cZ;JnsWL^5JUS7`H73u*d)8E1>`@L%s z^s05*AsgRf(VK#67;}Cu)!nT4$w6}E+cmtN^1!|W>FXqVDX0~>#zqQ%4BvYoOp90B zE+NGf+pRh1+4^I=t5?Mz$|2jIeSSVwc&<6XwiqK?`aH8OGbgWb&|jyQY%{CD4&U#g z@1)oqjhPy(In$Pzjdw?p>%+aTkE0)&ui&Q#75?47)dyQ#2kZO8i~J0th^~s^l#J#D zJK{JB*-+i@U#r|5xS5k-)$Aqb%mi z<@&o?<7Q5rS^4PphI_|%j9UHuJx8{6pEa0@95H6@ES%bN=g!8fj|=DGRmk+Uo!q&( z>Gb)X#0);~&JBy3HhR?DID9k>-}{-nWb~*Yt37jQPO9EOpyx0jl%GR5u8)Y1#!DE5 zu?@y%d^(9{vyIri5t*w=-Y+wY!p5%E1ih<657Q5;rlp!_LYoP!+5D`XtjBbA9X=?6wBZc#8<`<>MHM{wdXMmd@oy~*c=La0B zryHf8;3J!d)6bZKfbnQvovgS4kTFQs2pn69t1nct1NzA{(8cp0u7 zetJ15ZoCXvN#QUM3@wo9Rt+@h-@xNK96@FE3^3^91pAF7KR@`n)b%Lxqwi$DgLw@O zH=6vgcpyx;@f|AF3w*m*gkU4p&}4qkT#i9BV7eR4Q7<Pf;3pV)?042R*A+1lBK?aO%{l&Z4~Q3E(~Tl||Y;^`~m@)kMYR z%a>jJEw298iD1)-synL;cge6-2wM(QoHT@2t|YAk>Q8+Ge!^0jw_nmzn1e8$SYCgs zg|HaCghkZX>)+Zip17kPYg-mm0Q^oYKdjXwA88`iG?(}t)hqhJL*Nb%rVz2PZUK4| z^*6$D9F;EI2fDGOB9P3fQ6h(Q>v1*sC13)uBTlCUh)dYYC2V*6?xLq(e`+9p_tAg% zsqc;7o%r`q5o^@*cvqi#$`9#G4TnA)+bBQ4aD=~((nm2`;^c052OMjWj&b|0;iFI<=)YSS4@y1PqU2&**#rR$7!iPEHIok*2Dj**E zah&7ypxa1uP1GKGUZG6ovxImJhbq%xv_n7LzOo#!g9-K3tQ}H1<+vZ*heO3n7O{SD z@IN6fi5_&y(T7qXn|T%42 zy{w_oflrj9KUFpBKL^&JKCo0Zk>eEOcc}}%LvIgyYiL#mFJU%P#|(sC-9C57POkdF z#En8E;ZXK?T3Gx#Qu=q*a5E3Ap90%X(DNkY#(`h&!m&rwr279gkOhLq3I>5ZmCFkonoiv1v=U^Y_Q65v5&EM0&{#NG7(9PvvDt9p0sdL#2byW|`Jgz4ACe7JRfzJr6~j z4%B&{`c7_nD){C==^rOBW$1gG~NH8 z2c>c!;uYy>I{Q&7{Ny=FwVG3>KEoJ;Di$XW{&6f1*U!}koC)QaREzO_ik?Qbp7u8C z^9TRf=SQH=;k=CeH}7-nyJ+G(js&h}oTp6~H(?jq9x#SrOlzLr>?^dn9_=&{(q41a z&jFOHlJwl+v=KgYX2LnoF^*-6VDpI@A9coBt;)0!W$KRcU7o|8V`S|RO`O9~z8(;s}%-M+Jwyb0p&7P`smC8;XYBG#si8VoxNl zNu2p>9HohjYcO_kzq{IakT*dS*y<7&n5mz8l%xReD%6& z-ZqLOkWU7GdrQQcajkpR_;wMjkL@aa48l*O@JGE}dZ+A*bgk8u+Qf2ao3J3AJ8@VZlt|a9 zM#M)J_}aNhXt;QWzX{jUqFy^km(l{))F#{}X}AWj4~GYxW5_>+@jN5E^m`UUa>gKo ze!p592xPYuewDait#I+HfL|&6uH@)`hg=`VF^utyIgGb4ZeT29tYAFM_%-9NjHV&1 zpK%0ZIOC0sX^aJoC5&4cUuN9L_z7b@;}4AQ443J>%vi?wDB~K&0>)IvIK~jhUW}fM z`^eUC#~9Bu`V5u+{TRnECNO3&E@fQBxPfsS1F&<_-$LKSR?Z`Na(ZU$VcnhPQ z@ovUP8Os^pVyt02#`ra3`(T+~FUBE^V;Cng&Sp$wEMQ#4xPfsy8M8IDoM;<6OoJMisBxfDdoYq^acmoG7e=%n@fi@d2<1d__Jx z!xoWe&xn*O0%E8a{3!@8m0)E=eiB-QdCs6!lPk(J31{PTR3d_Y8(X-Sy!m^#oi#!TDEju+oL;qG<3_dc3JDZ=$ zHy(4hGHI;TPs`89OAjy1!zak*q^H^<(yVszv^H5U5U-aZo|WQ7TXXaDMK2-#W{?!RNUVYbGY5Gi=fZ_zK!e5s{IdYvT-0rIOCfbL8KfctpP4j;ZB| z>2}QRVI~-qv^=ZP$K5!OVmj%h zug1~0!o_%E6g>j#H5^5VsgvSjCPk~2#Agh#Hrqm{4UT$n8Kbgt^7EpwA|$Y8Bl9%} z>v-8qCKb#PyYQ8QT&x-uN-y!eAvQZVf4)vANenmH*>r8izsrJ%yG97A<=zs}<@ zJ-%44!kJhUnj@xb_>dg_MQT-4t-i+;iO#gP@NNp08((MC$_y(*;w9UHM_6Pq&mvXoXk)QV3M9d=R{UO>7o^jUn zT$`hS*l*z(2Lq6K(E8naYM5x>Vh+|k^=cDcT`aAd+Pz-!yp-f@eM{V}Ma@a{B@0n9 zvDxWVP-IeElIiKA5zLDm%-3_n0~BIS^g#9G$d4XI>c4ekiCeB-`mn6jnK|iMdaDtS zd&N7xnOI7{9rdtwqPI5vYmL+^;j=`W zjE13)8RAy`_vCbZ+ls+Jz#bC;wyyl3zJfd>ZBGOa!evCDLv6E~Q7J=EQhe-SM zNz>@}Ufp7`G`@;UUTUC=&vuM{B7{Oua;|O4&o$;EPdoH+R`6J}DGe+VuQ9j4G7Z4|zL*RSCdRVb^1ICi7y_@Hs@>v>U8#(IAZJZpoa$w@xjgJnt6wD@d`&*goIa%teLMS z`S<({(0k>p(SKjBylVNs*F*7f{jIN0{NHQ&pBMA5zD8zna~|yeqa~sCL{7yk5u=(J=8O{!mOl&s-z|5~du-H|^P96*+M{uw#JVbp z?(zSMmxfFICbmn&Uq$%0<)*7TsoH#1_irctKSIIPKg=n!cl8fP{*TE1e_1B3+AK^m z$*VRC^M6^E{yP=_Ki*6ECQ}4 z#HNKufpx%bK-x^R16T?4?Q0OruvpL$nABN_9zZPUhH;2jQ9@b_vE}U=Lv95FvsL zHiH?9g+st=F(iy5e_&TpB#a{o>x5ayA`t>iAUXRRM+wmr@&I5O{Hw6INa5>%wcxSX zSw#6N2i6n67WM#|Lxi{v`AYy!0+s@!fDJ&>N1If>5Xpv+F+yBJ@qsspBu^7IhTjTa z25f|$;BoYBk9ut*Bdz0ZF{FEzh0)qdooE2|2b#jMMvm~+K+0bOqd5ZghWfPtJ8AV= z!bE``=T@y6Ncr-cME0u_IlxR|Gt>ba5I$)#($$i+s_QYnYiB%)qp-#e|MIE0_eA{^ zMaz6O0AqV0UF;1xP4yOoCIU`-Oum7>;2A=!h5kxlZ>_=56i3g-*K4ISv1b6`lH%!| zJXWm|_#o;rVHVmC=>ryu6I#M-A?m;raChGbtOd44z8Vrxui7cid?VInpg$6)hA?0; z(9ho>OgCd499VjbY-f>)5_3|xoQ$!lLOh1_jq@-IDN+p;jA`>xFNS2TEM15}z={lN zmsZV^Da0HzQGJ}!ENiIV(Y}CXB*)&9UGPsUk?pV)SO@=# z2lRerFs;LLPv8Z>X^2+^3`2i3K8U?OX7q=Lgb=`jjY4>%pNU7PKO}3#z!@k{?PK(= z467FTxDXF(^@h65ScgOXSC&#cPu473sGeb8;5g{1+loC-z<_PCUl%df0oNj2!IMJV z2Q)rKdToXR;LXri3oJ!Ezkg!i63~LZC~3r>7Gf^282B(gidME=wvz_n43tOg!1x4t zDR8(*(yD>?A-_4;Z$jm*2G*0F=dky%H`*)kb&NB#02!sFdlbB&_d~4_7$%YokuRcM zB3=QI`e6xh8uZlxso#s2qfYh!VfRW&lyTCoZ@~Eo7 z%x{aYJp3wUzbOC?Lby`KDqt+a*X^P9lVT`;ONj5$e@piveZv`#%6DLYE-|nVRzFZ*`LaQ{V^_8 ze@x>7>h%*E&#c-;;7=mmqvQyMOV&g!)myTb!&nEri`w6(WcPZl2}t1^KcjKtlvegR z=|j3lvEK{%X#h3=%fG-l-v{+}jPjFY5XY$SJbaI-=pj|lo$RvzoQ=k(=MW2 z8!)b2rux8n0<$n*pxlkXwKR@vns^=Muo$qPgu;7jVkPX<=!HEK;3X|Jkpryq##%Tq zp%wO95O1T2Dqv$(F(RwS>RGp zQ$dS`ZBn^Vk-z`ys18~3oQk?lj8I5x* z-=MyShe^vp8p?iv%YpI$?gw0M!aM-JTxn7RP4S(guQFl2`OIYYG^sn--V5-c4C-%! zT*jIlD6heMP&>e+?Ly)}6ZHBsQ#s1FK)0GfCS^hWw86NK{ycD6Jp%WwIFqsk_ucV@ zZrYzTls3qp&rlQWK)p&txf}My2WL~_a6c2}G}LQAS&4T3I8$OC2zf+FyWu~~q&~-e zo<-sO?SXS8PlLWSD1%Th4QEVju){v1Ov;YyrZMOTrT;jn1j^X)pbN_S2__ZG=QwMU zhSDblXVeFEg41VGQ&2WcGO2MWVrNbsx6q4VnT3Jc7mor0zkzfVkQC}UC2 zf0_yM0)OL1mYwcy68vzM`z0XHv^hCf;e%cFThD6~G@p-=ycq29z)3 zxk>__DC)9bE?zK1+>HA?3vwN2_=`3=f4T=QJe7R-O4DEXY{hWqs1aPfMeOhM_6G99G{N=$(k6fUV7v>x*v?sEi~ zUJKhdqU?gw4XV-=r5WXUC<9P-Lm7<{g-aqYFu%5QKLPhG?cC?MneE(Xr)7M2N@}@L zd7Aq*r@3E$n){8Xxvzdh=9Y@?r@0R|piOG2-{zX;ONM>h!d`F()~mv#a7K&M^c ziM_Xza=4xB)@?I%kW6ry25q)+1)r=M*C#JkI+EEZ)!&dkhId3ky2 z>8GDom6et1z4zWzCr+GTY>J(t68_=>^2XrI<~^1RRnON}u7t~h%W65?Y`L&`<;vBo zS>dqkX|9$%$MN_zT)W}oFxujJyz0DO zS?$)u_6Z5-pDC5~_f#h&RAYR7t@}?%Sb+Pelgj>^@tg^DbbYpGT5ufK%^1J>Reg;? z*nb_z*XcOM7kRK6@S(Sa1j5%`{Zk#TjrDb6Vaiu-zut3guPJ+~QDY;jqyPLzhXu4$(AQ0GRolIa z_UO8Nj&|P!kCP@ezfjk&-t%_$&3l@ft1Xqh27Tn$=@m`Q`Z4Q+zK6CniGHl<&n_BT z9ut;VPog)FU^@M^`&a|tW2%%YzegBSfESpjL#-yEgGjO`JDA!+p(Ya zaw_hNMVXL0BX>jYiR>wqb9ZDj^4a8k#E20pDk@6dbkj`=Uv|>4LBO?)r>Cc@f`S6I zW5*6HPAVwE9oxTl~AElD-KsA%wMch!J>o;k<^N> zV;~9hQ@=YJvkM|ggGF4<}jI6!iyW;lgF)CHE zD|Sw2e@|~>|EVJ19zR~Dv=rZ^)LKtZaqs3SsnP zuG^-80|zQ!Utjn_{%ZXA@hUhtSVcrcXxdGiHcic#F+<&Y>#gd~fBv&tJ7|ubs^pcxfJipgZ)^uW{t8f4^)MJ9j~5TJ67F2H%x7uAEF*w9IPH&8l;|GH%1j( zBGjYnW~rw#lhmu5uU9qM^OesFqON{X)Ua1Y1->k5XAntQCqfbQC6!}<>cgO{+5-M zshvA_YFqZ)bI+;$FO{kF-;3JWDC(70UQw^V{9U z17+{g2sxFD)p!}EZkK6ll`K)2lA(6Wc6C7Z(u8VOT!Qvq(U|{}um(+szh4F)AMNKh z!WWQU>H)Ohg7zh7Uy1hn(Ef9m_Jh!#`!u&D;r)i`N^L5`-UZsf*odzypnWace~k8@ zq5UDWKaBRrT-y5$!u%EtS-%}V2lhY8Hes(G>+HstpoIH+DXAT%I@g7#i$?~nEq&^`+7(O2U2nJ`Sr9n+Njb%~OPGnDMwuH?YplkHUx6|FAA8wdKq zXYz&Gj#AZ-{6=`t((5)Q!;+LtyI0AQElM(;P_n&B$=*7b_GYvnh4xd>eiquNp#3Ja zx1;@&X#WD*zlQc0Gvhmi zSs^q6kMTbOA3{T8du*g7k9~KfEI59ZpYE~FH(5s*Gc|=rDaAa^u zXb9R9vY}V|`c81d$SRRhkw%(yxyNNOLs5Y}4Djk_?hha%*nebXL||mhmA(6T;5jPz zI4cZ_LHod9b`TO7b9wLHeOjJt7zcQn0c1pL|1pvDZroohv*>;s>O`+ z00_JS{dE7s#zuj}QDKo$8VL=fL4Uv>6FMx2AVfw*#q^ytsdw+*K%(_|%&7BXg8afT z#0_|U)uc%p%PU%+hxF{)GoXimSY&7vs65G2V|nEyr+5U82@bxtbI-2+fiY1rQ4wKL zEl|RQw4{iT5felFx|q(t!TEr+iU@_daz2l_&d_wfr1@U=E zAjCD&Xb$?3J`k{=3wrgK8WR!~I2N+j`WVlR6z<;Zx~O2#F~a3B=^v_V;Qr|0A!DH` zu8(#9;BRCo1Ti9PN7b*t z{;D#Li`oufVec<4cGj;SCc4p>=%APBvoX=F!bG>VQLHqmAekE3^Ity& z8n~&-@bjJM2N}s;orU<1O5E`hVYZQZ(onT{f2nD_vmpI{NX-cJ_84O z5A^Hh-gEFpxbEApzqdE)_Hy_1^Y!ypS9lHa@w@VTzh0=& z!)*2$;^*dep%2;&zPQIleLXH4;&*|Y+xgwQcD-r+Z@lq_V2@M)pOlxEYq+t8Bwu{-h1As4 z2=)|oJ+^_5!26%0&9?2^x378Xsi)S!?F+m6?z^WZB_)M!+O#Peel=kmjO=^aWtR=d z-o|a$UVANR=Mfqj>cf8V`FFwpN2k?+6%z6o;o3;JyM`RAYIhaZ0U@tt?xc^Y4j zufteBu?_J22lQ1t&1#K?r7v|?#|<~6f{2TQ_ZWWX#& z2G7OF#yz7v7x-U#>7|X}qZedi195rr#TPXWln?AT33-5hCM_?7kvah#*5&^F`-Pdh zL4L5Leo%MV=FKYwABz*Wy zu$KaUXdGyB$OF=q_>dQbouA47k3arcXMQHlNpHeSSs*TioAjrQW_=?P`mRXOK9RB5 zf**~T!SE{J@UqAd;4tWU5wB-NHhqZLKy&k!jsTB38h+S%F*4ydvcU_ZJQqEo9#CFx z`B-Fftw_i_pg(Z{4@SKyaxHLJ^d90J&DihWE8_ox7Ax#5b}041J}G5+V74TpG?-m>;;`>w+0P7gwfc<1i z0$zlF-hPn*&$XyUhoA@jF=vQj|0(~NkA&Ia2TQ_l$T{Vo{GeP>HVj;-)vX?!5_MSeRA=s(A8;2JCZc~u*DhJ;~xn=Y@(jg zwzP)_<=vGhhFq~E{rQ~p0^8EgxTYDOLr00jMa|RCo%;vg#(9HeI^c0eI81i z1e-b#HmM(Q@GNh!t>|W58yqzJUS3|U_8)UR=lS38gZj()z~Be0BAqP0vmbJj2 zd{`v*Yv9lzatm;vKGP;e!X_E|OdJe-CJu%^(mtKmQ17Jlj(=NI@{t(iC&xh)=s zuDIgC&%}Z0snqM_iPdAI4o(H@-%SRxps_{tsW~yz@e=^ z{}VPTV_~ptoIg?0=7q@Jb3-I`PN+P%{5JXcy&547u0Dp=XV@gB_TY!T*YXc|HW7Bh z#gZ@)Zt4T+Z199O-qk;*KSUfU|1YIYkmrFzIdIsq1~@<$sL#ZK`b?Yj2<$3xNGA?= zPL%r+fkQ&5teq34aUh)ZF*rXPKE_DQZ%Dt9f7%Sr!So}T9S(;S78VNqQX?DuV0k7z z!M?22JRT>z?;b7Btpg6wg(rbSsZ*b6lWfZ;$kwG3Br_>c9$FM6n->JjgTUcl;IJMz ztN{*8ZGDWtKsS(nV}A%gWI@Y6T{cRXY_vvKY1S>9!@?;ety38 zz39_me?V5OSka2ZneaFhPY4ff(hl0Bv*|N%U}}ff?cM4_I^nF%ZrD!o_V$*Tm>9v? z4543_oSZDnmMxQw8#ijb#+Hw^JM@Lx!-cfx+(27JUx@xF*8@z{=jz<$()=^l8`&%L z{OsyuJmS>nbm#(8d%zm^uvgI&zW&X)*6?3*%{8NNv}e(rIdi^-AFcV2m6at^r%n}| zEt0?e?QepZl05X#LmD2g?@4dsz`20)1m%jEYZ&4{drCRvTu2(d{9vf;%Om{99!A?w!cJJpBjR9W-Xpy!dtB$xw<9m857cR`&)@^+XG5Q9lU((A&r{pz2R>%d z`b_0@*IoAz&h+Sg5RUECWx*Lm!I>n%o|X(AJXqjcYgh;e{SDVl9E@z>!IJ$kQ>Ki* zXa>BwkRd|^XV)|zIA>7i>H8RX5MIJh znJ1oH8?dB4P)Dc_22WU02I-p^a}U?vTt7oze%Jmdb&K*3S+e20*a7M(BI2@o^=i!r z;y}BKy=9p>bEdve9uNm369;46%QYo^Y|1_1WJwt!ZiamEK7A_s+^{8o$KKN@oHN|a zHN?5VPq}BNEOEb?I1x7PIpge|*58nj5D5wj5}Z!aeBg7oF>qjM$d$npLrc?di<2o+ zrfAthOkTrJT9f9)g>X|3Xn%O0YXr&cfZqan(=xz0igaduW?mBq+ELP! z_Wy||o{*A~5*;8erT0j5%^OV1E(!?gP+_dvr8@(}(cV|DpYdK5pVQ z^^GNAA>U~iAg9`wK%7L|R_MB(-)L(n-;_DRPd%Vsk|&fC@__qUd~PIuX20YS`N8=d z_~rm^+D*m@JDN27hMcl=m0{vwWWq+gC_~hR`|i6>V6%ktv}+=4MrIq@8QLNy$`98> z%!HkIZQHhudyH1Rj_3EnP8>D-hW`(GZy*kaT{Lhpa9~M#(>7u5*6Rey6z5u3`Vv-N z6Mpi>@F5u2q1;oJNOS5CY(^&FI2YJk;dhmP+6Y&Ew8s8P3K-i7=2-ZmW*6BIJuHokx#`+g?&nC(x=OE%i9VDG-KWOVXAGz{`wB5C9 zmpuCDqgtLxZzB_L$_{mheiN@LGwhFZ0c>U-UfFc6bSLageF%^1yh!+sOdO1SCZ4bl zoQc-`P{!DXxDj5;Rt0P)aWPA>?x!9SL5TE)^3_*g)pnOOpuSSx2|x7!@@U8Ke^h2%G3Y~j8vCEH6RrC{#DQ}s{TJFl z$^qvi&Yh$&*PNt1`9NR5-~+E|^J!~o$FQc*aSh=7pN5@yX!&=Qjka=bTpN1pT2j_| z%{t@-Wu9~f!o6Fab4*w>;FiZ z>V#8TN|Uqf>@2%E%bw2C(^&>O%T#A6Ev1R258yE)2O*#1Ijt^eO8=Ib@_brf5u9F! zj9qX!hU}Vn&4^0d{XlfrZsJ%TSPS}N&X0_bkDmsAKNNfZ!MN{_OrQHw$ zit!!B{TX9pIvw)3@5wa))}V{9UbvIkXwt-qzxJJ4dTja2c-d{(yJftO zv9cYFu20+j%pjJlJm~Sc9;A=JOc6p?v35ne^nGg7s@0{M;FY!+?eSrw>e6IR8BR$U?oR z&6qRymd&`!0*{{kuCk zH!~YB(B~jc=v$Yp4biba#%H*X$i%Uk8GB{?m@z+BdE{9Z#%LJpV~mON==;2GddwJ~ zevlDglZ!a#YP(MBF|eqZYZ&(csH+r;TS zIX>bkaoAUw;R*-Grkt~6V!Vwp62=P|r`q*!(ph7MU;Qkyv<@*(z`*l7)CtCHnGF2t zSFBpKYEf!x>YemKnGHC&Z$eryaetk$CdN1zD`T9A@eRhw7-M2=gE2D3msZbi$^TkE zPY8pSN5ox?*f;xg9iKLUnZ6CPD;$J{{NH_FpfKL`1YqEvCilh}W1&1U&c*l&V~sk- zx3bX<3R_cR{qmBLmVOw-@BYaTfp@`{d3|?SV#kw`?BZhxFO?JjNLM3 z$k_8XXS|R)!T1Pcg%-r93>t7SR~yaPEmEueQ@8Nl-bJwgcTyK=cepO#+Kx1#?IAsm zG`ufw{3BY&3AjH^d1MTqd((zaFwVqSALC@HP8c{?bX`jv{|=1(;c?z+wLzZrFDZl6 zL;CzI2?sOj!Av+vC-UJ;Y5$9*I%djP-3GwGcxjvT8~excAAIn^xW^xVdl(qH>Px{!{uULfrmW8Cl=FOY6-J{** z{73swSs*y_LhaJO9JB(%<&v_=rV(5*SFY zZl#V*P|vUUm?CQO(x9Q(bihzH`Bi)YQ6^%}mvE%WEk*ZaUYBdPZ;jI|GGPI{33T;EX^ zNDso_HnC6C+XvZlooi1SpBIRWO^4p=bs*<@#$^~=f}bTg_pR|Z)?b8!ne-qWZFxiZ zIB(H*a9xV-MCW+) z>66h%GfK==KZBlkp|5R-SA2`HxfD2C((*?gWMA~{NN?`tQUuIHj{8xR8(xnz&*}3IG$_$NBlMp-=mwsvDqj2Oc?N<0Kt0&FIy!4kDQ=#7u9JXN4j*d-Tj%EuaOm30ZX@v!2n$6HS-{anH!9_{=PpcHY5a z`OWus&FXXsXSLXOy9x3bvTfV89f;8!gN^zQ=R*FTo11%z-9ZfCF#1?rP*9-nqfVFX z?Cc2aXTMNfT&($N%wgm+Z2G&dgE5|l^UvI8{RMs0L;j9X54kp`yp~`MaufVIo@)%m z+UhZ^1CgFG`e0qkG<`V3pD}cbwk!tgCa#&dhp-QB%?O<9L^Awi#rOfoqOQ>{F;iaI zmMIR)6Rv@uYH`W*J?EdJD0k5|QI7~KeFw^R&2x6iSlr?p&UUWfxPIjNm1_mARs6A5 zdIWGW*2*+Q<13Yp}-vGyC9LoNF)cMR486b!Pz9o!q0QPer(Cx0tww z;Tr16e=ZUB;hMgNiRGRS_Yt^eBmKxH&O!8pn4Y~iO#540e{k)=wGVw=t~ZLd+%N2d z>+o>ELH^<0Gdj*q8%8?Rzig9TmhpN$N#C7*JJ$_dJFSVItjD;iPQ-zAExzY1si~>@ zj0odun?Hs@}{Voqtxe%5v?3U)_Uot;^sJY6>Gus(B7)-_=x(`@GZF1zSA zu$l0yz%5sN{~F%xXJ#EY{r0uac-C^ocdvE98;y8n)^A?x>a2+?)H%;t)2!dR)(!7g zqq{dl-K1jhHw^D#xe0HNnW}EUbp-xS!e^%EyBvmZXwv^^kobvZARi)i!9Rd?N=kqi zX+nR?)KYx+dXbuo-sY(Vc=uiet`hhPVl`3M@mB%ZF2!%c!}*-0H%=qrj+j&k>MX?D zUE)xG6`+aZJD&9zH=!p!VQ*%H$@&v3{{>%dHMNc%iu=i^ zNtluV+dPc1s6%~*;r%%AXt^A%;s7VdX1Z2&cjglK0kV^P@7|@X%ecUlg$v9p@LO5H zU{UgT@8wGujY*!J7{4$s`I?3EW-ncwyf|UmHM18l921wkaOjF*-e&yD;=Ba>a^DQs z(a@FIJZ{<2<;nU-9-X~;ou#+oUb-DxCdbcSj^BY><-Eq7rSVIaW0d$g(M#v8z|VHg zjZbcU==wM;1x@)=dQ;+8#4j)};NS7yamf*jRxF+$ztr2jd|vSE+4yb5@!kn>3zFl# z&3>)@jPpDFDC7K2foYsyD{^Q)&d;D5F2J4trvFbUV8wifiJ~dXE!&*!lkJ}!kR6mg zGdngrF*`jwGuxV7mR*@$lU<+Pn62#YcC+2z9%PTU$J&$Z7JItgYA>@_+H36f_C~wP zanCX5_~!)WMCZijB;{Cg(sQgiWjU2OH97S;jX5gUJ=dJ;pBt1Log15*lxxXN&$Z^3 z?JWF1Bo;9y5uQIPDuRgCa4-=-t?C^I4Iiel0 zjwFZ0k?ycM${dxB8b`gO(V_C)^UeAG`9b;7`LX#)`Ih|jd~1GLer0}5etmvpzAA7p zFck24LiHj6D;hCt~~*K#&F)GV{@SrAgJJfO`NeJb;G}FbM!I z;lO4l@JR$lDZrxvm^1;Gj`eKKcL2k3;8+DLYk_A2Fl_>^Zot+9`1)v!!-4ZmV4VoO zQ-FCIaL)wxj)Jm+@`B2Os)Cw=+JgFmhJwa|rUF&yR_I>nQD`poDfBN4C=4nLFN`jn zSr}WGSeR6pQfMhmD@-rUEG#drDy=PTC~YcLc!_QmFSi`L|t%=qYYnnCF z>adnutE{!w25Xbm&E{eAu?5(|Z8L3&wiH{MEz{<(mD{RpwYCOZlg%y5Bg-c%AS*m; zW>#WWN>*A{W|kwXJgX|J7Qfd+DfNJq24sgrMiaABveO`=j_mU6s_feAhU}(nH@k=3 z#~xr0x6iaE+EeUl_Ds9OUT&|l*V-HGO?I~&j~t(zfSmB0nK_9$DLH95nK_P}@|>!i z+MI@*rX06ik6fSJfZXuhnYoF%DY4F^vX!Ot{DCYE~Tjw(kjINRiKgS7ZSQoorAROFE1gFx#B?mZE1@1I}Gj8CD4>%GIZX|*eY2bpRu)MIUu(q(Fu&L0k z$fL-oD4;03Xl7AjQA$x-QD%{&sJy7EsJ5t~sHw=U*rV8|IG{MZcxG{8aY}Joab~fj zxV*TkxVE^V81h*UK2lEG)s-?xN)4o>5fb7K>F|eSL_;c)AQ9<>*21#F%EFq$`ohLS zRpee|F7hu5DvBdWp59tfaD}rlh{4u|$=+mzqocOM^|8`%I0n}+x%@owrE?dEy-rFrQ58wGFzps##V1@w5crjEOVBBR!~-SR%}*MmL)4a z%bHb|Rhdrov08xj~c&yT#wK;LD`LQKSKz h256NJG%69gR0dtDgdQn-dv>g?tw9I>P5-MY@IO(Q*TVn+ literal 0 HcmV?d00001 diff --git a/src/fetchcode/vcs/pip/_vendor/distlib/w64.exe b/src/fetchcode/vcs/pip/_vendor/distlib/w64.exe index 46139dbf9400b7bc0b64e6756ce17b4eb5fd7436..5763076d2878093971a0ef9870e1cde7f556b18b 100644 GIT binary patch delta 30933 zcmeHwd017|_xCvmuFPCiF7t3@P!IWrc6vz91);S05JG{T={p0=beXq}Et-bczYtL)1z4pG> zqNUA>RyQl_$u}%Z=xg4x^5PeFAAJ3^DNNel_UYe<-<;t4^sEAXp1z^Lc2A!GK3Ul6 z=^enk54JvCuFwNC{s;|rRN&CbIe8?L`?pU<3dc=&y&31^_SbkM`76l{TnOhPICESr zqF<{yT>G&Wdxd}YT>_a|AaI{<2 z%8xbhj$JE%M5+e@iMCfbpZzg%TvmsI3E8G>j$70ntZC=0i5J0zh4$B9KhFy~g085Z}1MarL2TR$YfDY_0uZ;cN)IW#2X<62(LKl*c4|K6q5&^V0uK z_zPfAuk4g=SihT(s2Ya+KFDNEk4uTUnpT<{Cr4w*lB6RgRY*Gn7X90M z9T7`MhdGYYWMP@gsJZNq0q=-XMX6DK8=AAEYL2T^ zBXhRsMy zl_2iVbbT`N-T(mOU#a@!d+8;AngUWSPi!VHEJmB*Gw?nut&n7vTqTCb3d0ebp;Tp1 z3muM#SaQ!4AX=V+0V=>|$n#J+)Usm`VMa@^S(NsPrI{XLX}*tGI>$I7X?W5I(Ymn_ zf-v(Nup2_xxAe!_i2-9t9o9*j^s~ASwqaL`If(gs=Wi;n3VOMDJkQbU|Ls1FuA6{p znib1*Cr!=oX^~-;yMmf59hQDfD>aV}GujMk5QItU(jAF@3e8SX#J6EL~6`mX4_sv43QV?$*HT-FaHN%pwbZKAAL3suZp=1~b z@fe8=0Gr`G@J_TWSZ81Gswj=vD3+Wu6^J=K^xIR`p@(etV*H+hY%KJTA5mLP-uCht zAL_iZK55plr&utU+_O4~(j(eZEb*MlM>JdFX}MYApPF2;?8Q9~?Bul(RMEYTqNUmq zISS6wh9W7F0#C&87*j?5h!d4Mw8CcSiE7zu66_Fp#Tgs$3Fs}5P?4V=*lau0ei)@t z3`KX9K^~*D?y#&&e@M-ATLiH-Lo)G5ZqpHgwpJAJhi4Shiy+Y!lVLM_LKLeZQx#-@ zAZ#nlF`?d}t? zOEpiD&dJFrtdbfJ!FEa>v!OXkMHaa|@mmfE8ow*?8`R3lqI7v;a`)CIU;TC$qCfFe zL^xwe1SEtOUegk;DhVdXiOHp*l!%!wr(t5N^}CGvp-m_a3xgJQ<(pHm50IJh_caYp9%2F**{m_0rs+?4z*V=DoMQ%QwJwSs<$~6SJP#}-(Dj}c(OxoFSR&qXi`LFSi&~Lqm!(gdvKc-FsWniC0i#t(ik|US=*0p zP%Xkdp}u93bcRF}wqy(3y~oDF5KgKfyNtHi%O|lQx+esKD4j4NXBSVP^)5i%fQXhq z&;n~d4CIMPrLiLq=v>2b0+0uz&d>_5wbVY+Y##gD?6!N1bNy|ladP!7)=?kn6NfUS zV~P=Z2`OSuBX#EMw^+X3Png((E!F#YAE7Lm2I4K7jT)`gxLXUhQ}6G6NK5njXPT}B z`%@n|`3Dk6V^l{nPc+c3Ha{#;I+hOG(P}tNS&^_E2|JY$u)1}-#X)bvf;=NvI4Tok zK6LWrSXv*}EcD9`>Kr+<6(p+!n_=oHX7=z37$s0kAoo3VFN%etftI8KQX1G-9sz>O zDR$iBr4hgAD6hd=tsGbq_3s?O-1_<^>Q_1~EH16o6c(Lkb!!ifG_ny+V`-$7&JT-7 zD}^Z}XS+cF7E2N)qhim=uRdmbmbGj%aiTuB7&8gU@oqy}@LXYYanBx)iTC%|Y=y#| zz2csO9@m$JvG1C9acd28xtj9=ue`Gbvo#OT8jcMdLnW3bhh@-Aag%?6saefIGja9b zV`+3K43Up3B+7;+sH-*&RT>RxWo0o{^}#qiq)@{Z3%peu64^*f|8*x;+@h0v4)X7$ z?4>^P{pM_Mi!h;66uaJ{Yv6Y$XwG^cLk{eAq(fGL0z1QISaOv0Y#E+)_z0}7BS$JR zF5LCqI9W8|U!(LUyH36ot>mIjX@!Cwmj57o8g+P0i_g%YK!-QjYGmeJKii(|X<1}; z#@UIJ&O>)w-E3*l#t*yDG%z{R;q5OXgoA+E2~c3*IgJX-uVF_-kG+Wa@mR#?;g4ty z{Dc}~P1p!f?=DJq)G)yYvl7Zb{z}20q9tKHByy%;WPFcgTRqj%^khySOtNMOHp8Mc zcE+>E@F419OG0ZLvUzV@tOg+e_UVg=(wNpCZ3I7~TfR`hZbX%P zp0C8`9cD)ionCtDC>X77`2g~E81d+#r7CFwkv4!-mIAX(uMD5ek+oQIwXk)yjt^2B zp(b35q#fY6ycb3Spv`dZAj|X`<20!&hUKu##j@RAtwYyF@v!)8hO9%%8ugY#m6-;s zfA>M~)yRiiGY9XWV1`^Yr!hGG-Eq1w8RWXwAdx*KpKHyMz1s+rJFw~ALA`GI@|?*_ zDyl&jmFTuYmlit8Tpwz_9jv@{o=^eZ`#PHZMX9fcv{+$tTS{zBvQu|<(0hv8Z{g3@ z)I3*~=<{-mNl8#K+x4`!8ui<21QVv*VcCF= zMmcWvpUbyHsSsDnf9)A6N(bs&$T!jJ@~LOUm@0Peh=(Umi&ou!SRs4WGn1sN=*RAL zXk~tZTUf24; zlaz4Lht`iyo|{a~cl)*rMB{XNJsURE^OGJX<|gQEG5&l(FElCIk!Sko$XFA`OTS~P zk6zILdttAjKI)OLbU#!|(@F6s%LC5^EiHRIOCP)c35Tu~j5>ra|6}Gr5+vFZOFojji5JO2Sc4@o0#pZk5ehmoA~P$xyJ`q0XKZ zQloTRI&XPk!&EyzSaCY!V;IxIOfjdDD&D3gwIi7=?c(qI{%+c|KXOEyXw@gCN=GoG z<%eIfsxEQDj<49mE?s&&`Hl=MIzrV0BRQhYPzw~}=uu)Tbfwd~0~&RA5RxTp9qJ&) z9FZ^m$tHIVnD!6H!vjqEOb^oLLrD^j807l+nDA0vZzME18E9gU+pCxhNfoFZx5L|> zt4gwpyqpE69kvnQ889tNUxK9x50+cv}TUF_Gc zxgq3KN{P9UE}ck|4neIIddrc%&pLWH#IqU3eE}n3N@$&5k22Bu*I%T$ya>9U4Qk~r z3{Gd0T7?LABAC=FUD)Eyerh#DC{1JSLplp1(pYv#Tj5q4W(^sc`w?siO;itTEi{gP zvJAtsjOKy^q^oAfbesgpL?0a%r}lv2wGm-v;Z$!5^U?e_Kb0JuGb!G!O&gw@HBQWV z0PkD3HvIA*0)=~zSnv<6+k9C;y_@_ux;4N7c8-^P9p{N^jt=eeQ=k?jO(wQwatM3# z2UmCaT4qfabIwE1TdjGv@rN*>|08znhlr4;+Z9PqpdwuDY>KFE&qbvHWj=dDB3r96 z<)@xa#nh*xlAYq+_Ju=sq?mIOTJ20Y&vN#*GF$`~Y9McW#x18lV(;$lWH$vINB0IK zP4y=8VSbD&SLlZc7vr&GElggU;a61Wey#th(1*6<7}rW?YEN%;F0_Pxpyc{Hczr!= zhSA$u+kFwnU&3fc{b2bem~^o?^;D)P+3x|5hX>|Mz;;%!FDm#rs*X9}ARV~=Eru84 zL8Da`1DiniY8cOU?Mw4XBZ}o!)TR~GZc}9re9Mgc{o?%~Af@Eb!U>cFkK(r+cdxG_ zk0Z1ak3qa%6fMVW#MJu$lEl&&nCU-+@~m)wg54s(Z}tbbTG^SFfol7hW{EjJf+sB$ zTS=+?35zZqTz~JMX#W5zj*UCeiWx&=eLdol1uM`;UIZtqHp9H3L|?u*gskP%LpF`jVy8 zxdLLPKG%x~zlR|JUub`JC#-wIHpIYJ7%P7UEVXH))*e+{@}LFmMm2(+VyN- z*1dIF@0Z-4HLwAzgqBcyoX?#tk9d>L@<=GWB$w4=LE$(gVJ|o?|3#|`w;RW0j~gg( zU+eHm?wv@_NrCn7PT)2{Uu;2raLO-?5xWN-0MS$^#}5aoROg7DQocWwrp8<+xwV26 zjzZoqDK82hNKUzabjtp72nmITae>k{#mJXfMjKzgBb(A@XjmM&uPo*Qww#iC`gf`@ z&sxeJZqIkf4bTCsu1))>UZg~ zL-sPfca8FLw2%hiGLCiF(uVPDaHK(qy~-v=_MF)3xq#C+xJuRatsx-M)Xb9Y!__y) zv*AouG+7tuPF0gV0;wM3@Jy4z{&>C-WkkJ~9Ea8J0#dc2$m`u@x`I&UTaMe1@HKeRYu^6y z9;8qW-hIX)cW7XV)rM~99gk7))2^jCi9I$N7Y3hWGvAs1A`ud%?j(;)_rM%LlYPvE z`c7hwZuw4TY#VDHj;#t4CB%}D0)v}}q4N|9v=W8~C%+B#q?+M=m*MjN8;U|@_0w2S z6D#H^Cr^IHdh-%kzlEwPlNZD6#3FF6ge}6!oqwj?SKNdumFkwkmn>(Fth@30YVottOrwy?DGHL_s%>lFRGKli~Lzr(=aLnQhG=-oBjaZkI zIFk|=;ZMDv)M(iiwgqBv*g;A*h)3jC>e$q%Ac2i!QdB_eFHj4cp&L35lWHYeNE$U$ zvC@abM+@&@rfglqjz)E~%Y4c)Ta=GEehoBKh$t0@Erw*YB#g3IwWy^Eu3p_&5DS6M zia_0~PbgL#R;KdLM`E>ZD2PrlQmS=nz@TU-!@(N>s6ig^{TiNcGxYfkT2;3Te3g73 zaLPlW$jOHlf&{2ohW@n~+z_!OtN<5R=&Y;?xh?o?hFB!*pt6(|5%d|WZCBat2jy%L zgxrp+$U+N<_ds&qEh?zlXo8LMd@;SlyI>M*kzyWjpV_8>Y? zaCT-so%-tyFyypN<+DF*wyiTdrSKkKY$O*T3|X}Aw0!zSwzqSqktT(xQ;3J<`tL*r{^Z=#4p1xy88f9^d3ms*$KS$a2dVn_4{?E_A*NcvyHWW~K)l?Pe9 zd&!iQ_5VnZYD)K1(zTQDdAYvXa~TfLcyRPOg#?^gZ~o36b&FzU&Fxw1xDcV&3N|2a zypZ@i+Z<=qzxEKBt*$jDyNsS!sRkEBOV2BCMaq9enWNJk ztzvUI9sH;nwijH0Dwa9wc;f@`Q~$q&xhP7DPtY_&!|ZJ*>pA({@a6G_-ZtVX)n!o! zs)sgggKjTSwAAfI6}+T^(ExN=i)y73#TGvP12#NFX<^fQvNqeT->Guib7>zdfkWyv3KG2zgp zYjvGXPYgA0IZicwZilJkBBva|YPbrYJo+RHo5oj4pM3uwZA_EvFcL7#M5RyFo#zs4 zceS@oog}yZ;7~et1~NcG5H4x4#9T>78wx?gAc0?9Zd=mQoY!%BtNk|$wbjT z#XUAV(HZ~)9OmJjm5TIY>~*PhRsO3IM><{Zol_IXDT$ZKa+KdX#D?|h=X7BQvRM<( z>|j6j>FBgojcrn59x<1(R-ky$U>BR=NjdZF`;~C&BX*>3yimA~IVGnEUgudxa%=OT zVbpK%cNb!5SqIt42_Qg+xLFbo)xsvw7mr55Lfc~q(-&6&$AP=mE$pYeJh4~dytfTo z1C%E&jpVt%V?rLJoe#Sd(%QErM&-6^VriG~J9JDqYcn{%11daa*Gj5d7Pf_mJ`kZk z`-y`DmX+cenpX=$iSBK`>B~Vy7gpEz{GJ6m^sGHVN z8D~hD|CXUNBncNh8s)S~!9+8}fiJ1T!$>>eQTb~)fv$gq7B)Soq5E(XZe1#WgxG&4 zZulor<#S&~H09pjW(a?q@%<9pznV_hx@UN0on;`-nAWBQa60!Wr>$G+s zo6)bGf7>C-SeU#k(X%*oB8rg+v?Odi$G+*;t@D<*!CtE4{|3M;inJM4fYP(sEI&-U z`nyydj^as`(Cv!vKL2eN)W0(y$A5Oa3I6+h^uC%NGNP?NqFlnTi$Phe2!^0?>zGqvjcCQp&HXujC^GWxh19AQ~NUT z?vAnLVusMxhg}upg-?gER;lfUMMD5uvYb?-kUoT!q^=Wwt74G@x(H)0vaA7}x<_5W z%I`ymQ~w5=OQ9EZfP4-`G^q%kY~4?~Kqr7vEPbN_O+)PNPwdcu5TV_E_Sb-rplx8W zg#XrnG|Qu>h2hn@bHFhJu`U;vu?_=+f;K%Pj4Z=ubH?)MaiJ&NO<~4rU0DO0GO)X_ z=Op`RU|ZqId3JJOa`Y!}QRnN6f2RV`KqINyc^YxQLL3x8u1UHlk`H>al(Y)L_7}6I z1-du1q@fr-9e)B>*RrffWgCq)6m-uHb<(1TOx3^4m+x- z4uKudv9|_$wY+wc=od|N`yjuGB>5z(7~Ivc-x+N$^sFQeDmy_LSL?oXW_B4b2ouh+ zu^Fp{183Rej9ek=0?QfFR@}OqMpDAX2`9T`&^SuaYRWF+W$8}#%d&rPEEueCz`@GN zbN@k~zMGvL5@JqONg|TJFG6pKQOD&jDlLPABHCJwMw0|T70;w3t8OLDVr)QumdY$p zsAzs5NRmVh@}J+KqSjY4cf-yk_A$iXy?+DVnMiut^qegYsIN}m{~cQ})JORIJ61k4 zNRYl`+lM9zZ=Yiiheiv_PqTKzItg$6$i@u|6jtwIuMLY4T7Jj24(l&Wsb-Idbr*&? zvu?v<+5HIt-AhVvJjz6I)+semNI_5lxTnB&MMktE`EjR`t|-+Vff5yU9K0LS(Nec{ zcUW3(N5OiBEy(p2GVibzx$*6KgGG@=s%<=FgR;nz40e>&x|s6f!qM(*so~ab_9QpC z#pky{u~a+8pPs+CzL^xb>NXoTG0Nk`Vp;;G6fx(fXmy)^o4q-)gYdYR?VcDaw7JcG zofzl)js-b!nL;~nIu#=-JKDz<7CWhJV-c0nl*1nHcHii2j)$|L%sS2veh_1PV0);71-W9*9Vn>gXfv$x2F{cL zBfLTJU3th1CjjcIX0~b=v0;`+p7ZgVNp}^>iMZ6G)|(3{=5|~q-je5itDMfhM;xcK zJDhEA{adPihCKrjJohcRJfTFam?oXgTbKxFUB3>kJJoFab9~dp(tZbx3_-GR|jG89c<}=Hv-t54rIW2na zLiY^x5px>u?OB8q+#f%(j7&e_*pF;lW>UM;TTu&36%5x}le@fMjb$NL=z_WX)fi>3 zRy5?0G3K>cU@_X<=Y3YMH zHg|M@)85koH>nmag7LCLFfy$-WPN*7~Jq5Zr(YZUF9 zZmy7=Ce;p+dnvAWWLg+Kdfz<%`ifcE5nV zv5t)zyBx*+F*Z2fQm70v{I~~aDcHpB3oe?nvAss(k+8yz3f)VNb`(v1MGG>WO~J{F zLm?eflE<~})N2GT)Immu(~IB{)l--2YJh@GwCzocpDybcxneohqfe>q55{HrRwcp7 ze^vjjOCSC2VxD5y5?ab3CzwxG2x=Ce72r)B0RfAXJ9W1_v@2}B#7-@BMpi2!a5!6* z)w)FPr z-@%TI?lwa^NFn=v1x62z*B4uXpb^DIS_iO`*~S2>4(j5xB;~wP4>j6dZIj#h_8|pB zu|?7O1X-IQ>%{IOYE54Cu1&PtW;S}fA4-2^d`wQik5S7MobBf4Jz@PNCgrsC zTaS#$KVQjDTBm*#R^ZX?Sfs~uf`<%vCRZM(aH8*#9$x{yvllg^bYu^G&Q|d?T2h1c zJ30*afF^i@amk!CE@^Din55C;_OeM6BHT~n9NKh^lKVA#e}Z>lMi`1doF@GM$3fWN zSpT^9*^ejevs~q6huO&q!5uB(Nf%K+ctG&Po=zq3<6~FGeX4pBe^~B;ZHM;tLJzs` zVOBrMuf1g^CQT!ea4y|F9YqK+RoL^L(8U8$iF){f@GKv%JPZvNSm(S!osZzF2o@vW zCIM<43JaROe;qzi|<%a}}gHt@!a=?)u9A+fx zcRvwx?xE-^o57mPj^=lF3;_ckpyx98mjfOD#ub3g&?c9~yxdA~4Pql-t`ojaWpM=? z1jl!`UM_f75V|j8a|?sq-hqXxRl4m4wxKYTz41z0xA^DO6YJUTS3-sUE7-5EOm&&k zUs=CxN@sgKm8iG0QCd*p<0mg|`^*ceWHg4RlmFmWC zfhWIC@~~WoRusi2So3*P2QlX__433YrhgUXReN~W^B(@#$Qp`#0#=eZ(QQLtrulgkHA6zRtUx^;At#TP)ZxufdW_XCtB+}_9*TSnP;KZW`9tEGc`H@Zf# z9ge9=E)8t6kp&mW3MC`ih~g0~dyPbf92qSiLg$%r@OiTDiX$TIK*yehZ{Qu1W0t@c zk?tx3c^aE6J-I^yhIRi4=2p^9(3i9LlEA)=E1Qq>hgQ7@9y?O0t>g`xoA~DeQ(pzkGdHv9C0&GzL)p_M z!@SPDP3}r*F(Erzxfm;yWqLT88}9eGO={Q89B9u;juCRshMu(CJXc;BK|1LV)SqexUMaIphB_GXlD$H*SNxm6gS z#UouQ5q^7L}+SdQUfZHCwyP;C@HK zG0Kj}=gqjdtX?+Cu7}@-XUh3Y%yE%y0lJe(llDO0`*Z%zrFKTgY3N94017a+-8IO}Z{8e@b=^o{y|Ye&}a?$ywyVk;IZs3-$eeM5TAG z`eKGb#CHB75|o)VIRywM%%cQyUUwsIt06++I0ij62tsX!L}D}(gSX1ikVP4Yp{v4R zBnC>qPl(m9GjGd zTxaD`xb7TCxV!ujK#Z?8jZcfeV#1pVlpUZH;`SY-(3h`VQ`n2;scu6Q&f}(7o8jAJ zwx!&&O`Jw)X*0Y-1)>&o)Wb`2)J$@7LbA>9YBH-Y4-vjJG5yL;?u~1o&yJ;Uut6(D zVclA`a^*nx5&uZJ{yMw0a*lBP12*;DRv``Vpz^IGoeWs&_6x;Y%H^2r+6LY3b+-B4 zc0!L5cJbXgLRJzR`ChvJ0n7{a%;t3^%KJS zu<@%B`2B3{ssO=tF5A5-R#=|I?yl+;yC?~>y+O1NeE#a3fE(y?WxV^$xPvu2?5UPA3G zwq}jVe7OLHNdrCPvnt{vA5jsbyjMjc}QT#r>MB;&H?ir{al( zw^MQQ;>t!9C-1H7sp9#BJF55`!XLxegw_`k-k{=(3BQ`J^uJU_ky>PJ5%Y22*PFHW^_w5MJ}odafF{x zak^%b_o=u@c!i2*5WY#pGYMa-;#q{ht>SrvFII7S_94$#@i~O&t9TJ`vphkKET)7^ z6)z(^O~qFdo~Yt$36E9rjf6+2_!hzgRJ?+455i$t(O1IB@jkKXo8_Y9gl}LaNXl0e z{+MC2;aTHAeCNk4MkUpPMA`t9NR@IxrDTFKPNi&9De&RBjZi7;X0VL)M)TDrirBeo zx)JF=sgy}7B?6QaDrJO9fmhJ&fJ*7BQsO|_rBXVnltfUzRw=@duSiY6rx{D)9@INcsoFS5@LlB5K{M zy?sC@vl$!o?mJUs}Em|Jd zz0wz(VGK-ST1(g^FNpf&TIK6AY4qi;gRvm04#An-b=|UxmdK%&GNl4Xy zi93LUCP#VC4VHbht&p^XS&#a#-vT_?!EJ7?esI3f_9^KyXR}|nwU3-$X0x@5lluW^ zyY7YAiWBDg9=Jc#q(^DeuQ=`beaPazNgp}T6LrEV3g#S{MTanHk@(G8@{!xuCEZJQ za4ganZ>71Mhc?j->+rDAczIJr9c0x>;Ozc_=K#sIE>&{X>)+msqiAP#;TvCJerI;? zn`oi!S{CtbXW?~UHu~EL!PS?&{%x>O@54U%ws())oi(4Hz8JrDg1*HUiRl~KMiQM) z?|ASuiSD2?sn{$k0bd=o8G3YP#_ipN2W!|1+gk~jJF)rO#|l3Mvx|sr4`#OQ5klk| z7FiK4eBF^{R2YQ%_t~V1*M*$-*~!Wf=l;=ZYdhN5lS)6<>VVu#0L$@p=Cih4(zp{^YZ+1bTU z2TNCT>iGf1Q|^ifHoVbT){ecrGe9`pp1rlx(4(;w=Z!f?An6&fURTi*4!o6DIz7?k zl2o*S^V1i128%3#xta?=pUZ1Y+2x)7!nM)t(axmKiAb;*p>h=lawJ}V3x8A()rEKe z!$WoTUByEcYl1<)bXQQ2t=Qneb5gu=xl4%mbcL zufMMpt6y9K7@y!oMK6GJ3}58d9`?2XiQWjqU9Uat^-&)7l8bRX3?uDfuMb|eMS*ld zF&bdx{?V4Yx)5ILB4Fz4V;Exw`Oi0^sCVT9Z#1<*t^01Wvi8FA{Nh=H+-C+J{CdJO z;=uY&Zpnsz;w|+4a_iJjLV2OZG*-5;Z)p2*=;s@7-K*ZUwW@dR0tnhN%AdyWZHydL z*$v;fF8K(pq=kNAG6v9Uu%CWU0gD$pc;b)T+#M zyW)f7TdsG8ZxbxjyRqq?M)(lvUZ1!kQ#(2Ad3lvn*%zO_V&vTL9{Oi}J-)$yNX)@E zDO{^kbE-tUntu9ki=B*aEcvrW=bk8^s-Twz@b?^M{k)qn$D1Aa{D5%d753() zNy04Wt@k%geH?#KB6maB)6D_G zezBRq` zjd|(>-v_9v7g5$k#i|!p1bKwU|Rc>ruRf&HgUfJo6coY@bCweB}hJJk9r^uNfIfj!GI1=XD}7jo6EqzIeq8$&(lg7|=YhN}En%o-hgu>8|rY>+5$!SQU*%v{P-o%GX zGu)OnqwBKhF>Lzoz^;P67_$k@I4(C+*tNV5M*k!44MeoOl=aeRWQ~a0s<7BhxbHnK zf2{H2o$Yb?-GAj@GMZKG?jXGB${Kbj3Kbu)&fkyd=HyDH(pw_D7P>ot3K$)hFG|zG z=AbtcaThQa-#nAtI*lRcsAxIJi{0JYEnvI9pDg%hu*j-8!WAF3zAAwaV?S375Jvd0 z_ItvGzXq{2d%ChEds+w^2<^sN>W?tGyGmR#O?ZVvh0-91ku_s$PRf4d>nCax&7O zK#b7e0ho#r_!SiVZj*GDevpA4H{)`^o9(L}=avm6s4f0`FS6*G0Kp@b4XO$9@SH;X z){`Xt+tp%x7xyV_eoaKAD=ZG>J1OLPpf{{e@|ZLtr6}iZe=+JN4(;+;Cw8Vq2EVzBn8u$f@k7!!y~- z3%%K{qb-G!yX@pq&jB;_q4FBqu>_7&lBK{ znL|seh>Qjlc9M6)j@NE~x6Wi8kM%KEU#2Bv!xQ$Es<7Am>o3T2!Dorn6}rfSl^kx= z8azhCJHT13Qmk_HTF?S4oT|znwrX_dOdg_aZ!)O5E)(aqzw?~Gf z{C=(IZxiodF^ab1^vh0#;qvb=y{kDYNq(eQ;V?C4H)kHly@mF7Smf~_!D?oy$B*_q z1e;fz&f_4v0yZCZB1yM-7F06dq?ab*wXq}AQ5=@z=p3s4gQtbAax29d>oy>V#hz*- zG;hJiobq8wC;A3)klJKuzdcOZc_-k~z(33A`Ov4CZzO=L~h~V@kF=#${T|R`d{JzcbiaRO0dHNE6g?IyM04$Gf5u*lP$g@m8knA2hHhrNm#{g;;! zJE6q>?PWybO}6=TJ7ISnyLh^%=(!oc%A>9o5zinRPeE4Ljr@^I*A z6pe32RM-2ccJcGMFTXaJWu1x6yBeW9`^58EE-@ME4C?D=Oa^F=hhu5<_KE6udNG{W z6#`yD%uwQn(}uG{w9u4E>!U|K2myy{8K-+>}<;@%XdX6 zhcmgT%H=1VoXY3`h98bo_@n7(f8Yj_SNCHBfAkeTnu2Ve4$jCUFM+}FV^H9f|CXoH zfj9B)< z$R7=5!jI9N$GxHsHaz37UaM5kxU)QZs9H#LQx6!?15$dT+;=G3@ngSb(Jtn)5Nx^l ztkK{W4er)ptpbd_S2! znHtQ~ph<&88eFNtEgJkmgO@e&ey{Ow`~=4NlYG>k;JG!)Jv?*rdS&8a$)H0&UC`X8eiJx;8+A%{17QenguK z6&l>A!4(?3sKGzNc%J<%ga`I)%Oz4L7LQV)I8lX}?*J}q)zWR(Rh1AB1&AzdytH4J z+nO)h+wElRxjtaDZMbQPJ;i_7j9>53v zBhD5-3Tr{kG{Juj~mV2_jl&} zO&*;81U=^;>(2Qv_dL;}$|HmKaprs`xN<(jT{xdu=T4l@P$eP1xo0yp1Jy^hJv)0m ztM)g&Ic^_9Q?*(2V;`l~0-A9!5x?M^g#A$~bp~+DKjQzbACC9oxbp~20-HB)#x;+1 z=9+hQt8&SA&T8rrZ(Mf9Bdooq_Lq~Jj&qCkhEK)QpKF%SaZQZfQqUC$r%{QEDQx_c zptS_}EQrnWddYYTPCFIjLiXLH=TrV^+g1wyoMup{RY>%n;xxSnvX za?E$gYNp7ELmrxB$xT%hb2;uVLX&{zYD1!%(zoK&d=TNgrUF{nb1l&OEtb1vyt9iB zF!Rb8122o6Gx00rm|*IxL93MfSc4N> zkQeK6sG;GXYjM%b7%_x9S_M*pUqJ{}XRZUs)0`Rq&LDo|pr%R}EyNG4BlOZ{H?PzM z8UuABt)WcGSQna#+5~#OGOT#`@;ySYYwtYeLz+?NpRg~V$n$N=f9CnP@|OiZ>Y3Kc z{fOfhBaFIMzFFYgdJwmWF4m?)ba|b?8!?XMf*s$LU=KUKM|rIsAMgUP8nNyDPj&I# z&v640n%eQdTg~<0>3Bd}&E+rK@!sXp0`F747zGWWf@(1D|EGecpXRs^5t<6}ZRN~Iq@wxA)4dLt8{!oEY&00p2Os@=Zw4)@0;%vtc!5{<)@MrtAV! zVLs&=#VsYu@Z9P78QCU07%i`K=KHsQn`93&O;E){``D$;%$|@od3-^3!Q3P)Pac{2 z!DJpjH$S&r@51|fzeHL378Vrb&M+lUERdQ6hk1M~t+h1JDjGUPhax-R5o?*%@;Evf3 z&n=ifc?Rm1Q7|(H)xn-d#b_UL6OJm(C{Jp|d-KWVlSBA%;IAn+ zhVY4${J9yzZM0CKH z5x4-Jn*(?NVT%vX)c`Jqza1S;$*U|^_<8v(m&cm^QT@V5c05u)+)5H+E2?eRG5Q|;^l zw`%w{z`m{V{RlA42V93h1#AHP8KH;@0}OAY(jx#5YWNYrU*OQ6hP>q1pC7695MXCq zjZu%nlg7bGhCgOrLHZ&>txN{|L`mSejezGhd_p^p8;mh6hR_VaRS27aZvZ@iPyxIK z&=w6{3!Li&IS94DHv;ZQ$U?z2fE_xk4TuIDX~rKaI1BJ|EP6-$*6}95YY5aUGN5m) z$`b%6Aw+;@Dd1-aBy#qe)uvY>mGVnw|E1u;x0xtvHj6jvx0yrN22`XR$;AI34 zFP^IdT$QMH`)a^7Nid>8Ukm8chg`cDOV>W|4RJu~I1mF#Tw-EAx)6dDy!7a-)-~??5)GIz29QPFA zZQ$Gx-25VJF{8zVOhTwaLLOiV0=1?B@Q-2WcF^sIb6htBs!%fEWCY@w2H0u@S_U~` zfEy907YQCgAbx_jkth&!ZWMmW1%c@BBXj0zEm8w`DibvY(;2{e1S;SLVAu<)P=cd1 zoZxZ|H^Lc_jj#)O%K-lzjok})17OS;RZcA65)EGp*m|rY)67Kx*@i$}L~uhEYKPy? zAviW0Ctu)MfDH)AzzOC~z@!D92iR{Sx*a&Y;aolfReJ3th?uNSx&Xjg2-IQ|;B|zR z;J*R5X9~u@0+D@F;NPE)2?cy5pw|p^J8&PsRT{n;uoYYqL=OYpgb)#kW0PiDQ1y3Vj$6~cY(SQpPJW=Ulz@QSfCyape$}kzBeC7jgSjKU& zzzN1JR~s-K(3hdnpa(Fl{~>Rytsf3}9)ae;MZlKtsD*g~jzFN+WPO_-Md&2vpbxz`qe_t~3H#tJS$u1~~2%#y=4WS*O5+K;zd4c;PhE zD(DvhJDyPmMFYNtpmaOo=TK?zYy=Ff#c~850XQ9j^h7@3b`7rp9D81!XIX&$m!P1) zA3*<&S~((j5!nQod%7$u!DvZqcgoS>Q9)D$5oYB;&7^E8~`YZ^`tXY&3MKSA>J z5>Akuw1g8R2Q1+P{jaGwIbq3XNOXduRNO~F=x%{(O|Vi;*!r`eu0En?KqWlejl*EA}ATlu|ff_suktp-ofODoThlz3M>qGOPk`3A3&|3(M|FE z4M-1v9{0h|KWs#&HlePrKV4!nc|S(jHWqN(!3EZN>d#2yq|mT4fZ_W=l%WlUhj4NuIoKsi_g98b+37? zd#!b^b&eJ594eMORQ2WMmHh^kH~5OHSI_HQ*XA6P_t&j~e?_yk?q3QnulrTOn!2}u zZ*}-r-Adrb}3!E@&kqy5vubA>#|rQPDVAa?L- zjIM>_>V=h^0=&ALisqXjcjPPl@qjzeT!fHbzx4(QAubv265KyNX-KGh;Z)O9@Rq{E z#UW{Lpa+{Q1oToeMgcfCup)r}_@q(P(z≫3)+~fw|%E#Ia9=fDUPPAW>G%9=Let z*FrabWt`J0`*7j=P!?(5ZCFw#juWl!cOG(_&5+y(k90|tbnoTzoVcX6EJj>%)ihYN z8vcfWD4nnw+(^cvTcQcLD$G}|aa?nFj#c;DL$<+ws9*?Xe<1O)VM^lNl*DI2Mos)& z&VMBC->{c?f+1}zadLH zCclMt%aQ6)cvG;X`-+{@jTpG?0hPm0n~q4<+~mxJpK>fk$#Op0GhI5lq!wYTA);S> zP@ezDW)mgISTiDogmZFhD3v-lSsoL_o^}WdISYS|622gU-*1nWO_wfaNvATTqjIo6 zd)FasszE}O0$Wz}-=f`vrdQDLpQ9)3r4w?Umzr3*bw7IL!4l-ulCBpYW+v217v*tU z(izg>Ea{>sotC%xk^E^mO{W&cC2$+t~zzd_qmhE#E87R*TleDG5kx>d?aql zdukH|xYX;TiLL=I6)j+J>y z?xRJXTXM|glyJ^wScbf6D1Asp9i{?#h?WHrexpR|)`*I`7(t6}R5()=bX@{4<`RyG zmL(-l#JHDmF5#HXP=OS3teK7Jd#WfI%sGs=yuFA6&xvx;k`ZCF8L}aTaqPKhK|~(O zl{xOAT$_c2V;BgLt!S_1_ieUCxh-7WKhtYV$7a}b83|eq0T7kyoeGfvn6`}W3p#~BftP0&yg)Tt|O6MIg_fm!GGPiO-n}H!pj&=nYO_fhG!#chj77^^bmIr|NNF8}X?W(_4EHp0tob-h57t=`9_+N<+jTa?l+kLv!fxt2`@YjrDX>3Mh^HWH zwv)|}cZGHH2%hN=MtTjYd{H_huTXUr1UrGhq!z^|brpvUMeB2Z`rRFhF*-fdrLW~g zB&v)PG=-fy$CD5TF=U>^%|``I+>OL_9oY^KpBXoyn-IZ8G#?g=JaC8AnJQn6yyV);|Cn)Tr5ch}2) zUTnSR1g~p`wnxC>Dj@KUCVz#n9_BPl+k8L-qGNG-%E*Y2vTj+>K1S>K-{d|M zpr)19*^gd8L2{)dqNG&XA6XV#dW!{zMe7<$7Kv4$^Kz~B88*YO7uY863E98sh;K+$ zYk`<)`ZY%}HebGlidbevBOXUOZRkDIPrNQGEBH%~mGdIxHSM=FxQ7pIHq_?r%gwadAG7 zmAgc;Glt0g?VTvWOQfyxXoR`TqZJvn2G@K5ld5}Jm^`388Vaw@ZZwE|66~s~gxcm% zY+f>{C6pb&LS?ONg=Uf+LfA;3o*gbmU_U{*TKwdK2*!LOgm)v@9-m&px#wxDZ##=Z zU^0hr6xA_yA#H9yuOl0v^w1n zERDPfa>-Tw60{lWtd3f)cvm42$>gnw}K}p%lk5 zq4^gv^-<{*QPSHCeU+G(jg0qGdlhp|Hg+6?cVVo?E; zoG+pq&~+uT?A3tI;UztIjhAde(TG_k8Z*vPyCy9k%NuzrCdPaF*9Ig^i6QI95CqBHwr_BhEUsSvi;o;dTTiR4QNR0= zT-S%OpeZ5aV-fAST$JJk5&N?Mc^oE6Tg+hQ5zNXL#UMX~8Rcy;SP&7;$(b0>du&8a zpe=~xJ`qB+6F^He9jBVMP)%b}baN13!jL;5FF|`EANS^u}?x8Fj}T<6hjmnA)dV3~*faUlD$Q}lo+^busd&=8{ZHPv|@ z>L^`mc9V+&+p`s|Lt7T5P%33Qc_sFOq*8NHwNvuYzgR+8FX7C6_H@{AVS*z&5SHaW zJ)pfu{_7dO}cZ7B9 z+$kh8;cBkbkRu((uq*SGW09AZ?Oy0(b~@@v(0C9U?^gpVQEzJ+PaXPae+Yci)mOMR zkkxe!6IdK;?wTv?GO#|~h6%9)nYCNIa4(Y`?AA?K*@NBeHag`4@Ajm6fkaZx0dgz) zrHZQOh^;ev7Suv_bQSt-Vni~j;VFliUvaYk^j_#}^s%4Z{c$>?VJdlU-`8<)ADYA2{aZF#VF{G}J$V&*IlqLG(t@keidq9`sCc z%U9*f66MieEG^y;^duN{To@#qUFCSigBHPn_C(Il#uvoYf>0ZXDm2sF_TwsnLKDDSQed=Csq}^5i~NNcb4Oa6Hzt>ad%T(WSlK7xnmkgyW)GpC{gr^ zD<-!oL2e>{YTZ14aEi@+pYKn5Ast8ee?NJUHVIc3N; zcSeF~CI|CWKd!k&?v9=EJgIhM80bH^h~Yrk8^utCg7 z32p|O>=Ujw_Y@0s0ms>vdwA3)#j!KmWomF=SMljwJt?NxGL)D+{uWe>!#Noht=4as{^B#Ukl&H-}aIBp<$ z`XMWd42`m}%+vC85 z&7x^I<^-8ia$rhXbYm&WTHPa2YMWuj9(W|3M+e>qRmq4LOXqCp!hFSCBfEdVM)m1Y z-r$DOPbJHt^KK5#&|~QgosU8G`4&w7Fj1#Ulm1p72p*P=^AM72Tt zqoQ?-=mW+;Nk+^++KxH1=2E&y^iL*Z_mTqrMPv)rqO<)`lH{b?`AnL-fz z15k31GFD6-6oZ8*ow6BLfh}d*rz@7AEa^g)-5L2c7navI%5fwTl&f7>dEehWt-2%} zKn|;$Ye^?2v4J6-$|IAIrZ`5a?>*RbX(mTl^PDiBEWu6`7_A0=qXjzF1fU6;=P9P2 zy4w`dn-|T8;zaX7C((AoC2I1yklnyf-fJ~-U#A;ukV>PV7?lueS>A_Z=`TNi; z_De|W6K@ZKUAte56V+2E`;SGMl6x=8kA}$i7<}51fVL}E_aMqN^>XPGqE)xyF2|;X zc1}WpdA;-J0eWCvTbrgyb zXf;^r^p|?PGzrOiBH7gUiDKvdXf!4tkSb#t_qhBXs zu@76`Z;(C~8bebudvzSQ^M`(!yvO$tO0Tg~2GCbBo6LGY*>m(3g$Wtd`Gnj7bwvk| z@xw*G_jj-6At=}yMFg2CBF#Uyi+Di9vpE_f%M=kuO{D&3w*Sc|J*y|tB-bywjJ<|U zr*q+1hqMu)+ep79*l_$bruOR}hb@Qo2E{A5i+=MmvWm2*-W4bhZGdQ6i<04TC`}9R zKe2K-U^Aa11=oR-CE4G)lLl4n2D>nVo_@%#rTG*UD2q=$49`gKRf=hG!`APZX%sc& z-VInIY=+N}GU^$Eetud>-0hz1^MjJ~1WMW*@qlZNKmxgHXe-6F8Rj4wa!(Fr}+R)5b!pWJ0}ZQ*0ZoKNqzqMgeJ7q zYBT%@ic%A8h6a$Rjt4e6ZsrFz=+6^!_6C;{?Y8^{$avv>)40=rwJQcScd^d{de~0RjXUQXu+yYmuE6_MmI10M@mQe(p-D)8?0g7*?|m$$3&1=^W963sm*Zi zL-tcfOy0WRs03FaR?$=b_A2G(c8zwsik_K*GS2!C^0c^3pe~n| zHqhk3xVN{RZRX^~qgj5Y;mM&SWYN`93sT>Z>an=*A+N=qj8=fpRoDkm!X@_)h+~Uf z_5s_KIU@aM>`pc1QVOy;Eyo(!lO!bfZh)rc$QZ2dDys{i8P>5#AJwn7=b1P#QaEv# z%^BFQ*IGLz;!3~)THBx(^08cPp@}AEng8a?9207r9wH=n$;~pCbZf4i{O~gSVc?_= zHxHr9&tmB|O1hs9vD`u7!oEYSbkM@C3k4J%HTgGeQ^A)}&RGz!89Kku{KQD(EAOi+ z7;PuHZ@fur)@E|I8LGiIyI0+0lf*FLW-VJJ_7K+HWS@%tg_=9;j@YMs*Sjcz`)AN! zG$Gf&NA0fkw}^?gw;r2R`Po5jOjCm7E8dY4 z^_8eTH`O;^^*#EUQjsdq+$!Uy%Gf{*t1j>WAhjMDTmb(7wyzu%?;4ze+0%mqJnNbf zjBQXH$vDXCzGkltj&Z$y8_Q|{VlTSIP7Mw(Pg9a?Kc=-=+b*2LbolA6vj5A+D!W5z zlV3N}e&*P9^aKvVqSB`7j>VcunNe$-dP?r^LqciT*AM}$5M1wIj=9m&V|W8Rtzm0; zHYYnGsHYOS9`S94ND8&+Lc(eA>-6DlTXrwwwX^DsBx9J(khO`1Vp_z_Mw9G%rb zUnTH6TxVUCL(j6EIRl-gBZy}0S9yf>$?f4>p!y~%J`W~y+yNKr5y{nN$lb`E&)p%! zZ)Y(>QiU~J*^D8X!e`gm_930iAC07TgORTcQ_DKUPQD2Sl!&{z-^^n^Z;wuK5sjk#JclX7N9-J%6|4o49H`&pnF3QYHk?`Dpv!=mZT+r(}C=Q z%~1L_xUj+ex}>&aSvMpx4N0hXtyM^Y)eiLzFZ=^rD!P5`n26+UwW7JT@4P|85Wqkj z-;debLxTlz1lvEeP1rqgzv>{I3=y`=87W|75Ci9jwIDF7phjjh!!x@%IZaJ1o_Qkh8!Z5q?BQp)3 z8oc5vl~}o>MJ})hNLlt&KkP;NzQz6=o+n)KU>PG)g>j?Uq7kutJgXSt%?^(+3V#o0 zKabcbl+`iI$RuI^FRXTC&lKrrO#J?28VGE$xt94r1&H5Mx|)X2$-yT{-RP!4P_(>) zQwaifBY$SyMuiFcPOu@P!a_12VvfwjP~*%G9+gGb>3V_20K~N1{swz>R7i+hJL9D{ zu<Zy3yT)Nk6fkqch^pyiT(~U-=7Vi24~x%Wl+| zUnMZ*s?K)g>8q~(6L2ARd4p$ z*iJ&x0Cst7MEOwI2U~2;xHpkrrHW;xo`l*Au1AF}>xt1~v%B+hJqSU+7PmLsY!=kQd|$WFAQl|g z@6O0(3&xEM-P%amI8hngH-J{lFu0D&5WJ$1{XWj8e@r3<$k+4xCu!k^!=<&$~}zkkmT zObQla4zuqjB?y}ivyPJo3BR0R6DOw#Hi1=7PGq6QJ$;QjWPz=@LH-3t3~lE$_XqY= z@ig9%9WQ>(z43b-Jygr%j^XT?r@9(H1V^zvzJg-P>)^BaHNPY?5Lg?h4`-h}mE32z zDpZyri=Z+r#^$at%0wM?gra#=7A(hu+ir!7R6L3m^3z5ZHT`YDvX>p3K1+D|Rn~jP zldN>c0ed5}xnE^uBH5N28Jtl$HZF4Oa_e-V@a-{0Zhp)^P zHot|1%F3b)qu@ejnb}itDPa3%M#SD|#nk;iluq2@)5pe&c5a879Z`u0%#b_jPDYb& z!a3N?U}!#Qj!6otF1D$*b+zjbby7|Y@+^4UO z&9MvmuGtZD`pOlS@yvSH7qG*|AvIYZc9q?H#y|WJOc#3dS_~RojpRtJIQXz?n=R|0 zO=tckG+SDJ&&ayY4s!NIJj80b!iLO#Ixh;kLcK`|l%Kmom4WuaTNQ;w0l20{j14JQ zOgM-7^Rf+QF{HfjH9Co2%9bkc87Jdj#xgN3P4O`2ISIUx^_yeF-^p`)aws_6?uuwT zlxfXx%~+Zdmy%wAJ#+WAxXW_l`lR=%Hw8lr`!98AW2+JFAu7e`qn5vRjzKu{3cENb z!fnSIj3>JvAl_CfW|?q&_3-7rQ~_G^0%^ z@1f7;`{(1in&~I~XRiJA`zv|!VGGkyo_CSeKO2UUH9s5VOAUbpmS*%!G2gc<^IT@9 z<{DDkN%&wmOE2x*^IvA_@kYco*bKGsz*@*-_iH|3YdQ|qka}2})Ci3b%xt#7O6F@z z-$L@e<{G1iR4Uc&U^ARwOq71fYv?^xUSDYifl4H<(i(s*nrjTA;-DheOoQR>FvzGU}4ar`fHIw{azNGX)^`E~bQn&bv^`2anSA8Q;K| zJ(GK_2!jSa_QF`>2B7+ zN3a6)W52P~vTgx)zam8?HvrU$+dKv{@WK?1W94O`66kYYVXpZwV_-4E0Glab_PcYx(>0E(u1^9 zuJTxp{Z$s$|8E4g8T&1uk#S$TKlKt*zbbWC-bZ{bN5w+VoXC+5;c)iwHuN5@Hog>7 zAIx==zc|On%?o8mCWoi|QI1X=i>9#7tj9*Pe<#{z-N7*g8@Mtjl%VNp+5=j2jW+-~ zjKZ?d4tWIk{UAtB1&4cWqsKQd6a3oWavRrtxG8`ZqMy5 zSPa28*x4x=UKg9eG1obzezjm}vja?G&8$m)w8zUR6JlnH1;0bbRyDI}`Q3#LFR@km z;X-ON+nt{*OkK$S$TtY*USdvDdkFUyvXrR_P8B#n;i&l%Gfj8(e zO=sE$8MW&04CxFkt8}M?^Nr(6s8lP-;XG5i0L>;O)S>{+7KFL_!xwkjht-Xo^sO)wBwRv^XI1ad$Eb*<~}V-*K(xavkv}^;gvM2iOqg4 z!2bMr45a1Hh4fl~x~)X=PLN1S9ZCIBXmn_x{Fcg<%N`2~r&;53i9+7Uoj&tpc;Tfz zEaUk>;Y+q+e^Y}^W+qOImdqNu{ECs^2SZzV5>K-HN*{LQ`N_h(H7w$VaADY*omnsB z@Gd)#qlK_*lA_q#FJ^=lDc)O4Z2DLIp}{heee{(cAoQPg5Ur!C5 z;HcnxjHNDID!h}o^UH;cdHZn&)GWI^ObGm(%_;X2JWb3}p5pY~Uh0*Gx$Jm(u+t93 z^U+-PdwD0pFM@ehGz!ad+3kwW!tG5v-&pjz5F5D;V-=SkakS-FbVM;jwoMki8kPS# z*j}&22T`wZHl?tuN6v2)YlH_~FxJaH1KHZbInJkUVs6ySLLT!dN*vH~80X=HtLeDY zqPyhO;BstJ_0|e>T28@_*tOGscd)r{BWIdESn_VX(;tS}$!MH8dz6E;uE)3Rm7@M3 z&Dd@t{0D@rMTqTiQwjd6C;Fj>*sn#&%v{;c{px}Ca!x0aPGq7 z>LMo{V562y^shUG9Ud7k@buxt{`N>?_p{?mHVTFN+3c4i{Jx%v)+lJjsGN_DfO2;z z@5M^YK72XU>um%gqz{==(=gowhp?+J_j3LmEh&B3{Pjups5nrEYtF_mSf5v(@RMmG z!U$^7P5uHDV(vrWvM*TGE0Mk*sk4;0^A$~U0dqwDlAU}dFrX)iquSrTiN)O#83u2T zrAwVIw~6sf6MV$aamK_xee*tS5F}M*?LIbmsZsc3CY!aiM?|IzB|z^oOrAN)MLs3o z1sYG3@MafBm=QDCfu+GlMc2`rV53-Y59+lqnV@KIG!Fd~`ZdQqdWQ~k2UBTw*g#~C zBvKEXLLhy6ASH;I{q%IDsj=9nfWj*Xm?sb(h438d8tteK-qX?k2~&I91&&qp_OTy` zTCg{I$CTgMt4{FKKy`g;!kR@ByeKs|^8zOLBr=yuw?t+Ai0lQH`PWeKY1>%X^>6F< z=OjCszvoT8(PHb7V-<1lWF*6^>k3vK4kS`*a|d~Ld-@oe2naw;q`KyAfpVEvg4}iq z`3gpywwUB8>&V0@lMl12O9O?`ud~*r!-S4suxv|!knj|H$`USoTDx<(W=ZfY_^?axevTDMU4@VwRwecCZUB>HUWPt{%(mTFG_h!0EeoVoU>5@W_QQiM zTu7dfuY0dr?vHaUH8YK)SO`<6)*|9m&=skiDUSrVdb1~%`}i#Wg`0x;N}r!q@bjBsK-+g%kr zVAp!AJo9J{*+4Cfa`^!~N`NbKfek%{6i*Le4l5ReUE!S#| z8I3_MQorJZta0j7Q7dcP1-hl^ey`=Yd0|wMcdK{DQtnD> zJ#cn3A;t_f25g9-N{sbdjA6CxuW9+QbN^?IZw@GVAjYc6lt){PBL~>xf+x$RPobKv zQ|wfQ9t_RAut+Sp1tsyv3FL0>+oB9aw-)8WgNK~@4yF&(h=Z*TR{99@1MdYrA$){{ zCASv!ip*gW`P{du%s=GuuPV1{Z^MsUwOgF+aPtaP-TS)jz$8{)?4;4wPDDe;p`qD? zl~V*^5Zkvh!!Q36+NzpeZE`PAZQHg(A2YvI*M-?1GkZ2N{A3DD#CEg;=_EhIYOczk z169ZQp~vL9t?WgXB;-$EpRy5tuC%YQWHu^$n?7Ozyd0it8ggR z0Z2f8^bv}f-1MeBRSG(>7#k3DaJCw)M2i+i(LTpUP^yPAy^lDel-r1P8!YI659Ieh zVsEaFbUTuSWRwgYj3YUwKPOQ7yo%2)4L7S_q6~+p2d8k4Cb>wK0Z@(Q4*pN%(RBtkG!c zZ{KI{uI*$$J06{Id~K4D^d941kLVEi{^K5h|2>xQ`ZIxXm~>=HATxqwf2g9+$?1`# zmopO(^YhmiKk@koXnIuDWseVa{62W^=px^Rjm_k(HsD_Dv(j+wDahrUu_0?E1{!*uIN+fB`Skf;Tmu@G+$R2|ned=Yb6 zf3Exr&Vz~#H6ubym*gDj&m5^qUX8lI0Gkp#0`Z~t^-HkcK>@knnn4=WfL)EcwCgJw zPXUJG2vCNc=MWHNt$cw3Xb@f`j@7VMkvNSN)AW+%ooDfv%rm=!q(El(Mi%(0WI#!$k$PBeye>9%zukb-WVfv zTgH}eoGhVK@-_zvHx{zln-himS#0y>Clc>xVFt9|GU)M(Qt1UG{?8|-FMuH(g+_6rC#K0| zti$^;tpB@S!uA)~=y%r&M_*0BTB4m)dIQl9=TRck z7NT2J8Y7XDe^TihqA#g*Ezu`cx}NCrTGi7)o?R+^iRdjV-AMF0m6nOFQt1|=D^YoQ`%puB2SGPu$JhpDm|U(4Jy5b z=+!E{o#>@1T|@K&m98baM5XJAE>!6TqQ|QA2BLFR`VtRIhn%K*8Yv)ArDdX{RJw)e zAeC+<+C!x|sBGC@rJaawH7T|CAo_+%`w`u!(ng}st8^66^`Of!aOf+R@;)^nndt2* zO_zW2W|bC+UaQi1L@!tAu|!v>bUx80m7Y%Ybd@e4dZJ3tCpu51D?pda8LDR~1thC< z711#&y_RUBN^cI{|M$=su)a`7#W>e3P8EMhmS6Nq8RtZ=+D(9HWnGa5Xm9tynD9Q%RB$c^QWs=qbGfHJzRVJw)FauTQd}3;? zt360T8?%s)^&Rx=@f!8!Dwg_jwD9$+ozp+A5ZJ`MJ%ydycdpnQ&I|KiWV=2e5WZ+O zwrIPsxmN8~8&td1CNMO+RotuW>^>hBx3A-nZwJE)wFKKBEV(CSJIp-RYlP zVTzG&Sljb_-ikb*6KdVq;4gxNMmIM7i&mGs^eLfI^&S3G>GEC82pK8X7Z%#$`lBkjW5!Yfi^ z%aGRIn|L~k7d9v4;lHr7L%$05I_x^hla^^PX~SC0lYdG+ z(VB5|?s^zUbK`fJRlCV@pk_B&g}TFTGB35C9d;*kXaAdF^!8h9NL_&aL}bin)rF=; zw#SqEs7Z;cWwf4(2GeMe9HwP)e{owD@+%-kHhWJ%XXB{RW70>ak4hh%K8C);-GJAg zHF1H^3wN+Drk#bPi@mggzSDEn%_2RYJ@FqHhVjvd*N zTfP&w4`|+fj?@79%HvQq2wb#|%z^fe>31?5ohtN|J4pkrwkVrrT+C0zGqze9wpQI3 zNT!VQ9xS;vW=U>({p!Ou!<$NL>MIu_7WO*MX_a^Mm83CkhWQBUv>*b;q;Jvu6Ej)W zr(wdKH`%UF8C#$|>kmEDY%V<{FypOCv& z@$HFbi#|Io%!p>gc1H^_)$FC+{=V;YQ$~`j9S+VJsgKHQnw{`|yDR%-cTzyog7%&r z3=-d2sksM3kHOHD{k=PASf~;ejr|Z^vLHgl*9JDDYoZ|9G#`TWbRrLDlc!#z+S(sf zIw#HKnqIMx3(!}_Lqz`kMam)weIqx#$Y$0A3Y$i;S8CGZOAuf-VvnK-I78qcmi+^3 z{)c6M`x?cv?}u^0{Pu+O`T@IL?LzA=NF=3>-UJf&*WGWk1BzrdwA* zCBD3eer`m*SkG?N4Kns>muv2iY1BQ{efw&ZqO({)za6k;H5*?a>VEh6_FOBUXD`=> z3ImGS`}H9y6W&96SYs>aV?XWPJ88Hxn~WcGklaTV;v8Emnvd~f zio3^Bw(&pDDX=r6kwU9Dwk0}P_--nzi|*X1H`xTFjK6S148j}EW*<2T8!oMH2ZLlC zViG&7K>QhE0p8WeKDUBnc`*?}kEv{4OjOv*;Y$3!p;6uJ=;qQ#He=7CB|!TrRvQ!6 zwG;#mxWVmV&ABLB&s6s#uOjWPV!>IA<$bSm%r&-?Axe|x+NEXdT$Uc&Gk-Xw$(W7D z0NGOFFUn(7^MV97t}M|y)LGt%t#Xy>#}l)h1P|>c`Gig!clT1@x$yh8j0Cqm&?%h! z3Mz+Vp*~x3`vtQ89yUWnCw4A2%I>(8V}Hj6+W%z4a5oltv$*btP+t!tOccw0W*FzlhRw!ttyv^CyrDV!CEnr}G*3s&`bgoo#?th-s){ zCZ5a|lDy4e2hlqNU_N(K{SQKM(uK+Gv?BE6Q3Q#W1l&e{h~+vvI;cgg*(7z}D{qv# zi+7t@j~>x>GnaF0QjgG1Rbc7mCTX#N-XJ9+R+h!S!K{mf)LT7@?bERbV~%m1 z{QT(4HMFW|Hxg#sBit$8DPbG-_!?$N=%a;L(nLJoR<8oaOYG#Hfbv7oPMX!<55~CWM9BJTzxu%7vIkEAGyTJRLmfOV)XB52Dzsz>v=lliF>m#Pw>4t zZG)$~3_nI+7WGbsoc$?vmd*|N$c`|$;(^ffaT{#I7_aI`ix7;_o(q(J2pdCKgx=(?8@e( z=q^G|z1g{mVTfW!i{=N7^S+>M=T&gzL%8VM9vJ;DN`bvEY-PR}-1TxU{CE<6L0;Y# z@S+-!IFo(4Jt-g)Z1rUklw`jw{We@1;0Jo-9d(S~(M{;NoW<_&W5-;4*^#py{bu~E zj!0UQ@-nQ0@Z(Ony%_N``~9qc{4VU{(htz`X&nP=8-9QR4@xhBDJo;QtfxG4Mq6g` z6p#{|}<;-DHNqqY4{RPqr8(j zTuAWxbt(GvQ-rBwPJX|bWnYMMD+9+mB@AlHu9&@Y;aOp=H@kl!Q25G=`F#DWld~`*z?oUZ3eR^DQoNbI1b$ZN?6R54* z-uPWmefe!~{9>;w3l*7R~t)G0&@=gmu$c?$tQq`U|Z5szI>bU{7534^E!;IP3PtRyR+< z(@yCeHwnE;p#TjE2pH!M~vc{OL;bINH(@+t@pIKTL zjn;5}cy;<<9>3plwL{R2K3p2rFCte(d4h`HShzHj@^;(RsB%uGsvL^v;nVhWo>&C9 zT(}1OR8zx!_QtjD&T$#Ml4Qfqv)6o`%hMjJ@o6_kpT-BOpB4oARA_!Wz|bd7^Fu2t zpN9LYJk+f6Dbf5mZz><7=I^NaYwxM@*a6X}oXgV~*pI0n%^#q7t6J3vur(Mlz=9~`?+v_g>IbRNLS7;5%h5H2CrI= z65gKY?6DAgdOL95g-)C|((#V>Xiw*3SV{N6AsuB-bDsV#oc|G5&ff$qbmsgMojCvD zhK9EK9!%i4v56cvP0P3#^o#%S^Dg$BOQHkk67Nv*-wIv_8}M1Uwt_pm>p1sBU(TJc zJ}vMLK5Z4;)(fh$Hy*|ypOL+*uM7O}UTtB2V)}cfbKIL+8T`9DbHRm2xZuRYTrj_% z>rm@f;+n^Ceg&9qy6WL}eD}eg1}Dy7f-7|73@D!=-czX)k2X%l=oBKznDc6ge&wcgg3qxSs#0e_NXs59YWMxIZ+zdFYmKpD5&v_-f zb6%@l^Qs$p-cR7T>S70ebbMQkhB=sxB^iL#;(gjD;#m{N*_BbeY8OX7%z>)VU|${Lzz1}0>%HrXF%``mcTh`v2mFKJ|3mRt z$2;;}hSQ`mVbb8oDEvpZPnud!PY2Euv(Yo&z1FqFCBLn>juwuKhWqG8bpY_*n3;&{ zjkw;JaNff`s>e9;eX76L@xE#5+{TGG11M5??pJx2?g+W=Dn-aocSQ) zOm^l23~hDDs>S4lQ<}W`btgWQMD{xI0qt2?>+v-*IJE9EM5XZeocT~k>h4k@uDZew ziRbftRC|oy&fslx1ILZ}vHA_356Nrm1N!~BM7R>I!FG|%Ic&`}|NpL`e}qhzuO2sa z^*7FZuCD>pA7j_>X@?T`e78IvKab`-4K0x7!pG{BMuoD4jv$3wJ>P{7@}WkGv*$Sf zczezWdH)i>IMn7t7d|A&A3wklU`pfz5)=3UK8|nGJSAA*tLt6(;A&@AJ~WAvsKOaU z(a8N_w+NxRpngQ!n6zM_)ON?jT?*co>StYf-<5Ip|8Kwj|MuJezus>z?!eD89OVZU znFf@VmKM!6<(1AUD4IK$bE@{#^WH(+GYFeqR65mEG+@r`xpQU}4KS6?%A5yDuDXk! zw^o0x=Le=dVV5&!YGKZdX{A$3pHIiiZ9f7jK8_$xon1JqsFb_GWoSFXB9&GByay(t zgD3xjXNiDPj4+kXnElkq=SzylaMe4!_>P`4Rrb_UQ&~yY?BY3Ns`2ZqhU!C}e3g5o z{czK)QL~?$F}pBxes#Gg-?Ms|7ayhnjmw-}FsHC6Z_bR_rlQj7vtInI>hC)8L7@w% zU~1L_=9Hll+;g%y^-Q ztG?;Yhr6$)v__Z;mE@}34g3K2D~inUqV!p_<`h&9HSn3$wf?-(D=l9=dT4ruOUvKw z<->ns$69RmyMOZM_u1v-y`yfX|4*QLjLQ3c@a`cY{15i!C4!nge&jq*hmdU4dXO4Rxg5Bil657(9lUk zPQzACEnSO-vWAzcn+oJi`R_{P?zkJOrY8aF^$ZQS6f$tnVl<0Y&XE+XafyfLv3Xb?=f%i1}AQ=06rbXao<2@7H~J5J$}xh7Wf042k4)GfjHgzd6)BC5QxKYaR@j9 z{2p!{2>}Pks0GRay2h%Rxd9KMj~o3V4-D$TaruxZ>=lPSq94yC1DEvVxJvMs0^ft9 zG`9f9W`>b&q>Wqo)H)hoCa$5Hf*~;3$LTz$0)}Qo@HCO&C8^ z<<|oFJm>>o=m4N~IBw8DR{_6*qjb&#yNyuk7~oMjO0ym~a-`Zqgyk7HP?sPw;Ul=U zpt&(PCcsTZGZN}>Sl$P|6Y!S__bk`5YGox)3;csxni`xf~Fyr*YhJ3>Lyq;b^ez0{RxH3DLPF365%(3|s@Z7Bi&* z7*Ys}0unU>KQ6*_1zijLuo&{7w*ntMr528x&T-*$a1RT91n|j{a%2ooS_w279Az*a z_&Xd;#TMY9xu`v4jsVA-RQ^O@Y8hGr{Jy}saMZ^p;FoYFD6x9rw0Y1g0jM!>{~`=h zRI(PBvRF;0Z#jsia1@!)WJVJ~r~ChWUJE!taz=v?uT!inf z#`FTc1(>%EnS&+_dkcdTbOdlT9F3Jq;1M{gKqJum9h_|;LzoLk{1TwwMy!(H6B;)o zP5|^j5%0oLvupuA_pVCM2iCw*WWvQ;P-8@{1O|M7QH?$h0>1De5(d2hcpoke^h4nF zPf&2sH-KZesp|;gwCyUt5cp*|{!qq^z^WZsNDyHI@TpJH_n<3)Njp&p(8<7;;7Gm# zcovQZ+IirNU24nC0;V3om;%2q@GKn7`}4rvwOBF1uL1hiWAK75C*m&LLwH(&*~ioj za)2MfQSG-9eF6g$gNwfJmvRc~6Lbd90!MmcIq)luJ`Y@US{*}`!0uAc->VoykSBEc1_c7`226#cLiGKnoa4Cfpt})3=nY2$itrh@1>lzeZP&1tlT0JN z=nF^6XFhNz+!oNgfM3H=qAfs|@6nyml7wVJCz_Bf&Py?d2+4X~RgM567HR>6Z&CpG zgk%mTn((|v6aJ~ugzi76GG0Kky^=iP0FBNAlAV+IgjS6vEGE+#1rU;Tk!ZpujV24D z+m9-ra1*J->Dnm72vG_k{E_Z1x1sAIv|G$-~fR3t#Aa!0RhTKO`hk9;1mA~p9ep``U_-Z zKOb~;Z7<&2erY?kww>D2PSvzi_3hN9HmY0()%I!29KWK>wSw=o3fd1f`AH4nT2^C6 z1b&;}?=`f%(_??_dT2=0uq`}lGr}Vu`x}nI@A23_{|X<o|7_i@z6$sdc=k;`~rR~oJu(VTHZT=mdp z{LiuZ8>er)w6QG(Zj;j{O2xP-YE$y2v`ylsyiH>_[^-]+) -(?P\d+[^-]*) (-(?P\d+[^-]*))? @@ -78,7 +88,8 @@ def _derive_abi(): \.whl$ ''', re.IGNORECASE | re.VERBOSE) -NAME_VERSION_RE = re.compile(r''' +NAME_VERSION_RE = re.compile( + r''' (?P[^-]+) -(?P\d+[^-]*) (-(?P\d+[^-]*))?$ @@ -94,8 +105,35 @@ def _derive_abi(): else: to_posix = lambda o: o.replace(os.sep, '/') +if sys.version_info[0] < 3: + import imp +else: + imp = None + import importlib.machinery + import importlib.util + + +def _get_suffixes(): + if imp: + return [s[0] for s in imp.get_suffixes()] + else: + return importlib.machinery.EXTENSION_SUFFIXES + + +def _load_dynamic(name, path): + # https://docs.python.org/3/library/importlib.html#importing-a-source-file-directly + if imp: + return imp.load_dynamic(name, path) + else: + spec = importlib.util.spec_from_file_location(name, path) + module = importlib.util.module_from_spec(spec) + sys.modules[name] = module + spec.loader.exec_module(module) + return module + class Mounter(object): + def __init__(self): self.impure_wheels = {} self.libs = {} @@ -123,13 +161,14 @@ def load_module(self, fullname): else: if fullname not in self.libs: raise ImportError('unable to find extension for %s' % fullname) - result = imp.load_dynamic(fullname, self.libs[fullname]) + result = _load_dynamic(fullname, self.libs[fullname]) result.__loader__ = self parts = fullname.rsplit('.', 1) if len(parts) > 1: result.__package__ = parts[0] return result + _hook = Mounter() @@ -196,8 +235,8 @@ def filename(self): arch = '.'.join(self.arch) # replace - with _ as a local version separator version = self.version.replace('-', '_') - return '%s-%s%s-%s-%s-%s.whl' % (self.name, version, buildver, - pyver, abi, arch) + return '%s-%s%s-%s-%s-%s.whl' % (self.name, version, buildver, pyver, + abi, arch) @property def exists(self): @@ -218,13 +257,15 @@ def metadata(self): info_dir = '%s.dist-info' % name_ver wrapper = codecs.getreader('utf-8') with ZipFile(pathname, 'r') as zf: - wheel_metadata = self.get_wheel_metadata(zf) - wv = wheel_metadata['Wheel-Version'].split('.', 1) - file_version = tuple([int(i) for i in wv]) - if file_version < (1, 1): - fns = [WHEEL_METADATA_FILENAME, METADATA_FILENAME, 'METADATA'] - else: - fns = [WHEEL_METADATA_FILENAME, METADATA_FILENAME] + self.get_wheel_metadata(zf) + # wv = wheel_metadata['Wheel-Version'].split('.', 1) + # file_version = tuple([int(i) for i in wv]) + # if file_version < (1, 1): + # fns = [WHEEL_METADATA_FILENAME, METADATA_FILENAME, + # LEGACY_METADATA_FILENAME] + # else: + # fns = [WHEEL_METADATA_FILENAME, METADATA_FILENAME] + fns = [WHEEL_METADATA_FILENAME, LEGACY_METADATA_FILENAME] result = None for fn in fns: try: @@ -293,16 +334,15 @@ def get_hash(self, data, hash_kind=None): try: hasher = getattr(hashlib, hash_kind) except AttributeError: - raise DistlibException('Unsupported hash algorithm: %r' % hash_kind) + raise DistlibException('Unsupported hash algorithm: %r' % + hash_kind) result = hasher(data).digest() result = base64.urlsafe_b64encode(result).rstrip(b'=').decode('ascii') return hash_kind, result - def write_record(self, records, record_path, base): - records = list(records) # make a copy for sorting - p = to_posix(os.path.relpath(record_path, base)) - records.append((p, '', '')) - records.sort() + def write_record(self, records, record_path, archive_record_path): + records = list(records) # make a copy, as mutated + records.append((archive_record_path, '', '')) with CSVWriter(record_path) as writer: for row in records: writer.writerow(row) @@ -310,7 +350,7 @@ def write_record(self, records, record_path, base): def write_records(self, info, libdir, archive_paths): records = [] distinfo, info_dir = info - hasher = getattr(hashlib, self.hash_kind) + # hasher = getattr(hashlib, self.hash_kind) for ap, p in archive_paths: with open(p, 'rb') as f: data = f.read() @@ -319,8 +359,8 @@ def write_records(self, info, libdir, archive_paths): records.append((ap, digest, size)) p = os.path.join(distinfo, 'RECORD') - self.write_record(records, p, libdir) ap = to_posix(os.path.join(info_dir, 'RECORD')) + self.write_record(records, p, ap) archive_paths.append((ap, p)) def build_zip(self, pathname, archive_paths): @@ -425,6 +465,19 @@ def build(self, paths, tags=None, wheel_version=None): ap = to_posix(os.path.join(info_dir, 'WHEEL')) archive_paths.append((ap, p)) + # sort the entries by archive path. Not needed by any spec, but it + # keeps the archive listing and RECORD tidier than they would otherwise + # be. Use the number of path segments to keep directory entries together, + # and keep the dist-info stuff at the end. + def sorter(t): + ap = t[0] + n = ap.count('/') + if '.dist-info' in ap: + n += 10000 + return (n, ap) + + archive_paths = sorted(archive_paths, key=sorter) + # Now, at last, RECORD. # Paths in here are archive paths - nothing else makes sense. self.write_records((distinfo, info_dir), libdir, archive_paths) @@ -469,14 +522,15 @@ def install(self, paths, maker, **kwargs): dry_run = maker.dry_run warner = kwargs.get('warner') lib_only = kwargs.get('lib_only', False) - bc_hashed_invalidation = kwargs.get('bytecode_hashed_invalidation', False) + bc_hashed_invalidation = kwargs.get('bytecode_hashed_invalidation', + False) pathname = os.path.join(self.dirname, self.filename) name_ver = '%s-%s' % (self.name, self.version) data_dir = '%s.data' % name_ver info_dir = '%s.dist-info' % name_ver - metadata_name = posixpath.join(info_dir, METADATA_FILENAME) + metadata_name = posixpath.join(info_dir, LEGACY_METADATA_FILENAME) wheel_metadata_name = posixpath.join(info_dir, 'WHEEL') record_name = posixpath.join(info_dir, 'RECORD') @@ -510,11 +564,11 @@ def install(self, paths, maker, **kwargs): # make a new instance rather than a copy of maker's, # as we mutate it fileop = FileOperator(dry_run=dry_run) - fileop.record = True # so we can rollback if needed + fileop.record = True # so we can rollback if needed - bc = not sys.dont_write_bytecode # Double negatives. Lovely! + bc = not sys.dont_write_bytecode # Double negatives. Lovely! - outfiles = [] # for RECORD writing + outfiles = [] # for RECORD writing # for script copying/shebang processing workdir = tempfile.mkdtemp() @@ -562,6 +616,14 @@ def install(self, paths, maker, **kwargs): if not is_script: with zf.open(arcname) as bf: fileop.copy_stream(bf, outfile) + # Issue #147: permission bits aren't preserved. Using + # zf.extract(zinfo, libdir) should have worked, but didn't, + # see https://www.thetopsites.net/article/53834422.shtml + # So ... manually preserve permission bits as given in zinfo + if os.name == 'posix': + # just set the normal permission bits + os.chmod(outfile, + (zinfo.external_attr >> 16) & 0x1FF) outfiles.append(outfile) # Double check the digest of the written file if not dry_run and row[1]: @@ -574,8 +636,9 @@ def install(self, paths, maker, **kwargs): '%s' % outfile) if bc and outfile.endswith('.py'): try: - pyc = fileop.byte_compile(outfile, - hashed_invalidation=bc_hashed_invalidation) + pyc = fileop.byte_compile( + outfile, + hashed_invalidation=bc_hashed_invalidation) outfiles.append(pyc) except Exception: # Don't give up if byte-compilation fails, @@ -619,7 +682,7 @@ def install(self, paths, maker, **kwargs): for v in epdata[k].values(): s = '%s:%s' % (v.prefix, v.suffix) if v.flags: - s += ' %s' % v.flags + s += ' [%s]' % ','.join(v.flags) d[v.name] = s except Exception: logger.warning('Unable to read legacy script ' @@ -650,7 +713,7 @@ def install(self, paths, maker, **kwargs): fileop.set_executable_mode(filenames) if gui_scripts: - options = {'gui': True } + options = {'gui': True} for k, v in gui_scripts.items(): script = '%s = %s' % (k, v) filenames = maker.make(script, options) @@ -660,7 +723,7 @@ def install(self, paths, maker, **kwargs): dist = InstalledDistribution(p) # Write SHARED - paths = dict(paths) # don't change passed in dict + paths = dict(paths) # don't change passed in dict del paths['purelib'] del paths['platlib'] paths['lib'] = libdir @@ -711,7 +774,8 @@ def _get_extensions(self): extract = True else: file_time = os.stat(dest).st_mtime - file_time = datetime.datetime.fromtimestamp(file_time) + file_time = datetime.datetime.fromtimestamp( + file_time) info = zf.getinfo(relpath) wheel_time = datetime.datetime(*info.date_time) extract = wheel_time > file_time @@ -732,7 +796,7 @@ def is_mountable(self): """ Determine if a wheel is asserted as mountable by its metadata. """ - return True # for now - metadata details TBD + return True # for now - metadata details TBD def mount(self, append=False): pathname = os.path.abspath(os.path.join(self.dirname, self.filename)) @@ -770,10 +834,10 @@ def unmount(self): def verify(self): pathname = os.path.join(self.dirname, self.filename) name_ver = '%s-%s' % (self.name, self.version) - data_dir = '%s.data' % name_ver + # data_dir = '%s.data' % name_ver info_dir = '%s.dist-info' % name_ver - metadata_name = posixpath.join(info_dir, METADATA_FILENAME) + # metadata_name = posixpath.join(info_dir, LEGACY_METADATA_FILENAME) wheel_metadata_name = posixpath.join(info_dir, 'WHEEL') record_name = posixpath.join(info_dir, 'RECORD') @@ -782,9 +846,9 @@ def verify(self): with ZipFile(pathname, 'r') as zf: with zf.open(wheel_metadata_name) as bwf: wf = wrapper(bwf) - message = message_from_file(wf) - wv = message['Wheel-Version'].split('.', 1) - file_version = tuple([int(i) for i in wv]) + message_from_file(wf) + # wv = message['Wheel-Version'].split('.', 1) + # file_version = tuple([int(i) for i in wv]) # TODO version verification records = {} @@ -842,7 +906,7 @@ def update(self, modifier, dest_dir=None, **kwargs): def get_version(path_map, info_dir): version = path = None - key = '%s/%s' % (info_dir, METADATA_FILENAME) + key = '%s/%s' % (info_dir, LEGACY_METADATA_FILENAME) if key not in path_map: key = '%s/PKG-INFO' % info_dir if key in path_map: @@ -853,25 +917,25 @@ def get_version(path_map, info_dir): def update_version(version, path): updated = None try: - v = NormalizedVersion(version) + NormalizedVersion(version) i = version.find('-') if i < 0: updated = '%s+1' % version else: parts = [int(s) for s in version[i + 1:].split('.')] parts[-1] += 1 - updated = '%s+%s' % (version[:i], - '.'.join(str(i) for i in parts)) + updated = '%s+%s' % (version[:i], '.'.join( + str(i) for i in parts)) except UnsupportedVersionError: - logger.debug('Cannot update non-compliant (PEP-440) ' - 'version %r', version) + logger.debug( + 'Cannot update non-compliant (PEP-440) ' + 'version %r', version) if updated: md = Metadata(path=path) md.version = updated - legacy = not path.endswith(METADATA_FILENAME) + legacy = path.endswith(LEGACY_METADATA_FILENAME) md.write(path=path, legacy=legacy) - logger.debug('Version updated from %r to %r', version, - updated) + logger.debug('Version updated from %r to %r', version, updated) pathname = os.path.join(self.dirname, self.filename) name_ver = '%s-%s' % (self.name, self.version) @@ -913,7 +977,8 @@ def update_version(version, path): os.close(fd) else: if not os.path.isdir(dest_dir): - raise DistlibException('Not a directory: %r' % dest_dir) + raise DistlibException('Not a directory: %r' % + dest_dir) newpath = os.path.join(dest_dir, self.filename) archive_paths = list(path_map.items()) distinfo = os.path.join(workdir, info_dir) @@ -924,17 +989,29 @@ def update_version(version, path): shutil.copyfile(newpath, pathname) return modified + +def _get_glibc_version(): + import platform + ver = platform.libc_ver() + result = [] + if ver[0] == 'glibc': + for s in ver[1].split('.'): + result.append(int(s) if s.isdigit() else 0) + result = tuple(result) + return result + + def compatible_tags(): """ Return (pyver, abi, arch) tuples compatible with this Python. """ versions = [VER_SUFFIX] major = VER_SUFFIX[0] - for minor in range(sys.version_info[1] - 1, - 1, -1): + for minor in range(sys.version_info[1] - 1, -1, -1): versions.append(''.join([major, str(minor)])) abis = [] - for suffix, _, _ in imp.get_suffixes(): + for suffix in _get_suffixes(): if suffix.startswith('.abi'): abis.append(suffix.split('.', 2)[1]) abis.sort() @@ -963,7 +1040,7 @@ def compatible_tags(): while minor >= 0: for match in matches: s = '%s_%s_%s_%s' % (name, major, minor, match) - if s != ARCH: # already there + if s != ARCH: # already there arches.append(s) minor -= 1 @@ -971,6 +1048,23 @@ def compatible_tags(): for abi in abis: for arch in arches: result.append((''.join((IMP_PREFIX, versions[0])), abi, arch)) + # manylinux + if abi != 'none' and sys.platform.startswith('linux'): + arch = arch.replace('linux_', '') + parts = _get_glibc_version() + if len(parts) == 2: + if parts >= (2, 5): + result.append((''.join((IMP_PREFIX, versions[0])), abi, + 'manylinux1_%s' % arch)) + if parts >= (2, 12): + result.append((''.join((IMP_PREFIX, versions[0])), abi, + 'manylinux2010_%s' % arch)) + if parts >= (2, 17): + result.append((''.join((IMP_PREFIX, versions[0])), abi, + 'manylinux2014_%s' % arch)) + result.append( + (''.join((IMP_PREFIX, versions[0])), abi, + 'manylinux_%s_%s_%s' % (parts[0], parts[1], arch))) # where no ABI / arch dependency, but IMP_PREFIX dependency for i, version in enumerate(versions): @@ -983,6 +1077,7 @@ def compatible_tags(): result.append((''.join(('py', version)), 'none', 'any')) if i == 0: result.append((''.join(('py', version[0])), 'none', 'any')) + return set(result) @@ -993,7 +1088,7 @@ def compatible_tags(): def is_compatible(wheel, tags=None): if not isinstance(wheel, Wheel): - wheel = Wheel(wheel) # assume it's a filename + wheel = Wheel(wheel) # assume it's a filename result = False if tags is None: tags = COMPATIBLE_TAGS diff --git a/src/fetchcode/vcs/pip/_vendor/distro.pyi b/src/fetchcode/vcs/pip/_vendor/distro.pyi deleted file mode 100644 index c7ea94b3..00000000 --- a/src/fetchcode/vcs/pip/_vendor/distro.pyi +++ /dev/null @@ -1 +0,0 @@ -from distro import * \ No newline at end of file diff --git a/src/fetchcode/vcs/pip/_vendor/distro.LICENSE b/src/fetchcode/vcs/pip/_vendor/distro/LICENSE similarity index 100% rename from src/fetchcode/vcs/pip/_vendor/distro.LICENSE rename to src/fetchcode/vcs/pip/_vendor/distro/LICENSE diff --git a/src/fetchcode/vcs/pip/_vendor/distro/__init__.py b/src/fetchcode/vcs/pip/_vendor/distro/__init__.py new file mode 100644 index 00000000..7686fe85 --- /dev/null +++ b/src/fetchcode/vcs/pip/_vendor/distro/__init__.py @@ -0,0 +1,54 @@ +from .distro import ( + NORMALIZED_DISTRO_ID, + NORMALIZED_LSB_ID, + NORMALIZED_OS_ID, + LinuxDistribution, + __version__, + build_number, + codename, + distro_release_attr, + distro_release_info, + id, + info, + like, + linux_distribution, + lsb_release_attr, + lsb_release_info, + major_version, + minor_version, + name, + os_release_attr, + os_release_info, + uname_attr, + uname_info, + version, + version_parts, +) + +__all__ = [ + "NORMALIZED_DISTRO_ID", + "NORMALIZED_LSB_ID", + "NORMALIZED_OS_ID", + "LinuxDistribution", + "build_number", + "codename", + "distro_release_attr", + "distro_release_info", + "id", + "info", + "like", + "linux_distribution", + "lsb_release_attr", + "lsb_release_info", + "major_version", + "minor_version", + "name", + "os_release_attr", + "os_release_info", + "uname_attr", + "uname_info", + "version", + "version_parts", +] + +__version__ = __version__ diff --git a/src/fetchcode/vcs/pip/_vendor/distro/__main__.py b/src/fetchcode/vcs/pip/_vendor/distro/__main__.py new file mode 100644 index 00000000..0c01d5b0 --- /dev/null +++ b/src/fetchcode/vcs/pip/_vendor/distro/__main__.py @@ -0,0 +1,4 @@ +from .distro import main + +if __name__ == "__main__": + main() diff --git a/src/fetchcode/vcs/pip/_vendor/distro.py b/src/fetchcode/vcs/pip/_vendor/distro/distro.py similarity index 64% rename from src/fetchcode/vcs/pip/_vendor/distro.py rename to src/fetchcode/vcs/pip/_vendor/distro/distro.py index 0611b62a..78ccdfa4 100644 --- a/src/fetchcode/vcs/pip/_vendor/distro.py +++ b/src/fetchcode/vcs/pip/_vendor/distro/distro.py @@ -1,4 +1,5 @@ -# Copyright 2015,2016,2017 Nir Cohen +#!/usr/bin/env python +# Copyright 2015-2021 Nir Cohen # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. @@ -20,26 +21,60 @@ It is the recommended replacement for Python's original :py:func:`platform.linux_distribution` function, but it provides much more functionality. An alternative implementation became necessary because Python -3.5 deprecated this function, and Python 3.8 will remove it altogether. -Its predecessor function :py:func:`platform.dist` was already -deprecated since Python 2.6 and will also be removed in Python 3.8. -Still, there are many cases in which access to OS distribution information -is needed. See `Python issue 1322 `_ for -more information. +3.5 deprecated this function, and Python 3.8 removed it altogether. Its +predecessor function :py:func:`platform.dist` was already deprecated since +Python 2.6 and removed in Python 3.8. Still, there are many cases in which +access to OS distribution information is needed. See `Python issue 1322 +`_ for more information. """ +import argparse +import json +import logging import os import re -import sys -import json import shlex -import logging -import argparse import subprocess +import sys +import warnings +from typing import ( + Any, + Callable, + Dict, + Iterable, + Optional, + Sequence, + TextIO, + Tuple, + Type, +) + +try: + from typing import TypedDict +except ImportError: + # Python 3.7 + TypedDict = dict + +__version__ = "1.9.0" -_UNIXCONFDIR = os.environ.get('UNIXCONFDIR', '/etc') -_OS_RELEASE_BASENAME = 'os-release' +class VersionDict(TypedDict): + major: str + minor: str + build_number: str + + +class InfoDict(TypedDict): + id: str + version: str + version_parts: VersionDict + like: str + codename: str + + +_UNIXCONFDIR = os.environ.get("UNIXCONFDIR", "/etc") +_UNIXUSRLIBDIR = os.environ.get("UNIXUSRLIBDIR", "/usr/lib") +_OS_RELEASE_BASENAME = "os-release" #: Translation table for normalizing the "ID" attribute defined in os-release #: files, for use by the :func:`distro.id` method. @@ -49,7 +84,8 @@ #: #: * Value: Normalized value. NORMALIZED_OS_ID = { - 'ol': 'oracle', # Oracle Linux + "ol": "oracle", # Oracle Linux + "opensuse-leap": "opensuse", # Newer versions of OpenSuSE report as opensuse-leap } #: Translation table for normalizing the "Distributor ID" attribute returned by @@ -60,11 +96,11 @@ #: #: * Value: Normalized value. NORMALIZED_LSB_ID = { - 'enterpriseenterpriseas': 'oracle', # Oracle Enterprise Linux 4 - 'enterpriseenterpriseserver': 'oracle', # Oracle Linux 5 - 'redhatenterpriseworkstation': 'rhel', # RHEL 6, 7 Workstation - 'redhatenterpriseserver': 'rhel', # RHEL 6, 7 Server - 'redhatenterprisecomputenode': 'rhel', # RHEL 6 ComputeNode + "enterpriseenterpriseas": "oracle", # Oracle Enterprise Linux 4 + "enterpriseenterpriseserver": "oracle", # Oracle Linux 5 + "redhatenterpriseworkstation": "rhel", # RHEL 6, 7 Workstation + "redhatenterpriseserver": "rhel", # RHEL 6, 7 Server + "redhatenterprisecomputenode": "rhel", # RHEL 6 ComputeNode } #: Translation table for normalizing the distro ID derived from the file name @@ -75,30 +111,61 @@ #: #: * Value: Normalized value. NORMALIZED_DISTRO_ID = { - 'redhat': 'rhel', # RHEL 6.x, 7.x + "redhat": "rhel", # RHEL 6.x, 7.x } # Pattern for content of distro release file (reversed) _DISTRO_RELEASE_CONTENT_REVERSED_PATTERN = re.compile( - r'(?:[^)]*\)(.*)\()? *(?:STL )?([\d.+\-a-z]*\d) *(?:esaeler *)?(.+)') + r"(?:[^)]*\)(.*)\()? *(?:STL )?([\d.+\-a-z]*\d) *(?:esaeler *)?(.+)" +) # Pattern for base file name of distro release file -_DISTRO_RELEASE_BASENAME_PATTERN = re.compile( - r'(\w+)[-_](release|version)$') +_DISTRO_RELEASE_BASENAME_PATTERN = re.compile(r"(\w+)[-_](release|version)$") + +# Base file names to be looked up for if _UNIXCONFDIR is not readable. +_DISTRO_RELEASE_BASENAMES = [ + "SuSE-release", + "altlinux-release", + "arch-release", + "base-release", + "centos-release", + "fedora-release", + "gentoo-release", + "mageia-release", + "mandrake-release", + "mandriva-release", + "mandrivalinux-release", + "manjaro-release", + "oracle-release", + "redhat-release", + "rocky-release", + "sl-release", + "slackware-version", +] # Base file names to be ignored when searching for distro release file _DISTRO_RELEASE_IGNORE_BASENAMES = ( - 'debian_version', - 'lsb-release', - 'oem-release', + "debian_version", + "lsb-release", + "oem-release", _OS_RELEASE_BASENAME, - 'system-release', - 'plesk-release', + "system-release", + "plesk-release", + "iredmail-release", + "board-release", + "ec2_version", ) -def linux_distribution(full_distribution_name=True): +def linux_distribution(full_distribution_name: bool = True) -> Tuple[str, str, str]: """ + .. deprecated:: 1.6.0 + + :func:`distro.linux_distribution()` is deprecated. It should only be + used as a compatibility shim with Python's + :py:func:`platform.linux_distribution()`. Please use :func:`distro.id`, + :func:`distro.version` and :func:`distro.name` instead. + Return information about the current OS distribution as a tuple ``(id_name, version, codename)`` with items as follows: @@ -107,7 +174,8 @@ def linux_distribution(full_distribution_name=True): * ``version``: The result of :func:`distro.version`. - * ``codename``: The result of :func:`distro.codename`. + * ``codename``: The extra item (usually in parentheses) after the + os-release version number, or the result of :func:`distro.codename`. The interface of this function is compatible with the original :py:func:`platform.linux_distribution` function, supporting a subset of @@ -122,10 +190,17 @@ def linux_distribution(full_distribution_name=True): method normalizes the distro ID string to a reliable machine-readable value for a number of popular OS distributions. """ + warnings.warn( + "distro.linux_distribution() is deprecated. It should only be used as a " + "compatibility shim with Python's platform.linux_distribution(). Please use " + "distro.id(), distro.version() and distro.name() instead.", + DeprecationWarning, + stacklevel=2, + ) return _distro.linux_distribution(full_distribution_name) -def id(): +def id() -> str: """ Return the distro ID of the current distribution, as a machine-readable string. @@ -146,8 +221,9 @@ def id(): "fedora" Fedora "sles" SUSE Linux Enterprise Server "opensuse" openSUSE - "amazon" Amazon Linux + "amzn" Amazon Linux "arch" Arch Linux + "buildroot" Buildroot "cloudlinux" CloudLinux OS "exherbo" Exherbo Linux "gentoo" GenToo Linux @@ -167,6 +243,10 @@ def id(): "netbsd" NetBSD "freebsd" FreeBSD "midnightbsd" MidnightBSD + "rocky" Rocky Linux + "aix" AIX + "guix" Guix System + "altlinux" ALT Linux ============== ========================================= If you have a need to get distros for reliable IDs added into this set, @@ -204,7 +284,7 @@ def id(): return _distro.id() -def name(pretty=False): +def name(pretty: bool = False) -> str: """ Return the name of the current OS distribution, as a human-readable string. @@ -243,7 +323,7 @@ def name(pretty=False): return _distro.name(pretty) -def version(pretty=False, best=False): +def version(pretty: bool = False, best: bool = False) -> str: """ Return the version of the current OS distribution, as a human-readable string. @@ -259,6 +339,10 @@ def version(pretty=False, best=False): sources in a fixed priority order does not always yield the most precise version (e.g. for Debian 8.2, or CentOS 7.1). + Some other distributions may not provide this kind of information. In these + cases, an empty string would be returned. This behavior can be observed + with rolling releases distributions (e.g. Arch Linux). + The *best* parameter can be used to control the approach for the returned version: @@ -287,7 +371,7 @@ def version(pretty=False, best=False): return _distro.version(pretty, best) -def version_parts(best=False): +def version_parts(best: bool = False) -> Tuple[str, str, str]: """ Return the version of the current OS distribution as a tuple ``(major, minor, build_number)`` with items as follows: @@ -304,7 +388,7 @@ def version_parts(best=False): return _distro.version_parts(best) -def major_version(best=False): +def major_version(best: bool = False) -> str: """ Return the major version of the current OS distribution, as a string, if provided. @@ -317,7 +401,7 @@ def major_version(best=False): return _distro.major_version(best) -def minor_version(best=False): +def minor_version(best: bool = False) -> str: """ Return the minor version of the current OS distribution, as a string, if provided. @@ -330,7 +414,7 @@ def minor_version(best=False): return _distro.minor_version(best) -def build_number(best=False): +def build_number(best: bool = False) -> str: """ Return the build number of the current OS distribution, as a string, if provided. @@ -343,7 +427,7 @@ def build_number(best=False): return _distro.build_number(best) -def like(): +def like() -> str: """ Return a space-separated list of distro IDs of distributions that are closely related to the current OS distribution in regards to packaging @@ -360,7 +444,7 @@ def like(): return _distro.like() -def codename(): +def codename() -> str: """ Return the codename for the release of the current OS distribution, as a string. @@ -384,7 +468,7 @@ def codename(): return _distro.codename() -def info(pretty=False, best=False): +def info(pretty: bool = False, best: bool = False) -> InfoDict: """ Return certain machine-readable information items about the current OS distribution in a dictionary, as shown in the following example: @@ -428,7 +512,7 @@ def info(pretty=False, best=False): return _distro.info(pretty, best) -def os_release_info(): +def os_release_info() -> Dict[str, str]: """ Return a dictionary containing key-value pairs for the information items from the os-release file data source of the current OS distribution. @@ -438,7 +522,7 @@ def os_release_info(): return _distro.os_release_info() -def lsb_release_info(): +def lsb_release_info() -> Dict[str, str]: """ Return a dictionary containing key-value pairs for the information items from the lsb_release command data source of the current OS distribution. @@ -449,7 +533,7 @@ def lsb_release_info(): return _distro.lsb_release_info() -def distro_release_info(): +def distro_release_info() -> Dict[str, str]: """ Return a dictionary containing key-value pairs for the information items from the distro release file data source of the current OS distribution. @@ -459,7 +543,7 @@ def distro_release_info(): return _distro.distro_release_info() -def uname_info(): +def uname_info() -> Dict[str, str]: """ Return a dictionary containing key-value pairs for the information items from the distro release file data source of the current OS distribution. @@ -467,7 +551,7 @@ def uname_info(): return _distro.uname_info() -def os_release_attr(attribute): +def os_release_attr(attribute: str) -> str: """ Return a single named information item from the os-release file data source of the current OS distribution. @@ -486,7 +570,7 @@ def os_release_attr(attribute): return _distro.os_release_attr(attribute) -def lsb_release_attr(attribute): +def lsb_release_attr(attribute: str) -> str: """ Return a single named information item from the lsb_release command output data source of the current OS distribution. @@ -506,7 +590,7 @@ def lsb_release_attr(attribute): return _distro.lsb_release_attr(attribute) -def distro_release_attr(attribute): +def distro_release_attr(attribute: str) -> str: """ Return a single named information item from the distro release file data source of the current OS distribution. @@ -525,7 +609,7 @@ def distro_release_attr(attribute): return _distro.distro_release_attr(attribute) -def uname_attr(attribute): +def uname_attr(attribute: str) -> str: """ Return a single named information item from the distro release file data source of the current OS distribution. @@ -542,22 +626,27 @@ def uname_attr(attribute): return _distro.uname_attr(attribute) -class cached_property(object): - """A version of @property which caches the value. On access, it calls the - underlying function and sets the value in `__dict__` so future accesses - will not re-call the property. - """ - def __init__(self, f): - self._fname = f.__name__ - self._f = f +try: + from functools import cached_property +except ImportError: + # Python < 3.8 + class cached_property: # type: ignore + """A version of @property which caches the value. On access, it calls the + underlying function and sets the value in `__dict__` so future accesses + will not re-call the property. + """ - def __get__(self, obj, owner): - assert obj is not None, 'call {} on an instance'.format(self._fname) - ret = obj.__dict__[self._fname] = self._f(obj) - return ret + def __init__(self, f: Callable[[Any], Any]) -> None: + self._fname = f.__name__ + self._f = f + def __get__(self, obj: Any, owner: Type[Any]) -> Any: + assert obj is not None, f"call {self._fname} on an instance" + ret = obj.__dict__[self._fname] = self._f(obj) + return ret -class LinuxDistribution(object): + +class LinuxDistribution: """ Provides information about a OS distribution. @@ -575,11 +664,15 @@ class LinuxDistribution(object): lsb_release command. """ - def __init__(self, - include_lsb=True, - os_release_file='', - distro_release_file='', - include_uname=True): + def __init__( + self, + include_lsb: Optional[bool] = None, + os_release_file: str = "", + distro_release_file: str = "", + include_uname: Optional[bool] = None, + root_dir: Optional[str] = None, + include_oslevel: Optional[bool] = None, + ) -> None: """ The initialization method of this class gathers information from the available data sources, and stores that in private instance attributes. @@ -618,6 +711,15 @@ def __init__(self, the program execution path the data source for the uname command will be empty. + * ``root_dir`` (string): The absolute path to the root directory to use + to find distro-related information files. Note that ``include_*`` + parameters must not be enabled in combination with ``root_dir``. + + * ``include_oslevel`` (bool): Controls whether (AIX) oslevel command + output is included as a data source. If the oslevel command is not + available in the program execution path the data source will be + empty. + Public instance attributes: * ``os_release_file`` (string): The path name of the @@ -635,40 +737,86 @@ def __init__(self, parameter. This controls whether the uname information will be loaded. + * ``include_oslevel`` (bool): The result of the ``include_oslevel`` + parameter. This controls whether (AIX) oslevel information will be + loaded. + + * ``root_dir`` (string): The result of the ``root_dir`` parameter. + The absolute path to the root directory to use to find distro-related + information files. + Raises: - * :py:exc:`IOError`: Some I/O issue with an os-release file or distro - release file. + * :py:exc:`ValueError`: Initialization parameters combination is not + supported. - * :py:exc:`subprocess.CalledProcessError`: The lsb_release command had - some issue (other than not being available in the program execution - path). + * :py:exc:`OSError`: Some I/O issue with an os-release file or distro + release file. * :py:exc:`UnicodeError`: A data source has unexpected characters or uses an unexpected encoding. """ - self.os_release_file = os_release_file or \ - os.path.join(_UNIXCONFDIR, _OS_RELEASE_BASENAME) - self.distro_release_file = distro_release_file or '' # updated later - self.include_lsb = include_lsb - self.include_uname = include_uname - - def __repr__(self): - """Return repr of all info - """ - return \ - "LinuxDistribution(" \ - "os_release_file={self.os_release_file!r}, " \ - "distro_release_file={self.distro_release_file!r}, " \ - "include_lsb={self.include_lsb!r}, " \ - "include_uname={self.include_uname!r}, " \ - "_os_release_info={self._os_release_info!r}, " \ - "_lsb_release_info={self._lsb_release_info!r}, " \ - "_distro_release_info={self._distro_release_info!r}, " \ - "_uname_info={self._uname_info!r})".format( - self=self) - - def linux_distribution(self, full_distribution_name=True): + self.root_dir = root_dir + self.etc_dir = os.path.join(root_dir, "etc") if root_dir else _UNIXCONFDIR + self.usr_lib_dir = ( + os.path.join(root_dir, "usr/lib") if root_dir else _UNIXUSRLIBDIR + ) + + if os_release_file: + self.os_release_file = os_release_file + else: + etc_dir_os_release_file = os.path.join(self.etc_dir, _OS_RELEASE_BASENAME) + usr_lib_os_release_file = os.path.join( + self.usr_lib_dir, _OS_RELEASE_BASENAME + ) + + # NOTE: The idea is to respect order **and** have it set + # at all times for API backwards compatibility. + if os.path.isfile(etc_dir_os_release_file) or not os.path.isfile( + usr_lib_os_release_file + ): + self.os_release_file = etc_dir_os_release_file + else: + self.os_release_file = usr_lib_os_release_file + + self.distro_release_file = distro_release_file or "" # updated later + + is_root_dir_defined = root_dir is not None + if is_root_dir_defined and (include_lsb or include_uname or include_oslevel): + raise ValueError( + "Including subprocess data sources from specific root_dir is disallowed" + " to prevent false information" + ) + self.include_lsb = ( + include_lsb if include_lsb is not None else not is_root_dir_defined + ) + self.include_uname = ( + include_uname if include_uname is not None else not is_root_dir_defined + ) + self.include_oslevel = ( + include_oslevel if include_oslevel is not None else not is_root_dir_defined + ) + + def __repr__(self) -> str: + """Return repr of all info""" + return ( + "LinuxDistribution(" + "os_release_file={self.os_release_file!r}, " + "distro_release_file={self.distro_release_file!r}, " + "include_lsb={self.include_lsb!r}, " + "include_uname={self.include_uname!r}, " + "include_oslevel={self.include_oslevel!r}, " + "root_dir={self.root_dir!r}, " + "_os_release_info={self._os_release_info!r}, " + "_lsb_release_info={self._lsb_release_info!r}, " + "_distro_release_info={self._distro_release_info!r}, " + "_uname_info={self._uname_info!r}, " + "_oslevel_info={self._oslevel_info!r})".format(self=self) + ) + + def linux_distribution( + self, full_distribution_name: bool = True + ) -> Tuple[str, str, str]: """ Return information about the OS distribution that is compatible with Python's :func:`platform.linux_distribution`, supporting a subset @@ -679,92 +827,103 @@ def linux_distribution(self, full_distribution_name=True): return ( self.name() if full_distribution_name else self.id(), self.version(), - self.codename() + self._os_release_info.get("release_codename") or self.codename(), ) - def id(self): + def id(self) -> str: """Return the distro ID of the OS distribution, as a string. For details, see :func:`distro.id`. """ - def normalize(distro_id, table): - distro_id = distro_id.lower().replace(' ', '_') + + def normalize(distro_id: str, table: Dict[str, str]) -> str: + distro_id = distro_id.lower().replace(" ", "_") return table.get(distro_id, distro_id) - distro_id = self.os_release_attr('id') + distro_id = self.os_release_attr("id") if distro_id: return normalize(distro_id, NORMALIZED_OS_ID) - distro_id = self.lsb_release_attr('distributor_id') + distro_id = self.lsb_release_attr("distributor_id") if distro_id: return normalize(distro_id, NORMALIZED_LSB_ID) - distro_id = self.distro_release_attr('id') + distro_id = self.distro_release_attr("id") if distro_id: return normalize(distro_id, NORMALIZED_DISTRO_ID) - distro_id = self.uname_attr('id') + distro_id = self.uname_attr("id") if distro_id: return normalize(distro_id, NORMALIZED_DISTRO_ID) - return '' + return "" - def name(self, pretty=False): + def name(self, pretty: bool = False) -> str: """ Return the name of the OS distribution, as a string. For details, see :func:`distro.name`. """ - name = self.os_release_attr('name') \ - or self.lsb_release_attr('distributor_id') \ - or self.distro_release_attr('name') \ - or self.uname_attr('name') + name = ( + self.os_release_attr("name") + or self.lsb_release_attr("distributor_id") + or self.distro_release_attr("name") + or self.uname_attr("name") + ) if pretty: - name = self.os_release_attr('pretty_name') \ - or self.lsb_release_attr('description') + name = self.os_release_attr("pretty_name") or self.lsb_release_attr( + "description" + ) if not name: - name = self.distro_release_attr('name') \ - or self.uname_attr('name') + name = self.distro_release_attr("name") or self.uname_attr("name") version = self.version(pretty=True) if version: - name = name + ' ' + version - return name or '' + name = f"{name} {version}" + return name or "" - def version(self, pretty=False, best=False): + def version(self, pretty: bool = False, best: bool = False) -> str: """ Return the version of the OS distribution, as a string. For details, see :func:`distro.version`. """ versions = [ - self.os_release_attr('version_id'), - self.lsb_release_attr('release'), - self.distro_release_attr('version_id'), - self._parse_distro_release_content( - self.os_release_attr('pretty_name')).get('version_id', ''), + self.os_release_attr("version_id"), + self.lsb_release_attr("release"), + self.distro_release_attr("version_id"), + self._parse_distro_release_content(self.os_release_attr("pretty_name")).get( + "version_id", "" + ), self._parse_distro_release_content( - self.lsb_release_attr('description')).get('version_id', ''), - self.uname_attr('release') + self.lsb_release_attr("description") + ).get("version_id", ""), + self.uname_attr("release"), ] - version = '' + if self.uname_attr("id").startswith("aix"): + # On AIX platforms, prefer oslevel command output. + versions.insert(0, self.oslevel_info()) + elif self.id() == "debian" or "debian" in self.like().split(): + # On Debian-like, add debian_version file content to candidates list. + versions.append(self._debian_version) + version = "" if best: # This algorithm uses the last version in priority order that has # the best precision. If the versions are not in conflict, that # does not matter; otherwise, using the last one instead of the # first one might be considered a surprise. for v in versions: - if v.count(".") > version.count(".") or version == '': + if v.count(".") > version.count(".") or version == "": version = v else: for v in versions: - if v != '': + if v != "": version = v break if pretty and version and self.codename(): - version = '{0} ({1})'.format(version, self.codename()) + version = f"{version} ({self.codename()})" return version - def version_parts(self, best=False): + def version_parts(self, best: bool = False) -> Tuple[str, str, str]: """ Return the version of the OS distribution, as a tuple of version numbers. @@ -773,14 +932,14 @@ def version_parts(self, best=False): """ version_str = self.version(best=best) if version_str: - version_regex = re.compile(r'(\d+)\.?(\d+)?\.?(\d+)?') + version_regex = re.compile(r"(\d+)\.?(\d+)?\.?(\d+)?") matches = version_regex.match(version_str) if matches: major, minor, build_number = matches.groups() - return major, minor or '', build_number or '' - return '', '', '' + return major, minor or "", build_number or "" + return "", "", "" - def major_version(self, best=False): + def major_version(self, best: bool = False) -> str: """ Return the major version number of the current distribution. @@ -788,7 +947,7 @@ def major_version(self, best=False): """ return self.version_parts(best)[0] - def minor_version(self, best=False): + def minor_version(self, best: bool = False) -> str: """ Return the minor version number of the current distribution. @@ -796,7 +955,7 @@ def minor_version(self, best=False): """ return self.version_parts(best)[1] - def build_number(self, best=False): + def build_number(self, best: bool = False) -> str: """ Return the build number of the current distribution. @@ -804,15 +963,15 @@ def build_number(self, best=False): """ return self.version_parts(best)[2] - def like(self): + def like(self) -> str: """ Return the IDs of distributions that are like the OS distribution. For details, see :func:`distro.like`. """ - return self.os_release_attr('id_like') or '' + return self.os_release_attr("id_like") or "" - def codename(self): + def codename(self) -> str: """ Return the codename of the OS distribution. @@ -821,32 +980,34 @@ def codename(self): try: # Handle os_release specially since distros might purposefully set # this to empty string to have no codename - return self._os_release_info['codename'] + return self._os_release_info["codename"] except KeyError: - return self.lsb_release_attr('codename') \ - or self.distro_release_attr('codename') \ - or '' + return ( + self.lsb_release_attr("codename") + or self.distro_release_attr("codename") + or "" + ) - def info(self, pretty=False, best=False): + def info(self, pretty: bool = False, best: bool = False) -> InfoDict: """ Return certain machine-readable information about the OS distribution. For details, see :func:`distro.info`. """ - return dict( + return InfoDict( id=self.id(), version=self.version(pretty, best), - version_parts=dict( + version_parts=VersionDict( major=self.major_version(best), minor=self.minor_version(best), - build_number=self.build_number(best) + build_number=self.build_number(best), ), like=self.like(), codename=self.codename(), ) - def os_release_info(self): + def os_release_info(self) -> Dict[str, str]: """ Return a dictionary containing key-value pairs for the information items from the os-release file data source of the OS distribution. @@ -855,7 +1016,7 @@ def os_release_info(self): """ return self._os_release_info - def lsb_release_info(self): + def lsb_release_info(self) -> Dict[str, str]: """ Return a dictionary containing key-value pairs for the information items from the lsb_release command data source of the OS @@ -865,7 +1026,7 @@ def lsb_release_info(self): """ return self._lsb_release_info - def distro_release_info(self): + def distro_release_info(self) -> Dict[str, str]: """ Return a dictionary containing key-value pairs for the information items from the distro release file data source of the OS @@ -875,7 +1036,7 @@ def distro_release_info(self): """ return self._distro_release_info - def uname_info(self): + def uname_info(self) -> Dict[str, str]: """ Return a dictionary containing key-value pairs for the information items from the uname command data source of the OS distribution. @@ -884,44 +1045,50 @@ def uname_info(self): """ return self._uname_info - def os_release_attr(self, attribute): + def oslevel_info(self) -> str: + """ + Return AIX' oslevel command output. + """ + return self._oslevel_info + + def os_release_attr(self, attribute: str) -> str: """ Return a single named information item from the os-release file data source of the OS distribution. For details, see :func:`distro.os_release_attr`. """ - return self._os_release_info.get(attribute, '') + return self._os_release_info.get(attribute, "") - def lsb_release_attr(self, attribute): + def lsb_release_attr(self, attribute: str) -> str: """ Return a single named information item from the lsb_release command output data source of the OS distribution. For details, see :func:`distro.lsb_release_attr`. """ - return self._lsb_release_info.get(attribute, '') + return self._lsb_release_info.get(attribute, "") - def distro_release_attr(self, attribute): + def distro_release_attr(self, attribute: str) -> str: """ Return a single named information item from the distro release file data source of the OS distribution. For details, see :func:`distro.distro_release_attr`. """ - return self._distro_release_info.get(attribute, '') + return self._distro_release_info.get(attribute, "") - def uname_attr(self, attribute): + def uname_attr(self, attribute: str) -> str: """ Return a single named information item from the uname command output data source of the OS distribution. - For details, see :func:`distro.uname_release_attr`. + For details, see :func:`distro.uname_attr`. """ - return self._uname_info.get(attribute, '') + return self._uname_info.get(attribute, "") @cached_property - def _os_release_info(self): + def _os_release_info(self) -> Dict[str, str]: """ Get the information items from the specified os-release file. @@ -929,12 +1096,12 @@ def _os_release_info(self): A dictionary containing all information items. """ if os.path.isfile(self.os_release_file): - with open(self.os_release_file) as release_file: + with open(self.os_release_file, encoding="utf-8") as release_file: return self._parse_os_release_content(release_file) return {} @staticmethod - def _parse_os_release_content(lines): + def _parse_os_release_content(lines: TextIO) -> Dict[str, str]: """ Parse the lines of an os-release file. @@ -951,16 +1118,6 @@ def _parse_os_release_content(lines): lexer = shlex.shlex(lines, posix=True) lexer.whitespace_split = True - # The shlex module defines its `wordchars` variable using literals, - # making it dependent on the encoding of the Python source file. - # In Python 2.6 and 2.7, the shlex source file is encoded in - # 'iso-8859-1', and the `wordchars` variable is defined as a byte - # string. This causes a UnicodeDecodeError to be raised when the - # parsed content is a unicode object. The following fix resolves that - # (... but it should be fixed in shlex...): - if sys.version_info[0] == 2 and isinstance(lexer.wordchars, bytes): - lexer.wordchars = lexer.wordchars.decode('iso-8859-1') - tokens = list(lexer) for token in tokens: # At this point, all shell-like parsing has been done (i.e. @@ -969,37 +1126,32 @@ def _parse_os_release_content(lines): # stripped, etc.), so the tokens are now either: # * variable assignments: var=value # * commands or their arguments (not allowed in os-release) - if '=' in token: - k, v = token.split('=', 1) + # Ignore any tokens that are not variable assignments + if "=" in token: + k, v = token.split("=", 1) props[k.lower()] = v - else: - # Ignore any tokens that are not variable assignments - pass - if 'version_codename' in props: + if "version" in props: + # extract release codename (if any) from version attribute + match = re.search(r"\((\D+)\)|,\s*(\D+)", props["version"]) + if match: + release_codename = match.group(1) or match.group(2) + props["codename"] = props["release_codename"] = release_codename + + if "version_codename" in props: # os-release added a version_codename field. Use that in # preference to anything else Note that some distros purposefully # do not have code names. They should be setting # version_codename="" - props['codename'] = props['version_codename'] - elif 'ubuntu_codename' in props: + props["codename"] = props["version_codename"] + elif "ubuntu_codename" in props: # Same as above but a non-standard field name used on older Ubuntus - props['codename'] = props['ubuntu_codename'] - elif 'version' in props: - # If there is no version_codename, parse it from the version - codename = re.search(r'(\(\D+\))|,(\s+)?\D+', props['version']) - if codename: - codename = codename.group() - codename = codename.strip('()') - codename = codename.strip(',') - codename = codename.strip() - # codename appears within paranthese. - props['codename'] = codename + props["codename"] = props["ubuntu_codename"] return props @cached_property - def _lsb_release_info(self): + def _lsb_release_info(self) -> Dict[str, str]: """ Get the information items from the lsb_release command output. @@ -1008,17 +1160,17 @@ def _lsb_release_info(self): """ if not self.include_lsb: return {} - with open(os.devnull, 'w') as devnull: - try: - cmd = ('lsb_release', '-a') - stdout = subprocess.check_output(cmd, stderr=devnull) - except OSError: # Command not found - return {} + try: + cmd = ("lsb_release", "-a") + stdout = subprocess.check_output(cmd, stderr=subprocess.DEVNULL) + # Command not found or lsb_release returned error + except (OSError, subprocess.CalledProcessError): + return {} content = self._to_str(stdout).splitlines() return self._parse_lsb_release_content(content) @staticmethod - def _parse_lsb_release_content(lines): + def _parse_lsb_release_content(lines: Iterable[str]) -> Dict[str, str]: """ Parse the output of the lsb_release command. @@ -1033,58 +1185,72 @@ def _parse_lsb_release_content(lines): """ props = {} for line in lines: - kv = line.strip('\n').split(':', 1) + kv = line.strip("\n").split(":", 1) if len(kv) != 2: # Ignore lines without colon. continue k, v = kv - props.update({k.replace(' ', '_').lower(): v.strip()}) + props.update({k.replace(" ", "_").lower(): v.strip()}) return props @cached_property - def _uname_info(self): - with open(os.devnull, 'w') as devnull: - try: - cmd = ('uname', '-rs') - stdout = subprocess.check_output(cmd, stderr=devnull) - except OSError: - return {} + def _uname_info(self) -> Dict[str, str]: + if not self.include_uname: + return {} + try: + cmd = ("uname", "-rs") + stdout = subprocess.check_output(cmd, stderr=subprocess.DEVNULL) + except OSError: + return {} content = self._to_str(stdout).splitlines() return self._parse_uname_content(content) + @cached_property + def _oslevel_info(self) -> str: + if not self.include_oslevel: + return "" + try: + stdout = subprocess.check_output("oslevel", stderr=subprocess.DEVNULL) + except (OSError, subprocess.CalledProcessError): + return "" + return self._to_str(stdout).strip() + + @cached_property + def _debian_version(self) -> str: + try: + with open( + os.path.join(self.etc_dir, "debian_version"), encoding="ascii" + ) as fp: + return fp.readline().rstrip() + except FileNotFoundError: + return "" + @staticmethod - def _parse_uname_content(lines): + def _parse_uname_content(lines: Sequence[str]) -> Dict[str, str]: + if not lines: + return {} props = {} - match = re.search(r'^([^\s]+)\s+([\d\.]+)', lines[0].strip()) + match = re.search(r"^([^\s]+)\s+([\d\.]+)", lines[0].strip()) if match: name, version = match.groups() # This is to prevent the Linux kernel version from # appearing as the 'best' version on otherwise # identifiable distributions. - if name == 'Linux': + if name == "Linux": return {} - props['id'] = name.lower() - props['name'] = name - props['release'] = version + props["id"] = name.lower() + props["name"] = name + props["release"] = version return props @staticmethod - def _to_str(text): + def _to_str(bytestring: bytes) -> str: encoding = sys.getfilesystemencoding() - encoding = 'utf-8' if encoding == 'ascii' else encoding - - if sys.version_info[0] >= 3: - if isinstance(text, bytes): - return text.decode(encoding) - else: - if isinstance(text, unicode): # noqa - return text.encode(encoding) - - return text + return bytestring.decode(encoding) @cached_property - def _distro_release_info(self): + def _distro_release_info(self) -> Dict[str, str]: """ Get the information items from the specified distro release file. @@ -1094,23 +1260,21 @@ def _distro_release_info(self): if self.distro_release_file: # If it was specified, we use it and parse what we can, even if # its file name or content does not match the expected pattern. - distro_info = self._parse_distro_release_file( - self.distro_release_file) + distro_info = self._parse_distro_release_file(self.distro_release_file) basename = os.path.basename(self.distro_release_file) # The file name pattern for user-specified distro release files # is somewhat more tolerant (compared to when searching for the # file), because we want to use what was specified as best as # possible. match = _DISTRO_RELEASE_BASENAME_PATTERN.match(basename) - if 'name' in distro_info \ - and 'cloudlinux' in distro_info['name'].lower(): - distro_info['id'] = 'cloudlinux' - elif match: - distro_info['id'] = match.group(1) - return distro_info else: try: - basenames = os.listdir(_UNIXCONFDIR) + basenames = [ + basename + for basename in os.listdir(self.etc_dir) + if basename not in _DISTRO_RELEASE_IGNORE_BASENAMES + and os.path.isfile(os.path.join(self.etc_dir, basename)) + ] # We sort for repeatability in cases where there are multiple # distro specific files; e.g. CentOS, Oracle, Enterprise all # containing `redhat-release` on top of their own. @@ -1120,38 +1284,31 @@ def _distro_release_info(self): # sure about the *-release files. Check common entries of # /etc for information. If they turn out to not be there the # error is handled in `_parse_distro_release_file()`. - basenames = ['SuSE-release', - 'arch-release', - 'base-release', - 'centos-release', - 'fedora-release', - 'gentoo-release', - 'mageia-release', - 'mandrake-release', - 'mandriva-release', - 'mandrivalinux-release', - 'manjaro-release', - 'oracle-release', - 'redhat-release', - 'sl-release', - 'slackware-version'] + basenames = _DISTRO_RELEASE_BASENAMES for basename in basenames: - if basename in _DISTRO_RELEASE_IGNORE_BASENAMES: - continue match = _DISTRO_RELEASE_BASENAME_PATTERN.match(basename) - if match: - filepath = os.path.join(_UNIXCONFDIR, basename) - distro_info = self._parse_distro_release_file(filepath) - if 'name' in distro_info: - # The name is always present if the pattern matches - self.distro_release_file = filepath - distro_info['id'] = match.group(1) - if 'cloudlinux' in distro_info['name'].lower(): - distro_info['id'] = 'cloudlinux' - return distro_info - return {} + if match is None: + continue + filepath = os.path.join(self.etc_dir, basename) + distro_info = self._parse_distro_release_file(filepath) + # The name is always present if the pattern matches. + if "name" not in distro_info: + continue + self.distro_release_file = filepath + break + else: # the loop didn't "break": no candidate. + return {} - def _parse_distro_release_file(self, filepath): + if match is not None: + distro_info["id"] = match.group(1) + + # CloudLinux < 7: manually enrich info with proper id. + if "cloudlinux" in distro_info.get("name", "").lower(): + distro_info["id"] = "cloudlinux" + + return distro_info + + def _parse_distro_release_file(self, filepath: str) -> Dict[str, str]: """ Parse a distro release file. @@ -1163,18 +1320,18 @@ def _parse_distro_release_file(self, filepath): A dictionary containing all information items. """ try: - with open(filepath) as fp: + with open(filepath, encoding="utf-8") as fp: # Only parse the first line. For instance, on SLES there # are multiple lines. We don't want them... return self._parse_distro_release_content(fp.readline()) - except (OSError, IOError): + except OSError: # Ignore not being able to read a specific, seemingly version # related file. - # See https://github.com/nir0s/distro/issues/162 + # See https://github.com/python-distro/distro/issues/162 return {} @staticmethod - def _parse_distro_release_content(line): + def _parse_distro_release_content(line: str) -> Dict[str, str]: """ Parse a line from a distro release file. @@ -1185,46 +1342,62 @@ def _parse_distro_release_content(line): Returns: A dictionary containing all information items. """ - matches = _DISTRO_RELEASE_CONTENT_REVERSED_PATTERN.match( - line.strip()[::-1]) + matches = _DISTRO_RELEASE_CONTENT_REVERSED_PATTERN.match(line.strip()[::-1]) distro_info = {} if matches: # regexp ensures non-None - distro_info['name'] = matches.group(3)[::-1] + distro_info["name"] = matches.group(3)[::-1] if matches.group(2): - distro_info['version_id'] = matches.group(2)[::-1] + distro_info["version_id"] = matches.group(2)[::-1] if matches.group(1): - distro_info['codename'] = matches.group(1)[::-1] + distro_info["codename"] = matches.group(1)[::-1] elif line: - distro_info['name'] = line.strip() + distro_info["name"] = line.strip() return distro_info _distro = LinuxDistribution() -def main(): +def main() -> None: logger = logging.getLogger(__name__) logger.setLevel(logging.DEBUG) logger.addHandler(logging.StreamHandler(sys.stdout)) parser = argparse.ArgumentParser(description="OS distro info tool") parser.add_argument( - '--json', - '-j', - help="Output in machine readable format", - action="store_true") + "--json", "-j", help="Output in machine readable format", action="store_true" + ) + + parser.add_argument( + "--root-dir", + "-r", + type=str, + dest="root_dir", + help="Path to the root filesystem directory (defaults to /)", + ) + args = parser.parse_args() + if args.root_dir: + dist = LinuxDistribution( + include_lsb=False, + include_uname=False, + include_oslevel=False, + root_dir=args.root_dir, + ) + else: + dist = _distro + if args.json: - logger.info(json.dumps(info(), indent=4, sort_keys=True)) + logger.info(json.dumps(dist.info(), indent=4, sort_keys=True)) else: - logger.info('Name: %s', name(pretty=True)) - distribution_version = version(pretty=True) - logger.info('Version: %s', distribution_version) - distribution_codename = codename() - logger.info('Codename: %s', distribution_codename) + logger.info("Name: %s", dist.name(pretty=True)) + distribution_version = dist.version(pretty=True) + logger.info("Version: %s", distribution_version) + distribution_codename = dist.codename() + logger.info("Codename: %s", distribution_codename) -if __name__ == '__main__': +if __name__ == "__main__": main() diff --git a/src/fetchcode/vcs/pip/_vendor/distro/py.typed b/src/fetchcode/vcs/pip/_vendor/distro/py.typed new file mode 100644 index 00000000..e69de29b diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib.pyi b/src/fetchcode/vcs/pip/_vendor/html5lib.pyi deleted file mode 100644 index 9bc9af95..00000000 --- a/src/fetchcode/vcs/pip/_vendor/html5lib.pyi +++ /dev/null @@ -1 +0,0 @@ -from html5lib import * \ No newline at end of file diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib/LICENSE b/src/fetchcode/vcs/pip/_vendor/html5lib/LICENSE deleted file mode 100644 index c87fa7a0..00000000 --- a/src/fetchcode/vcs/pip/_vendor/html5lib/LICENSE +++ /dev/null @@ -1,20 +0,0 @@ -Copyright (c) 2006-2013 James Graham and other contributors - -Permission is hereby granted, free of charge, to any person obtaining -a copy of this software and associated documentation files (the -"Software"), to deal in the Software without restriction, including -without limitation the rights to use, copy, modify, merge, publish, -distribute, sublicense, and/or sell copies of the Software, and to -permit persons to whom the Software is furnished to do so, subject to -the following conditions: - -The above copyright notice and this permission notice shall be -included in all copies or substantial portions of the Software. - -THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, -EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF -MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND -NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE -LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION -OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION -WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib/__init__.py b/src/fetchcode/vcs/pip/_vendor/html5lib/__init__.py deleted file mode 100644 index 315c9dec..00000000 --- a/src/fetchcode/vcs/pip/_vendor/html5lib/__init__.py +++ /dev/null @@ -1,35 +0,0 @@ -""" -HTML parsing library based on the `WHATWG HTML specification -`_. The parser is designed to be compatible with -existing HTML found in the wild and implements well-defined error recovery that -is largely compatible with modern desktop web browsers. - -Example usage:: - - from fetchcode.vcs.pip._vendor import html5lib - with open("my_document.html", "rb") as f: - tree = html5lib.parse(f) - -For convenience, this module re-exports the following names: - -* :func:`~.html5parser.parse` -* :func:`~.html5parser.parseFragment` -* :class:`~.html5parser.HTMLParser` -* :func:`~.treebuilders.getTreeBuilder` -* :func:`~.treewalkers.getTreeWalker` -* :func:`~.serializer.serialize` -""" - -from __future__ import absolute_import, division, unicode_literals - -from .html5parser import HTMLParser, parse, parseFragment -from .treebuilders import getTreeBuilder -from .treewalkers import getTreeWalker -from .serializer import serialize - -__all__ = ["HTMLParser", "parse", "parseFragment", "getTreeBuilder", - "getTreeWalker", "serialize"] - -# this has to be at the top level, see how setup.py parses this -#: Distribution version number. -__version__ = "1.0.1" diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib/_ihatexml.py b/src/fetchcode/vcs/pip/_vendor/html5lib/_ihatexml.py deleted file mode 100644 index 4c77717b..00000000 --- a/src/fetchcode/vcs/pip/_vendor/html5lib/_ihatexml.py +++ /dev/null @@ -1,288 +0,0 @@ -from __future__ import absolute_import, division, unicode_literals - -import re -import warnings - -from .constants import DataLossWarning - -baseChar = """ -[#x0041-#x005A] | [#x0061-#x007A] | [#x00C0-#x00D6] | [#x00D8-#x00F6] | -[#x00F8-#x00FF] | [#x0100-#x0131] | [#x0134-#x013E] | [#x0141-#x0148] | -[#x014A-#x017E] | [#x0180-#x01C3] | [#x01CD-#x01F0] | [#x01F4-#x01F5] | -[#x01FA-#x0217] | [#x0250-#x02A8] | [#x02BB-#x02C1] | #x0386 | -[#x0388-#x038A] | #x038C | [#x038E-#x03A1] | [#x03A3-#x03CE] | -[#x03D0-#x03D6] | #x03DA | #x03DC | #x03DE | #x03E0 | [#x03E2-#x03F3] | -[#x0401-#x040C] | [#x040E-#x044F] | [#x0451-#x045C] | [#x045E-#x0481] | -[#x0490-#x04C4] | [#x04C7-#x04C8] | [#x04CB-#x04CC] | [#x04D0-#x04EB] | -[#x04EE-#x04F5] | [#x04F8-#x04F9] | [#x0531-#x0556] | #x0559 | -[#x0561-#x0586] | [#x05D0-#x05EA] | [#x05F0-#x05F2] | [#x0621-#x063A] | -[#x0641-#x064A] | [#x0671-#x06B7] | [#x06BA-#x06BE] | [#x06C0-#x06CE] | -[#x06D0-#x06D3] | #x06D5 | [#x06E5-#x06E6] | [#x0905-#x0939] | #x093D | -[#x0958-#x0961] | [#x0985-#x098C] | [#x098F-#x0990] | [#x0993-#x09A8] | -[#x09AA-#x09B0] | #x09B2 | [#x09B6-#x09B9] | [#x09DC-#x09DD] | -[#x09DF-#x09E1] | [#x09F0-#x09F1] | [#x0A05-#x0A0A] | [#x0A0F-#x0A10] | -[#x0A13-#x0A28] | [#x0A2A-#x0A30] | [#x0A32-#x0A33] | [#x0A35-#x0A36] | -[#x0A38-#x0A39] | [#x0A59-#x0A5C] | #x0A5E | [#x0A72-#x0A74] | -[#x0A85-#x0A8B] | #x0A8D | [#x0A8F-#x0A91] | [#x0A93-#x0AA8] | -[#x0AAA-#x0AB0] | [#x0AB2-#x0AB3] | [#x0AB5-#x0AB9] | #x0ABD | #x0AE0 | -[#x0B05-#x0B0C] | [#x0B0F-#x0B10] | [#x0B13-#x0B28] | [#x0B2A-#x0B30] | -[#x0B32-#x0B33] | [#x0B36-#x0B39] | #x0B3D | [#x0B5C-#x0B5D] | -[#x0B5F-#x0B61] | [#x0B85-#x0B8A] | [#x0B8E-#x0B90] | [#x0B92-#x0B95] | -[#x0B99-#x0B9A] | #x0B9C | [#x0B9E-#x0B9F] | [#x0BA3-#x0BA4] | -[#x0BA8-#x0BAA] | [#x0BAE-#x0BB5] | [#x0BB7-#x0BB9] | [#x0C05-#x0C0C] | -[#x0C0E-#x0C10] | [#x0C12-#x0C28] | [#x0C2A-#x0C33] | [#x0C35-#x0C39] | -[#x0C60-#x0C61] | [#x0C85-#x0C8C] | [#x0C8E-#x0C90] | [#x0C92-#x0CA8] | -[#x0CAA-#x0CB3] | [#x0CB5-#x0CB9] | #x0CDE | [#x0CE0-#x0CE1] | -[#x0D05-#x0D0C] | [#x0D0E-#x0D10] | [#x0D12-#x0D28] | [#x0D2A-#x0D39] | -[#x0D60-#x0D61] | [#x0E01-#x0E2E] | #x0E30 | [#x0E32-#x0E33] | -[#x0E40-#x0E45] | [#x0E81-#x0E82] | #x0E84 | [#x0E87-#x0E88] | #x0E8A | -#x0E8D | [#x0E94-#x0E97] | [#x0E99-#x0E9F] | [#x0EA1-#x0EA3] | #x0EA5 | -#x0EA7 | [#x0EAA-#x0EAB] | [#x0EAD-#x0EAE] | #x0EB0 | [#x0EB2-#x0EB3] | -#x0EBD | [#x0EC0-#x0EC4] | [#x0F40-#x0F47] | [#x0F49-#x0F69] | -[#x10A0-#x10C5] | [#x10D0-#x10F6] | #x1100 | [#x1102-#x1103] | -[#x1105-#x1107] | #x1109 | [#x110B-#x110C] | [#x110E-#x1112] | #x113C | -#x113E | #x1140 | #x114C | #x114E | #x1150 | [#x1154-#x1155] | #x1159 | -[#x115F-#x1161] | #x1163 | #x1165 | #x1167 | #x1169 | [#x116D-#x116E] | -[#x1172-#x1173] | #x1175 | #x119E | #x11A8 | #x11AB | [#x11AE-#x11AF] | -[#x11B7-#x11B8] | #x11BA | [#x11BC-#x11C2] | #x11EB | #x11F0 | #x11F9 | -[#x1E00-#x1E9B] | [#x1EA0-#x1EF9] | [#x1F00-#x1F15] | [#x1F18-#x1F1D] | -[#x1F20-#x1F45] | [#x1F48-#x1F4D] | [#x1F50-#x1F57] | #x1F59 | #x1F5B | -#x1F5D | [#x1F5F-#x1F7D] | [#x1F80-#x1FB4] | [#x1FB6-#x1FBC] | #x1FBE | -[#x1FC2-#x1FC4] | [#x1FC6-#x1FCC] | [#x1FD0-#x1FD3] | [#x1FD6-#x1FDB] | -[#x1FE0-#x1FEC] | [#x1FF2-#x1FF4] | [#x1FF6-#x1FFC] | #x2126 | -[#x212A-#x212B] | #x212E | [#x2180-#x2182] | [#x3041-#x3094] | -[#x30A1-#x30FA] | [#x3105-#x312C] | [#xAC00-#xD7A3]""" - -ideographic = """[#x4E00-#x9FA5] | #x3007 | [#x3021-#x3029]""" - -combiningCharacter = """ -[#x0300-#x0345] | [#x0360-#x0361] | [#x0483-#x0486] | [#x0591-#x05A1] | -[#x05A3-#x05B9] | [#x05BB-#x05BD] | #x05BF | [#x05C1-#x05C2] | #x05C4 | -[#x064B-#x0652] | #x0670 | [#x06D6-#x06DC] | [#x06DD-#x06DF] | -[#x06E0-#x06E4] | [#x06E7-#x06E8] | [#x06EA-#x06ED] | [#x0901-#x0903] | -#x093C | [#x093E-#x094C] | #x094D | [#x0951-#x0954] | [#x0962-#x0963] | -[#x0981-#x0983] | #x09BC | #x09BE | #x09BF | [#x09C0-#x09C4] | -[#x09C7-#x09C8] | [#x09CB-#x09CD] | #x09D7 | [#x09E2-#x09E3] | #x0A02 | -#x0A3C | #x0A3E | #x0A3F | [#x0A40-#x0A42] | [#x0A47-#x0A48] | -[#x0A4B-#x0A4D] | [#x0A70-#x0A71] | [#x0A81-#x0A83] | #x0ABC | -[#x0ABE-#x0AC5] | [#x0AC7-#x0AC9] | [#x0ACB-#x0ACD] | [#x0B01-#x0B03] | -#x0B3C | [#x0B3E-#x0B43] | [#x0B47-#x0B48] | [#x0B4B-#x0B4D] | -[#x0B56-#x0B57] | [#x0B82-#x0B83] | [#x0BBE-#x0BC2] | [#x0BC6-#x0BC8] | -[#x0BCA-#x0BCD] | #x0BD7 | [#x0C01-#x0C03] | [#x0C3E-#x0C44] | -[#x0C46-#x0C48] | [#x0C4A-#x0C4D] | [#x0C55-#x0C56] | [#x0C82-#x0C83] | -[#x0CBE-#x0CC4] | [#x0CC6-#x0CC8] | [#x0CCA-#x0CCD] | [#x0CD5-#x0CD6] | -[#x0D02-#x0D03] | [#x0D3E-#x0D43] | [#x0D46-#x0D48] | [#x0D4A-#x0D4D] | -#x0D57 | #x0E31 | [#x0E34-#x0E3A] | [#x0E47-#x0E4E] | #x0EB1 | -[#x0EB4-#x0EB9] | [#x0EBB-#x0EBC] | [#x0EC8-#x0ECD] | [#x0F18-#x0F19] | -#x0F35 | #x0F37 | #x0F39 | #x0F3E | #x0F3F | [#x0F71-#x0F84] | -[#x0F86-#x0F8B] | [#x0F90-#x0F95] | #x0F97 | [#x0F99-#x0FAD] | -[#x0FB1-#x0FB7] | #x0FB9 | [#x20D0-#x20DC] | #x20E1 | [#x302A-#x302F] | -#x3099 | #x309A""" - -digit = """ -[#x0030-#x0039] | [#x0660-#x0669] | [#x06F0-#x06F9] | [#x0966-#x096F] | -[#x09E6-#x09EF] | [#x0A66-#x0A6F] | [#x0AE6-#x0AEF] | [#x0B66-#x0B6F] | -[#x0BE7-#x0BEF] | [#x0C66-#x0C6F] | [#x0CE6-#x0CEF] | [#x0D66-#x0D6F] | -[#x0E50-#x0E59] | [#x0ED0-#x0ED9] | [#x0F20-#x0F29]""" - -extender = """ -#x00B7 | #x02D0 | #x02D1 | #x0387 | #x0640 | #x0E46 | #x0EC6 | #x3005 | -#[#x3031-#x3035] | [#x309D-#x309E] | [#x30FC-#x30FE]""" - -letter = " | ".join([baseChar, ideographic]) - -# Without the -name = " | ".join([letter, digit, ".", "-", "_", combiningCharacter, - extender]) -nameFirst = " | ".join([letter, "_"]) - -reChar = re.compile(r"#x([\d|A-F]{4,4})") -reCharRange = re.compile(r"\[#x([\d|A-F]{4,4})-#x([\d|A-F]{4,4})\]") - - -def charStringToList(chars): - charRanges = [item.strip() for item in chars.split(" | ")] - rv = [] - for item in charRanges: - foundMatch = False - for regexp in (reChar, reCharRange): - match = regexp.match(item) - if match is not None: - rv.append([hexToInt(item) for item in match.groups()]) - if len(rv[-1]) == 1: - rv[-1] = rv[-1] * 2 - foundMatch = True - break - if not foundMatch: - assert len(item) == 1 - - rv.append([ord(item)] * 2) - rv = normaliseCharList(rv) - return rv - - -def normaliseCharList(charList): - charList = sorted(charList) - for item in charList: - assert item[1] >= item[0] - rv = [] - i = 0 - while i < len(charList): - j = 1 - rv.append(charList[i]) - while i + j < len(charList) and charList[i + j][0] <= rv[-1][1] + 1: - rv[-1][1] = charList[i + j][1] - j += 1 - i += j - return rv - -# We don't really support characters above the BMP :( -max_unicode = int("FFFF", 16) - - -def missingRanges(charList): - rv = [] - if charList[0] != 0: - rv.append([0, charList[0][0] - 1]) - for i, item in enumerate(charList[:-1]): - rv.append([item[1] + 1, charList[i + 1][0] - 1]) - if charList[-1][1] != max_unicode: - rv.append([charList[-1][1] + 1, max_unicode]) - return rv - - -def listToRegexpStr(charList): - rv = [] - for item in charList: - if item[0] == item[1]: - rv.append(escapeRegexp(chr(item[0]))) - else: - rv.append(escapeRegexp(chr(item[0])) + "-" + - escapeRegexp(chr(item[1]))) - return "[%s]" % "".join(rv) - - -def hexToInt(hex_str): - return int(hex_str, 16) - - -def escapeRegexp(string): - specialCharacters = (".", "^", "$", "*", "+", "?", "{", "}", - "[", "]", "|", "(", ")", "-") - for char in specialCharacters: - string = string.replace(char, "\\" + char) - - return string - -# output from the above -nonXmlNameBMPRegexp = re.compile('[\x00-,/:-@\\[-\\^`\\{-\xb6\xb8-\xbf\xd7\xf7\u0132-\u0133\u013f-\u0140\u0149\u017f\u01c4-\u01cc\u01f1-\u01f3\u01f6-\u01f9\u0218-\u024f\u02a9-\u02ba\u02c2-\u02cf\u02d2-\u02ff\u0346-\u035f\u0362-\u0385\u038b\u038d\u03a2\u03cf\u03d7-\u03d9\u03db\u03dd\u03df\u03e1\u03f4-\u0400\u040d\u0450\u045d\u0482\u0487-\u048f\u04c5-\u04c6\u04c9-\u04ca\u04cd-\u04cf\u04ec-\u04ed\u04f6-\u04f7\u04fa-\u0530\u0557-\u0558\u055a-\u0560\u0587-\u0590\u05a2\u05ba\u05be\u05c0\u05c3\u05c5-\u05cf\u05eb-\u05ef\u05f3-\u0620\u063b-\u063f\u0653-\u065f\u066a-\u066f\u06b8-\u06b9\u06bf\u06cf\u06d4\u06e9\u06ee-\u06ef\u06fa-\u0900\u0904\u093a-\u093b\u094e-\u0950\u0955-\u0957\u0964-\u0965\u0970-\u0980\u0984\u098d-\u098e\u0991-\u0992\u09a9\u09b1\u09b3-\u09b5\u09ba-\u09bb\u09bd\u09c5-\u09c6\u09c9-\u09ca\u09ce-\u09d6\u09d8-\u09db\u09de\u09e4-\u09e5\u09f2-\u0a01\u0a03-\u0a04\u0a0b-\u0a0e\u0a11-\u0a12\u0a29\u0a31\u0a34\u0a37\u0a3a-\u0a3b\u0a3d\u0a43-\u0a46\u0a49-\u0a4a\u0a4e-\u0a58\u0a5d\u0a5f-\u0a65\u0a75-\u0a80\u0a84\u0a8c\u0a8e\u0a92\u0aa9\u0ab1\u0ab4\u0aba-\u0abb\u0ac6\u0aca\u0ace-\u0adf\u0ae1-\u0ae5\u0af0-\u0b00\u0b04\u0b0d-\u0b0e\u0b11-\u0b12\u0b29\u0b31\u0b34-\u0b35\u0b3a-\u0b3b\u0b44-\u0b46\u0b49-\u0b4a\u0b4e-\u0b55\u0b58-\u0b5b\u0b5e\u0b62-\u0b65\u0b70-\u0b81\u0b84\u0b8b-\u0b8d\u0b91\u0b96-\u0b98\u0b9b\u0b9d\u0ba0-\u0ba2\u0ba5-\u0ba7\u0bab-\u0bad\u0bb6\u0bba-\u0bbd\u0bc3-\u0bc5\u0bc9\u0bce-\u0bd6\u0bd8-\u0be6\u0bf0-\u0c00\u0c04\u0c0d\u0c11\u0c29\u0c34\u0c3a-\u0c3d\u0c45\u0c49\u0c4e-\u0c54\u0c57-\u0c5f\u0c62-\u0c65\u0c70-\u0c81\u0c84\u0c8d\u0c91\u0ca9\u0cb4\u0cba-\u0cbd\u0cc5\u0cc9\u0cce-\u0cd4\u0cd7-\u0cdd\u0cdf\u0ce2-\u0ce5\u0cf0-\u0d01\u0d04\u0d0d\u0d11\u0d29\u0d3a-\u0d3d\u0d44-\u0d45\u0d49\u0d4e-\u0d56\u0d58-\u0d5f\u0d62-\u0d65\u0d70-\u0e00\u0e2f\u0e3b-\u0e3f\u0e4f\u0e5a-\u0e80\u0e83\u0e85-\u0e86\u0e89\u0e8b-\u0e8c\u0e8e-\u0e93\u0e98\u0ea0\u0ea4\u0ea6\u0ea8-\u0ea9\u0eac\u0eaf\u0eba\u0ebe-\u0ebf\u0ec5\u0ec7\u0ece-\u0ecf\u0eda-\u0f17\u0f1a-\u0f1f\u0f2a-\u0f34\u0f36\u0f38\u0f3a-\u0f3d\u0f48\u0f6a-\u0f70\u0f85\u0f8c-\u0f8f\u0f96\u0f98\u0fae-\u0fb0\u0fb8\u0fba-\u109f\u10c6-\u10cf\u10f7-\u10ff\u1101\u1104\u1108\u110a\u110d\u1113-\u113b\u113d\u113f\u1141-\u114b\u114d\u114f\u1151-\u1153\u1156-\u1158\u115a-\u115e\u1162\u1164\u1166\u1168\u116a-\u116c\u116f-\u1171\u1174\u1176-\u119d\u119f-\u11a7\u11a9-\u11aa\u11ac-\u11ad\u11b0-\u11b6\u11b9\u11bb\u11c3-\u11ea\u11ec-\u11ef\u11f1-\u11f8\u11fa-\u1dff\u1e9c-\u1e9f\u1efa-\u1eff\u1f16-\u1f17\u1f1e-\u1f1f\u1f46-\u1f47\u1f4e-\u1f4f\u1f58\u1f5a\u1f5c\u1f5e\u1f7e-\u1f7f\u1fb5\u1fbd\u1fbf-\u1fc1\u1fc5\u1fcd-\u1fcf\u1fd4-\u1fd5\u1fdc-\u1fdf\u1fed-\u1ff1\u1ff5\u1ffd-\u20cf\u20dd-\u20e0\u20e2-\u2125\u2127-\u2129\u212c-\u212d\u212f-\u217f\u2183-\u3004\u3006\u3008-\u3020\u3030\u3036-\u3040\u3095-\u3098\u309b-\u309c\u309f-\u30a0\u30fb\u30ff-\u3104\u312d-\u4dff\u9fa6-\uabff\ud7a4-\uffff]') # noqa - -nonXmlNameFirstBMPRegexp = re.compile('[\x00-@\\[-\\^`\\{-\xbf\xd7\xf7\u0132-\u0133\u013f-\u0140\u0149\u017f\u01c4-\u01cc\u01f1-\u01f3\u01f6-\u01f9\u0218-\u024f\u02a9-\u02ba\u02c2-\u0385\u0387\u038b\u038d\u03a2\u03cf\u03d7-\u03d9\u03db\u03dd\u03df\u03e1\u03f4-\u0400\u040d\u0450\u045d\u0482-\u048f\u04c5-\u04c6\u04c9-\u04ca\u04cd-\u04cf\u04ec-\u04ed\u04f6-\u04f7\u04fa-\u0530\u0557-\u0558\u055a-\u0560\u0587-\u05cf\u05eb-\u05ef\u05f3-\u0620\u063b-\u0640\u064b-\u0670\u06b8-\u06b9\u06bf\u06cf\u06d4\u06d6-\u06e4\u06e7-\u0904\u093a-\u093c\u093e-\u0957\u0962-\u0984\u098d-\u098e\u0991-\u0992\u09a9\u09b1\u09b3-\u09b5\u09ba-\u09db\u09de\u09e2-\u09ef\u09f2-\u0a04\u0a0b-\u0a0e\u0a11-\u0a12\u0a29\u0a31\u0a34\u0a37\u0a3a-\u0a58\u0a5d\u0a5f-\u0a71\u0a75-\u0a84\u0a8c\u0a8e\u0a92\u0aa9\u0ab1\u0ab4\u0aba-\u0abc\u0abe-\u0adf\u0ae1-\u0b04\u0b0d-\u0b0e\u0b11-\u0b12\u0b29\u0b31\u0b34-\u0b35\u0b3a-\u0b3c\u0b3e-\u0b5b\u0b5e\u0b62-\u0b84\u0b8b-\u0b8d\u0b91\u0b96-\u0b98\u0b9b\u0b9d\u0ba0-\u0ba2\u0ba5-\u0ba7\u0bab-\u0bad\u0bb6\u0bba-\u0c04\u0c0d\u0c11\u0c29\u0c34\u0c3a-\u0c5f\u0c62-\u0c84\u0c8d\u0c91\u0ca9\u0cb4\u0cba-\u0cdd\u0cdf\u0ce2-\u0d04\u0d0d\u0d11\u0d29\u0d3a-\u0d5f\u0d62-\u0e00\u0e2f\u0e31\u0e34-\u0e3f\u0e46-\u0e80\u0e83\u0e85-\u0e86\u0e89\u0e8b-\u0e8c\u0e8e-\u0e93\u0e98\u0ea0\u0ea4\u0ea6\u0ea8-\u0ea9\u0eac\u0eaf\u0eb1\u0eb4-\u0ebc\u0ebe-\u0ebf\u0ec5-\u0f3f\u0f48\u0f6a-\u109f\u10c6-\u10cf\u10f7-\u10ff\u1101\u1104\u1108\u110a\u110d\u1113-\u113b\u113d\u113f\u1141-\u114b\u114d\u114f\u1151-\u1153\u1156-\u1158\u115a-\u115e\u1162\u1164\u1166\u1168\u116a-\u116c\u116f-\u1171\u1174\u1176-\u119d\u119f-\u11a7\u11a9-\u11aa\u11ac-\u11ad\u11b0-\u11b6\u11b9\u11bb\u11c3-\u11ea\u11ec-\u11ef\u11f1-\u11f8\u11fa-\u1dff\u1e9c-\u1e9f\u1efa-\u1eff\u1f16-\u1f17\u1f1e-\u1f1f\u1f46-\u1f47\u1f4e-\u1f4f\u1f58\u1f5a\u1f5c\u1f5e\u1f7e-\u1f7f\u1fb5\u1fbd\u1fbf-\u1fc1\u1fc5\u1fcd-\u1fcf\u1fd4-\u1fd5\u1fdc-\u1fdf\u1fed-\u1ff1\u1ff5\u1ffd-\u2125\u2127-\u2129\u212c-\u212d\u212f-\u217f\u2183-\u3006\u3008-\u3020\u302a-\u3040\u3095-\u30a0\u30fb-\u3104\u312d-\u4dff\u9fa6-\uabff\ud7a4-\uffff]') # noqa - -# Simpler things -nonPubidCharRegexp = re.compile("[^\x20\x0D\x0Aa-zA-Z0-9\\-'()+,./:=?;!*#@$_%]") - - -class InfosetFilter(object): - replacementRegexp = re.compile(r"U[\dA-F]{5,5}") - - def __init__(self, - dropXmlnsLocalName=False, - dropXmlnsAttrNs=False, - preventDoubleDashComments=False, - preventDashAtCommentEnd=False, - replaceFormFeedCharacters=True, - preventSingleQuotePubid=False): - - self.dropXmlnsLocalName = dropXmlnsLocalName - self.dropXmlnsAttrNs = dropXmlnsAttrNs - - self.preventDoubleDashComments = preventDoubleDashComments - self.preventDashAtCommentEnd = preventDashAtCommentEnd - - self.replaceFormFeedCharacters = replaceFormFeedCharacters - - self.preventSingleQuotePubid = preventSingleQuotePubid - - self.replaceCache = {} - - def coerceAttribute(self, name, namespace=None): - if self.dropXmlnsLocalName and name.startswith("xmlns:"): - warnings.warn("Attributes cannot begin with xmlns", DataLossWarning) - return None - elif (self.dropXmlnsAttrNs and - namespace == "http://www.w3.org/2000/xmlns/"): - warnings.warn("Attributes cannot be in the xml namespace", DataLossWarning) - return None - else: - return self.toXmlName(name) - - def coerceElement(self, name): - return self.toXmlName(name) - - def coerceComment(self, data): - if self.preventDoubleDashComments: - while "--" in data: - warnings.warn("Comments cannot contain adjacent dashes", DataLossWarning) - data = data.replace("--", "- -") - if data.endswith("-"): - warnings.warn("Comments cannot end in a dash", DataLossWarning) - data += " " - return data - - def coerceCharacters(self, data): - if self.replaceFormFeedCharacters: - for _ in range(data.count("\x0C")): - warnings.warn("Text cannot contain U+000C", DataLossWarning) - data = data.replace("\x0C", " ") - # Other non-xml characters - return data - - def coercePubid(self, data): - dataOutput = data - for char in nonPubidCharRegexp.findall(data): - warnings.warn("Coercing non-XML pubid", DataLossWarning) - replacement = self.getReplacementCharacter(char) - dataOutput = dataOutput.replace(char, replacement) - if self.preventSingleQuotePubid and dataOutput.find("'") >= 0: - warnings.warn("Pubid cannot contain single quote", DataLossWarning) - dataOutput = dataOutput.replace("'", self.getReplacementCharacter("'")) - return dataOutput - - def toXmlName(self, name): - nameFirst = name[0] - nameRest = name[1:] - m = nonXmlNameFirstBMPRegexp.match(nameFirst) - if m: - warnings.warn("Coercing non-XML name", DataLossWarning) - nameFirstOutput = self.getReplacementCharacter(nameFirst) - else: - nameFirstOutput = nameFirst - - nameRestOutput = nameRest - replaceChars = set(nonXmlNameBMPRegexp.findall(nameRest)) - for char in replaceChars: - warnings.warn("Coercing non-XML name", DataLossWarning) - replacement = self.getReplacementCharacter(char) - nameRestOutput = nameRestOutput.replace(char, replacement) - return nameFirstOutput + nameRestOutput - - def getReplacementCharacter(self, char): - if char in self.replaceCache: - replacement = self.replaceCache[char] - else: - replacement = self.escapeChar(char) - return replacement - - def fromXmlName(self, name): - for item in set(self.replacementRegexp.findall(name)): - name = name.replace(item, self.unescapeChar(item)) - return name - - def escapeChar(self, char): - replacement = "U%05X" % ord(char) - self.replaceCache[char] = replacement - return replacement - - def unescapeChar(self, charcode): - return chr(int(charcode[1:], 16)) diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib/_inputstream.py b/src/fetchcode/vcs/pip/_vendor/html5lib/_inputstream.py deleted file mode 100644 index 6ec50d2f..00000000 --- a/src/fetchcode/vcs/pip/_vendor/html5lib/_inputstream.py +++ /dev/null @@ -1,923 +0,0 @@ -from __future__ import absolute_import, division, unicode_literals - -from fetchcode.vcs.pip._vendor.six import text_type, binary_type -from fetchcode.vcs.pip._vendor.six.moves import http_client, urllib - -import codecs -import re - -from fetchcode.vcs.pip._vendor import webencodings - -from .constants import EOF, spaceCharacters, asciiLetters, asciiUppercase -from .constants import _ReparseException -from . import _utils - -from io import StringIO - -try: - from io import BytesIO -except ImportError: - BytesIO = StringIO - -# Non-unicode versions of constants for use in the pre-parser -spaceCharactersBytes = frozenset([item.encode("ascii") for item in spaceCharacters]) -asciiLettersBytes = frozenset([item.encode("ascii") for item in asciiLetters]) -asciiUppercaseBytes = frozenset([item.encode("ascii") for item in asciiUppercase]) -spacesAngleBrackets = spaceCharactersBytes | frozenset([b">", b"<"]) - - -invalid_unicode_no_surrogate = "[\u0001-\u0008\u000B\u000E-\u001F\u007F-\u009F\uFDD0-\uFDEF\uFFFE\uFFFF\U0001FFFE\U0001FFFF\U0002FFFE\U0002FFFF\U0003FFFE\U0003FFFF\U0004FFFE\U0004FFFF\U0005FFFE\U0005FFFF\U0006FFFE\U0006FFFF\U0007FFFE\U0007FFFF\U0008FFFE\U0008FFFF\U0009FFFE\U0009FFFF\U000AFFFE\U000AFFFF\U000BFFFE\U000BFFFF\U000CFFFE\U000CFFFF\U000DFFFE\U000DFFFF\U000EFFFE\U000EFFFF\U000FFFFE\U000FFFFF\U0010FFFE\U0010FFFF]" # noqa - -if _utils.supports_lone_surrogates: - # Use one extra step of indirection and create surrogates with - # eval. Not using this indirection would introduce an illegal - # unicode literal on platforms not supporting such lone - # surrogates. - assert invalid_unicode_no_surrogate[-1] == "]" and invalid_unicode_no_surrogate.count("]") == 1 - invalid_unicode_re = re.compile(invalid_unicode_no_surrogate[:-1] + - eval('"\\uD800-\\uDFFF"') + # pylint:disable=eval-used - "]") -else: - invalid_unicode_re = re.compile(invalid_unicode_no_surrogate) - -non_bmp_invalid_codepoints = set([0x1FFFE, 0x1FFFF, 0x2FFFE, 0x2FFFF, 0x3FFFE, - 0x3FFFF, 0x4FFFE, 0x4FFFF, 0x5FFFE, 0x5FFFF, - 0x6FFFE, 0x6FFFF, 0x7FFFE, 0x7FFFF, 0x8FFFE, - 0x8FFFF, 0x9FFFE, 0x9FFFF, 0xAFFFE, 0xAFFFF, - 0xBFFFE, 0xBFFFF, 0xCFFFE, 0xCFFFF, 0xDFFFE, - 0xDFFFF, 0xEFFFE, 0xEFFFF, 0xFFFFE, 0xFFFFF, - 0x10FFFE, 0x10FFFF]) - -ascii_punctuation_re = re.compile("[\u0009-\u000D\u0020-\u002F\u003A-\u0040\u005C\u005B-\u0060\u007B-\u007E]") - -# Cache for charsUntil() -charsUntilRegEx = {} - - -class BufferedStream(object): - """Buffering for streams that do not have buffering of their own - - The buffer is implemented as a list of chunks on the assumption that - joining many strings will be slow since it is O(n**2) - """ - - def __init__(self, stream): - self.stream = stream - self.buffer = [] - self.position = [-1, 0] # chunk number, offset - - def tell(self): - pos = 0 - for chunk in self.buffer[:self.position[0]]: - pos += len(chunk) - pos += self.position[1] - return pos - - def seek(self, pos): - assert pos <= self._bufferedBytes() - offset = pos - i = 0 - while len(self.buffer[i]) < offset: - offset -= len(self.buffer[i]) - i += 1 - self.position = [i, offset] - - def read(self, bytes): - if not self.buffer: - return self._readStream(bytes) - elif (self.position[0] == len(self.buffer) and - self.position[1] == len(self.buffer[-1])): - return self._readStream(bytes) - else: - return self._readFromBuffer(bytes) - - def _bufferedBytes(self): - return sum([len(item) for item in self.buffer]) - - def _readStream(self, bytes): - data = self.stream.read(bytes) - self.buffer.append(data) - self.position[0] += 1 - self.position[1] = len(data) - return data - - def _readFromBuffer(self, bytes): - remainingBytes = bytes - rv = [] - bufferIndex = self.position[0] - bufferOffset = self.position[1] - while bufferIndex < len(self.buffer) and remainingBytes != 0: - assert remainingBytes > 0 - bufferedData = self.buffer[bufferIndex] - - if remainingBytes <= len(bufferedData) - bufferOffset: - bytesToRead = remainingBytes - self.position = [bufferIndex, bufferOffset + bytesToRead] - else: - bytesToRead = len(bufferedData) - bufferOffset - self.position = [bufferIndex, len(bufferedData)] - bufferIndex += 1 - rv.append(bufferedData[bufferOffset:bufferOffset + bytesToRead]) - remainingBytes -= bytesToRead - - bufferOffset = 0 - - if remainingBytes: - rv.append(self._readStream(remainingBytes)) - - return b"".join(rv) - - -def HTMLInputStream(source, **kwargs): - # Work around Python bug #20007: read(0) closes the connection. - # http://bugs.python.org/issue20007 - if (isinstance(source, http_client.HTTPResponse) or - # Also check for addinfourl wrapping HTTPResponse - (isinstance(source, urllib.response.addbase) and - isinstance(source.fp, http_client.HTTPResponse))): - isUnicode = False - elif hasattr(source, "read"): - isUnicode = isinstance(source.read(0), text_type) - else: - isUnicode = isinstance(source, text_type) - - if isUnicode: - encodings = [x for x in kwargs if x.endswith("_encoding")] - if encodings: - raise TypeError("Cannot set an encoding with a unicode input, set %r" % encodings) - - return HTMLUnicodeInputStream(source, **kwargs) - else: - return HTMLBinaryInputStream(source, **kwargs) - - -class HTMLUnicodeInputStream(object): - """Provides a unicode stream of characters to the HTMLTokenizer. - - This class takes care of character encoding and removing or replacing - incorrect byte-sequences and also provides column and line tracking. - - """ - - _defaultChunkSize = 10240 - - def __init__(self, source): - """Initialises the HTMLInputStream. - - HTMLInputStream(source, [encoding]) -> Normalized stream from source - for use by html5lib. - - source can be either a file-object, local filename or a string. - - The optional encoding parameter must be a string that indicates - the encoding. If specified, that encoding will be used, - regardless of any BOM or later declaration (such as in a meta - element) - - """ - - if not _utils.supports_lone_surrogates: - # Such platforms will have already checked for such - # surrogate errors, so no need to do this checking. - self.reportCharacterErrors = None - elif len("\U0010FFFF") == 1: - self.reportCharacterErrors = self.characterErrorsUCS4 - else: - self.reportCharacterErrors = self.characterErrorsUCS2 - - # List of where new lines occur - self.newLines = [0] - - self.charEncoding = (lookupEncoding("utf-8"), "certain") - self.dataStream = self.openStream(source) - - self.reset() - - def reset(self): - self.chunk = "" - self.chunkSize = 0 - self.chunkOffset = 0 - self.errors = [] - - # number of (complete) lines in previous chunks - self.prevNumLines = 0 - # number of columns in the last line of the previous chunk - self.prevNumCols = 0 - - # Deal with CR LF and surrogates split over chunk boundaries - self._bufferedCharacter = None - - def openStream(self, source): - """Produces a file object from source. - - source can be either a file object, local filename or a string. - - """ - # Already a file object - if hasattr(source, 'read'): - stream = source - else: - stream = StringIO(source) - - return stream - - def _position(self, offset): - chunk = self.chunk - nLines = chunk.count('\n', 0, offset) - positionLine = self.prevNumLines + nLines - lastLinePos = chunk.rfind('\n', 0, offset) - if lastLinePos == -1: - positionColumn = self.prevNumCols + offset - else: - positionColumn = offset - (lastLinePos + 1) - return (positionLine, positionColumn) - - def position(self): - """Returns (line, col) of the current position in the stream.""" - line, col = self._position(self.chunkOffset) - return (line + 1, col) - - def char(self): - """ Read one character from the stream or queue if available. Return - EOF when EOF is reached. - """ - # Read a new chunk from the input stream if necessary - if self.chunkOffset >= self.chunkSize: - if not self.readChunk(): - return EOF - - chunkOffset = self.chunkOffset - char = self.chunk[chunkOffset] - self.chunkOffset = chunkOffset + 1 - - return char - - def readChunk(self, chunkSize=None): - if chunkSize is None: - chunkSize = self._defaultChunkSize - - self.prevNumLines, self.prevNumCols = self._position(self.chunkSize) - - self.chunk = "" - self.chunkSize = 0 - self.chunkOffset = 0 - - data = self.dataStream.read(chunkSize) - - # Deal with CR LF and surrogates broken across chunks - if self._bufferedCharacter: - data = self._bufferedCharacter + data - self._bufferedCharacter = None - elif not data: - # We have no more data, bye-bye stream - return False - - if len(data) > 1: - lastv = ord(data[-1]) - if lastv == 0x0D or 0xD800 <= lastv <= 0xDBFF: - self._bufferedCharacter = data[-1] - data = data[:-1] - - if self.reportCharacterErrors: - self.reportCharacterErrors(data) - - # Replace invalid characters - data = data.replace("\r\n", "\n") - data = data.replace("\r", "\n") - - self.chunk = data - self.chunkSize = len(data) - - return True - - def characterErrorsUCS4(self, data): - for _ in range(len(invalid_unicode_re.findall(data))): - self.errors.append("invalid-codepoint") - - def characterErrorsUCS2(self, data): - # Someone picked the wrong compile option - # You lose - skip = False - for match in invalid_unicode_re.finditer(data): - if skip: - continue - codepoint = ord(match.group()) - pos = match.start() - # Pretty sure there should be endianness issues here - if _utils.isSurrogatePair(data[pos:pos + 2]): - # We have a surrogate pair! - char_val = _utils.surrogatePairToCodepoint(data[pos:pos + 2]) - if char_val in non_bmp_invalid_codepoints: - self.errors.append("invalid-codepoint") - skip = True - elif (codepoint >= 0xD800 and codepoint <= 0xDFFF and - pos == len(data) - 1): - self.errors.append("invalid-codepoint") - else: - skip = False - self.errors.append("invalid-codepoint") - - def charsUntil(self, characters, opposite=False): - """ Returns a string of characters from the stream up to but not - including any character in 'characters' or EOF. 'characters' must be - a container that supports the 'in' method and iteration over its - characters. - """ - - # Use a cache of regexps to find the required characters - try: - chars = charsUntilRegEx[(characters, opposite)] - except KeyError: - if __debug__: - for c in characters: - assert(ord(c) < 128) - regex = "".join(["\\x%02x" % ord(c) for c in characters]) - if not opposite: - regex = "^%s" % regex - chars = charsUntilRegEx[(characters, opposite)] = re.compile("[%s]+" % regex) - - rv = [] - - while True: - # Find the longest matching prefix - m = chars.match(self.chunk, self.chunkOffset) - if m is None: - # If nothing matched, and it wasn't because we ran out of chunk, - # then stop - if self.chunkOffset != self.chunkSize: - break - else: - end = m.end() - # If not the whole chunk matched, return everything - # up to the part that didn't match - if end != self.chunkSize: - rv.append(self.chunk[self.chunkOffset:end]) - self.chunkOffset = end - break - # If the whole remainder of the chunk matched, - # use it all and read the next chunk - rv.append(self.chunk[self.chunkOffset:]) - if not self.readChunk(): - # Reached EOF - break - - r = "".join(rv) - return r - - def unget(self, char): - # Only one character is allowed to be ungotten at once - it must - # be consumed again before any further call to unget - if char is not None: - if self.chunkOffset == 0: - # unget is called quite rarely, so it's a good idea to do - # more work here if it saves a bit of work in the frequently - # called char and charsUntil. - # So, just prepend the ungotten character onto the current - # chunk: - self.chunk = char + self.chunk - self.chunkSize += 1 - else: - self.chunkOffset -= 1 - assert self.chunk[self.chunkOffset] == char - - -class HTMLBinaryInputStream(HTMLUnicodeInputStream): - """Provides a unicode stream of characters to the HTMLTokenizer. - - This class takes care of character encoding and removing or replacing - incorrect byte-sequences and also provides column and line tracking. - - """ - - def __init__(self, source, override_encoding=None, transport_encoding=None, - same_origin_parent_encoding=None, likely_encoding=None, - default_encoding="windows-1252", useChardet=True): - """Initialises the HTMLInputStream. - - HTMLInputStream(source, [encoding]) -> Normalized stream from source - for use by html5lib. - - source can be either a file-object, local filename or a string. - - The optional encoding parameter must be a string that indicates - the encoding. If specified, that encoding will be used, - regardless of any BOM or later declaration (such as in a meta - element) - - """ - # Raw Stream - for unicode objects this will encode to utf-8 and set - # self.charEncoding as appropriate - self.rawStream = self.openStream(source) - - HTMLUnicodeInputStream.__init__(self, self.rawStream) - - # Encoding Information - # Number of bytes to use when looking for a meta element with - # encoding information - self.numBytesMeta = 1024 - # Number of bytes to use when using detecting encoding using chardet - self.numBytesChardet = 100 - # Things from args - self.override_encoding = override_encoding - self.transport_encoding = transport_encoding - self.same_origin_parent_encoding = same_origin_parent_encoding - self.likely_encoding = likely_encoding - self.default_encoding = default_encoding - - # Determine encoding - self.charEncoding = self.determineEncoding(useChardet) - assert self.charEncoding[0] is not None - - # Call superclass - self.reset() - - def reset(self): - self.dataStream = self.charEncoding[0].codec_info.streamreader(self.rawStream, 'replace') - HTMLUnicodeInputStream.reset(self) - - def openStream(self, source): - """Produces a file object from source. - - source can be either a file object, local filename or a string. - - """ - # Already a file object - if hasattr(source, 'read'): - stream = source - else: - stream = BytesIO(source) - - try: - stream.seek(stream.tell()) - except: # pylint:disable=bare-except - stream = BufferedStream(stream) - - return stream - - def determineEncoding(self, chardet=True): - # BOMs take precedence over everything - # This will also read past the BOM if present - charEncoding = self.detectBOM(), "certain" - if charEncoding[0] is not None: - return charEncoding - - # If we've been overriden, we've been overriden - charEncoding = lookupEncoding(self.override_encoding), "certain" - if charEncoding[0] is not None: - return charEncoding - - # Now check the transport layer - charEncoding = lookupEncoding(self.transport_encoding), "certain" - if charEncoding[0] is not None: - return charEncoding - - # Look for meta elements with encoding information - charEncoding = self.detectEncodingMeta(), "tentative" - if charEncoding[0] is not None: - return charEncoding - - # Parent document encoding - charEncoding = lookupEncoding(self.same_origin_parent_encoding), "tentative" - if charEncoding[0] is not None and not charEncoding[0].name.startswith("utf-16"): - return charEncoding - - # "likely" encoding - charEncoding = lookupEncoding(self.likely_encoding), "tentative" - if charEncoding[0] is not None: - return charEncoding - - # Guess with chardet, if available - if chardet: - try: - from fetchcode.vcs.pip._vendor.chardet.universaldetector import UniversalDetector - except ImportError: - pass - else: - buffers = [] - detector = UniversalDetector() - while not detector.done: - buffer = self.rawStream.read(self.numBytesChardet) - assert isinstance(buffer, bytes) - if not buffer: - break - buffers.append(buffer) - detector.feed(buffer) - detector.close() - encoding = lookupEncoding(detector.result['encoding']) - self.rawStream.seek(0) - if encoding is not None: - return encoding, "tentative" - - # Try the default encoding - charEncoding = lookupEncoding(self.default_encoding), "tentative" - if charEncoding[0] is not None: - return charEncoding - - # Fallback to html5lib's default if even that hasn't worked - return lookupEncoding("windows-1252"), "tentative" - - def changeEncoding(self, newEncoding): - assert self.charEncoding[1] != "certain" - newEncoding = lookupEncoding(newEncoding) - if newEncoding is None: - return - if newEncoding.name in ("utf-16be", "utf-16le"): - newEncoding = lookupEncoding("utf-8") - assert newEncoding is not None - elif newEncoding == self.charEncoding[0]: - self.charEncoding = (self.charEncoding[0], "certain") - else: - self.rawStream.seek(0) - self.charEncoding = (newEncoding, "certain") - self.reset() - raise _ReparseException("Encoding changed from %s to %s" % (self.charEncoding[0], newEncoding)) - - def detectBOM(self): - """Attempts to detect at BOM at the start of the stream. If - an encoding can be determined from the BOM return the name of the - encoding otherwise return None""" - bomDict = { - codecs.BOM_UTF8: 'utf-8', - codecs.BOM_UTF16_LE: 'utf-16le', codecs.BOM_UTF16_BE: 'utf-16be', - codecs.BOM_UTF32_LE: 'utf-32le', codecs.BOM_UTF32_BE: 'utf-32be' - } - - # Go to beginning of file and read in 4 bytes - string = self.rawStream.read(4) - assert isinstance(string, bytes) - - # Try detecting the BOM using bytes from the string - encoding = bomDict.get(string[:3]) # UTF-8 - seek = 3 - if not encoding: - # Need to detect UTF-32 before UTF-16 - encoding = bomDict.get(string) # UTF-32 - seek = 4 - if not encoding: - encoding = bomDict.get(string[:2]) # UTF-16 - seek = 2 - - # Set the read position past the BOM if one was found, otherwise - # set it to the start of the stream - if encoding: - self.rawStream.seek(seek) - return lookupEncoding(encoding) - else: - self.rawStream.seek(0) - return None - - def detectEncodingMeta(self): - """Report the encoding declared by the meta element - """ - buffer = self.rawStream.read(self.numBytesMeta) - assert isinstance(buffer, bytes) - parser = EncodingParser(buffer) - self.rawStream.seek(0) - encoding = parser.getEncoding() - - if encoding is not None and encoding.name in ("utf-16be", "utf-16le"): - encoding = lookupEncoding("utf-8") - - return encoding - - -class EncodingBytes(bytes): - """String-like object with an associated position and various extra methods - If the position is ever greater than the string length then an exception is - raised""" - def __new__(self, value): - assert isinstance(value, bytes) - return bytes.__new__(self, value.lower()) - - def __init__(self, value): - # pylint:disable=unused-argument - self._position = -1 - - def __iter__(self): - return self - - def __next__(self): - p = self._position = self._position + 1 - if p >= len(self): - raise StopIteration - elif p < 0: - raise TypeError - return self[p:p + 1] - - def next(self): - # Py2 compat - return self.__next__() - - def previous(self): - p = self._position - if p >= len(self): - raise StopIteration - elif p < 0: - raise TypeError - self._position = p = p - 1 - return self[p:p + 1] - - def setPosition(self, position): - if self._position >= len(self): - raise StopIteration - self._position = position - - def getPosition(self): - if self._position >= len(self): - raise StopIteration - if self._position >= 0: - return self._position - else: - return None - - position = property(getPosition, setPosition) - - def getCurrentByte(self): - return self[self.position:self.position + 1] - - currentByte = property(getCurrentByte) - - def skip(self, chars=spaceCharactersBytes): - """Skip past a list of characters""" - p = self.position # use property for the error-checking - while p < len(self): - c = self[p:p + 1] - if c not in chars: - self._position = p - return c - p += 1 - self._position = p - return None - - def skipUntil(self, chars): - p = self.position - while p < len(self): - c = self[p:p + 1] - if c in chars: - self._position = p - return c - p += 1 - self._position = p - return None - - def matchBytes(self, bytes): - """Look for a sequence of bytes at the start of a string. If the bytes - are found return True and advance the position to the byte after the - match. Otherwise return False and leave the position alone""" - p = self.position - data = self[p:p + len(bytes)] - rv = data.startswith(bytes) - if rv: - self.position += len(bytes) - return rv - - def jumpTo(self, bytes): - """Look for the next sequence of bytes matching a given sequence. If - a match is found advance the position to the last byte of the match""" - newPosition = self[self.position:].find(bytes) - if newPosition > -1: - # XXX: This is ugly, but I can't see a nicer way to fix this. - if self._position == -1: - self._position = 0 - self._position += (newPosition + len(bytes) - 1) - return True - else: - raise StopIteration - - -class EncodingParser(object): - """Mini parser for detecting character encoding from meta elements""" - - def __init__(self, data): - """string - the data to work on for encoding detection""" - self.data = EncodingBytes(data) - self.encoding = None - - def getEncoding(self): - methodDispatch = ( - (b"") - - def handleMeta(self): - if self.data.currentByte not in spaceCharactersBytes: - # if we have ") - - def getAttribute(self): - """Return a name,value pair for the next attribute in the stream, - if one is found, or None""" - data = self.data - # Step 1 (skip chars) - c = data.skip(spaceCharactersBytes | frozenset([b"/"])) - assert c is None or len(c) == 1 - # Step 2 - if c in (b">", None): - return None - # Step 3 - attrName = [] - attrValue = [] - # Step 4 attribute name - while True: - if c == b"=" and attrName: - break - elif c in spaceCharactersBytes: - # Step 6! - c = data.skip() - break - elif c in (b"/", b">"): - return b"".join(attrName), b"" - elif c in asciiUppercaseBytes: - attrName.append(c.lower()) - elif c is None: - return None - else: - attrName.append(c) - # Step 5 - c = next(data) - # Step 7 - if c != b"=": - data.previous() - return b"".join(attrName), b"" - # Step 8 - next(data) - # Step 9 - c = data.skip() - # Step 10 - if c in (b"'", b'"'): - # 10.1 - quoteChar = c - while True: - # 10.2 - c = next(data) - # 10.3 - if c == quoteChar: - next(data) - return b"".join(attrName), b"".join(attrValue) - # 10.4 - elif c in asciiUppercaseBytes: - attrValue.append(c.lower()) - # 10.5 - else: - attrValue.append(c) - elif c == b">": - return b"".join(attrName), b"" - elif c in asciiUppercaseBytes: - attrValue.append(c.lower()) - elif c is None: - return None - else: - attrValue.append(c) - # Step 11 - while True: - c = next(data) - if c in spacesAngleBrackets: - return b"".join(attrName), b"".join(attrValue) - elif c in asciiUppercaseBytes: - attrValue.append(c.lower()) - elif c is None: - return None - else: - attrValue.append(c) - - -class ContentAttrParser(object): - def __init__(self, data): - assert isinstance(data, bytes) - self.data = data - - def parse(self): - try: - # Check if the attr name is charset - # otherwise return - self.data.jumpTo(b"charset") - self.data.position += 1 - self.data.skip() - if not self.data.currentByte == b"=": - # If there is no = sign keep looking for attrs - return None - self.data.position += 1 - self.data.skip() - # Look for an encoding between matching quote marks - if self.data.currentByte in (b'"', b"'"): - quoteMark = self.data.currentByte - self.data.position += 1 - oldPosition = self.data.position - if self.data.jumpTo(quoteMark): - return self.data[oldPosition:self.data.position] - else: - return None - else: - # Unquoted value - oldPosition = self.data.position - try: - self.data.skipUntil(spaceCharactersBytes) - return self.data[oldPosition:self.data.position] - except StopIteration: - # Return the whole remaining value - return self.data[oldPosition:] - except StopIteration: - return None - - -def lookupEncoding(encoding): - """Return the python codec name corresponding to an encoding or None if the - string doesn't correspond to a valid encoding.""" - if isinstance(encoding, binary_type): - try: - encoding = encoding.decode("ascii") - except UnicodeDecodeError: - return None - - if encoding is not None: - try: - return webencodings.lookup(encoding) - except AttributeError: - return None - else: - return None diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib/_tokenizer.py b/src/fetchcode/vcs/pip/_vendor/html5lib/_tokenizer.py deleted file mode 100644 index 26553180..00000000 --- a/src/fetchcode/vcs/pip/_vendor/html5lib/_tokenizer.py +++ /dev/null @@ -1,1721 +0,0 @@ -from __future__ import absolute_import, division, unicode_literals - -from fetchcode.vcs.pip._vendor.six import unichr as chr - -from collections import deque - -from .constants import spaceCharacters -from .constants import entities -from .constants import asciiLetters, asciiUpper2Lower -from .constants import digits, hexDigits, EOF -from .constants import tokenTypes, tagTokenTypes -from .constants import replacementCharacters - -from ._inputstream import HTMLInputStream - -from ._trie import Trie - -entitiesTrie = Trie(entities) - - -class HTMLTokenizer(object): - """ This class takes care of tokenizing HTML. - - * self.currentToken - Holds the token that is currently being processed. - - * self.state - Holds a reference to the method to be invoked... XXX - - * self.stream - Points to HTMLInputStream object. - """ - - def __init__(self, stream, parser=None, **kwargs): - - self.stream = HTMLInputStream(stream, **kwargs) - self.parser = parser - - # Setup the initial tokenizer state - self.escapeFlag = False - self.lastFourChars = [] - self.state = self.dataState - self.escape = False - - # The current token being created - self.currentToken = None - super(HTMLTokenizer, self).__init__() - - def __iter__(self): - """ This is where the magic happens. - - We do our usually processing through the states and when we have a token - to return we yield the token which pauses processing until the next token - is requested. - """ - self.tokenQueue = deque([]) - # Start processing. When EOF is reached self.state will return False - # instead of True and the loop will terminate. - while self.state(): - while self.stream.errors: - yield {"type": tokenTypes["ParseError"], "data": self.stream.errors.pop(0)} - while self.tokenQueue: - yield self.tokenQueue.popleft() - - def consumeNumberEntity(self, isHex): - """This function returns either U+FFFD or the character based on the - decimal or hexadecimal representation. It also discards ";" if present. - If not present self.tokenQueue.append({"type": tokenTypes["ParseError"]}) is invoked. - """ - - allowed = digits - radix = 10 - if isHex: - allowed = hexDigits - radix = 16 - - charStack = [] - - # Consume all the characters that are in range while making sure we - # don't hit an EOF. - c = self.stream.char() - while c in allowed and c is not EOF: - charStack.append(c) - c = self.stream.char() - - # Convert the set of characters consumed to an int. - charAsInt = int("".join(charStack), radix) - - # Certain characters get replaced with others - if charAsInt in replacementCharacters: - char = replacementCharacters[charAsInt] - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "illegal-codepoint-for-numeric-entity", - "datavars": {"charAsInt": charAsInt}}) - elif ((0xD800 <= charAsInt <= 0xDFFF) or - (charAsInt > 0x10FFFF)): - char = "\uFFFD" - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "illegal-codepoint-for-numeric-entity", - "datavars": {"charAsInt": charAsInt}}) - else: - # Should speed up this check somehow (e.g. move the set to a constant) - if ((0x0001 <= charAsInt <= 0x0008) or - (0x000E <= charAsInt <= 0x001F) or - (0x007F <= charAsInt <= 0x009F) or - (0xFDD0 <= charAsInt <= 0xFDEF) or - charAsInt in frozenset([0x000B, 0xFFFE, 0xFFFF, 0x1FFFE, - 0x1FFFF, 0x2FFFE, 0x2FFFF, 0x3FFFE, - 0x3FFFF, 0x4FFFE, 0x4FFFF, 0x5FFFE, - 0x5FFFF, 0x6FFFE, 0x6FFFF, 0x7FFFE, - 0x7FFFF, 0x8FFFE, 0x8FFFF, 0x9FFFE, - 0x9FFFF, 0xAFFFE, 0xAFFFF, 0xBFFFE, - 0xBFFFF, 0xCFFFE, 0xCFFFF, 0xDFFFE, - 0xDFFFF, 0xEFFFE, 0xEFFFF, 0xFFFFE, - 0xFFFFF, 0x10FFFE, 0x10FFFF])): - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": - "illegal-codepoint-for-numeric-entity", - "datavars": {"charAsInt": charAsInt}}) - try: - # Try/except needed as UCS-2 Python builds' unichar only works - # within the BMP. - char = chr(charAsInt) - except ValueError: - v = charAsInt - 0x10000 - char = chr(0xD800 | (v >> 10)) + chr(0xDC00 | (v & 0x3FF)) - - # Discard the ; if present. Otherwise, put it back on the queue and - # invoke parseError on parser. - if c != ";": - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "numeric-entity-without-semicolon"}) - self.stream.unget(c) - - return char - - def consumeEntity(self, allowedChar=None, fromAttribute=False): - # Initialise to the default output for when no entity is matched - output = "&" - - charStack = [self.stream.char()] - if (charStack[0] in spaceCharacters or charStack[0] in (EOF, "<", "&") or - (allowedChar is not None and allowedChar == charStack[0])): - self.stream.unget(charStack[0]) - - elif charStack[0] == "#": - # Read the next character to see if it's hex or decimal - hex = False - charStack.append(self.stream.char()) - if charStack[-1] in ("x", "X"): - hex = True - charStack.append(self.stream.char()) - - # charStack[-1] should be the first digit - if (hex and charStack[-1] in hexDigits) \ - or (not hex and charStack[-1] in digits): - # At least one digit found, so consume the whole number - self.stream.unget(charStack[-1]) - output = self.consumeNumberEntity(hex) - else: - # No digits found - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "expected-numeric-entity"}) - self.stream.unget(charStack.pop()) - output = "&" + "".join(charStack) - - else: - # At this point in the process might have named entity. Entities - # are stored in the global variable "entities". - # - # Consume characters and compare to these to a substring of the - # entity names in the list until the substring no longer matches. - while (charStack[-1] is not EOF): - if not entitiesTrie.has_keys_with_prefix("".join(charStack)): - break - charStack.append(self.stream.char()) - - # At this point we have a string that starts with some characters - # that may match an entity - # Try to find the longest entity the string will match to take care - # of ¬i for instance. - try: - entityName = entitiesTrie.longest_prefix("".join(charStack[:-1])) - entityLength = len(entityName) - except KeyError: - entityName = None - - if entityName is not None: - if entityName[-1] != ";": - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "named-entity-without-semicolon"}) - if (entityName[-1] != ";" and fromAttribute and - (charStack[entityLength] in asciiLetters or - charStack[entityLength] in digits or - charStack[entityLength] == "=")): - self.stream.unget(charStack.pop()) - output = "&" + "".join(charStack) - else: - output = entities[entityName] - self.stream.unget(charStack.pop()) - output += "".join(charStack[entityLength:]) - else: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "expected-named-entity"}) - self.stream.unget(charStack.pop()) - output = "&" + "".join(charStack) - - if fromAttribute: - self.currentToken["data"][-1][1] += output - else: - if output in spaceCharacters: - tokenType = "SpaceCharacters" - else: - tokenType = "Characters" - self.tokenQueue.append({"type": tokenTypes[tokenType], "data": output}) - - def processEntityInAttribute(self, allowedChar): - """This method replaces the need for "entityInAttributeValueState". - """ - self.consumeEntity(allowedChar=allowedChar, fromAttribute=True) - - def emitCurrentToken(self): - """This method is a generic handler for emitting the tags. It also sets - the state to "data" because that's what's needed after a token has been - emitted. - """ - token = self.currentToken - # Add token to the queue to be yielded - if (token["type"] in tagTokenTypes): - token["name"] = token["name"].translate(asciiUpper2Lower) - if token["type"] == tokenTypes["EndTag"]: - if token["data"]: - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "attributes-in-end-tag"}) - if token["selfClosing"]: - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "self-closing-flag-on-end-tag"}) - self.tokenQueue.append(token) - self.state = self.dataState - - # Below are the various tokenizer states worked out. - def dataState(self): - data = self.stream.char() - if data == "&": - self.state = self.entityDataState - elif data == "<": - self.state = self.tagOpenState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.tokenQueue.append({"type": tokenTypes["Characters"], - "data": "\u0000"}) - elif data is EOF: - # Tokenization ends. - return False - elif data in spaceCharacters: - # Directly after emitting a token you switch back to the "data - # state". At that point spaceCharacters are important so they are - # emitted separately. - self.tokenQueue.append({"type": tokenTypes["SpaceCharacters"], "data": - data + self.stream.charsUntil(spaceCharacters, True)}) - # No need to update lastFourChars here, since the first space will - # have already been appended to lastFourChars and will have broken - # any sequences - else: - chars = self.stream.charsUntil(("&", "<", "\u0000")) - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": - data + chars}) - return True - - def entityDataState(self): - self.consumeEntity() - self.state = self.dataState - return True - - def rcdataState(self): - data = self.stream.char() - if data == "&": - self.state = self.characterReferenceInRcdata - elif data == "<": - self.state = self.rcdataLessThanSignState - elif data == EOF: - # Tokenization ends. - return False - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.tokenQueue.append({"type": tokenTypes["Characters"], - "data": "\uFFFD"}) - elif data in spaceCharacters: - # Directly after emitting a token you switch back to the "data - # state". At that point spaceCharacters are important so they are - # emitted separately. - self.tokenQueue.append({"type": tokenTypes["SpaceCharacters"], "data": - data + self.stream.charsUntil(spaceCharacters, True)}) - # No need to update lastFourChars here, since the first space will - # have already been appended to lastFourChars and will have broken - # any sequences - else: - chars = self.stream.charsUntil(("&", "<", "\u0000")) - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": - data + chars}) - return True - - def characterReferenceInRcdata(self): - self.consumeEntity() - self.state = self.rcdataState - return True - - def rawtextState(self): - data = self.stream.char() - if data == "<": - self.state = self.rawtextLessThanSignState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.tokenQueue.append({"type": tokenTypes["Characters"], - "data": "\uFFFD"}) - elif data == EOF: - # Tokenization ends. - return False - else: - chars = self.stream.charsUntil(("<", "\u0000")) - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": - data + chars}) - return True - - def scriptDataState(self): - data = self.stream.char() - if data == "<": - self.state = self.scriptDataLessThanSignState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.tokenQueue.append({"type": tokenTypes["Characters"], - "data": "\uFFFD"}) - elif data == EOF: - # Tokenization ends. - return False - else: - chars = self.stream.charsUntil(("<", "\u0000")) - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": - data + chars}) - return True - - def plaintextState(self): - data = self.stream.char() - if data == EOF: - # Tokenization ends. - return False - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.tokenQueue.append({"type": tokenTypes["Characters"], - "data": "\uFFFD"}) - else: - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": - data + self.stream.charsUntil("\u0000")}) - return True - - def tagOpenState(self): - data = self.stream.char() - if data == "!": - self.state = self.markupDeclarationOpenState - elif data == "/": - self.state = self.closeTagOpenState - elif data in asciiLetters: - self.currentToken = {"type": tokenTypes["StartTag"], - "name": data, "data": [], - "selfClosing": False, - "selfClosingAcknowledged": False} - self.state = self.tagNameState - elif data == ">": - # XXX In theory it could be something besides a tag name. But - # do we really care? - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "expected-tag-name-but-got-right-bracket"}) - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "<>"}) - self.state = self.dataState - elif data == "?": - # XXX In theory it could be something besides a tag name. But - # do we really care? - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "expected-tag-name-but-got-question-mark"}) - self.stream.unget(data) - self.state = self.bogusCommentState - else: - # XXX - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "expected-tag-name"}) - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "<"}) - self.stream.unget(data) - self.state = self.dataState - return True - - def closeTagOpenState(self): - data = self.stream.char() - if data in asciiLetters: - self.currentToken = {"type": tokenTypes["EndTag"], "name": data, - "data": [], "selfClosing": False} - self.state = self.tagNameState - elif data == ">": - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "expected-closing-tag-but-got-right-bracket"}) - self.state = self.dataState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "expected-closing-tag-but-got-eof"}) - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "": - self.emitCurrentToken() - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-tag-name"}) - self.state = self.dataState - elif data == "/": - self.state = self.selfClosingStartTagState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.currentToken["name"] += "\uFFFD" - else: - self.currentToken["name"] += data - # (Don't use charsUntil here, because tag names are - # very short and it's faster to not do anything fancy) - return True - - def rcdataLessThanSignState(self): - data = self.stream.char() - if data == "/": - self.temporaryBuffer = "" - self.state = self.rcdataEndTagOpenState - else: - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "<"}) - self.stream.unget(data) - self.state = self.rcdataState - return True - - def rcdataEndTagOpenState(self): - data = self.stream.char() - if data in asciiLetters: - self.temporaryBuffer += data - self.state = self.rcdataEndTagNameState - else: - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "" and appropriate: - self.currentToken = {"type": tokenTypes["EndTag"], - "name": self.temporaryBuffer, - "data": [], "selfClosing": False} - self.emitCurrentToken() - self.state = self.dataState - elif data in asciiLetters: - self.temporaryBuffer += data - else: - self.tokenQueue.append({"type": tokenTypes["Characters"], - "data": "" and appropriate: - self.currentToken = {"type": tokenTypes["EndTag"], - "name": self.temporaryBuffer, - "data": [], "selfClosing": False} - self.emitCurrentToken() - self.state = self.dataState - elif data in asciiLetters: - self.temporaryBuffer += data - else: - self.tokenQueue.append({"type": tokenTypes["Characters"], - "data": "" and appropriate: - self.currentToken = {"type": tokenTypes["EndTag"], - "name": self.temporaryBuffer, - "data": [], "selfClosing": False} - self.emitCurrentToken() - self.state = self.dataState - elif data in asciiLetters: - self.temporaryBuffer += data - else: - self.tokenQueue.append({"type": tokenTypes["Characters"], - "data": "": - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": ">"}) - self.state = self.scriptDataState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.tokenQueue.append({"type": tokenTypes["Characters"], - "data": "\uFFFD"}) - self.state = self.scriptDataEscapedState - elif data == EOF: - self.state = self.dataState - else: - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": data}) - self.state = self.scriptDataEscapedState - return True - - def scriptDataEscapedLessThanSignState(self): - data = self.stream.char() - if data == "/": - self.temporaryBuffer = "" - self.state = self.scriptDataEscapedEndTagOpenState - elif data in asciiLetters: - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "<" + data}) - self.temporaryBuffer = data - self.state = self.scriptDataDoubleEscapeStartState - else: - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "<"}) - self.stream.unget(data) - self.state = self.scriptDataEscapedState - return True - - def scriptDataEscapedEndTagOpenState(self): - data = self.stream.char() - if data in asciiLetters: - self.temporaryBuffer = data - self.state = self.scriptDataEscapedEndTagNameState - else: - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "" and appropriate: - self.currentToken = {"type": tokenTypes["EndTag"], - "name": self.temporaryBuffer, - "data": [], "selfClosing": False} - self.emitCurrentToken() - self.state = self.dataState - elif data in asciiLetters: - self.temporaryBuffer += data - else: - self.tokenQueue.append({"type": tokenTypes["Characters"], - "data": ""))): - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": data}) - if self.temporaryBuffer.lower() == "script": - self.state = self.scriptDataDoubleEscapedState - else: - self.state = self.scriptDataEscapedState - elif data in asciiLetters: - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": data}) - self.temporaryBuffer += data - else: - self.stream.unget(data) - self.state = self.scriptDataEscapedState - return True - - def scriptDataDoubleEscapedState(self): - data = self.stream.char() - if data == "-": - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "-"}) - self.state = self.scriptDataDoubleEscapedDashState - elif data == "<": - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "<"}) - self.state = self.scriptDataDoubleEscapedLessThanSignState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.tokenQueue.append({"type": tokenTypes["Characters"], - "data": "\uFFFD"}) - elif data == EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-script-in-script"}) - self.state = self.dataState - else: - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": data}) - return True - - def scriptDataDoubleEscapedDashState(self): - data = self.stream.char() - if data == "-": - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "-"}) - self.state = self.scriptDataDoubleEscapedDashDashState - elif data == "<": - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "<"}) - self.state = self.scriptDataDoubleEscapedLessThanSignState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.tokenQueue.append({"type": tokenTypes["Characters"], - "data": "\uFFFD"}) - self.state = self.scriptDataDoubleEscapedState - elif data == EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-script-in-script"}) - self.state = self.dataState - else: - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": data}) - self.state = self.scriptDataDoubleEscapedState - return True - - def scriptDataDoubleEscapedDashDashState(self): - data = self.stream.char() - if data == "-": - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "-"}) - elif data == "<": - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "<"}) - self.state = self.scriptDataDoubleEscapedLessThanSignState - elif data == ">": - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": ">"}) - self.state = self.scriptDataState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.tokenQueue.append({"type": tokenTypes["Characters"], - "data": "\uFFFD"}) - self.state = self.scriptDataDoubleEscapedState - elif data == EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-script-in-script"}) - self.state = self.dataState - else: - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": data}) - self.state = self.scriptDataDoubleEscapedState - return True - - def scriptDataDoubleEscapedLessThanSignState(self): - data = self.stream.char() - if data == "/": - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "/"}) - self.temporaryBuffer = "" - self.state = self.scriptDataDoubleEscapeEndState - else: - self.stream.unget(data) - self.state = self.scriptDataDoubleEscapedState - return True - - def scriptDataDoubleEscapeEndState(self): - data = self.stream.char() - if data in (spaceCharacters | frozenset(("/", ">"))): - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": data}) - if self.temporaryBuffer.lower() == "script": - self.state = self.scriptDataEscapedState - else: - self.state = self.scriptDataDoubleEscapedState - elif data in asciiLetters: - self.tokenQueue.append({"type": tokenTypes["Characters"], "data": data}) - self.temporaryBuffer += data - else: - self.stream.unget(data) - self.state = self.scriptDataDoubleEscapedState - return True - - def beforeAttributeNameState(self): - data = self.stream.char() - if data in spaceCharacters: - self.stream.charsUntil(spaceCharacters, True) - elif data in asciiLetters: - self.currentToken["data"].append([data, ""]) - self.state = self.attributeNameState - elif data == ">": - self.emitCurrentToken() - elif data == "/": - self.state = self.selfClosingStartTagState - elif data in ("'", '"', "=", "<"): - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "invalid-character-in-attribute-name"}) - self.currentToken["data"].append([data, ""]) - self.state = self.attributeNameState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.currentToken["data"].append(["\uFFFD", ""]) - self.state = self.attributeNameState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "expected-attribute-name-but-got-eof"}) - self.state = self.dataState - else: - self.currentToken["data"].append([data, ""]) - self.state = self.attributeNameState - return True - - def attributeNameState(self): - data = self.stream.char() - leavingThisState = True - emitToken = False - if data == "=": - self.state = self.beforeAttributeValueState - elif data in asciiLetters: - self.currentToken["data"][-1][0] += data +\ - self.stream.charsUntil(asciiLetters, True) - leavingThisState = False - elif data == ">": - # XXX If we emit here the attributes are converted to a dict - # without being checked and when the code below runs we error - # because data is a dict not a list - emitToken = True - elif data in spaceCharacters: - self.state = self.afterAttributeNameState - elif data == "/": - self.state = self.selfClosingStartTagState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.currentToken["data"][-1][0] += "\uFFFD" - leavingThisState = False - elif data in ("'", '"', "<"): - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": - "invalid-character-in-attribute-name"}) - self.currentToken["data"][-1][0] += data - leavingThisState = False - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "eof-in-attribute-name"}) - self.state = self.dataState - else: - self.currentToken["data"][-1][0] += data - leavingThisState = False - - if leavingThisState: - # Attributes are not dropped at this stage. That happens when the - # start tag token is emitted so values can still be safely appended - # to attributes, but we do want to report the parse error in time. - self.currentToken["data"][-1][0] = ( - self.currentToken["data"][-1][0].translate(asciiUpper2Lower)) - for name, _ in self.currentToken["data"][:-1]: - if self.currentToken["data"][-1][0] == name: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "duplicate-attribute"}) - break - # XXX Fix for above XXX - if emitToken: - self.emitCurrentToken() - return True - - def afterAttributeNameState(self): - data = self.stream.char() - if data in spaceCharacters: - self.stream.charsUntil(spaceCharacters, True) - elif data == "=": - self.state = self.beforeAttributeValueState - elif data == ">": - self.emitCurrentToken() - elif data in asciiLetters: - self.currentToken["data"].append([data, ""]) - self.state = self.attributeNameState - elif data == "/": - self.state = self.selfClosingStartTagState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.currentToken["data"].append(["\uFFFD", ""]) - self.state = self.attributeNameState - elif data in ("'", '"', "<"): - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "invalid-character-after-attribute-name"}) - self.currentToken["data"].append([data, ""]) - self.state = self.attributeNameState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "expected-end-of-tag-but-got-eof"}) - self.state = self.dataState - else: - self.currentToken["data"].append([data, ""]) - self.state = self.attributeNameState - return True - - def beforeAttributeValueState(self): - data = self.stream.char() - if data in spaceCharacters: - self.stream.charsUntil(spaceCharacters, True) - elif data == "\"": - self.state = self.attributeValueDoubleQuotedState - elif data == "&": - self.state = self.attributeValueUnQuotedState - self.stream.unget(data) - elif data == "'": - self.state = self.attributeValueSingleQuotedState - elif data == ">": - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "expected-attribute-value-but-got-right-bracket"}) - self.emitCurrentToken() - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.currentToken["data"][-1][1] += "\uFFFD" - self.state = self.attributeValueUnQuotedState - elif data in ("=", "<", "`"): - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "equals-in-unquoted-attribute-value"}) - self.currentToken["data"][-1][1] += data - self.state = self.attributeValueUnQuotedState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "expected-attribute-value-but-got-eof"}) - self.state = self.dataState - else: - self.currentToken["data"][-1][1] += data - self.state = self.attributeValueUnQuotedState - return True - - def attributeValueDoubleQuotedState(self): - data = self.stream.char() - if data == "\"": - self.state = self.afterAttributeValueState - elif data == "&": - self.processEntityInAttribute('"') - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.currentToken["data"][-1][1] += "\uFFFD" - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-attribute-value-double-quote"}) - self.state = self.dataState - else: - self.currentToken["data"][-1][1] += data +\ - self.stream.charsUntil(("\"", "&", "\u0000")) - return True - - def attributeValueSingleQuotedState(self): - data = self.stream.char() - if data == "'": - self.state = self.afterAttributeValueState - elif data == "&": - self.processEntityInAttribute("'") - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.currentToken["data"][-1][1] += "\uFFFD" - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-attribute-value-single-quote"}) - self.state = self.dataState - else: - self.currentToken["data"][-1][1] += data +\ - self.stream.charsUntil(("'", "&", "\u0000")) - return True - - def attributeValueUnQuotedState(self): - data = self.stream.char() - if data in spaceCharacters: - self.state = self.beforeAttributeNameState - elif data == "&": - self.processEntityInAttribute(">") - elif data == ">": - self.emitCurrentToken() - elif data in ('"', "'", "=", "<", "`"): - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-character-in-unquoted-attribute-value"}) - self.currentToken["data"][-1][1] += data - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.currentToken["data"][-1][1] += "\uFFFD" - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-attribute-value-no-quotes"}) - self.state = self.dataState - else: - self.currentToken["data"][-1][1] += data + self.stream.charsUntil( - frozenset(("&", ">", '"', "'", "=", "<", "`", "\u0000")) | spaceCharacters) - return True - - def afterAttributeValueState(self): - data = self.stream.char() - if data in spaceCharacters: - self.state = self.beforeAttributeNameState - elif data == ">": - self.emitCurrentToken() - elif data == "/": - self.state = self.selfClosingStartTagState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-EOF-after-attribute-value"}) - self.stream.unget(data) - self.state = self.dataState - else: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-character-after-attribute-value"}) - self.stream.unget(data) - self.state = self.beforeAttributeNameState - return True - - def selfClosingStartTagState(self): - data = self.stream.char() - if data == ">": - self.currentToken["selfClosing"] = True - self.emitCurrentToken() - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": - "unexpected-EOF-after-solidus-in-tag"}) - self.stream.unget(data) - self.state = self.dataState - else: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-character-after-solidus-in-tag"}) - self.stream.unget(data) - self.state = self.beforeAttributeNameState - return True - - def bogusCommentState(self): - # Make a new comment token and give it as value all the characters - # until the first > or EOF (charsUntil checks for EOF automatically) - # and emit it. - data = self.stream.charsUntil(">") - data = data.replace("\u0000", "\uFFFD") - self.tokenQueue.append( - {"type": tokenTypes["Comment"], "data": data}) - - # Eat the character directly after the bogus comment which is either a - # ">" or an EOF. - self.stream.char() - self.state = self.dataState - return True - - def markupDeclarationOpenState(self): - charStack = [self.stream.char()] - if charStack[-1] == "-": - charStack.append(self.stream.char()) - if charStack[-1] == "-": - self.currentToken = {"type": tokenTypes["Comment"], "data": ""} - self.state = self.commentStartState - return True - elif charStack[-1] in ('d', 'D'): - matched = True - for expected in (('o', 'O'), ('c', 'C'), ('t', 'T'), - ('y', 'Y'), ('p', 'P'), ('e', 'E')): - charStack.append(self.stream.char()) - if charStack[-1] not in expected: - matched = False - break - if matched: - self.currentToken = {"type": tokenTypes["Doctype"], - "name": "", - "publicId": None, "systemId": None, - "correct": True} - self.state = self.doctypeState - return True - elif (charStack[-1] == "[" and - self.parser is not None and - self.parser.tree.openElements and - self.parser.tree.openElements[-1].namespace != self.parser.tree.defaultNamespace): - matched = True - for expected in ["C", "D", "A", "T", "A", "["]: - charStack.append(self.stream.char()) - if charStack[-1] != expected: - matched = False - break - if matched: - self.state = self.cdataSectionState - return True - - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "expected-dashes-or-doctype"}) - - while charStack: - self.stream.unget(charStack.pop()) - self.state = self.bogusCommentState - return True - - def commentStartState(self): - data = self.stream.char() - if data == "-": - self.state = self.commentStartDashState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.currentToken["data"] += "\uFFFD" - elif data == ">": - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "incorrect-comment"}) - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-comment"}) - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - self.currentToken["data"] += data - self.state = self.commentState - return True - - def commentStartDashState(self): - data = self.stream.char() - if data == "-": - self.state = self.commentEndState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.currentToken["data"] += "-\uFFFD" - elif data == ">": - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "incorrect-comment"}) - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-comment"}) - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - self.currentToken["data"] += "-" + data - self.state = self.commentState - return True - - def commentState(self): - data = self.stream.char() - if data == "-": - self.state = self.commentEndDashState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.currentToken["data"] += "\uFFFD" - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "eof-in-comment"}) - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - self.currentToken["data"] += data + \ - self.stream.charsUntil(("-", "\u0000")) - return True - - def commentEndDashState(self): - data = self.stream.char() - if data == "-": - self.state = self.commentEndState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.currentToken["data"] += "-\uFFFD" - self.state = self.commentState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-comment-end-dash"}) - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - self.currentToken["data"] += "-" + data - self.state = self.commentState - return True - - def commentEndState(self): - data = self.stream.char() - if data == ">": - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.currentToken["data"] += "--\uFFFD" - self.state = self.commentState - elif data == "!": - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-bang-after-double-dash-in-comment"}) - self.state = self.commentEndBangState - elif data == "-": - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-dash-after-double-dash-in-comment"}) - self.currentToken["data"] += data - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-comment-double-dash"}) - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - # XXX - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-char-in-comment"}) - self.currentToken["data"] += "--" + data - self.state = self.commentState - return True - - def commentEndBangState(self): - data = self.stream.char() - if data == ">": - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - elif data == "-": - self.currentToken["data"] += "--!" - self.state = self.commentEndDashState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.currentToken["data"] += "--!\uFFFD" - self.state = self.commentState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-comment-end-bang-state"}) - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - self.currentToken["data"] += "--!" + data - self.state = self.commentState - return True - - def doctypeState(self): - data = self.stream.char() - if data in spaceCharacters: - self.state = self.beforeDoctypeNameState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "expected-doctype-name-but-got-eof"}) - self.currentToken["correct"] = False - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "need-space-after-doctype"}) - self.stream.unget(data) - self.state = self.beforeDoctypeNameState - return True - - def beforeDoctypeNameState(self): - data = self.stream.char() - if data in spaceCharacters: - pass - elif data == ">": - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "expected-doctype-name-but-got-right-bracket"}) - self.currentToken["correct"] = False - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.currentToken["name"] = "\uFFFD" - self.state = self.doctypeNameState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "expected-doctype-name-but-got-eof"}) - self.currentToken["correct"] = False - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - self.currentToken["name"] = data - self.state = self.doctypeNameState - return True - - def doctypeNameState(self): - data = self.stream.char() - if data in spaceCharacters: - self.currentToken["name"] = self.currentToken["name"].translate(asciiUpper2Lower) - self.state = self.afterDoctypeNameState - elif data == ">": - self.currentToken["name"] = self.currentToken["name"].translate(asciiUpper2Lower) - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.currentToken["name"] += "\uFFFD" - self.state = self.doctypeNameState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-doctype-name"}) - self.currentToken["correct"] = False - self.currentToken["name"] = self.currentToken["name"].translate(asciiUpper2Lower) - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - self.currentToken["name"] += data - return True - - def afterDoctypeNameState(self): - data = self.stream.char() - if data in spaceCharacters: - pass - elif data == ">": - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - elif data is EOF: - self.currentToken["correct"] = False - self.stream.unget(data) - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-doctype"}) - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - if data in ("p", "P"): - matched = True - for expected in (("u", "U"), ("b", "B"), ("l", "L"), - ("i", "I"), ("c", "C")): - data = self.stream.char() - if data not in expected: - matched = False - break - if matched: - self.state = self.afterDoctypePublicKeywordState - return True - elif data in ("s", "S"): - matched = True - for expected in (("y", "Y"), ("s", "S"), ("t", "T"), - ("e", "E"), ("m", "M")): - data = self.stream.char() - if data not in expected: - matched = False - break - if matched: - self.state = self.afterDoctypeSystemKeywordState - return True - - # All the characters read before the current 'data' will be - # [a-zA-Z], so they're garbage in the bogus doctype and can be - # discarded; only the latest character might be '>' or EOF - # and needs to be ungetted - self.stream.unget(data) - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "expected-space-or-right-bracket-in-doctype", "datavars": - {"data": data}}) - self.currentToken["correct"] = False - self.state = self.bogusDoctypeState - - return True - - def afterDoctypePublicKeywordState(self): - data = self.stream.char() - if data in spaceCharacters: - self.state = self.beforeDoctypePublicIdentifierState - elif data in ("'", '"'): - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-char-in-doctype"}) - self.stream.unget(data) - self.state = self.beforeDoctypePublicIdentifierState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-doctype"}) - self.currentToken["correct"] = False - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - self.stream.unget(data) - self.state = self.beforeDoctypePublicIdentifierState - return True - - def beforeDoctypePublicIdentifierState(self): - data = self.stream.char() - if data in spaceCharacters: - pass - elif data == "\"": - self.currentToken["publicId"] = "" - self.state = self.doctypePublicIdentifierDoubleQuotedState - elif data == "'": - self.currentToken["publicId"] = "" - self.state = self.doctypePublicIdentifierSingleQuotedState - elif data == ">": - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-end-of-doctype"}) - self.currentToken["correct"] = False - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-doctype"}) - self.currentToken["correct"] = False - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-char-in-doctype"}) - self.currentToken["correct"] = False - self.state = self.bogusDoctypeState - return True - - def doctypePublicIdentifierDoubleQuotedState(self): - data = self.stream.char() - if data == "\"": - self.state = self.afterDoctypePublicIdentifierState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.currentToken["publicId"] += "\uFFFD" - elif data == ">": - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-end-of-doctype"}) - self.currentToken["correct"] = False - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-doctype"}) - self.currentToken["correct"] = False - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - self.currentToken["publicId"] += data - return True - - def doctypePublicIdentifierSingleQuotedState(self): - data = self.stream.char() - if data == "'": - self.state = self.afterDoctypePublicIdentifierState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.currentToken["publicId"] += "\uFFFD" - elif data == ">": - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-end-of-doctype"}) - self.currentToken["correct"] = False - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-doctype"}) - self.currentToken["correct"] = False - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - self.currentToken["publicId"] += data - return True - - def afterDoctypePublicIdentifierState(self): - data = self.stream.char() - if data in spaceCharacters: - self.state = self.betweenDoctypePublicAndSystemIdentifiersState - elif data == ">": - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - elif data == '"': - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-char-in-doctype"}) - self.currentToken["systemId"] = "" - self.state = self.doctypeSystemIdentifierDoubleQuotedState - elif data == "'": - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-char-in-doctype"}) - self.currentToken["systemId"] = "" - self.state = self.doctypeSystemIdentifierSingleQuotedState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-doctype"}) - self.currentToken["correct"] = False - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-char-in-doctype"}) - self.currentToken["correct"] = False - self.state = self.bogusDoctypeState - return True - - def betweenDoctypePublicAndSystemIdentifiersState(self): - data = self.stream.char() - if data in spaceCharacters: - pass - elif data == ">": - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - elif data == '"': - self.currentToken["systemId"] = "" - self.state = self.doctypeSystemIdentifierDoubleQuotedState - elif data == "'": - self.currentToken["systemId"] = "" - self.state = self.doctypeSystemIdentifierSingleQuotedState - elif data == EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-doctype"}) - self.currentToken["correct"] = False - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-char-in-doctype"}) - self.currentToken["correct"] = False - self.state = self.bogusDoctypeState - return True - - def afterDoctypeSystemKeywordState(self): - data = self.stream.char() - if data in spaceCharacters: - self.state = self.beforeDoctypeSystemIdentifierState - elif data in ("'", '"'): - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-char-in-doctype"}) - self.stream.unget(data) - self.state = self.beforeDoctypeSystemIdentifierState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-doctype"}) - self.currentToken["correct"] = False - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - self.stream.unget(data) - self.state = self.beforeDoctypeSystemIdentifierState - return True - - def beforeDoctypeSystemIdentifierState(self): - data = self.stream.char() - if data in spaceCharacters: - pass - elif data == "\"": - self.currentToken["systemId"] = "" - self.state = self.doctypeSystemIdentifierDoubleQuotedState - elif data == "'": - self.currentToken["systemId"] = "" - self.state = self.doctypeSystemIdentifierSingleQuotedState - elif data == ">": - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-char-in-doctype"}) - self.currentToken["correct"] = False - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-doctype"}) - self.currentToken["correct"] = False - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-char-in-doctype"}) - self.currentToken["correct"] = False - self.state = self.bogusDoctypeState - return True - - def doctypeSystemIdentifierDoubleQuotedState(self): - data = self.stream.char() - if data == "\"": - self.state = self.afterDoctypeSystemIdentifierState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.currentToken["systemId"] += "\uFFFD" - elif data == ">": - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-end-of-doctype"}) - self.currentToken["correct"] = False - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-doctype"}) - self.currentToken["correct"] = False - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - self.currentToken["systemId"] += data - return True - - def doctypeSystemIdentifierSingleQuotedState(self): - data = self.stream.char() - if data == "'": - self.state = self.afterDoctypeSystemIdentifierState - elif data == "\u0000": - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - self.currentToken["systemId"] += "\uFFFD" - elif data == ">": - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-end-of-doctype"}) - self.currentToken["correct"] = False - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-doctype"}) - self.currentToken["correct"] = False - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - self.currentToken["systemId"] += data - return True - - def afterDoctypeSystemIdentifierState(self): - data = self.stream.char() - if data in spaceCharacters: - pass - elif data == ">": - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - elif data is EOF: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "eof-in-doctype"}) - self.currentToken["correct"] = False - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - self.tokenQueue.append({"type": tokenTypes["ParseError"], "data": - "unexpected-char-in-doctype"}) - self.state = self.bogusDoctypeState - return True - - def bogusDoctypeState(self): - data = self.stream.char() - if data == ">": - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - elif data is EOF: - # XXX EMIT - self.stream.unget(data) - self.tokenQueue.append(self.currentToken) - self.state = self.dataState - else: - pass - return True - - def cdataSectionState(self): - data = [] - while True: - data.append(self.stream.charsUntil("]")) - data.append(self.stream.charsUntil(">")) - char = self.stream.char() - if char == EOF: - break - else: - assert char == ">" - if data[-1][-2:] == "]]": - data[-1] = data[-1][:-2] - break - else: - data.append(char) - - data = "".join(data) # pylint:disable=redefined-variable-type - # Deal with null here rather than in the parser - nullCount = data.count("\u0000") - if nullCount > 0: - for _ in range(nullCount): - self.tokenQueue.append({"type": tokenTypes["ParseError"], - "data": "invalid-codepoint"}) - data = data.replace("\u0000", "\uFFFD") - if data: - self.tokenQueue.append({"type": tokenTypes["Characters"], - "data": data}) - self.state = self.dataState - return True diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib/_trie/__init__.py b/src/fetchcode/vcs/pip/_vendor/html5lib/_trie/__init__.py deleted file mode 100644 index a5ba4bf1..00000000 --- a/src/fetchcode/vcs/pip/_vendor/html5lib/_trie/__init__.py +++ /dev/null @@ -1,14 +0,0 @@ -from __future__ import absolute_import, division, unicode_literals - -from .py import Trie as PyTrie - -Trie = PyTrie - -# pylint:disable=wrong-import-position -try: - from .datrie import Trie as DATrie -except ImportError: - pass -else: - Trie = DATrie -# pylint:enable=wrong-import-position diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib/_trie/_base.py b/src/fetchcode/vcs/pip/_vendor/html5lib/_trie/_base.py deleted file mode 100644 index 6b71975f..00000000 --- a/src/fetchcode/vcs/pip/_vendor/html5lib/_trie/_base.py +++ /dev/null @@ -1,40 +0,0 @@ -from __future__ import absolute_import, division, unicode_literals - -try: - from collections.abc import Mapping -except ImportError: # Python 2.7 - from collections import Mapping - - -class Trie(Mapping): - """Abstract base class for tries""" - - def keys(self, prefix=None): - # pylint:disable=arguments-differ - keys = super(Trie, self).keys() - - if prefix is None: - return set(keys) - - return {x for x in keys if x.startswith(prefix)} - - def has_keys_with_prefix(self, prefix): - for key in self.keys(): - if key.startswith(prefix): - return True - - return False - - def longest_prefix(self, prefix): - if prefix in self: - return prefix - - for i in range(1, len(prefix) + 1): - if prefix[:-i] in self: - return prefix[:-i] - - raise KeyError(prefix) - - def longest_prefix_item(self, prefix): - lprefix = self.longest_prefix(prefix) - return (lprefix, self[lprefix]) diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib/_trie/datrie.py b/src/fetchcode/vcs/pip/_vendor/html5lib/_trie/datrie.py deleted file mode 100644 index 0237622b..00000000 --- a/src/fetchcode/vcs/pip/_vendor/html5lib/_trie/datrie.py +++ /dev/null @@ -1,44 +0,0 @@ -from __future__ import absolute_import, division, unicode_literals - -from datrie import Trie as DATrie -from fetchcode.vcs.pip._vendor.six import text_type - -from ._base import Trie as ABCTrie - - -class Trie(ABCTrie): - def __init__(self, data): - chars = set() - for key in data.keys(): - if not isinstance(key, text_type): - raise TypeError("All keys must be strings") - for char in key: - chars.add(char) - - self._data = DATrie("".join(chars)) - for key, value in data.items(): - self._data[key] = value - - def __contains__(self, key): - return key in self._data - - def __len__(self): - return len(self._data) - - def __iter__(self): - raise NotImplementedError() - - def __getitem__(self, key): - return self._data[key] - - def keys(self, prefix=None): - return self._data.keys(prefix) - - def has_keys_with_prefix(self, prefix): - return self._data.has_keys_with_prefix(prefix) - - def longest_prefix(self, prefix): - return self._data.longest_prefix(prefix) - - def longest_prefix_item(self, prefix): - return self._data.longest_prefix_item(prefix) diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib/_trie/py.py b/src/fetchcode/vcs/pip/_vendor/html5lib/_trie/py.py deleted file mode 100644 index b256949f..00000000 --- a/src/fetchcode/vcs/pip/_vendor/html5lib/_trie/py.py +++ /dev/null @@ -1,67 +0,0 @@ -from __future__ import absolute_import, division, unicode_literals -from fetchcode.vcs.pip._vendor.six import text_type - -from bisect import bisect_left - -from ._base import Trie as ABCTrie - - -class Trie(ABCTrie): - def __init__(self, data): - if not all(isinstance(x, text_type) for x in data.keys()): - raise TypeError("All keys must be strings") - - self._data = data - self._keys = sorted(data.keys()) - self._cachestr = "" - self._cachepoints = (0, len(data)) - - def __contains__(self, key): - return key in self._data - - def __len__(self): - return len(self._data) - - def __iter__(self): - return iter(self._data) - - def __getitem__(self, key): - return self._data[key] - - def keys(self, prefix=None): - if prefix is None or prefix == "" or not self._keys: - return set(self._keys) - - if prefix.startswith(self._cachestr): - lo, hi = self._cachepoints - start = i = bisect_left(self._keys, prefix, lo, hi) - else: - start = i = bisect_left(self._keys, prefix) - - keys = set() - if start == len(self._keys): - return keys - - while self._keys[i].startswith(prefix): - keys.add(self._keys[i]) - i += 1 - - self._cachestr = prefix - self._cachepoints = (start, i) - - return keys - - def has_keys_with_prefix(self, prefix): - if prefix in self._data: - return True - - if prefix.startswith(self._cachestr): - lo, hi = self._cachepoints - i = bisect_left(self._keys, prefix, lo, hi) - else: - i = bisect_left(self._keys, prefix) - - if i == len(self._keys): - return False - - return self._keys[i].startswith(prefix) diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib/_utils.py b/src/fetchcode/vcs/pip/_vendor/html5lib/_utils.py deleted file mode 100644 index 46ff3468..00000000 --- a/src/fetchcode/vcs/pip/_vendor/html5lib/_utils.py +++ /dev/null @@ -1,124 +0,0 @@ -from __future__ import absolute_import, division, unicode_literals - -from types import ModuleType - -from fetchcode.vcs.pip._vendor.six import text_type - -try: - import xml.etree.cElementTree as default_etree -except ImportError: - import xml.etree.ElementTree as default_etree - - -__all__ = ["default_etree", "MethodDispatcher", "isSurrogatePair", - "surrogatePairToCodepoint", "moduleFactoryFactory", - "supports_lone_surrogates"] - - -# Platforms not supporting lone surrogates (\uD800-\uDFFF) should be -# caught by the below test. In general this would be any platform -# using UTF-16 as its encoding of unicode strings, such as -# Jython. This is because UTF-16 itself is based on the use of such -# surrogates, and there is no mechanism to further escape such -# escapes. -try: - _x = eval('"\\uD800"') # pylint:disable=eval-used - if not isinstance(_x, text_type): - # We need this with u"" because of http://bugs.jython.org/issue2039 - _x = eval('u"\\uD800"') # pylint:disable=eval-used - assert isinstance(_x, text_type) -except: # pylint:disable=bare-except - supports_lone_surrogates = False -else: - supports_lone_surrogates = True - - -class MethodDispatcher(dict): - """Dict with 2 special properties: - - On initiation, keys that are lists, sets or tuples are converted to - multiple keys so accessing any one of the items in the original - list-like object returns the matching value - - md = MethodDispatcher({("foo", "bar"):"baz"}) - md["foo"] == "baz" - - A default value which can be set through the default attribute. - """ - - def __init__(self, items=()): - # Using _dictEntries instead of directly assigning to self is about - # twice as fast. Please do careful performance testing before changing - # anything here. - _dictEntries = [] - for name, value in items: - if isinstance(name, (list, tuple, frozenset, set)): - for item in name: - _dictEntries.append((item, value)) - else: - _dictEntries.append((name, value)) - dict.__init__(self, _dictEntries) - assert len(self) == len(_dictEntries) - self.default = None - - def __getitem__(self, key): - return dict.get(self, key, self.default) - - -# Some utility functions to deal with weirdness around UCS2 vs UCS4 -# python builds - -def isSurrogatePair(data): - return (len(data) == 2 and - ord(data[0]) >= 0xD800 and ord(data[0]) <= 0xDBFF and - ord(data[1]) >= 0xDC00 and ord(data[1]) <= 0xDFFF) - - -def surrogatePairToCodepoint(data): - char_val = (0x10000 + (ord(data[0]) - 0xD800) * 0x400 + - (ord(data[1]) - 0xDC00)) - return char_val - -# Module Factory Factory (no, this isn't Java, I know) -# Here to stop this being duplicated all over the place. - - -def moduleFactoryFactory(factory): - moduleCache = {} - - def moduleFactory(baseModule, *args, **kwargs): - if isinstance(ModuleType.__name__, type("")): - name = "_%s_factory" % baseModule.__name__ - else: - name = b"_%s_factory" % baseModule.__name__ - - kwargs_tuple = tuple(kwargs.items()) - - try: - return moduleCache[name][args][kwargs_tuple] - except KeyError: - mod = ModuleType(name) - objs = factory(baseModule, *args, **kwargs) - mod.__dict__.update(objs) - if "name" not in moduleCache: - moduleCache[name] = {} - if "args" not in moduleCache[name]: - moduleCache[name][args] = {} - if "kwargs" not in moduleCache[name][args]: - moduleCache[name][args][kwargs_tuple] = {} - moduleCache[name][args][kwargs_tuple] = mod - return mod - - return moduleFactory - - -def memoize(func): - cache = {} - - def wrapped(*args, **kwargs): - key = (tuple(args), tuple(kwargs.items())) - if key not in cache: - cache[key] = func(*args, **kwargs) - return cache[key] - - return wrapped diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib/constants.py b/src/fetchcode/vcs/pip/_vendor/html5lib/constants.py deleted file mode 100644 index 1ff80419..00000000 --- a/src/fetchcode/vcs/pip/_vendor/html5lib/constants.py +++ /dev/null @@ -1,2947 +0,0 @@ -from __future__ import absolute_import, division, unicode_literals - -import string - -EOF = None - -E = { - "null-character": - "Null character in input stream, replaced with U+FFFD.", - "invalid-codepoint": - "Invalid codepoint in stream.", - "incorrectly-placed-solidus": - "Solidus (/) incorrectly placed in tag.", - "incorrect-cr-newline-entity": - "Incorrect CR newline entity, replaced with LF.", - "illegal-windows-1252-entity": - "Entity used with illegal number (windows-1252 reference).", - "cant-convert-numeric-entity": - "Numeric entity couldn't be converted to character " - "(codepoint U+%(charAsInt)08x).", - "illegal-codepoint-for-numeric-entity": - "Numeric entity represents an illegal codepoint: " - "U+%(charAsInt)08x.", - "numeric-entity-without-semicolon": - "Numeric entity didn't end with ';'.", - "expected-numeric-entity-but-got-eof": - "Numeric entity expected. Got end of file instead.", - "expected-numeric-entity": - "Numeric entity expected but none found.", - "named-entity-without-semicolon": - "Named entity didn't end with ';'.", - "expected-named-entity": - "Named entity expected. Got none.", - "attributes-in-end-tag": - "End tag contains unexpected attributes.", - 'self-closing-flag-on-end-tag': - "End tag contains unexpected self-closing flag.", - "expected-tag-name-but-got-right-bracket": - "Expected tag name. Got '>' instead.", - "expected-tag-name-but-got-question-mark": - "Expected tag name. Got '?' instead. (HTML doesn't " - "support processing instructions.)", - "expected-tag-name": - "Expected tag name. Got something else instead", - "expected-closing-tag-but-got-right-bracket": - "Expected closing tag. Got '>' instead. Ignoring ''.", - "expected-closing-tag-but-got-eof": - "Expected closing tag. Unexpected end of file.", - "expected-closing-tag-but-got-char": - "Expected closing tag. Unexpected character '%(data)s' found.", - "eof-in-tag-name": - "Unexpected end of file in the tag name.", - "expected-attribute-name-but-got-eof": - "Unexpected end of file. Expected attribute name instead.", - "eof-in-attribute-name": - "Unexpected end of file in attribute name.", - "invalid-character-in-attribute-name": - "Invalid character in attribute name", - "duplicate-attribute": - "Dropped duplicate attribute on tag.", - "expected-end-of-tag-name-but-got-eof": - "Unexpected end of file. Expected = or end of tag.", - "expected-attribute-value-but-got-eof": - "Unexpected end of file. Expected attribute value.", - "expected-attribute-value-but-got-right-bracket": - "Expected attribute value. Got '>' instead.", - 'equals-in-unquoted-attribute-value': - "Unexpected = in unquoted attribute", - 'unexpected-character-in-unquoted-attribute-value': - "Unexpected character in unquoted attribute", - "invalid-character-after-attribute-name": - "Unexpected character after attribute name.", - "unexpected-character-after-attribute-value": - "Unexpected character after attribute value.", - "eof-in-attribute-value-double-quote": - "Unexpected end of file in attribute value (\").", - "eof-in-attribute-value-single-quote": - "Unexpected end of file in attribute value (').", - "eof-in-attribute-value-no-quotes": - "Unexpected end of file in attribute value.", - "unexpected-EOF-after-solidus-in-tag": - "Unexpected end of file in tag. Expected >", - "unexpected-character-after-solidus-in-tag": - "Unexpected character after / in tag. Expected >", - "expected-dashes-or-doctype": - "Expected '--' or 'DOCTYPE'. Not found.", - "unexpected-bang-after-double-dash-in-comment": - "Unexpected ! after -- in comment", - "unexpected-space-after-double-dash-in-comment": - "Unexpected space after -- in comment", - "incorrect-comment": - "Incorrect comment.", - "eof-in-comment": - "Unexpected end of file in comment.", - "eof-in-comment-end-dash": - "Unexpected end of file in comment (-)", - "unexpected-dash-after-double-dash-in-comment": - "Unexpected '-' after '--' found in comment.", - "eof-in-comment-double-dash": - "Unexpected end of file in comment (--).", - "eof-in-comment-end-space-state": - "Unexpected end of file in comment.", - "eof-in-comment-end-bang-state": - "Unexpected end of file in comment.", - "unexpected-char-in-comment": - "Unexpected character in comment found.", - "need-space-after-doctype": - "No space after literal string 'DOCTYPE'.", - "expected-doctype-name-but-got-right-bracket": - "Unexpected > character. Expected DOCTYPE name.", - "expected-doctype-name-but-got-eof": - "Unexpected end of file. Expected DOCTYPE name.", - "eof-in-doctype-name": - "Unexpected end of file in DOCTYPE name.", - "eof-in-doctype": - "Unexpected end of file in DOCTYPE.", - "expected-space-or-right-bracket-in-doctype": - "Expected space or '>'. Got '%(data)s'", - "unexpected-end-of-doctype": - "Unexpected end of DOCTYPE.", - "unexpected-char-in-doctype": - "Unexpected character in DOCTYPE.", - "eof-in-innerhtml": - "XXX innerHTML EOF", - "unexpected-doctype": - "Unexpected DOCTYPE. Ignored.", - "non-html-root": - "html needs to be the first start tag.", - "expected-doctype-but-got-eof": - "Unexpected End of file. Expected DOCTYPE.", - "unknown-doctype": - "Erroneous DOCTYPE.", - "expected-doctype-but-got-chars": - "Unexpected non-space characters. Expected DOCTYPE.", - "expected-doctype-but-got-start-tag": - "Unexpected start tag (%(name)s). Expected DOCTYPE.", - "expected-doctype-but-got-end-tag": - "Unexpected end tag (%(name)s). Expected DOCTYPE.", - "end-tag-after-implied-root": - "Unexpected end tag (%(name)s) after the (implied) root element.", - "expected-named-closing-tag-but-got-eof": - "Unexpected end of file. Expected end tag (%(name)s).", - "two-heads-are-not-better-than-one": - "Unexpected start tag head in existing head. Ignored.", - "unexpected-end-tag": - "Unexpected end tag (%(name)s). Ignored.", - "unexpected-start-tag-out-of-my-head": - "Unexpected start tag (%(name)s) that can be in head. Moved.", - "unexpected-start-tag": - "Unexpected start tag (%(name)s).", - "missing-end-tag": - "Missing end tag (%(name)s).", - "missing-end-tags": - "Missing end tags (%(name)s).", - "unexpected-start-tag-implies-end-tag": - "Unexpected start tag (%(startName)s) " - "implies end tag (%(endName)s).", - "unexpected-start-tag-treated-as": - "Unexpected start tag (%(originalName)s). Treated as %(newName)s.", - "deprecated-tag": - "Unexpected start tag %(name)s. Don't use it!", - "unexpected-start-tag-ignored": - "Unexpected start tag %(name)s. Ignored.", - "expected-one-end-tag-but-got-another": - "Unexpected end tag (%(gotName)s). " - "Missing end tag (%(expectedName)s).", - "end-tag-too-early": - "End tag (%(name)s) seen too early. Expected other end tag.", - "end-tag-too-early-named": - "Unexpected end tag (%(gotName)s). Expected end tag (%(expectedName)s).", - "end-tag-too-early-ignored": - "End tag (%(name)s) seen too early. Ignored.", - "adoption-agency-1.1": - "End tag (%(name)s) violates step 1, " - "paragraph 1 of the adoption agency algorithm.", - "adoption-agency-1.2": - "End tag (%(name)s) violates step 1, " - "paragraph 2 of the adoption agency algorithm.", - "adoption-agency-1.3": - "End tag (%(name)s) violates step 1, " - "paragraph 3 of the adoption agency algorithm.", - "adoption-agency-4.4": - "End tag (%(name)s) violates step 4, " - "paragraph 4 of the adoption agency algorithm.", - "unexpected-end-tag-treated-as": - "Unexpected end tag (%(originalName)s). Treated as %(newName)s.", - "no-end-tag": - "This element (%(name)s) has no end tag.", - "unexpected-implied-end-tag-in-table": - "Unexpected implied end tag (%(name)s) in the table phase.", - "unexpected-implied-end-tag-in-table-body": - "Unexpected implied end tag (%(name)s) in the table body phase.", - "unexpected-char-implies-table-voodoo": - "Unexpected non-space characters in " - "table context caused voodoo mode.", - "unexpected-hidden-input-in-table": - "Unexpected input with type hidden in table context.", - "unexpected-form-in-table": - "Unexpected form in table context.", - "unexpected-start-tag-implies-table-voodoo": - "Unexpected start tag (%(name)s) in " - "table context caused voodoo mode.", - "unexpected-end-tag-implies-table-voodoo": - "Unexpected end tag (%(name)s) in " - "table context caused voodoo mode.", - "unexpected-cell-in-table-body": - "Unexpected table cell start tag (%(name)s) " - "in the table body phase.", - "unexpected-cell-end-tag": - "Got table cell end tag (%(name)s) " - "while required end tags are missing.", - "unexpected-end-tag-in-table-body": - "Unexpected end tag (%(name)s) in the table body phase. Ignored.", - "unexpected-implied-end-tag-in-table-row": - "Unexpected implied end tag (%(name)s) in the table row phase.", - "unexpected-end-tag-in-table-row": - "Unexpected end tag (%(name)s) in the table row phase. Ignored.", - "unexpected-select-in-select": - "Unexpected select start tag in the select phase " - "treated as select end tag.", - "unexpected-input-in-select": - "Unexpected input start tag in the select phase.", - "unexpected-start-tag-in-select": - "Unexpected start tag token (%(name)s in the select phase. " - "Ignored.", - "unexpected-end-tag-in-select": - "Unexpected end tag (%(name)s) in the select phase. Ignored.", - "unexpected-table-element-start-tag-in-select-in-table": - "Unexpected table element start tag (%(name)s) in the select in table phase.", - "unexpected-table-element-end-tag-in-select-in-table": - "Unexpected table element end tag (%(name)s) in the select in table phase.", - "unexpected-char-after-body": - "Unexpected non-space characters in the after body phase.", - "unexpected-start-tag-after-body": - "Unexpected start tag token (%(name)s)" - " in the after body phase.", - "unexpected-end-tag-after-body": - "Unexpected end tag token (%(name)s)" - " in the after body phase.", - "unexpected-char-in-frameset": - "Unexpected characters in the frameset phase. Characters ignored.", - "unexpected-start-tag-in-frameset": - "Unexpected start tag token (%(name)s)" - " in the frameset phase. Ignored.", - "unexpected-frameset-in-frameset-innerhtml": - "Unexpected end tag token (frameset) " - "in the frameset phase (innerHTML).", - "unexpected-end-tag-in-frameset": - "Unexpected end tag token (%(name)s)" - " in the frameset phase. Ignored.", - "unexpected-char-after-frameset": - "Unexpected non-space characters in the " - "after frameset phase. Ignored.", - "unexpected-start-tag-after-frameset": - "Unexpected start tag (%(name)s)" - " in the after frameset phase. Ignored.", - "unexpected-end-tag-after-frameset": - "Unexpected end tag (%(name)s)" - " in the after frameset phase. Ignored.", - "unexpected-end-tag-after-body-innerhtml": - "Unexpected end tag after body(innerHtml)", - "expected-eof-but-got-char": - "Unexpected non-space characters. Expected end of file.", - "expected-eof-but-got-start-tag": - "Unexpected start tag (%(name)s)" - ". Expected end of file.", - "expected-eof-but-got-end-tag": - "Unexpected end tag (%(name)s)" - ". Expected end of file.", - "eof-in-table": - "Unexpected end of file. Expected table content.", - "eof-in-select": - "Unexpected end of file. Expected select content.", - "eof-in-frameset": - "Unexpected end of file. Expected frameset content.", - "eof-in-script-in-script": - "Unexpected end of file. Expected script content.", - "eof-in-foreign-lands": - "Unexpected end of file. Expected foreign content", - "non-void-element-with-trailing-solidus": - "Trailing solidus not allowed on element %(name)s", - "unexpected-html-element-in-foreign-content": - "Element %(name)s not allowed in a non-html context", - "unexpected-end-tag-before-html": - "Unexpected end tag (%(name)s) before html.", - "unexpected-inhead-noscript-tag": - "Element %(name)s not allowed in a inhead-noscript context", - "eof-in-head-noscript": - "Unexpected end of file. Expected inhead-noscript content", - "char-in-head-noscript": - "Unexpected non-space character. Expected inhead-noscript content", - "XXX-undefined-error": - "Undefined error (this sucks and should be fixed)", -} - -namespaces = { - "html": "http://www.w3.org/1999/xhtml", - "mathml": "http://www.w3.org/1998/Math/MathML", - "svg": "http://www.w3.org/2000/svg", - "xlink": "http://www.w3.org/1999/xlink", - "xml": "http://www.w3.org/XML/1998/namespace", - "xmlns": "http://www.w3.org/2000/xmlns/" -} - -scopingElements = frozenset([ - (namespaces["html"], "applet"), - (namespaces["html"], "caption"), - (namespaces["html"], "html"), - (namespaces["html"], "marquee"), - (namespaces["html"], "object"), - (namespaces["html"], "table"), - (namespaces["html"], "td"), - (namespaces["html"], "th"), - (namespaces["mathml"], "mi"), - (namespaces["mathml"], "mo"), - (namespaces["mathml"], "mn"), - (namespaces["mathml"], "ms"), - (namespaces["mathml"], "mtext"), - (namespaces["mathml"], "annotation-xml"), - (namespaces["svg"], "foreignObject"), - (namespaces["svg"], "desc"), - (namespaces["svg"], "title"), -]) - -formattingElements = frozenset([ - (namespaces["html"], "a"), - (namespaces["html"], "b"), - (namespaces["html"], "big"), - (namespaces["html"], "code"), - (namespaces["html"], "em"), - (namespaces["html"], "font"), - (namespaces["html"], "i"), - (namespaces["html"], "nobr"), - (namespaces["html"], "s"), - (namespaces["html"], "small"), - (namespaces["html"], "strike"), - (namespaces["html"], "strong"), - (namespaces["html"], "tt"), - (namespaces["html"], "u") -]) - -specialElements = frozenset([ - (namespaces["html"], "address"), - (namespaces["html"], "applet"), - (namespaces["html"], "area"), - (namespaces["html"], "article"), - (namespaces["html"], "aside"), - (namespaces["html"], "base"), - (namespaces["html"], "basefont"), - (namespaces["html"], "bgsound"), - (namespaces["html"], "blockquote"), - (namespaces["html"], "body"), - (namespaces["html"], "br"), - (namespaces["html"], "button"), - (namespaces["html"], "caption"), - (namespaces["html"], "center"), - (namespaces["html"], "col"), - (namespaces["html"], "colgroup"), - (namespaces["html"], "command"), - (namespaces["html"], "dd"), - (namespaces["html"], "details"), - (namespaces["html"], "dir"), - (namespaces["html"], "div"), - (namespaces["html"], "dl"), - (namespaces["html"], "dt"), - (namespaces["html"], "embed"), - (namespaces["html"], "fieldset"), - (namespaces["html"], "figure"), - (namespaces["html"], "footer"), - (namespaces["html"], "form"), - (namespaces["html"], "frame"), - (namespaces["html"], "frameset"), - (namespaces["html"], "h1"), - (namespaces["html"], "h2"), - (namespaces["html"], "h3"), - (namespaces["html"], "h4"), - (namespaces["html"], "h5"), - (namespaces["html"], "h6"), - (namespaces["html"], "head"), - (namespaces["html"], "header"), - (namespaces["html"], "hr"), - (namespaces["html"], "html"), - (namespaces["html"], "iframe"), - # Note that image is commented out in the spec as "this isn't an - # element that can end up on the stack, so it doesn't matter," - (namespaces["html"], "image"), - (namespaces["html"], "img"), - (namespaces["html"], "input"), - (namespaces["html"], "isindex"), - (namespaces["html"], "li"), - (namespaces["html"], "link"), - (namespaces["html"], "listing"), - (namespaces["html"], "marquee"), - (namespaces["html"], "menu"), - (namespaces["html"], "meta"), - (namespaces["html"], "nav"), - (namespaces["html"], "noembed"), - (namespaces["html"], "noframes"), - (namespaces["html"], "noscript"), - (namespaces["html"], "object"), - (namespaces["html"], "ol"), - (namespaces["html"], "p"), - (namespaces["html"], "param"), - (namespaces["html"], "plaintext"), - (namespaces["html"], "pre"), - (namespaces["html"], "script"), - (namespaces["html"], "section"), - (namespaces["html"], "select"), - (namespaces["html"], "style"), - (namespaces["html"], "table"), - (namespaces["html"], "tbody"), - (namespaces["html"], "td"), - (namespaces["html"], "textarea"), - (namespaces["html"], "tfoot"), - (namespaces["html"], "th"), - (namespaces["html"], "thead"), - (namespaces["html"], "title"), - (namespaces["html"], "tr"), - (namespaces["html"], "ul"), - (namespaces["html"], "wbr"), - (namespaces["html"], "xmp"), - (namespaces["svg"], "foreignObject") -]) - -htmlIntegrationPointElements = frozenset([ - (namespaces["mathml"], "annotation-xml"), - (namespaces["svg"], "foreignObject"), - (namespaces["svg"], "desc"), - (namespaces["svg"], "title") -]) - -mathmlTextIntegrationPointElements = frozenset([ - (namespaces["mathml"], "mi"), - (namespaces["mathml"], "mo"), - (namespaces["mathml"], "mn"), - (namespaces["mathml"], "ms"), - (namespaces["mathml"], "mtext") -]) - -adjustSVGAttributes = { - "attributename": "attributeName", - "attributetype": "attributeType", - "basefrequency": "baseFrequency", - "baseprofile": "baseProfile", - "calcmode": "calcMode", - "clippathunits": "clipPathUnits", - "contentscripttype": "contentScriptType", - "contentstyletype": "contentStyleType", - "diffuseconstant": "diffuseConstant", - "edgemode": "edgeMode", - "externalresourcesrequired": "externalResourcesRequired", - "filterres": "filterRes", - "filterunits": "filterUnits", - "glyphref": "glyphRef", - "gradienttransform": "gradientTransform", - "gradientunits": "gradientUnits", - "kernelmatrix": "kernelMatrix", - "kernelunitlength": "kernelUnitLength", - "keypoints": "keyPoints", - "keysplines": "keySplines", - "keytimes": "keyTimes", - "lengthadjust": "lengthAdjust", - "limitingconeangle": "limitingConeAngle", - "markerheight": "markerHeight", - "markerunits": "markerUnits", - "markerwidth": "markerWidth", - "maskcontentunits": "maskContentUnits", - "maskunits": "maskUnits", - "numoctaves": "numOctaves", - "pathlength": "pathLength", - "patterncontentunits": "patternContentUnits", - "patterntransform": "patternTransform", - "patternunits": "patternUnits", - "pointsatx": "pointsAtX", - "pointsaty": "pointsAtY", - "pointsatz": "pointsAtZ", - "preservealpha": "preserveAlpha", - "preserveaspectratio": "preserveAspectRatio", - "primitiveunits": "primitiveUnits", - "refx": "refX", - "refy": "refY", - "repeatcount": "repeatCount", - "repeatdur": "repeatDur", - "requiredextensions": "requiredExtensions", - "requiredfeatures": "requiredFeatures", - "specularconstant": "specularConstant", - "specularexponent": "specularExponent", - "spreadmethod": "spreadMethod", - "startoffset": "startOffset", - "stddeviation": "stdDeviation", - "stitchtiles": "stitchTiles", - "surfacescale": "surfaceScale", - "systemlanguage": "systemLanguage", - "tablevalues": "tableValues", - "targetx": "targetX", - "targety": "targetY", - "textlength": "textLength", - "viewbox": "viewBox", - "viewtarget": "viewTarget", - "xchannelselector": "xChannelSelector", - "ychannelselector": "yChannelSelector", - "zoomandpan": "zoomAndPan" -} - -adjustMathMLAttributes = {"definitionurl": "definitionURL"} - -adjustForeignAttributes = { - "xlink:actuate": ("xlink", "actuate", namespaces["xlink"]), - "xlink:arcrole": ("xlink", "arcrole", namespaces["xlink"]), - "xlink:href": ("xlink", "href", namespaces["xlink"]), - "xlink:role": ("xlink", "role", namespaces["xlink"]), - "xlink:show": ("xlink", "show", namespaces["xlink"]), - "xlink:title": ("xlink", "title", namespaces["xlink"]), - "xlink:type": ("xlink", "type", namespaces["xlink"]), - "xml:base": ("xml", "base", namespaces["xml"]), - "xml:lang": ("xml", "lang", namespaces["xml"]), - "xml:space": ("xml", "space", namespaces["xml"]), - "xmlns": (None, "xmlns", namespaces["xmlns"]), - "xmlns:xlink": ("xmlns", "xlink", namespaces["xmlns"]) -} - -unadjustForeignAttributes = dict([((ns, local), qname) for qname, (prefix, local, ns) in - adjustForeignAttributes.items()]) - -spaceCharacters = frozenset([ - "\t", - "\n", - "\u000C", - " ", - "\r" -]) - -tableInsertModeElements = frozenset([ - "table", - "tbody", - "tfoot", - "thead", - "tr" -]) - -asciiLowercase = frozenset(string.ascii_lowercase) -asciiUppercase = frozenset(string.ascii_uppercase) -asciiLetters = frozenset(string.ascii_letters) -digits = frozenset(string.digits) -hexDigits = frozenset(string.hexdigits) - -asciiUpper2Lower = dict([(ord(c), ord(c.lower())) - for c in string.ascii_uppercase]) - -# Heading elements need to be ordered -headingElements = ( - "h1", - "h2", - "h3", - "h4", - "h5", - "h6" -) - -voidElements = frozenset([ - "base", - "command", - "event-source", - "link", - "meta", - "hr", - "br", - "img", - "embed", - "param", - "area", - "col", - "input", - "source", - "track" -]) - -cdataElements = frozenset(['title', 'textarea']) - -rcdataElements = frozenset([ - 'style', - 'script', - 'xmp', - 'iframe', - 'noembed', - 'noframes', - 'noscript' -]) - -booleanAttributes = { - "": frozenset(["irrelevant", "itemscope"]), - "style": frozenset(["scoped"]), - "img": frozenset(["ismap"]), - "audio": frozenset(["autoplay", "controls"]), - "video": frozenset(["autoplay", "controls"]), - "script": frozenset(["defer", "async"]), - "details": frozenset(["open"]), - "datagrid": frozenset(["multiple", "disabled"]), - "command": frozenset(["hidden", "disabled", "checked", "default"]), - "hr": frozenset(["noshade"]), - "menu": frozenset(["autosubmit"]), - "fieldset": frozenset(["disabled", "readonly"]), - "option": frozenset(["disabled", "readonly", "selected"]), - "optgroup": frozenset(["disabled", "readonly"]), - "button": frozenset(["disabled", "autofocus"]), - "input": frozenset(["disabled", "readonly", "required", "autofocus", "checked", "ismap"]), - "select": frozenset(["disabled", "readonly", "autofocus", "multiple"]), - "output": frozenset(["disabled", "readonly"]), - "iframe": frozenset(["seamless"]), -} - -# entitiesWindows1252 has to be _ordered_ and needs to have an index. It -# therefore can't be a frozenset. -entitiesWindows1252 = ( - 8364, # 0x80 0x20AC EURO SIGN - 65533, # 0x81 UNDEFINED - 8218, # 0x82 0x201A SINGLE LOW-9 QUOTATION MARK - 402, # 0x83 0x0192 LATIN SMALL LETTER F WITH HOOK - 8222, # 0x84 0x201E DOUBLE LOW-9 QUOTATION MARK - 8230, # 0x85 0x2026 HORIZONTAL ELLIPSIS - 8224, # 0x86 0x2020 DAGGER - 8225, # 0x87 0x2021 DOUBLE DAGGER - 710, # 0x88 0x02C6 MODIFIER LETTER CIRCUMFLEX ACCENT - 8240, # 0x89 0x2030 PER MILLE SIGN - 352, # 0x8A 0x0160 LATIN CAPITAL LETTER S WITH CARON - 8249, # 0x8B 0x2039 SINGLE LEFT-POINTING ANGLE QUOTATION MARK - 338, # 0x8C 0x0152 LATIN CAPITAL LIGATURE OE - 65533, # 0x8D UNDEFINED - 381, # 0x8E 0x017D LATIN CAPITAL LETTER Z WITH CARON - 65533, # 0x8F UNDEFINED - 65533, # 0x90 UNDEFINED - 8216, # 0x91 0x2018 LEFT SINGLE QUOTATION MARK - 8217, # 0x92 0x2019 RIGHT SINGLE QUOTATION MARK - 8220, # 0x93 0x201C LEFT DOUBLE QUOTATION MARK - 8221, # 0x94 0x201D RIGHT DOUBLE QUOTATION MARK - 8226, # 0x95 0x2022 BULLET - 8211, # 0x96 0x2013 EN DASH - 8212, # 0x97 0x2014 EM DASH - 732, # 0x98 0x02DC SMALL TILDE - 8482, # 0x99 0x2122 TRADE MARK SIGN - 353, # 0x9A 0x0161 LATIN SMALL LETTER S WITH CARON - 8250, # 0x9B 0x203A SINGLE RIGHT-POINTING ANGLE QUOTATION MARK - 339, # 0x9C 0x0153 LATIN SMALL LIGATURE OE - 65533, # 0x9D UNDEFINED - 382, # 0x9E 0x017E LATIN SMALL LETTER Z WITH CARON - 376 # 0x9F 0x0178 LATIN CAPITAL LETTER Y WITH DIAERESIS -) - -xmlEntities = frozenset(['lt;', 'gt;', 'amp;', 'apos;', 'quot;']) - -entities = { - "AElig": "\xc6", - "AElig;": "\xc6", - "AMP": "&", - "AMP;": "&", - "Aacute": "\xc1", - "Aacute;": "\xc1", - "Abreve;": "\u0102", - "Acirc": "\xc2", - "Acirc;": "\xc2", - "Acy;": "\u0410", - "Afr;": "\U0001d504", - "Agrave": "\xc0", - "Agrave;": "\xc0", - "Alpha;": "\u0391", - "Amacr;": "\u0100", - "And;": "\u2a53", - "Aogon;": "\u0104", - "Aopf;": "\U0001d538", - "ApplyFunction;": "\u2061", - "Aring": "\xc5", - "Aring;": "\xc5", - "Ascr;": "\U0001d49c", - "Assign;": "\u2254", - "Atilde": "\xc3", - "Atilde;": "\xc3", - "Auml": "\xc4", - "Auml;": "\xc4", - "Backslash;": "\u2216", - "Barv;": "\u2ae7", - "Barwed;": "\u2306", - "Bcy;": "\u0411", - "Because;": "\u2235", - "Bernoullis;": "\u212c", - "Beta;": "\u0392", - "Bfr;": "\U0001d505", - "Bopf;": "\U0001d539", - "Breve;": "\u02d8", - "Bscr;": "\u212c", - "Bumpeq;": "\u224e", - "CHcy;": "\u0427", - "COPY": "\xa9", - "COPY;": "\xa9", - "Cacute;": "\u0106", - "Cap;": "\u22d2", - "CapitalDifferentialD;": "\u2145", - "Cayleys;": "\u212d", - "Ccaron;": "\u010c", - "Ccedil": "\xc7", - "Ccedil;": "\xc7", - "Ccirc;": "\u0108", - "Cconint;": "\u2230", - "Cdot;": "\u010a", - "Cedilla;": "\xb8", - "CenterDot;": "\xb7", - "Cfr;": "\u212d", - "Chi;": "\u03a7", - "CircleDot;": "\u2299", - "CircleMinus;": "\u2296", - "CirclePlus;": "\u2295", - "CircleTimes;": "\u2297", - "ClockwiseContourIntegral;": "\u2232", - "CloseCurlyDoubleQuote;": "\u201d", - "CloseCurlyQuote;": "\u2019", - "Colon;": "\u2237", - "Colone;": "\u2a74", - "Congruent;": "\u2261", - "Conint;": "\u222f", - "ContourIntegral;": "\u222e", - "Copf;": "\u2102", - "Coproduct;": "\u2210", - "CounterClockwiseContourIntegral;": "\u2233", - "Cross;": "\u2a2f", - "Cscr;": "\U0001d49e", - "Cup;": "\u22d3", - "CupCap;": "\u224d", - "DD;": "\u2145", - "DDotrahd;": "\u2911", - "DJcy;": "\u0402", - "DScy;": "\u0405", - "DZcy;": "\u040f", - "Dagger;": "\u2021", - "Darr;": "\u21a1", - "Dashv;": "\u2ae4", - "Dcaron;": "\u010e", - "Dcy;": "\u0414", - "Del;": "\u2207", - "Delta;": "\u0394", - "Dfr;": "\U0001d507", - "DiacriticalAcute;": "\xb4", - "DiacriticalDot;": "\u02d9", - "DiacriticalDoubleAcute;": "\u02dd", - "DiacriticalGrave;": "`", - "DiacriticalTilde;": "\u02dc", - "Diamond;": "\u22c4", - "DifferentialD;": "\u2146", - "Dopf;": "\U0001d53b", - "Dot;": "\xa8", - "DotDot;": "\u20dc", - "DotEqual;": "\u2250", - "DoubleContourIntegral;": "\u222f", - "DoubleDot;": "\xa8", - "DoubleDownArrow;": "\u21d3", - "DoubleLeftArrow;": "\u21d0", - "DoubleLeftRightArrow;": "\u21d4", - "DoubleLeftTee;": "\u2ae4", - "DoubleLongLeftArrow;": "\u27f8", - "DoubleLongLeftRightArrow;": "\u27fa", - "DoubleLongRightArrow;": "\u27f9", - "DoubleRightArrow;": "\u21d2", - "DoubleRightTee;": "\u22a8", - "DoubleUpArrow;": "\u21d1", - "DoubleUpDownArrow;": "\u21d5", - "DoubleVerticalBar;": "\u2225", - "DownArrow;": "\u2193", - "DownArrowBar;": "\u2913", - "DownArrowUpArrow;": "\u21f5", - "DownBreve;": "\u0311", - "DownLeftRightVector;": "\u2950", - "DownLeftTeeVector;": "\u295e", - "DownLeftVector;": "\u21bd", - "DownLeftVectorBar;": "\u2956", - "DownRightTeeVector;": "\u295f", - "DownRightVector;": "\u21c1", - "DownRightVectorBar;": "\u2957", - "DownTee;": "\u22a4", - "DownTeeArrow;": "\u21a7", - "Downarrow;": "\u21d3", - "Dscr;": "\U0001d49f", - "Dstrok;": "\u0110", - "ENG;": "\u014a", - "ETH": "\xd0", - "ETH;": "\xd0", - "Eacute": "\xc9", - "Eacute;": "\xc9", - "Ecaron;": "\u011a", - "Ecirc": "\xca", - "Ecirc;": "\xca", - "Ecy;": "\u042d", - "Edot;": "\u0116", - "Efr;": "\U0001d508", - "Egrave": "\xc8", - "Egrave;": "\xc8", - "Element;": "\u2208", - "Emacr;": "\u0112", - "EmptySmallSquare;": "\u25fb", - "EmptyVerySmallSquare;": "\u25ab", - "Eogon;": "\u0118", - "Eopf;": "\U0001d53c", - "Epsilon;": "\u0395", - "Equal;": "\u2a75", - "EqualTilde;": "\u2242", - "Equilibrium;": "\u21cc", - "Escr;": "\u2130", - "Esim;": "\u2a73", - "Eta;": "\u0397", - "Euml": "\xcb", - "Euml;": "\xcb", - "Exists;": "\u2203", - "ExponentialE;": "\u2147", - "Fcy;": "\u0424", - "Ffr;": "\U0001d509", - "FilledSmallSquare;": "\u25fc", - "FilledVerySmallSquare;": "\u25aa", - "Fopf;": "\U0001d53d", - "ForAll;": "\u2200", - "Fouriertrf;": "\u2131", - "Fscr;": "\u2131", - "GJcy;": "\u0403", - "GT": ">", - "GT;": ">", - "Gamma;": "\u0393", - "Gammad;": "\u03dc", - "Gbreve;": "\u011e", - "Gcedil;": "\u0122", - "Gcirc;": "\u011c", - "Gcy;": "\u0413", - "Gdot;": "\u0120", - "Gfr;": "\U0001d50a", - "Gg;": "\u22d9", - "Gopf;": "\U0001d53e", - "GreaterEqual;": "\u2265", - "GreaterEqualLess;": "\u22db", - "GreaterFullEqual;": "\u2267", - "GreaterGreater;": "\u2aa2", - "GreaterLess;": "\u2277", - "GreaterSlantEqual;": "\u2a7e", - "GreaterTilde;": "\u2273", - "Gscr;": "\U0001d4a2", - "Gt;": "\u226b", - "HARDcy;": "\u042a", - "Hacek;": "\u02c7", - "Hat;": "^", - "Hcirc;": "\u0124", - "Hfr;": "\u210c", - "HilbertSpace;": "\u210b", - "Hopf;": "\u210d", - "HorizontalLine;": "\u2500", - "Hscr;": "\u210b", - "Hstrok;": "\u0126", - "HumpDownHump;": "\u224e", - "HumpEqual;": "\u224f", - "IEcy;": "\u0415", - "IJlig;": "\u0132", - "IOcy;": "\u0401", - "Iacute": "\xcd", - "Iacute;": "\xcd", - "Icirc": "\xce", - "Icirc;": "\xce", - "Icy;": "\u0418", - "Idot;": "\u0130", - "Ifr;": "\u2111", - "Igrave": "\xcc", - "Igrave;": "\xcc", - "Im;": "\u2111", - "Imacr;": "\u012a", - "ImaginaryI;": "\u2148", - "Implies;": "\u21d2", - "Int;": "\u222c", - "Integral;": "\u222b", - "Intersection;": "\u22c2", - "InvisibleComma;": "\u2063", - "InvisibleTimes;": "\u2062", - "Iogon;": "\u012e", - "Iopf;": "\U0001d540", - "Iota;": "\u0399", - "Iscr;": "\u2110", - "Itilde;": "\u0128", - "Iukcy;": "\u0406", - "Iuml": "\xcf", - "Iuml;": "\xcf", - "Jcirc;": "\u0134", - "Jcy;": "\u0419", - "Jfr;": "\U0001d50d", - "Jopf;": "\U0001d541", - "Jscr;": "\U0001d4a5", - "Jsercy;": "\u0408", - "Jukcy;": "\u0404", - "KHcy;": "\u0425", - "KJcy;": "\u040c", - "Kappa;": "\u039a", - "Kcedil;": "\u0136", - "Kcy;": "\u041a", - "Kfr;": "\U0001d50e", - "Kopf;": "\U0001d542", - "Kscr;": "\U0001d4a6", - "LJcy;": "\u0409", - "LT": "<", - "LT;": "<", - "Lacute;": "\u0139", - "Lambda;": "\u039b", - "Lang;": "\u27ea", - "Laplacetrf;": "\u2112", - "Larr;": "\u219e", - "Lcaron;": "\u013d", - "Lcedil;": "\u013b", - "Lcy;": "\u041b", - "LeftAngleBracket;": "\u27e8", - "LeftArrow;": "\u2190", - "LeftArrowBar;": "\u21e4", - "LeftArrowRightArrow;": "\u21c6", - "LeftCeiling;": "\u2308", - "LeftDoubleBracket;": "\u27e6", - "LeftDownTeeVector;": "\u2961", - "LeftDownVector;": "\u21c3", - "LeftDownVectorBar;": "\u2959", - "LeftFloor;": "\u230a", - "LeftRightArrow;": "\u2194", - "LeftRightVector;": "\u294e", - "LeftTee;": "\u22a3", - "LeftTeeArrow;": "\u21a4", - "LeftTeeVector;": "\u295a", - "LeftTriangle;": "\u22b2", - "LeftTriangleBar;": "\u29cf", - "LeftTriangleEqual;": "\u22b4", - "LeftUpDownVector;": "\u2951", - "LeftUpTeeVector;": "\u2960", - "LeftUpVector;": "\u21bf", - "LeftUpVectorBar;": "\u2958", - "LeftVector;": "\u21bc", - "LeftVectorBar;": "\u2952", - "Leftarrow;": "\u21d0", - "Leftrightarrow;": "\u21d4", - "LessEqualGreater;": "\u22da", - "LessFullEqual;": "\u2266", - "LessGreater;": "\u2276", - "LessLess;": "\u2aa1", - "LessSlantEqual;": "\u2a7d", - "LessTilde;": "\u2272", - "Lfr;": "\U0001d50f", - "Ll;": "\u22d8", - "Lleftarrow;": "\u21da", - "Lmidot;": "\u013f", - "LongLeftArrow;": "\u27f5", - "LongLeftRightArrow;": "\u27f7", - "LongRightArrow;": "\u27f6", - "Longleftarrow;": "\u27f8", - "Longleftrightarrow;": "\u27fa", - "Longrightarrow;": "\u27f9", - "Lopf;": "\U0001d543", - "LowerLeftArrow;": "\u2199", - "LowerRightArrow;": "\u2198", - "Lscr;": "\u2112", - "Lsh;": "\u21b0", - "Lstrok;": "\u0141", - "Lt;": "\u226a", - "Map;": "\u2905", - "Mcy;": "\u041c", - "MediumSpace;": "\u205f", - "Mellintrf;": "\u2133", - "Mfr;": "\U0001d510", - "MinusPlus;": "\u2213", - "Mopf;": "\U0001d544", - "Mscr;": "\u2133", - "Mu;": "\u039c", - "NJcy;": "\u040a", - "Nacute;": "\u0143", - "Ncaron;": "\u0147", - "Ncedil;": "\u0145", - "Ncy;": "\u041d", - "NegativeMediumSpace;": "\u200b", - "NegativeThickSpace;": "\u200b", - "NegativeThinSpace;": "\u200b", - "NegativeVeryThinSpace;": "\u200b", - "NestedGreaterGreater;": "\u226b", - "NestedLessLess;": "\u226a", - "NewLine;": "\n", - "Nfr;": "\U0001d511", - "NoBreak;": "\u2060", - "NonBreakingSpace;": "\xa0", - "Nopf;": "\u2115", - "Not;": "\u2aec", - "NotCongruent;": "\u2262", - "NotCupCap;": "\u226d", - "NotDoubleVerticalBar;": "\u2226", - "NotElement;": "\u2209", - "NotEqual;": "\u2260", - "NotEqualTilde;": "\u2242\u0338", - "NotExists;": "\u2204", - "NotGreater;": "\u226f", - "NotGreaterEqual;": "\u2271", - "NotGreaterFullEqual;": "\u2267\u0338", - "NotGreaterGreater;": "\u226b\u0338", - "NotGreaterLess;": "\u2279", - "NotGreaterSlantEqual;": "\u2a7e\u0338", - "NotGreaterTilde;": "\u2275", - "NotHumpDownHump;": "\u224e\u0338", - "NotHumpEqual;": "\u224f\u0338", - "NotLeftTriangle;": "\u22ea", - "NotLeftTriangleBar;": "\u29cf\u0338", - "NotLeftTriangleEqual;": "\u22ec", - "NotLess;": "\u226e", - "NotLessEqual;": "\u2270", - "NotLessGreater;": "\u2278", - "NotLessLess;": "\u226a\u0338", - "NotLessSlantEqual;": "\u2a7d\u0338", - "NotLessTilde;": "\u2274", - "NotNestedGreaterGreater;": "\u2aa2\u0338", - "NotNestedLessLess;": "\u2aa1\u0338", - "NotPrecedes;": "\u2280", - "NotPrecedesEqual;": "\u2aaf\u0338", - "NotPrecedesSlantEqual;": "\u22e0", - "NotReverseElement;": "\u220c", - "NotRightTriangle;": "\u22eb", - "NotRightTriangleBar;": "\u29d0\u0338", - "NotRightTriangleEqual;": "\u22ed", - "NotSquareSubset;": "\u228f\u0338", - "NotSquareSubsetEqual;": "\u22e2", - "NotSquareSuperset;": "\u2290\u0338", - "NotSquareSupersetEqual;": "\u22e3", - "NotSubset;": "\u2282\u20d2", - "NotSubsetEqual;": "\u2288", - "NotSucceeds;": "\u2281", - "NotSucceedsEqual;": "\u2ab0\u0338", - "NotSucceedsSlantEqual;": "\u22e1", - "NotSucceedsTilde;": "\u227f\u0338", - "NotSuperset;": "\u2283\u20d2", - "NotSupersetEqual;": "\u2289", - "NotTilde;": "\u2241", - "NotTildeEqual;": "\u2244", - "NotTildeFullEqual;": "\u2247", - "NotTildeTilde;": "\u2249", - "NotVerticalBar;": "\u2224", - "Nscr;": "\U0001d4a9", - "Ntilde": "\xd1", - "Ntilde;": "\xd1", - "Nu;": "\u039d", - "OElig;": "\u0152", - "Oacute": "\xd3", - "Oacute;": "\xd3", - "Ocirc": "\xd4", - "Ocirc;": "\xd4", - "Ocy;": "\u041e", - "Odblac;": "\u0150", - "Ofr;": "\U0001d512", - "Ograve": "\xd2", - "Ograve;": "\xd2", - "Omacr;": "\u014c", - "Omega;": "\u03a9", - "Omicron;": "\u039f", - "Oopf;": "\U0001d546", - "OpenCurlyDoubleQuote;": "\u201c", - "OpenCurlyQuote;": "\u2018", - "Or;": "\u2a54", - "Oscr;": "\U0001d4aa", - "Oslash": "\xd8", - "Oslash;": "\xd8", - "Otilde": "\xd5", - "Otilde;": "\xd5", - "Otimes;": "\u2a37", - "Ouml": "\xd6", - "Ouml;": "\xd6", - "OverBar;": "\u203e", - "OverBrace;": "\u23de", - "OverBracket;": "\u23b4", - "OverParenthesis;": "\u23dc", - "PartialD;": "\u2202", - "Pcy;": "\u041f", - "Pfr;": "\U0001d513", - "Phi;": "\u03a6", - "Pi;": "\u03a0", - "PlusMinus;": "\xb1", - "Poincareplane;": "\u210c", - "Popf;": "\u2119", - "Pr;": "\u2abb", - "Precedes;": "\u227a", - "PrecedesEqual;": "\u2aaf", - "PrecedesSlantEqual;": "\u227c", - "PrecedesTilde;": "\u227e", - "Prime;": "\u2033", - "Product;": "\u220f", - "Proportion;": "\u2237", - "Proportional;": "\u221d", - "Pscr;": "\U0001d4ab", - "Psi;": "\u03a8", - "QUOT": "\"", - "QUOT;": "\"", - "Qfr;": "\U0001d514", - "Qopf;": "\u211a", - "Qscr;": "\U0001d4ac", - "RBarr;": "\u2910", - "REG": "\xae", - "REG;": "\xae", - "Racute;": "\u0154", - "Rang;": "\u27eb", - "Rarr;": "\u21a0", - "Rarrtl;": "\u2916", - "Rcaron;": "\u0158", - "Rcedil;": "\u0156", - "Rcy;": "\u0420", - "Re;": "\u211c", - "ReverseElement;": "\u220b", - "ReverseEquilibrium;": "\u21cb", - "ReverseUpEquilibrium;": "\u296f", - "Rfr;": "\u211c", - "Rho;": "\u03a1", - "RightAngleBracket;": "\u27e9", - "RightArrow;": "\u2192", - "RightArrowBar;": "\u21e5", - "RightArrowLeftArrow;": "\u21c4", - "RightCeiling;": "\u2309", - "RightDoubleBracket;": "\u27e7", - "RightDownTeeVector;": "\u295d", - "RightDownVector;": "\u21c2", - "RightDownVectorBar;": "\u2955", - "RightFloor;": "\u230b", - "RightTee;": "\u22a2", - "RightTeeArrow;": "\u21a6", - "RightTeeVector;": "\u295b", - "RightTriangle;": "\u22b3", - "RightTriangleBar;": "\u29d0", - "RightTriangleEqual;": "\u22b5", - "RightUpDownVector;": "\u294f", - "RightUpTeeVector;": "\u295c", - "RightUpVector;": "\u21be", - "RightUpVectorBar;": "\u2954", - "RightVector;": "\u21c0", - "RightVectorBar;": "\u2953", - "Rightarrow;": "\u21d2", - "Ropf;": "\u211d", - "RoundImplies;": "\u2970", - "Rrightarrow;": "\u21db", - "Rscr;": "\u211b", - "Rsh;": "\u21b1", - "RuleDelayed;": "\u29f4", - "SHCHcy;": "\u0429", - "SHcy;": "\u0428", - "SOFTcy;": "\u042c", - "Sacute;": "\u015a", - "Sc;": "\u2abc", - "Scaron;": "\u0160", - "Scedil;": "\u015e", - "Scirc;": "\u015c", - "Scy;": "\u0421", - "Sfr;": "\U0001d516", - "ShortDownArrow;": "\u2193", - "ShortLeftArrow;": "\u2190", - "ShortRightArrow;": "\u2192", - "ShortUpArrow;": "\u2191", - "Sigma;": "\u03a3", - "SmallCircle;": "\u2218", - "Sopf;": "\U0001d54a", - "Sqrt;": "\u221a", - "Square;": "\u25a1", - "SquareIntersection;": "\u2293", - "SquareSubset;": "\u228f", - "SquareSubsetEqual;": "\u2291", - "SquareSuperset;": "\u2290", - "SquareSupersetEqual;": "\u2292", - "SquareUnion;": "\u2294", - "Sscr;": "\U0001d4ae", - "Star;": "\u22c6", - "Sub;": "\u22d0", - "Subset;": "\u22d0", - "SubsetEqual;": "\u2286", - "Succeeds;": "\u227b", - "SucceedsEqual;": "\u2ab0", - "SucceedsSlantEqual;": "\u227d", - "SucceedsTilde;": "\u227f", - "SuchThat;": "\u220b", - "Sum;": "\u2211", - "Sup;": "\u22d1", - "Superset;": "\u2283", - "SupersetEqual;": "\u2287", - "Supset;": "\u22d1", - "THORN": "\xde", - "THORN;": "\xde", - "TRADE;": "\u2122", - "TSHcy;": "\u040b", - "TScy;": "\u0426", - "Tab;": "\t", - "Tau;": "\u03a4", - "Tcaron;": "\u0164", - "Tcedil;": "\u0162", - "Tcy;": "\u0422", - "Tfr;": "\U0001d517", - "Therefore;": "\u2234", - "Theta;": "\u0398", - "ThickSpace;": "\u205f\u200a", - "ThinSpace;": "\u2009", - "Tilde;": "\u223c", - "TildeEqual;": "\u2243", - "TildeFullEqual;": "\u2245", - "TildeTilde;": "\u2248", - "Topf;": "\U0001d54b", - "TripleDot;": "\u20db", - "Tscr;": "\U0001d4af", - "Tstrok;": "\u0166", - "Uacute": "\xda", - "Uacute;": "\xda", - "Uarr;": "\u219f", - "Uarrocir;": "\u2949", - "Ubrcy;": "\u040e", - "Ubreve;": "\u016c", - "Ucirc": "\xdb", - "Ucirc;": "\xdb", - "Ucy;": "\u0423", - "Udblac;": "\u0170", - "Ufr;": "\U0001d518", - "Ugrave": "\xd9", - "Ugrave;": "\xd9", - "Umacr;": "\u016a", - "UnderBar;": "_", - "UnderBrace;": "\u23df", - "UnderBracket;": "\u23b5", - "UnderParenthesis;": "\u23dd", - "Union;": "\u22c3", - "UnionPlus;": "\u228e", - "Uogon;": "\u0172", - "Uopf;": "\U0001d54c", - "UpArrow;": "\u2191", - "UpArrowBar;": "\u2912", - "UpArrowDownArrow;": "\u21c5", - "UpDownArrow;": "\u2195", - "UpEquilibrium;": "\u296e", - "UpTee;": "\u22a5", - "UpTeeArrow;": "\u21a5", - "Uparrow;": "\u21d1", - "Updownarrow;": "\u21d5", - "UpperLeftArrow;": "\u2196", - "UpperRightArrow;": "\u2197", - "Upsi;": "\u03d2", - "Upsilon;": "\u03a5", - "Uring;": "\u016e", - "Uscr;": "\U0001d4b0", - "Utilde;": "\u0168", - "Uuml": "\xdc", - "Uuml;": "\xdc", - "VDash;": "\u22ab", - "Vbar;": "\u2aeb", - "Vcy;": "\u0412", - "Vdash;": "\u22a9", - "Vdashl;": "\u2ae6", - "Vee;": "\u22c1", - "Verbar;": "\u2016", - "Vert;": "\u2016", - "VerticalBar;": "\u2223", - "VerticalLine;": "|", - "VerticalSeparator;": "\u2758", - "VerticalTilde;": "\u2240", - "VeryThinSpace;": "\u200a", - "Vfr;": "\U0001d519", - "Vopf;": "\U0001d54d", - "Vscr;": "\U0001d4b1", - "Vvdash;": "\u22aa", - "Wcirc;": "\u0174", - "Wedge;": "\u22c0", - "Wfr;": "\U0001d51a", - "Wopf;": "\U0001d54e", - "Wscr;": "\U0001d4b2", - "Xfr;": "\U0001d51b", - "Xi;": "\u039e", - "Xopf;": "\U0001d54f", - "Xscr;": "\U0001d4b3", - "YAcy;": "\u042f", - "YIcy;": "\u0407", - "YUcy;": "\u042e", - "Yacute": "\xdd", - "Yacute;": "\xdd", - "Ycirc;": "\u0176", - "Ycy;": "\u042b", - "Yfr;": "\U0001d51c", - "Yopf;": "\U0001d550", - "Yscr;": "\U0001d4b4", - "Yuml;": "\u0178", - "ZHcy;": "\u0416", - "Zacute;": "\u0179", - "Zcaron;": "\u017d", - "Zcy;": "\u0417", - "Zdot;": "\u017b", - "ZeroWidthSpace;": "\u200b", - "Zeta;": "\u0396", - "Zfr;": "\u2128", - "Zopf;": "\u2124", - "Zscr;": "\U0001d4b5", - "aacute": "\xe1", - "aacute;": "\xe1", - "abreve;": "\u0103", - "ac;": "\u223e", - "acE;": "\u223e\u0333", - "acd;": "\u223f", - "acirc": "\xe2", - "acirc;": "\xe2", - "acute": "\xb4", - "acute;": "\xb4", - "acy;": "\u0430", - "aelig": "\xe6", - "aelig;": "\xe6", - "af;": "\u2061", - "afr;": "\U0001d51e", - "agrave": "\xe0", - "agrave;": "\xe0", - "alefsym;": "\u2135", - "aleph;": "\u2135", - "alpha;": "\u03b1", - "amacr;": "\u0101", - "amalg;": "\u2a3f", - "amp": "&", - "amp;": "&", - "and;": "\u2227", - "andand;": "\u2a55", - "andd;": "\u2a5c", - "andslope;": "\u2a58", - "andv;": "\u2a5a", - "ang;": "\u2220", - "ange;": "\u29a4", - "angle;": "\u2220", - "angmsd;": "\u2221", - "angmsdaa;": "\u29a8", - "angmsdab;": "\u29a9", - "angmsdac;": "\u29aa", - "angmsdad;": "\u29ab", - "angmsdae;": "\u29ac", - "angmsdaf;": "\u29ad", - "angmsdag;": "\u29ae", - "angmsdah;": "\u29af", - "angrt;": "\u221f", - "angrtvb;": "\u22be", - "angrtvbd;": "\u299d", - "angsph;": "\u2222", - "angst;": "\xc5", - "angzarr;": "\u237c", - "aogon;": "\u0105", - "aopf;": "\U0001d552", - "ap;": "\u2248", - "apE;": "\u2a70", - "apacir;": "\u2a6f", - "ape;": "\u224a", - "apid;": "\u224b", - "apos;": "'", - "approx;": "\u2248", - "approxeq;": "\u224a", - "aring": "\xe5", - "aring;": "\xe5", - "ascr;": "\U0001d4b6", - "ast;": "*", - "asymp;": "\u2248", - "asympeq;": "\u224d", - "atilde": "\xe3", - "atilde;": "\xe3", - "auml": "\xe4", - "auml;": "\xe4", - "awconint;": "\u2233", - "awint;": "\u2a11", - "bNot;": "\u2aed", - "backcong;": "\u224c", - "backepsilon;": "\u03f6", - "backprime;": "\u2035", - "backsim;": "\u223d", - "backsimeq;": "\u22cd", - "barvee;": "\u22bd", - "barwed;": "\u2305", - "barwedge;": "\u2305", - "bbrk;": "\u23b5", - "bbrktbrk;": "\u23b6", - "bcong;": "\u224c", - "bcy;": "\u0431", - "bdquo;": "\u201e", - "becaus;": "\u2235", - "because;": "\u2235", - "bemptyv;": "\u29b0", - "bepsi;": "\u03f6", - "bernou;": "\u212c", - "beta;": "\u03b2", - "beth;": "\u2136", - "between;": "\u226c", - "bfr;": "\U0001d51f", - "bigcap;": "\u22c2", - "bigcirc;": "\u25ef", - "bigcup;": "\u22c3", - "bigodot;": "\u2a00", - "bigoplus;": "\u2a01", - "bigotimes;": "\u2a02", - "bigsqcup;": "\u2a06", - "bigstar;": "\u2605", - "bigtriangledown;": "\u25bd", - "bigtriangleup;": "\u25b3", - "biguplus;": "\u2a04", - "bigvee;": "\u22c1", - "bigwedge;": "\u22c0", - "bkarow;": "\u290d", - "blacklozenge;": "\u29eb", - "blacksquare;": "\u25aa", - "blacktriangle;": "\u25b4", - "blacktriangledown;": "\u25be", - "blacktriangleleft;": "\u25c2", - "blacktriangleright;": "\u25b8", - "blank;": "\u2423", - "blk12;": "\u2592", - "blk14;": "\u2591", - "blk34;": "\u2593", - "block;": "\u2588", - "bne;": "=\u20e5", - "bnequiv;": "\u2261\u20e5", - "bnot;": "\u2310", - "bopf;": "\U0001d553", - "bot;": "\u22a5", - "bottom;": "\u22a5", - "bowtie;": "\u22c8", - "boxDL;": "\u2557", - "boxDR;": "\u2554", - "boxDl;": "\u2556", - "boxDr;": "\u2553", - "boxH;": "\u2550", - "boxHD;": "\u2566", - "boxHU;": "\u2569", - "boxHd;": "\u2564", - "boxHu;": "\u2567", - "boxUL;": "\u255d", - "boxUR;": "\u255a", - "boxUl;": "\u255c", - "boxUr;": "\u2559", - "boxV;": "\u2551", - "boxVH;": "\u256c", - "boxVL;": "\u2563", - "boxVR;": "\u2560", - "boxVh;": "\u256b", - "boxVl;": "\u2562", - "boxVr;": "\u255f", - "boxbox;": "\u29c9", - "boxdL;": "\u2555", - "boxdR;": "\u2552", - "boxdl;": "\u2510", - "boxdr;": "\u250c", - "boxh;": "\u2500", - "boxhD;": "\u2565", - "boxhU;": "\u2568", - "boxhd;": "\u252c", - "boxhu;": "\u2534", - "boxminus;": "\u229f", - "boxplus;": "\u229e", - "boxtimes;": "\u22a0", - "boxuL;": "\u255b", - "boxuR;": "\u2558", - "boxul;": "\u2518", - "boxur;": "\u2514", - "boxv;": "\u2502", - "boxvH;": "\u256a", - "boxvL;": "\u2561", - "boxvR;": "\u255e", - "boxvh;": "\u253c", - "boxvl;": "\u2524", - "boxvr;": "\u251c", - "bprime;": "\u2035", - "breve;": "\u02d8", - "brvbar": "\xa6", - "brvbar;": "\xa6", - "bscr;": "\U0001d4b7", - "bsemi;": "\u204f", - "bsim;": "\u223d", - "bsime;": "\u22cd", - "bsol;": "\\", - "bsolb;": "\u29c5", - "bsolhsub;": "\u27c8", - "bull;": "\u2022", - "bullet;": "\u2022", - "bump;": "\u224e", - "bumpE;": "\u2aae", - "bumpe;": "\u224f", - "bumpeq;": "\u224f", - "cacute;": "\u0107", - "cap;": "\u2229", - "capand;": "\u2a44", - "capbrcup;": "\u2a49", - "capcap;": "\u2a4b", - "capcup;": "\u2a47", - "capdot;": "\u2a40", - "caps;": "\u2229\ufe00", - "caret;": "\u2041", - "caron;": "\u02c7", - "ccaps;": "\u2a4d", - "ccaron;": "\u010d", - "ccedil": "\xe7", - "ccedil;": "\xe7", - "ccirc;": "\u0109", - "ccups;": "\u2a4c", - "ccupssm;": "\u2a50", - "cdot;": "\u010b", - "cedil": "\xb8", - "cedil;": "\xb8", - "cemptyv;": "\u29b2", - "cent": "\xa2", - "cent;": "\xa2", - "centerdot;": "\xb7", - "cfr;": "\U0001d520", - "chcy;": "\u0447", - "check;": "\u2713", - "checkmark;": "\u2713", - "chi;": "\u03c7", - "cir;": "\u25cb", - "cirE;": "\u29c3", - "circ;": "\u02c6", - "circeq;": "\u2257", - "circlearrowleft;": "\u21ba", - "circlearrowright;": "\u21bb", - "circledR;": "\xae", - "circledS;": "\u24c8", - "circledast;": "\u229b", - "circledcirc;": "\u229a", - "circleddash;": "\u229d", - "cire;": "\u2257", - "cirfnint;": "\u2a10", - "cirmid;": "\u2aef", - "cirscir;": "\u29c2", - "clubs;": "\u2663", - "clubsuit;": "\u2663", - "colon;": ":", - "colone;": "\u2254", - "coloneq;": "\u2254", - "comma;": ",", - "commat;": "@", - "comp;": "\u2201", - "compfn;": "\u2218", - "complement;": "\u2201", - "complexes;": "\u2102", - "cong;": "\u2245", - "congdot;": "\u2a6d", - "conint;": "\u222e", - "copf;": "\U0001d554", - "coprod;": "\u2210", - "copy": "\xa9", - "copy;": "\xa9", - "copysr;": "\u2117", - "crarr;": "\u21b5", - "cross;": "\u2717", - "cscr;": "\U0001d4b8", - "csub;": "\u2acf", - "csube;": "\u2ad1", - "csup;": "\u2ad0", - "csupe;": "\u2ad2", - "ctdot;": "\u22ef", - "cudarrl;": "\u2938", - "cudarrr;": "\u2935", - "cuepr;": "\u22de", - "cuesc;": "\u22df", - "cularr;": "\u21b6", - "cularrp;": "\u293d", - "cup;": "\u222a", - "cupbrcap;": "\u2a48", - "cupcap;": "\u2a46", - "cupcup;": "\u2a4a", - "cupdot;": "\u228d", - "cupor;": "\u2a45", - "cups;": "\u222a\ufe00", - "curarr;": "\u21b7", - "curarrm;": "\u293c", - "curlyeqprec;": "\u22de", - "curlyeqsucc;": "\u22df", - "curlyvee;": "\u22ce", - "curlywedge;": "\u22cf", - "curren": "\xa4", - "curren;": "\xa4", - "curvearrowleft;": "\u21b6", - "curvearrowright;": "\u21b7", - "cuvee;": "\u22ce", - "cuwed;": "\u22cf", - "cwconint;": "\u2232", - "cwint;": "\u2231", - "cylcty;": "\u232d", - "dArr;": "\u21d3", - "dHar;": "\u2965", - "dagger;": "\u2020", - "daleth;": "\u2138", - "darr;": "\u2193", - "dash;": "\u2010", - "dashv;": "\u22a3", - "dbkarow;": "\u290f", - "dblac;": "\u02dd", - "dcaron;": "\u010f", - "dcy;": "\u0434", - "dd;": "\u2146", - "ddagger;": "\u2021", - "ddarr;": "\u21ca", - "ddotseq;": "\u2a77", - "deg": "\xb0", - "deg;": "\xb0", - "delta;": "\u03b4", - "demptyv;": "\u29b1", - "dfisht;": "\u297f", - "dfr;": "\U0001d521", - "dharl;": "\u21c3", - "dharr;": "\u21c2", - "diam;": "\u22c4", - "diamond;": "\u22c4", - "diamondsuit;": "\u2666", - "diams;": "\u2666", - "die;": "\xa8", - "digamma;": "\u03dd", - "disin;": "\u22f2", - "div;": "\xf7", - "divide": "\xf7", - "divide;": "\xf7", - "divideontimes;": "\u22c7", - "divonx;": "\u22c7", - "djcy;": "\u0452", - "dlcorn;": "\u231e", - "dlcrop;": "\u230d", - "dollar;": "$", - "dopf;": "\U0001d555", - "dot;": "\u02d9", - "doteq;": "\u2250", - "doteqdot;": "\u2251", - "dotminus;": "\u2238", - "dotplus;": "\u2214", - "dotsquare;": "\u22a1", - "doublebarwedge;": "\u2306", - "downarrow;": "\u2193", - "downdownarrows;": "\u21ca", - "downharpoonleft;": "\u21c3", - "downharpoonright;": "\u21c2", - "drbkarow;": "\u2910", - "drcorn;": "\u231f", - "drcrop;": "\u230c", - "dscr;": "\U0001d4b9", - "dscy;": "\u0455", - "dsol;": "\u29f6", - "dstrok;": "\u0111", - "dtdot;": "\u22f1", - "dtri;": "\u25bf", - "dtrif;": "\u25be", - "duarr;": "\u21f5", - "duhar;": "\u296f", - "dwangle;": "\u29a6", - "dzcy;": "\u045f", - "dzigrarr;": "\u27ff", - "eDDot;": "\u2a77", - "eDot;": "\u2251", - "eacute": "\xe9", - "eacute;": "\xe9", - "easter;": "\u2a6e", - "ecaron;": "\u011b", - "ecir;": "\u2256", - "ecirc": "\xea", - "ecirc;": "\xea", - "ecolon;": "\u2255", - "ecy;": "\u044d", - "edot;": "\u0117", - "ee;": "\u2147", - "efDot;": "\u2252", - "efr;": "\U0001d522", - "eg;": "\u2a9a", - "egrave": "\xe8", - "egrave;": "\xe8", - "egs;": "\u2a96", - "egsdot;": "\u2a98", - "el;": "\u2a99", - "elinters;": "\u23e7", - "ell;": "\u2113", - "els;": "\u2a95", - "elsdot;": "\u2a97", - "emacr;": "\u0113", - "empty;": "\u2205", - "emptyset;": "\u2205", - "emptyv;": "\u2205", - "emsp13;": "\u2004", - "emsp14;": "\u2005", - "emsp;": "\u2003", - "eng;": "\u014b", - "ensp;": "\u2002", - "eogon;": "\u0119", - "eopf;": "\U0001d556", - "epar;": "\u22d5", - "eparsl;": "\u29e3", - "eplus;": "\u2a71", - "epsi;": "\u03b5", - "epsilon;": "\u03b5", - "epsiv;": "\u03f5", - "eqcirc;": "\u2256", - "eqcolon;": "\u2255", - "eqsim;": "\u2242", - "eqslantgtr;": "\u2a96", - "eqslantless;": "\u2a95", - "equals;": "=", - "equest;": "\u225f", - "equiv;": "\u2261", - "equivDD;": "\u2a78", - "eqvparsl;": "\u29e5", - "erDot;": "\u2253", - "erarr;": "\u2971", - "escr;": "\u212f", - "esdot;": "\u2250", - "esim;": "\u2242", - "eta;": "\u03b7", - "eth": "\xf0", - "eth;": "\xf0", - "euml": "\xeb", - "euml;": "\xeb", - "euro;": "\u20ac", - "excl;": "!", - "exist;": "\u2203", - "expectation;": "\u2130", - "exponentiale;": "\u2147", - "fallingdotseq;": "\u2252", - "fcy;": "\u0444", - "female;": "\u2640", - "ffilig;": "\ufb03", - "fflig;": "\ufb00", - "ffllig;": "\ufb04", - "ffr;": "\U0001d523", - "filig;": "\ufb01", - "fjlig;": "fj", - "flat;": "\u266d", - "fllig;": "\ufb02", - "fltns;": "\u25b1", - "fnof;": "\u0192", - "fopf;": "\U0001d557", - "forall;": "\u2200", - "fork;": "\u22d4", - "forkv;": "\u2ad9", - "fpartint;": "\u2a0d", - "frac12": "\xbd", - "frac12;": "\xbd", - "frac13;": "\u2153", - "frac14": "\xbc", - "frac14;": "\xbc", - "frac15;": "\u2155", - "frac16;": "\u2159", - "frac18;": "\u215b", - "frac23;": "\u2154", - "frac25;": "\u2156", - "frac34": "\xbe", - "frac34;": "\xbe", - "frac35;": "\u2157", - "frac38;": "\u215c", - "frac45;": "\u2158", - "frac56;": "\u215a", - "frac58;": "\u215d", - "frac78;": "\u215e", - "frasl;": "\u2044", - "frown;": "\u2322", - "fscr;": "\U0001d4bb", - "gE;": "\u2267", - "gEl;": "\u2a8c", - "gacute;": "\u01f5", - "gamma;": "\u03b3", - "gammad;": "\u03dd", - "gap;": "\u2a86", - "gbreve;": "\u011f", - "gcirc;": "\u011d", - "gcy;": "\u0433", - "gdot;": "\u0121", - "ge;": "\u2265", - "gel;": "\u22db", - "geq;": "\u2265", - "geqq;": "\u2267", - "geqslant;": "\u2a7e", - "ges;": "\u2a7e", - "gescc;": "\u2aa9", - "gesdot;": "\u2a80", - "gesdoto;": "\u2a82", - "gesdotol;": "\u2a84", - "gesl;": "\u22db\ufe00", - "gesles;": "\u2a94", - "gfr;": "\U0001d524", - "gg;": "\u226b", - "ggg;": "\u22d9", - "gimel;": "\u2137", - "gjcy;": "\u0453", - "gl;": "\u2277", - "glE;": "\u2a92", - "gla;": "\u2aa5", - "glj;": "\u2aa4", - "gnE;": "\u2269", - "gnap;": "\u2a8a", - "gnapprox;": "\u2a8a", - "gne;": "\u2a88", - "gneq;": "\u2a88", - "gneqq;": "\u2269", - "gnsim;": "\u22e7", - "gopf;": "\U0001d558", - "grave;": "`", - "gscr;": "\u210a", - "gsim;": "\u2273", - "gsime;": "\u2a8e", - "gsiml;": "\u2a90", - "gt": ">", - "gt;": ">", - "gtcc;": "\u2aa7", - "gtcir;": "\u2a7a", - "gtdot;": "\u22d7", - "gtlPar;": "\u2995", - "gtquest;": "\u2a7c", - "gtrapprox;": "\u2a86", - "gtrarr;": "\u2978", - "gtrdot;": "\u22d7", - "gtreqless;": "\u22db", - "gtreqqless;": "\u2a8c", - "gtrless;": "\u2277", - "gtrsim;": "\u2273", - "gvertneqq;": "\u2269\ufe00", - "gvnE;": "\u2269\ufe00", - "hArr;": "\u21d4", - "hairsp;": "\u200a", - "half;": "\xbd", - "hamilt;": "\u210b", - "hardcy;": "\u044a", - "harr;": "\u2194", - "harrcir;": "\u2948", - "harrw;": "\u21ad", - "hbar;": "\u210f", - "hcirc;": "\u0125", - "hearts;": "\u2665", - "heartsuit;": "\u2665", - "hellip;": "\u2026", - "hercon;": "\u22b9", - "hfr;": "\U0001d525", - "hksearow;": "\u2925", - "hkswarow;": "\u2926", - "hoarr;": "\u21ff", - "homtht;": "\u223b", - "hookleftarrow;": "\u21a9", - "hookrightarrow;": "\u21aa", - "hopf;": "\U0001d559", - "horbar;": "\u2015", - "hscr;": "\U0001d4bd", - "hslash;": "\u210f", - "hstrok;": "\u0127", - "hybull;": "\u2043", - "hyphen;": "\u2010", - "iacute": "\xed", - "iacute;": "\xed", - "ic;": "\u2063", - "icirc": "\xee", - "icirc;": "\xee", - "icy;": "\u0438", - "iecy;": "\u0435", - "iexcl": "\xa1", - "iexcl;": "\xa1", - "iff;": "\u21d4", - "ifr;": "\U0001d526", - "igrave": "\xec", - "igrave;": "\xec", - "ii;": "\u2148", - "iiiint;": "\u2a0c", - "iiint;": "\u222d", - "iinfin;": "\u29dc", - "iiota;": "\u2129", - "ijlig;": "\u0133", - "imacr;": "\u012b", - "image;": "\u2111", - "imagline;": "\u2110", - "imagpart;": "\u2111", - "imath;": "\u0131", - "imof;": "\u22b7", - "imped;": "\u01b5", - "in;": "\u2208", - "incare;": "\u2105", - "infin;": "\u221e", - "infintie;": "\u29dd", - "inodot;": "\u0131", - "int;": "\u222b", - "intcal;": "\u22ba", - "integers;": "\u2124", - "intercal;": "\u22ba", - "intlarhk;": "\u2a17", - "intprod;": "\u2a3c", - "iocy;": "\u0451", - "iogon;": "\u012f", - "iopf;": "\U0001d55a", - "iota;": "\u03b9", - "iprod;": "\u2a3c", - "iquest": "\xbf", - "iquest;": "\xbf", - "iscr;": "\U0001d4be", - "isin;": "\u2208", - "isinE;": "\u22f9", - "isindot;": "\u22f5", - "isins;": "\u22f4", - "isinsv;": "\u22f3", - "isinv;": "\u2208", - "it;": "\u2062", - "itilde;": "\u0129", - "iukcy;": "\u0456", - "iuml": "\xef", - "iuml;": "\xef", - "jcirc;": "\u0135", - "jcy;": "\u0439", - "jfr;": "\U0001d527", - "jmath;": "\u0237", - "jopf;": "\U0001d55b", - "jscr;": "\U0001d4bf", - "jsercy;": "\u0458", - "jukcy;": "\u0454", - "kappa;": "\u03ba", - "kappav;": "\u03f0", - "kcedil;": "\u0137", - "kcy;": "\u043a", - "kfr;": "\U0001d528", - "kgreen;": "\u0138", - "khcy;": "\u0445", - "kjcy;": "\u045c", - "kopf;": "\U0001d55c", - "kscr;": "\U0001d4c0", - "lAarr;": "\u21da", - "lArr;": "\u21d0", - "lAtail;": "\u291b", - "lBarr;": "\u290e", - "lE;": "\u2266", - "lEg;": "\u2a8b", - "lHar;": "\u2962", - "lacute;": "\u013a", - "laemptyv;": "\u29b4", - "lagran;": "\u2112", - "lambda;": "\u03bb", - "lang;": "\u27e8", - "langd;": "\u2991", - "langle;": "\u27e8", - "lap;": "\u2a85", - "laquo": "\xab", - "laquo;": "\xab", - "larr;": "\u2190", - "larrb;": "\u21e4", - "larrbfs;": "\u291f", - "larrfs;": "\u291d", - "larrhk;": "\u21a9", - "larrlp;": "\u21ab", - "larrpl;": "\u2939", - "larrsim;": "\u2973", - "larrtl;": "\u21a2", - "lat;": "\u2aab", - "latail;": "\u2919", - "late;": "\u2aad", - "lates;": "\u2aad\ufe00", - "lbarr;": "\u290c", - "lbbrk;": "\u2772", - "lbrace;": "{", - "lbrack;": "[", - "lbrke;": "\u298b", - "lbrksld;": "\u298f", - "lbrkslu;": "\u298d", - "lcaron;": "\u013e", - "lcedil;": "\u013c", - "lceil;": "\u2308", - "lcub;": "{", - "lcy;": "\u043b", - "ldca;": "\u2936", - "ldquo;": "\u201c", - "ldquor;": "\u201e", - "ldrdhar;": "\u2967", - "ldrushar;": "\u294b", - "ldsh;": "\u21b2", - "le;": "\u2264", - "leftarrow;": "\u2190", - "leftarrowtail;": "\u21a2", - "leftharpoondown;": "\u21bd", - "leftharpoonup;": "\u21bc", - "leftleftarrows;": "\u21c7", - "leftrightarrow;": "\u2194", - "leftrightarrows;": "\u21c6", - "leftrightharpoons;": "\u21cb", - "leftrightsquigarrow;": "\u21ad", - "leftthreetimes;": "\u22cb", - "leg;": "\u22da", - "leq;": "\u2264", - "leqq;": "\u2266", - "leqslant;": "\u2a7d", - "les;": "\u2a7d", - "lescc;": "\u2aa8", - "lesdot;": "\u2a7f", - "lesdoto;": "\u2a81", - "lesdotor;": "\u2a83", - "lesg;": "\u22da\ufe00", - "lesges;": "\u2a93", - "lessapprox;": "\u2a85", - "lessdot;": "\u22d6", - "lesseqgtr;": "\u22da", - "lesseqqgtr;": "\u2a8b", - "lessgtr;": "\u2276", - "lesssim;": "\u2272", - "lfisht;": "\u297c", - "lfloor;": "\u230a", - "lfr;": "\U0001d529", - "lg;": "\u2276", - "lgE;": "\u2a91", - "lhard;": "\u21bd", - "lharu;": "\u21bc", - "lharul;": "\u296a", - "lhblk;": "\u2584", - "ljcy;": "\u0459", - "ll;": "\u226a", - "llarr;": "\u21c7", - "llcorner;": "\u231e", - "llhard;": "\u296b", - "lltri;": "\u25fa", - "lmidot;": "\u0140", - "lmoust;": "\u23b0", - "lmoustache;": "\u23b0", - "lnE;": "\u2268", - "lnap;": "\u2a89", - "lnapprox;": "\u2a89", - "lne;": "\u2a87", - "lneq;": "\u2a87", - "lneqq;": "\u2268", - "lnsim;": "\u22e6", - "loang;": "\u27ec", - "loarr;": "\u21fd", - "lobrk;": "\u27e6", - "longleftarrow;": "\u27f5", - "longleftrightarrow;": "\u27f7", - "longmapsto;": "\u27fc", - "longrightarrow;": "\u27f6", - "looparrowleft;": "\u21ab", - "looparrowright;": "\u21ac", - "lopar;": "\u2985", - "lopf;": "\U0001d55d", - "loplus;": "\u2a2d", - "lotimes;": "\u2a34", - "lowast;": "\u2217", - "lowbar;": "_", - "loz;": "\u25ca", - "lozenge;": "\u25ca", - "lozf;": "\u29eb", - "lpar;": "(", - "lparlt;": "\u2993", - "lrarr;": "\u21c6", - "lrcorner;": "\u231f", - "lrhar;": "\u21cb", - "lrhard;": "\u296d", - "lrm;": "\u200e", - "lrtri;": "\u22bf", - "lsaquo;": "\u2039", - "lscr;": "\U0001d4c1", - "lsh;": "\u21b0", - "lsim;": "\u2272", - "lsime;": "\u2a8d", - "lsimg;": "\u2a8f", - "lsqb;": "[", - "lsquo;": "\u2018", - "lsquor;": "\u201a", - "lstrok;": "\u0142", - "lt": "<", - "lt;": "<", - "ltcc;": "\u2aa6", - "ltcir;": "\u2a79", - "ltdot;": "\u22d6", - "lthree;": "\u22cb", - "ltimes;": "\u22c9", - "ltlarr;": "\u2976", - "ltquest;": "\u2a7b", - "ltrPar;": "\u2996", - "ltri;": "\u25c3", - "ltrie;": "\u22b4", - "ltrif;": "\u25c2", - "lurdshar;": "\u294a", - "luruhar;": "\u2966", - "lvertneqq;": "\u2268\ufe00", - "lvnE;": "\u2268\ufe00", - "mDDot;": "\u223a", - "macr": "\xaf", - "macr;": "\xaf", - "male;": "\u2642", - "malt;": "\u2720", - "maltese;": "\u2720", - "map;": "\u21a6", - "mapsto;": "\u21a6", - "mapstodown;": "\u21a7", - "mapstoleft;": "\u21a4", - "mapstoup;": "\u21a5", - "marker;": "\u25ae", - "mcomma;": "\u2a29", - "mcy;": "\u043c", - "mdash;": "\u2014", - "measuredangle;": "\u2221", - "mfr;": "\U0001d52a", - "mho;": "\u2127", - "micro": "\xb5", - "micro;": "\xb5", - "mid;": "\u2223", - "midast;": "*", - "midcir;": "\u2af0", - "middot": "\xb7", - "middot;": "\xb7", - "minus;": "\u2212", - "minusb;": "\u229f", - "minusd;": "\u2238", - "minusdu;": "\u2a2a", - "mlcp;": "\u2adb", - "mldr;": "\u2026", - "mnplus;": "\u2213", - "models;": "\u22a7", - "mopf;": "\U0001d55e", - "mp;": "\u2213", - "mscr;": "\U0001d4c2", - "mstpos;": "\u223e", - "mu;": "\u03bc", - "multimap;": "\u22b8", - "mumap;": "\u22b8", - "nGg;": "\u22d9\u0338", - "nGt;": "\u226b\u20d2", - "nGtv;": "\u226b\u0338", - "nLeftarrow;": "\u21cd", - "nLeftrightarrow;": "\u21ce", - "nLl;": "\u22d8\u0338", - "nLt;": "\u226a\u20d2", - "nLtv;": "\u226a\u0338", - "nRightarrow;": "\u21cf", - "nVDash;": "\u22af", - "nVdash;": "\u22ae", - "nabla;": "\u2207", - "nacute;": "\u0144", - "nang;": "\u2220\u20d2", - "nap;": "\u2249", - "napE;": "\u2a70\u0338", - "napid;": "\u224b\u0338", - "napos;": "\u0149", - "napprox;": "\u2249", - "natur;": "\u266e", - "natural;": "\u266e", - "naturals;": "\u2115", - "nbsp": "\xa0", - "nbsp;": "\xa0", - "nbump;": "\u224e\u0338", - "nbumpe;": "\u224f\u0338", - "ncap;": "\u2a43", - "ncaron;": "\u0148", - "ncedil;": "\u0146", - "ncong;": "\u2247", - "ncongdot;": "\u2a6d\u0338", - "ncup;": "\u2a42", - "ncy;": "\u043d", - "ndash;": "\u2013", - "ne;": "\u2260", - "neArr;": "\u21d7", - "nearhk;": "\u2924", - "nearr;": "\u2197", - "nearrow;": "\u2197", - "nedot;": "\u2250\u0338", - "nequiv;": "\u2262", - "nesear;": "\u2928", - "nesim;": "\u2242\u0338", - "nexist;": "\u2204", - "nexists;": "\u2204", - "nfr;": "\U0001d52b", - "ngE;": "\u2267\u0338", - "nge;": "\u2271", - "ngeq;": "\u2271", - "ngeqq;": "\u2267\u0338", - "ngeqslant;": "\u2a7e\u0338", - "nges;": "\u2a7e\u0338", - "ngsim;": "\u2275", - "ngt;": "\u226f", - "ngtr;": "\u226f", - "nhArr;": "\u21ce", - "nharr;": "\u21ae", - "nhpar;": "\u2af2", - "ni;": "\u220b", - "nis;": "\u22fc", - "nisd;": "\u22fa", - "niv;": "\u220b", - "njcy;": "\u045a", - "nlArr;": "\u21cd", - "nlE;": "\u2266\u0338", - "nlarr;": "\u219a", - "nldr;": "\u2025", - "nle;": "\u2270", - "nleftarrow;": "\u219a", - "nleftrightarrow;": "\u21ae", - "nleq;": "\u2270", - "nleqq;": "\u2266\u0338", - "nleqslant;": "\u2a7d\u0338", - "nles;": "\u2a7d\u0338", - "nless;": "\u226e", - "nlsim;": "\u2274", - "nlt;": "\u226e", - "nltri;": "\u22ea", - "nltrie;": "\u22ec", - "nmid;": "\u2224", - "nopf;": "\U0001d55f", - "not": "\xac", - "not;": "\xac", - "notin;": "\u2209", - "notinE;": "\u22f9\u0338", - "notindot;": "\u22f5\u0338", - "notinva;": "\u2209", - "notinvb;": "\u22f7", - "notinvc;": "\u22f6", - "notni;": "\u220c", - "notniva;": "\u220c", - "notnivb;": "\u22fe", - "notnivc;": "\u22fd", - "npar;": "\u2226", - "nparallel;": "\u2226", - "nparsl;": "\u2afd\u20e5", - "npart;": "\u2202\u0338", - "npolint;": "\u2a14", - "npr;": "\u2280", - "nprcue;": "\u22e0", - "npre;": "\u2aaf\u0338", - "nprec;": "\u2280", - "npreceq;": "\u2aaf\u0338", - "nrArr;": "\u21cf", - "nrarr;": "\u219b", - "nrarrc;": "\u2933\u0338", - "nrarrw;": "\u219d\u0338", - "nrightarrow;": "\u219b", - "nrtri;": "\u22eb", - "nrtrie;": "\u22ed", - "nsc;": "\u2281", - "nsccue;": "\u22e1", - "nsce;": "\u2ab0\u0338", - "nscr;": "\U0001d4c3", - "nshortmid;": "\u2224", - "nshortparallel;": "\u2226", - "nsim;": "\u2241", - "nsime;": "\u2244", - "nsimeq;": "\u2244", - "nsmid;": "\u2224", - "nspar;": "\u2226", - "nsqsube;": "\u22e2", - "nsqsupe;": "\u22e3", - "nsub;": "\u2284", - "nsubE;": "\u2ac5\u0338", - "nsube;": "\u2288", - "nsubset;": "\u2282\u20d2", - "nsubseteq;": "\u2288", - "nsubseteqq;": "\u2ac5\u0338", - "nsucc;": "\u2281", - "nsucceq;": "\u2ab0\u0338", - "nsup;": "\u2285", - "nsupE;": "\u2ac6\u0338", - "nsupe;": "\u2289", - "nsupset;": "\u2283\u20d2", - "nsupseteq;": "\u2289", - "nsupseteqq;": "\u2ac6\u0338", - "ntgl;": "\u2279", - "ntilde": "\xf1", - "ntilde;": "\xf1", - "ntlg;": "\u2278", - "ntriangleleft;": "\u22ea", - "ntrianglelefteq;": "\u22ec", - "ntriangleright;": "\u22eb", - "ntrianglerighteq;": "\u22ed", - "nu;": "\u03bd", - "num;": "#", - "numero;": "\u2116", - "numsp;": "\u2007", - "nvDash;": "\u22ad", - "nvHarr;": "\u2904", - "nvap;": "\u224d\u20d2", - "nvdash;": "\u22ac", - "nvge;": "\u2265\u20d2", - "nvgt;": ">\u20d2", - "nvinfin;": "\u29de", - "nvlArr;": "\u2902", - "nvle;": "\u2264\u20d2", - "nvlt;": "<\u20d2", - "nvltrie;": "\u22b4\u20d2", - "nvrArr;": "\u2903", - "nvrtrie;": "\u22b5\u20d2", - "nvsim;": "\u223c\u20d2", - "nwArr;": "\u21d6", - "nwarhk;": "\u2923", - "nwarr;": "\u2196", - "nwarrow;": "\u2196", - "nwnear;": "\u2927", - "oS;": "\u24c8", - "oacute": "\xf3", - "oacute;": "\xf3", - "oast;": "\u229b", - "ocir;": "\u229a", - "ocirc": "\xf4", - "ocirc;": "\xf4", - "ocy;": "\u043e", - "odash;": "\u229d", - "odblac;": "\u0151", - "odiv;": "\u2a38", - "odot;": "\u2299", - "odsold;": "\u29bc", - "oelig;": "\u0153", - "ofcir;": "\u29bf", - "ofr;": "\U0001d52c", - "ogon;": "\u02db", - "ograve": "\xf2", - "ograve;": "\xf2", - "ogt;": "\u29c1", - "ohbar;": "\u29b5", - "ohm;": "\u03a9", - "oint;": "\u222e", - "olarr;": "\u21ba", - "olcir;": "\u29be", - "olcross;": "\u29bb", - "oline;": "\u203e", - "olt;": "\u29c0", - "omacr;": "\u014d", - "omega;": "\u03c9", - "omicron;": "\u03bf", - "omid;": "\u29b6", - "ominus;": "\u2296", - "oopf;": "\U0001d560", - "opar;": "\u29b7", - "operp;": "\u29b9", - "oplus;": "\u2295", - "or;": "\u2228", - "orarr;": "\u21bb", - "ord;": "\u2a5d", - "order;": "\u2134", - "orderof;": "\u2134", - "ordf": "\xaa", - "ordf;": "\xaa", - "ordm": "\xba", - "ordm;": "\xba", - "origof;": "\u22b6", - "oror;": "\u2a56", - "orslope;": "\u2a57", - "orv;": "\u2a5b", - "oscr;": "\u2134", - "oslash": "\xf8", - "oslash;": "\xf8", - "osol;": "\u2298", - "otilde": "\xf5", - "otilde;": "\xf5", - "otimes;": "\u2297", - "otimesas;": "\u2a36", - "ouml": "\xf6", - "ouml;": "\xf6", - "ovbar;": "\u233d", - "par;": "\u2225", - "para": "\xb6", - "para;": "\xb6", - "parallel;": "\u2225", - "parsim;": "\u2af3", - "parsl;": "\u2afd", - "part;": "\u2202", - "pcy;": "\u043f", - "percnt;": "%", - "period;": ".", - "permil;": "\u2030", - "perp;": "\u22a5", - "pertenk;": "\u2031", - "pfr;": "\U0001d52d", - "phi;": "\u03c6", - "phiv;": "\u03d5", - "phmmat;": "\u2133", - "phone;": "\u260e", - "pi;": "\u03c0", - "pitchfork;": "\u22d4", - "piv;": "\u03d6", - "planck;": "\u210f", - "planckh;": "\u210e", - "plankv;": "\u210f", - "plus;": "+", - "plusacir;": "\u2a23", - "plusb;": "\u229e", - "pluscir;": "\u2a22", - "plusdo;": "\u2214", - "plusdu;": "\u2a25", - "pluse;": "\u2a72", - "plusmn": "\xb1", - "plusmn;": "\xb1", - "plussim;": "\u2a26", - "plustwo;": "\u2a27", - "pm;": "\xb1", - "pointint;": "\u2a15", - "popf;": "\U0001d561", - "pound": "\xa3", - "pound;": "\xa3", - "pr;": "\u227a", - "prE;": "\u2ab3", - "prap;": "\u2ab7", - "prcue;": "\u227c", - "pre;": "\u2aaf", - "prec;": "\u227a", - "precapprox;": "\u2ab7", - "preccurlyeq;": "\u227c", - "preceq;": "\u2aaf", - "precnapprox;": "\u2ab9", - "precneqq;": "\u2ab5", - "precnsim;": "\u22e8", - "precsim;": "\u227e", - "prime;": "\u2032", - "primes;": "\u2119", - "prnE;": "\u2ab5", - "prnap;": "\u2ab9", - "prnsim;": "\u22e8", - "prod;": "\u220f", - "profalar;": "\u232e", - "profline;": "\u2312", - "profsurf;": "\u2313", - "prop;": "\u221d", - "propto;": "\u221d", - "prsim;": "\u227e", - "prurel;": "\u22b0", - "pscr;": "\U0001d4c5", - "psi;": "\u03c8", - "puncsp;": "\u2008", - "qfr;": "\U0001d52e", - "qint;": "\u2a0c", - "qopf;": "\U0001d562", - "qprime;": "\u2057", - "qscr;": "\U0001d4c6", - "quaternions;": "\u210d", - "quatint;": "\u2a16", - "quest;": "?", - "questeq;": "\u225f", - "quot": "\"", - "quot;": "\"", - "rAarr;": "\u21db", - "rArr;": "\u21d2", - "rAtail;": "\u291c", - "rBarr;": "\u290f", - "rHar;": "\u2964", - "race;": "\u223d\u0331", - "racute;": "\u0155", - "radic;": "\u221a", - "raemptyv;": "\u29b3", - "rang;": "\u27e9", - "rangd;": "\u2992", - "range;": "\u29a5", - "rangle;": "\u27e9", - "raquo": "\xbb", - "raquo;": "\xbb", - "rarr;": "\u2192", - "rarrap;": "\u2975", - "rarrb;": "\u21e5", - "rarrbfs;": "\u2920", - "rarrc;": "\u2933", - "rarrfs;": "\u291e", - "rarrhk;": "\u21aa", - "rarrlp;": "\u21ac", - "rarrpl;": "\u2945", - "rarrsim;": "\u2974", - "rarrtl;": "\u21a3", - "rarrw;": "\u219d", - "ratail;": "\u291a", - "ratio;": "\u2236", - "rationals;": "\u211a", - "rbarr;": "\u290d", - "rbbrk;": "\u2773", - "rbrace;": "}", - "rbrack;": "]", - "rbrke;": "\u298c", - "rbrksld;": "\u298e", - "rbrkslu;": "\u2990", - "rcaron;": "\u0159", - "rcedil;": "\u0157", - "rceil;": "\u2309", - "rcub;": "}", - "rcy;": "\u0440", - "rdca;": "\u2937", - "rdldhar;": "\u2969", - "rdquo;": "\u201d", - "rdquor;": "\u201d", - "rdsh;": "\u21b3", - "real;": "\u211c", - "realine;": "\u211b", - "realpart;": "\u211c", - "reals;": "\u211d", - "rect;": "\u25ad", - "reg": "\xae", - "reg;": "\xae", - "rfisht;": "\u297d", - "rfloor;": "\u230b", - "rfr;": "\U0001d52f", - "rhard;": "\u21c1", - "rharu;": "\u21c0", - "rharul;": "\u296c", - "rho;": "\u03c1", - "rhov;": "\u03f1", - "rightarrow;": "\u2192", - "rightarrowtail;": "\u21a3", - "rightharpoondown;": "\u21c1", - "rightharpoonup;": "\u21c0", - "rightleftarrows;": "\u21c4", - "rightleftharpoons;": "\u21cc", - "rightrightarrows;": "\u21c9", - "rightsquigarrow;": "\u219d", - "rightthreetimes;": "\u22cc", - "ring;": "\u02da", - "risingdotseq;": "\u2253", - "rlarr;": "\u21c4", - "rlhar;": "\u21cc", - "rlm;": "\u200f", - "rmoust;": "\u23b1", - "rmoustache;": "\u23b1", - "rnmid;": "\u2aee", - "roang;": "\u27ed", - "roarr;": "\u21fe", - "robrk;": "\u27e7", - "ropar;": "\u2986", - "ropf;": "\U0001d563", - "roplus;": "\u2a2e", - "rotimes;": "\u2a35", - "rpar;": ")", - "rpargt;": "\u2994", - "rppolint;": "\u2a12", - "rrarr;": "\u21c9", - "rsaquo;": "\u203a", - "rscr;": "\U0001d4c7", - "rsh;": "\u21b1", - "rsqb;": "]", - "rsquo;": "\u2019", - "rsquor;": "\u2019", - "rthree;": "\u22cc", - "rtimes;": "\u22ca", - "rtri;": "\u25b9", - "rtrie;": "\u22b5", - "rtrif;": "\u25b8", - "rtriltri;": "\u29ce", - "ruluhar;": "\u2968", - "rx;": "\u211e", - "sacute;": "\u015b", - "sbquo;": "\u201a", - "sc;": "\u227b", - "scE;": "\u2ab4", - "scap;": "\u2ab8", - "scaron;": "\u0161", - "sccue;": "\u227d", - "sce;": "\u2ab0", - "scedil;": "\u015f", - "scirc;": "\u015d", - "scnE;": "\u2ab6", - "scnap;": "\u2aba", - "scnsim;": "\u22e9", - "scpolint;": "\u2a13", - "scsim;": "\u227f", - "scy;": "\u0441", - "sdot;": "\u22c5", - "sdotb;": "\u22a1", - "sdote;": "\u2a66", - "seArr;": "\u21d8", - "searhk;": "\u2925", - "searr;": "\u2198", - "searrow;": "\u2198", - "sect": "\xa7", - "sect;": "\xa7", - "semi;": ";", - "seswar;": "\u2929", - "setminus;": "\u2216", - "setmn;": "\u2216", - "sext;": "\u2736", - "sfr;": "\U0001d530", - "sfrown;": "\u2322", - "sharp;": "\u266f", - "shchcy;": "\u0449", - "shcy;": "\u0448", - "shortmid;": "\u2223", - "shortparallel;": "\u2225", - "shy": "\xad", - "shy;": "\xad", - "sigma;": "\u03c3", - "sigmaf;": "\u03c2", - "sigmav;": "\u03c2", - "sim;": "\u223c", - "simdot;": "\u2a6a", - "sime;": "\u2243", - "simeq;": "\u2243", - "simg;": "\u2a9e", - "simgE;": "\u2aa0", - "siml;": "\u2a9d", - "simlE;": "\u2a9f", - "simne;": "\u2246", - "simplus;": "\u2a24", - "simrarr;": "\u2972", - "slarr;": "\u2190", - "smallsetminus;": "\u2216", - "smashp;": "\u2a33", - "smeparsl;": "\u29e4", - "smid;": "\u2223", - "smile;": "\u2323", - "smt;": "\u2aaa", - "smte;": "\u2aac", - "smtes;": "\u2aac\ufe00", - "softcy;": "\u044c", - "sol;": "/", - "solb;": "\u29c4", - "solbar;": "\u233f", - "sopf;": "\U0001d564", - "spades;": "\u2660", - "spadesuit;": "\u2660", - "spar;": "\u2225", - "sqcap;": "\u2293", - "sqcaps;": "\u2293\ufe00", - "sqcup;": "\u2294", - "sqcups;": "\u2294\ufe00", - "sqsub;": "\u228f", - "sqsube;": "\u2291", - "sqsubset;": "\u228f", - "sqsubseteq;": "\u2291", - "sqsup;": "\u2290", - "sqsupe;": "\u2292", - "sqsupset;": "\u2290", - "sqsupseteq;": "\u2292", - "squ;": "\u25a1", - "square;": "\u25a1", - "squarf;": "\u25aa", - "squf;": "\u25aa", - "srarr;": "\u2192", - "sscr;": "\U0001d4c8", - "ssetmn;": "\u2216", - "ssmile;": "\u2323", - "sstarf;": "\u22c6", - "star;": "\u2606", - "starf;": "\u2605", - "straightepsilon;": "\u03f5", - "straightphi;": "\u03d5", - "strns;": "\xaf", - "sub;": "\u2282", - "subE;": "\u2ac5", - "subdot;": "\u2abd", - "sube;": "\u2286", - "subedot;": "\u2ac3", - "submult;": "\u2ac1", - "subnE;": "\u2acb", - "subne;": "\u228a", - "subplus;": "\u2abf", - "subrarr;": "\u2979", - "subset;": "\u2282", - "subseteq;": "\u2286", - "subseteqq;": "\u2ac5", - "subsetneq;": "\u228a", - "subsetneqq;": "\u2acb", - "subsim;": "\u2ac7", - "subsub;": "\u2ad5", - "subsup;": "\u2ad3", - "succ;": "\u227b", - "succapprox;": "\u2ab8", - "succcurlyeq;": "\u227d", - "succeq;": "\u2ab0", - "succnapprox;": "\u2aba", - "succneqq;": "\u2ab6", - "succnsim;": "\u22e9", - "succsim;": "\u227f", - "sum;": "\u2211", - "sung;": "\u266a", - "sup1": "\xb9", - "sup1;": "\xb9", - "sup2": "\xb2", - "sup2;": "\xb2", - "sup3": "\xb3", - "sup3;": "\xb3", - "sup;": "\u2283", - "supE;": "\u2ac6", - "supdot;": "\u2abe", - "supdsub;": "\u2ad8", - "supe;": "\u2287", - "supedot;": "\u2ac4", - "suphsol;": "\u27c9", - "suphsub;": "\u2ad7", - "suplarr;": "\u297b", - "supmult;": "\u2ac2", - "supnE;": "\u2acc", - "supne;": "\u228b", - "supplus;": "\u2ac0", - "supset;": "\u2283", - "supseteq;": "\u2287", - "supseteqq;": "\u2ac6", - "supsetneq;": "\u228b", - "supsetneqq;": "\u2acc", - "supsim;": "\u2ac8", - "supsub;": "\u2ad4", - "supsup;": "\u2ad6", - "swArr;": "\u21d9", - "swarhk;": "\u2926", - "swarr;": "\u2199", - "swarrow;": "\u2199", - "swnwar;": "\u292a", - "szlig": "\xdf", - "szlig;": "\xdf", - "target;": "\u2316", - "tau;": "\u03c4", - "tbrk;": "\u23b4", - "tcaron;": "\u0165", - "tcedil;": "\u0163", - "tcy;": "\u0442", - "tdot;": "\u20db", - "telrec;": "\u2315", - "tfr;": "\U0001d531", - "there4;": "\u2234", - "therefore;": "\u2234", - "theta;": "\u03b8", - "thetasym;": "\u03d1", - "thetav;": "\u03d1", - "thickapprox;": "\u2248", - "thicksim;": "\u223c", - "thinsp;": "\u2009", - "thkap;": "\u2248", - "thksim;": "\u223c", - "thorn": "\xfe", - "thorn;": "\xfe", - "tilde;": "\u02dc", - "times": "\xd7", - "times;": "\xd7", - "timesb;": "\u22a0", - "timesbar;": "\u2a31", - "timesd;": "\u2a30", - "tint;": "\u222d", - "toea;": "\u2928", - "top;": "\u22a4", - "topbot;": "\u2336", - "topcir;": "\u2af1", - "topf;": "\U0001d565", - "topfork;": "\u2ada", - "tosa;": "\u2929", - "tprime;": "\u2034", - "trade;": "\u2122", - "triangle;": "\u25b5", - "triangledown;": "\u25bf", - "triangleleft;": "\u25c3", - "trianglelefteq;": "\u22b4", - "triangleq;": "\u225c", - "triangleright;": "\u25b9", - "trianglerighteq;": "\u22b5", - "tridot;": "\u25ec", - "trie;": "\u225c", - "triminus;": "\u2a3a", - "triplus;": "\u2a39", - "trisb;": "\u29cd", - "tritime;": "\u2a3b", - "trpezium;": "\u23e2", - "tscr;": "\U0001d4c9", - "tscy;": "\u0446", - "tshcy;": "\u045b", - "tstrok;": "\u0167", - "twixt;": "\u226c", - "twoheadleftarrow;": "\u219e", - "twoheadrightarrow;": "\u21a0", - "uArr;": "\u21d1", - "uHar;": "\u2963", - "uacute": "\xfa", - "uacute;": "\xfa", - "uarr;": "\u2191", - "ubrcy;": "\u045e", - "ubreve;": "\u016d", - "ucirc": "\xfb", - "ucirc;": "\xfb", - "ucy;": "\u0443", - "udarr;": "\u21c5", - "udblac;": "\u0171", - "udhar;": "\u296e", - "ufisht;": "\u297e", - "ufr;": "\U0001d532", - "ugrave": "\xf9", - "ugrave;": "\xf9", - "uharl;": "\u21bf", - "uharr;": "\u21be", - "uhblk;": "\u2580", - "ulcorn;": "\u231c", - "ulcorner;": "\u231c", - "ulcrop;": "\u230f", - "ultri;": "\u25f8", - "umacr;": "\u016b", - "uml": "\xa8", - "uml;": "\xa8", - "uogon;": "\u0173", - "uopf;": "\U0001d566", - "uparrow;": "\u2191", - "updownarrow;": "\u2195", - "upharpoonleft;": "\u21bf", - "upharpoonright;": "\u21be", - "uplus;": "\u228e", - "upsi;": "\u03c5", - "upsih;": "\u03d2", - "upsilon;": "\u03c5", - "upuparrows;": "\u21c8", - "urcorn;": "\u231d", - "urcorner;": "\u231d", - "urcrop;": "\u230e", - "uring;": "\u016f", - "urtri;": "\u25f9", - "uscr;": "\U0001d4ca", - "utdot;": "\u22f0", - "utilde;": "\u0169", - "utri;": "\u25b5", - "utrif;": "\u25b4", - "uuarr;": "\u21c8", - "uuml": "\xfc", - "uuml;": "\xfc", - "uwangle;": "\u29a7", - "vArr;": "\u21d5", - "vBar;": "\u2ae8", - "vBarv;": "\u2ae9", - "vDash;": "\u22a8", - "vangrt;": "\u299c", - "varepsilon;": "\u03f5", - "varkappa;": "\u03f0", - "varnothing;": "\u2205", - "varphi;": "\u03d5", - "varpi;": "\u03d6", - "varpropto;": "\u221d", - "varr;": "\u2195", - "varrho;": "\u03f1", - "varsigma;": "\u03c2", - "varsubsetneq;": "\u228a\ufe00", - "varsubsetneqq;": "\u2acb\ufe00", - "varsupsetneq;": "\u228b\ufe00", - "varsupsetneqq;": "\u2acc\ufe00", - "vartheta;": "\u03d1", - "vartriangleleft;": "\u22b2", - "vartriangleright;": "\u22b3", - "vcy;": "\u0432", - "vdash;": "\u22a2", - "vee;": "\u2228", - "veebar;": "\u22bb", - "veeeq;": "\u225a", - "vellip;": "\u22ee", - "verbar;": "|", - "vert;": "|", - "vfr;": "\U0001d533", - "vltri;": "\u22b2", - "vnsub;": "\u2282\u20d2", - "vnsup;": "\u2283\u20d2", - "vopf;": "\U0001d567", - "vprop;": "\u221d", - "vrtri;": "\u22b3", - "vscr;": "\U0001d4cb", - "vsubnE;": "\u2acb\ufe00", - "vsubne;": "\u228a\ufe00", - "vsupnE;": "\u2acc\ufe00", - "vsupne;": "\u228b\ufe00", - "vzigzag;": "\u299a", - "wcirc;": "\u0175", - "wedbar;": "\u2a5f", - "wedge;": "\u2227", - "wedgeq;": "\u2259", - "weierp;": "\u2118", - "wfr;": "\U0001d534", - "wopf;": "\U0001d568", - "wp;": "\u2118", - "wr;": "\u2240", - "wreath;": "\u2240", - "wscr;": "\U0001d4cc", - "xcap;": "\u22c2", - "xcirc;": "\u25ef", - "xcup;": "\u22c3", - "xdtri;": "\u25bd", - "xfr;": "\U0001d535", - "xhArr;": "\u27fa", - "xharr;": "\u27f7", - "xi;": "\u03be", - "xlArr;": "\u27f8", - "xlarr;": "\u27f5", - "xmap;": "\u27fc", - "xnis;": "\u22fb", - "xodot;": "\u2a00", - "xopf;": "\U0001d569", - "xoplus;": "\u2a01", - "xotime;": "\u2a02", - "xrArr;": "\u27f9", - "xrarr;": "\u27f6", - "xscr;": "\U0001d4cd", - "xsqcup;": "\u2a06", - "xuplus;": "\u2a04", - "xutri;": "\u25b3", - "xvee;": "\u22c1", - "xwedge;": "\u22c0", - "yacute": "\xfd", - "yacute;": "\xfd", - "yacy;": "\u044f", - "ycirc;": "\u0177", - "ycy;": "\u044b", - "yen": "\xa5", - "yen;": "\xa5", - "yfr;": "\U0001d536", - "yicy;": "\u0457", - "yopf;": "\U0001d56a", - "yscr;": "\U0001d4ce", - "yucy;": "\u044e", - "yuml": "\xff", - "yuml;": "\xff", - "zacute;": "\u017a", - "zcaron;": "\u017e", - "zcy;": "\u0437", - "zdot;": "\u017c", - "zeetrf;": "\u2128", - "zeta;": "\u03b6", - "zfr;": "\U0001d537", - "zhcy;": "\u0436", - "zigrarr;": "\u21dd", - "zopf;": "\U0001d56b", - "zscr;": "\U0001d4cf", - "zwj;": "\u200d", - "zwnj;": "\u200c", -} - -replacementCharacters = { - 0x0: "\uFFFD", - 0x0d: "\u000D", - 0x80: "\u20AC", - 0x81: "\u0081", - 0x82: "\u201A", - 0x83: "\u0192", - 0x84: "\u201E", - 0x85: "\u2026", - 0x86: "\u2020", - 0x87: "\u2021", - 0x88: "\u02C6", - 0x89: "\u2030", - 0x8A: "\u0160", - 0x8B: "\u2039", - 0x8C: "\u0152", - 0x8D: "\u008D", - 0x8E: "\u017D", - 0x8F: "\u008F", - 0x90: "\u0090", - 0x91: "\u2018", - 0x92: "\u2019", - 0x93: "\u201C", - 0x94: "\u201D", - 0x95: "\u2022", - 0x96: "\u2013", - 0x97: "\u2014", - 0x98: "\u02DC", - 0x99: "\u2122", - 0x9A: "\u0161", - 0x9B: "\u203A", - 0x9C: "\u0153", - 0x9D: "\u009D", - 0x9E: "\u017E", - 0x9F: "\u0178", -} - -tokenTypes = { - "Doctype": 0, - "Characters": 1, - "SpaceCharacters": 2, - "StartTag": 3, - "EndTag": 4, - "EmptyTag": 5, - "Comment": 6, - "ParseError": 7 -} - -tagTokenTypes = frozenset([tokenTypes["StartTag"], tokenTypes["EndTag"], - tokenTypes["EmptyTag"]]) - - -prefixes = dict([(v, k) for k, v in namespaces.items()]) -prefixes["http://www.w3.org/1998/Math/MathML"] = "math" - - -class DataLossWarning(UserWarning): - """Raised when the current tree is unable to represent the input data""" - pass - - -class _ReparseException(Exception): - pass diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib/filters/alphabeticalattributes.py b/src/fetchcode/vcs/pip/_vendor/html5lib/filters/alphabeticalattributes.py deleted file mode 100644 index 5ba926e3..00000000 --- a/src/fetchcode/vcs/pip/_vendor/html5lib/filters/alphabeticalattributes.py +++ /dev/null @@ -1,29 +0,0 @@ -from __future__ import absolute_import, division, unicode_literals - -from . import base - -from collections import OrderedDict - - -def _attr_key(attr): - """Return an appropriate key for an attribute for sorting - - Attributes have a namespace that can be either ``None`` or a string. We - can't compare the two because they're different types, so we convert - ``None`` to an empty string first. - - """ - return (attr[0][0] or ''), attr[0][1] - - -class Filter(base.Filter): - """Alphabetizes attributes for elements""" - def __iter__(self): - for token in base.Filter.__iter__(self): - if token["type"] in ("StartTag", "EmptyTag"): - attrs = OrderedDict() - for name, value in sorted(token["data"].items(), - key=_attr_key): - attrs[name] = value - token["data"] = attrs - yield token diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib/filters/base.py b/src/fetchcode/vcs/pip/_vendor/html5lib/filters/base.py deleted file mode 100644 index c7dbaed0..00000000 --- a/src/fetchcode/vcs/pip/_vendor/html5lib/filters/base.py +++ /dev/null @@ -1,12 +0,0 @@ -from __future__ import absolute_import, division, unicode_literals - - -class Filter(object): - def __init__(self, source): - self.source = source - - def __iter__(self): - return iter(self.source) - - def __getattr__(self, name): - return getattr(self.source, name) diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib/filters/inject_meta_charset.py b/src/fetchcode/vcs/pip/_vendor/html5lib/filters/inject_meta_charset.py deleted file mode 100644 index aefb5c84..00000000 --- a/src/fetchcode/vcs/pip/_vendor/html5lib/filters/inject_meta_charset.py +++ /dev/null @@ -1,73 +0,0 @@ -from __future__ import absolute_import, division, unicode_literals - -from . import base - - -class Filter(base.Filter): - """Injects ```` tag into head of document""" - def __init__(self, source, encoding): - """Creates a Filter - - :arg source: the source token stream - - :arg encoding: the encoding to set - - """ - base.Filter.__init__(self, source) - self.encoding = encoding - - def __iter__(self): - state = "pre_head" - meta_found = (self.encoding is None) - pending = [] - - for token in base.Filter.__iter__(self): - type = token["type"] - if type == "StartTag": - if token["name"].lower() == "head": - state = "in_head" - - elif type == "EmptyTag": - if token["name"].lower() == "meta": - # replace charset with actual encoding - has_http_equiv_content_type = False - for (namespace, name), value in token["data"].items(): - if namespace is not None: - continue - elif name.lower() == 'charset': - token["data"][(namespace, name)] = self.encoding - meta_found = True - break - elif name == 'http-equiv' and value.lower() == 'content-type': - has_http_equiv_content_type = True - else: - if has_http_equiv_content_type and (None, "content") in token["data"]: - token["data"][(None, "content")] = 'text/html; charset=%s' % self.encoding - meta_found = True - - elif token["name"].lower() == "head" and not meta_found: - # insert meta into empty head - yield {"type": "StartTag", "name": "head", - "data": token["data"]} - yield {"type": "EmptyTag", "name": "meta", - "data": {(None, "charset"): self.encoding}} - yield {"type": "EndTag", "name": "head"} - meta_found = True - continue - - elif type == "EndTag": - if token["name"].lower() == "head" and pending: - # insert meta into head (if necessary) and flush pending queue - yield pending.pop(0) - if not meta_found: - yield {"type": "EmptyTag", "name": "meta", - "data": {(None, "charset"): self.encoding}} - while pending: - yield pending.pop(0) - meta_found = True - state = "post_head" - - if state == "in_head": - pending.append(token) - else: - yield token diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib/filters/lint.py b/src/fetchcode/vcs/pip/_vendor/html5lib/filters/lint.py deleted file mode 100644 index 11947354..00000000 --- a/src/fetchcode/vcs/pip/_vendor/html5lib/filters/lint.py +++ /dev/null @@ -1,93 +0,0 @@ -from __future__ import absolute_import, division, unicode_literals - -from fetchcode.vcs.pip._vendor.six import text_type - -from . import base -from ..constants import namespaces, voidElements - -from ..constants import spaceCharacters -spaceCharacters = "".join(spaceCharacters) - - -class Filter(base.Filter): - """Lints the token stream for errors - - If it finds any errors, it'll raise an ``AssertionError``. - - """ - def __init__(self, source, require_matching_tags=True): - """Creates a Filter - - :arg source: the source token stream - - :arg require_matching_tags: whether or not to require matching tags - - """ - super(Filter, self).__init__(source) - self.require_matching_tags = require_matching_tags - - def __iter__(self): - open_elements = [] - for token in base.Filter.__iter__(self): - type = token["type"] - if type in ("StartTag", "EmptyTag"): - namespace = token["namespace"] - name = token["name"] - assert namespace is None or isinstance(namespace, text_type) - assert namespace != "" - assert isinstance(name, text_type) - assert name != "" - assert isinstance(token["data"], dict) - if (not namespace or namespace == namespaces["html"]) and name in voidElements: - assert type == "EmptyTag" - else: - assert type == "StartTag" - if type == "StartTag" and self.require_matching_tags: - open_elements.append((namespace, name)) - for (namespace, name), value in token["data"].items(): - assert namespace is None or isinstance(namespace, text_type) - assert namespace != "" - assert isinstance(name, text_type) - assert name != "" - assert isinstance(value, text_type) - - elif type == "EndTag": - namespace = token["namespace"] - name = token["name"] - assert namespace is None or isinstance(namespace, text_type) - assert namespace != "" - assert isinstance(name, text_type) - assert name != "" - if (not namespace or namespace == namespaces["html"]) and name in voidElements: - assert False, "Void element reported as EndTag token: %(tag)s" % {"tag": name} - elif self.require_matching_tags: - start = open_elements.pop() - assert start == (namespace, name) - - elif type == "Comment": - data = token["data"] - assert isinstance(data, text_type) - - elif type in ("Characters", "SpaceCharacters"): - data = token["data"] - assert isinstance(data, text_type) - assert data != "" - if type == "SpaceCharacters": - assert data.strip(spaceCharacters) == "" - - elif type == "Doctype": - name = token["name"] - assert name is None or isinstance(name, text_type) - assert token["publicId"] is None or isinstance(name, text_type) - assert token["systemId"] is None or isinstance(name, text_type) - - elif type == "Entity": - assert isinstance(token["name"], text_type) - - elif type == "SerializerError": - assert isinstance(token["data"], text_type) - - else: - assert False, "Unknown token type: %(type)s" % {"type": type} - - yield token diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib/filters/optionaltags.py b/src/fetchcode/vcs/pip/_vendor/html5lib/filters/optionaltags.py deleted file mode 100644 index 4a865012..00000000 --- a/src/fetchcode/vcs/pip/_vendor/html5lib/filters/optionaltags.py +++ /dev/null @@ -1,207 +0,0 @@ -from __future__ import absolute_import, division, unicode_literals - -from . import base - - -class Filter(base.Filter): - """Removes optional tags from the token stream""" - def slider(self): - previous1 = previous2 = None - for token in self.source: - if previous1 is not None: - yield previous2, previous1, token - previous2 = previous1 - previous1 = token - if previous1 is not None: - yield previous2, previous1, None - - def __iter__(self): - for previous, token, next in self.slider(): - type = token["type"] - if type == "StartTag": - if (token["data"] or - not self.is_optional_start(token["name"], previous, next)): - yield token - elif type == "EndTag": - if not self.is_optional_end(token["name"], next): - yield token - else: - yield token - - def is_optional_start(self, tagname, previous, next): - type = next and next["type"] or None - if tagname in 'html': - # An html element's start tag may be omitted if the first thing - # inside the html element is not a space character or a comment. - return type not in ("Comment", "SpaceCharacters") - elif tagname == 'head': - # A head element's start tag may be omitted if the first thing - # inside the head element is an element. - # XXX: we also omit the start tag if the head element is empty - if type in ("StartTag", "EmptyTag"): - return True - elif type == "EndTag": - return next["name"] == "head" - elif tagname == 'body': - # A body element's start tag may be omitted if the first thing - # inside the body element is not a space character or a comment, - # except if the first thing inside the body element is a script - # or style element and the node immediately preceding the body - # element is a head element whose end tag has been omitted. - if type in ("Comment", "SpaceCharacters"): - return False - elif type == "StartTag": - # XXX: we do not look at the preceding event, so we never omit - # the body element's start tag if it's followed by a script or - # a style element. - return next["name"] not in ('script', 'style') - else: - return True - elif tagname == 'colgroup': - # A colgroup element's start tag may be omitted if the first thing - # inside the colgroup element is a col element, and if the element - # is not immediately preceded by another colgroup element whose - # end tag has been omitted. - if type in ("StartTag", "EmptyTag"): - # XXX: we do not look at the preceding event, so instead we never - # omit the colgroup element's end tag when it is immediately - # followed by another colgroup element. See is_optional_end. - return next["name"] == "col" - else: - return False - elif tagname == 'tbody': - # A tbody element's start tag may be omitted if the first thing - # inside the tbody element is a tr element, and if the element is - # not immediately preceded by a tbody, thead, or tfoot element - # whose end tag has been omitted. - if type == "StartTag": - # omit the thead and tfoot elements' end tag when they are - # immediately followed by a tbody element. See is_optional_end. - if previous and previous['type'] == 'EndTag' and \ - previous['name'] in ('tbody', 'thead', 'tfoot'): - return False - return next["name"] == 'tr' - else: - return False - return False - - def is_optional_end(self, tagname, next): - type = next and next["type"] or None - if tagname in ('html', 'head', 'body'): - # An html element's end tag may be omitted if the html element - # is not immediately followed by a space character or a comment. - return type not in ("Comment", "SpaceCharacters") - elif tagname in ('li', 'optgroup', 'tr'): - # A li element's end tag may be omitted if the li element is - # immediately followed by another li element or if there is - # no more content in the parent element. - # An optgroup element's end tag may be omitted if the optgroup - # element is immediately followed by another optgroup element, - # or if there is no more content in the parent element. - # A tr element's end tag may be omitted if the tr element is - # immediately followed by another tr element, or if there is - # no more content in the parent element. - if type == "StartTag": - return next["name"] == tagname - else: - return type == "EndTag" or type is None - elif tagname in ('dt', 'dd'): - # A dt element's end tag may be omitted if the dt element is - # immediately followed by another dt element or a dd element. - # A dd element's end tag may be omitted if the dd element is - # immediately followed by another dd element or a dt element, - # or if there is no more content in the parent element. - if type == "StartTag": - return next["name"] in ('dt', 'dd') - elif tagname == 'dd': - return type == "EndTag" or type is None - else: - return False - elif tagname == 'p': - # A p element's end tag may be omitted if the p element is - # immediately followed by an address, article, aside, - # blockquote, datagrid, dialog, dir, div, dl, fieldset, - # footer, form, h1, h2, h3, h4, h5, h6, header, hr, menu, - # nav, ol, p, pre, section, table, or ul, element, or if - # there is no more content in the parent element. - if type in ("StartTag", "EmptyTag"): - return next["name"] in ('address', 'article', 'aside', - 'blockquote', 'datagrid', 'dialog', - 'dir', 'div', 'dl', 'fieldset', 'footer', - 'form', 'h1', 'h2', 'h3', 'h4', 'h5', 'h6', - 'header', 'hr', 'menu', 'nav', 'ol', - 'p', 'pre', 'section', 'table', 'ul') - else: - return type == "EndTag" or type is None - elif tagname == 'option': - # An option element's end tag may be omitted if the option - # element is immediately followed by another option element, - # or if it is immediately followed by an optgroup - # element, or if there is no more content in the parent - # element. - if type == "StartTag": - return next["name"] in ('option', 'optgroup') - else: - return type == "EndTag" or type is None - elif tagname in ('rt', 'rp'): - # An rt element's end tag may be omitted if the rt element is - # immediately followed by an rt or rp element, or if there is - # no more content in the parent element. - # An rp element's end tag may be omitted if the rp element is - # immediately followed by an rt or rp element, or if there is - # no more content in the parent element. - if type == "StartTag": - return next["name"] in ('rt', 'rp') - else: - return type == "EndTag" or type is None - elif tagname == 'colgroup': - # A colgroup element's end tag may be omitted if the colgroup - # element is not immediately followed by a space character or - # a comment. - if type in ("Comment", "SpaceCharacters"): - return False - elif type == "StartTag": - # XXX: we also look for an immediately following colgroup - # element. See is_optional_start. - return next["name"] != 'colgroup' - else: - return True - elif tagname in ('thead', 'tbody'): - # A thead element's end tag may be omitted if the thead element - # is immediately followed by a tbody or tfoot element. - # A tbody element's end tag may be omitted if the tbody element - # is immediately followed by a tbody or tfoot element, or if - # there is no more content in the parent element. - # A tfoot element's end tag may be omitted if the tfoot element - # is immediately followed by a tbody element, or if there is no - # more content in the parent element. - # XXX: we never omit the end tag when the following element is - # a tbody. See is_optional_start. - if type == "StartTag": - return next["name"] in ['tbody', 'tfoot'] - elif tagname == 'tbody': - return type == "EndTag" or type is None - else: - return False - elif tagname == 'tfoot': - # A tfoot element's end tag may be omitted if the tfoot element - # is immediately followed by a tbody element, or if there is no - # more content in the parent element. - # XXX: we never omit the end tag when the following element is - # a tbody. See is_optional_start. - if type == "StartTag": - return next["name"] == 'tbody' - else: - return type == "EndTag" or type is None - elif tagname in ('td', 'th'): - # A td element's end tag may be omitted if the td element is - # immediately followed by a td or th element, or if there is - # no more content in the parent element. - # A th element's end tag may be omitted if the th element is - # immediately followed by a td or th element, or if there is - # no more content in the parent element. - if type == "StartTag": - return next["name"] in ('td', 'th') - else: - return type == "EndTag" or type is None - return False diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib/filters/sanitizer.py b/src/fetchcode/vcs/pip/_vendor/html5lib/filters/sanitizer.py deleted file mode 100644 index 4a8bc3c8..00000000 --- a/src/fetchcode/vcs/pip/_vendor/html5lib/filters/sanitizer.py +++ /dev/null @@ -1,896 +0,0 @@ -from __future__ import absolute_import, division, unicode_literals - -import re -from xml.sax.saxutils import escape, unescape - -from fetchcode.vcs.pip._vendor.six.moves import urllib_parse as urlparse - -from . import base -from ..constants import namespaces, prefixes - -__all__ = ["Filter"] - - -allowed_elements = frozenset(( - (namespaces['html'], 'a'), - (namespaces['html'], 'abbr'), - (namespaces['html'], 'acronym'), - (namespaces['html'], 'address'), - (namespaces['html'], 'area'), - (namespaces['html'], 'article'), - (namespaces['html'], 'aside'), - (namespaces['html'], 'audio'), - (namespaces['html'], 'b'), - (namespaces['html'], 'big'), - (namespaces['html'], 'blockquote'), - (namespaces['html'], 'br'), - (namespaces['html'], 'button'), - (namespaces['html'], 'canvas'), - (namespaces['html'], 'caption'), - (namespaces['html'], 'center'), - (namespaces['html'], 'cite'), - (namespaces['html'], 'code'), - (namespaces['html'], 'col'), - (namespaces['html'], 'colgroup'), - (namespaces['html'], 'command'), - (namespaces['html'], 'datagrid'), - (namespaces['html'], 'datalist'), - (namespaces['html'], 'dd'), - (namespaces['html'], 'del'), - (namespaces['html'], 'details'), - (namespaces['html'], 'dfn'), - (namespaces['html'], 'dialog'), - (namespaces['html'], 'dir'), - (namespaces['html'], 'div'), - (namespaces['html'], 'dl'), - (namespaces['html'], 'dt'), - (namespaces['html'], 'em'), - (namespaces['html'], 'event-source'), - (namespaces['html'], 'fieldset'), - (namespaces['html'], 'figcaption'), - (namespaces['html'], 'figure'), - (namespaces['html'], 'footer'), - (namespaces['html'], 'font'), - (namespaces['html'], 'form'), - (namespaces['html'], 'header'), - (namespaces['html'], 'h1'), - (namespaces['html'], 'h2'), - (namespaces['html'], 'h3'), - (namespaces['html'], 'h4'), - (namespaces['html'], 'h5'), - (namespaces['html'], 'h6'), - (namespaces['html'], 'hr'), - (namespaces['html'], 'i'), - (namespaces['html'], 'img'), - (namespaces['html'], 'input'), - (namespaces['html'], 'ins'), - (namespaces['html'], 'keygen'), - (namespaces['html'], 'kbd'), - (namespaces['html'], 'label'), - (namespaces['html'], 'legend'), - (namespaces['html'], 'li'), - (namespaces['html'], 'm'), - (namespaces['html'], 'map'), - (namespaces['html'], 'menu'), - (namespaces['html'], 'meter'), - (namespaces['html'], 'multicol'), - (namespaces['html'], 'nav'), - (namespaces['html'], 'nextid'), - (namespaces['html'], 'ol'), - (namespaces['html'], 'output'), - (namespaces['html'], 'optgroup'), - (namespaces['html'], 'option'), - (namespaces['html'], 'p'), - (namespaces['html'], 'pre'), - (namespaces['html'], 'progress'), - (namespaces['html'], 'q'), - (namespaces['html'], 's'), - (namespaces['html'], 'samp'), - (namespaces['html'], 'section'), - (namespaces['html'], 'select'), - (namespaces['html'], 'small'), - (namespaces['html'], 'sound'), - (namespaces['html'], 'source'), - (namespaces['html'], 'spacer'), - (namespaces['html'], 'span'), - (namespaces['html'], 'strike'), - (namespaces['html'], 'strong'), - (namespaces['html'], 'sub'), - (namespaces['html'], 'sup'), - (namespaces['html'], 'table'), - (namespaces['html'], 'tbody'), - (namespaces['html'], 'td'), - (namespaces['html'], 'textarea'), - (namespaces['html'], 'time'), - (namespaces['html'], 'tfoot'), - (namespaces['html'], 'th'), - (namespaces['html'], 'thead'), - (namespaces['html'], 'tr'), - (namespaces['html'], 'tt'), - (namespaces['html'], 'u'), - (namespaces['html'], 'ul'), - (namespaces['html'], 'var'), - (namespaces['html'], 'video'), - (namespaces['mathml'], 'maction'), - (namespaces['mathml'], 'math'), - (namespaces['mathml'], 'merror'), - (namespaces['mathml'], 'mfrac'), - (namespaces['mathml'], 'mi'), - (namespaces['mathml'], 'mmultiscripts'), - (namespaces['mathml'], 'mn'), - (namespaces['mathml'], 'mo'), - (namespaces['mathml'], 'mover'), - (namespaces['mathml'], 'mpadded'), - (namespaces['mathml'], 'mphantom'), - (namespaces['mathml'], 'mprescripts'), - (namespaces['mathml'], 'mroot'), - (namespaces['mathml'], 'mrow'), - (namespaces['mathml'], 'mspace'), - (namespaces['mathml'], 'msqrt'), - (namespaces['mathml'], 'mstyle'), - (namespaces['mathml'], 'msub'), - (namespaces['mathml'], 'msubsup'), - (namespaces['mathml'], 'msup'), - (namespaces['mathml'], 'mtable'), - (namespaces['mathml'], 'mtd'), - (namespaces['mathml'], 'mtext'), - (namespaces['mathml'], 'mtr'), - (namespaces['mathml'], 'munder'), - (namespaces['mathml'], 'munderover'), - (namespaces['mathml'], 'none'), - (namespaces['svg'], 'a'), - (namespaces['svg'], 'animate'), - (namespaces['svg'], 'animateColor'), - (namespaces['svg'], 'animateMotion'), - (namespaces['svg'], 'animateTransform'), - (namespaces['svg'], 'clipPath'), - (namespaces['svg'], 'circle'), - (namespaces['svg'], 'defs'), - (namespaces['svg'], 'desc'), - (namespaces['svg'], 'ellipse'), - (namespaces['svg'], 'font-face'), - (namespaces['svg'], 'font-face-name'), - (namespaces['svg'], 'font-face-src'), - (namespaces['svg'], 'g'), - (namespaces['svg'], 'glyph'), - (namespaces['svg'], 'hkern'), - (namespaces['svg'], 'linearGradient'), - (namespaces['svg'], 'line'), - (namespaces['svg'], 'marker'), - (namespaces['svg'], 'metadata'), - (namespaces['svg'], 'missing-glyph'), - (namespaces['svg'], 'mpath'), - (namespaces['svg'], 'path'), - (namespaces['svg'], 'polygon'), - (namespaces['svg'], 'polyline'), - (namespaces['svg'], 'radialGradient'), - (namespaces['svg'], 'rect'), - (namespaces['svg'], 'set'), - (namespaces['svg'], 'stop'), - (namespaces['svg'], 'svg'), - (namespaces['svg'], 'switch'), - (namespaces['svg'], 'text'), - (namespaces['svg'], 'title'), - (namespaces['svg'], 'tspan'), - (namespaces['svg'], 'use'), -)) - -allowed_attributes = frozenset(( - # HTML attributes - (None, 'abbr'), - (None, 'accept'), - (None, 'accept-charset'), - (None, 'accesskey'), - (None, 'action'), - (None, 'align'), - (None, 'alt'), - (None, 'autocomplete'), - (None, 'autofocus'), - (None, 'axis'), - (None, 'background'), - (None, 'balance'), - (None, 'bgcolor'), - (None, 'bgproperties'), - (None, 'border'), - (None, 'bordercolor'), - (None, 'bordercolordark'), - (None, 'bordercolorlight'), - (None, 'bottompadding'), - (None, 'cellpadding'), - (None, 'cellspacing'), - (None, 'ch'), - (None, 'challenge'), - (None, 'char'), - (None, 'charoff'), - (None, 'choff'), - (None, 'charset'), - (None, 'checked'), - (None, 'cite'), - (None, 'class'), - (None, 'clear'), - (None, 'color'), - (None, 'cols'), - (None, 'colspan'), - (None, 'compact'), - (None, 'contenteditable'), - (None, 'controls'), - (None, 'coords'), - (None, 'data'), - (None, 'datafld'), - (None, 'datapagesize'), - (None, 'datasrc'), - (None, 'datetime'), - (None, 'default'), - (None, 'delay'), - (None, 'dir'), - (None, 'disabled'), - (None, 'draggable'), - (None, 'dynsrc'), - (None, 'enctype'), - (None, 'end'), - (None, 'face'), - (None, 'for'), - (None, 'form'), - (None, 'frame'), - (None, 'galleryimg'), - (None, 'gutter'), - (None, 'headers'), - (None, 'height'), - (None, 'hidefocus'), - (None, 'hidden'), - (None, 'high'), - (None, 'href'), - (None, 'hreflang'), - (None, 'hspace'), - (None, 'icon'), - (None, 'id'), - (None, 'inputmode'), - (None, 'ismap'), - (None, 'keytype'), - (None, 'label'), - (None, 'leftspacing'), - (None, 'lang'), - (None, 'list'), - (None, 'longdesc'), - (None, 'loop'), - (None, 'loopcount'), - (None, 'loopend'), - (None, 'loopstart'), - (None, 'low'), - (None, 'lowsrc'), - (None, 'max'), - (None, 'maxlength'), - (None, 'media'), - (None, 'method'), - (None, 'min'), - (None, 'multiple'), - (None, 'name'), - (None, 'nohref'), - (None, 'noshade'), - (None, 'nowrap'), - (None, 'open'), - (None, 'optimum'), - (None, 'pattern'), - (None, 'ping'), - (None, 'point-size'), - (None, 'poster'), - (None, 'pqg'), - (None, 'preload'), - (None, 'prompt'), - (None, 'radiogroup'), - (None, 'readonly'), - (None, 'rel'), - (None, 'repeat-max'), - (None, 'repeat-min'), - (None, 'replace'), - (None, 'required'), - (None, 'rev'), - (None, 'rightspacing'), - (None, 'rows'), - (None, 'rowspan'), - (None, 'rules'), - (None, 'scope'), - (None, 'selected'), - (None, 'shape'), - (None, 'size'), - (None, 'span'), - (None, 'src'), - (None, 'start'), - (None, 'step'), - (None, 'style'), - (None, 'summary'), - (None, 'suppress'), - (None, 'tabindex'), - (None, 'target'), - (None, 'template'), - (None, 'title'), - (None, 'toppadding'), - (None, 'type'), - (None, 'unselectable'), - (None, 'usemap'), - (None, 'urn'), - (None, 'valign'), - (None, 'value'), - (None, 'variable'), - (None, 'volume'), - (None, 'vspace'), - (None, 'vrml'), - (None, 'width'), - (None, 'wrap'), - (namespaces['xml'], 'lang'), - # MathML attributes - (None, 'actiontype'), - (None, 'align'), - (None, 'columnalign'), - (None, 'columnalign'), - (None, 'columnalign'), - (None, 'columnlines'), - (None, 'columnspacing'), - (None, 'columnspan'), - (None, 'depth'), - (None, 'display'), - (None, 'displaystyle'), - (None, 'equalcolumns'), - (None, 'equalrows'), - (None, 'fence'), - (None, 'fontstyle'), - (None, 'fontweight'), - (None, 'frame'), - (None, 'height'), - (None, 'linethickness'), - (None, 'lspace'), - (None, 'mathbackground'), - (None, 'mathcolor'), - (None, 'mathvariant'), - (None, 'mathvariant'), - (None, 'maxsize'), - (None, 'minsize'), - (None, 'other'), - (None, 'rowalign'), - (None, 'rowalign'), - (None, 'rowalign'), - (None, 'rowlines'), - (None, 'rowspacing'), - (None, 'rowspan'), - (None, 'rspace'), - (None, 'scriptlevel'), - (None, 'selection'), - (None, 'separator'), - (None, 'stretchy'), - (None, 'width'), - (None, 'width'), - (namespaces['xlink'], 'href'), - (namespaces['xlink'], 'show'), - (namespaces['xlink'], 'type'), - # SVG attributes - (None, 'accent-height'), - (None, 'accumulate'), - (None, 'additive'), - (None, 'alphabetic'), - (None, 'arabic-form'), - (None, 'ascent'), - (None, 'attributeName'), - (None, 'attributeType'), - (None, 'baseProfile'), - (None, 'bbox'), - (None, 'begin'), - (None, 'by'), - (None, 'calcMode'), - (None, 'cap-height'), - (None, 'class'), - (None, 'clip-path'), - (None, 'color'), - (None, 'color-rendering'), - (None, 'content'), - (None, 'cx'), - (None, 'cy'), - (None, 'd'), - (None, 'dx'), - (None, 'dy'), - (None, 'descent'), - (None, 'display'), - (None, 'dur'), - (None, 'end'), - (None, 'fill'), - (None, 'fill-opacity'), - (None, 'fill-rule'), - (None, 'font-family'), - (None, 'font-size'), - (None, 'font-stretch'), - (None, 'font-style'), - (None, 'font-variant'), - (None, 'font-weight'), - (None, 'from'), - (None, 'fx'), - (None, 'fy'), - (None, 'g1'), - (None, 'g2'), - (None, 'glyph-name'), - (None, 'gradientUnits'), - (None, 'hanging'), - (None, 'height'), - (None, 'horiz-adv-x'), - (None, 'horiz-origin-x'), - (None, 'id'), - (None, 'ideographic'), - (None, 'k'), - (None, 'keyPoints'), - (None, 'keySplines'), - (None, 'keyTimes'), - (None, 'lang'), - (None, 'marker-end'), - (None, 'marker-mid'), - (None, 'marker-start'), - (None, 'markerHeight'), - (None, 'markerUnits'), - (None, 'markerWidth'), - (None, 'mathematical'), - (None, 'max'), - (None, 'min'), - (None, 'name'), - (None, 'offset'), - (None, 'opacity'), - (None, 'orient'), - (None, 'origin'), - (None, 'overline-position'), - (None, 'overline-thickness'), - (None, 'panose-1'), - (None, 'path'), - (None, 'pathLength'), - (None, 'points'), - (None, 'preserveAspectRatio'), - (None, 'r'), - (None, 'refX'), - (None, 'refY'), - (None, 'repeatCount'), - (None, 'repeatDur'), - (None, 'requiredExtensions'), - (None, 'requiredFeatures'), - (None, 'restart'), - (None, 'rotate'), - (None, 'rx'), - (None, 'ry'), - (None, 'slope'), - (None, 'stemh'), - (None, 'stemv'), - (None, 'stop-color'), - (None, 'stop-opacity'), - (None, 'strikethrough-position'), - (None, 'strikethrough-thickness'), - (None, 'stroke'), - (None, 'stroke-dasharray'), - (None, 'stroke-dashoffset'), - (None, 'stroke-linecap'), - (None, 'stroke-linejoin'), - (None, 'stroke-miterlimit'), - (None, 'stroke-opacity'), - (None, 'stroke-width'), - (None, 'systemLanguage'), - (None, 'target'), - (None, 'text-anchor'), - (None, 'to'), - (None, 'transform'), - (None, 'type'), - (None, 'u1'), - (None, 'u2'), - (None, 'underline-position'), - (None, 'underline-thickness'), - (None, 'unicode'), - (None, 'unicode-range'), - (None, 'units-per-em'), - (None, 'values'), - (None, 'version'), - (None, 'viewBox'), - (None, 'visibility'), - (None, 'width'), - (None, 'widths'), - (None, 'x'), - (None, 'x-height'), - (None, 'x1'), - (None, 'x2'), - (namespaces['xlink'], 'actuate'), - (namespaces['xlink'], 'arcrole'), - (namespaces['xlink'], 'href'), - (namespaces['xlink'], 'role'), - (namespaces['xlink'], 'show'), - (namespaces['xlink'], 'title'), - (namespaces['xlink'], 'type'), - (namespaces['xml'], 'base'), - (namespaces['xml'], 'lang'), - (namespaces['xml'], 'space'), - (None, 'y'), - (None, 'y1'), - (None, 'y2'), - (None, 'zoomAndPan'), -)) - -attr_val_is_uri = frozenset(( - (None, 'href'), - (None, 'src'), - (None, 'cite'), - (None, 'action'), - (None, 'longdesc'), - (None, 'poster'), - (None, 'background'), - (None, 'datasrc'), - (None, 'dynsrc'), - (None, 'lowsrc'), - (None, 'ping'), - (namespaces['xlink'], 'href'), - (namespaces['xml'], 'base'), -)) - -svg_attr_val_allows_ref = frozenset(( - (None, 'clip-path'), - (None, 'color-profile'), - (None, 'cursor'), - (None, 'fill'), - (None, 'filter'), - (None, 'marker'), - (None, 'marker-start'), - (None, 'marker-mid'), - (None, 'marker-end'), - (None, 'mask'), - (None, 'stroke'), -)) - -svg_allow_local_href = frozenset(( - (None, 'altGlyph'), - (None, 'animate'), - (None, 'animateColor'), - (None, 'animateMotion'), - (None, 'animateTransform'), - (None, 'cursor'), - (None, 'feImage'), - (None, 'filter'), - (None, 'linearGradient'), - (None, 'pattern'), - (None, 'radialGradient'), - (None, 'textpath'), - (None, 'tref'), - (None, 'set'), - (None, 'use') -)) - -allowed_css_properties = frozenset(( - 'azimuth', - 'background-color', - 'border-bottom-color', - 'border-collapse', - 'border-color', - 'border-left-color', - 'border-right-color', - 'border-top-color', - 'clear', - 'color', - 'cursor', - 'direction', - 'display', - 'elevation', - 'float', - 'font', - 'font-family', - 'font-size', - 'font-style', - 'font-variant', - 'font-weight', - 'height', - 'letter-spacing', - 'line-height', - 'overflow', - 'pause', - 'pause-after', - 'pause-before', - 'pitch', - 'pitch-range', - 'richness', - 'speak', - 'speak-header', - 'speak-numeral', - 'speak-punctuation', - 'speech-rate', - 'stress', - 'text-align', - 'text-decoration', - 'text-indent', - 'unicode-bidi', - 'vertical-align', - 'voice-family', - 'volume', - 'white-space', - 'width', -)) - -allowed_css_keywords = frozenset(( - 'auto', - 'aqua', - 'black', - 'block', - 'blue', - 'bold', - 'both', - 'bottom', - 'brown', - 'center', - 'collapse', - 'dashed', - 'dotted', - 'fuchsia', - 'gray', - 'green', - '!important', - 'italic', - 'left', - 'lime', - 'maroon', - 'medium', - 'none', - 'navy', - 'normal', - 'nowrap', - 'olive', - 'pointer', - 'purple', - 'red', - 'right', - 'solid', - 'silver', - 'teal', - 'top', - 'transparent', - 'underline', - 'white', - 'yellow', -)) - -allowed_svg_properties = frozenset(( - 'fill', - 'fill-opacity', - 'fill-rule', - 'stroke', - 'stroke-width', - 'stroke-linecap', - 'stroke-linejoin', - 'stroke-opacity', -)) - -allowed_protocols = frozenset(( - 'ed2k', - 'ftp', - 'http', - 'https', - 'irc', - 'mailto', - 'news', - 'gopher', - 'nntp', - 'telnet', - 'webcal', - 'xmpp', - 'callto', - 'feed', - 'urn', - 'aim', - 'rsync', - 'tag', - 'ssh', - 'sftp', - 'rtsp', - 'afs', - 'data', -)) - -allowed_content_types = frozenset(( - 'image/png', - 'image/jpeg', - 'image/gif', - 'image/webp', - 'image/bmp', - 'text/plain', -)) - - -data_content_type = re.compile(r''' - ^ - # Match a content type / - (?P[-a-zA-Z0-9.]+/[-a-zA-Z0-9.]+) - # Match any character set and encoding - (?:(?:;charset=(?:[-a-zA-Z0-9]+)(?:;(?:base64))?) - |(?:;(?:base64))?(?:;charset=(?:[-a-zA-Z0-9]+))?) - # Assume the rest is data - ,.* - $ - ''', - re.VERBOSE) - - -class Filter(base.Filter): - """Sanitizes token stream of XHTML+MathML+SVG and of inline style attributes""" - def __init__(self, - source, - allowed_elements=allowed_elements, - allowed_attributes=allowed_attributes, - allowed_css_properties=allowed_css_properties, - allowed_css_keywords=allowed_css_keywords, - allowed_svg_properties=allowed_svg_properties, - allowed_protocols=allowed_protocols, - allowed_content_types=allowed_content_types, - attr_val_is_uri=attr_val_is_uri, - svg_attr_val_allows_ref=svg_attr_val_allows_ref, - svg_allow_local_href=svg_allow_local_href): - """Creates a Filter - - :arg allowed_elements: set of elements to allow--everything else will - be escaped - - :arg allowed_attributes: set of attributes to allow in - elements--everything else will be stripped - - :arg allowed_css_properties: set of CSS properties to allow--everything - else will be stripped - - :arg allowed_css_keywords: set of CSS keywords to allow--everything - else will be stripped - - :arg allowed_svg_properties: set of SVG properties to allow--everything - else will be removed - - :arg allowed_protocols: set of allowed protocols for URIs - - :arg allowed_content_types: set of allowed content types for ``data`` URIs. - - :arg attr_val_is_uri: set of attributes that have URI values--values - that have a scheme not listed in ``allowed_protocols`` are removed - - :arg svg_attr_val_allows_ref: set of SVG attributes that can have - references - - :arg svg_allow_local_href: set of SVG elements that can have local - hrefs--these are removed - - """ - super(Filter, self).__init__(source) - self.allowed_elements = allowed_elements - self.allowed_attributes = allowed_attributes - self.allowed_css_properties = allowed_css_properties - self.allowed_css_keywords = allowed_css_keywords - self.allowed_svg_properties = allowed_svg_properties - self.allowed_protocols = allowed_protocols - self.allowed_content_types = allowed_content_types - self.attr_val_is_uri = attr_val_is_uri - self.svg_attr_val_allows_ref = svg_attr_val_allows_ref - self.svg_allow_local_href = svg_allow_local_href - - def __iter__(self): - for token in base.Filter.__iter__(self): - token = self.sanitize_token(token) - if token: - yield token - - # Sanitize the +html+, escaping all elements not in ALLOWED_ELEMENTS, and - # stripping out all attributes not in ALLOWED_ATTRIBUTES. Style attributes - # are parsed, and a restricted set, specified by ALLOWED_CSS_PROPERTIES and - # ALLOWED_CSS_KEYWORDS, are allowed through. attributes in ATTR_VAL_IS_URI - # are scanned, and only URI schemes specified in ALLOWED_PROTOCOLS are - # allowed. - # - # sanitize_html('') - # => <script> do_nasty_stuff() </script> - # sanitize_html('Click here for $100') - # => Click here for $100 - def sanitize_token(self, token): - - # accommodate filters which use token_type differently - token_type = token["type"] - if token_type in ("StartTag", "EndTag", "EmptyTag"): - name = token["name"] - namespace = token["namespace"] - if ((namespace, name) in self.allowed_elements or - (namespace is None and - (namespaces["html"], name) in self.allowed_elements)): - return self.allowed_token(token) - else: - return self.disallowed_token(token) - elif token_type == "Comment": - pass - else: - return token - - def allowed_token(self, token): - if "data" in token: - attrs = token["data"] - attr_names = set(attrs.keys()) - - # Remove forbidden attributes - for to_remove in (attr_names - self.allowed_attributes): - del token["data"][to_remove] - attr_names.remove(to_remove) - - # Remove attributes with disallowed URL values - for attr in (attr_names & self.attr_val_is_uri): - assert attr in attrs - # I don't have a clue where this regexp comes from or why it matches those - # characters, nor why we call unescape. I just know it's always been here. - # Should you be worried by this comment in a sanitizer? Yes. On the other hand, all - # this will do is remove *more* than it otherwise would. - val_unescaped = re.sub("[`\x00-\x20\x7f-\xa0\\s]+", '', - unescape(attrs[attr])).lower() - # remove replacement characters from unescaped characters - val_unescaped = val_unescaped.replace("\ufffd", "") - try: - uri = urlparse.urlparse(val_unescaped) - except ValueError: - uri = None - del attrs[attr] - if uri and uri.scheme: - if uri.scheme not in self.allowed_protocols: - del attrs[attr] - if uri.scheme == 'data': - m = data_content_type.match(uri.path) - if not m: - del attrs[attr] - elif m.group('content_type') not in self.allowed_content_types: - del attrs[attr] - - for attr in self.svg_attr_val_allows_ref: - if attr in attrs: - attrs[attr] = re.sub(r'url\s*\(\s*[^#\s][^)]+?\)', - ' ', - unescape(attrs[attr])) - if (token["name"] in self.svg_allow_local_href and - (namespaces['xlink'], 'href') in attrs and re.search(r'^\s*[^#\s].*', - attrs[(namespaces['xlink'], 'href')])): - del attrs[(namespaces['xlink'], 'href')] - if (None, 'style') in attrs: - attrs[(None, 'style')] = self.sanitize_css(attrs[(None, 'style')]) - token["data"] = attrs - return token - - def disallowed_token(self, token): - token_type = token["type"] - if token_type == "EndTag": - token["data"] = "" % token["name"] - elif token["data"]: - assert token_type in ("StartTag", "EmptyTag") - attrs = [] - for (ns, name), v in token["data"].items(): - attrs.append(' %s="%s"' % (name if ns is None else "%s:%s" % (prefixes[ns], name), escape(v))) - token["data"] = "<%s%s>" % (token["name"], ''.join(attrs)) - else: - token["data"] = "<%s>" % token["name"] - if token.get("selfClosing"): - token["data"] = token["data"][:-1] + "/>" - - token["type"] = "Characters" - - del token["name"] - return token - - def sanitize_css(self, style): - # disallow urls - style = re.compile(r'url\s*\(\s*[^\s)]+?\s*\)\s*').sub(' ', style) - - # gauntlet - if not re.match(r"""^([:,;#%.\sa-zA-Z0-9!]|\w-\w|'[\s\w]+'|"[\s\w]+"|\([\d,\s]+\))*$""", style): - return '' - if not re.match(r"^\s*([-\w]+\s*:[^:;]*(;\s*|$))*$", style): - return '' - - clean = [] - for prop, value in re.findall(r"([-\w]+)\s*:\s*([^:;]*)", style): - if not value: - continue - if prop.lower() in self.allowed_css_properties: - clean.append(prop + ': ' + value + ';') - elif prop.split('-')[0].lower() in ['background', 'border', 'margin', - 'padding']: - for keyword in value.split(): - if keyword not in self.allowed_css_keywords and \ - not re.match(r"^(#[0-9a-fA-F]+|rgb\(\d+%?,\d*%?,?\d*%?\)?|\d{0,2}\.?\d{0,2}(cm|em|ex|in|mm|pc|pt|px|%|,|\))?)$", keyword): # noqa - break - else: - clean.append(prop + ': ' + value + ';') - elif prop.lower() in self.allowed_svg_properties: - clean.append(prop + ': ' + value + ';') - - return ' '.join(clean) diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib/filters/whitespace.py b/src/fetchcode/vcs/pip/_vendor/html5lib/filters/whitespace.py deleted file mode 100644 index 0d12584b..00000000 --- a/src/fetchcode/vcs/pip/_vendor/html5lib/filters/whitespace.py +++ /dev/null @@ -1,38 +0,0 @@ -from __future__ import absolute_import, division, unicode_literals - -import re - -from . import base -from ..constants import rcdataElements, spaceCharacters -spaceCharacters = "".join(spaceCharacters) - -SPACES_REGEX = re.compile("[%s]+" % spaceCharacters) - - -class Filter(base.Filter): - """Collapses whitespace except in pre, textarea, and script elements""" - spacePreserveElements = frozenset(["pre", "textarea"] + list(rcdataElements)) - - def __iter__(self): - preserve = 0 - for token in base.Filter.__iter__(self): - type = token["type"] - if type == "StartTag" \ - and (preserve or token["name"] in self.spacePreserveElements): - preserve += 1 - - elif type == "EndTag" and preserve: - preserve -= 1 - - elif not preserve and type == "SpaceCharacters" and token["data"]: - # Test on token["data"] above to not introduce spaces where there were not - token["data"] = " " - - elif not preserve and type == "Characters": - token["data"] = collapse_spaces(token["data"]) - - yield token - - -def collapse_spaces(text): - return SPACES_REGEX.sub(' ', text) diff --git a/src/fetchcode/vcs/pip/_vendor/html5lib/html5parser.py b/src/fetchcode/vcs/pip/_vendor/html5lib/html5parser.py deleted file mode 100644 index 8f691e58..00000000 --- a/src/fetchcode/vcs/pip/_vendor/html5lib/html5parser.py +++ /dev/null @@ -1,2791 +0,0 @@ -from __future__ import absolute_import, division, unicode_literals -from fetchcode.vcs.pip._vendor.six import with_metaclass, viewkeys - -import types -from collections import OrderedDict - -from . import _inputstream -from . import _tokenizer - -from . import treebuilders -from .treebuilders.base import Marker - -from . import _utils -from .constants import ( - spaceCharacters, asciiUpper2Lower, - specialElements, headingElements, cdataElements, rcdataElements, - tokenTypes, tagTokenTypes, - namespaces, - htmlIntegrationPointElements, mathmlTextIntegrationPointElements, - adjustForeignAttributes as adjustForeignAttributesMap, - adjustMathMLAttributes, adjustSVGAttributes, - E, - _ReparseException -) - - -def parse(doc, treebuilder="etree", namespaceHTMLElements=True, **kwargs): - """Parse an HTML document as a string or file-like object into a tree - - :arg doc: the document to parse as a string or file-like object - - :arg treebuilder: the treebuilder to use when parsing - - :arg namespaceHTMLElements: whether or not to namespace HTML elements - - :returns: parsed tree - - Example: - - >>> from html5lib.html5parser import parse - >>> parse('

This is a doc

') - - - """ - tb = treebuilders.getTreeBuilder(treebuilder) - p = HTMLParser(tb, namespaceHTMLElements=namespaceHTMLElements) - return p.parse(doc, **kwargs) - - -def parseFragment(doc, container="div", treebuilder="etree", namespaceHTMLElements=True, **kwargs): - """Parse an HTML fragment as a string or file-like object into a tree - - :arg doc: the fragment to parse as a string or file-like object - - :arg container: the container context to parse the fragment in - - :arg treebuilder: the treebuilder to use when parsing - - :arg namespaceHTMLElements: whether or not to namespace HTML elements - - :returns: parsed tree - - Example: - - >>> from html5lib.html5libparser import parseFragment - >>> parseFragment('this is a fragment') - - - """ - tb = treebuilders.getTreeBuilder(treebuilder) - p = HTMLParser(tb, namespaceHTMLElements=namespaceHTMLElements) - return p.parseFragment(doc, container=container, **kwargs) - - -def method_decorator_metaclass(function): - class Decorated(type): - def __new__(meta, classname, bases, classDict): - for attributeName, attribute in classDict.items(): - if isinstance(attribute, types.FunctionType): - attribute = function(attribute) - - classDict[attributeName] = attribute - return type.__new__(meta, classname, bases, classDict) - return Decorated - - -class HTMLParser(object): - """HTML parser - - Generates a tree structure from a stream of (possibly malformed) HTML. - - """ - - def __init__(self, tree=None, strict=False, namespaceHTMLElements=True, debug=False): - """ - :arg tree: a treebuilder class controlling the type of tree that will be - returned. Built in treebuilders can be accessed through - html5lib.treebuilders.getTreeBuilder(treeType) - - :arg strict: raise an exception when a parse error is encountered - - :arg namespaceHTMLElements: whether or not to namespace HTML elements - - :arg debug: whether or not to enable debug mode which logs things - - Example: - - >>> from html5lib.html5parser import HTMLParser - >>> parser = HTMLParser() # generates parser with etree builder - >>> parser = HTMLParser('lxml', strict=True) # generates parser with lxml builder which is strict - - """ - - # Raise an exception on the first error encountered - self.strict = strict - - if tree is None: - tree = treebuilders.getTreeBuilder("etree") - self.tree = tree(namespaceHTMLElements) - self.errors = [] - - self.phases = dict([(name, cls(self, self.tree)) for name, cls in - getPhases(debug).items()]) - - def _parse(self, stream, innerHTML=False, container="div", scripting=False, **kwargs): - - self.innerHTMLMode = innerHTML - self.container = container - self.scripting = scripting - self.tokenizer = _tokenizer.HTMLTokenizer(stream, parser=self, **kwargs) - self.reset() - - try: - self.mainLoop() - except _ReparseException: - self.reset() - self.mainLoop() - - def reset(self): - self.tree.reset() - self.firstStartTag = False - self.errors = [] - self.log = [] # only used with debug mode - # "quirks" / "limited quirks" / "no quirks" - self.compatMode = "no quirks" - - if self.innerHTMLMode: - self.innerHTML = self.container.lower() - - if self.innerHTML in cdataElements: - self.tokenizer.state = self.tokenizer.rcdataState - elif self.innerHTML in rcdataElements: - self.tokenizer.state = self.tokenizer.rawtextState - elif self.innerHTML == 'plaintext': - self.tokenizer.state = self.tokenizer.plaintextState - else: - # state already is data state - # self.tokenizer.state = self.tokenizer.dataState - pass - self.phase = self.phases["beforeHtml"] - self.phase.insertHtmlElement() - self.resetInsertionMode() - else: - self.innerHTML = False # pylint:disable=redefined-variable-type - self.phase = self.phases["initial"] - - self.lastPhase = None - - self.beforeRCDataPhase = None - - self.framesetOK = True - - @property - def documentEncoding(self): - """Name of the character encoding that was used to decode the input stream, or - :obj:`None` if that is not determined yet - - """ - if not hasattr(self, 'tokenizer'): - return None - return self.tokenizer.stream.charEncoding[0].name - - def isHTMLIntegrationPoint(self, element): - if (element.name == "annotation-xml" and - element.namespace == namespaces["mathml"]): - return ("encoding" in element.attributes and - element.attributes["encoding"].translate( - asciiUpper2Lower) in - ("text/html", "application/xhtml+xml")) - else: - return (element.namespace, element.name) in htmlIntegrationPointElements - - def isMathMLTextIntegrationPoint(self, element): - return (element.namespace, element.name) in mathmlTextIntegrationPointElements - - def mainLoop(self): - CharactersToken = tokenTypes["Characters"] - SpaceCharactersToken = tokenTypes["SpaceCharacters"] - StartTagToken = tokenTypes["StartTag"] - EndTagToken = tokenTypes["EndTag"] - CommentToken = tokenTypes["Comment"] - DoctypeToken = tokenTypes["Doctype"] - ParseErrorToken = tokenTypes["ParseError"] - - for token in self.normalizedTokens(): - prev_token = None - new_token = token - while new_token is not None: - prev_token = new_token - currentNode = self.tree.openElements[-1] if self.tree.openElements else None - currentNodeNamespace = currentNode.namespace if currentNode else None - currentNodeName = currentNode.name if currentNode else None - - type = new_token["type"] - - if type == ParseErrorToken: - self.parseError(new_token["data"], new_token.get("datavars", {})) - new_token = None - else: - if (len(self.tree.openElements) == 0 or - currentNodeNamespace == self.tree.defaultNamespace or - (self.isMathMLTextIntegrationPoint(currentNode) and - ((type == StartTagToken and - token["name"] not in frozenset(["mglyph", "malignmark"])) or - type in (CharactersToken, SpaceCharactersToken))) or - (currentNodeNamespace == namespaces["mathml"] and - currentNodeName == "annotation-xml" and - type == StartTagToken and - token["name"] == "svg") or - (self.isHTMLIntegrationPoint(currentNode) and - type in (StartTagToken, CharactersToken, SpaceCharactersToken))): - phase = self.phase - else: - phase = self.phases["inForeignContent"] - - if type == CharactersToken: - new_token = phase.processCharacters(new_token) - elif type == SpaceCharactersToken: - new_token = phase.processSpaceCharacters(new_token) - elif type == StartTagToken: - new_token = phase.processStartTag(new_token) - elif type == EndTagToken: - new_token = phase.processEndTag(new_token) - elif type == CommentToken: - new_token = phase.processComment(new_token) - elif type == DoctypeToken: - new_token = phase.processDoctype(new_token) - - if (type == StartTagToken and prev_token["selfClosing"] and - not prev_token["selfClosingAcknowledged"]): - self.parseError("non-void-element-with-trailing-solidus", - {"name": prev_token["name"]}) - - # When the loop finishes it's EOF - reprocess = True - phases = [] - while reprocess: - phases.append(self.phase) - reprocess = self.phase.processEOF() - if reprocess: - assert self.phase not in phases - - def normalizedTokens(self): - for token in self.tokenizer: - yield self.normalizeToken(token) - - def parse(self, stream, *args, **kwargs): - """Parse a HTML document into a well-formed tree - - :arg stream: a file-like object or string containing the HTML to be parsed - - The optional encoding parameter must be a string that indicates - the encoding. If specified, that encoding will be used, - regardless of any BOM or later declaration (such as in a meta - element). - - :arg scripting: treat noscript elements as if JavaScript was turned on - - :returns: parsed tree - - Example: - - >>> from html5lib.html5parser import HTMLParser - >>> parser = HTMLParser() - >>> parser.parse('

This is a doc

') - - - """ - self._parse(stream, False, None, *args, **kwargs) - return self.tree.getDocument() - - def parseFragment(self, stream, *args, **kwargs): - """Parse a HTML fragment into a well-formed tree fragment - - :arg container: name of the element we're setting the innerHTML - property if set to None, default to 'div' - - :arg stream: a file-like object or string containing the HTML to be parsed - - The optional encoding parameter must be a string that indicates - the encoding. If specified, that encoding will be used, - regardless of any BOM or later declaration (such as in a meta - element) - - :arg scripting: treat noscript elements as if JavaScript was turned on - - :returns: parsed tree - - Example: - - >>> from html5lib.html5libparser import HTMLParser - >>> parser = HTMLParser() - >>> parser.parseFragment('this is a fragment') - - - """ - self._parse(stream, True, *args, **kwargs) - return self.tree.getFragment() - - def parseError(self, errorcode="XXX-undefined-error", datavars=None): - # XXX The idea is to make errorcode mandatory. - if datavars is None: - datavars = {} - self.errors.append((self.tokenizer.stream.position(), errorcode, datavars)) - if self.strict: - raise ParseError(E[errorcode] % datavars) - - def normalizeToken(self, token): - # HTML5 specific normalizations to the token stream - if token["type"] == tokenTypes["StartTag"]: - raw = token["data"] - token["data"] = OrderedDict(raw) - if len(raw) > len(token["data"]): - # we had some duplicated attribute, fix so first wins - token["data"].update(raw[::-1]) - - return token - - def adjustMathMLAttributes(self, token): - adjust_attributes(token, adjustMathMLAttributes) - - def adjustSVGAttributes(self, token): - adjust_attributes(token, adjustSVGAttributes) - - def adjustForeignAttributes(self, token): - adjust_attributes(token, adjustForeignAttributesMap) - - def reparseTokenNormal(self, token): - # pylint:disable=unused-argument - self.parser.phase() - - def resetInsertionMode(self): - # The name of this method is mostly historical. (It's also used in the - # specification.) - last = False - newModes = { - "select": "inSelect", - "td": "inCell", - "th": "inCell", - "tr": "inRow", - "tbody": "inTableBody", - "thead": "inTableBody", - "tfoot": "inTableBody", - "caption": "inCaption", - "colgroup": "inColumnGroup", - "table": "inTable", - "head": "inBody", - "body": "inBody", - "frameset": "inFrameset", - "html": "beforeHead" - } - for node in self.tree.openElements[::-1]: - nodeName = node.name - new_phase = None - if node == self.tree.openElements[0]: - assert self.innerHTML - last = True - nodeName = self.innerHTML - # Check for conditions that should only happen in the innerHTML - # case - if nodeName in ("select", "colgroup", "head", "html"): - assert self.innerHTML - - if not last and node.namespace != self.tree.defaultNamespace: - continue - - if nodeName in newModes: - new_phase = self.phases[newModes[nodeName]] - break - elif last: - new_phase = self.phases["inBody"] - break - - self.phase = new_phase - - def parseRCDataRawtext(self, token, contentType): - # Generic RCDATA/RAWTEXT Parsing algorithm - assert contentType in ("RAWTEXT", "RCDATA") - - self.tree.insertElement(token) - - if contentType == "RAWTEXT": - self.tokenizer.state = self.tokenizer.rawtextState - else: - self.tokenizer.state = self.tokenizer.rcdataState - - self.originalPhase = self.phase - - self.phase = self.phases["text"] - - -@_utils.memoize -def getPhases(debug): - def log(function): - """Logger that records which phase processes each token""" - type_names = dict((value, key) for key, value in - tokenTypes.items()) - - def wrapped(self, *args, **kwargs): - if function.__name__.startswith("process") and len(args) > 0: - token = args[0] - try: - info = {"type": type_names[token['type']]} - except: - raise - if token['type'] in tagTokenTypes: - info["name"] = token['name'] - - self.parser.log.append((self.parser.tokenizer.state.__name__, - self.parser.phase.__class__.__name__, - self.__class__.__name__, - function.__name__, - info)) - return function(self, *args, **kwargs) - else: - return function(self, *args, **kwargs) - return wrapped - - def getMetaclass(use_metaclass, metaclass_func): - if use_metaclass: - return method_decorator_metaclass(metaclass_func) - else: - return type - - # pylint:disable=unused-argument - class Phase(with_metaclass(getMetaclass(debug, log))): - """Base class for helper object that implements each phase of processing - """ - - def __init__(self, parser, tree): - self.parser = parser - self.tree = tree - - def processEOF(self): - raise NotImplementedError - - def processComment(self, token): - # For most phases the following is correct. Where it's not it will be - # overridden. - self.tree.insertComment(token, self.tree.openElements[-1]) - - def processDoctype(self, token): - self.parser.parseError("unexpected-doctype") - - def processCharacters(self, token): - self.tree.insertText(token["data"]) - - def processSpaceCharacters(self, token): - self.tree.insertText(token["data"]) - - def processStartTag(self, token): - return self.startTagHandler[token["name"]](token) - - def startTagHtml(self, token): - if not self.parser.firstStartTag and token["name"] == "html": - self.parser.parseError("non-html-root") - # XXX Need a check here to see if the first start tag token emitted is - # this token... If it's not, invoke self.parser.parseError(). - for attr, value in token["data"].items(): - if attr not in self.tree.openElements[0].attributes: - self.tree.openElements[0].attributes[attr] = value - self.parser.firstStartTag = False - - def processEndTag(self, token): - return self.endTagHandler[token["name"]](token) - - class InitialPhase(Phase): - def processSpaceCharacters(self, token): - pass - - def processComment(self, token): - self.tree.insertComment(token, self.tree.document) - - def processDoctype(self, token): - name = token["name"] - publicId = token["publicId"] - systemId = token["systemId"] - correct = token["correct"] - - if (name != "html" or publicId is not None or - systemId is not None and systemId != "about:legacy-compat"): - self.parser.parseError("unknown-doctype") - - if publicId is None: - publicId = "" - - self.tree.insertDoctype(token) - - if publicId != "": - publicId = publicId.translate(asciiUpper2Lower) - - if (not correct or token["name"] != "html" or - publicId.startswith( - ("+//silmaril//dtd html pro v0r11 19970101//", - "-//advasoft ltd//dtd html 3.0 aswedit + extensions//", - "-//as//dtd html 3.0 aswedit + extensions//", - "-//ietf//dtd html 2.0 level 1//", - "-//ietf//dtd html 2.0 level 2//", - "-//ietf//dtd html 2.0 strict level 1//", - "-//ietf//dtd html 2.0 strict level 2//", - "-//ietf//dtd html 2.0 strict//", - "-//ietf//dtd html 2.0//", - "-//ietf//dtd html 2.1e//", - "-//ietf//dtd html 3.0//", - "-//ietf//dtd html 3.2 final//", - "-//ietf//dtd html 3.2//", - "-//ietf//dtd html 3//", - "-//ietf//dtd html level 0//", - "-//ietf//dtd html level 1//", - "-//ietf//dtd html level 2//", - "-//ietf//dtd html level 3//", - "-//ietf//dtd html strict level 0//", - "-//ietf//dtd html strict level 1//", - "-//ietf//dtd html strict level 2//", - "-//ietf//dtd html strict level 3//", - "-//ietf//dtd html strict//", - "-//ietf//dtd html//", - "-//metrius//dtd metrius presentational//", - "-//microsoft//dtd internet explorer 2.0 html strict//", - "-//microsoft//dtd internet explorer 2.0 html//", - "-//microsoft//dtd internet explorer 2.0 tables//", - "-//microsoft//dtd internet explorer 3.0 html strict//", - "-//microsoft//dtd internet explorer 3.0 html//", - "-//microsoft//dtd internet explorer 3.0 tables//", - "-//netscape comm. corp.//dtd html//", - "-//netscape comm. corp.//dtd strict html//", - "-//o'reilly and associates//dtd html 2.0//", - "-//o'reilly and associates//dtd html extended 1.0//", - "-//o'reilly and associates//dtd html extended relaxed 1.0//", - "-//softquad software//dtd hotmetal pro 6.0::19990601::extensions to html 4.0//", - "-//softquad//dtd hotmetal pro 4.0::19971010::extensions to html 4.0//", - "-//spyglass//dtd html 2.0 extended//", - "-//sq//dtd html 2.0 hotmetal + extensions//", - "-//sun microsystems corp.//dtd hotjava html//", - "-//sun microsystems corp.//dtd hotjava strict html//", - "-//w3c//dtd html 3 1995-03-24//", - "-//w3c//dtd html 3.2 draft//", - "-//w3c//dtd html 3.2 final//", - "-//w3c//dtd html 3.2//", - "-//w3c//dtd html 3.2s draft//", - "-//w3c//dtd html 4.0 frameset//", - "-//w3c//dtd html 4.0 transitional//", - "-//w3c//dtd html experimental 19960712//", - "-//w3c//dtd html experimental 970421//", - "-//w3c//dtd w3 html//", - "-//w3o//dtd w3 html 3.0//", - "-//webtechs//dtd mozilla html 2.0//", - "-//webtechs//dtd mozilla html//")) or - publicId in ("-//w3o//dtd w3 html strict 3.0//en//", - "-/w3c/dtd html 4.0 transitional/en", - "html") or - publicId.startswith( - ("-//w3c//dtd html 4.01 frameset//", - "-//w3c//dtd html 4.01 transitional//")) and - systemId is None or - systemId and systemId.lower() == "http://www.ibm.com/data/dtd/v11/ibmxhtml1-transitional.dtd"): - self.parser.compatMode = "quirks" - elif (publicId.startswith( - ("-//w3c//dtd xhtml 1.0 frameset//", - "-//w3c//dtd xhtml 1.0 transitional//")) or - publicId.startswith( - ("-//w3c//dtd html 4.01 frameset//", - "-//w3c//dtd html 4.01 transitional//")) and - systemId is not None): - self.parser.compatMode = "limited quirks" - - self.parser.phase = self.parser.phases["beforeHtml"] - - def anythingElse(self): - self.parser.compatMode = "quirks" - self.parser.phase = self.parser.phases["beforeHtml"] - - def processCharacters(self, token): - self.parser.parseError("expected-doctype-but-got-chars") - self.anythingElse() - return token - - def processStartTag(self, token): - self.parser.parseError("expected-doctype-but-got-start-tag", - {"name": token["name"]}) - self.anythingElse() - return token - - def processEndTag(self, token): - self.parser.parseError("expected-doctype-but-got-end-tag", - {"name": token["name"]}) - self.anythingElse() - return token - - def processEOF(self): - self.parser.parseError("expected-doctype-but-got-eof") - self.anythingElse() - return True - - class BeforeHtmlPhase(Phase): - # helper methods - def insertHtmlElement(self): - self.tree.insertRoot(impliedTagToken("html", "StartTag")) - self.parser.phase = self.parser.phases["beforeHead"] - - # other - def processEOF(self): - self.insertHtmlElement() - return True - - def processComment(self, token): - self.tree.insertComment(token, self.tree.document) - - def processSpaceCharacters(self, token): - pass - - def processCharacters(self, token): - self.insertHtmlElement() - return token - - def processStartTag(self, token): - if token["name"] == "html": - self.parser.firstStartTag = True - self.insertHtmlElement() - return token - - def processEndTag(self, token): - if token["name"] not in ("head", "body", "html", "br"): - self.parser.parseError("unexpected-end-tag-before-html", - {"name": token["name"]}) - else: - self.insertHtmlElement() - return token - - class BeforeHeadPhase(Phase): - def __init__(self, parser, tree): - Phase.__init__(self, parser, tree) - - self.startTagHandler = _utils.MethodDispatcher([ - ("html", self.startTagHtml), - ("head", self.startTagHead) - ]) - self.startTagHandler.default = self.startTagOther - - self.endTagHandler = _utils.MethodDispatcher([ - (("head", "body", "html", "br"), self.endTagImplyHead) - ]) - self.endTagHandler.default = self.endTagOther - - def processEOF(self): - self.startTagHead(impliedTagToken("head", "StartTag")) - return True - - def processSpaceCharacters(self, token): - pass - - def processCharacters(self, token): - self.startTagHead(impliedTagToken("head", "StartTag")) - return token - - def startTagHtml(self, token): - return self.parser.phases["inBody"].processStartTag(token) - - def startTagHead(self, token): - self.tree.insertElement(token) - self.tree.headPointer = self.tree.openElements[-1] - self.parser.phase = self.parser.phases["inHead"] - - def startTagOther(self, token): - self.startTagHead(impliedTagToken("head", "StartTag")) - return token - - def endTagImplyHead(self, token): - self.startTagHead(impliedTagToken("head", "StartTag")) - return token - - def endTagOther(self, token): - self.parser.parseError("end-tag-after-implied-root", - {"name": token["name"]}) - - class InHeadPhase(Phase): - def __init__(self, parser, tree): - Phase.__init__(self, parser, tree) - - self.startTagHandler = _utils.MethodDispatcher([ - ("html", self.startTagHtml), - ("title", self.startTagTitle), - (("noframes", "style"), self.startTagNoFramesStyle), - ("noscript", self.startTagNoscript), - ("script", self.startTagScript), - (("base", "basefont", "bgsound", "command", "link"), - self.startTagBaseLinkCommand), - ("meta", self.startTagMeta), - ("head", self.startTagHead) - ]) - self.startTagHandler.default = self.startTagOther - - self.endTagHandler = _utils.MethodDispatcher([ - ("head", self.endTagHead), - (("br", "html", "body"), self.endTagHtmlBodyBr) - ]) - self.endTagHandler.default = self.endTagOther - - # the real thing - def processEOF(self): - self.anythingElse() - return True - - def processCharacters(self, token): - self.anythingElse() - return token - - def startTagHtml(self, token): - return self.parser.phases["inBody"].processStartTag(token) - - def startTagHead(self, token): - self.parser.parseError("two-heads-are-not-better-than-one") - - def startTagBaseLinkCommand(self, token): - self.tree.insertElement(token) - self.tree.openElements.pop() - token["selfClosingAcknowledged"] = True - - def startTagMeta(self, token): - self.tree.insertElement(token) - self.tree.openElements.pop() - token["selfClosingAcknowledged"] = True - - attributes = token["data"] - if self.parser.tokenizer.stream.charEncoding[1] == "tentative": - if "charset" in attributes: - self.parser.tokenizer.stream.changeEncoding(attributes["charset"]) - elif ("content" in attributes and - "http-equiv" in attributes and - attributes["http-equiv"].lower() == "content-type"): - # Encoding it as UTF-8 here is a hack, as really we should pass - # the abstract Unicode string, and just use the - # ContentAttrParser on that, but using UTF-8 allows all chars - # to be encoded and as a ASCII-superset works. - data = _inputstream.EncodingBytes(attributes["content"].encode("utf-8")) - parser = _inputstream.ContentAttrParser(data) - codec = parser.parse() - self.parser.tokenizer.stream.changeEncoding(codec) - - def startTagTitle(self, token): - self.parser.parseRCDataRawtext(token, "RCDATA") - - def startTagNoFramesStyle(self, token): - # Need to decide whether to implement the scripting-disabled case - self.parser.parseRCDataRawtext(token, "RAWTEXT") - - def startTagNoscript(self, token): - if self.parser.scripting: - self.parser.parseRCDataRawtext(token, "RAWTEXT") - else: - self.tree.insertElement(token) - self.parser.phase = self.parser.phases["inHeadNoscript"] - - def startTagScript(self, token): - self.tree.insertElement(token) - self.parser.tokenizer.state = self.parser.tokenizer.scriptDataState - self.parser.originalPhase = self.parser.phase - self.parser.phase = self.parser.phases["text"] - - def startTagOther(self, token): - self.anythingElse() - return token - - def endTagHead(self, token): - node = self.parser.tree.openElements.pop() - assert node.name == "head", "Expected head got %s" % node.name - self.parser.phase = self.parser.phases["afterHead"] - - def endTagHtmlBodyBr(self, token): - self.anythingElse() - return token - - def endTagOther(self, token): - self.parser.parseError("unexpected-end-tag", {"name": token["name"]}) - - def anythingElse(self): - self.endTagHead(impliedTagToken("head")) - - class InHeadNoscriptPhase(Phase): - def __init__(self, parser, tree): - Phase.__init__(self, parser, tree) - - self.startTagHandler = _utils.MethodDispatcher([ - ("html", self.startTagHtml), - (("basefont", "bgsound", "link", "meta", "noframes", "style"), self.startTagBaseLinkCommand), - (("head", "noscript"), self.startTagHeadNoscript), - ]) - self.startTagHandler.default = self.startTagOther - - self.endTagHandler = _utils.MethodDispatcher([ - ("noscript", self.endTagNoscript), - ("br", self.endTagBr), - ]) - self.endTagHandler.default = self.endTagOther - - def processEOF(self): - self.parser.parseError("eof-in-head-noscript") - self.anythingElse() - return True - - def processComment(self, token): - return self.parser.phases["inHead"].processComment(token) - - def processCharacters(self, token): - self.parser.parseError("char-in-head-noscript") - self.anythingElse() - return token - - def processSpaceCharacters(self, token): - return self.parser.phases["inHead"].processSpaceCharacters(token) - - def startTagHtml(self, token): - return self.parser.phases["inBody"].processStartTag(token) - - def startTagBaseLinkCommand(self, token): - return self.parser.phases["inHead"].processStartTag(token) - - def startTagHeadNoscript(self, token): - self.parser.parseError("unexpected-start-tag", {"name": token["name"]}) - - def startTagOther(self, token): - self.parser.parseError("unexpected-inhead-noscript-tag", {"name": token["name"]}) - self.anythingElse() - return token - - def endTagNoscript(self, token): - node = self.parser.tree.openElements.pop() - assert node.name == "noscript", "Expected noscript got %s" % node.name - self.parser.phase = self.parser.phases["inHead"] - - def endTagBr(self, token): - self.parser.parseError("unexpected-inhead-noscript-tag", {"name": token["name"]}) - self.anythingElse() - return token - - def endTagOther(self, token): - self.parser.parseError("unexpected-end-tag", {"name": token["name"]}) - - def anythingElse(self): - # Caller must raise parse error first! - self.endTagNoscript(impliedTagToken("noscript")) - - class AfterHeadPhase(Phase): - def __init__(self, parser, tree): - Phase.__init__(self, parser, tree) - - self.startTagHandler = _utils.MethodDispatcher([ - ("html", self.startTagHtml), - ("body", self.startTagBody), - ("frameset", self.startTagFrameset), - (("base", "basefont", "bgsound", "link", "meta", "noframes", "script", - "style", "title"), - self.startTagFromHead), - ("head", self.startTagHead) - ]) - self.startTagHandler.default = self.startTagOther - self.endTagHandler = _utils.MethodDispatcher([(("body", "html", "br"), - self.endTagHtmlBodyBr)]) - self.endTagHandler.default = self.endTagOther - - def processEOF(self): - self.anythingElse() - return True - - def processCharacters(self, token): - self.anythingElse() - return token - - def startTagHtml(self, token): - return self.parser.phases["inBody"].processStartTag(token) - - def startTagBody(self, token): - self.parser.framesetOK = False - self.tree.insertElement(token) - self.parser.phase = self.parser.phases["inBody"] - - def startTagFrameset(self, token): - self.tree.insertElement(token) - self.parser.phase = self.parser.phases["inFrameset"] - - def startTagFromHead(self, token): - self.parser.parseError("unexpected-start-tag-out-of-my-head", - {"name": token["name"]}) - self.tree.openElements.append(self.tree.headPointer) - self.parser.phases["inHead"].processStartTag(token) - for node in self.tree.openElements[::-1]: - if node.name == "head": - self.tree.openElements.remove(node) - break - - def startTagHead(self, token): - self.parser.parseError("unexpected-start-tag", {"name": token["name"]}) - - def startTagOther(self, token): - self.anythingElse() - return token - - def endTagHtmlBodyBr(self, token): - self.anythingElse() - return token - - def endTagOther(self, token): - self.parser.parseError("unexpected-end-tag", {"name": token["name"]}) - - def anythingElse(self): - self.tree.insertElement(impliedTagToken("body", "StartTag")) - self.parser.phase = self.parser.phases["inBody"] - self.parser.framesetOK = True - - class InBodyPhase(Phase): - # http://www.whatwg.org/specs/web-apps/current-work/#parsing-main-inbody - # the really-really-really-very crazy mode - def __init__(self, parser, tree): - Phase.__init__(self, parser, tree) - - # Set this to the default handler - self.processSpaceCharacters = self.processSpaceCharactersNonPre - - self.startTagHandler = _utils.MethodDispatcher([ - ("html", self.startTagHtml), - (("base", "basefont", "bgsound", "command", "link", "meta", - "script", "style", "title"), - self.startTagProcessInHead), - ("body", self.startTagBody), - ("frameset", self.startTagFrameset), - (("address", "article", "aside", "blockquote", "center", "details", - "dir", "div", "dl", "fieldset", "figcaption", "figure", - "footer", "header", "hgroup", "main", "menu", "nav", "ol", "p", - "section", "summary", "ul"), - self.startTagCloseP), - (headingElements, self.startTagHeading), - (("pre", "listing"), self.startTagPreListing), - ("form", self.startTagForm), - (("li", "dd", "dt"), self.startTagListItem), - ("plaintext", self.startTagPlaintext), - ("a", self.startTagA), - (("b", "big", "code", "em", "font", "i", "s", "small", "strike", - "strong", "tt", "u"), self.startTagFormatting), - ("nobr", self.startTagNobr), - ("button", self.startTagButton), - (("applet", "marquee", "object"), self.startTagAppletMarqueeObject), - ("xmp", self.startTagXmp), - ("table", self.startTagTable), - (("area", "br", "embed", "img", "keygen", "wbr"), - self.startTagVoidFormatting), - (("param", "source", "track"), self.startTagParamSource), - ("input", self.startTagInput), - ("hr", self.startTagHr), - ("image", self.startTagImage), - ("isindex", self.startTagIsIndex), - ("textarea", self.startTagTextarea), - ("iframe", self.startTagIFrame), - ("noscript", self.startTagNoscript), - (("noembed", "noframes"), self.startTagRawtext), - ("select", self.startTagSelect), - (("rp", "rt"), self.startTagRpRt), - (("option", "optgroup"), self.startTagOpt), - (("math"), self.startTagMath), - (("svg"), self.startTagSvg), - (("caption", "col", "colgroup", "frame", "head", - "tbody", "td", "tfoot", "th", "thead", - "tr"), self.startTagMisplaced) - ]) - self.startTagHandler.default = self.startTagOther - - self.endTagHandler = _utils.MethodDispatcher([ - ("body", self.endTagBody), - ("html", self.endTagHtml), - (("address", "article", "aside", "blockquote", "button", "center", - "details", "dialog", "dir", "div", "dl", "fieldset", "figcaption", "figure", - "footer", "header", "hgroup", "listing", "main", "menu", "nav", "ol", "pre", - "section", "summary", "ul"), self.endTagBlock), - ("form", self.endTagForm), - ("p", self.endTagP), - (("dd", "dt", "li"), self.endTagListItem), - (headingElements, self.endTagHeading), - (("a", "b", "big", "code", "em", "font", "i", "nobr", "s", "small", - "strike", "strong", "tt", "u"), self.endTagFormatting), - (("applet", "marquee", "object"), self.endTagAppletMarqueeObject), - ("br", self.endTagBr), - ]) - self.endTagHandler.default = self.endTagOther - - def isMatchingFormattingElement(self, node1, node2): - return (node1.name == node2.name and - node1.namespace == node2.namespace and - node1.attributes == node2.attributes) - - # helper - def addFormattingElement(self, token): - self.tree.insertElement(token) - element = self.tree.openElements[-1] - - matchingElements = [] - for node in self.tree.activeFormattingElements[::-1]: - if node is Marker: - break - elif self.isMatchingFormattingElement(node, element): - matchingElements.append(node) - - assert len(matchingElements) <= 3 - if len(matchingElements) == 3: - self.tree.activeFormattingElements.remove(matchingElements[-1]) - self.tree.activeFormattingElements.append(element) - - # the real deal - def processEOF(self): - allowed_elements = frozenset(("dd", "dt", "li", "p", "tbody", "td", - "tfoot", "th", "thead", "tr", "body", - "html")) - for node in self.tree.openElements[::-1]: - if node.name not in allowed_elements: - self.parser.parseError("expected-closing-tag-but-got-eof") - break - # Stop parsing - - def processSpaceCharactersDropNewline(self, token): - # Sometimes (start of
, , and