diff --git a/poetry.lock b/poetry.lock index 6d73e34..8e348af 100644 --- a/poetry.lock +++ b/poetry.lock @@ -5701,13 +5701,13 @@ stats = ["scipy (>=1.7)", "statsmodels (>=0.12)"] [[package]] name = "selenium" -version = "4.14.0" +version = "4.15.0" description = "" optional = false python-versions = ">=3.8" files = [ - {file = "selenium-4.14.0-py3-none-any.whl", hash = "sha256:be9824a9354a7fe288e3fad9ceb6a9c65ddc7c44545d23ad0ebf4ce202b19893"}, - {file = "selenium-4.14.0.tar.gz", hash = "sha256:0d14b0d9842366f38fb5f8f842cf7c042bcfa062affc6a0a86e4d634bdd0fe54"}, + {file = "selenium-4.15.0-py3-none-any.whl", hash = "sha256:c566dd3b20765dad64e65edca19a52f421f601ed1739f87dd4c5c07aae5dae6f"}, + {file = "selenium-4.15.0.tar.gz", hash = "sha256:1d339cb4577a2c617122ebe6342b7e9bca4cb4588a2d322c898f5df29c91df02"}, ] [package.dependencies] @@ -7426,4 +7426,4 @@ web-server = ["dash", "dash-auth", "dash-bootstrap-components", "matplotlib", "s [metadata] lock-version = "2.0" python-versions = "^3.11" -content-hash = "dfb1bd40b554a5522a9db4fbd991a7038b4ed6479ad1e1fcbf40dfe7168a3e3d" +content-hash = "a6ba453d93d19326e80413983acc11f3307363c14e863c18cd0eace421de70b0" diff --git a/pyproject.toml b/pyproject.toml index 7a233a3..d515a77 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -67,7 +67,7 @@ python = "^3.11" python-dotenv = "^1.0.0" rapidfuzz = "^3.5.1" seaborn = "^0.13.0" -selenium = "^4.12.0" +selenium = "^4.15.0" spacy = "^3.6.1" spacy-sentiws = "^3.0.0" torch = {version = "*", source = "torch-cpu"} diff --git a/src/aki_prj23_transparenzregister/utils/data_extraction/unternehmensregister/extract.py b/src/aki_prj23_transparenzregister/utils/data_extraction/unternehmensregister/extract.py index 6fd3174..c37b260 100644 --- a/src/aki_prj23_transparenzregister/utils/data_extraction/unternehmensregister/extract.py +++ b/src/aki_prj23_transparenzregister/utils/data_extraction/unternehmensregister/extract.py @@ -1,4 +1,5 @@ """Unternehmensregister Scraping.""" +# ruff: noqa: B023 import glob import multiprocessing import os @@ -105,7 +106,9 @@ def scrape(query: str, download_dir: list[str]) -> None: driver.find_element(By.CLASS_NAME, "download-wrapper").click() try: - wait.until(wait_for_download_condition(download_path, num_files)) + wait.until( + lambda: wait_for_download_condition(download_path, num_files) # type: ignore + ) file_name = "".join(e for e in company_name if e.isalnum()) + ".xml" rename_latest_file( download_path,