diff --git a/CHANGELOG.md b/CHANGELOG.md index be829667..fd691e07 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,3 +1,7 @@ +## 0.0.51 + +* Bump unstructured to 0.10.19 + ## 0.0.50 * Bump unstructured to 0.10.18 diff --git a/prepline_general/api/app.py b/prepline_general/api/app.py index 40461ee7..5be5d0eb 100644 --- a/prepline_general/api/app.py +++ b/prepline_general/api/app.py @@ -11,7 +11,7 @@ app = FastAPI( title="Unstructured Pipeline API", description="""""", - version="0.0.50", + version="0.0.51", docs_url="/general/docs", openapi_url="/general/openapi.json", ) diff --git a/prepline_general/api/general.py b/prepline_general/api/general.py index 6ac98769..bb219977 100644 --- a/prepline_general/api/general.py +++ b/prepline_general/api/general.py @@ -569,7 +569,7 @@ def return_content_type(filename): @router.post("/general/v0/general") -@router.post("/general/v0.0.50/general") +@router.post("/general/v0.0.51/general") def pipeline_1( request: Request, gz_uncompressed_content_type: Optional[str] = Form(default=None), diff --git a/preprocessing-pipeline-family.yaml b/preprocessing-pipeline-family.yaml index 53562207..b03f29fa 100644 --- a/preprocessing-pipeline-family.yaml +++ b/preprocessing-pipeline-family.yaml @@ -1,2 +1,2 @@ name: general -version: 0.0.50 +version: 0.0.51 diff --git a/requirements/base.txt b/requirements/base.txt index 28f2dd86..78f1d614 100644 --- a/requirements/base.txt +++ b/requirements/base.txt @@ -11,7 +11,7 @@ anyio==3.7.1 # fastapi # starlette backoff==2.2.1 - # via -r base.in + # via -r requirements/base.in beautifulsoup4==4.12.2 # via unstructured certifi==2023.7.22 @@ -26,11 +26,9 @@ charset-normalizer==3.3.0 # requests click==8.1.3 # via - # -r base.in + # -r requirements/base.in # nltk # uvicorn -cmake==3.27.6 - # via triton coloredlogs==15.0.1 # via onnxruntime contourpy==1.1.1 @@ -49,8 +47,10 @@ emoji==2.8.0 # via unstructured et-xmlfile==1.1.0 # via openpyxl +exceptiongroup==1.1.3 + # via anyio fastapi==0.103.2 - # via -r base.in + # via -r requirements/base.in filelock==3.12.4 # via # huggingface-hub @@ -64,12 +64,15 @@ flatbuffers==23.5.26 fonttools==4.43.0 # via matplotlib fsspec==2023.9.2 - # via huggingface-hub + # via + # huggingface-hub + # torch h11==0.14.0 # via uvicorn -huggingface-hub==0.17.3 +huggingface-hub==0.16.4 # via # timm + # tokenizers # transformers # unstructured-inference humanfriendly==10.0 @@ -78,6 +81,10 @@ idna==3.4 # via # anyio # requests +importlib-metadata==6.8.0 + # via markdown +importlib-resources==6.1.0 + # via matplotlib iopath==0.1.10 # via layoutparser jinja2==3.1.2 @@ -90,8 +97,6 @@ langdetect==1.0.9 # via unstructured layoutparser[layoutmodels,tesseract]==0.3.4 # via unstructured-inference -lit==17.0.2 - # via triton lxml==4.9.3 # via # ebooklib @@ -130,30 +135,36 @@ numpy==1.24.4 # torchvision # transformers # unstructured -nvidia-cublas-cu11==11.10.3.66 +nvidia-cublas-cu12==12.1.3.1 # via - # nvidia-cudnn-cu11 - # nvidia-cusolver-cu11 + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 # torch -nvidia-cuda-cupti-cu11==11.7.101 - # via torch -nvidia-cuda-nvrtc-cu11==11.7.99 +nvidia-cuda-cupti-cu12==12.1.105 # via torch -nvidia-cuda-runtime-cu11==11.7.99 +nvidia-cuda-nvrtc-cu12==12.1.105 # via torch -nvidia-cudnn-cu11==8.5.0.96 +nvidia-cuda-runtime-cu12==12.1.105 # via torch -nvidia-cufft-cu11==10.9.0.58 +nvidia-cudnn-cu12==8.9.2.26 # via torch -nvidia-curand-cu11==10.2.10.91 +nvidia-cufft-cu12==11.0.2.54 # via torch -nvidia-cusolver-cu11==11.4.0.1 +nvidia-curand-cu12==10.3.2.106 # via torch -nvidia-cusparse-cu11==11.7.4.91 +nvidia-cusolver-cu12==11.4.5.107 # via torch -nvidia-nccl-cu11==2.14.3 +nvidia-cusparse-cu12==12.1.0.106 + # via + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.18.1 # via torch -nvidia-nvtx-cu11==11.7.91 +nvidia-nvjitlink-cu12==12.2.140 + # via + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 # via torch olefile==0.46 # via msg-parser @@ -204,28 +215,28 @@ pillow==10.0.1 # unstructured-pytesseract portalocker==2.8.2 # via iopath -protobuf==4.24.3 +protobuf==4.24.4 # via # onnx # onnxruntime psutil==5.9.5 - # via -r base.in + # via -r requirements/base.in pycocotools==2.0.7 # via effdet pycparser==2.21 # via cffi pycryptodome==3.19.0 - # via -r base.in + # via -r requirements/base.in pydantic==1.10.13 # via - # -r base.in + # -r requirements/base.in # fastapi pypandoc==1.11 # via unstructured pyparsing==3.1.1 # via matplotlib pypdf==3.16.2 - # via -r base.in + # via -r requirements/base.in pypdfium2==4.20.0 # via pdfplumber pytesseract==0.3.10 @@ -256,21 +267,21 @@ pyyaml==6.0.1 rapidfuzz==3.3.1 # via unstructured-inference ratelimit==2.2.1 - # via -r base.in -regex==2023.8.8 + # via -r requirements/base.in +regex==2023.10.3 # via # nltk # transformers requests==2.31.0 # via - # -r base.in + # -r requirements/base.in # huggingface-hub # torchvision # transformers # unstructured safetensors==0.3.2 # via - # -c constraints.in + # -c requirements/constraints.in # timm # transformers scipy==1.10.1 @@ -294,16 +305,15 @@ tabulate==0.9.0 # via unstructured timm==0.9.7 # via effdet -tokenizers==0.13.3 +tokenizers==0.14.0 # via transformers -torch==2.0.1 +torch==2.1.0 # via # effdet # layoutparser # timm # torchvision - # triton -torchvision==0.15.2 +torchvision==0.16.0 # via # effdet # layoutparser @@ -314,42 +324,43 @@ tqdm==4.66.1 # iopath # nltk # transformers -transformers==4.33.3 +transformers==4.34.0 # via unstructured-inference -triton==2.0.0 +triton==2.1.0 # via torch typing-extensions==4.8.0 # via + # annotated-types # fastapi # huggingface-hub # iopath # onnx # pydantic + # pydantic-core + # pypdf + # starlette # torch # typing-inspect + # uvicorn typing-inspect==0.9.0 # via dataclasses-json tzdata==2023.3 # via pandas unstructured[local-inference]==0.10.19 - # via -r base.in + # via -r requirements/base.in unstructured-inference==0.6.6 # via unstructured unstructured-pytesseract==0.3.12 # via unstructured -urllib3==2.0.5 +urllib3==2.0.6 # via requests uvicorn==0.23.2 - # via -r base.in -wheel==0.41.2 - # via - # nvidia-cublas-cu11 - # nvidia-cuda-cupti-cu11 - # nvidia-cuda-runtime-cu11 - # nvidia-curand-cu11 - # nvidia-cusparse-cu11 - # nvidia-nvtx-cu11 + # via -r requirements/base.in xlrd==2.0.1 # via unstructured xlsxwriter==3.1.6 # via python-pptx +zipp==3.17.0 + # via + # importlib-metadata + # importlib-resources diff --git a/requirements/test.txt b/requirements/test.txt index 0ffc88b3..1b902a94 100644 --- a/requirements/test.txt +++ b/requirements/test.txt @@ -15,10 +15,6 @@ anyio==3.7.1 # httpcore # jupyter-server # starlette -appnope==0.1.3 - # via - # ipykernel - # ipython argon2-cffi==23.1.0 # via jupyter-server argon2-cffi-bindings==21.2.0 @@ -37,7 +33,7 @@ attrs==23.1.0 # via # jsonschema # referencing -babel==2.12.1 +babel==2.13.0 # via jupyterlab-server backcall==0.2.0 # via ipython @@ -91,7 +87,7 @@ contourpy==1.1.1 # via # -r requirements/base.txt # matplotlib -coverage[toml]==7.3.1 +coverage[toml]==7.3.2 # via pytest-cov cryptography==41.0.4 # via @@ -151,6 +147,7 @@ filelock==3.12.4 # huggingface-hub # torch # transformers + # triton filetype==1.2.0 # via # -r requirements/base.txt @@ -171,6 +168,7 @@ fsspec==2023.9.2 # via # -r requirements/base.txt # huggingface-hub + # torch ghapi==1.0.4 # via nbdev h11==0.14.0 @@ -182,10 +180,11 @@ httpcore==0.18.0 # via httpx httpx==0.25.0 # via -r requirements/test.in -huggingface-hub==0.17.3 +huggingface-hub==0.16.4 # via # -r requirements/base.txt # timm + # tokenizers # transformers # unstructured-inference humanfriendly==10.0 @@ -371,7 +370,7 @@ mypy-extensions==1.0.0 # typing-inspect nbclient==0.8.0 # via nbconvert -nbconvert==7.8.0 +nbconvert==7.9.2 # via # jupyter # jupyter-server @@ -413,6 +412,58 @@ numpy==1.24.4 # torchvision # transformers # unstructured +nvidia-cublas-cu12==12.1.3.1 + # via + # -r requirements/base.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -r requirements/base.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -r requirements/base.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -r requirements/base.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -r requirements/base.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -r requirements/base.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -r requirements/base.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -r requirements/base.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -r requirements/base.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.18.1 + # via + # -r requirements/base.txt + # torch +nvidia-nvjitlink-cu12==12.2.140 + # via + # -r requirements/base.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -r requirements/base.txt + # torch olefile==0.46 # via # -r requirements/base.txt @@ -503,7 +554,7 @@ pillow==10.0.1 # unstructured-pytesseract pkgutil-resolve-name==1.3.10 # via jsonschema -platformdirs==3.10.0 +platformdirs==3.11.0 # via # black # jupyter-core @@ -519,7 +570,7 @@ prompt-toolkit==3.0.39 # via # ipython # jupyter-console -protobuf==4.24.3 +protobuf==4.24.4 # via # -r requirements/base.txt # onnx @@ -650,7 +701,7 @@ referencing==0.30.2 # jsonschema # jsonschema-specifications # jupyter-events -regex==2023.8.8 +regex==2023.10.3 # via # -r requirements/base.txt # nltk @@ -671,7 +722,7 @@ rfc3986-validator==0.1.1 # via # jsonschema # jupyter-events -rpds-py==0.10.3 +rpds-py==0.10.4 # via # jsonschema # referencing @@ -732,7 +783,7 @@ timm==0.9.7 # effdet tinycss2==1.2.1 # via nbconvert -tokenizers==0.13.3 +tokenizers==0.14.0 # via # -r requirements/base.txt # transformers @@ -743,14 +794,14 @@ tomli==2.0.1 # jupyterlab # mypy # pytest -torch==2.0.1 +torch==2.1.0 # via # -r requirements/base.txt # effdet # layoutparser # timm # torchvision -torchvision==0.15.2 +torchvision==0.16.0 # via # -r requirements/base.txt # effdet @@ -771,7 +822,7 @@ tqdm==4.66.1 # iopath # nltk # transformers -traitlets==5.10.1 +traitlets==5.11.2 # via # comm # ipykernel @@ -788,10 +839,14 @@ traitlets==5.10.1 # nbconvert # nbformat # qtconsole -transformers==4.33.3 +transformers==4.34.0 # via # -r requirements/base.txt # unstructured-inference +triton==2.1.0 + # via + # -r requirements/base.txt + # torch types-python-dateutil==2.8.19.14 # via arrow typing-extensions==4.8.0 @@ -819,9 +874,9 @@ tzdata==2023.3 # via # -r requirements/base.txt # pandas -unstructured[local-inference]==0.10.18 +unstructured[local-inference]==0.10.19 # via -r requirements/base.txt -unstructured-inference==0.5.31 +unstructured-inference==0.6.6 # via # -r requirements/base.txt # unstructured @@ -831,7 +886,7 @@ unstructured-pytesseract==0.3.12 # unstructured uri-template==1.3.0 # via jsonschema -urllib3==2.0.5 +urllib3==2.0.6 # via # -r requirements/base.txt # requests