-
Notifications
You must be signed in to change notification settings - Fork 9
/
pyproject.toml
130 lines (112 loc) · 4.4 KB
/
pyproject.toml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
[tool.poetry]
name = "vision-agent-tools"
version = "0.1.1"
description = "Toolbox for vision tasks"
authors = ["Landing AI <[email protected]>"]
readme = "README.md"
[tool.poetry.urls]
"homepage" = "https://landing.ai"
"repository" = "https://github.com/landing-ai/vision-agent-tools"
"documentation" = "https://github.com/landing-ai/vision-agent-tools"
[build-system]
requires = ["poetry-core>=1.8.0", "setuptools", "packaging", "wheel", "torch"]
build-backend = "poetry.core.masonry.api"
[tool.poetry.dependencies]
python = ">=3.10,<3.13"
pydantic = "^2.7.4"
numpy = ">=1.17,<2.0"
pillow = ">=10.0.1,<=15.0"
gdown = "^5.1.0"
wget = "^3.2"
torch = { version = "2.3.1", source = "torch_cuda118" }
torchvision = { version = "0.18.1", source = "torch_cuda118" }
flash-attn = [
{ url = "https://github.com/Dao-AILab/flash-attention/releases/download/v2.6.1/flash_attn-2.6.1+cu118torch2.3cxx11abiFALSE-cp310-cp310-linux_x86_64.whl", python = "~3.10", optional = true },
{ url = "https://github.com/Dao-AILab/flash-attention/releases/download/v2.6.1/flash_attn-2.6.1+cu118torch2.3cxx11abiFALSE-cp311-cp311-linux_x86_64.whl", python = "~3.11", optional = true },
{ url = "https://github.com/Dao-AILab/flash-attention/releases/download/v2.6.1/flash_attn-2.6.1+cu118torch2.3cxx11abiFALSE-cp312-cp312-linux_x86_64.whl", python = "~3.12", optional = true },
]
qreader = { version = "^3.14", optional = true }
scipy = { version = "^1.13.1", optional = true }
peft = { version = "^0.11.1", optional = true }
timm = { version = "^0.6.7", optional = true }
einops = { version = "^0.8.0", optional = true }
transformers = { extras = ["torch"], version = "^4.45.0", optional = true }
controlnet-aux = { version = "^0.0.9", optional = true }
lmdeploy = { version = "^0.5.3", optional = true }
sentencepiece = { version = "^0.2.0", optional = true }
protobuf = { version = "^5.28.2", optional = true }
qwen_vl_utils = { version = "^0.0.8", optional = true }
huggingface-hub = { version = "^0.26.2", optional = true }
# this dependency is just here due to the ixc-25 model and because it's imported in one util file
# that we don't use: https://huggingface.co/internlm/internlm-xcomposer2d5-7b/blob/main/ixc_utils.py#L9
# as soon as they remove it we can extinguish from our side too
decord = { version = "^0.6.0", optional = true }
diffusers = { git = "https://github.com/huggingface/diffusers.git", branch = "main", optional = true }
loca = { git = "https://github.com/landing-ai/loca.git", branch = "main", optional = true }
sam-2 = { git = "https://github.com/landing-ai/segment-anything-2.git", branch = "main", optional = true }
depth-anything-v2 = { git = "https://github.com/landing-ai/depth-anything-v2.git", branch = "main", optional = true }
[tool.poetry.group.dev.dependencies]
pre-commit = "^3.7.1"
pytest = "^8.2.2"
ruff = "^0.5.0"
mypy = "^1.10.1"
ipykernel = "^6.29.5"
pytest-asyncio = "^0.23.7"
opencv-stubs = "^0.0.10"
opencv-python = "^4.10.0.84"
# notebooks
av = "^13.1.0"
matplotlib = "^3.9.2"
[tool.poetry.group.docs.dependencies]
mkdocstrings = { extras = ["python"], version = "^0.25.1" }
mkdocs-material = "^9.5.28"
griffe-fieldz = "^0.2.0"
opencv-python = "^4.10.0.84"
[tool.poetry.extras]
all = [
"flash-attn",
"qreader",
"transformers",
"scipy",
"loca",
"depth-anything-v2",
"timm",
"peft",
"einops",
"controlnet-aux",
"lmdeploy",
"sam-2",
"decord",
"diffusers",
"sentencepiece",
"protobuf",
"qwen_vl_utils",
]
qr-reader = ["qreader"]
owlv2 = ["transformers", "scipy"]
loca-model = ["loca"]
depth-anything-v2-model = ["depth-anything-v2"]
nsfw-classification = ["transformers", "scipy"]
controlnet-aux = ["controlnet-aux"]
florence2 = ["flash-attn", "transformers", "timm", "einops", "peft"]
florence2-qa = ["flash-attn", "transformers", "timm", "einops", "peft"]
florence2-sam2 = [
"flash-attn",
"transformers",
"timm",
"einops",
"peft",
"sam-2",
]
clip-media-sim = ["transformers"]
ixc-25 = ["transformers", "lmdeploy", "decord"]
flux1 = ["transformers", "diffusers", "sentencepiece", "protobuf"]
qwen2-vl = ["qwen_vl_utils", "transformers", "flash-attn"]
siglip = ["flash-attn", "transformers", "sentencepiece", "protobuf"]
sam2-model = ["sam-2", "huggingface-hub"]
[[tool.poetry.source]]
name = "torch_cuda118"
url = "https://download.pytorch.org/whl/cu118"
priority = "explicit"
[tool.poetry.scripts]
install-dependencies = "scripts.install_dependencies:install_dependencies"