You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The procedure described above should create a correct conda env for running test_correspondence.ipynb
What happened
Some scary warnings: In the first cell (imports) in test_correspondence.ipynb,
WARNING[XFORMERS]: xFormers can't load C++/CUDA extensions. xFormers was built for:
PyTorch 2.1.0.post100 with CUDA None (you have 2.1.0+cu121)
Python 3.10.13 (you have 3.10.14)
Please reinstall xformers (see https://github.com/facebookresearch/xformers#installing-xformers)
Memory-efficient attention, SwiGLU, sparse and more won't be available.
Set XFORMERS_MORE_DETAILS=1 for more details
in cell 7 due to xformers not being installed correctly
File ~/miniconda3/envs/diff3f/lib/python3.10/site-packages/xformers/ops/fmha/dispatch.py:63, in _run_priority_list(name, priority_list, inp)
61 for op, not_supported in zip(priority_list, not_supported_reasons):
62 msg += "\n" + _format_not_supported_reasons(op, not_supported)
---> 63 raise NotImplementedError(msg)
NotImplementedError: No operator found for `memory_efficient_attention_forward` with inputs:
query : shape=(1, 1370, 12, 64) (torch.float32)
key : shape=(1, 1370, 12, 64) (torch.float32)
value : shape=(1, 1370, 12, 64) (torch.float32)
attn_bias : <class 'NoneType'>
p : 0.0
`decoderF` is not supported because:
xFormers wasn't build with CUDA support
requires device with capability > (8, 0) but your GPU has capability (7, 5) (too old)
attn_bias type is <class 'NoneType'>
operator wasn't built - see `python -m xformers.info` for more info
`[email protected]` is not supported because:
xFormers wasn't build with CUDA support
requires device with capability > (8, 0) but your GPU has capability (7, 5) (too old)
dtype=torch.float32 (supported: {torch.bfloat16, torch.float16})
operator wasn't built - see `python -m xformers.info` for more info
`tritonflashattF` is not supported because:
xFormers wasn't build with CUDA support
requires device with capability > (8, 0) but your GPU has capability (7, 5) (too old)
dtype=torch.float32 (supported: {torch.bfloat16, torch.float16})
operator wasn't built - see `python -m xformers.info` for more info
triton is not available
requires GPU with sm80 minimum compute capacity, e.g., A100/H100/L4
Only work on pre-MLIR triton for now
`cutlassF` is not supported because:
xFormers wasn't build with CUDA support
operator wasn't built - see `python -m xformers.info` for more info
`smallkF` is not supported because:
max(query.shape[-1] != value.shape[-1]) > 32
xFormers wasn't build with CUDA support
operator wasn't built - see `python -m xformers.info` for more info
unsupported embed per head: 64
What I think is the problem
The setup does not configure accelerate and xformers correctly.
Suggested solution
Install accelerate and xformers correctly. I configured a docker image to build such environment.
The text was updated successfully, but these errors were encountered:
Thanks for sharing the code. I struggled with the environment setup for a while, so I'm sharing my experience here and describing how I solved it.
What I did
I followed README: 1. create env with
conda env create -f environment.yaml
. 2. install pytorch3d withI'm using Arch Linux and miniconda3.
What is expected
The procedure described above should create a correct conda env for running test_correspondence.ipynb
What happened
in cell 5 due to absence of
accelerate
package.in cell 7 due to xformers not being installed correctly
What I think is the problem
The setup does not configure accelerate and xformers correctly.
Suggested solution
Install accelerate and xformers correctly. I configured a docker image to build such environment.
The text was updated successfully, but these errors were encountered: