-
Notifications
You must be signed in to change notification settings - Fork 53
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
INTERNAL ASSERT FAILED: Allocations must be based on constant integers for local memory #2792
Comments
Another instance of #2702? CC: @jacobhinkle |
Yes I think this is a dupe of #2702. I verified that #2714 fixes it as long as I address the known issue with |
IvanYashchuk
added a commit
to Lightning-AI/lightning-thunder
that referenced
this issue
Aug 16, 2024
…st of executors (#974) We should use an Inductor-based concatenation executor (through torch.compile) by default since it gives us a perf improvement for sections of the network that include the torch.cat operation (mainly RoPE). Previously we didn't enable it by default because there were memory leaks that were fixed (see Lightning-AI/lit-thunder-LEGACY#2194 if you have access). In addition, this PR also avoids hitting a problem with nvFuser (NVIDIA/Fuser#2792). The backward of RoPE was generating a nvFuser fusion in between TorchCompile region because the torchcompile_cat executor wasn't marked to be able to execute torch.sum. I also added a test that verifies that only one fusion region is created. In the (NVIDIA-internal) Mixology dashboard Llama-3-8B, Mistral-7B-v0.1, and stablecode-completion-alpha-3b models do not work with Thunder + Inductor concatenation executor and this problem should be fixed in this PR
Fixed by #2714 |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
I get the following error when running one of Thunder benchmarks (
pytest thunder/benchmarks/targets.py -k "test_litgpt_qkv_split_rope[Llama-3-8B-backward-bs1-thunder+nvfuser+torch.compile]"
):The text was updated successfully, but these errors were encountered: