-
Notifications
You must be signed in to change notification settings - Fork 450
use constraints-dev.txt in e2e tests
#3320
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
For instructlab, "pip install ." does not install vllm, but it does install an uncapped torch (2.7.0 currently). When we install vllm later, we compile a binary flash_attn wheel against torch 2.7.0. vllm 0.8.4 requires torch==2.6.0, so we downgrade torch, and then we use that with the incompatible flash_attn binary wheel. To resolve this, use constraints-dev.txt in the first pip install operation. This restricts torch to 2.6.0 immediately when we first install instructlab, so that we will compile flash_attn against that torch version. Signed-off-by: Ken Dreyer <[email protected]>
|
E2E (NVIDIA L40S x4) workflow launched on this PR: View run |
|
e2e workflow failed on this PR: View run, please investigate. |
|
This fixes the flash-attn problem. The e2e tests get further, but they still fail in NCCL timeouts. I've filed #3321 to track that separately. |
|
Force merging since Ken confirmed this improves situation even if it doesn't make CI green yet. |
|
@mergify backport release-v0.26 |
✅ Backports have been createdDetails
|
…-3320 use `constraints-dev.txt` in e2e tests (backport #3320)
This is a port of instructlab/instructlab#3320 over to the SDG repository. While doing so, I noticed we also were not using "-DGGML_CUDA=ON" so updated that as well, since it's the same pip install line in the file. Signed-off-by: Ben Browning <[email protected]>
This is a port of instructlab/instructlab#3320 over to the SDG repository. While doing so, I noticed we also were not using "-DGGML_CUDA=ON" so updated that as well, since it's the same pip install line in the file. Signed-off-by: Ben Browning <[email protected]> (cherry picked from commit 225612c)
For instructlab,
pip install .does not installvllm, but it does install an uncappedtorch(2.7.0currently).When we install
vllmlater, we compile a binaryflash_attnwheel againsttorch 2.7.0.vllm 0.8.4requirestorch==2.6.0, so we downgradetorch, and then we use that with the incompatibleflash_attnbinary wheel.ImportErrorlooks like:To resolve this, use
constraints-dev.txtin the firstpip installoperation. This restrictstorchto2.6.0immediately when we first install instructlab, so that we will compileflash_attnagainst thattorchversion.