Thanks to visit codestin.com
Credit goes to github.com

Skip to content

HIP: disable rocwmma on gfx12 by default until rocm 7.0 #14202

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Jun 16, 2025

Conversation

IMbackK
Copy link
Collaborator

@IMbackK IMbackK commented Jun 15, 2025

Currently rocwmma dose not support gfx12 in any released version of rocm, causeing build failures when llamacpp is built with GGML_HIP_ROCWMMA_FATTN enabled against gfx12.

This causes issues for one of our downstreams (koboldcpp) who want to build a fat binary with gfx8-gfx12 support and simultaneously use rocwmma where applicable. This detail has also caused various issues #14193 #13110 to be filed against llamacpp by confused users.

This pr thus disables rocwmma on gfx12 even when GGML_HIP_ROCWMMA_FATTN is set if rocm < 7.0 is used.
I have it on good authority that gfx12 rocwmma support will land in rocm 7.0 which is scheduled for release later this year.

As the rocwmma git repo has contained gfx12 support for some time, it has been possible to compile llamacpp against git rocwmma to get a working build for some time, to facilitate this, this pr also adds a GGML_HIP_FORCE_ROCWMMA_FATTN_GFX12 option to override the check for rocm 7.0 and try to use rocwmma regardless.

@github-actions github-actions bot added Nvidia GPU Issues specific to Nvidia GPUs ggml changes relating to the ggml tensor library for machine learning labels Jun 15, 2025
@IMbackK IMbackK requested a review from JohannesGaessler June 15, 2025 20:35
@LostRuins
Copy link
Collaborator

LostRuins commented Jun 16, 2025

Nice, hopefully amd will add it in soon. Eventually as rocm updates this can probably be changed again.

@henk717
Copy link

henk717 commented Jun 16, 2025

With the version check I expect it to be fine since people will be running older versions for a while, 7.0 users would just get it automatically.

@IMbackK IMbackK merged commit 7d6d91b into ggml-org:master Jun 16, 2025
87 of 88 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ggml changes relating to the ggml tensor library for machine learning Nvidia GPU Issues specific to Nvidia GPUs
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants