Thanks to visit codestin.com
Credit goes to github.com

Skip to content

Vulkan Use Mmap (disable Vulkan MMap opt out)#14525

Draft
inforithmics wants to merge 1 commit intoollama:mainfrom
inforithmics:VulkanUseMMap
Draft

Vulkan Use Mmap (disable Vulkan MMap opt out)#14525
inforithmics wants to merge 1 commit intoollama:mainfrom
inforithmics:VulkanUseMMap

Conversation

@inforithmics
Copy link
Contributor

@inforithmics inforithmics commented Mar 1, 2026

Vulkan Use Mmap (disable Vulkan MMAp opt out)

Draft until new Vendor sync because in newer llama.cpp Versions mmap works without problems in Vulkan (It shows that it uses more memory but in reality it doesn't) It shows more used memory in Process and gpu but in effect there is more free memory.

The reason it seems that the Memory is duplicated is that the Shared Memory used in the GPU is also reported to be used by ollama (But it is the same memory) So for example 6GB Shared Memory used Ollama although "uses" 6GB Memory. This hapens on iGPU where the shared Memory is used for GPU offload.

@inforithmics inforithmics marked this pull request as draft March 1, 2026 11:01
@inforithmics inforithmics changed the title Vulkan Use Mmap (disable Vulkan MMAp opt out) Vulkan Use Mmap (disable Vulkan MMap opt out) Mar 1, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant