You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have separately cloned llama-cpp from the GitHub repo and have built it according to the following post Ollama issues forum. That successfully runs any LLM on my GPU.
However when I try to build llama-cpp-python, it recognizes the Vulkan backend but still throws the same metal error and the same ValueError: Failed to create llama_context
Prerequisites
Please answer the following questions for yourself before submitting an issue.
Expected Behavior
I am trying to run a local LLM:
`from llama_cpp import Llama
llm = Llama(model_path="/Users/mfzainulabideen/Downloads/Llama/Llama-3.2-3B-Instruct/Llama-3.2-3B-Instruct-F16.gguf")`
It should execute the code and complete the cell in ipynb so that I can figure out what I plan to do next with the LLM.
Current Behavior
The code doesn't run and throws an error.
Environment and Context
I am trying to execute this on a 2019 16" MBP with a 2.4GHz i9, 32GB RAM and a 1TB SSD running on macOS 15.3.2.
I am using the latest miniconda env with
Python 3.11.11
.GNU Make 3.81
built fori386-apple-darwin11.3.0
.Failure Information (for bugs)
Python throws the following error:
Steps to Reproduce
then in a notebook, with the
llm
env as the kernel, run the following code:Obviously change the model_path.
Failure Logs
PFA the verbose log upto the point of failure:
llama-cpp-python-verbose-log.txt
My environment info:
The text was updated successfully, but these errors were encountered: