-
Notifications
You must be signed in to change notification settings - Fork 1.1k
Issues: abetlen/llama-cpp-python
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Can't install with GPU support with Cuda toolkit 12.9 and Cuda 12.9
#2013
opened May 5, 2025 by
hunainahmedj
TypeError: 'NoneType' object is not callable in __del__() when exiting
#2002
opened Apr 12, 2025 by
gbutiri
4 tasks done
Safeerchalil:codespace-automatic-barnacle-5gv4qx4j775wf47g
#2000
opened Apr 11, 2025 by
Safeerchalil
How to use *chat_template* with .gguf models ? (tokenizer_name not implemented)
#1999
opened Apr 11, 2025 by
Bobchenyx
Running basic example from docs results in
TypeError: 'NoneType' object is not callable
#1998
opened Apr 11, 2025 by
nchammas
4 tasks done
Intel Mac (i9 - 5500M) (macOS 15.3.2) - ValueError: Failed to create llama_context - llama_init_from_model: failed to initialize Metal backend
#1988
opened Mar 29, 2025 by
starkAhmed43
4 tasks done
Stream last block not return token usage info when create_chat_completion_openai_v1 or create_chat_completion but server does
#1984
opened Mar 27, 2025 by
hh23485
4 tasks done
Slow when logits_all=True, inconsistent logprobs and solutions
#1983
opened Mar 26, 2025 by
For-rest2005
HIP enabled binary, or allow me to replace with my own llama binary
#1982
opened Mar 25, 2025 by
madprops
Set GGML_BUILD_NUMBER to the correct version when building from pypi tarball
#1979
opened Mar 22, 2025 by
booxter
4 tasks done
Previous Next
ProTip!
Type g p on any issue or pull request to go back to the pull request listing page.