You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+4-1
Original file line number
Diff line number
Diff line change
@@ -163,7 +163,7 @@ Below is a short example demonstrating how to use the high-level API to for basi
163
163
)
164
164
>>> output = llm(
165
165
"Q: Name the planets in the solar system? A: ", # Prompt
166
-
max_tokens=32, # Generate up to 32 tokens
166
+
max_tokens=32, # Generate up to 32 tokens, set to None to generate up to the end of the context window
167
167
stop=["Q:", "\n"], # Stop generating just before the model would generate a new question
168
168
echo=True# Echo the prompt back in the output
169
169
) # Generate a completion, can also call create_completion
@@ -425,6 +425,9 @@ pip install -e .[all]
425
425
make clean
426
426
```
427
427
428
+
You can also test out specific commits of `lama.cpp` by checking out the desired commit in the `vendor/llama.cpp` submodule and then running `make clean` and `pip install -e .` again. Any changes in the `llama.h` API will require
429
+
changes to the `llama_cpp/llama_cpp.py` file to match the new API (additional changes may be required elsewhere).
430
+
428
431
## FAQ
429
432
430
433
### Are there pre-built binaries / binary wheels available?
0 commit comments