You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
ggml-backend : fix async copy from CPU (ggml-org#8897)
* ggml-backend : fix async copy from CPU
* cuda : more reliable async copy, fix stream used when the devices are the same
llama : add support for llama 3.1 rope scaling factors (ggml-org#8676)
* Add llama 3.1 rope scaling factors to llama conversion and inference
This commit generates the rope factors on conversion and adds them to the resulting model as a tensor. At inference time, these factors are passed to the `ggml_rope_ext` rope oepration, improving results for context windows above 8192
* Update convert_hf_to_gguf.py
Co-authored-by: compilade <[email protected]>
* address comments
* address comments
* Update src/llama.cpp
Co-authored-by: compilade <[email protected]>
* Update convert_hf_to_gguf.py
Co-authored-by: compilade <[email protected]>
---------
Co-authored-by: compilade <[email protected]>