Thanks to visit codestin.com Credit goes to github.com
We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 45bf5ae commit 62aad61Copy full SHA for 62aad61
llama_cpp/_internals.py
@@ -730,12 +730,14 @@ def sample(
730
if len(self.prev) > 0:
731
nl_token = ctx_main.model.token_nl()
732
nl_logit = logits_array[nl_token]
733
- if self.params.penalty_last_n > 0:
+ last_tokens = self.prev[-self.params.penalty_last_n:]
734
+ last_tokens_size = min(len(last_tokens), self.params.penalty_last_n)
735
+ if last_tokens_size > 0:
736
+ last_tokens_p = (llama_cpp.llama_token * len(last_tokens))(*last_tokens)
737
ctx_main.sample_repetition_penalties(
738
token_data_array,
- # TODO: Only create this once
- (llama_cpp.llama_token * len(self.prev))(*self.prev),
- self.params.penalty_last_n,
739
+ last_tokens_p,
740
+ last_tokens_size,
741
self.params.penalty_repeat,
742
self.params.penalty_freq,
743
self.params.penalty_present,
0 commit comments