Skip to content

Commit

Permalink
Merge branch 'llama-cpp-logits-fix'
Browse files Browse the repository at this point in the history
  • Loading branch information
lbeurerkellner committed Jul 28, 2023
2 parents ed14ec3 + e3b7766 commit 3abc049
Showing 1 changed file with 3 additions and 3 deletions.
6 changes: 3 additions & 3 deletions src/lmql/models/lmtp/backends/llama_cpp_model.py
Original file line number Diff line number Diff line change
Expand Up @@ -40,9 +40,9 @@ def score(self, input_ids, attention_mask, **model_kwargs):
self.llm.n_tokens = longest_prefix

self.llm.eval(tokens)
scores = np.array([self.llm.scores[j][i] for j,i in enumerate(input_ids[0])])
scores = nputil.log_softmax(scores, axis=-1)
# print("llama_cpp_model: score() took", time.time() - s, "seconds", file=sys.stderr)
logits = np.array(self.llm.scores)
logits = nputil.log_softmax(logits, axis=-1)
scores = np.array([logits[j][i] for j,i in enumerate(input_ids[0])])

return scores.reshape(1, -1)

Expand Down

0 comments on commit 3abc049

Please sign in to comment.