Skip to content

I'm always getting n_tokens <= n_batch and then it fails running #76

Closed Answered by giladgd
rossjackson asked this question in Q&A
Discussion options

You must be logged in to vote

@rossjackson Try setting batchSize to the same value as the contextSize you have set on a LlamaContext.
The limit on a single prompt evaluation tokens count is the batchSize (including the tokens generated by the chat prompt wrapper)

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by giladgd
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants