Skip to content

llamacpp chat/completions response unrelated to prompt on cpu local deploy #1919

Closed Answered by mudler
semsion asked this question in Q&A
Discussion options

You must be logged in to vote

Don't use the model file as model in the request unless you want to handle the prompt template for yourself.

Just use the model names like you would do with OpenAI. For instance gpt-4-vision-preview, or gpt-4 are already present in the AIO images, just use those as model when doing the curl calls.

Replies: 2 comments 1 reply

Comment options

You must be logged in to vote
0 replies
Answer selected by semsion
Comment options

You must be logged in to vote
1 reply
@mudler
Comment options

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants
Converted from issue

This discussion was converted from issue #1918 on March 28, 2024 10:58.