-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Request]: koboldcpp-cu12 (new version using CUDA 12) #13275
Labels
package-request
Request for adding an app
Comments
Guess this belongs rather to Versions bucket. |
@aliesbelik You're absolutely correct. It totally escaped me. I'll move it over there. |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Prerequisites
Criteria
Name
koboldcpp-cu12
Description
"A simple one-file way to run various GGML models like LLAMA, ALPACA, VICUNA. This version uses the newer CUDA 12 binaries. If you have a newer NVIDIA GPU and don't mind larger files, you may get increased speeds by using this new version."
Homepage
https://github.com/lostruins/koboldcpp
Download Link(s)
https://github.com/LostRuins/koboldcpp/releases/download/{latest-tag}/koboldcpp_cu12.exe
Eg.: https://github.com/LostRuins/koboldcpp/releases/download/v1.65/koboldcpp_cu12.exe
Some Indication of Popularity/Repute
3.9k stars
282 forks
Same regular install as the standard
koboldcpp
, just a new executable with newer cuda binaries.The text was updated successfully, but these errors were encountered: