Here are
8 public repositories
matching this topic...
Running Llama 2 and other Open-Source LLMs on CPU Inference Locally for Document Q&A
Updated
Nov 6, 2023
Python
Runs LLaMA with Extremely HIGH speed
Updated
Mar 10, 2024
Fortran
The bare metal in my basement
Portable LLM - A rust library for LLM inference
Updated
Apr 13, 2024
Rust
Wrapper for simplified use of Llama2 GGUF quantized models.
Updated
Jan 14, 2024
Python
Simple large language model playground app
Updated
May 13, 2024
Python
Simple bot that transcribes Telegram voice messages. Powered by go-telegram-bot-api & whisper.cpp Go bindings.
Improve this page
Add a description, image, and links to the
cpu-inference
topic page so that developers can more easily learn about it.
Curate this topic
Add this topic to your repo
To associate your repository with the
cpu-inference
topic, visit your repo's landing page and select "manage topics."
Learn more
You can’t perform that action at this time.