∇ Valyu LLManager simplifies and scales LLM application deployment, reducing infrastructure complexity and costs.
-
Updated
May 17, 2024 - Python
∇ Valyu LLManager simplifies and scales LLM application deployment, reducing infrastructure complexity and costs.
A framework for few-shot evaluation of autoregressive language models.
Automating the deployment of the Takeoff Server on AWS for LLMs
Deep learning environment setups
Okik is a command-line interface (CLI) tool for LLM, RAG and model serving.
Streaming of LLM responses in realtime using Fastapi and Streamlit.
Building Static Web Applications using Large Language Model. From hand sketched documents, images and screenshots to proper web pages.
internet llm - access your ollama (or any other local llm) instance from across the internet
You can run any large language model on your local machine with this repository.
A Production-Ready, Scalable RAG-powered LLM-based Context-Aware QA App
A guide on how to run LLMs on intel CPUs
Lightweight and extensible LLM Inference serving benchmark tool written in Rust.
An Easy-to-use Knowledge Editing Framework for LLMs.
Hinglish Chatbot powered by Azure Cognitive Services, Google Translate and Open AI
Deno LLM API Service
This repository demonstrates LLM execution on CPUs using packages like llamafile, emphasizing low-latency, high-throughput, and cost-effective benefits for inference and serving.
It is a comprehensive resource hub compiling all LLM papers accepted at the International Conference on Learning Representations (ICLR) in 2024.
A self-hosted personal chatbot API with FastAPI. It allows you to interact with the Llama2 LLM (and other open-source LLMs) to have natural language conversations, generate text, and perform various language-related tasks.
Add a description, image, and links to the llm-serving topic page so that developers can more easily learn about it.
To associate your repository with the llm-serving topic, visit your repo's landing page and select "manage topics."