SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.
-
Updated
May 23, 2024 - Python
SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.
An efficient, flexible and full-featured toolkit for fine-tuning large models (InternLM2, Llama3, Phi3, Qwen, Mistral, ...)
Linux LiveCD for offline AI training and inference.
Low-code framework for building custom LLMs, neural networks, and other AI models
Backend for the AI-copilot
DLRover: An Automatic Distributed Deep Learning System
SkyPilot: Run LLMs, AI, and Batch jobs on any cloud. Get maximum savings, highest GPU availability, and managed execution—all with a simple interface.
Collection of best practices, reference architectures, model training examples and utilities to train large models on AWS.
H2O LLM Studio - a framework and no-code GUI for fine-tuning LLMs. Documentation: https://h2oai.github.io/h2o-llmstudio/
A Python package for automatically training and comparing language models.
LLM-PowerHouse: Unleash LLMs' potential through curated tutorials, best practices, and ready-to-use code for custom training and inferencing.
Popular Large Language Models from scratch - 2024
Nvidia GPU exporter for prometheus using nvidia-smi binary
Run GPU inference and training jobs on serverless infrastructure that scales with you.
The Arcee client for executing domain-adpated language model routines
Fine-tuning of Flan-5T LLM for text classification
LLM (Large Language Model) FineTuning
ICLR 2024 论文和开源项目合集
Add a description, image, and links to the llm-training topic page so that developers can more easily learn about it.
To associate your repository with the llm-training topic, visit your repo's landing page and select "manage topics."