Large Language Model Text Generation Inference
-
Updated
May 22, 2024 - Python
Large Language Model Text Generation Inference
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Tevatron - A flexible toolkit for neural retrieval research and development.
A solid foundational understanding of XAI, primarily emphasizing how XAI methodologies can expose latent biases in datasets and reveal valuable insights.
A high-throughput and memory-efficient inference and serving engine for LLMs
A high-performance inference system for large language models, designed for production environments.
This repository contains the python package for Helical
JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs welcome).
An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.
Pytorch implementation of "Genie: Generative Interactive Environments", Bruce et al. (2024).
Official implementation of DrugGEN
Implementations of Deep Learning Techniques
An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites
A framework for few-shot evaluation of language models.
3D-sViT-UNET: An Effective framework for Enhanced Brain glioma Segmentation
Transformer for galaxy images (and general astronomy)
Port of OpenAI's Whisper model in C/C++
Self-Created Tools to convert ONNX files (NCHW) to TensorFlow/TFLite/Keras format (NHWC). The purpose of this tool is to solve the massive Transpose extrapolation problem in onnx-tensorflow (onnx-tf). I don't need a Star, but give me a pull request.
Neural network potentials
Add a description, image, and links to the transformer topic page so that developers can more easily learn about it.
To associate your repository with the transformer topic, visit your repo's landing page and select "manage topics."