Open deep learning compiler stack for cpu, gpu and specialized accelerators
-
Updated
May 29, 2024 - Python
Open deep learning compiler stack for cpu, gpu and specialized accelerators
Universal LLM Deployment Engine with ML Compilation
High-performance In-browser LLM Inference Engine
Bringing stable diffusion models to web browsers. Everything runs inside the browser with no server support.
yolort is a runtime stack for yolov5 on specialized accelerators such as tensorrt, libtorch, onnxruntime, tvm and ncnn.
TVM Documentation in Chinese Simplified / TVM 中文文档
TON Foundation invites talent to imagine and realize projects that have the potential to integrate with the daily lives of users.
AutoKernel 是一个简单易用,低门槛的自动算子优化工具,提高深度学习算法部署效率。
Quantization library for PyTorch. Support low-precision and mixed-precision quantization, with hardware implementation through TVM.
A home for the final text of all TVM RFCs.
Solidity compiler for TVM
Open, Modular, Deep Learning Accelerator
FlashInfer: Kernel Library for LLM Serving
比做算法的懂工程落地,比做工程的懂算法模型。
Optimizing Mobile Deep Learning on ARM GPU with TVM
Benchmark scripts for TVM
Large input size REAL-TIME Face Detector on Cpp. It can also support face verification using MobileFaceNet+Arcface with real-time inference. 480P Over 30FPS on CPU
Add a description, image, and links to the tvm topic page so that developers can more easily learn about it.
To associate your repository with the tvm topic, visit your repo's landing page and select "manage topics."