Stars
A high-throughput and memory-efficient inference and serving engine for LLMs
The Torch-MLIR project aims to provide first class support from the PyTorch ecosystem to the MLIR ecosystem.
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
IREE's PyTorch Frontend, based on Torch Dynamo.
LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalability, and high-speed performance.
Llama中文社区,实时汇总最新Llama学习资料,构建最好的中文Llama大模型开源生态,完全开源可商用
A retargetable MLIR-based machine learning compiler and runtime toolkit.
Development repository for the Triton language and compiler
Shared Middle-Layer for Triton Compilation
Development repository for the Triton-Linalg conversion
Fast and memory-efficient exact attention
Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.
An MLIR-based compiler framework bridges DSLs (domain-specific languages) to DSAs (domain-specific architectures).
Machine learning compiler based on MLIR for Sophgo TPU.


