A High-Performance LLM Inference Engine with vLLM-Style Continuous Batching
-
Updated
Jan 2, 2026 - C++
A High-Performance LLM Inference Engine with vLLM-Style Continuous Batching
gLLM: Global Balanced Pipeline Parallelism System for Distributed LLM Serving with Token Throttling
OpenAI-compatible server with continuous batching for MLX on Apple Silicon
Fork of OpenAI and Anthropic compatible server for Apple Silicon. Native MLX backend, 500+ tok/s. Run LLMs and vision-language models with continuous batching, MCP tool calling, and multimodal support.
Add a description, image, and links to the continuous-batching topic page so that developers can more easily learn about it.
To associate your repository with the continuous-batching topic, visit your repo's landing page and select "manage topics."