pypi.org : ai-dynamo-vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
Registry
-
Source
- Documentation
- JSON
purl: pkg:pypi/ai-dynamo-vllm
Keywords:
amd
, cuda
, deepseek
, gpt
, hpu
, inference
, inferentia
, llama
, llm
, llm-serving
, llmops
, mlops
, model-serving
, pytorch
, qwen
, rocm
, tpu
, trainium
, transformer
, xpu
License: Apache-2.0
Latest release: about 1 month ago
First release: about 1 month ago
Downloads: 1,851 last month
Stars: 44,312 on GitHub
Forks: 6,781 on GitHub
Total Commits: 4106
Committers: 701
Average commits per author: 5.857
Development Distribution Score (DDS): 0.882
More commit stats: commits.ecosyste.ms
See more repository details: repos.ecosyste.ms
Funding links: https://github.com/sponsors/vllm-project, https://opencollective.com/vllm
Last synced: 2 days ago