Skip to content

NX-AI/vllm

About

A high-throughput and memory-efficient inference and serving engine for LLMs

Resources

License

Code of conduct

Contributing

Security policy

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 85.1%
  • Cuda 9.5%
  • C++ 3.8%
  • Shell 0.7%
  • C 0.5%
  • CMake 0.3%
  • Other 0.1%