Ecosyste.ms: OpenCollective

An open API service for software projects hosted on Open Collective.

vLLM

vLLM is a high-throughput and memory-efficient inference and serving engine for large language models (LLMs).
Collective - Host: opensource - https://opencollective.com/vllm - Code: https://github.com/vllm-project/vllm

[Kernel] Use flashinfer for decoding

github.com/vllm-project/vllm - LiuXiaoxuanPKU opened this pull request 6 months ago
[Kernel] Support Fp8 Checkpoints (Dynamic + Static)

github.com/vllm-project/vllm - robertgshaw2-neuralmagic opened this pull request 6 months ago
[Misc] Upgrade outlines to v0.0.41

github.com/vllm-project/vllm - psykhi opened this pull request 6 months ago
Add logger extra

github.com/vllm-project/vllm - olehviniarchyk opened this pull request 6 months ago
[Core] Consolidate prompt arguments to LLM engines

github.com/vllm-project/vllm - DarkLight1337 opened this pull request 6 months ago
[CI] check size of the wheels

github.com/vllm-project/vllm - simon-mo opened this pull request 6 months ago
[New Model]: Support Phi-3

github.com/vllm-project/vllm - alexkreidler opened this issue 6 months ago
[New Model]: Llama 3 8B Instruct

github.com/vllm-project/vllm - K-Mistele opened this issue 6 months ago
[Speculative decoding] CUDA graph support

github.com/vllm-project/vllm - heeju-kim2 opened this pull request 6 months ago
[WIP] Infrastructure for encoder/decoder support

github.com/vllm-project/vllm - afeldman-nm opened this pull request 6 months ago
add standalone_api_server

github.com/vllm-project/vllm - alex-k-cart opened this pull request 6 months ago
[CI/Build] AMD CI pipeline with extended set of tests.

github.com/vllm-project/vllm - Alexei-V-Ivanov-AMD opened this pull request 6 months ago
[Speculative decoding] Fix async executing

github.com/vllm-project/vllm - zxdvd opened this pull request 6 months ago
[Bug]: Ray memory leak

github.com/vllm-project/vllm - saattrupdan opened this issue 6 months ago
[Model] Add moondream vision language model

github.com/vllm-project/vllm - vikhyat opened this pull request 6 months ago
[Bugfix] Fix marlin kernel crash on H100

github.com/vllm-project/vllm - alexm-neuralmagic opened this pull request 6 months ago
[Usage]: Make request to LLAVA server.

github.com/vllm-project/vllm - premg16 opened this issue 6 months ago
[Frontend] Support GPT-4V Chat Completions API

github.com/vllm-project/vllm - DarkLight1337 opened this pull request 6 months ago
[Model] Initial support for LLaVA-NeXT

github.com/vllm-project/vllm - DarkLight1337 opened this pull request 6 months ago
[Core] Support image processor

github.com/vllm-project/vllm - DarkLight1337 opened this pull request 6 months ago
[Misc]: optimize eager mode host time

github.com/vllm-project/vllm - functionxu123 opened this pull request 6 months ago
Adding max queue time parameter

github.com/vllm-project/vllm - KrishnaM251 opened this pull request 6 months ago
[Usage]: Llama 3 8B Instruct Inference

github.com/vllm-project/vllm - aliozts opened this issue 6 months ago
[Feature]: AMD ROCm 6.1 Support

github.com/vllm-project/vllm - kannan-scalers-ai opened this issue 6 months ago
[Feature]: Phi2 LoRA support

github.com/vllm-project/vllm - zero-or-one opened this issue 6 months ago
[Misc]Add customized information for models

github.com/vllm-project/vllm - jeejeelee opened this pull request 6 months ago
[Bug]: Invalid Device Ordinal on ROCm

github.com/vllm-project/vllm - Bellk17 opened this issue 6 months ago
[Misc] [CI]: AMD test flaky on main CI

github.com/vllm-project/vllm - cadedaniel opened this issue 6 months ago
[Model] Jamba support

github.com/vllm-project/vllm - mzusman opened this pull request 6 months ago
[CI/BUILD] enable intel queue for longer CPU tests

github.com/vllm-project/vllm - zhouyuan opened this pull request 6 months ago
[Bug]: --engine-use-ray is broken. #4100

github.com/vllm-project/vllm - jdinalt opened this pull request 6 months ago
[Bug]: guided_json bad output for llama2-13b

github.com/vllm-project/vllm - pseudotensor opened this issue 6 months ago
[Model] Adding support for MiniCPM-V

github.com/vllm-project/vllm - HwwwwwwwH opened this pull request 6 months ago
[Bug]: vllm_C is missing.

github.com/vllm-project/vllm - Calvinnncy97 opened this issue 6 months ago
[Model] Add support for 360zhinao

github.com/vllm-project/vllm - garycaokai opened this pull request 6 months ago
[Bug]: RuntimeError: Unknown layout

github.com/vllm-project/vllm - zzlgreat opened this issue 6 months ago
[Usage]: Problem when loading my trained model.

github.com/vllm-project/vllm - hummingbird2030 opened this issue 6 months ago
[Feature]: bitsandbytes support

github.com/vllm-project/vllm - orellavie1212 opened this issue 6 months ago
[Frontend] Refactor prompt processing

github.com/vllm-project/vllm - DarkLight1337 opened this pull request 6 months ago
[Bug]: start api server stuck

github.com/vllm-project/vllm - QianguoS opened this issue 6 months ago
[Core] Support LoRA on quantized models

github.com/vllm-project/vllm - jeejeelee opened this pull request 6 months ago