Ecosyste.ms: OpenCollective

An open API service for software projects hosted on Open Collective.

vLLM

vLLM is a high-throughput and memory-efficient inference and serving engine for large language models (LLMs).
Collective - Host: opensource - https://opencollective.com/vllm - Code: https://github.com/vllm-project/vllm

[Bug]: ray error when tp>=2

github.com/vllm-project/vllm - Jimmy-Lu opened this issue 4 months ago
[vlm] Remove vision language config.

github.com/vllm-project/vllm - xwjiang2010 opened this pull request 4 months ago
[ Misc ] Expand Fp8 MoE Support to Qwen

github.com/vllm-project/vllm - robertgshaw2-neuralmagic opened this pull request 4 months ago
[ Misc ] Refactor Marlin Python Utilities

github.com/vllm-project/vllm - robertgshaw2-neuralmagic opened this pull request 4 months ago
[RFC]: Priority Scheduling

github.com/vllm-project/vllm - apatke opened this issue 4 months ago
[Doc] Reinstate doc dependencies

github.com/vllm-project/vllm - DarkLight1337 opened this pull request 4 months ago
[Kernel] Unify the kernel used in flash attention backend

github.com/vllm-project/vllm - LiuXiaoxuanPKU opened this pull request 4 months ago
[Kernel][Model] logits_soft_cap for Gemma2 with flashinfer

github.com/vllm-project/vllm - LiuXiaoxuanPKU opened this pull request 4 months ago
Benchmark: add H100 suite

github.com/vllm-project/vllm - simon-mo opened this pull request 4 months ago
Support for quantized kv cache (compressed-tensors)

github.com/vllm-project/vllm - dbogunowicz opened this pull request 4 months ago
model test with cache

github.com/vllm-project/vllm - khluu opened this pull request 4 months ago
[New Model]: facebook/seamless-m4t-v2-large

github.com/vllm-project/vllm - frittentheke opened this issue 4 months ago
Test HF cache

github.com/vllm-project/vllm - khluu opened this pull request 4 months ago
[ci][distributed] fix phi-3v test failure

github.com/vllm-project/vllm - youkaichao opened this pull request 4 months ago
[CI/Build] Reuse code for checking output consistency

github.com/vllm-project/vllm - DarkLight1337 opened this pull request 4 months ago
[Frontend] Bad words sampling parameter

github.com/vllm-project/vllm - Alvant opened this pull request 4 months ago
about the RotaryEmbedding

github.com/vllm-project/vllm - tricky61 opened this issue 4 months ago
[Misc] Update Phi-3-Vision Example

github.com/vllm-project/vllm - ywang96 opened this pull request 4 months ago
[misc][doc] try to add warning for latest html

github.com/vllm-project/vllm - youkaichao opened this pull request 4 months ago
[Bugfix][TPU] Fix TPU sampler output

github.com/vllm-project/vllm - WoosukKwon opened this pull request 4 months ago
[Bugfix][TPU] Fix pad slot id

github.com/vllm-project/vllm - WoosukKwon opened this pull request 4 months ago
test yum install

github.com/vllm-project/vllm - khluu opened this pull request 4 months ago
[Misc] Fix `get_min_capability`

github.com/vllm-project/vllm - dsikka opened this pull request 4 months ago
[ Misc ] Isolate Fp8Moe From Mixtral

github.com/vllm-project/vllm - robertgshaw2-neuralmagic opened this pull request 4 months ago
[CI/Build] [3/3] Reorganize entrypoints tests

github.com/vllm-project/vllm - DarkLight1337 opened this pull request 4 months ago
Whisper support

github.com/vllm-project/vllm - huseinzol05 opened this pull request 4 months ago
[Bugfix] Fix Engine Failing After Invalid Request - AsyncEngineDeadError

github.com/vllm-project/vllm - robertgshaw2-neuralmagic opened this pull request 4 months ago
Unmark more files as executable

github.com/vllm-project/vllm - tlrmchlsmth opened this pull request 4 months ago
[Core] Adding Priority Scheduling

github.com/vllm-project/vllm - apatke opened this pull request 4 months ago
[Bug]: qwen1.5-32b-chat no response

github.com/vllm-project/vllm - linpan opened this issue 4 months ago
Add support for multi-node on CI

github.com/vllm-project/vllm - khluu opened this pull request 4 months ago
[Bugfix] Support `eos_token_id` from `config.json`

github.com/vllm-project/vllm - DarkLight1337 opened this pull request 4 months ago
Gemma2 models from google

github.com/vllm-project/vllm - bks5881 opened this issue 4 months ago
add FAQ doc under 'serving'

github.com/vllm-project/vllm - llmpros opened this pull request 4 months ago
[Usage]: can I save log to a file?

github.com/vllm-project/vllm - chenchunhui97 opened this issue 4 months ago
[Kernel] Add per-tensor and per-token AZP epilogues

github.com/vllm-project/vllm - ProExpertProg opened this pull request 4 months ago
[New Model]: Florence-2

github.com/vllm-project/vllm - localbarrage opened this issue 4 months ago
[Misc] Extend vLLM Metrics logging API

github.com/vllm-project/vllm - SolitaryThinker opened this pull request 4 months ago
[ Bugfix ] Enabling Loading Models With Fused QKV/MLP on Disk with FP8

github.com/vllm-project/vllm - robertgshaw2-neuralmagic opened this pull request 4 months ago
[Model] Initial support for BLIP-2

github.com/vllm-project/vllm - DarkLight1337 opened this pull request 4 months ago
[Bugfix] Fix precisions in Gemma 1

github.com/vllm-project/vllm - WoosukKwon opened this pull request 4 months ago
[Model] Add Gemma 2

github.com/vllm-project/vllm - WoosukKwon opened this pull request 4 months ago
[Bug]: TRACKING ISSUE: CUDA OOM with Logprobs

github.com/vllm-project/vllm - robertgshaw2-neuralmagic opened this issue 4 months ago
[Bug]: TRACKING ISSUE: `AsyncEngineDeadError`

github.com/vllm-project/vllm - robertgshaw2-neuralmagic opened this issue 4 months ago
wikivular

github.com/vllm-project/vllm - Wikivu opened this pull request 4 months ago