Ecosyste.ms: OpenCollective

An open API service for software projects hosted on Open Collective.

vLLM

vLLM is a high-throughput and memory-efficient inference and serving engine for large language models (LLMs).
Collective - Host: opensource - https://opencollective.com/vllm - Code: https://github.com/vllm-project/vllm

[Bug]: 2 nodes serving hanging

github.com/vllm-project/vllm - AlvL1225 opened this issue about 1 month ago
[plugin][torch.compile] allow to add custom compile backend

github.com/vllm-project/vllm - youkaichao opened this pull request about 1 month ago
[Bugfix] Fix code for downloading models from modelscope

github.com/vllm-project/vllm - tastelikefeet opened this pull request about 1 month ago
[BugFix] fix group_topk

github.com/vllm-project/vllm - dsikka opened this pull request about 1 month ago
[Kernel] Factor registrations

github.com/vllm-project/vllm - bnellnm opened this pull request about 1 month ago
[Model] Refactor BLIP/BLIP-2 to support composite model loading

github.com/vllm-project/vllm - DarkLight1337 opened this pull request about 1 month ago
[Misc]: Memory Order in Custom Allreduce

github.com/vllm-project/vllm - HydraQYH opened this issue about 1 month ago
[New Model]: qwen2-audio

github.com/vllm-project/vllm - seetimee opened this issue about 1 month ago
[Kernel] Factor registrations

github.com/vllm-project/vllm - bnellnm opened this pull request about 1 month ago
[Model] Support Solar Model

github.com/vllm-project/vllm - shing100 opened this pull request about 1 month ago
[Core] Multi-Step + Single Step Prefills via Chunked Prefill code path

github.com/vllm-project/vllm - varun-sundar-rabindranath opened this pull request about 1 month ago
[Hardware][intel GPU] bump up ipex version to 2.3

github.com/vllm-project/vllm - jikunshang opened this pull request about 1 month ago
[Usage]: how to shutdown vllm server

github.com/vllm-project/vllm - wiluen opened this issue about 1 month ago
[bugfix] torch profiler bug for single gpu with GPUExecutor

github.com/vllm-project/vllm - SolitaryThinker opened this pull request about 1 month ago
[CI/Build]: Add Bandit security check to workflow

github.com/vllm-project/vllm - ChengyuZhu6 opened this pull request about 1 month ago
[Core][VLM] Add support for placeholder token content hashes

github.com/vllm-project/vllm - petersalas opened this pull request about 1 month ago
[Core][VLM] Add precise multi-modal placeholder tracking

github.com/vllm-project/vllm - petersalas opened this pull request about 1 month ago
Restoring missing CI file.

github.com/vllm-project/vllm - Alexei-V-Ivanov-AMD opened this pull request about 1 month ago
[Feature] Add support for Llama 3.1 and 3.2 tool use

github.com/vllm-project/vllm - maxdebayser opened this pull request about 1 month ago
[Gemma2] add bitsandbytes support for Gemma2

github.com/vllm-project/vllm - blueyo0 opened this pull request about 1 month ago
[misc] CUDA Time Layerwise Profiler

github.com/vllm-project/vllm - LucasWilkinson opened this pull request about 1 month ago
Add output streaming support to multi-step + async

github.com/vllm-project/vllm - alexm-neuralmagic opened this pull request about 1 month ago
[Misc] Skip loading extra bias for Qwen2-MOE GPTQ models

github.com/vllm-project/vllm - jeejeelee opened this pull request about 1 month ago
[CI/Build] Buildkite pipeline generator

github.com/vllm-project/vllm - khluu opened this pull request about 1 month ago
Do vLLM support `input_embeds` as input while using LLama?

github.com/vllm-project/vllm - OswaldoBornemann opened this issue about 1 month ago
[Bug]: 段错误 (核心已转储)

github.com/vllm-project/vllm - LIUKAI0815 opened this issue about 1 month ago
Fix verify tokens with the correct bonus token

github.com/vllm-project/vllm - jiqing-feng opened this pull request about 1 month ago
[Speculative Decoding] Test refactor

github.com/vllm-project/vllm - LiuXiaoxuanPKU opened this pull request about 1 month ago
[Frontend] Clean up type annotations for mistral tokenizer

github.com/vllm-project/vllm - DarkLight1337 opened this pull request about 1 month ago
[Misc] Benchmark for awq_triton kernels

github.com/vllm-project/vllm - rasmith opened this pull request about 1 month ago
Fix ppc64le buildkite job

github.com/vllm-project/vllm - sumitd2 opened this pull request about 1 month ago
[MISC] Dump model runner inputs when crashing

github.com/vllm-project/vllm - comaniac opened this pull request about 1 month ago
[Model] support minicpm3

github.com/vllm-project/vllm - SUDA-HLT-ywfang opened this pull request about 1 month ago
Correct adapter usage for cohere

github.com/vllm-project/vllm - vladislavkruglikov opened this pull request about 1 month ago
[Bugfix] Mapping physical device indices for e2e test utils

github.com/vllm-project/vllm - ShangmingCai opened this pull request about 1 month ago
[Usage]: Throughput and quality issue with vllm 0.6.0.

github.com/vllm-project/vllm - Agrawalchitranshu opened this issue about 1 month ago
[RFC]: More functionality for API control

github.com/vllm-project/vllm - paulcx opened this issue about 1 month ago
[New Model]: Reflection-Llama-3.1-70B

github.com/vllm-project/vllm - sekh77 opened this issue about 1 month ago
[Bugfix] Fix async postprocessor in case of preemption

github.com/vllm-project/vllm - alexm-neuralmagic opened this pull request about 1 month ago
[Bug]: how to set gpu id in code?

github.com/vllm-project/vllm - cqray1990 opened this issue about 1 month ago
[New Model]: Support for Idefics3 8B Llama3

github.com/vllm-project/vllm - costelter opened this issue about 1 month ago