Ecosyste.ms: OpenCollective

An open API service for software projects hosted on Open Collective.

vLLM

vLLM is a high-throughput and memory-efficient inference and serving engine for large language models (LLMs).
Collective - Host: opensource - https://opencollective.com/vllm - Code: https://github.com/vllm-project/vllm

[Kernel][Misc] dynamo support for ScalarType

github.com/vllm-project/vllm - bnellnm opened this pull request 2 months ago
[Bug]: Gemma 2 9b errors

github.com/vllm-project/vllm - nivibilla opened this issue 2 months ago
[Kernel] register punica functions as torch ops

github.com/vllm-project/vllm - bnellnm opened this pull request 2 months ago
[do-not-merge] test PR

github.com/vllm-project/vllm - mgoin opened this pull request 2 months ago
Merge main

github.com/vllm-project/vllm - niuzheng168 opened this pull request 2 months ago
[misc] use nvml to get consistent device name

github.com/vllm-project/vllm - youkaichao opened this pull request 2 months ago
[Bug]: run quantized model error

github.com/vllm-project/vllm - soulzzz opened this issue 2 months ago
[Bug]: vllm运行卡住

github.com/vllm-project/vllm - backtime1 opened this issue 2 months ago
[Core.aDAG] Temporarily turn off NCCL in aDAG tests

github.com/vllm-project/vllm - ruisearch42 opened this pull request 2 months ago
[Core] Use uvloop with zmq-decoupled front-end

github.com/vllm-project/vllm - njhill opened this pull request 2 months ago
[Prototype] Create and use custom NCCL group for aDAG

github.com/vllm-project/vllm - ruisearch42 opened this pull request 2 months ago
Varun/multi step chunked prefill

github.com/vllm-project/vllm - varun-sundar-rabindranath opened this pull request 2 months ago
[Bugfix] neuron: enable tensor parallelism

github.com/vllm-project/vllm - omrishiv opened this pull request 2 months ago
[model] Support for Llava-Next-Video model

github.com/vllm-project/vllm - TKONIY opened this pull request 2 months ago
[RFC]: Support for video input

github.com/vllm-project/vllm - TKONIY opened this issue 2 months ago
[Bugfix]Add sharded_state to load format

github.com/vllm-project/vllm - tjandy98 opened this pull request 2 months ago
[Usage]: Access weight of model with tp=2

github.com/vllm-project/vllm - floatingbigcat opened this issue 2 months ago
[do-not-merge] test PR

github.com/vllm-project/vllm - khluu opened this pull request 2 months ago
[Misc] Revert `compressed-tensors` code reuse

github.com/vllm-project/vllm - kylesayrs opened this pull request 2 months ago
[Feature]: Context Parallelism

github.com/vllm-project/vllm - huseinzol05 opened this issue 2 months ago
[Bug]: AutoAWQ marlin methods error

github.com/vllm-project/vllm - MichoChan opened this issue 2 months ago
[Hardware][CPU] Support AWQ for CPU backend

github.com/vllm-project/vllm - bigPYJ1151 opened this pull request 2 months ago
[Bug]: Docker.xpu build failed

github.com/vllm-project/vllm - liuxingbin opened this issue 2 months ago
support tqdm in notebooks

github.com/vllm-project/vllm - fzyzcjy opened this pull request 2 months ago
[ci] fix model tests

github.com/vllm-project/vllm - youkaichao opened this pull request 2 months ago
[doc] update test script to include cudagraph

github.com/vllm-project/vllm - youkaichao opened this pull request 2 months ago
[CI] Fix crashes of performance benchmark

github.com/vllm-project/vllm - KuntaiDu opened this pull request 2 months ago
[Bug]: Dockerfile Build breaks in local

github.com/vllm-project/vllm - palash-fin opened this issue 2 months ago
add causal parameter for flash attention

github.com/vllm-project/vllm - WanXiaopei opened this pull request 2 months ago
[Bugfix][Docs] Update list of mock imports

github.com/vllm-project/vllm - DarkLight1337 opened this pull request 2 months ago
[misc][ci] fix cpu test with plugins

github.com/vllm-project/vllm - youkaichao opened this pull request 2 months ago
[Bugfix][CI] Import ray under guard

github.com/vllm-project/vllm - WoosukKwon opened this pull request 2 months ago
Announce NVIDIA Meetup

github.com/vllm-project/vllm - simon-mo opened this pull request 2 months ago
Release v0.5.5

github.com/vllm-project/vllm - simon-mo opened this issue 2 months ago
[Bug]: aqlm test failing on H100

github.com/vllm-project/vllm - bnellnm opened this issue 2 months ago
[Bug]: Support Falcon Mamba

github.com/vllm-project/vllm - hahmad2008 opened this issue 2 months ago
[CI/Build] Add text-only test for Qwen models

github.com/vllm-project/vllm - alex-jw-brooks opened this pull request 2 months ago
[Feature]: ROCm 6.2 support & FP8 Support

github.com/vllm-project/vllm - ferrybaltimore opened this issue 2 months ago
[TPU] Support multi-host inference

github.com/vllm-project/vllm - WoosukKwon opened this pull request 2 months ago