Ecosyste.ms: OpenCollective

An open API service for software projects hosted on Open Collective.

vLLM

vLLM is a high-throughput and memory-efficient inference and serving engine for large language models (LLMs).
Collective - Host: opensource - https://opencollective.com/vllm - Code: https://github.com/vllm-project/vllm

[Installation]: LGPL license in dependencies

github.com/vllm-project/vllm - laurens-gs opened this issue about 2 months ago
[MODEL] Qwen Multimodal Support (Qwen-VL / Qwen-VL-Chat)

github.com/vllm-project/vllm - alex-jw-brooks opened this pull request about 2 months ago
ppc64le: Dockerfile fixed, and a script for buildkite

github.com/vllm-project/vllm - sumitd2 opened this pull request about 2 months ago
[Bug]: Trailing newline as outputs

github.com/vllm-project/vllm - dawu415 opened this issue about 2 months ago
[Core][Kernel][Misc] Support external swapper for vllm

github.com/vllm-project/vllm - zeroorhero opened this pull request about 2 months ago
Neuron cache blocks must be 1 more than max num seqs

github.com/vllm-project/vllm - ajayvohra2005 opened this pull request about 2 months ago
[RFC]: Build `vllm-flash-attn` from source

github.com/vllm-project/vllm - ProExpertProg opened this issue about 2 months ago
[WIP] Multi Step Chunked Prefill - Prefill Steps

github.com/vllm-project/vllm - varun-sundar-rabindranath opened this pull request about 2 months ago
[New Model]: LlavaQwen2ForCausalLM

github.com/vllm-project/vllm - Chuyun-Shen opened this issue about 2 months ago
[Misc] Update `GPTQ` to use `vLLMParameters`

github.com/vllm-project/vllm - dsikka opened this pull request about 2 months ago
[Misc] Update fbgemmfp8 to use `vLLMParameters`

github.com/vllm-project/vllm - dsikka opened this pull request about 2 months ago
[Misc] Remove `SqueezeLLM`

github.com/vllm-project/vllm - dsikka opened this pull request about 2 months ago
Roberta embedding

github.com/vllm-project/vllm - maxdebayser opened this pull request about 2 months ago
[Bug]: Multistep with n>1 Fails

github.com/vllm-project/vllm - robertgshaw2-neuralmagic opened this issue about 2 months ago
[Model] Add Ultravox support for multiple audio chunks

github.com/vllm-project/vllm - petersalas opened this pull request about 2 months ago
[Bug]: segfault when loading MoE models

github.com/vllm-project/vllm - nivibilla opened this issue about 2 months ago
[Feature]: Context Caching

github.com/vllm-project/vllm - RonanKMcGovern opened this issue about 2 months ago
[Performance]: vLLM version issue.

github.com/vllm-project/vllm - zjjznw123 opened this issue about 2 months ago
[Bugfix][VLM] Fix incompatibility between #7902 and #7230

github.com/vllm-project/vllm - DarkLight1337 opened this pull request about 2 months ago
[Bug]: deploy multi lora by vllm mode error

github.com/vllm-project/vllm - askcs517 opened this issue about 2 months ago
[ci][test] fix pp test failure

github.com/vllm-project/vllm - youkaichao opened this pull request about 2 months ago
[Bug]:reset LLM for each inference

github.com/vllm-project/vllm - victorzhz111 opened this issue about 2 months ago
[misc] [doc] [frontend] LLM torch profiler support

github.com/vllm-project/vllm - SolitaryThinker opened this pull request about 2 months ago
[Model] EXAONE 3.0 model support

github.com/vllm-project/vllm - Deepfocused opened this pull request about 2 months ago
[Bug]: Is vllm compatible with torchrun?

github.com/vllm-project/vllm - HwwwwwwwH opened this issue about 2 months ago
[Misc] Use ray[adag] dependency instead of cuda

github.com/vllm-project/vllm - ruisearch42 opened this pull request about 2 months ago
[Doc] fix the autoAWQ example

github.com/vllm-project/vllm - stas00 opened this pull request about 2 months ago
[TPU] Align worker index with node boundary

github.com/vllm-project/vllm - WoosukKwon opened this pull request about 2 months ago
[Core] Add support for recursively loading weights by model ID

github.com/vllm-project/vllm - petersalas opened this pull request about 2 months ago
[mypy][CI/Build] Fix mypy errors

github.com/vllm-project/vllm - DarkLight1337 opened this pull request about 2 months ago
[multi-step] add flashinfer backend

github.com/vllm-project/vllm - SolitaryThinker opened this pull request about 2 months ago
[hardware][rocm] allow rocm to override default env var

github.com/vllm-project/vllm - youkaichao opened this pull request about 2 months ago
[Model] Add OLMoE

github.com/vllm-project/vllm - Muennighoff opened this pull request about 2 months ago
[Core] Combine async postprocessor and multi-step

github.com/vllm-project/vllm - alexm-neuralmagic opened this pull request about 2 months ago
Remove request.max_tokens assertion in serving_completion.py

github.com/vllm-project/vllm - zifeitong opened this pull request about 2 months ago
[benchmark] Update TGI version

github.com/vllm-project/vllm - philschmid opened this pull request about 2 months ago
[Core] Enable Memory Tiering for vLLM

github.com/vllm-project/vllm - PanJason opened this pull request about 2 months ago
[Feature]: Lora for MiniCPM_2_6

github.com/vllm-project/vllm - tristan279 opened this issue about 2 months ago
[Core] Adding Control Vector Support

github.com/vllm-project/vllm - raywanb opened this pull request about 2 months ago
[Model][VLM] Add Qwen2-VL model support

github.com/vllm-project/vllm - fyabc opened this pull request about 2 months ago
[Model] EXAONE 3.0 model support - closed

github.com/vllm-project/vllm - Deepfocused opened this pull request about 2 months ago
[CI/Build][VLM] Cleanup multiple images inputs model test

github.com/vllm-project/vllm - Isotr0py opened this pull request about 2 months ago
extend cuda graph size for H200

github.com/vllm-project/vllm - kushanam opened this pull request about 2 months ago
[Bug]: CUDA_VISIBLE_DEVICES not detected

github.com/vllm-project/vllm - paolovic opened this issue about 2 months ago
[Doc]: Update tensorizer docs to include vllm[tensorizer]

github.com/vllm-project/vllm - sethkimmel3 opened this pull request about 2 months ago
Adding new cutlass configurations for llama70B

github.com/vllm-project/vllm - kushanam opened this pull request about 2 months ago
[Performance]: Prefix-caching aware scheduling

github.com/vllm-project/vllm - comaniac opened this issue about 2 months ago
[Bugfix] Fix single output condition in output processor

github.com/vllm-project/vllm - WoosukKwon opened this pull request about 2 months ago
[CI/Build] Add linting for github actions workflows

github.com/vllm-project/vllm - russellb opened this pull request about 2 months ago