Ecosyste.ms: OpenCollective

An open API service for software projects hosted on Open Collective.

vLLM

vLLM is a high-throughput and memory-efficient inference and serving engine for large language models (LLMs).
Collective - Host: opensource - https://opencollective.com/vllm - Code: https://github.com/vllm-project/vllm

[Misc]Further reduce BNB static variable

github.com/vllm-project/vllm - jeejeelee opened this pull request about 1 month ago
[Interleaved ATTN] Support for Mistral-8B

github.com/vllm-project/vllm - patrickvonplaten opened this pull request about 1 month ago
[Bug]: Duplicate request_id breaks the engine

github.com/vllm-project/vllm - tjohnson31415 opened this issue about 1 month ago
[Core] Update to outlines >= 0.1.8

github.com/vllm-project/vllm - russellb opened this pull request about 1 month ago
[Hardware][Intel-Gaudi] Enable LoRA support for Intel Gaudi (HPU)

github.com/vllm-project/vllm - SanjuCSudhakaran opened this pull request about 1 month ago
[Docs] Add dedicated tool calling page to docs

github.com/vllm-project/vllm - mgoin opened this pull request about 1 month ago
[CI][Installation] Avoid uploading CUDA 11.8 wheel

github.com/vllm-project/vllm - cermeng opened this pull request about 1 month ago
[Usage]: Fail to load config.json

github.com/vllm-project/vllm - dequeueing opened this issue about 1 month ago
Add Sageattention backend

github.com/vllm-project/vllm - flozi00 opened this pull request about 1 month ago
[8/N] enable cli flag without a space

github.com/vllm-project/vllm - youkaichao opened this pull request about 1 month ago
[V1] Fix Compilation config & Enable CUDA graph by default

github.com/vllm-project/vllm - WoosukKwon opened this pull request about 1 month ago
[Bug]: Gemma2 becomes a fool.

github.com/vllm-project/vllm - Foreist opened this issue about 1 month ago
[Kernel] Register punica ops directly

github.com/vllm-project/vllm - jeejeelee opened this pull request about 1 month ago
[platforms] improve error message for unspecified platforms

github.com/vllm-project/vllm - youkaichao opened this pull request about 1 month ago
[Feature]: Manually inject Prefix KV Cache

github.com/vllm-project/vllm - toilaluan opened this issue about 1 month ago
[Model]: Add support for Aria model

github.com/vllm-project/vllm - xffxff opened this pull request about 1 month ago
[Doc] fix a small typo in docstring of llama_tool_parser

github.com/vllm-project/vllm - FerdinandZhong opened this pull request about 1 month ago
[Feature]: Multimodel prefix-caching features

github.com/vllm-project/vllm - justzhanghong opened this issue about 1 month ago
[Usage]:

github.com/vllm-project/vllm - Lukas-123 opened this issue about 1 month ago
[Platforms] Add `device_type` in `Platform`

github.com/vllm-project/vllm - MengqingCao opened this pull request about 1 month ago
Need to update the jax and jaxlib version

github.com/vllm-project/vllm - vanbasten23 opened this pull request about 1 month ago
Turn on V1 for H200 build

github.com/vllm-project/vllm - simon-mo opened this pull request about 1 month ago
Metrics model name when using multiple loras

github.com/vllm-project/vllm - mces89 opened this issue about 1 month ago
[Model] Add OLMo November 2024 model

github.com/vllm-project/vllm - 2015aroras opened this pull request about 1 month ago
[Core] Implement disagg prefill by StatelessProcessGroup

github.com/vllm-project/vllm - KuntaiDu opened this pull request about 1 month ago
Support softcap in ROCm Flash Attention

github.com/vllm-project/vllm - hliuca opened this pull request about 1 month ago
[CI/Build] Dockerfile build for ARM64 / GH200

github.com/vllm-project/vllm - drikster80 opened this pull request about 1 month ago
[Bugfix] GPU memory profiling should be per LLM instance

github.com/vllm-project/vllm - tjohnson31415 opened this pull request about 1 month ago
[Frontend] Add Command-R and Llama-3 chat template

github.com/vllm-project/vllm - ccs96307 opened this pull request about 1 month ago
[Misc] Increase default video fetch timeout

github.com/vllm-project/vllm - DarkLight1337 opened this pull request about 1 month ago
[V1] Replace traversal search with lookup table

github.com/vllm-project/vllm - Abatom opened this pull request about 1 month ago
[torch.compile] limit inductor threads and lazy import quant

github.com/vllm-project/vllm - youkaichao opened this pull request about 1 month ago
[Usage]: VSCode debugger is hanging

github.com/vllm-project/vllm - jeejeelee opened this issue about 1 month ago
[Usage]: Cant use vllm on a multiGPU node

github.com/vllm-project/vllm - 4k1s opened this issue about 1 month ago
[Core] Add Sliding Window Support with Flashinfer

github.com/vllm-project/vllm - pavanimajety opened this pull request about 1 month ago
[Pixtral-Large] Pixtral actually has no bias in vision-lang adapter

github.com/vllm-project/vllm - patrickvonplaten opened this pull request about 1 month ago
[misc][plugin] improve plugin loading

github.com/vllm-project/vllm - youkaichao opened this pull request about 1 month ago
[Misc] Avoid misleading warning messages

github.com/vllm-project/vllm - jeejeelee opened this pull request about 1 month ago
[6/N] torch.compile rollout to users

github.com/vllm-project/vllm - youkaichao opened this pull request about 2 months ago
[ci/build] Have dependabot ignore all patch update

github.com/vllm-project/vllm - khluu opened this pull request about 2 months ago
Compressed tensors w8a8 tpu

github.com/vllm-project/vllm - robertgshaw2-neuralmagic opened this pull request about 2 months ago
[CI/Build] Update Dockerfile.rocm

github.com/vllm-project/vllm - Alexei-V-Ivanov-AMD opened this pull request about 2 months ago
[ci][bugfix] fix kernel tests

github.com/vllm-project/vllm - youkaichao opened this pull request about 2 months ago
[Bugfix] Guard for negative counter metrics to prevent crash

github.com/vllm-project/vllm - tjohnson31415 opened this pull request about 2 months ago
[Bug]: rocm issue

github.com/vllm-project/vllm - YYXLN opened this issue about 2 months ago
[Doc]: Pages were moved without a redirect

github.com/vllm-project/vllm - shannonxtreme opened this issue about 2 months ago
[Doc]: Migrate to Markdown

github.com/vllm-project/vllm - rafvasq opened this issue about 2 months ago
Fix open_collective value in FUNDING.yml

github.com/vllm-project/vllm - andrew opened this pull request about 2 months ago
[Doc] Update doc for LoRA support in GLM-4V

github.com/vllm-project/vllm - B-201 opened this pull request about 2 months ago
[Misc] Reduce medusa weight

github.com/vllm-project/vllm - skylee-01 opened this pull request about 2 months ago
Fix: Build error seen on Power Architecture

github.com/vllm-project/vllm - mikejuliet13 opened this pull request about 2 months ago
[Model][LoRA]LoRA support added for glm-4v

github.com/vllm-project/vllm - B-201 opened this pull request about 2 months ago
[Bugfix]Fix Phi-3 BNB online quantization

github.com/vllm-project/vllm - jeejeelee opened this pull request about 2 months ago
[Model] Remove transformers attention porting in VITs

github.com/vllm-project/vllm - Isotr0py opened this pull request about 2 months ago
Bump the patch-update group with 2 updates

github.com/vllm-project/vllm - dependabot[bot] opened this pull request about 2 months ago