Ecosyste.ms: OpenCollective

An open API service for software projects hosted on Open Collective.

vLLM

vLLM is a high-throughput and memory-efficient inference and serving engine for large language models (LLMs).
Collective - Host: opensource - https://opencollective.com/vllm - Code: https://github.com/vllm-project/vllm

[core] [3/N] multi-step args and sequence.py

github.com/vllm-project/vllm - SolitaryThinker opened this pull request 2 months ago
[misc] Add Torch profiler support

github.com/vllm-project/vllm - SolitaryThinker opened this pull request 2 months ago
support bitsandbytes 8-bit and FP4 quantized models

github.com/vllm-project/vllm - chenqianfzh opened this pull request 2 months ago
[Bug][Frontend] Add and test client timeouts

github.com/vllm-project/vllm - joerunde opened this pull request 2 months ago
[Core][Model][Frontend] Model architecture plugins

github.com/vllm-project/vllm - NadavShmayo opened this pull request 2 months ago
[Misc] update fp8 to use `vLLMParameter`

github.com/vllm-project/vllm - dsikka opened this pull request 2 months ago
[Model] Adding Granite model.

github.com/vllm-project/vllm - shawntan opened this pull request 2 months ago
[Usage]: GPTQ quantization behavior

github.com/vllm-project/vllm - onlinex opened this issue 2 months ago
Simplify Jamba state management

github.com/vllm-project/vllm - tlrmchlsmth opened this pull request 2 months ago
[misc][plugin] add plugin system implementation

github.com/vllm-project/vllm - youkaichao opened this pull request 2 months ago
[New Model]: LLaVA-OneVision

github.com/vllm-project/vllm - EthanZoneCoding opened this issue 2 months ago
[Misc]: How to use intel-gpu in openvino

github.com/vllm-project/vllm - liuxingbin opened this issue 2 months ago
[Kernel] W8A16 Int8 inside FusedMoE

github.com/vllm-project/vllm - mzusman opened this pull request 2 months ago
[VLM][Model] Add test for InternViT vision encoder

github.com/vllm-project/vllm - Isotr0py opened this pull request 2 months ago
[CI/Build] Minor refactoring for vLLM assets

github.com/vllm-project/vllm - ywang96 opened this pull request 2 months ago
[misc] add commit id in collect env

github.com/vllm-project/vllm - youkaichao opened this pull request 2 months ago
[Usage]: KV Cache Warning for `gemma2`

github.com/vllm-project/vllm - robertgshaw2-neuralmagic opened this issue 2 months ago
[Core] Move detokenization to front-end process

github.com/vllm-project/vllm - njhill opened this pull request 2 months ago
[Bug]: Dockerfile build error

github.com/vllm-project/vllm - palash-fin opened this issue 2 months ago
[Bugfix][Frontend] Fix Issues Under High Load With `zeromq` Frontend

github.com/vllm-project/vllm - robertgshaw2-neuralmagic opened this pull request 2 months ago
[Core]RequestMetrics add preempt metrics

github.com/vllm-project/vllm - zeroorhero opened this pull request 2 months ago
[Bugfix] Fix ITL recording in serving benchmark

github.com/vllm-project/vllm - ywang96 opened this pull request 2 months ago
[CI/Build][ROCm] Enabling LoRA tests on ROCm

github.com/vllm-project/vllm - alexeykondrat opened this pull request 2 months ago
[Installation]: git clone cutlass fails

github.com/vllm-project/vllm - paolovic opened this issue 2 months ago
[FrontEnd] Keep RPC server tcp protocol

github.com/vllm-project/vllm - esmeetu opened this pull request 2 months ago
[Performance] e2e overheads reduction: Small followup diff

github.com/vllm-project/vllm - alexm-neuralmagic opened this pull request 2 months ago
Create speculative decode dynamic parallel strategy

github.com/vllm-project/vllm - vladislavkruglikov opened this pull request 2 months ago
Create parallel scorer

github.com/vllm-project/vllm - vladislavkruglikov opened this pull request 2 months ago
Allow model executor to return many next tokens

github.com/vllm-project/vllm - vladislavkruglikov opened this pull request 2 months ago
[Bugfix] Fix reinit procedure in ModelInputForGPUBuilder

github.com/vllm-project/vllm - alexm-neuralmagic opened this pull request 2 months ago
Create draft from random tokens from promt

github.com/vllm-project/vllm - vladislavkruglikov opened this pull request 2 months ago
Save speculative decoding states

github.com/vllm-project/vllm - vladislavkruglikov opened this pull request 2 months ago
Create speculative decode dynamic parallel strategy

github.com/vllm-project/vllm - vladislavkruglikov opened this issue 2 months ago
[Bug]: internvl2-8b提问无限循环

github.com/vllm-project/vllm - haoduoyu1203 opened this issue 2 months ago
[Bug]: internvl2-8b 提问无限循环回答

github.com/vllm-project/vllm - haoduoyu1203 opened this issue 2 months ago
[Core] RequestMetrics add preempt metrics

github.com/vllm-project/vllm - zeroorhero opened this pull request 2 months ago
[Bug]: LLama3 LoRA load failed

github.com/vllm-project/vllm - victorlwchen opened this issue 2 months ago
[TPU] Use mark_dynamic to reduce compilation time

github.com/vllm-project/vllm - WoosukKwon opened this pull request 2 months ago
[Misc] Update Fused MoE weight loading

github.com/vllm-project/vllm - dsikka opened this pull request 2 months ago
[Core] Factor out input preprocessing to a separate class

github.com/vllm-project/vllm - DarkLight1337 opened this pull request 2 months ago
[CI/Build] Add e2e correctness in oai

github.com/vllm-project/vllm - robertgshaw2-neuralmagic opened this pull request 2 months ago
[Misc/Testing] Use `torch.testing.assert_close`

github.com/vllm-project/vllm - jon-chuang opened this pull request 2 months ago