Ecosyste.ms: OpenCollective

An open API service for software projects hosted on Open Collective.

vLLM

vLLM is a high-throughput and memory-efficient inference and serving engine for large language models (LLMs).
Collective - Host: opensource - https://opencollective.com/vllm - Code: https://github.com/vllm-project/vllm

[Usage]: flash_attn vs xformers

github.com/vllm-project/vllm - VeryVery opened this issue 6 months ago
[CI/Build] Reduce race condition in docker build

github.com/vllm-project/vllm - youkaichao opened this pull request 6 months ago
[Bug]: StableLM 12b head size incorrect

github.com/vllm-project/vllm - bjoernpl opened this issue 6 months ago
[Model] LoRA gptbigcode implementation

github.com/vllm-project/vllm - raywanb opened this pull request 6 months ago
[Model] Initialize Fuyu-8B support

github.com/vllm-project/vllm - Isotr0py opened this pull request 6 months ago
[Kernel] PyTorch Labs Fused MoE Kernel Integration

github.com/vllm-project/vllm - robertgshaw2-neuralmagic opened this pull request 6 months ago
[Bug]: killed due to high memory usage

github.com/vllm-project/vllm - xiewf1990 opened this issue 6 months ago
[Bug]: Cannot load lora adapters in WSL 2

github.com/vllm-project/vllm - invokeinnovation opened this issue 6 months ago
[Roadmap] vLLM Roadmap Q2 2024

github.com/vllm-project/vllm - simon-mo opened this issue 7 months ago
[Frontend] openAI entrypoint dynamic adapter load

github.com/vllm-project/vllm - DavidPeleg6 opened this pull request 7 months ago
[Misc]: Implement CPU/GPU swapping in BlockManagerV2

github.com/vllm-project/vllm - Kaiyang-Chen opened this pull request 7 months ago
[Model] Cohere CommandR+

github.com/vllm-project/vllm - saurabhdash2512 opened this pull request 7 months ago
[Bug]: YI:34B在使用上无法停止。

github.com/vllm-project/vllm - cat2353050774 opened this issue 7 months ago
[Feature]: cuda12.2 support

github.com/vllm-project/vllm - s-natsubori opened this issue 7 months ago
[Bug]: CUDA error: invalid argument

github.com/vllm-project/vllm - qingjiaozyn opened this issue 7 months ago
[Bug]: Custom all reduce not work.

github.com/vllm-project/vllm - esmeetu opened this issue 7 months ago
[RFC] Initial Support for CPUs

github.com/vllm-project/vllm - bigPYJ1151 opened this issue 7 months ago
[Kernel] Use flash-attn for decoding

github.com/vllm-project/vllm - skrider opened this pull request 7 months ago
[RFC] Initial Support for Cloud TPUs

github.com/vllm-project/vllm - WoosukKwon opened this issue 7 months ago
[BugFix] Fix Falcon tied embeddings

github.com/vllm-project/vllm - WoosukKwon opened this pull request 7 months ago
[Feature]: Offload Model Weights to CPU

github.com/vllm-project/vllm - chenqianfzh opened this issue 7 months ago
[New Model]: Phi-2 support for LoRA

github.com/vllm-project/vllm - andykhanna opened this issue 7 months ago
[Kernel] Full Tensor Parallelism for LoRA Layers

github.com/vllm-project/vllm - FurtherAI opened this pull request 7 months ago
[Bug]: vllm slows down after a long run

github.com/vllm-project/vllm - momomobinx opened this issue 7 months ago
[New Model]: Please support CogVLM

github.com/vllm-project/vllm - kietna1809 opened this issue 7 months ago
[Misc] Add attention sinks

github.com/vllm-project/vllm - felixzhu555 opened this pull request 7 months ago
[Bug]: Use of LoRAReqeust

github.com/vllm-project/vllm - meiru-cam opened this issue 7 months ago
[Core] Add generic typing to `LRUCache`

github.com/vllm-project/vllm - njhill opened this pull request 7 months ago
[Usage]: Set dtype for VLLM using YAML

github.com/vllm-project/vllm - telekoteko opened this issue 7 months ago
[Misc] add HOST_IP env var

github.com/vllm-project/vllm - youkaichao opened this pull request 7 months ago
Fixes #1556 double free

github.com/vllm-project/vllm - br3no opened this pull request 7 months ago
lm-evaluation-harness broken on master

github.com/vllm-project/vllm - pcmoritz opened this issue 7 months ago
Enable scaled FP8 (e4m3fn) KV cache on ROCm (AMD GPU)

github.com/vllm-project/vllm - AdrianAbeyta opened this pull request 7 months ago