Ecosyste.ms: OpenCollective

An open API service for software projects hosted on Open Collective.

vLLM

vLLM is a high-throughput and memory-efficient inference and serving engine for large language models (LLMs).
Collective - Host: opensource - https://opencollective.com/vllm - Code: https://github.com/vllm-project/vllm

[Bugfix] Fix illegal memory access for lora

github.com/vllm-project/vllm - sfc-gh-zhwang opened this pull request 5 months ago
[Bugfix] Fix KeyError: 1 When Using LoRA adapters

github.com/vllm-project/vllm - BlackBird-Coding opened this pull request 5 months ago
[Feature]: Linear adapter support for Mixtral

github.com/vllm-project/vllm - DhruvaBansal00 opened this issue 5 months ago
[Kernel] Update Cutlass fp8 configs

github.com/vllm-project/vllm - varun-sundar-rabindranath opened this pull request 5 months ago
[CI/Build] Test buildkite monorepo plugin

github.com/vllm-project/vllm - dgoupil opened this pull request 5 months ago
[CI/Build] increase wheel size limit to 200 MB

github.com/vllm-project/vllm - youkaichao opened this pull request 5 months ago
[Feature]:

github.com/vllm-project/vllm - double-vin opened this issue 5 months ago
[Doc] Use intersphinx and update entrypoints docs

github.com/vllm-project/vllm - DarkLight1337 opened this pull request 5 months ago
[New Model]: LLaVA-NeXT-Video support

github.com/vllm-project/vllm - AmazDeng opened this issue 5 months ago
[Bug]: The tail problem

github.com/vllm-project/vllm - ZixinxinWang opened this issue 5 months ago
Add gptq_marlin test to cover bug report #5088

github.com/vllm-project/vllm - alexm-neuralmagic opened this pull request 5 months ago
[Bugfix] Avoid Warnings in SparseML Activation Quantization

github.com/vllm-project/vllm - robertgshaw2-neuralmagic opened this pull request 5 months ago
[Bugfix] Automatically Detect SparseML models

github.com/vllm-project/vllm - robertgshaw2-neuralmagic opened this pull request 5 months ago
New CI template on AWS stack

github.com/vllm-project/vllm - khluu opened this pull request 5 months ago
[Doc][Build] update after removing vllm-nccl

github.com/vllm-project/vllm - youkaichao opened this pull request 5 months ago
[Speculative Decoding] Enable arbitrary model inputs

github.com/vllm-project/vllm - abhigoyal1997 opened this pull request 5 months ago
[CI/Build] Simplify OpenAI server setup in tests

github.com/vllm-project/vllm - DarkLight1337 opened this pull request 5 months ago
[Misc] Add vLLM version getter to utils

github.com/vllm-project/vllm - DarkLight1337 opened this pull request 5 months ago
[Bug]: Gemma model fails with GPTQ marlin

github.com/vllm-project/vllm - arunpatala opened this issue 5 months ago
[Installation]: Error when importing LLM from vllm

github.com/vllm-project/vllm - manishkumar0709 opened this issue 5 months ago
[RFC]: OpenAI Triton-only backend

github.com/vllm-project/vllm - bringlein opened this issue 5 months ago
[Model] Support MAP-NEO model

github.com/vllm-project/vllm - xingweiqu opened this pull request 5 months ago
[Usage]: quantization option usage

github.com/vllm-project/vllm - Juelianqvq opened this issue 5 months ago
[Model] Add support for falcon-11B

github.com/vllm-project/vllm - Isotr0py opened this pull request 5 months ago
Heterogeneous Speculative Decoding (CPU + GPU)

github.com/vllm-project/vllm - jiqing-feng opened this pull request 5 months ago
[Model] Add Internlm2 LoRA support

github.com/vllm-project/vllm - Isotr0py opened this pull request 5 months ago
[Core] Allow AQLM on Pascal

github.com/vllm-project/vllm - sasha0552 opened this pull request 5 months ago
[Bug]: Cannot build cpu docker image

github.com/vllm-project/vllm - licryle opened this issue 5 months ago
[Feature]: multi-steps model_runner?

github.com/vllm-project/vllm - leiwen83 opened this issue 5 months ago
[Frontend] Add tokenize/detokenize endpoints

github.com/vllm-project/vllm - sasha0552 opened this pull request 5 months ago
[Bugfix] Adds outlines performance improvement

github.com/vllm-project/vllm - lynkz-matt-psaltis opened this pull request 5 months ago
[Feature]: Add num_requests_preempted metric

github.com/vllm-project/vllm - sathyanarays opened this issue 5 months ago
Chat method for offline llm

github.com/vllm-project/vllm - nunjunj opened this pull request 5 months ago
[Installation]:

github.com/vllm-project/vllm - Kastycupra opened this issue 5 months ago
Bump version to v0.4.3

github.com/vllm-project/vllm - simon-mo opened this pull request 5 months ago
[Misc] add logging level env var

github.com/vllm-project/vllm - youkaichao opened this pull request 5 months ago
[Misc] Make Serving Benchmark More User-friendly

github.com/vllm-project/vllm - ywang96 opened this pull request 5 months ago