Ecosyste.ms: OpenCollective

An open API service for software projects hosted on Open Collective.

vLLM

vLLM is a high-throughput and memory-efficient inference and serving engine for large language models (LLMs).
Collective - Host: opensource - https://opencollective.com/vllm - Code: https://github.com/vllm-project/vllm

[Kernel] add bfloat16 support for gptq kernel

github.com/vllm-project/vllm - jinzhen-lin opened this pull request 5 months ago
[core] SequenceController in SamplingParams

github.com/vllm-project/vllm - mmoskal opened this pull request 5 months ago
Sync huggingface modifications of qwen Moe model

github.com/vllm-project/vllm - eigen2017 opened this pull request 5 months ago
[Misc] Logits processor plugins

github.com/vllm-project/vllm - NadavShmayo opened this pull request 5 months ago
[Misc]Easier access to the nccl library

github.com/vllm-project/vllm - Cyuchuan opened this pull request 5 months ago
[Bugfix] Fix call to init_logger in openai server

github.com/vllm-project/vllm - NadavShmayo opened this pull request 5 months ago
[Core][Bugfix]: fix prefix caching for blockv2

github.com/vllm-project/vllm - leiwen83 opened this pull request 5 months ago
[Core][Bugfix]: fix prefix caching for blockv2

github.com/vllm-project/vllm - leiwen83 opened this pull request 5 months ago
[Core][Bugfix]: fix prefix caching for blockv2

github.com/vllm-project/vllm - leiwen83 opened this pull request 5 months ago
[Frontend] Re-enable custom roles in Chat Completions API

github.com/vllm-project/vllm - DarkLight1337 opened this pull request 5 months ago
Add TensorizerArgs to client api server

github.com/vllm-project/vllm - vrdn-23 opened this pull request 5 months ago
[Misc] Enhance attention selector

github.com/vllm-project/vllm - WoosukKwon opened this pull request 5 months ago
[Kernel] Add w8a8 CUTLASS kernels

github.com/vllm-project/vllm - tlrmchlsmth opened this pull request 5 months ago
[CI] Nits for bad initialization of SeqGroup in testing

github.com/vllm-project/vllm - robertgshaw2-neuralmagic opened this pull request 5 months ago
[New Model]: Blip2 Support required

github.com/vllm-project/vllm - anisingh1 opened this issue 5 months ago
[CI/Build] use setuptools-scm to set __version__

github.com/vllm-project/vllm - dtrifiro opened this pull request 5 months ago
[Bugfix] Fix CLI arguments in OpenAI server docs

github.com/vllm-project/vllm - AllenDou opened this pull request 5 months ago
[Core]fix type annotation for `swap_blocks`

github.com/vllm-project/vllm - jikunshang opened this pull request 5 months ago
[Speculative decoding] Improve n-gram efficiency

github.com/vllm-project/vllm - comaniac opened this pull request 5 months ago
[Misc] Added devcontainer to help vscode dev setup

github.com/vllm-project/vllm - ElefHead opened this pull request 5 months ago
[Misc] Apply a couple g++ cleanups

github.com/vllm-project/vllm - stevegrubb opened this pull request 5 months ago
[CORE] Improvement in ranks code

github.com/vllm-project/vllm - SwapnilDreams100 opened this pull request 5 months ago
[Bugfix] Add logs for all model dtype casting

github.com/vllm-project/vllm - mgoin opened this pull request 5 months ago
[Bug]: Not able to do lora inference with phi-3

github.com/vllm-project/vllm - WeiXiaoSummer opened this issue 5 months ago
[CI/Build] Tweak Marlin Nondeterminism Issues

github.com/vllm-project/vllm - robertgshaw2-neuralmagic opened this pull request 5 months ago
add `TypeLogitsProcessor`

github.com/vllm-project/vllm - eitanturok opened this pull request 5 months ago
[Bugfix] Update grafana.json

github.com/vllm-project/vllm - robertgshaw2-neuralmagic opened this pull request 5 months ago
[Doc] Add API reference for offline inference

github.com/vllm-project/vllm - DarkLight1337 opened this pull request 5 months ago
[Bugfix] Fix CLI arguments in OpenAI server docs

github.com/vllm-project/vllm - DarkLight1337 opened this pull request 5 months ago
[Misc] Remove unnecessary ModelRunner imports

github.com/vllm-project/vllm - WoosukKwon opened this pull request 5 months ago
[Bug]: VLLM + tritonserver

github.com/vllm-project/vllm - dlopes78 opened this issue 5 months ago
Remove Ray health check

github.com/vllm-project/vllm - Yard1 opened this pull request 5 months ago
Installation with CPU with errors

github.com/vllm-project/vllm - ming-ddtechcg opened this issue 5 months ago
[Core] Implement sharded state loader

github.com/vllm-project/vllm - aurickq opened this pull request 5 months ago
[Misc] Add OpenTelemetry support

github.com/vllm-project/vllm - ronensc opened this pull request 5 months ago
[Doc]: API reference for LLM class

github.com/vllm-project/vllm - zplizzi opened this issue 5 months ago
[Feature]: support for aixcoder

github.com/vllm-project/vllm - chucksylar opened this issue 5 months ago
[Feature]: vAttention

github.com/vllm-project/vllm - nivibilla opened this issue 5 months ago
[Frontend] Move async logic outside of constructor

github.com/vllm-project/vllm - DarkLight1337 opened this pull request 5 months ago
[Usage]: prefix-caching

github.com/vllm-project/vllm - chenchunhui97 opened this issue 5 months ago
[RFC]: Inline Golden (Expected) Tests

github.com/vllm-project/vllm - youkaichao opened this issue 5 months ago