Prefix-Aware Attention for LLM Decoding
-
Updated
Jan 5, 2026 - Python
Prefix-Aware Attention for LLM Decoding
A curated list of plugins built on top of vLLM
vLLM Plugins for additional features like decoding strategies, monitoring, models etc
A manager to load vllm plugins without rebuilding image for each new plugin.
Add a description, image, and links to the vllm-plugins topic page so that developers can more easily learn about it.
To associate your repository with the vllm-plugins topic, visit your repo's landing page and select "manage topics."