PAT: Prefix-Aware Attention for LLM Decoding
-
Updated
Dec 16, 2025 - Python
PAT: Prefix-Aware Attention for LLM Decoding
A curated list of plugins built on top of vLLM
Add a description, image, and links to the vllm-plugins topic page so that developers can more easily learn about it.
To associate your repository with the vllm-plugins topic, visit your repo's landing page and select "manage topics."