We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent ac60dc7 commit 1b7624bCopy full SHA for 1b7624b
vllm/envs.py
@@ -321,6 +321,7 @@ def maybe_convert_int(value: Optional[str]) -> Optional[int]:
321
# - "XFORMERS": use XFormers
322
# - "ROCM_FLASH": use ROCmFlashAttention
323
# - "FLASHINFER": use flashinfer
324
+ # - "FLASHMLA": use FlashMLA
325
"VLLM_ATTENTION_BACKEND":
326
lambda: os.getenv("VLLM_ATTENTION_BACKEND", None),
327
0 commit comments