Skip to content

Commit

Permalink
Fix cutlass_fp8_supported flag set on HPU (#550)
Browse files Browse the repository at this point in the history
  • Loading branch information
nirda7 authored Nov 26, 2024
2 parents 0f513bd + 7133502 commit 38c2d10
Showing 1 changed file with 1 addition and 0 deletions.
1 change: 1 addition & 0 deletions vllm/model_executor/layers/quantization/fp8.py
Original file line number Diff line number Diff line change
Expand Up @@ -117,6 +117,7 @@ class Fp8LinearMethod(LinearMethodBase):

def __init__(self, quant_config: Fp8Config):
self.quant_config = quant_config
self.cutlass_fp8_supported = False
if current_platform.is_cuda_alike():
self.cutlass_fp8_supported = cutlass_fp8_supported()

Expand Down

0 comments on commit 38c2d10

Please sign in to comment.