Fixing rocm gptq by using triton code too (renamed cuda into triton). (#2691)
This commit is contained in:
parent
43df056eee
commit
cece8635f8
|
@ -10,8 +10,8 @@ from text_generation_server.utils.weights import Weight, Weights, WeightsLoader
|
||||||
|
|
||||||
if SYSTEM == "ipex":
|
if SYSTEM == "ipex":
|
||||||
from .ipex import QuantLinear
|
from .ipex import QuantLinear
|
||||||
elif SYSTEM == "cuda":
|
elif SYSTEM in {"cuda", "rocm"}:
|
||||||
from .cuda import QuantLinear
|
from .triton import QuantLinear
|
||||||
|
|
||||||
|
|
||||||
@dataclass
|
@dataclass
|
||||||
|
|
Loading…
Reference in New Issue