[Misc] Fused MoE Marlin support for GPTQ (#8217)
This commit is contained in:
3
tests/weight_loading/models-large.txt
Normal file
3
tests/weight_loading/models-large.txt
Normal file
@ -0,0 +1,3 @@
|
||||
compressed-tensors, nm-testing/Mixtral-8x7B-Instruct-v0.1-W4A16-quantized, main
|
||||
compressed-tensors, nm-testing/Mixtral-8x7B-Instruct-v0.1-W4A16-channel-quantized, main
|
||||
gptq_marlin, TheBloke/Mixtral-8x7B-v0.1-GPTQ, main
|
||||
Reference in New Issue
Block a user