Configuration Parsing Warning: In config.json: "quantization_config.bits" must be an integer

2.33 bpw EXL3 quant of GLM-4.6

Base quants provided by MikeRoz

This is just a quick mix of the 2.25 bpw quant with attention, dense layers and shared experts in 4.0 bpw.

Downloads last month
332
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support