EDUCATIONAL and EXPERIMENTAL while waiting for llama.cpp integration: run Qwen3-Next-80B FP8+FP4 model on consumer hardware using Intel's 4-bit AutoRound quantization, CPU, IPEX, GPU, Flash Linear ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results