Strix Halo Llama-bench Results (GLM-4.5-Air)
Analysis
This post on r/LocalLLaMA shares benchmark results for the GLM-4.5-Air model running on a Strix Halo (EVO-X2) system with 128GB of RAM. The user is seeking to optimize their setup and is requesting comparisons from others. The benchmarks include various configurations of the GLM4moe 106B model with Q4_K quantization, using ROCm 7.10. The data presented includes model size, parameters, backend, number of GPU layers (ngl), threads, n_ubatch, type_k, type_v, fa, mmap, test type, and tokens per second (t/s). The user is specifically interested in optimizing for use with Cline.
Key Takeaways
- •Strix Halo performance with GLM-4.5-Air is being benchmarked.
- •The user is seeking optimization advice and comparative data.
- •ROCm 7.10 is used as the backend for the benchmarks.
“Looking for anyone who has some benchmarks they would like to share. I am trying to optimize my EVO-X2 (Strix Halo) 128GB box using GLM-4.5-Air for use with Cline.”