Research#llm📝 BlogAnalyzed: Dec 27, 2025 08:31

Strix Halo Llama-bench Results (GLM-4.5-Air)

Published:Dec 27, 2025 05:16
1 min read
r/LocalLLaMA

Analysis

This post on r/LocalLLaMA shares benchmark results for the GLM-4.5-Air model running on a Strix Halo (EVO-X2) system with 128GB of RAM. The user is seeking to optimize their setup and is requesting comparisons from others. The benchmarks include various configurations of the GLM4moe 106B model with Q4_K quantization, using ROCm 7.10. The data presented includes model size, parameters, backend, number of GPU layers (ngl), threads, n_ubatch, type_k, type_v, fa, mmap, test type, and tokens per second (t/s). The user is specifically interested in optimizing for use with Cline.

Key Takeaways

Reference

Looking for anyone who has some benchmarks they would like to share. I am trying to optimize my EVO-X2 (Strix Halo) 128GB box using GLM-4.5-Air for use with Cline.