Self-hosting LLM on Multi-CPU and System RAM
Analysis
Key Takeaways
- •The post explores the viability of running large LLMs on older server hardware with significant RAM.
- •The author is specifically considering a dual-socket Xeon system with 256GB of RAM.
- •The primary concern is whether the system will provide acceptable performance for running open-source LLMs.
“I was thinking about buying a bunch more sys ram to it and self host larger LLMs, maybe in the future I could run some good models on it.”