Old hardware, New (AI) problems

Jay MiracolaJay Miracola
1 min read

What do we say to buying bleeding edge hardware for running AI workloads?

Not today! I have an old HP Z600 (2009!) and GPU that I wanted to use to run #Kubernetes, #Ollama, Open WebUI, and utilize NVIDIA’s gpu-operator. It has been a solid machine through the years with dual socket Xeons, loads of ECC ram, and simply wont quit. It has run several hypervisors, OpenStack, OpenShift, and more! When I decided to plug in a GPU, and load up my AI stack, I had no idea the rabbit hole I would go down. Here is the short story; Ollama’s GPU runner by default uses the AVX instruction set which is not available in old CPUs. I briefly thought it was time to retire my old machine and buy something a little newer, but no! The kind Ollama devs added a build argument in their Dockerfile --build-arg CUSTOM_CPU_FLAGS= . Leaving the flag’s values empty builds the GPU runner without AVX allowing my beloved Z600 to live on, continuing to serve modern workloads.

Moral of the story? With a little ingenuity (and a helpful open-source community), old hardware can still punch above its weight in the AI era!

0
Subscribe to my newsletter

Read articles from Jay Miracola directly inside your inbox. Subscribe to the newsletter, and don't miss out.

Written by

Jay Miracola
Jay Miracola