The user built a workstation to run large language models (LLMs) locally, using a combination of second-hand parts including a HP Z440 workstation, two Tesla server GPUs, and a third video card to keep the BIOS happy. The setup allows for running medium-sized models under the user's control, with a tradeoff between noise, temperature, and power consumption, and the user is happy with the ...