The problem is simple: consumer motherboards don’t have that many PCIe slots, and consumer CPUs don’t have enough lanes to run 3+ GPUs at full PCIe gen 3 or gen 4 speeds.
My idea was to buy 3-4 computers for cheap, slot a GPU into each of them and use 4 of them in tandem. I imagine this will require some sort of agent running on each node which will be connected through a 10Gbe network. I can get a 10Gbe network running for this project.
Does Ollama or any other local AI project support this? Getting a server motherboard with CPU is going to get expensive very quickly, but this would be a great alternative.
Thanks
Never heard of it. What is it about?
It’s a way to do distributed parallel computing using consumer-grade hardware. I don’t actually know a ton about them, so you’d be better served by looking up information about them.
https://en.wikipedia.org/wiki/Beowulf_cluster
Thanks
deleted by creator