24GB VRAM will easily let you run medium-sized models with good context length, and if you’re a gamer the XTX is a beast for raster performance and has good price/performance.
If you want to get serious about LLMs also keep in mind that most models and tools scale well across multiple GPUs, so you might buy one today (even a lesser one with “only” 16 or 12GB) and add another later. Just make sure your motherboard can fit 2, and you have a CPU, RAM and power supply that can handle it.
Here’s a good example from a guy who glued two much more modest cards together with decent results: adamniederer.com/blog/rocm-cross-arch.html
=> More informations about this toot | View the thread | More toots from will_a113@lemmy.ml
=> View sith@lemmy.zip profile
text/gemini
This content has been proxied by September (ba2dc).