Can someone tell me if I should buy an RTX 4090 or try to rent an A100 if I want to fine-tune one of those big language models? Sorry if this is a really basic question but I am totally lost with all these numbers and letters lol. I want to build a specialized bot for my plant nursery here in Seattle to help customers with care instructions and I have about 1800 to 2000 dollars saved up for a new computer this month. I saw the 4090 in a shop and it looks huge but then I read online that the A100 has way more memory or something? I dont really get what VRAM is or why it matters so much for the training part. If I buy the 4090 will it just break if I try to run a big model on it? Or is the A100 only for like giant companies because it seems really expensive to rent hourly. I just want something that wont take like a week to learn my data sets. My friend said I might need two cards but I definitely cant afford that right now. Just trying to figure out the smartest way to spend my money before I go and buy the wrong thing...
I remember picking up an ASUS ROG Strix GeForce RTX 4090 24GB last year for a similar project. Honestly, it was a massive letdown.
^ This. Also, if you're really trying to stretch that 2000 dollar budget, buying a brand new NVIDIA GeForce RTX 4090 24GB is basically the fastest way to go broke. It's a beast for gaming, but for LLMs, you're paying a massive premium for speed you might not fully use if the memory bandwidth is the bottleneck anyway.
Honestly, I tried building my own rig for this and it was such a letdown. My setup kept crashing because the memory wasnt enough for the models I wanted. Unfortunately, consumer gear feels really limited for this.
Ngl I'm kinda worried about those 12VHPWR connectors melting during 24/7 training. I think I heard about safety issues there... definitely watch your temps and power draw if you go local.
I spent way too much on a local build and it was a massive disappointment.
Yep, this is the way
tl;dr: I have been stuck with this exact same dilemma for months now and it is honestly driving me crazy. I have the same problem and keep looking for a clear answer but everywhere I turn there is just more conflicting info. Tbh I have been building PCs for a long time but this whole LLM hardware requirement thing is a total mess to figure out... I keep looking at my savings and then looking at the 4090, then getting scared I will just be wasting money if the memory isnt enough for what I want to do. It is so frustrating because you want to own your hardware but the risk of it being totally useless for training in a few months is just sitting there in the back of my mind. Still looking for a real answer myself.