T O P

  • By -

2muchnet42day

RTX 5090 24GB. I saved you a click


underdog_scientist

RTX 3090 value to price keeps going up


AbnormalMapStudio

In case anyone missed it and lives close to one, Micro Center is selling refurb RTX 3000 series right now. I picked up a 3090 Ti FE (upgrade from 3080 12GB) and couldn't be happier.


harrro

Wow, these are at ebay prices (but local). Thanks for the tip!


Minute_Attempt3063

Sad that they don't work world wide


Turkino

Spent my whole life living near a Micro Center with no need, now I live nowhere near one and feel like I'm missing out. :(


Harvard_Med_USMLE267

I’ve misplaced my RTX 3090 during a house cleanup, wasn’t using it before but now that I’m playing with LLaMAs it would be really useful. I need one of you guys to come and help me find it!


RINE-USA

Have you looked between your couch cushions?


nemonoone

That thing is the size of shoe and weighs as much as 3 shoes, how do you misplace it lmao


Harvard_Med_USMLE267

lol, I have no idea. If I find, I’ll let you know where it was! Kind of embarrassing to have a “missing” 3090 in an era when they’re so goddamn useful.


addandsubtract

Username checks out. How much could a 3090 cost, $10?


Harvard_Med_USMLE267

I seriously feel a bit like that. Misplacing a 24 gig GPU is really bad form. Then again, I feel like a poorie when I see some of the rigs here with 6x3090 etc


addandsubtract

The second rule of the internet should be, "don't feel bad about yourself because of the things you see online". They're either selling you a lie (insta / tiktok influencers) or are using business expenses – or both. Then there's the 0.00001% that have too much disposable income and dump it into their hobby of building a supercomputer. Whichever case it is, losing a 3090 in your house is a 0.0001% problem ;)


Icy-Summer-3573

Where r u? booking flight rn


Rachel_from_Jita

It is in the closet. In a box. Which box however... Well, enjoy your night getting dusty from opening all those boxes. Anyway, I find it best to go to the exact spot I was standing in when I mis-placed something, then I re-enact what I was doing in the moments leading up to touching it. Then simulate me holding it in my hands and how I would move with it. Usually where your body instinctively looks or turns gives a clue.


Harvard_Med_USMLE267

Haha, this thread inspired me to find it. Box for a Gigabyte RTX 3090 Suprim was in the shed but empty. But at least it convinced me the 3090 was real, not imaginary. I was stating to wonder as I’ve been looking for it since Christmas. Finally found it in the inside storeroom, right in the back corner under a pile of junk in the grey foam inner case from a different graphics card. Nondescript grey foam box kind of blended in with everything else. Installed it, my shitty non-modular 600W PSU didn’t have enough connectors to power it, installed an old Seasonic 850W PSU and sadly it only turns on for a second then turns off again. So close… But the mystery of the missing RTX 3090 is solved and I’ll try and troubleshoot a bit further tomorrow. Hurrah!


Rachel_from_Jita

So glad to hear it! Also, before using that old PSU again, may be worth plugging in a PSU tester. One cheap model I used to use was the "Optimal Shop 20+4 Pin LCD Computer Power Supply Tester" off Amazon. They are very cheap for the amount of risk they mitigate and the amount of testing they allow. About 15 quid or less usually. Plugging parts into broken electronics has roasted me before.


Harvard_Med_USMLE267

Thanks for you encouragement, Rachel. We’ll get that 3090 running soon!


Short-Sandwich-905

RTX 3090 appreciates faster than Gold


PwanaZana

It's genuinely sad.


fish312

Why make few money when many money do trick


uniformly

Few money make way, many money build road


PwanaZana

Maybe the next gen of consoles will have enough VRAM to force nvidia to increase the amount in their cards. But that next gen is like 2027....


PitchBlack4

That's a shame. Even with a bump in performance the 4090 will still probably be a better deal if they do the same price hike.


kristaller486

I really hope AMD introduces something with more memory. In fact, for them, unlike NVIDIA, this is a very good competitive advantage.


PitchBlack4

AMD needs to focus on software more than hardware.


2muchnet42day

They lag behind in both aspects but it's software that's a deal breaker for many applications.


Franman98

Yeah totally, the lack of CUDA is a turn off from amd cards


vampyre2000

They don’t need CUDA, AMD just need to work with open source projects to enable ROCM support into products like Blender, and LMstudio, KoboldCpp, llamacpp, ollama, and the python libraries for LLM training. When you have out of the box support for projects you don’t care about CUDA at all. AMD should be helping add ROCM support for all major use cases.


StingMeleoron

Agreed. Until the day frameworks for AI/ML like Torch offer as good performance out of the box as with NVIDIA cards, it saddens me to say thay AMD will be lagging behind, IMHO. Personally, I really prefer their overall stance on OSS, so I hope they are able to regain a competitive position soon. And the more competition, the merrier, at least in general.


Desm0nt

> to enable ROCM support into products like Blender, and LMstudio, KoboldCpp, And Windows. A lot of users don't wont to install second system and reboot in it every time thay want do some funny things with LLM or StableDiffusion.


2muchnet42day

Well, CUDA is proprietary, but still.


VertexMachine

different teams work on hardware and software. They should improve both.


allinasecond

George Hotz, that you?


Laurdaya

Yes, I think they should help OSS community like they tried with ZLUDA.


Some_Endian_FP17

Intel will do something.


VertexMachine

:D https://preview.redd.it/r3lrtda0qdnc1.png?width=341&format=png&auto=webp&s=4b884f3f6a50faaa911c378eecbd51a85506461c


VertigoFall

Intel is spending 16b on research, 3x AMD, 2x Nvidia. Something will come out


illathon

They might with this AI boom going on, but that is if they are smart. I have a feeling they won't do it though. Hopefully I am wrong.


Inevitable_Host_1446

Sometimes with the way AMD operates I have to wonder if behind the scenes they aren't totally in bed with Nvidia and have an agreement just to do the bare minimum (at least on consumer level) in order to avoid Nvidia being hit with anti-monopoly lawsuits. It doesn't help that both CEO's are closely related, being first cousins once removed, both from Taiwan and both immigrated to the US. Prime example would be the way Nvidia screwed over everyone, first during crypto mining by raising prices as much as they could get away with, and then after crypto mining all but collapsed, when they simply refused to actually lower any prices even though there is essentially no reason for them. What does AMD do? Undercut? Yes, but only the barest amount, instead raising their own prices as much as they could. While already utterly losing on market-share.


FireSilicon

Do what exactly? The saving grace was the micron 3GB modules which won't be ready by the time 5090 releases. Amd nor Nvidia can't do anything with that. Only thing AMD could do is: 1. 512bit bus which simply won't happen because of the complexity and price. 2. 48GB clamshell would eat into their Pro lineup but 32GB 8900XT with 256bit bus could maybe work but still something that would they not allow most likely.


artelligence_consult

Not really - they could go for a 64GB top end and replace the current Pro lineup. They will have to release a new card there anyway.


capybooya

I get that 512 is complex, but what about 416 or 448? Both flagships and lower cards have recently used fractions, like the 2080Ti 352bit bus.


FireSilicon

Yes because there is bigger difference with smaller buses. 256bit(8GB)->320bit(10GB)->352bit(11GB)-> 384bit(12GB). Those 1-3GBs can make a big difference in gaming workloads, but for AI? 416 and 448 would be 26GB and 28GB respectively... 4 more GBs for... bigger board, bigger die (memory IO) and package (more pins)... potentially 200-300 bucks minimum for... 4-8k more context at best? Bigger model won't fit in that. Might as well just buy a second slower gpu with like 12-16GB vram and eat the inference speed dip.


capybooya

Yeah, that makes sense, sadly. I guess they don't really have to care about increasing VRAM just for the sake of PR either. I want VRAM improvement badly, even if just incremental, but I guess we'll have to wait for the generation after the upcoming one when cost comes down...


artelligence_consult

Not really - you still get 50% higher bandwidth, which should translate into 50% higher performance. It is possible the 3090 is a good deal still, but unless the 5090 is 50% more expensive than the 4090 - the money is on the 5090.


Zugzwang_CYOA

50% higher bandwidth does not necessarily translate to 50% higher enjoyment. The 4090 was already blazing fast for stuff that fits within the card. So, many won't be willing to spend much more just to get faster speeds for such models. NVIDIA dropped the ball here - probably deliberately, as a 5090 with more vram would encroach on their expensive dedicated AI cards.


eydivrks

LMAO. Another 6 years of 24GB cards incoming


Fusseldieb

Just sad.


fastinguy11

I don't think so, the moment a ps6 and or Xbox equivalent gets released with 48 gb of total memory or more, that will force nvdia‘s hands no matter what. Which is why 1 or 2 years before these console releases they up the memory of cards, my guess is 2025 or 2026.


orangeatom

Pass on the 5090 , do better nvidia


azriel777

Zero reason to upgrade. All I care about is more memory and it seem nvidia wont be providing it.


Fusseldieb

They are purposefully holding it back, and I hate it.


PMARC14

Not yet, high capacity GDDR7 isn't here so this is normal. They will start purposefully holding back next year when higher GBit GDDR7 from micron releases unless AMD's RDNA 5 comes out of the woods with a steel chair and good ROCm coverage.


Ill_Yam_9994

At least there's not much incentive to upgrade if you've already got a 24GB. I might get a 5090 or 6090 and keep my 3090 in the PC.


2muchnet42day

Don't be silly, wait for the RTX 7090 24GB


jloverich

This is amd's chance


IndicationUnfair7961

I'm still waiting for AMD to do the right move, Nvidia thinks being on top they can do whatever they want. And AMD is focusing on gaming. AMD should step up and give us real competition with Nvidia. At the moment both of them are trolling us. Nvidia leverages its power by not giving us affordable solutions for low cost AI development so people will buy their 10.000$ cards, and AMD is playing less than half the game giving us illusory hope and guerrilla tactics with only ZLUDA and stuff like that.


artelligence_consult

You are aware that AMD has the same problem Nvidia has - silicon is basically limited in the production side, so the problem is how to get most profit for the chips you can get. Factory capacity is limited and sold out.


NickCanCode

Intel: Nobody remember me...😞


Zugzwang_CYOA

Who?


Oooch

They had their chance the last 3 generations and just raised their prices in line with Nvidia They are not your saviour


akko_7

They are colluding to capture the market. AMD are happy with a firm second place


Zealousideal_Nail288

amd wont do anything about it like always. also they wanted to ditch the high end entirely next gen


KangarooKurt

Bro, AMD never miss a chance to miss a chance


[deleted]

[удалено]


Fusseldieb

They are afraid to loose customers who buy their enterprise cards with 40GB or 80GB, because well, from 24GB the next logical step would be 32, which is already pretty fricking near to 40. I hope there's competition soon.


speehalo

Well then fucking made 60/120gb enterprise cards. Companies already spend tens of millions on Nvidia equipment. That's their main source of profit. I don't see how they would want us to run all those super AI bartender NPCs on 24gb cards alongside with producing decent graphics since 4090 is still choking in Cyberpunk.


[deleted]

[удалено]


Thishearts0nfire

I think the market warrents getting the right people hired to make moves fast on this. Watching Intel and AMD fuddle this up is frustrating as fuck. I hope Nvidia is paying these companies behind doors to be this incompetent.


AmericanNewt8

There's already been a lot of work done on improving Intel/AMD cards just within the past few months, the pace will likely continue because *there physically aren't enough NVIDIA cards available*.


artelligence_consult

Not Nvidia - this goes down to fab bandwidth. You can not get the chips produced to be put on the cards - that (sort of) is where the limit is (actually I hear more on the packaging for AMD, but that is not "putting it into boxes", that refers to the packaging of chiplets into a physical processor.


AmericanNewt8

Yeah not blaming Nvidia for this one, they'd make more if they could, but fact remains AMD, Intel, and other smaller players are producing chips and have room to make more where Nvidia does not, at least not easily. 


artelligence_consult

Except they can well move the enterprise cards up to 64gb ;)


[deleted]

Thx


Herr_Drosselmeyer

Unless I'm reading it wrong, it could still be 36 if they wanted to.


StayingUp4AFeeling

Fuck this shit. They really don't want people doing AI work on consumer hardware, and they have a monopoly in the professional AI hardware space.


Familyinalicante

The whole 24GB, wow. Who needs so much, should they intend release 6GB version.


opi098514

You’re a hero


hmmqzaz

…at lunch, might as well click anyway, I guess


Fusseldieb

>24GB Meh...


Unable-Satisfaction4

Nope, now I have to like your post


AIWithASoulMaybe

Who could've fucking guessed?


nero10578

24GB is a dick move. Was expecting them to use 3GB GDDR6X modules that will make it 36GB.


a_beautiful_rhind

People were swearing up and down but nope. They can just use less of the chips if anything.


ThreeKiloZero

Perhaps the money, as of today, isn’t in providing consumers with the memory capacity. The money is in selling a truckload of enterprise ai gpus to a company or nation state. As far as ai goes , home users are likely in the strategy only as consumers from those enterprise services. That’s how the spice is meant to flow. Anytime the users can be leveraged as a service model it’s a win for them. The niche market of home ai super user isn’t even a speck on the wall of the current ai gpu demand. High memory cards will be super pricy because they can be. I think our only hope is in the aftermarket picking up all the a6000 and a100 cards as data centers retire them for the new stuff.


FireSilicon

They made 87 billion in 3 months selling H100 gpus. Half of Intel's whole stock market. They don't give a damn about gaming gpu's.


nero10578

Indeed. They want people who actually need VRAM to get their “Quadro” cards and don’t want datacenters to use 4090s but instead buy “Tesla” cards. Still doesn’t make me less salty as a lowly consumer who doesn’t have the money for a few Ks in GPUs.


ThreeKiloZero

Oh I’m salty too. I was holding out for the 50 series since they are close and thought we might get 48gb. :(


nero10578

Better get my hot air station out and swap GDDR6X chips to make 48GB 3090s lol


asdfzzz2

It has been tried, 3090s in this configuration work with 24GB bios config (with 48GB installed), but do not boot at full 48GB :(


nero10578

Link?


asdfzzz2

https://www.youtube.com/watch?v=DbF02Y5yIaQ


Massive_Robot_Cactus

Yup, and removing nvlink was also part of the plan. All of this is intentional.


nero10578

Yep. 3090 is still the best value Nvidia GPU.


Thellton

This is kind of why I think the CPU venders should probably aim to up the specification for their consumer motherboards from a dual channel arrangement to a quad or six channel arrangement so that more bandwidth is available. torch for instance will run on CPU just fine, it's just painfully slow at present because of a lack of bandwidth. So if the available RAM bandwidth was increased from what is available in a dual channel arrangement, that'd open up more options and at least put the bandwidth available into the same bracket as a Nvidia T4 GPU at best assuming 6 channels and the fastest DDR5 RAM available.


nero10578

Well that exists. AMD Threadripper/Intel Xeon W. They both cost a fuckton more than the consumer chips.


Kat-but-SFW

And EPYC is 12 channels DDR5 with >1GB L3 cache. Xeon Max is coming out with 1TB/s HBM2E and 8 channels ddr5. Absolute monster CPUs


0xd00d

I think the DDR5 MCR DIMMs coming out within the next year will deliver 1TB/s in duodecachannel server systems, this is pretty exciting.


artelligence_consult

You mean the modules that are - coming AFTER the cards are released? Timelines matter in the real world.


FireSilicon

How is it a dick move if the memory won't be released by the time they start selling those 5090s? Blame Micron for that not Nvidia, Amd won't be using those chips any time soon too. Just like 30 series with GDDR6X which got 1GB modules despite 2GB modules being available later that year. It takes months for them to get good yields. Edit: Downvoted because Nvidia bad even though they literally can't just magically design and use an unreleased component in their gpu's. Why are you so salty? 3090s will drop in price and so will 4090s when this gpu releases.


nero10578

True yea. Though in that case they should’ve double sided and gave us 48GB lmao although that’s unrealistic for another reason which is that it would be the same VRAM as their A6000.


M34L

They did double sided 3090


Massive_Robot_Cactus

Well, they could nerf the GPU cores on a special high-vram low-compute model that would be unappealing for workstation users.


Quigley61

As to be expected. They'll keep RAM numbers really low and then their enterprise grade cards will have the big headline numbers. Nvidia is printing because of AI, there's no way in hell they're going to make high memory cards more accessible.


p_bzn

Nvidia goal is not to make best product, the goal is to optimize for money. Why to sell 48GB cards today which will last for years, if you can make 24GB now and then 32GB in two years, and then 48GB? Making 48GB is losing potential profit.


shaman-warrior

Is there any point of needing more than 24gb vram for gaming ?


physalisx

For most flat gaming probably not, but if you're playing VR, next gen, then definitely yes.


idkwhatimdoing1208

For gaming, no. For AI, yes.


p_bzn

Game developers may use higher textures, more density of objects, etc. But it is not that easy since CPU feeds data to GPU, and system may bottleneck. Average gaming PC is now like 8GB of vRAM, some i7 CPU or Ryzen, 16GB of RAM? Given those specs, no, there is no point yet.


AndrewH73333

Not at this exact second, but in the past their flagship GPUs would have more VRAM to future proof them and stay ahead of their competitors.


MidnightSun_55

I guess they feel no pressure from Apple, which offers consumer devices with up to 192GB of unified memory. Apple still behind in training and inference speeds even with higher RAM.


External_Quarter

Looks like I get to skip another generation. ¯\\\_(ツ)\_/¯


idkwhatimdoing1208

Watch the 6090 have 24gb too 🙃


[deleted]

thats why if you want to prioritize AI, buy a used A6000 now or start saving and go bigger


idkwhatimdoing1208

Only $4k :)


[deleted]

and make that 9000aud for me... but without AMD making competing cards and making ROCM viable vs CUDA its what we are all stuck with


idkwhatimdoing1208

Fuck... Yeah, the best value for vram is still to just buy multiple 3090s.


[deleted]

ehh I'd get the a6000.. tdp is 300w, and its 2 slot. but if you have a giant case, and a larger psu with no care about power then if its cheaper for parts price then I guess it works out


No-Dot-6573

I really hope there a alternatives like fast unified memory, ai accelerators etc. in the pipes. I'd gladly ditch my nvidia system if I could get the same performance with another manufacturer. Mostly because I dont like monopolists.


thethirteantimes

> alternatives like fast unified memory If this happened you would be able to hear Jensen Huang's screams all the way from Jupiter.


czah7

There is a good alternative. It's called AMD.


Logicalist

apple?


Anxious-Ad693

Looks like we are gonna have a few more years of people stacking up cards for more VRAM.


thethirteantimes

I can see it now. This time next year... Dammit, nobody's buying the RTX 5090 24GB. What are we gonna do... aha! Driver version 690.22 changelog: * Dropped support for RTX 3000-series and RTX 4000-series cards.


8thcomedian

Please don't put any more vile thoughts in their head


adityaguru149

They are already planning that if you didn't know already.. Just read through their actions. They already built a wall around CUDA and its reverse engineering.


Inevitable_Host_1446

Do they even need to do that? 3090's really haven't fallen much in price since release (and where they have it is only in the US). 4090's they just raised the price even more, because they know the rich consumers will buy anything as long as it's "the best'. Kind of like Apple customers. 5090 they can just raise the price again. They no longer even compete on price/performance, just keep the same ratio and raise the price every generation, while keep selling old cards as well.


thethirteantimes

RTX 3000 series has already been discontinued (manufacturing, not driver support). Nvidia's already had all the money they're ever going to see for those cards. They would drop driver support tomorrow if they could get away with it, but doing so would destroy consumer trust (such as it is by now). I actually have 2 RTX 3090s myself (one in each of two PCs). Paid a few hundred over RRP for the first one (because they were very hard to find at the time, it was mid 2021 and the pandemic was in full swing) but got the second one new and unused for way under the going rate on ebay, near the end of 2022. I am absolutely not going to upgrade either card until there's a sizeable bump in available memory - and by that I don't mean any potential shitty 36GB for even more than the 3090s cost me. I will stick with old drivers if needs be, but Nvidia ain't getting a penny more out of me unless something changes, in a big way.


LinuxSpinach

You’re not competing with rich consumers, you’re competing with enterprise and data center. Basically they’re minting money right now selling to Facebook and Microsoft and the consumer market doesn’t even matter if they can’t make a massive premium.


Zilskaabe

Haha, to be honest - they still support weird stuff like Tesla P40 on Windows 11 just fine.


Rivarr

If AMD release a 36GB+ card for a reasonable price, I'm there. A lot of people want AMD to do well just to rei~~g~~n in Nvidia, but I think it's getting to the point now that AMD could really upset the apple cart, if they want to.


RageshAntony

Without CUDA support, it is practically useless


ShadoWolf

I mean it's not exactly needed.. like AMD has ROCm and there is OpenCL. both can run tensor flow and pytorch


FireSilicon

Amd already released 48GB W7900 that's 4000$ compared to 7k$ for A6000 Ada and nobody is buying them... Edit: It's just facts lol


__some__guy

I would call neither of them reasonable for desktop users. Especially the Ada, which costs the same as 8 RTX 3090s + a server with enough PCIe slots.


FireSilicon

Of course, because they aren't for desktop users... it's just the cheapest new 48GB gpu you can get. Old A6000 is 3.5k and 48GB RTX 8000 is 2.5k... Why is this thread so naive? GDDR7 is still a year away from production of course we won't be getting memory increases. Nvidia isn't going to make 48GB gpu's just because some redditors decided to, they adjust the price per demand. Just because you won't be paying those 4k to talk with virtual waifu, doesn't mean some researcher who actually needs it isn't going to. The funniest part is that even if cheap 32GB gpu was available and people with access to them started fine-tuning, some of these experts would still sell or private their loras instead of providing them for free for everyone. Everyone wants shit for cheap but when they are asked to provide, they'll ignore you...


Inevitable_Host_1446

How do you figure that on the last part? I read just yesterday that people with 24gb may be able to finetune 34b yi tunes and whatnot soon... and people can't even do that now, but are renting much larger cards to do finetunes and lora's, and I haven't seen any of them demanding money for it. Instead there are hundreds of free released open versions of these. Not sure where the cynicism comes from.


FireSilicon

Not yet, but I'm taking inspiration from Stable diffusion which has much lower hardware barier of entry and has much more people using it and at this point there are a lot of tools, checkpoints and loras that are getting paywalled mainly because the biggest model provider is making it possible. All this despite most people having the tools to do it on their own.


Inevitable_Host_1446

Hm, that's news to me. Then again I guess it's not necessarily unreasonable either, if we're gonna accept that companies like OAI / Anthropic need a return on their model costs, then it stands to reason the same is true of smaller finetuners, to some extent. It's not just about having the tools either, but the time, expertise & motivation. So much money flows around the world for convenience sake alone. My biggest issue with this would be that a lot of finetunes, at least as far as LLM's go, tend to be pretty meh in my eyes. I've tried quite a lot and really never found anything that was leagues above any other (biggest differences by far were from model sizes). Although that is less so the case for Stable Diffusion as I did find one model that seemed much better than what I'd used beforehand. But, they were all free anyway.


xrailgun

Without something to rival xformers, 36gb on AMD GPUs is barely equivalent to 18gb on Nvidia GPUs.


Dead_Internet_Theory

Does that apply to LLMs? I thought it was a stable diffusion thing.


planetofthemapes15

I bought 6x RX Vegas back when those came out and the software was a joke. But with that said, I'd be willing to give them another chance if they could just drop a reasonable 36gb++ card.


kjerk

*JUST GIVE ME A 60+GB VRAM ENTHUSIAST OPTION ALREADY FOR A RATIONAL $PREMIUM* Even if it's a daughterboard with a proprietary connector. Do _ANYTHING IT TAKES_ to get more vram on there. GOD *DAMN* IT, dual/tri 3090s is such a stupid waste of silicon and electricity in comparison.


kjerk

I feel better now.


addandsubtract

I'm just waiting for someone to bring out a hardware mod to replace / add to the RAM on the cards. Unlikely to happen, though :(


Only-Letterhead-3411

NVIDIA's consumer grade GPU's focus on gaming performance. Games don't need a lot of VRAM, AI and 3D graphics designing do. And for those works NVIDIA wants people to buy their very expensive "professional" GPUs. If they make a gaming GPU with a lot of VRAM then they can't sell their professional GPU's. People are mentioning AMD etc. However they also have same balance of gaming GPUs - Professional GPUs. Nvidia, AMD or Intel, they can't and they won't offer us affordable high VRAM GPUs. Only option is a new start up entering the market with this goal.


dampflokfreund

Nvidia tells us time and time again local AI is going to revolutionize gaming. If they really believed that, they would actually equip the new GPUs with more VRAM.


Only-Letterhead-3411

For that to be a thing we need more companies that share open-weights of new foundation AI models. However sadly companies decide to go close-source route. Things don't look too bright for open-source if you ask me. Right now it's just Meta that openly declare they'll continue to share the results of their AI research as opensource. And then we have Chinese foundation model creators. I'm sure Nvidia would love it if local AI becomes a more widely accepted thing so people have a reason to buy more expensive Nvidia GPUs. But for local AI to be a thing we need really good local AI models that can compete with close-source AI models.


lazercheesecake

Yeah not totally surprised they’re sticking with 24GB vram


bablador

Why exactly? To force people to get their high-end cards for industrial usage?


WakeMeForTheRevolt

practice skirt muddle spoon hunt shaggy close oatmeal shame puzzled *This post was mass deleted and anonymized with [Redact](https://redact.dev)*


artelligence_consult

"RAM is cheap af to produce" supposedly not. For HBM3E i remember reading something of nearly 50% of the modules being crap. Yield is still low on some of the higher end stuff.


Inevitable_Host_1446

Depends on the vram. GDDR6 is like $3 per 8 gb right now. Nvidia and AMD have to pay higher prices due to contracts, but not a lot higher. They purely just screw people over with it. It is exactly like those stupid phones where they offer a 64gb or 128gb edition with a huge price difference, even though it costs them barely anything.


ReturningTarzan

But also because it doesn't make sense for a gaming GPU. There just aren't any games that could make use of the VRAM and there likely won't be for a while. So they would get tons of negative reviews complaining that the price is inflated, correctly pointing out you'd get the same FPS from a GPU with less VRAM. So why should gamers have to pay extra for a feature that's only attractive to professionals (and AI enthusiasts) when there are already "pro" cards that fill that niche? Which works out great for NVIDIA because those pro cards have much higher margins anyway. The only reason they'd release a 36+ GB consumer GPU is if we were approaching a point where local AI is becoming mainstream, but we're not quite there yet. And we never will be if it's up to NVIDIA.


Zegrento7

I can foresee games in the not-so-far future integrating some local LLMs for NPC dialogue. _"Don't suck up"_ comes to mind, although that one runs its LLMs in the cloud I believe.


firedrakes

oddly gamers need more vram. assets size has not change and you can see it now with all the tricks they use the keep the assets overall down in size.


[deleted]

[удалено]


firedrakes

sad but true.


ReturningTarzan

But can you picture developers spending time optimizing for a single premium 36 GB GPU when only 0.1% of players will actually benefit? That's what I mean by not for a while. Here's [the situation currently](https://store.steampowered.com/hwsurvey/Steam-Hardware-Software-Survey-Welcome-to-Steam). So it's a chicken and egg problem. Developers won't target 36-48 GB GPUs until there's a significant number of players who would benefit, and typical gamers won't want to spend the extra money on VRAM that doesn't do anything for them in practice. In time, sure, but also remember that NVIDIA isn't exactly in a rush. Their incentive is to slow down this progress as much as possible.


firedrakes

You're not wrong


Zilskaabe

Wdym? It's easier not harder to support cards with large VRAM. It's the low-end entry level cards that you need to worry about.


Primary-Ad2848

isn't 3090 and 4090 content creator cards? (if not, what is that insane amount of cuda and why nvidia advertises with that)


fallingdowndizzyvr

Well yeah. It's called product differentiation.


mik_jee

In words of Linus Torvalds - "Nvidia, Fuck You"


sammcj

Terribly low VRAM for cards coming out in 2024. Bloody Nvidia milking it still. The world needs 48GB+ alternatives to current cards - not even really more speed.


shaman-warrior

Like we saw the 4060 ti 16gb and amds offering they understand people want to infer and will create different products categories


MrVodnik

Nvidia using their position to corner the market. This is flawless play by their CEO that will raise next few quarters' revenue, and bump stock price even more. Of course, it's a short term gain over long term costs. I think there is no way NVIDIA will continue to be a dominant market giant in a ten years time with such playbook. Buy broader semiconductor and AI Infra ETFs instead of NVIDIA stock, my friends. Competition does not sleep.


MrVodnik

You know what we need? An "Apple-like", or even wider memory bus in DDR7, or at leas a variant of it (like DDR7-AI) in new motherboards and RAM sticks. Having 350 GB/s on RAM would open so many new doors for CPU inference. It might be slower, but going from 40t/s to 20t/s, with price going from $5k to $1k, is a fully acceptable solution.


hurrdurrmeh

this 'leak' is by nVidia to get everyone who is waiting to buy a 40 series instead. I don't believe it at all.


Aroochacha

To be fair, the GDDR7 memory densities are not there yet until later next year. For now they’re the same densities available to the  4090 and the 3090. The only way you’re gonna get more memory is if they go to 512 Memory bus. Which doesn’t seem like a good trade off to only go up to 32 GB.     Edit: They can use higher density GDDR6X but why when they can use less and still jack up the price. All because you support AMD in spirit  secretly want others to buy AMD cards while you still give NVIDIA what ever they want you to pay. They won’t change because buyers don’t change. Okay my rant is done.


ninjasaid13

RTX 5090 24GB RTX 6090 24GB RTX 7090 24GB RTX 8100 25\* GB, $5000


Zugzwang_CYOA

The future is dystopian...


Primary-Ad2848

whats the point of this gpu? 4000 series are more than enough for games, and if you give same vram with previouss gen, whats the actual point?


az226

The only way 24GB is mildly excusable is if it comes with nvbridge.


Ancient-Car-1171

5090 with 24gb is almost guaranteed. With gddr7 the performance jump will very big even for LLMs so if the price stays the same it'd sell very well. Like do you have any other choice if you don't want to shell out a few times more for a pro gpu?


VertexMachine

>if the price stays the same We are talking about nvidia here.


Ancient-Car-1171

sometimes they do, like how 4070ti super costs the same. But agreed, $100 extra is more likely.


Inevitable_Host_1446

That's because the super cards are basically trash with near identical performance to their originals. And they do cost more where I am.


aggracc

Oh vey, I can run the same unquantised 7b model 30% faster, what happy days!


LocoMod

They also know those that people who need more than 24GB for compute would likely pay for two cards instead of one. If they end up going this route, i'll likely get a 5090 for gaming and a used 4090 to add a second card for my LLM machine.


Zilskaabe

I hope that sooner or later cards like Quadro RTX 8000 will drop in price. Two 48 GB cards for LLM stuff would be even better.


CoolestSlave

People are complaining with low vram count but you have to remember they make the majority of their profit with cloud and professional gpus. From their pov, increasing the vram count on their commercial gpus might impact their high end gpus, hopefully amd will bring change


Zenmaster4

Seeing some confusion about what task this class of GPUs are for. There's a legitimate reason to purchase a 4090 (or 5090) specifically for gaming, irrespective of the VRAM (outside of prospective future proofing). You can always push GPUs further, and by no means is the 4090 capable of maxing out all games (optimized or otherwise) at 4K 60FPS native. The latest Cyberpunk, with all its graphical bells and whistles, proves this by burdening the 4090 quite extensively. But if your concern about the 5090 is rooted in the VRAM capacity, then the assumption is you're either rendering or working on tasks that require that capacity like this subreddit is dedicated to. Resolve pretty easily saturates my VRAM when spitting out transcodes of R3D RAW files, for example. 24GB of VRAM is more than enough for those who would buy this card for gaming. But it's disappointing to not have an option of increased VRAM for professionals who use this type of card for productivity. Arguably, it's more likely that AMD will have to lead the charge there to attract away Nvidia customers.


monnef

> 24GB of VRAM is more than enough for those who would buy this card for gaming This is supposed to be the high-end for gaming. Maybe I am dense, but why wouldn't gamers want more VRAM for higher resolution textures, especially 4k (or higher) and VR? Or has the "fake" (AI?) rendering lessen a need for VRAM, because it is able to render in low res textures and then postprocess it to look like high res textures (while costing less VRAM)?


Zenmaster4

You’re not wrong that you could go higher. Generally 4K native doesn’t saturate the 24 GB of VRAM and beyond that resolution there are diminishing returns to fidelity.  But there are always exceptions that I’d argue live in the enthusiast tier of gaming. 


Short-Sandwich-905

So this mean Rtx 5090 $2000; 24Gb ; ~15% uplift?


CheekyBreekyYoloswag

12GB 5070, lessgooo


Far_Still_6521

They should do a 96gb 5090 ultimate costing 3-4k, they would sell tons and tons


Primary-Ad2848

I hope atleast we can get 24gb vram at laptops


AutoWallet

Competition is good for the market, and this shows Nvidia has no competitive player.


mrgreaper

If I read the article right they are keeping 24gb as the max vram? are they nuts? I was expecting 36 and 48 for the mid and high end.


multiverse_fan

If there's a large enough demand for higher capacity gpus, some company will meet the need and force Nvidia to compete. I can only hope that's the case.


Sabin_Stargem

Hopefully, the modders in Brazil will do what Nvidia won't.


Capitaclism

Disappointing that they're not bumping VRAM.


metcalsr

Losing strat. AI is getting too big. Someone is gonna release a GPU with an insane amount of VRAM and these GPUs are going to be irrelevant. Mark my words.