NVidia GeForce RTX 50x0 cards - 70% performance increase, but AI > you

jayrebb

Naxxramas 1.0 Raider
15,127
15,076
Start making 4 series again and put the 4090 back into production is the best apology.
 
  • 4Like
Reactions: 3 users

jayrebb

Naxxramas 1.0 Raider
15,127
15,076
nvidia really does not give a fuck anymore about making GPUs, zotacs are shipping out with fewer cores than its supposed to lol


1740162784679.png
 
  • 5Worf
Reactions: 4 users

Grez

Trakanon Raider
1,273
1,236
Start making 4 series again and put the 4090 back into production is the best apology.
Won't happen. From what I understand, the 4090s and 5090 use the same equipment or machinery or whatever. They're not going to discontinue the 5090 or reduce the production rate of the 5090 to make 4090s.
 

Malakriss

Golden Baronet of the Realm
12,880
12,268
Christ if this snuck by even for the cards they sent reviewers, how bad is it in the wild
 
  • 2WTF
  • 1Worf
Reactions: 2 users

Borzak

Silver Baron of the Realm
26,646
35,473
Hmm premium price for a non premium chip/card. Sounds like a winning combination for a company.
 
  • 3Like
  • 1Clown World
Reactions: 3 users

Neranja

<Bronze Donator>
2,784
4,498
I'm honestly waiting until they just say fuck it and nuke everything but the flagship card and let amd/intel have all the stuff below that. Unless for some reason their AI/Datacenter business crashes.
That probably won't happen, because Jensens ego won't allow to admit defeat in a field he once dominated. That man probably hasn't heard the word "no" in years.

Also, especially after DeepSeek we are approaching the "trough of disillusionment" in regards to "AI"--especially when people realize how it fundamentally works, and why it's not going to get much better, regardless how much money you throw at it.

I can explain this in detail, but it will probably a long rant, and I don't want to type it out if no one's interested.
 
  • 1Like
Reactions: 1 user

Hateyou

Not Great, Not Terrible
<Bronze Donator>
17,779
48,114
That probably won't happen, because Jensens ego won't allow to admit defeat in a field he once dominated. That man probably hasn't heard the word "no" in years.

Also, especially after DeepSeek we are approaching the "trough of disillusionment" in regards to "AI"--especially when people realize how it fundamentally works, and why it's not going to get much better, regardless how much money you throw at it.

I can explain this in detail, but it will probably a long rant, and I don't want to type it out if no one's interested.
Go ahead
 
  • 3Like
Reactions: 2 users

Neranja

<Bronze Donator>
2,784
4,498
This is an ELI5 version of how AI works: If you enter a prompt, the "AI" -- or more precise, the "language model" responds with a chain of words. The AI puts these together according to the statistical probability of what the next word according to a) the input and b) the training data would be (probability vector, or stochastic vector).

The AI bros claim that to improve the quality of the output and make the model more "intelligent", we just need to build a bigger model, with more parameters. Which requires more memory and computational power:

1740242954905.png


This is why OpenAI wants $500 billion: They dangle the "singularity" (when "AI is becoming smarter than humans" and improves itself) at the end of the curve above, like a carrot in front of a donkey. "Trust us, we just need a bigger boat."

The problem with that assumption is: The curve goes probably more like this:

1740244175058.png


And here's a video you can now watch:
 
  • 8Like
Reactions: 7 users

Tripamang

Naxxramas 1.0 Raider
5,591
34,899
This is an ELI5 version of how AI works: If you enter a prompt, the "AI" -- or more precise, the "language model" responds with a chain of words. The AI puts these together according to the statistical probability of what the next word according to a) the input and b) the training data would be (probability vector, or stochastic vector).

The AI bros claim that to improve the quality of the output and make the model more "intelligent", we just need to build a bigger model, with more parameters. Which requires more memory and computational power:

View attachment 575114

This is why OpenAI wants $500 billion: They dangle the "singularity" (when "AI is becoming smarter than humans" and improves itself) at the end of the curve above, like a carrot in front of a donkey. "Trust us, we just need a bigger boat."

The problem with that assumption is: The curve goes probably more like this:

View attachment 575118

And here's a video you can now watch:


Old news bro, thinking models have proven the wall is a lie and the more compute you give the model to spend on a question at inference time the more accurate the response. In fact the models are likely too big as they are now and we could spend more time training them on logic and less on sheer data as it makes them more accurate. For example if there are too concepts in completely different topics but the same name the model will give less satisfactory answers for both. If they are smart enough to understand a topic but not necessarily trained on it they can give better answers by researching the topics before answering then creating the output. You're pushing the compute from training to inference time, and getting better responses.

Tldr there is no ai wall and we're well into exponential performance improvements and getting close to logarithmic.
 

Neranja

<Bronze Donator>
2,784
4,498
First of all, let me preface this by stating that I didn't claim AI models were useless, just that we are approaching the "trough of disillusionment". In fact, we already made very useful things 20 years ago with similar statistical models, like email SPAM filters using Naive Bayes classifiers and Markov networks.

How do you think SPAM filters work? They get trained. Still, the accuracy for such a seemingly simple task compared to full-blown LLMs never got to 100%.

Old news bro, thinking models have proven the wall is a lie and the more compute you give the model to spend on a question at inference time the more accurate the response.
So this is still the current cope? Just spend more time on compute? You realize we are quickly approaching the end of Moore's Law, as improvements in computational power seem to be tied more to node shrinks instead of architectural design nowadays (*cough* Blackwell *cough*), but there is a physical limit on how much further we can shrink? Production is getting more complex and expensive as well, killing yields?

Ponder about this for a while: 3 nm is around 27 silicon atoms.

No, ponder about it some more. Now after some pondering, answer this: How much more do you expect to shrink before yields fall off a cliff, single atom defects make a trace and therefore the complete die unusable, or even worse: the field effects for transistors just outright stop working?

In fact the models are likely too big as they are now and we could spend more time training them on logic and less on sheer data as it makes them more accurate.
Funny enough, this is what DeepSeek innovated: They have a bag that contains multiple models that they call Mixture-of-Experts.

OpenAI's first response ("they stole our training data/models") was very telling.

If they are smart enough to understand a topic but not necessarily trained on it they can give better answers by researching the topics before answering then creating the output. You're pushing the compute from training to inference time, and getting better responses.
And here we have Ouroboros eating its tail. What you describe, in layman terms, is just "let me google that for you, and also build a model real quick".

a) Building models is a pretty expensive operation
b) Search engines themselves are starting to use AI to rank and filter results, feeding the model tainted data, so you need to build your own Google for untainted data
c) But the internet is for porn, so 80% of it is trite shit. Even worse, now more and more of shit is AI generated, which feeds into the models.
d) Everyone is starting to get realy pissed about AI bros collecting as much data as possible, with hammering servers, ignoring robots.txt, and outright pirating 82 TB of books.

Expect a lot more laws and legal precedents to fall into place to regulate AI, now that everyone realizes the AI bros behaved like a bull in a china shop to acquire "training data", and the "it's just training data" excuse stops working because people realize the underlying statistical nature of AI.

- * -​

There is an even more fundamental challenge in regards to AI: it gets trained on "human knowledge." We can expect--from experience--that at any point in time, 50% of human knowledge is probably to most likely wrong.
Do you still believe the geocentric model, or that smoking doesn't cause cancer? Did you have a lobotomy yet? It's an easy procedure, you know:It won the 1949 Nobel Prize for physiology or medicine. No? Why don't you trust the science?

The current state of human knowledge is fucked, because we incentivized "publish or perish", which lead to loads of junk papers and the cherry-picking or even outright falsifying of data, and are now facing a replication crisis, even in fields that should--by definition of being a STEM field, or by ethical standards--be very much immunie to this, like medicine. If you have time you can go down some deep Chinese rabbit holes.

People who go around clamoring "trust the science" have no idea how science works. And you want to train AI models on all of this?
 
  • 3Like
Reactions: 2 users

gak

<Gold Donor>
2,002
10,544
hub.gif

NVIDIA GeForce RTX 5070 Ti Appears With Fewer ROPs; Problem Now Extends To Other High-End RTX 50 GPUs

If you think the reports of fewer ROPs on the RTX 5090 have stopped surfacing, then you should check another one, but this time, it's the GeForce RTX 5070 Ti. It should be obvious now, as NVIDIA itself confirmed the RTX 5070 Ti to be one of those affected GPUs, but it doesn't seem to be as small as NVIDIA wants to show.

A user on Facebook posted a screenshot of the GPU-Z (via @GawroskiT), revealing the reduced ROP count on an RTX 5070 Ti edition. While we don't know the exact GPU edition and to which vendor it belongs, this confirms the presence of such RTX 5070 Ti GPUs in the market. Compared to the 96 ROPs specified by NVIDIA in the official spec sheet for the GeForce RTX 5070 Ti, this one brings only 88 ROPs. This is an 8.4% reduction, which is noticeable and will definitely impact the gaming performance.

While NVIDIA says that the impact is 4% on average, it could be higher or lower depending on the games being played. Nonetheless, this figure is still high and cannot be excused since customers are paying the price in full. In most cases, they are paying even more than what the GPU actually costs, given the inventory situation.

Now, if you take a quick look at some other specs of this particular RTX 5070 Ti edition, you can clearly see that the reduction in ROPs has impacted the Pixel Fillrate, which is directly related to ROPs. Compared to 287.7 GPixel/s on the RTX 5070 Ti with 96 ROPs, the nerfed RTX 5070 Ti delivers only 223.7 GPixel/s of Pixel Fillrate. However, do keep in mind that the boost clock on the latter is much lower. Keeping the boost clock at 2.99 GHz on the affected GPU still delivers 263.12 GPixel/s of Pixel Fillrate, which is roughly 9% lower.

NVIDIA and its board partners are misleading its customers by not notifying them of inferior specs even though these cards pass QA testing before reaching the shelves. Despite claiming that these account for up to only 0.5% of all the manufactured units, the number can quickly go into the thousands.

n.gif
 
  • 4Worf
Reactions: 3 users

Big Phoenix

Pronouns: zie/zhem/zer
<Gold Donor>
47,663
102,831
View attachment 575258
NVIDIA GeForce RTX 5070 Ti Appears With Fewer ROPs; Problem Now Extends To Other High-End RTX 50 GPUs

If you think the reports of fewer ROPs on the RTX 5090 have stopped surfacing, then you should check another one, but this time, it's the GeForce RTX 5070 Ti. It should be obvious now, as NVIDIA itself confirmed the RTX 5070 Ti to be one of those affected GPUs, but it doesn't seem to be as small as NVIDIA wants to show.

A user on Facebook posted a screenshot of the GPU-Z (via @GawroskiT), revealing the reduced ROP count on an RTX 5070 Ti edition. While we don't know the exact GPU edition and to which vendor it belongs, this confirms the presence of such RTX 5070 Ti GPUs in the market. Compared to the 96 ROPs specified by NVIDIA in the official spec sheet for the GeForce RTX 5070 Ti, this one brings only 88 ROPs. This is an 8.4% reduction, which is noticeable and will definitely impact the gaming performance.

While NVIDIA says that the impact is 4% on average, it could be higher or lower depending on the games being played. Nonetheless, this figure is still high and cannot be excused since customers are paying the price in full. In most cases, they are paying even more than what the GPU actually costs, given the inventory situation.

Now, if you take a quick look at some other specs of this particular RTX 5070 Ti edition, you can clearly see that the reduction in ROPs has impacted the Pixel Fillrate, which is directly related to ROPs. Compared to 287.7 GPixel/s on the RTX 5070 Ti with 96 ROPs, the nerfed RTX 5070 Ti delivers only 223.7 GPixel/s of Pixel Fillrate. However, do keep in mind that the boost clock on the latter is much lower. Keeping the boost clock at 2.99 GHz on the affected GPU still delivers 263.12 GPixel/s of Pixel Fillrate, which is roughly 9% lower.

NVIDIA and its board partners are misleading its customers by not notifying them of inferior specs even though these cards pass QA testing before reaching the shelves. Despite claiming that these account for up to only 0.5% of all the manufactured units, the number can quickly go into the thousands.

View attachment 575263
wheeze.png