There have been multiple accounts created with the sole purpose of posting advertisement posts or replies containing unsolicited advertising.

Accounts which solely post advertisements, or persistently post them may be terminated.

_sideffect ,

And only needs 600w!

drasglaf ,
@drasglaf@sh.itjust.works avatar

And it will cost 3000€

wreckedcarzz ,
@wreckedcarzz@lemmy.world avatar

WHO’S THAT POKÉMON?

TomAwsm ,

Rotom RTX?

ramble81 ,

Cool cool…. What about the price? That’s all I care about at this point.

Crashumbc ,

3k

ArtVandelay ,
@ArtVandelay@lemmy.world avatar

No no 5090 is the price, not the model

neidu2 , (edited )

Yeah, about clock speeds… remember when they were front and center 20 years ago while marketing CPUs? Intel started marketing CPUs by their clock speeds in the 90’s, hilighting that as a selling point over their competitors that usually ran at slightly lower clock speeds.

But Intel painted themselves into a corner: Clock speeds don’t matter - instruction sets and floating point ops per seconds do. In the mid 2000s they had to slowly phase out the clock speed marketing, as clock speeds had reached such levels that further increases would be detrimental to performance, so they had to change their marketing and branding strategy.

As soon as clock speed marketing had been phased out, Intel CPUs actually ran at lower speeds than the previous generation, while still outperforming them.

I’m curious to see whether nvidia is about to do the same thing.

deegeese ,

GPU code is more amenable to high clock speeds because it doesn’t have the branch prediction and data prefetch problems of general purpose CPU code.

Intel stopped chasing clock speed because it required them to make their pipelines extremely long and extremely vulnerable to a cache miss.

Dudewitbow ,

also to bring a rudamentary comparison:

a cpu is a few very complicated cores, a gpu is thousands of dumb cores.

its easier to make something doing something low in instructions(gpu) faster than something that has a shit ton of instructions(cpu) due to like you mention, branch prediction.

modern cpu performance gains is focusing more on paralellism and in the case of efficiency cores, scheduling to optimize for performance.

GPU wise, its really something as simple as GPUs are typically memory bottlenecked. memory bandwidth (memory speed x bus width with a few caveats with cache lowering requirements based on hits) its the major indicator on GPU performance. bus width is fixed on a hardware chip design, so the simpilist method to increase general performance is clocks.

SuckMyWang ,

It could be, yes of course

  • All
  • Subscribed
  • Moderated
  • Favorites
  • [email protected]
  • random
  • lifeLocal
  • goranko
  • All magazines