• Hey, guest user. Hope you're enjoying NeoGAF! Have you considered registering for an account? Come join us and add your take to the daily discourse.

|OT| Nvidia RTX 30XX |OT|

Rikkori

Member
May 9, 2020
1,179
1,902
385
Ampere cards use GDDR6X so that might make some kind of impact. I'm not an expert on the matter but I'm reading that it provides a power advantage over standard GDDR6. How much and to what practical effect I'm not certain, but something to keep in mind.
Oh, it's going to be fine for this game for sure (with DLSS on especially), but I wanted to note the trend because some people are still in denial about vram requirements going up.
 
  • Like
Reactions: Aidah and pawel86ck

Malakhov

Member
Jun 6, 2004
7,338
1,365
1,780
Ampere cards use GDDR6X so that might make some kind of impact. I'm not an expert on the matter but I'm reading that it provides a power advantage over standard GDDR6. How much and to what practical effect I'm not certain, but something to keep in mind.
Gddr6x has double the bandwidth of gddr6 but so far they haven't been used that way. Mostly used to get the same speed but using less power for it, less heat etc.. also easier to overclock. Possibly reach double bandwidth someday
 
Last edited:

cucuchu

Member
Dec 1, 2016
444
181
340
Oh, it's going to be fine for this game for sure (with DLSS on especially), but I wanted to note the trend because some people are still in denial about vram requirements going up.
Yeah its most likely not gonna be enough to max out everything at Ultra at 4k going into late 2021. 1440p is my resolution of choice for PC gaming though so I'm not too worried about it. Even though I'm gonna try my best to get a 3080 at launch, part of me won't be that disappointed that I'm forced to see what AMD has to show and if Nvidia responds with a 20gb 3080.
 
Last edited:

Rikkori

Member
May 9, 2020
1,179
1,902
385
Yeah its most likely not gonna be enough to max out everything at Ultra at 4k going into late 2021. 1440p is my resolution of choice for PC gaming though so I'm not too worried about it. Even though I'm gonna try my best to get a 3080 at launch, part of me won't be that disappointed that I'm forced to see what AMD has to show and if Nvidia responds with a 20gb 3080.
It's a deal with the devil, I'm doing the same. Want a new card ASAP and will just deal with re-selling 2 years down the line for an upgrade, rather than waiting around for it all to settle. I'll do a more long term upgrade then to a proper flagship (4090 or whatever) in conjunction with probably an 8K miniLED TV. Still on 4K 60/1440p 120 until then so the 3080 will manage it easier.
 

YodaBB

Member
Nov 19, 2019
136
192
230
I think I'm going to wait for either a 12gb+ 3080Ti or maybe a AMD 6000. Best of luck to everyone hoping to snag a card this week! :goog_horns:
 
  • Like
Reactions: Rygeist

Rygeist

Member
Jan 28, 2019
112
120
240
Super disappointed with some of the leaked benchmarks for the 3080. It doesn't seem to OC well at all. Think I'm just gonna wait for the 3080ti
 
Last edited:
Jul 29, 2013
2,398
4,453
940
US
Fortnite is getting RT, DLSS, and Reflex support on Sept. 17th. Gonna be the new battleground for Ampere vs Big Navi. We'll see how AMD does with RT. Either way this game is gonna fly with DLSS enabled.

Watch Dogs Legion requirements...


"Additionally, the game will support ray-traced reflections and NVIDIA’s DLSS Ultra Performance mode, built to accelerate 8K gaming."

This is the game they'll be bundling with some Ampere cards. Looks like 2080ti can do 4K/Ultra with RT. Ampere cards will put the beatdown on this game.
 
Last edited:

bohrdom

Member
Mar 19, 2020
87
128
210
C'mon guys 10 GB is enough, why do you want more? Didn't you hear what Nvidia said and 'le allocation izuh notuh utilisation'? :p



They are right. Allocation isn't utilization. Engine devs will optimize their engines if they need to. I wouldn't be worried about the 10 GB VRAM requirement since they developed this card for games to run at 4K @ 60 fps. Consoles are most likely not gonna use more than 10 gb of VRAM anyways.
 

blly155

Banned
Aug 3, 2014
11,983
5,310
745
8GB VRAM for 1440p and 11GB for 4K...ouch lol.

RTX 3090 it is. if this is what Watch Dogs needs then i dread to see what Cyberpunk 2077 will need.
 
Last edited:

Ironbunny

Member
Jun 11, 2019
201
366
295
Finland
8GB VRAM for 1440p and 11GB for 4K...ouch lol.

RTX 3090 it is. if this is what watch Dogs needs then i dread to see what Cyberpunk 2077 will need.
Yea that 10GB on 3080 really is a bummer. :/ Might still have to go for the 3080 and switch to super or Ti later myself. 3090 price makes my eyes water. Interesting to see if the rumours about FE having better overclocking ability compared to 3rd party is true. For the 3080 versus 3090 you are paying for the memory which imo might very well be worth it if you play at 4k or over.
 

Jimmy Rustle

Member
Jul 11, 2013
306
68
485
8GB VRAM for 1440p and 11GB for 4K...ouch lol.

RTX 3090 it is. if this is what Watch Dogs needs then i dread to see what Cyberpunk 2077 will need.
Watch Dogs won’t use the full 11gb of VRAM.

The more VRAM the GPU has available, the more the game will use up.

It will run just fine with 10gb. 11gb is recommended only because that’s what the 2080ti has, which at the time of those recommendations, is still the most powerful card on the market.
 
  • Like
Reactions: Nydus

Gen.Grievous

Member
Mar 11, 2010
1,043
49
770
Watch Dogs won’t use the full 11gb of VRAM.

The more VRAM the GPU has available, the more the game will use up.

It will run just fine with 10gb. 11gb is recommended only because that’s what the 2080ti has, which at the time of those recommendations, is still the most powerful card on the market.
Pretty obvious they just list the amount of VRAM the recommended card for that settings has.

WD Legions is not going to use 16 GB of system ram either.
 
Last edited:

Rikkori

Member
May 9, 2020
1,179
1,902
385
They are right. Allocation isn't utilization. Engine devs will optimize their engines if they need to. I wouldn't be worried about the 10 GB VRAM requirement since they developed this card for games to run at 4K @ 60 fps. Consoles are most likely not gonna use more than 10 gb of VRAM anyways.
Engine devs are gonna "optimize" their engines for a particular card? lol. Fucking Ubisoft doesn't even care to optimise games for the platform as whole simply pushing out consoles builds to PCs (ala Assassin's Creed), and you think they care if you have to turn textures down because your card has "only" 10 GB? :messenger_tears_of_joy:

Damn, I wish I had your optimism.
 

Sentenza

Member
Dec 3, 2011
14,162
910
815
Some people seem to be unaware of the fact that just because a game MAY fill as much VRAM as it's available, it doesn't mean it NEEDS that much to work properly.
Same goes with system RAM, by the way.
 

BluRayHiDef

Member
Aug 21, 2015
2,176
2,860
715
Gddr6x has double the bandwidth of gddr6 but so far they haven't been used that way. Mostly used to get the same speed but using less power for it, less heat etc.. also easier to overclock. Possibly reach double bandwidth someday
Since GDDR6X isn't an official iteration of GDDR, as it hasn't been officialized by JDEC and was co-developed by Nvidia specifically for their graphics cards, will it ever become/i] official or will its improvements be incorporated into a new iteration (GDDR7?) that will be official? Is the latter what happened with GDDR5X?
 

BluRayHiDef

Member
Aug 21, 2015
2,176
2,860
715
Since GDDR6X isn't an official iteration of GDDR, as it hasn't been officialized by JEDEC and was co-developed by Nvidia specifically for their graphics cards, will it ever become official or will its improvements be incorporated into a new iteration (GDDR7?) that will be official? Is the latter what happened with GDDR5X?
 

Rikkori

Member
May 9, 2020
1,179
1,902
385
The tl;dr is 25% over 2080 Ti, but worse OCer, so OC for OC it's more like 15% over the 2080 Ti.

What I haven't seen from anyone yet is undervolting tests, which is where I think the card will shine more than Turing. And on a side note, this bodes well for AMD, as they can absolutely match this in terms of raw performance. 7nm TSMC just shits on Samsung.
 
Jan 7, 2018
1,489
1,573
445
The cooling on FE cards are whats the word inadequate ? 57 dBA? That is AMD cooler territory. (n)

Now lets wait for AIB card reviews so we'll have a comparison. Maybe it's not just the FE cooler that is terrible, but the additional power draw over previous cards makes that cooler look much worse than it would on previous ~250 W cards. Most importantly the huge performance increase is here just not 2x 2080.
 

Rickyiez

Member
Jan 20, 2020
226
225
335
Looks like the 2X from 2080 to 3080 was (like many expected) just the best case scenario in one game with RTX enabled. Real performance increase looks to be at around 60-65% over the 2080, which is still impressive
It's almost 2 times faster in Doom Eternal 4k . Anyhow , it's a card that is 80-90% faster than my 1080Ti with Ray tracing , good enough upgrade for me

No review thread btw ?
 
Last edited:
Jul 29, 2013
2,398
4,453
940
US
Looks like the Chinese 3DMark leaks were accurate with 3080 getting a 11449 score in Port Royale RT benchmark compared to 7969 for the 2080ti. Looks like 3080 is about 25-30% more powerful than a 2080ti.

Quick look at some benchies:
-RDR2-
3080 = 4K/69fps/Max Settings
2080ti = 4K/49fps/Max Settings

-FS 2020-
3080 = 4K/42fps/Ultra Settings
2080ti = 4K/35fps/Ultra Settings

-Death Stranding-
3080 = 4K/106fps/VHQ Settings
2080ti = 4K/81fps/VHQ Settings




 

Ellery

Member
Feb 17, 2019
1,230
2,257
505
34
Manches, Amsterdam, Berlin
It's almost 2 times faster in Doom Eternal 4k . Anyhow , it's a card that is 80-90% faster than my 1080Ti with Ray tracing , good enough upgrade for me

No review thread btw ?
Yup, but I am looking at averages throughout many games. That is where the 65% number came from.

Holy shit!

415 watts for 3080



Terrible
It is for the entire system
 
Last edited:
  • Like
Reactions: CrustyBritches

Rickyiez

Member
Jan 20, 2020
226
225
335
It is for the entire system
I swear some people doesn't read the context of the graph or information . They just deduce the conclusion they wanted , and have a chance to hate or say something like "terrible" . Same thing with the Watchdog Legion requirement and vram section
 
Jun 15, 2020
120
193
210
so the absolute latest and greatest 30tflop card cant even get 45fps on a new release game in 4k?

I know FS2000 melts PCs but i would have thought a brand new 30tflop card would have been able to handle it better.

with thats said, i assumed a 2080ti could do better than 35fps.

Im assuming it would get 60fps in 1440p end DLSS bringing it up to 4k