• Hey, guest user. Hope you're enjoying NeoGAF! Have you considered registering for an account? Come join us and add your take to the daily discourse.

PS5 Die Shot has been revealed

silent head

Member
You want the simplest answer?

Series X:
10GB of GPU Optimal Memory at 560GB/s (Series X GPU is using this, its performance is great)
6GB of Standard Memory at 336GB/s (If the GPU touches any of this RAM performance WILL suffer)

Meanwhile on PS5:
16GB at 448GB/s

That is the primary reason we see PS5 winning ANY performance battles so far. It's developers having games that are organized in their memory access patterns in a certain way, and it's just plain simpler to work with the PS5's fully symmetric memory design. There's less to plan for or think about. Series X's design requires more careful planning, or use of a feature like Sampler Feedback Streaming to greatly mitigate the just 10GB of faster RAM for the GPU. If the Series X GPU touches that other 3.5GB reserved for games past that 10GB, the GPU's performance will dip because now you have a 52 Compute Unit 1825MHz GPU now trying to survive on just 336GB/s of memory bandwidth. It becomes bandwidth starved.

Some developers handle this better than others. In cases where series x and ps5 perf are identical a dev has just opted for parity (don't have an issue here so long as the game is solid on both), or the additional headroom offered by Series X allows it to JUST match what PS5 is outputting or brings it slightly below. Control is clear evidence of there being more headroom for higher performance on Series X. That additional performance headroom on Series X existing is how you get cases like in Hitman 3.

That is the reason PS5 sometimes performs better. It's down to the memory design causing the Series X GPU to be bandwidth starved in momentary blips. There are solutions to this.
Hitman 3?
Series X ~ 9.7TF rdna1
0rHnf4l.jpg
 

quest

Not Banned from OT
I think the lack of dram could cause issues with relying on fast streaming. Aren't both ps5 and Series X TLC ssds? TLC tends to have bad performance, sometimes a portion of the TLC ssd is converted into SLC to serve as cache and to improve performance. Not sure if either of the consoles are doing that.

I thought for the most part the dram on drive was mostly for writes to keep them from queueing up a basically a cache. Reads larger arrays like more drives to spread the data over helps speed. I could be misinterpreted things.
 

FranXico

Member
I think the lack of dram could cause issues with relying on fast streaming. Aren't both ps5 and Series X TLC ssds? TLC tends to have bad performance, sometimes a portion of the TLC ssd is converted into SLC to serve as cache and to improve performance. Not sure if either of the consoles are doing that.
The PS5 SSD uses a DRAM, but that is located closer to the SSD, you could see it in the teardown.
 

Proelite

Member
If you're referring to the CPU it's the same setup as on Xbox right?
The only difference is the cut down FPU.


Playstation 5's processor is a custom one based on the Zen 2 architecture from AMD, though according to some of my own information, the difference primarily between the Zen 2 cache of the Desktop processors and the PS5 is the unified L3 cache. I was told that this cache is only 8MB for the PS5, but the unified design helps drastically reduce latency.
 
Last edited:

ethomaz

Banned
I will say the three things I don't understand/unknown in the die shot:

- The cut FPU
- The big silicon space bettween the Physical GDDR6 Memory Interface (in both Series S|X and AMD APUs it is very close).
- Most of the right side of the die shot.
 
Last edited:

assurdum

Banned
Hitman 3?
Series X ~ 9.7TF rdna1
0rHnf4l.jpg
"Hitman 3? You see that's a solid prove, an undisputable fact, the war is over, wider margin of improvement is possible without any doubt in every game on series X, who said otherwise denies the only true truth. The rest of multiplat? Meh, means nothing, early tools, bad optimization, rushed port, Sony easy peasy tools, nothing to discuss there."
Basically.
At least wait to see more games?
Maybe things could change, who knows, but Jeez that's ridiculous all that evangelion for Hitman 3. Furthermore Hitman 3 specs required are ridiculous for RAM and CPU, that could also explain why they can push so much the GPU on series X.
 
Last edited:
Wow, big props to MBG for owing up to his part in spreading some of the RDNA 3 "secret sauce" rumors; tbf he always said it was just stuff he was hearing from other people and warned viewers to take it with a grain of salt. He actually went over his part (intentionally or not) in it and rather thoroughly, too.

Quite better than RGT tbh, who just kind of skimmed over any specifics that were challenged/debunked with the die shots (skipping right over Infinity Cache and just saying "a lot of cache", for example). Tho I think he's recovering from being sick so they might not feel like clarifying their 'sources' and such right now.

Now just waiting on MLID :giggle:

Proelite Proelite Lol probably wouldn't go that far, but he definitely rush right over any IC commentary with a quickness.
 
Last edited:
Wow, big props to MBG for owing up to his part in spreading some of the RDNA 3 "secret sauce" rumors; tbf he always said it was just stuff he was hearing from other people and warned viewers to take it with a grain of salt. He actually went over his part (intentionally or not) in it and rather thoroughly, too.

Quite better than RGT tbh, who just kind of skimmed over any specifics that were challenged/debunked with the die shots (skipping right over Infinity Cache and just saying "a lot of cache", for example). Tho I think he's recovering from being sick so they might not feel like clarifying their 'sources' and such right now.

Now just waiting on MLID :giggle:

Proelite Proelite Lol probably wouldn't go that far, but he definitely rush right over any IC commentary with a quickness.

Hopefully the channels that were spreading that negativity about the PS5 also come clean and apologize. For a moment I thought the die shot was extremely bad news for the PS5 but there's actually positive things it that I've missed due to the FUD.
 

Proelite

Member
Hopefully the channels that were spreading that negativity about the PS5 also come clean and apologize. For a moment I thought the die shot was extremely bad news for the PS5 but there's actually positive things it that I've missed due to the FUD.

Lesson here is that youtube channels spread FUD in general, sooner or later. Their main goal is to get their numbers up.
 
Last edited:

Loxus

Member


After spending some time trying to understand the cache differences in Ryzen.
I realized something about the PS5.

Notice how in Zen 2 the Global Memory Interconnect separates the two Zen 2 core clusters.
yAtRwuE.jpg

While in Zen 3, nothing is separating the Core clusters.
KjY5f5R.jpg


The same can be said for PS5, nothing is separating the Core clusters.
Pretty similar to Zen 3.
So my question is after seeing Core complex 1 basically touching Core complex 2.
Can Core complex 1 access Cache from Core complex 2 making it Unified Cache?
yPIGM5g.jpg


The same can't be said for XBSX, as the Core cluster are separate like in Zen 2.
cgjkQat.jpg
 
After spending some time trying to understand the cache differences in Ryzen.
I realized something about the PS5.

Notice how in Zen 2 the Global Memory Interconnect separates the two Zen 2 core clusters.
yAtRwuE.jpg

While in Zen 3, nothing is separating the Core clusters.
KjY5f5R.jpg


The same can be said for PS5, nothing is separating the Core clusters.
Pretty similar to Zen 3.
So my question is after seeing Core complex 1 basically touching Core complex 2.
Can Core complex 1 access Cache from Core complex 2 making it Unified Cache?
yPIGM5g.jpg


The same can't be said for XBSX, as the Core cluster are separate like in Zen 2.
cgjkQat.jpg

It looks like Sony did that so they could fit all those cores on a small die. Doesn't look like unified cache to me.
 

Proelite

Member
After spending some time trying to understand the cache differences in Ryzen.
I realized something about the PS5.

Notice how in Zen 2 the Global Memory Interconnect separates the two Zen 2 core clusters.
yAtRwuE.jpg

While in Zen 3, nothing is separating the Core clusters.
KjY5f5R.jpg


The same can be said for PS5, nothing is separating the Core clusters.
Pretty similar to Zen 3.
So my question is after seeing Core complex 1 basically touching Core complex 2.
Can Core complex 1 access Cache from Core complex 2 making it Unified Cache?
yPIGM5g.jpg


The same can't be said for XBSX, as the Core cluster are separate like in Zen 2.
cgjkQat.jpg

You can see in the Zen 3 die shot the entire cache is unified.
 

Xyphie

Member
If it was a unified cache you would have all 8 MB as one chunk and the cores surrounding it. We've known since Summer of 2019 that it was 2x4MB when either AMD or Sony leaked benchmarks of the SoC on the net, why do people keep going on about this?
 

Proelite

Member
If it was a unified cache you would have all 8 MB as one chunk and the cores surrounding it. We've known since Summer of 2019 that it was 2x4MB when either AMD or Sony leaked benchmarks of the SoC on the net, why do people keep going on about this?

I predict we'll still be debating in 2026 whether or not PS5 has unified L3 and/or hw VRS. Whether or Xss is holding back the gen.
 
No, it's not running on magic. We need to stop this now. There are a few very clear possibilities here and all of them are with info we've known for several months now:

[For PlayStation 5]

1: The PS5 GPU cache scrubbers are being leveraged​
2: The PS5 GPU's faster clock speed benefits more current-gen engines or engines affected by clock speeds more than wider GPU designs​
3: The PS5's I/O memory subsystem is more robust both on paper and in practice with keeping and feeding data to memory as needed​

[Against Series X]

1: There is a larger bandwidth penalty on the segmented fast/slow memory pools than anticipated (either due to design or due to software OS/kernel and/or GDK API settings that have to be optimized)​
2: The I/O memory subsystem is suboptimal either out of design or due to lack of feature availability (DirectStorage is not readily available for PC developers until quite later in 2021, so many 3P devs might not be leveraging that part of XvA. And if they aren't leveraging that, they also aren't leveraging SFS)​
3: The CPU is not fast enough to keep up with drawcalls for GPU while also needing to handle data transfer between the fast/slow memory pools and some overhead (1/10th of a core according to MS) for the XvA I/O memory subsystem when SMT is enabled (this can be "solved" by simply clocking the CPU higher, if the thermals allow for it, through a firmware patch)​
4: The GPU clock is 405 MHz slower than PS5's plus lacks hardware cache scrubbers, which can have a negative impact on game engine that are built with narrower GPU designs in mind and more reliant on faster GPU clocks (majority of current-gen game engines are this way)​

...if we start to see consistent performance in favor of Series X later in the year, then those advantages can simply be explained as:

[For Series X]

1: The system is able to maintain data transference between fast/slow memory pools optimally​
2: GPU-bound memory has a 112 GB/s bandwidth advantage over PS5's GDDR6 memory bandwidth​
3: The GPU is wider, so more work can be issued and processed simultaneously per issue compared to PS5​
4: The I/O memory subsystem has all components readily available and in use; even if I/O memory subsystem performance in​
isolation still favors PS5, the gulf could be at a point where the difference is not large enough to negatively impact the other​
aforementioned advantages (higher bandwidth in GPU-optimized pool, 100 MHz faster CPU clock, lowered overhead of hardware​
resources by system in general, etc.)​
Also worth adding that while Series X's audio might be somewhat less sophisticated than Tempest Engine, it also requires less memory
bandwidth most likely. Meaning less of a squeeze by the audio and that leaves even more for the GPU, under intense audio workloads. Of
course, all of these points are IF the performance metrics swing in favor for Microsoft longer-term; currently that is not the case.
I miss your underboobs... I mean, this is a surprisingly reasonable assessment.
 
Good old Colonel James it's nice to meet you again after more than 10 years and see you still fighting for your team .
Proven wrong by what? just to get me a culture about it
I'm not sure why you act like the guy is fighting, do we need to act like the seriex is somehow in a league of its own in therms of performance compared to the PS5? Bown down to statues of the Spencer?

Just watch any of the analysis and comparisons released so far, none has a 'clear winner'.
 
"Hitman 3? You see that's a solid prove, an undisputable fact, the war is over, wider margin of improvement is possible without any doubt in every game on series X, who said otherwise denies the only true truth. The rest of multiplat? Meh, means nothing, early tools, bad optimization, rushed port, Sony easy peasy tools, nothing to discuss there."
Basically.
At least wait to see more games?
Maybe things could change, who knows, but Jeez that's ridiculous all that evangelion for Hitman 3. Furthermore Hitman 3 specs required are ridiculous for RAM and CPU, that could also explain why they can push so much the GPU on series X.
That Hitman 3 comparison was done in a cutscene. Not gameplay.
 

jroc74

Phone reception is more important to me than human rights



The 128bits theory doesn’t fit.
The remove of FADD doesn’t fit.
They don’t know what AMD/Sony did here.

People that knows are in NDA.

I remember Matt Hargett mentioning CCX before. 1 vs 2? Separate vs combined? I can't recall exactly what it was about.

As always, the plot thickens...
 

Lysandros

Member
After spending some time trying to understand the cache differences in Ryzen.
I realized something about the PS5.

Notice how in Zen 2 the Global Memory Interconnect separates the two Zen 2 core clusters.
yAtRwuE.jpg

While in Zen 3, nothing is separating the Core clusters.
KjY5f5R.jpg


The same can be said for PS5, nothing is separating the Core clusters.
Pretty similar to Zen 3.
So my question is after seeing Core complex 1 basically touching Core complex 2.
Can Core complex 1 access Cache from Core complex 2 making it Unified Cache?
yPIGM5g.jpg


The same can't be said for XBSX, as the Core cluster are separate like in Zen 2.
cgjkQat.jpg
I think it's very possible that this can have some latency implications. Very nice post.
 
Last edited:
I miss your underboobs... I mean, this is a surprisingly reasonable assessment.

Boobs and booties help keep a man well grounded đź‘Ť

I don't get why some dev just not leaks the beans. Not like they can do anything about it.

They'd probably get blacklisted out of the industry if they're under NDA, and publishers would probably want to avoid them too. Not worth the risk unless you're already planning to leave the industry altogether or are retired.
 
Last edited:



The 128bits theory doesn’t fit.
The remove of FADD doesn’t fit.
They don’t know what AMD/Sony did here.

People that knows are in NDA.

The tyranny of the NDA must end. The secret sauce has to be revealed.

Don't know why they're focusing so much on such a simple topic, instead of the more complex ones like the GE, cache scrubbers and I/O complex within that soc. Then again probably cause it's easier to compare to something they know about instead of custom hardware they have no source material to go by (and even then, they clearly got it wrong on first sight). We make no progress on Day 2, other than.... nobody knows what the f we're looking at. :messenger_grinning_smiling:
 

Kenpachii

Member
Boobs and booties help keep a man well grounded đź‘Ť



They'd probably get blacklisted out of the industry if they're under NDA, and publishers would probably want to avoid them too. Not worth the risk unless you're already planning to leave the industry altogether or are retired.

They can leak it without even exposing themselves perfectly fine. The same as people leak information towards normally media. How would sony know what company did so?
 
Indeed, that is one of the few things posted in this thread that is actually worth something lol.
I think it's very possible that this can have some latency implications. Very nice post.
Yes they maybe reduced the latency between both CCXs. On PS4 the CCXs could read each others L2 cache but it was with much higher latency (190 cycles).

0t0O5D0.png
 
But Hitman 3 is performing better on PS5. What the hell you are babbling about.

Nah it doesn't, cut it out. Lower resolution and lower quality shadows, whereas Series X is Native 4K. A single section in the entire game dipping on Xbox when the rest of the game runs at a flawless 60fps with the exception of that benchmark cutscene DF used that drops on all platforms qualifies as the PS5 running the game better? Don't go making up stuff now to make yourself feel better. Stick to the facts.
 
After spending some time trying to understand the cache differences in Ryzen.
I realized something about the PS5.

Notice how in Zen 2 the Global Memory Interconnect separates the two Zen 2 core clusters.
yAtRwuE.jpg

While in Zen 3, nothing is separating the Core clusters.
KjY5f5R.jpg


The same can be said for PS5, nothing is separating the Core clusters.
Pretty similar to Zen 3.
So my question is after seeing Core complex 1 basically touching Core complex 2.
Can Core complex 1 access Cache from Core complex 2 making it Unified Cache?
yPIGM5g.jpg


The same can't be said for XBSX, as the Core cluster are separate like in Zen 2.
cgjkQat.jpg


No way. Cross communication between complexes is one of the absolute most disastrous means of killing performance that exists on these processors if it isn't unified. You don't have one core complex dip over into the cache of the other. It's not unified if there are still two separate L3 caches per complex, which there are.

The massive performance penalty isn't worth snooping another core complex's cache. It will always make more sense to organize your data so you're never in a position where you're needing to cross communicate in such a fashion. It'll never be a performance win over making sure the right data is inside the local L3 when needed.
 

Shmunter

Member
Nah it doesn't, cut it out. Lower resolution and lower quality shadows, whereas Series X is Native 4K. A single section in the entire game dipping on Xbox when the rest of the game runs at a flawless 60fps with the exception of that benchmark cutscene DF used that drops on all platforms qualifies as the PS5 running the game better? Don't go making up stuff now to make yourself feel better. Stick to the facts.
Yeah but Hitman 3 is also only 30fps on One X vs 60fps on Pro. What does that indicate?

The game has a bizzare setup, hanging ones hat on it as the definitive benchmark is flawed. And even then the performance delta is there, XsX is clearly aiming beyond it’s capability in parts.
 
Last edited:

kyliethicc

Member
After spending some time trying to understand the cache differences in Ryzen.
I realized something about the PS5.

Notice how in Zen 2 the Global Memory Interconnect separates the two Zen 2 core clusters.
yAtRwuE.jpg

While in Zen 3, nothing is separating the Core clusters.
KjY5f5R.jpg


The same can be said for PS5, nothing is separating the Core clusters.
Pretty similar to Zen 3.
So my question is after seeing Core complex 1 basically touching Core complex 2.
Can Core complex 1 access Cache from Core complex 2 making it Unified Cache?
yPIGM5g.jpg


The same can't be said for XBSX, as the Core cluster are separate like in Zen 2.
cgjkQat.jpg
This is something I've wondered too. Seems like it could help lower latency since all 8 PS5 CPUs are adjacent and uninterrupted.

Its not the same as Zen3 unified L3 cache, where all 8 cores can access the entire central cache, but it could improve performance.

Might be why the PS5 seems to perform better in CPU tests. Its clearly a customized Zen2 8 core.
 

Kerlurk

Banned
 
Last edited:
Yeah but Hitman 3 is also only 30fps on One X vs 60fps on Pro. What does that indicate?

The game has a bizzare setup, hanging ones hat on it as the definitive benchmark is flawed. And even then the performance delta is there, XsX is clearly aiming beyond it’s capability in parts.
It only took one game out of the sample not conforming to the trend for the hybernators to go berserk to the same old nonsense. It's like trying to flip the reality of the situation on its face - that's, the PS5 being the best console for multiplatform games performance on average (without even resorting to experience differentiators like the controller). It's more desperation than straw clutching, something these folks are very familiar with over these past 7 years. These folk are working very hard to reverse the narrative reversal that occurred at the start of the gen (not even with DF's aid can they reserve the damage - but some are oblivious to that fact, or the fact that even DF is taking reputational damage in the process). It's funny to witness how hard they're working at it.... see how far they take the rope with which to hang themselves. Then again they clearly don't care how they're perceived, as it should btw. Life is too short, specially on the internet. Likewise for those that clown them to oblivion for their clownery - all is fair.
 
Last edited:
Hitman 3?
Series X ~ 9.7TF rdna1
0rHnf4l.jpg

This is a specific scene. It still runs native 4K with high textures and high shadows at a flawless 60fps across what, 98% of the game? And do you actually believe what these early games showcase are the full picture? This was also a heavy memory bandwidth test also as there are cases where series x outperformed the 2070 Super, which as you can see from this minimum fps test actually outperformed the 5700XT handily. You think an additional .22 TFLOPS alone are responsible for Series X in parts of this test beating a GPU 2070 Super when the minimum fps suggest Series X only hitting 89% of it? Things aren't quite as simple as you're trying to make them seem. You basically missed the entire point of this benchmark.

 
Yeah but Hitman 3 is also only 30fps on One X vs 60fps on Pro. What does that indicate?

The game has a bizzare setup, hanging ones hat on it as the definitive benchmark is flawed. And even then the performance delta is there, XsX is clearly aiming beyond it’s capability in parts.

It definitely isn't the definitive anything, but it so far represents the biggest gap between the two next gen consoles. Beyond its capabilities? That same section with the flowers could be fixed on Series X to run flawlessly, you know that, right? Will they bother? I don't know, but that's a data management thing. And excellent example with ps4 pro being 60fps. That's such a large gap that it's clearly an oversight. If the PS5 version of Hitman ended up 30fps I'd consider that an obvious oversight also because the Series X wouldn't have THAT big an edge.

But we are going off topic, so I'll end the series x/hitman talk since this isn't the thread
 
Last edited:
Top Bottom