Why are you hoping for this thread ends? This threads should stay up and remind everyone that Digital Foundry failed their readership hard.This should be the thread ender, but unfortunately we all know that wont be the case.
Why are you hoping for this thread ends? This threads should stay up and remind everyone that Digital Foundry failed their readership hard.This should be the thread ender, but unfortunately we all know that wont be the case.
* That's being generous and giving the 133GB/s bandwidth number the benefit of the doubt.
Someone correct me if I'm wrong?
Doesn't matter if devs use some of the GPU CUs for compute on PS4. Yes that leaves 'less' for graphics but then those devs will also want to use a similar amount of compute on Xbox one, so if anything the relative differential for just graphics would be even more disparate
Eg 12 vs 18 just for graphics is a 50% improvement
Use 4CU equivalent for compute (yes I realise it would be shared across CUs but just for illustration) as suggested by some
8 vs 14 left for graphics is now a 75% improvement
You can spin the numbers with GTX 580 vs 680, its the same result, of course one is more powerful, but how this difference translates to games vs your last GPU being a Geforce 7800 GT is so insignificant given diminishing returns and developer talent, budget and time, it's just ridiculous.
if I have to wait 2 years I'll either get the hardware then for cheaper, or just start gaming on PC at that point which would be orders of magnitude more powerful and console specs won't even matter at that point.
you are still spouting out the same bullet points and yet you can't even own up to a lot of misleading information and straight bullshitting you did on this thread yesterday??Microsoft should be like Nintendo and not talk about Specs at all or be like Sony and put out in the open.
They issue random PRs and expect every one to simply say "oh great, I will blindly accept what you say"
- Cloud = 3x Xbox one power
- Cloud will offload physics
- Cloud = 300,000 Servers = Entire world compute power of 1999
- 5 Billion transistors
- $700M cloud data center, (300,000 Server network will be built in 2015, they just got the land lease)
- Titled Textures (even though it has been in OpenGL for more than a year)
- 192 Gbps eSRAM
They want to lure early adopters. But when they ask MS "what, what did you say"... then the answer from Xbox boss is... "Don't bother specs doesn't matter"
Then every single forum erupts. It is very simple. Either be ready to answer or don't talk about specs at all.
Just... Just stop lmfao.If you were truly being generous and really accounting "Theoretical" for both it would be 192GB/s not 133GB/s which is what's expected to be attainable, 176GB/s for the PS4 is purely theoretical, not what's attainable.
You can spin the numbers with GTX 580 vs 680, its the same result, of course one is more powerful, but how this difference translates to games vs your last GPU being a Geforce 7800 GT is so insignificant given diminishing returns and developer talent, budget and time, it's just ridiculous.
If Forza 5 can't compete with Drive Club, then I'll stand corrected, and don't argue that later down the road, blah blah blah, I'm buying a console at launch and I expect the upgrade at that time, if I have to wait 2 years I'll either get the hardware then for cheaper, or just start gaming on PC at that point which would be orders of magnitude more powerful and console specs won't even matter at that point.
People LOOOVE using percentages because it REALLY blows up the difference.
Check this out!!
GTX 680 is 200% more powerful than GTX 580 OMG WTF BBQ! is that better? My point still stands, when coming from a Geforce 7800GT, going to a GTX 580 or 680 is still a massive upgrade and both players should be happy with the results and should both be enjoying current Gen PC status with all the bells and whistles.
PS4 can do compute with out impacting graphics. Someone actually posted a quote from Cerny explaining it a few pages back. The 4 compute CUs is a pre PS4 reveal rumor.
If you were truly being generous and really accounting "Theoretical" for both it would be 192GB/s not 133GB/s which is what's expected to be attainable, 176GB/s for the PS4 is purely theoretical, not what's attainable.
You can spin the numbers with GTX 580 vs 680, its the same result, of course one is more powerful, but how this difference translates to games vs your last GPU being a Geforce 7800 GT is so insignificant given diminishing returns and developer talent, budget and time, it's just ridiculous.
If Forza 5 can't compete with Drive Club, then I'll stand corrected, and don't argue that later down the road, blah blah blah, I'm buying a console at launch and I expect the upgrade at that time, if I have to wait 2 years I'll either get the hardware then for cheaper, or just start gaming on PC at that point which would be orders of magnitude more powerful and console specs won't even matter at that point.
Microsoft should be like Nintendo and not talk about Specs at all or be like Sony and put out in the open.
It seems like you do not understand math or you'd realize that I simply explained why the example you used is grade-A fanboy "I am making stuff up as I go" bullshit...
Assuming that on PS4 a dev uses 14 CU for graphic and 4 for computing, when the same game will be ported to the Xbone it means you will only have 8 CUs for graphic. On top of that, the Xbone reserves 2 entire CPU cores for the OS, while the PS4 may reserve less than one core due to a much lighter OS. Thus, I think that multiplatform games will look better on PS4.
Assuming that on PS4 a dev uses 14 CU for graphic and 4 for computing, when the same game will be ported to the Xbone it means you will only have 8 CUs for graphic. On top of that, the Xbone reserves 2 entire CPU cores for the OS, while the PS4 may reserve less than one core due to a much lighter OS. Thus, I think that multiplatform games will look better on PS4.
GTX 580 = 1.56 TF
GTX 680 = 3.1 TF
Excuse my math if that doesn't make the 680 roughly 2x more or 200% more powerful on paper. My point is that games will not look the same order of magnitude better, it's just relative, but compare how those 2 cards DEMOLISH games running on a GeForce 7800GT which is basically what we're getting with the next gen console upgrade, then the difference is not that big of a deal, both cards are running what everyone considers "Next gen" visuals.
This is an even greater example but even still I think someone with a GTX 580 will do just fine.
Actually I truly believe that the difference between Xbox One and PS4 is being greatly overstated around here.
When looking at specs on paper, one could say that the GTX 680 is a lot more powerful than the GTX 670 and they would be right, lets look at the numbers:
GTX 670 = 2.46 Teraflops
GTX 680= 3.1 TF
Difference of 640MF. Difference between Xbox One and PS4 = 600MF
Depending on perspective, one person could say: Wow that is a significant difference, but to me it's not, in terms of what that will translate to what I see and play, it won't be that significant, especially if I'm upgrading from a Geforce 7800GT, whether I go with a 670 or 680, I'm still getting a substantial upgrade, and whether the 680 is a lot more powerful on paper, when coming for a 7800GT, going to 680 instead of 670 for an additional 600 Mflops is really not significant, the person with the 670 should be equally excited and will be playing with the same visual fidelity minus a little AA and other settings tweaks to keep the framerate on par, but nothing to make a considerable difference visually.
Yes, as long as we are assuming the rest of the system is equally potent. And yes I believe that multiplats will generally look better on PS4 - the question is if anyone (besides people with very good eyes or DF) will give a f*ck.. Which might pose the other question why so many people are arguing about the specs so much - let's wait and see what happens.
Totally agree. I just dont agree that the gap will be as big as certain people here think.
This is an even greater example but even still I think someone with a GTX 580 will do just fine.
Totally agree. I just dont agree that the gap will be as big as certain people here think.
They don't release the specs because they know the console is a lot weaker
Lol?There we go again. Any source (apart from numbers / graphs yada yada yada) on this one?
Something we can all agree on. The whole thread summed up in 3 sentences.Xbone has a path to its ESRAN that is faster than the PS4 unifed solution. Which I thought was always true but apparently not. PS4s memory setup is still better by a larger margin.
There we go again. Any source (apart from numbers / graphs yada yada yada) on this one?
I haven't followed this thread much, but I watched Cerny's Road to Playstation presentation and in it he talks about them considering a RAM solution similar to the 360's using eSRAM, with him admitting that despite it being exotic, in the long run you'd get much better bandwidth results. Could this end up being the case with the 360? I know they went with much slower GDDR3 RAM as a solution so that will impede from an exact comparison.
It's the new argument you make for the Xbox One. Albert Penello (Marketing and Leads planning for Xbox One) also used that in a recent post.Lol?
Cerny talks about a small pool of eDRAM with a theoretical bandwidth of 1000GB/s in addition to to GDDR5 (albeit with a lower buswidth) for a total of 1088 memory bandwidth (only theoretically possible for that small pool of eDRAM.)I haven't followed this thread much, but I watched Cerny's Road to Playstation presentation and in it he talks about them considering a RAM solution similar to the 360's using eSRAM, with him admitting that despite it being exotic, in the long run you'd get much better bandwidth results. Could this end up being the case with the 360? I know they went with much slower GDDR3 RAM as a solution so that will impede from an exact comparison.
There we go again. Any source (apart from numbers / graphs yada yada yada) on this one?
The leaks on both the PS4 and Xbox One has been nearly accurate so far, so I trust what we have. Plus Microsoft has been amusingly quiet about the specs, and hasn't even tried to deny it.
Overall, I'll give the rumors a 90% chance of being true.
sustain a peak theoretical throughput of 102GB/s, Microsoft has revised its own figures upwards significantly, telling developers that 192GB/s is now theoretically possible.
the original bandwidth claim derives from a pretty basic calculation - 128 bytes per block multiplied by the GPU speed of 800MHz offers up the previous max throughput of 102.4GB/s. It's believed that this calculation remains true for separate read/write operations from and to the ESRAM. Microsoft techs have found that the hardware is capable of reading and writing simultaneously. Apparently, there are spare processing cycle "holes" that can be utilised for additional operations.
Ok then just for you:
Xbone has 109% of the memory bandwidth, 50% of the ROP performance, and 67% of the shader performance of PS4.
Memory banwidth:
Xbone: 192GB/s
PS4: 176GB/s
ROP's:
Xbone: 16 ROP's
PS4: 32 ROP's
Shader units:
Xbone: 768
PS4: 1152
Even theoretically Xbone still looks to be lacking quite considerably?
The leaks on both the PS4 and Xbox One has been nearly accurate so far, so I trust what we have. Plus Microsoft has been amusingly quiet about the specs, and hasn't even tried to deny it.
Overall, I'll give the rumors a 90% chance of being true.
They dont have to, once the Xbone is out the capable folks will figure out how much weaker the specs are.Somehow I fear that Microsoft will never publish specs and this crap will go on for the rest of next generation.
it is possible that given the specs revealed by Sony, MS is scrambling to squeeze out more performance from the hardware before spilling it out?
is it possible that they will overclock the gpu and CPU to make up some of the difference? Given their larger hardware they could go for a stronger cooling solution to cope with the extra heat from an OC.
it is possible that given the specs revealed by Sony, MS is scrambling to squeeze out more performance from the hardware before spilling it out?
is it possible that they will overclock the gpu and CPU to make up some of the difference? Given their larger hardware they could go for a stronger cooling solution to cope with the extra heat from an OC.
that is speculation on my part. However I am curious to hear others take on it.
I would personally be against such a route this late in the dev cycle, but if it is feasible from a hardware perspective I don't see why they couldn't explore it.
There we go again. Any source (apart from numbers / graphs yada yada yada) on this one?
Ask yourself this question; did people give a fuck this gen about minute differences? How many meltdowns did we have when DF face offs constantly favored 360 versions of multiplats?Yes, as long as we are assuming the rest of the system is equally potent. And yes I believe that multiplats will generally look better on PS4 - the question is if anyone (besides people with very good eyes or DF) will give a f*ck.. Which might pose the other question why so many people are arguing about the specs so much - let's wait and see what happens.
They dont have to, once the Xbone is out the capable folks will figure out how much weaker the specs are.
The problem is that this is framed with the bit about this being a discovery by Microsoft.The more I look at this and read what others are saying, the more I think the answer is right in front of us. The eSRAM bus cannot suddenly read and write at the same time, it isn't possible and they are not claiming it, look at the way they say "from and to ESRAM".
it is possible that given the specs revealed by Sony, MS is scrambling to squeeze out more performance from the hardware before spilling it out?
is it possible that they will overclock the gpu and CPU to make up some of the difference? Given their larger hardware they could go for a stronger cooling solution to cope with the extra heat from an OC.
GAF will give a fuck. Non GAF people will care for the price. Both perspectives favour PS4. Well, until MS announces a price cut which I expect now.Yes, as long as we are assuming the rest of the system is equally potent. And yes I believe that multiplats will generally look better on PS4 - the question is if anyone (besides people with very good eyes or DF) will give a f*ck.. Which might pose the other question why so many people are arguing about the specs so much - let's wait and see what happens.
In that case MS is only prolonging the inevitable.Somehow I fear that Microsoft will never publish specs and this crap will go on for the rest of next generation.
Not necessarily. For instance, we still don't know much about the Wii U's inner workings. There is always room enough for debates like in this thread to happen. I also don't think that multi-platform games will change much. People will either downplay the technical differences (You can't see a difference between 720p and 1080p), or argue that developers are to lazy to leverage the full potential of ESRAM (if the Xbox version is inferior), 6 additional compute units (if the PS4 version is not better), or whatever, while the other camp will, of course, argue for the opposite case.
Not necessarily. For instance, we still don't know much about the Wii U's inner workings. There is always room enough for debates like in this thread to happen. I also don't think that multi-platform games will change much. People will either downplay the technical differences (You can't see a difference between 720p and 1080p), or argue that developers are to lazy to leverage the full potential of ESRAM (if the Xbox version is inferior), 6 additional compute units (if the PS4 version is not better), or whatever, while the other camp will, of course, argue for the opposite case.
Show your source for these, or stop spewing bullshit.The slide didn't imply there was a hardware division-there isn't. It was simply a suggested division by Sony, for the reasons in my last post.
To be clear, there are 18 CU's. Sony suggests you use 14 for graphics and 4 for compute. Their research showed more than 14 CU's for graphics helped little (which means the system is bottlenecked somewhere else,
The problem is that this is framed with the bit about this being a discovery by Microsoft.
A way old vgleaks leak already spoke about having higher theoretical bandwidth than 102GB/s.
It just means reading from eSRAM and writing to DRAM and vice versa.
To give an example: Reading from DRAM is at 68GB/s and writing to eSRAM would also be a 68GB/s as it's limited to by how fast it can copy.
In total that is 136GB/s bandwidth. But it's all fancy accounting and has been known since the first memory system leak.
GAF will give a fuck. Non GAF people will care for the price. Both perspectives favour PS4. Well, until MS announces a price cut which I expect now.