Albert Pennello Takes To NeoGAF To Defend The Performance Of Xbox One, Again!

/ 4 years ago


Albert Pennello really knows how to brave a storm, it seems the poor guy can’t say a single thing without it being twisted and turned into something negative by people online, of course it is really his own fault since the he seems intent on correcting people who are mostly Sony fanboys, trolls and generally people who really don’t care, eg, non Xbox consumers.

There is no doubt that the Xbox One is significantly more powerful than the Xbox 360, but there appears to be some speculation on the PlayStation 4 being significantly more powerful than the Xbox One, Pennello disagrees and he’s brought along his fact book to prove it.

Here is what he had to say on NeoGaf, in full.

I see my statements the other day caused more of a stir than I had intended. I saw threads locking down as fast as they pop up, so I apologize for the delayed response.

I was hoping my comments would lead the discussion to be more about the games (and the fact that games on both systems look great) as a sign of my point about performance, but unfortunately I saw more discussion of my credibility.

So I thought I would add more detail to what I said the other day, that perhaps people can debate those individual merits instead of making personal attacks. This should hopefully dismiss the notion I’m simply creating FUD or spin.

I do want to be super clear: I’m not disparaging Sony. I’m not trying to diminish them, or their launch or what they have said. But I do need to draw comparisons since I am trying to explain that the way people are calculating the differences between the two machines isn’t completely accurate. I think I’ve been upfront I have nothing but respect for those guys, but I’m not a fan of the mis-information about our performance.

So, here are couple of points about some of the individual parts for people to consider:

• 18 CU’s vs. 12 CU’s =/= 50% more performance. Multi-core processors have inherent inefficiency with more CU’s, so it’s simply incorrect to say 50% more GPU.
• Adding to that, each of our CU’s is running 6% faster. It’s not simply a 6% clock speed increase overall.
• We have more memory bandwidth. 176gb/sec is peak on paper for GDDR5. Our peak on paper is 272gb/sec. (68gb/sec DDR3 + 204gb/sec on ESRAM). ESRAM can do read/write cycles simultaneously so I see this number mis-quoted.
• We have at least 10% more CPU. Not only a faster processor, but a better audio chip also offloading CPU cycles.
• We understand GPGPU and its importance very well. Microsoft invented Direct Compute, and have been using GPGPU in a shipping product since 2010 – it’s called Kinect.
• Speaking of GPGPU – we have 3X the coherent bandwidth for GPGPU at 30gb/sec which significantly improves our ability for the CPU to efficiently read data generated by the GPU.

Hopefully with some of those more specific points people will understand where we have reduced bottlenecks in the system. I’m sure this will get debated endlessly but at least you can see I’m backing up my points.

I still I believe that we get little credit for the fact that, as a SW company, the people designing our system are some of the smartest graphics engineers around – they understand how to architect and balance a system for graphics performance. Each company has their strengths, and I feel that our strength is overlooked when evaluating both boxes.

Given this continued belief of a significant gap, we’re working with our most senior graphics and silicon engineers to get into more depth on this topic. They will be more credible then I am, and can talk in detail about some of the benchmarking we’ve done and how we balanced our system.

Thanks again for letting my participate. Hope this gives people more background on my claims.

He does bring up some interesting points but I fear he’s preaching to a bunch of people who just don’t want to hear it anyway. When it really boils down to it, the PlayStation 4 is most likely more powerful than the Xbox One, but not by as much as you would expect given the specifications, there really are too many variables and as with any hardware review we perform here at eTeknix, everything from the CPU cooler to the motherboard, to the drivers and OS can drastically effect performance, so only time will tell who really packs a bigger performance punch.

Thank you NeoGaf for providing us with this information.

Topics: , , , ,


7 Responses to “Albert Pennello Takes To NeoGAF To Defend The Performance Of Xbox One, Again!”
  1. Toby Kirkby says:

    whats that sound???…. sounds like someone grasping at straws

    • Brant says:

      Sounds like an explanation and response to all of the ‘facts’ that are just rumors created in echo chambers by people who think they know more about a machine they had no part in making.

  2. Reality says:

    can we stop comparing dicks and just play the games now?

  3. TerryxMasters says:

    Except for that part where Mark Cerny debunked him on damn near everything he just mentioned ages ago.

  4. Etnos says:

    Interesting points, seems like reductive logic and ignorance of how actually computer electronics works keep ruling fanboy discussions. Hopefully both systems do well and we all can play some good games.

  5. tipoo2 says:

    “Adding to that, each of our CU’s is running 6% faster. It’s not simply a 6% clock speed increase overall.”

    Right, it’s less than that. Less CUs in total, so the 6% increase as compared to the PS4s CUs means the total increase is even lower. It’s not cumulative. If you have two processors with four cores and equivalent performance at 3GHz, and then you increase the clock speed of one by 100MHz, you can’t say that’s 100MHz times four, as the number of cores is comparable. Going even further, if the slower clocked processor had eight cores, that effective increase for the higher clocked one is halved relative to the eight core one.

    He tried to spin a 6% clock increase as something that’s cumulative times the number of CUs. That’s silly.

Speak Your Mind

Tell us what you're thinking...
and oh, if you want a pic to show with your comment, go get a gravatar!

Optimized with PageSpeed Ninja