ZEN2 8c/16t with NAVI GPU Spotted - AMD Console chip?

Discussion in 'Frontpage news' started by Hilbert Hagedoorn, Jul 24, 2019.

  1. Hilbert Hagedoorn

    Hilbert Hagedoorn Don Vito Corleone Staff Member

    Messages:
    39,555
    Likes Received:
    8,222
    GPU:
    AMD | NVIDIA
  2. slyphnier

    slyphnier Master Guru

    Messages:
    771
    Likes Received:
    60
    GPU:
    GTX1070
    sony silent or not, it going to be similar SoC to Xbox anyway, probably some slight change/difference but it still same amd-soc-based
    or is there any potential alternative ?
     
  3. Backstabak

    Backstabak Master Guru

    Messages:
    671
    Likes Received:
    267
    GPU:
    Gigabyte Rx 5700xt
    I think there is basically no alternative on the console maker, as Intel still has no GPU or APU to talk about and I doubt that anyone wants to work with NVIDIA after first xbox. So it's very likely that new consoles will have very similar, or even exactly the same, HW. And again, as in current generation, the biggest difference will be available games and exclusives.
     
    waltc3 and MonstroMart like this.
  4. AlbertX

    AlbertX Member

    Messages:
    20
    Likes Received:
    6
    GPU:
    2 x GTX 1070
    After the first Xbox, Ps3 had an Nvidia GPU and now Nintendo does.
     

  5. Evildead666

    Evildead666 Maha Guru

    Messages:
    1,302
    Likes Received:
    273
    GPU:
    Vega64/EKWB/Noctua
    Yeah, and after working with Nvidia on the PS3 Sony didn't want to work with them again.

    There wasn't really anything else available for the switch at the time.
     
  6. fantaskarsef

    fantaskarsef Ancient Guru

    Messages:
    11,884
    Likes Received:
    3,908
    GPU:
    2080Ti @h2o
    So we're seeing more medium clocked cores (this time with HT) and a mid range GPU? Why am I not impressed...
     
    airbud7 likes this.
  7. Undying

    Undying Ancient Guru

    Messages:
    14,218
    Likes Received:
    3,391
    GPU:
    Aorus RX580 XTR 8GB
    It makes 8c/16t standard. This just changes my mind of getting an R5 3600 and looking at R7 3700 instead.
     
  8. Picolete

    Picolete Master Guru

    Messages:
    318
    Likes Received:
    88
    GPU:
    R9 290 Sapphire Tri-x
    TDP cant be to high on a console
     
    airbud7 and vbetts like this.
  9. DeskStar

    DeskStar Maha Guru

    Messages:
    1,068
    Likes Received:
    151
    GPU:
    EVGA 2080Ti FTW3 HC
    Looking pretty good from where I'm sitting. To make an eight core sixteen thread CPU a standard for consoles I'm all about it. Can only help out PC gaming to most extent....that is hardware adoption and proper usage of said hardware in practice.

    Still awaiting the scalping price of GPU's from a particular team to come down in pricing.
     
  10. vbetts

    vbetts Don Vincenzo Staff Member

    Messages:
    15,073
    Likes Received:
    1,637
    GPU:
    GTX 1080 Ti
    Considering the move from Jaguar and a GCN/Polaris based platform performance definitely will be a lot better for the console market.
     
    airbud7 likes this.

  11. schmidtbag

    schmidtbag Ancient Guru

    Messages:
    5,493
    Likes Received:
    2,032
    GPU:
    HIS R9 290
    Consoles never had impressive specs. I'd say this is actually a pretty substantial upgrade over last gen.
    Remember:
    The performance you get in console hardware does not correlate with PC hardware. Build a PC with the same specs as the original XB1 and I'm sure you'll struggle to sustain 1080p@30FPS.

    Don't forget - these are low-power cores. In multithreaded tasks, a single 3600 core will most likely outperform 2 cores for these consoles. I'm sure they're going to reserve a core or 2 for the OS like they did with last-gen. Considering how much more people are demanding of their consoles (like streaming or recording), It's reasonable they'll still reserve 2 cores. So, unless you expect to do that stuff yourself, you can most likely get by with the 3600 and not see any problems.

    Besides, look at all of the current console games ported to PC. It's an 8-core CPU, but practically none of these games will bog down something like a 8600K. In most cases, an overclocked 7600K is sufficient. Remember: we're often playing games with at least twice the framerate.

    If you don't want to take any risks and have money to spare, the 3700 is probably the smarter choice, but I'm willing to bet the 3600 will be sufficient for anyone who wants to play next-gen games and doesn't intend to run intensive background processes (like streaming). Alternatively, if all you care about is making sure you have enough cores to match consoles, you could get the 2700. Those are a steal, and they offer more than enough performance.
     
    airbud7, Venix and Undying like this.
  12. Denial

    Denial Ancient Guru

    Messages:
    13,160
    Likes Received:
    2,650
    GPU:
    EVGA RTX 3080
    I'm actually not sure if this is true. I've been playing around with power limiting/scaling on my 3900 based on a suggestion by @Fox2232 and the scaling on this processor is extremely weird. I can go down to like 70w capped and only lose like 10% sustained performance. It doesn't make any sense. I think the TDP on the console is obviously going to be low and balanced towards the GPU but it wouldn't surprise of the CPU performance is within 10-15% of a full 3700 or 3600 whatever the 8 core is.
     
    Fox2232 likes this.
  13. schmidtbag

    schmidtbag Ancient Guru

    Messages:
    5,493
    Likes Received:
    2,032
    GPU:
    HIS R9 290
    I'm not sure I understand what you're saying. What are you doing to cap off the wattage? The IPC should remain pretty constant regardless of what your frequency is, but, just because the wattage is lower, that doesn't mean the frequency isn't affected (doesn't mean it is, either). It's worth pointing out that the Ryzen 3000 series seems to sort of have a mind of its own, where for example if you lower the voltage too much, it automatically underclocks (but it might not report that to Task Manager). Gamers Nexus has an article on that behavior that goes into more depth.
    All that being said, when you account for the actual current frequency, a 3.6GHz core of the exact same architecture should be approximately twice as fast as a 1.8GHz core. In the case of comparing these consoles to a PC, there will be some deviation due to stuff like other IRQ'ing devices, RAM speed, background processes, and cache size, but, that's why I said "approximately".
     
  14. vbetts

    vbetts Don Vincenzo Staff Member

    Messages:
    15,073
    Likes Received:
    1,637
    GPU:
    GTX 1080 Ti
    Lower wattage is usually obtained by lowering the clocked rate, that or limiting boosts by clocks or what cores can boost.
     
    airbud7 likes this.
  15. schmidtbag

    schmidtbag Ancient Guru

    Messages:
    5,493
    Likes Received:
    2,032
    GPU:
    HIS R9 290
    Right, but that's why I find it a bit confusing. If you lower the clock rate then of course the performance will drop. Denial didn't specify what the frequency was. Ryzen from the very beginning seems to exponentially increase in wattage as you get closer and closer to the limit of the chip (which so far has typically been somewhere between 4.0-4.4GHz, depending on silicon lottery).
    So - without much other information, it's not much surprise that you could chop off 30% of the wattage for a 10% performance loss.

    That being said, the part I find confusing is whether or not he's saying the IPC is being affected. So hypothetically, let's say there was a 30% drop in frequency, but benchmarks only showed a 10% loss in performance, that would be very weird.
     
    airbud7 likes this.

  16. Denial

    Denial Ancient Guru

    Messages:
    13,160
    Likes Received:
    2,650
    GPU:
    EVGA RTX 3080
    Nah not the IPC - the clocks definitely automatically drop but like you said the power is exponential. I just don't think this chip is ever going to need to run at it's base frequency of 1.8 (picture actually says 1.6 not sure where the 1.8 is coming from in the article). I use PPT to control the power and the 10% performance loss comes from actual benchmarks not just looking at frequency - but the frequencies basically drop 10%.
     
    airbud7 likes this.
  17. schmidtbag

    schmidtbag Ancient Guru

    Messages:
    5,493
    Likes Received:
    2,032
    GPU:
    HIS R9 290
    Well yeah, I wasn't really questioning power consumption though. All I'm saying is core-per core and per-clock, the 3600 is roughly twice as fast as what the next-gen consoles will be. Even if you're right that the consoles won't typically run at base clocks, the same goes for a PC user running a Zen2 CPU as well. To your point though, it will be much more power-hungry. On the other hand, the 3700 Undying was considering isn't going to be a whole lot better in that regard.
    Just a side note:
    Since 1.8GHz is the minimum spec, it's worth pointing out that a [good] game developer is basically obligated to assume that's the frequency they have to work with. Since it is very possible the system will clock that low, they must ensure that the game will run smoothly at that speed. That being said, 1.8GHz, in most cases, will be the fastest we should expect of that CPU when all cores are in use. For games that don't need all available cores, that's where the boost clocks come in handy, without compromising TDP.
    Yup, definitely. Haha I've been in this sort of dichotomy with my CPU, where I have some headroom to overclock by another 150MHz. But the added heat (and therefore fan noise) and power consumption just isn't worth it since I'm already so close to the ceiling. On the other hand, I really don't care about preserving this CPU since I bought it knowing it was on the verge of obsolescence, so part of me is just like "let's just crank up until the crank breaks".
    Meanwhile, I could actually try to preserve this CPU and undervolt+underclock just a little bit, where it could make for a decent home server.
     
    airbud7 likes this.
  18. fantaskarsef

    fantaskarsef Ancient Guru

    Messages:
    11,884
    Likes Received:
    3,908
    GPU:
    2080Ti @h2o
    While you are absolutely right of course, it also makes it predictable. So that's why I'm simply not surprised.


    Sure... 5 years later it's easy to do that as well.


    Which is not saying the performance will be bad, too low, or whatever you think. It's just... a substantial upgrade, like you said @schmidtbag , but this is no revolution. Which still holds my humble opinion true, I'm simply not impressed. Whatever you want to interpret is not my saying. :)
     
    airbud7 likes this.
  19. schmidtbag

    schmidtbag Ancient Guru

    Messages:
    5,493
    Likes Received:
    2,032
    GPU:
    HIS R9 290
    But that's my point - no console's hardware was ever a revolution, so why should this be any different? I suppose you could say the original GameBoy was revolutionary, since it was the first truly portable gaming system that was actually decent, but people were more marveling that it was portable, rather than its level of performance. After all, it was just black and white.
    EDIT:
    Same could be said about the first Atari.
     
    airbud7 likes this.
  20. fantaskarsef

    fantaskarsef Ancient Guru

    Messages:
    11,884
    Likes Received:
    3,908
    GPU:
    2080Ti @h2o
    Yes I do remember my GameBoy. Should get it out of the basement ;)

    But yeah, when we're both saying the same, as expected / not impressed, why do you find any need to answer to my post? I already said you are right in the exact quote you put into your's, my fellow guru
     
    airbud7 likes this.

Share This Page