1. This site uses cookies. By continuing to use this site, you are agreeing to our use of cookies. Learn More.

AMD announces Radeon VII (7nm)

Discussion in 'Frontpage news' started by Hilbert Hagedoorn, Jan 9, 2019.

  1. Noisiv

    Noisiv Ancient Guru

    Messages:
    6,655
    Likes Received:
    496
    GPU:
    2070 Super
  2. HWgeek

    HWgeek Master Guru

    Messages:
    439
    Likes Received:
    314
    GPU:
    Gigabyte 6200 Turbo Fotce @500/600 8x1p
    fantaskarsef likes this.
  3. fantaskarsef

    fantaskarsef Ancient Guru

    Messages:
    10,776
    Likes Received:
    2,937
    GPU:
    2080Ti @h2o
  4. HWgeek

    HWgeek Master Guru

    Messages:
    439
    Likes Received:
    314
    GPU:
    Gigabyte 6200 Turbo Fotce @500/600 8x1p
    Interesting! News:

    From 1 Side(Game Dev's) we know:
    https://www.hardocp.com/article/2019/01/14/amd_radeon_vii_interview_scott_herkelman/2

    From other side(AMD) we now hear this:
    https://www.overclock3d.net/news/gp..._supports_directml_-_an_alternative_to_dlss/1

    I think AMD got some more trick in their sleeves for 2019 :)

    P/S - why do I have a feeling the DLSS going to be another "G-Sync VS Free-Sync?"
    NV mage Gamers to pay for it before MS released DirectML for Free?

    P.S2: I have read on Anandtech that DirectML will Use FP16 for better performance
    https://www.anandtech.com/show/12549/microsoft-windows-machine-learning-for-gaming

    look about this chart:
    ALL pascal Based GPU's offer 1/64 of FP32 performance:
    1080TI- 177.2 GFLOPS (1:64)!!
    while Vega offers 2X (25,166 GFLOPS (2:1)) and Polaris offers 1:1!(6,589 GFLOPS (1:1))
    [​IMG]

    [​IMG]
     
    Last edited: Jan 16, 2019
    BlackZero, Valken and Maddness like this.

  5. Stormyandcold

    Stormyandcold Ancient Guru

    Messages:
    5,201
    Likes Received:
    138
    GPU:
    MSI GTX1070 GamingX
    HWgeek, AMD already uses their FP16 in games like Far Cry 5 and Wolfenstein 2. It's no surprise that they'd extend the usefulness of FP16.

    However, when we look at those games, then it's clear that GTX1080ti still trounces the Vega 64.

    Finally, DirectML works across the board, so, it's not going to be a unique technology for AMD. Where as Nvidia now has the option of both their DLSS implementation and DirectML.

    No tricks up the sleeve if the competition already has it...
     
  6. alanm

    alanm Ancient Guru

    Messages:
    8,898
    Likes Received:
    1,273
    GPU:
    Asus 2080 Dual OC
  7. Lavcat

    Lavcat Master Guru

    Messages:
    435
    Likes Received:
    14
    GPU:
    AMD Radeon VII
    Not sure if this is old news -- however I just noticed on AMD's Radeon VII product page the disclaimer: "GPU specifications and features may vary by OEM configuration." I don't remember that the wording was there before.

    This implies the OEM cards could be different from the reference design.
     
  8. Fox2232

    Fox2232 Ancient Guru

    Messages:
    9,739
    Likes Received:
    2,199
    GPU:
    5700XT+AW@240Hz
    As I wrote before, 8GB variant of Radeon 7 could be $150 cheaper. And at that price, even I would seriously consider it. 8GB of VRAM is plenty more than I need for my 1080p gaming.
     
  9. alanm

    alanm Ancient Guru

    Messages:
    8,898
    Likes Received:
    1,273
    GPU:
    Asus 2080 Dual OC
    The way I understand it is that its not so easy to do that without a redesign and refitting of a new interposer, which would make it just as costly as leaving it with 16gb. 8gb HBM2 would make sense on an entirely new design of a card rather than redesigning and reconfiguring an existing 16gb for 8gb. Other types of vram would not be a problem.
     
  10. Fox2232

    Fox2232 Ancient Guru

    Messages:
    9,739
    Likes Received:
    2,199
    GPU:
    5700XT+AW@240Hz
    It would still be 1024pins per HBM2 chip and same bandwidth. I did post slide from one of HBM2 manufacturers showing 2GB per chip having exactly same bandwidth and pin count as 4GB chips.

    It should be noted that HBM uses 1 die which is sort of controller and above it you have multiple dies which are actual memory.
     

  11. alanm

    alanm Ancient Guru

    Messages:
    8,898
    Likes Received:
    1,273
    GPU:
    Asus 2080 Dual OC
    Well, why havent they offered an 8gb version for the R VII? If they did, @ $150 cheaper, would be a killer card for the money.
     
    anticupidon likes this.
  12. anticupidon

    anticupidon Ancient Guru

    Messages:
    3,972
    Likes Received:
    746
    GPU:
    integrated
    Sounds appealing. Where i can get it? :D
     
  13. nevcairiel

    nevcairiel Master Guru

    Messages:
    601
    Likes Received:
    190
    GPU:
    MSI 1080 Gaming X
    NVIDIA has been pushing into machine learning (ML) hard. Calling that a trick up their sleeves for AMD is not very accurate.
    On top of 2x FP16 performance on Turing, NVIDIA also has the Tensor Cores, which can do a lot of ML work "for free", while on AMD it costs shader performance. More Machine Learning tasks in mainstream software/games only plays to NVIDIAs advantages in Turing, if anything.
     
  14. Fox2232

    Fox2232 Ancient Guru

    Messages:
    9,739
    Likes Received:
    2,199
    GPU:
    5700XT+AW@240Hz
    We all seen the marketing. But nobody did show actual reproducible benchmark. So, all those Tensor cores, by how many TFLOPs it boosts 2080Ti FP16 (27TFLOPs). As actual total FP16 performance of card 30TFLOPs? I do not think so, otherwise it would be marketed as such.

    Or in reverse. If you take general workload for ML and run it on 27TFLOPs GPU and then run it via Tensor Cores. How much faster it will be?
    [​IMG]
    Titan V has 30TFLOPs of FP16. And 640 Tensor Cores. They account for additional throughput which would be equal to 10.7 TFLOPs of FP16.

    RTX 2060 has 240 Tensor Cores which would be around 4 TFLOPs of FP16 used for ML. Totaling FP16 ML of card as 16.9 TFLOPs.
    RTX 2070 has 288 Tensor Cores which would be around 4.8 TFLOPs of FP16 used for ML. Totaling FP16 ML of card as 19.7 TFLOPs.
    RTX 2080 has 368 Tensor Cores which would be around 6.1 TFLOPs of FP16 used for ML. Totaling FP16 ML of card as 26.2 TFLOPs.
    RTX 2080Ti has 544 Tensor Cores which would be around 9 TFLOPs of FP16 used for ML. Totaling FP16 ML of card as 35.9 TFLOPs.

    Those are not exactly crazy high values in comparison to Vega 64 with 25 TFLOPs of FP16. And Radeon 7 having 27 TFLOPs of FP16.
    = = = =

    Now as for actual statement that running something on Tensor cores comes "for free". It kind of does, but how much would it cost AMD?
    RTX 2080 which is comparable in price to Radeon 7 has 6.1 TFLOPs of FP16 "for free" outside shaders. For Radeon 7 it means, it will have 20.9TFLOPs left after doing same ML workload.
    (Sacrifice of 22.6% of shader time/performance.)
     
    Last edited: Jan 27, 2019
    INSTG8R likes this.
  15. Embra

    Embra Master Guru

    Messages:
    898
    Likes Received:
    200
    GPU:
    Vega 64 Nitro+LE
    Could they not also do a 12gb card as well? 12gb would have even to cover needs longer, for a bit more than 8gb.
     

  16. Maddness

    Maddness Master Guru

    Messages:
    948
    Likes Received:
    214
    GPU:
    EVGA RTX 2080Ti FTW
    I think the issue is no-one is making 2 or 3gb stacks of HBM2 memory. As far as i know it's 4gb only. I have no idea of the costs involved, but it might end up being to costly to change and that is why we are only getting 16gb HBM2 at this stage.
     

Share This Page