Results 1 to 10 of 10
  1. #1
    Spur-taaaa TDMVPDPOY's Avatar
    My Team
    San Antonio Spurs
    Join Date
    Feb 2005
    Post Count
    41,330
    lmao amd

    here are the voltra specs http://wccftech.com/sk-hynix-gddr6-n...-gpu-gtc-2017/

    moving to ddr6
    while amd vega10 currently is still on ddr5, and vega11 will be hbm2 memory below 1080 performance....

  2. #2
    Take the fcking keys away baseline bum's Avatar
    My Team
    San Antonio Spurs
    Join Date
    Mar 2003
    Post Count
    93,154
    Holy the Volta V100 chip has a die size of 815mm^2.

  3. #3
    Veteran Wild Cobra's Avatar
    My Team
    Portland Trailblazers
    Join Date
    May 2007
    Post Count
    43,117
    Holy the Volta V100 chip has a die size of 815mm^2.
    Over 1 sq inch.

    I've seen larger chips, but that was when they were still using 0.25 micron technology.

  4. #4
    Spur-taaaa TDMVPDPOY's Avatar
    My Team
    San Antonio Spurs
    Join Date
    Feb 2005
    Post Count
    41,330
    if the chips continue to get smaller, why not make it a dual card since they are going away from sli?

  5. #5
    Veteran Wild Cobra's Avatar
    My Team
    Portland Trailblazers
    Join Date
    May 2007
    Post Count
    43,117
    if the chips continue to get smaller, why not make it a dual card since they are going away from sli?
    That one chip consumes as much as 300 watts. Smaller is not easy to do. Do you realize what is involved in 12 micron production?

  6. #6
    Veteran illusioNtEk's Avatar
    My Team
    San Antonio Spurs
    Join Date
    Oct 2012
    Post Count
    1,942
    Can't wait for next gen laptops... finally going to have the same power as desktops have

  7. #7
    Take the fcking keys away baseline bum's Avatar
    My Team
    San Antonio Spurs
    Join Date
    Mar 2003
    Post Count
    93,154
    if the chips continue to get smaller, why not make it a dual card since they are going away from sli?
    Dual gpu cards still use SLI man.

  8. #8
    Veteran Wild Cobra's Avatar
    My Team
    Portland Trailblazers
    Join Date
    May 2007
    Post Count
    43,117
    Dual gpu cards still use SLI man.
    Yes, but on-board would be far less susceptible to signal to noise ratios.

  9. #9
    🏆🏆🏆🏆🏆 ElNono's Avatar
    My Team
    San Antonio Spurs
    Join Date
    Apr 2007
    Post Count
    152,607
    The biggest issue with SLI is that the biggest performance gain is with AFR (Alternate Frame Rendering, where one card draws even frames and the other odd frames) and it suffers from micro-stuttering, which is inherent to multi-GPU (especially dual) configurations.

    The reason is that one of the cards always has to transmit the full frame buffer over to the master for display post-rendering. Latest SLI bridges can do up to 650 MHz, which is good enough for up to 5K resolutions, but the transfers are unavoidable, and because frame rendering time is not constant or exact, a small delay can really cause partial frame renders, thus the stutter.

    G-Sync or FreeSync would correct for that, but then you have dynamic refresh rates...

    Now if you put them on the same card, you could optimize by using dedicated memory per framebuffer, but then you run into cache coherency issues, etc. Multi-CPU configurations are always difficult when you have to eventually do serial work (one frame at a time).

  10. #10
    Spur-taaaa TDMVPDPOY's Avatar
    My Team
    San Antonio Spurs
    Join Date
    Feb 2005
    Post Count
    41,330
    was watching the NVidia computex 17 conference...

    regarding volta and whatever....seems like NVidia wants to challenge intel or whoever is making cpus, since the volta gpu is just as powerful as a cpu, why the need for a cpu from a compe or when NVidia could make their own cpu = gpu volta...I wonder if they jump to cpu, would they give intel/amd a run for their money with cpu compe ion? the way the ceo was talkin about volta, looked like he wanted volta to be mainstream....

    hence was watching the NVidia laptop segment how they have a gaming laptop 18mm with a 1080gtx comparing it to ps4pro makes no fkn sense....how old is a ps4pro vs this new laptop, hence that laptop can always be upgraded, while the ps4pro u cant upgrade the core component like cpu/gpu...they should've compared it to other gaming laptops manufactures

Thread Information

Users Browsing this Thread

There are currently 1 users browsing this thread. (0 members and 1 guests)

Posting Permissions

  • You may not post new threads
  • You may not post replies
  • You may not post attachments
  • You may not edit your posts
  •