• We’re currently investigating an issue related to the forum theme and styling that is impacting page layout and visual formatting. The problem has been identified, and we are actively working on a resolution. There is no impact to user data or functionality, this is strictly a front-end display issue. We’ll post an update once the fix has been deployed. Thanks for your patience while we get this sorted.

Why GeForce 10 series are using x16 PCIe?

alphajoza

Junior Member
GeForce 10 series are employing PCIe 3.0 x16 and since PCIe lanes are limited in CPUs and motherboards, users are forced to use 2 x8 configuration if they want to use two GPUs; some wonders what limitation (especially in bandwidth) the GPU will encounter when forcing it to use x8 instead of x16; but tests and benchmarks shows that there's almost no difference in performance when going from x16 to x8!

My questions is that how is this possible? and if the GPU can have the same performance even in x8 mode, then why advertise it as PCIe 3.0 x16?
 
GeForce 10 series are employing PCIe 3.0 x16 and since PCIe lanes are limited in CPUs and motherboards, users are forced to use 2 x8 configuration if they want to use two GPUs; some wonders what limitation (especially in bandwidth) the GPU will encounter when forcing it to use x8 instead of x16; but tests and benchmarks shows that there's almost no difference in performance when going from x16 to x8!

My questions is that how is this possible? and if the GPU can have the same performance even in x8 mode, then why advertise it as PCIe 3.0 x16?
because the card's interface is a PCIe 3.0 x16 slot. the fact that PCIe can self configure the amount of lanes and GPUs are perfectly happy to work in reduced numbers of lanes doesn't change that.
 
Plus, it's marketing. It's been around since... well, forever. AGP at least.

"Well, our new card is AGP4X, it's up to twice as fast as AGP2X!" - When in reality, even back then AGP4X was a minor difference from 2X.

The story repeated at AGP8X vs 4X
And again for first gen PCIE vs AGP
And so on and so forth.

The issue isn't that the faster interface speeds aren't useful- they are. The issue is that a major reason the faster interface speeds were originally invented was to allow DMA (direct memory access) for the graphics card to talk to main system memory. At the time, video cards had relatively tiny amounts of memory and the idea of simply 'sharing' the main memory on the system sounded super amazing.

The problem was, even back then onboard memory on graphics cards was way faster and lower latency than system memory; any card that actually resorted to using system RAM performed *horribly*. This can even be seen today, when you try and play a game on a card with too little VRAM and you configure the settings too high, you get *horrible* hitching and hiccups in the framerate. Congratulations, you've found the spots where your GPU borrows system RAM to make up for its lack of VRAM!

And though system RAM has gotten much faster, and the interface has gotten faster, it's not mattered because of two factors:
1. GPU memory has gotten much faster as well
2. GPUs come with lots *more* memory now, so why would they need to borrow system RAM?

So today, of course they say "Well, our card is PCIE 3.0 X16!" Because if they didn't, someone else would say "Our card is X16, theirs is only X8- obviously ours is faster!"

Though, not everyone does this. The AMD RX460 cards are only X8.
 
Please allow Sanic to reply on my behalf:

hqdefault.jpg


It's really just that simple. Faster = better
 
Plus, it's marketing. It's been around since... well, forever. AGP at least.

"Well, our new card is AGP4X, it's up to twice as fast as AGP2X!" - When in reality, even back then AGP4X was a minor difference from 2X.

The story repeated at AGP8X vs 4X
And again for first gen PCIE vs AGP
And so on and so forth.

The issue isn't that the faster interface speeds aren't useful- they are. The issue is that a major reason the faster interface speeds were originally invented was to allow DMA (direct memory access) for the graphics card to talk to main system memory. At the time, video cards had relatively tiny amounts of memory and the idea of simply 'sharing' the main memory on the system sounded super amazing.

The problem was, even back then onboard memory on graphics cards was way faster and lower latency than system memory; any card that actually resorted to using system RAM performed *horribly*. This can even be seen today, when you try and play a game on a card with too little VRAM and you configure the settings too high, you get *horrible* hitching and hiccups in the framerate. Congratulations, you've found the spots where your GPU borrows system RAM to make up for its lack of VRAM!

And though system RAM has gotten much faster, and the interface has gotten faster, it's not mattered because of two factors:
1. GPU memory has gotten much faster as well
2. GPUs come with lots *more* memory now, so why would they need to borrow system RAM?

So today, of course they say "Well, our card is PCIE 3.0 X16!" Because if they didn't, someone else would say "Our card is X16, theirs is only X8- obviously ours is faster!"

Though, not everyone does this. The AMD RX460 cards are only X8.

In absolute sense, it is the case that faster speeds reduce latency when sending data with serial data transfers like pci-e. And there is indeed a situation where other factors are slowing the system down more than the pcie latency as can be seen with x8 and x16. But indeed, now there is so much vram that almost everything is preloaded into vram before the level can be played. So, the largest sets of data like textures are already present in vram and compressed too. And a list of drawcalls send to the gpu is much smaller. And there is of course tricks like compression to send data over the pci-e to the gpu and let the gpu decompress it. But that would only be useful when doing it at the loading of the level and with static data. I do wonder if compressing and decompressing can now be done in a fashion that it is also useful with a draw call. Is compressing and decompressing less time consuming than the current pci-e latency ? Is that possible ?
 
Back
Top