Compare any two graphics cards:
GeForce GTX Titan vs Radeon HD 5870
IntroThe GeForce GTX Titan features a GPU clock speed of 837 MHz, and the 6144 MB of GDDR5 memory runs at 1502 MHz through a 384-bit bus. It also is made up of 2688 Stream Processors, 224 Texture Address Units, and 48 Raster Operation Units.
Compare all of that to the Radeon HD 5870, which makes use of a 40 nm design. AMD has set the core frequency at 850 MHz. The GDDR5 memory runs at a speed of 1200 MHz on this specific card. It features 1600(320x5) SPUs as well as 80 Texture Address Units and 32 ROPs.
Power Usage and Theoretical Benchmarks
Power Consumption (Max TDP)
The GeForce GTX Titan should theoretically perform quite a bit faster than the Radeon HD 5870 overall. (explain)
Texel RateThe GeForce GTX Titan should be a lot (approximately 176%) better at AF than the Radeon HD 5870. (explain)
Pixel RateThe GeForce GTX Titan is much (more or less 48%) more effective at full screen anti-aliasing than the Radeon HD 5870, and also should be able to handle higher screen resolutions better. (explain)
Please note that the above 'benchmarks' are all just theoretical - the results were calculated based on the card's specifications, and real-world performance may (and probably will) vary at least a bit.
Please note that the price comparisons are based on search keywords - sometimes it might show cards with very similar names that are not exactly the same as the one chosen in the comparison. We do try to filter out the wrong results as best we can, though.
Memory Bandwidth: Memory bandwidth is the max amount of information (counted in megabytes per second) that can be moved past the external memory interface within a second. It's worked out by multiplying the card's interface width by its memory clock speed. If it uses DDR RAM, the result should be multiplied by 2 once again. If it uses DDR5, multiply by ANOTHER 2x. The better the bandwidth is, the better the card will be in general. It especially helps with anti-aliasing, HDR and higher screen resolutions.
Texel Rate: Texel rate is the maximum texture map elements (texels) that are applied per second. This number is worked out by multiplying the total amount of texture units by the core speed of the chip. The higher the texel rate, the better the video card will be at handling texture filtering (anisotropic filtering - AF). It is measured in millions of texels applied in one second.
Pixel Rate: Pixel rate is the most pixels the graphics card could possibly write to the local memory in one second - measured in millions of pixels per second. Pixel rate is worked out by multiplying the number of Render Output Units by the the core clock speed. ROPs (Raster Operations Pipelines - also called Render Output Units) are responsible for filling the screen with pixels (the image). The actual pixel output rate also depends on quite a few other factors, most notably the memory bandwidth - the lower the bandwidth is, the lower the ability to reach the maximum fill rate.