Compare any two graphics cards:
GeForce GTX 280 vs Radeon HD 5770
IntroThe GeForce GTX 280 comes with a GPU clock speed of 602 MHz, and the 1024 MB of GDDR3 RAM runs at 1107 MHz through a 512-bit bus. It also is made up of 240 SPUs, 80 TAUs, and 32 Raster Operation Units.
Compare all of that to the Radeon HD 5770, which comes with core clock speeds of 850 MHz on the GPU, and 1200 MHz on the 1024 MB of GDDR5 memory. It features 800(160x5) SPUs as well as 40 TAUs and 16 Rasterization Operator Units.
Power Usage and Theoretical Benchmarks
Power Consumption (Max TDP)
Performance-wise, the GeForce GTX 280 should theoretically be quite a bit superior to the Radeon HD 5770 overall. (explain)
Texel RateThe GeForce GTX 280 should be a lot (more or less 42%) more effective at texture filtering than the Radeon HD 5770. (explain)
Pixel RateIf running with high levels of AA is important to you, then the GeForce GTX 280 is a better choice, by a large margin. (explain)
Please note that the above 'benchmarks' are all just theoretical - the results were calculated based on the card's specifications, and real-world performance may (and probably will) vary at least a bit.
Please note that the price comparisons are based on search keywords - sometimes it might show cards with very similar names that are not exactly the same as the one chosen in the comparison. We do try to filter out the wrong results as best we can, though.
Memory Bandwidth: Bandwidth is the maximum amount of information (counted in MB per second) that can be moved past the external memory interface in a second. The number is worked out by multiplying the card's interface width by its memory speed. In the case of DDR type memory, it must be multiplied by 2 again. If it uses DDR5, multiply by 4 instead. The higher the memory bandwidth, the faster the card will be in general. It especially helps with AA, High Dynamic Range and high resolutions.
Texel Rate: Texel rate is the maximum number of texture map elements (texels) that are processed in one second. This number is calculated by multiplying the total amount of texture units of the card by the core clock speed of the chip. The better the texel rate, the better the video card will be at handling texture filtering (anisotropic filtering - AF). It is measured in millions of texels per second.
Pixel Rate: Pixel rate is the most pixels the video card can possibly write to the local memory per second - measured in millions of pixels per second. Pixel rate is worked out by multiplying the number of Raster Operations Pipelines by the the core speed of the card. ROPs (Raster Operations Pipelines - aka Render Output Units) are responsible for filling the screen with pixels (the image). The actual pixel rate is also dependant on quite a few other factors, especially the memory bandwidth - the lower the bandwidth is, the lower the ability to reach the max fill rate.