|May26-11, 03:40 AM||#1|
High-resolution gaming reduces the need for anti-aliasing?
I have a 1920x1080 monitor and a powerful enough rig to playably run any of my games on that resolution, even Crysis. I heard that AA (anti-aliasing) isn't necessary at high resolutions. And that adding AA at high-res is almost completely unnoticeable, and it only murders your framerate.
The only function of AA is to make jagged edges appear smoother. I believe high resolutions has this same effect, eliminating the need for AA. Although high-res AA might have some effect if you look very close. But otherwise it is mostly unnoticeable.
I play games on 1080p with 0x AA and I do not see any jagged edges. But when I reduce the resolution, the jagged edges become very obvious.
Anisotropic filtering (AF) is a different story.
So low resolution + AA or high resolution - AA?
Not to mention that turning off all AA (especially at high-res) significantly improves your framerates.
|May26-11, 03:40 PM||#2|
It appears to boil down to how sharp your eyesight is. If you don't notice the aliasing, you don't need the AA software.
|May27-11, 03:55 AM||#3|
|Similar Threads for: High-resolution gaming reduces the need for anti-aliasing?|
|High resolution CCD camera||Electrical Engineering||1|
|Anti-Aliasing Filter||Electrical Engineering||2|
|High resolution = tiny words and graphics||Computing & Technology||5|
|VERY high resolution Cassini photos?||General Astronomy||2|
|High-Resolution Brain Atlases in CyberSpace||Biology||0|