AMD's Radeon HD 5870: Bringing About the Next Generation Of GPUs
by Ryan Smith on September 23, 2009 9:00 AM EST- Posted in
- GPUs
Eyefinity
Somewhere around 2006 - 2007 ATI was working on the overall specifications for what would eventually turn into the RV870 GPU. These GPUs are designed by combining the views of ATI's engineers with the demands of the developers, end-users and OEMs. In the case of Eyefinity, the initial demand came directly from the OEMs.
ATI was working on the mobile version of its RV870 architecture and realized that it had a number of DisplayPort (DP) outputs at the request of OEMs. The OEMs wanted up to six DP outputs from the GPU, but with only two active at a time. The six came from two for internal panel use (if an OEM wanted to do a dual-monitor notebook, which has happened since), two for external outputs (one DP and one DVI/VGA/HDMI for example), and two for passing through to a docking station. Again, only two had to be active at once so the GPU only had six sets of DP lanes but the display engines to drive two simultaneously.
ATI looked at the effort required to enable all six outputs at the same time and made it so, thus the RV870 GPU can output to a maximum of six displays at the same time. Not all cards support this as you first need to have the requisite number of display outputs on the card itself. The standard Radeon HD 5870 can drive three outputs simultaneously: any combination of the DVI and HDMI ports for up to 2 monitors, and a DisplayPort output independent of DVI/HDMI. Later this year you'll see a version of the card with six mini-DisplayPort outputs for driving six monitors.
It's not just hardware, there's a software component as well. The Radeon HD 5000 series driver allows you to combine all of these display outputs into one single large surface, visible to Windows and your games as a single display with tremendous resolution.
I set up a group of three Dell 24" displays (U2410s). This isn't exactly what Eyefinity was designed for since each display costs $600, but the point is that you could group three $200 1920 x 1080 panels together and potentially have a more immersive gaming experience (for less money) than a single 30" panel.
For our Eyefinity tests I chose to use every single type of output on the card, that's one DVI, one HDMI and one DisplayPort:
With all three outputs connected, Windows defaults to cloning the display across all monitors. Going into ATI's Catalyst Control Center lets you configure your Eyefinity groups:
With three displays connected I could create a single 1x3 or 3x1 arrangement of displays. I also had the ability to rotate the displays first so they were in portrait mode.
You can create smaller groups, although the ability to do so disappeared after I created my first Eyefinity setup (even after deleting it and trying to recreate it). Once you've selected the type of Eyefinity display you'd like to create, the driver will make a guess as to the arrangement of your panels.
If it guessed correctly, just click Yes and you're good to go. Otherwise ATI has a handy way of determining the location of your monitors:
With the software side taken care of, you now have a Single Large Surface as ATI likes to call it. The display appears as one contiguous panel with a ridiculous resolution to the OS and all applications/games:
Three 24" panels in a row give us 5760 x 1200
The screenshot above should clue you into the first problem with an Eyefinity setup: aspect ratio. While the Windows desktop simply expands to provide you with more screen real estate, some games may not increase how much you can see - they may just stretch the viewport to fill all of the horizontal resolution. The resolution is correctly listed in Batman Arkham Asylum, but the aspect ratio is not (5760:1200 !~ 16:9). In these situations my Eyefinity setup made me feel downright sick; the weird stretching of characters as they moved towards the outer edges of my vision left me feeling ill.
Dispite Oblivion's support for ultra wide aspect ratio gaming, by default the game stretches to occupy all horizontal resolution
Other games have their own quirks. Resident Evil 5 correctly identified the resolution but appeared to maintain a 16:9 aspect ratio without stretching. In other words, while my display was only 1200 pixels high, the game rendered as if it were 3240 pixels high and only fit what it could onto my screens. This resulted in unusable menus and a game that wasn't actually playable once you got into it.
Games with pre-rendered cutscenes generally don't mesh well with Eyefinity either. In fact, anything that's not rendered on the fly tends to only occupy the middle portion of the screens. Game menus are a perfect example of this:
There are other issues with Eyefinity that go beyond just properly taking advantage of the resolution. While the three-monitor setup pictured above is great for games, it's not ideal in Windows. You'd want your main screen to be the one in the center, however since it's a single large display your start menu would actually appear on the leftmost panel. The same applies to games that have a HUD located in the lower left or lower right corners of the display. In Oblivion your health, magic and endurance bars all appear in the lower left, which in the case above means that the far left corner of the left panel is where you have to look for your vitals. Given that each panel is nearly two feet wide, that's a pretty far distance to look.
The biggest issue that everyone worried about was bezel thickness hurting the experience. To be honest, bezel thickness was only an issue for me when I oriented the monitors in portrait mode. Sitting close to an array of wide enough panels, the bezel thickness isn't that big of a deal. Which brings me to the next point: immersion.
The game that sold me on Eyefinity was actually one that I don't play: World of Warcraft. The game handled the ultra wide resolution perfectly, it didn't stretch any content, it just expanded my viewport. With the left and right displays tilted inwards slightly, WoW was more immersive. It's not so much that I could see what was going on around me, but that whenever I moved forward I I had the game world in more of my peripheral vision than I usually do. Running through a field felt more like running through a field, since there was more field in my vision. It's the only example where I actually felt like this was the first step towards the holy grail of creating the Holodeck. The effect was pretty impressive, although costly given that I only really attained it in a single game.
Before using Eyefinity for myself I thought I would hate the bezel thickness of the Dell U2410 monitors and I felt that the experience wouldn't be any more engaging. I was wrong on both counts, but I was also wrong to assume that all games would just work perfectly. Out of the four that I tried, only WoW worked flawlessly - the rest either had issues rendering at the unusually wide resolution or simply stretched the content and didn't give me as much additional viewspace to really make the feature useful. Will this all change given that in six months ATI's entire graphics lineup will support three displays? I'd say that's more than likely. The last company to attempt something similar was Matrox and it unfortunately didn't have the graphics horsepower to back it up.
The Radeon HD 5870 itself is fast enough to render many games at 5760 x 1200 even at full detail settings. I managed 48 fps in World of Warcraft and a staggering 66 fps in Batman Arkham Asylum without AA enabled. It's absolutely playable.
327 Comments
View All Comments
Wreckage - Wednesday, September 23, 2009 - link
Hot, loud, huge power draw and it barely beats a 285.A disappointment for sure.
SiliconDoc - Thursday, September 24, 2009 - link
Thank you Wreckage, now, I was going to say draw up your shields, but it's too late, the attackers have already had at it.--
Thanks for saying what everyone was thinking. You are now "a hated fanboy", "a paid shill" for the "corporate greedy monster rip off machine", according to the real fanboy club, the ones who can't tell the truth, no matter what, and prefer their fantasiacal spins and lies.
Zstream - Wednesday, September 23, 2009 - link
They still allow you to post?yacoub - Wednesday, September 23, 2009 - link
He's right in the first sentence but went all fanboy in the second.Griswold - Wednesday, September 23, 2009 - link
Not really, he's a throughbred fanboy with everything he said. Even on the "loud" claim compared to what previous reference designs vom ATI were like...SiliconDoc - Wednesday, September 30, 2009 - link
So if YOU compare one loud design of ati's fan to another fan and as loud ati card( they're all quieter than 5870* but we'll make believe for you for now),
and they're both loud, anyone complaining about one of them being loud is "an nvidia fanboy" because he isn't aware of the other loud as heck ati cards, which of course, make another loud one "just great" and "not loud". LOL
It's just amazing, and if it was NV:
" This bleepity bleep fan and card are like a leaf blower again, just like the last brute force monster core power hog but this **tard is a hurricane with no eye."
But since it's the red cards that are loud, as YOU pointed out in the plural, not singular like the commenter, according to you HE's the FANBOY, because he doesn't like it. lol
ULTIMATE CONCLUSION: The commenter just told the truth, he was hoping for more, but was disappointed. YOU, the raging red, jumped his case, and pointed out the ati cards are loud "vom" prior.. and so he has no right to be a big green whining fanboy...
ROFLMAO
I bet he's a "racist against reds" every time he validly criticizes their cards, too.
---
the 5870 is THE LOUDEST ATI CARD ON THE CHART,AND THE LOUDEST SINGLE CORE CARD.
--
Next, the clucking rooster will whiplash around and flap the stubby wings at me, claiming at idle it only draws 27 watts and is therefore quiet.
As usual, the sane would them mention it will be nice not playing any 3d games with a 3d gaming card, and enjoying the whispery hush.
--
In any case:
Congratulations, you've just won the simpleton's red rooster raving rager thread contest medal and sarcastic unity award.(It's as real as any points you've made)
Anyhow thanks, you made me notice THE 5870 IS THE LOUDEST CARD ON THE CHARTS. I was too busy pointing out the dozen plus other major fibboes to notice.
It's the loudest ati card, ever.
GourdFreeMan - Wednesday, September 23, 2009 - link
I thought the technical portion of your review was well written. It is clear, concise and written to the level of understanding of your target audience. However, I am less than impressed with your choice of benchmarks. Why is everything run at 4xAA, 16xAF? Speaking for most PC gamers, I would have maxed the settings in Crysis Warhead before adding AA and AF. Also, why so many console ports? Neither I, nor anyone else I personally know have much interest in console ports (excluding RPGs from Bethesda). Where is Stalker: Clear Sky? As you note its sequel will be out soon. Given the short amount of time they had to work with DX11, I imagine it will run similarly to Stalker: Call of Pripyat. Also, where is ArmA II? Other than Crysis and Stalker it is the game most likely to be constrained by the GPU.I don't want to sound conspiratorial, but your choice of games and AA/AF settings closely mirror AMD's leaked marketing material. It is good that you put their claims to the test, as I trust Anandtech as an unbiased review site, but I don't think the games you covered properly cover the interests of PC gamers.
Ryan Smith - Wednesday, September 23, 2009 - link
For the settings we use, we generally try to use the highest settings possible. That's why everything except Crysis is at max quality and 4xAA 16xAF (bear in mind that AF is practically free these days). Crysis is the exception because of its terrible performance; Enthusiast level shaders aren't too expensive and improve the quality more than anything else, without driving performance right off a cliff. As far as playing the game goes, we would rather have AA than the rest of the Enthusiast features.As for our choice of games, I will note that due to IDF and chasing down these crazy AA bugs, we didn't get to run everything we wanted to. GRID and Wolfenstein (our OpenGL title) didn't make the cut. As for Stalker, we've had issues in the past getting repeatable results, so it's not a very reliable benchmark. It also takes quite a bit of time to run, and I would have had to drop (at least) 2 games to run it.
Overall our game selection is based upon several factors. We want to use good games, we want to use popular games so that the results are relevant for the most people, we want to use games that give reliable results, and ideally we want to use games that we can benchmark in a reasonable period of time (which means usually having playback/benchmark tools). We can't cover every last game, so we try to get what we can using the criteria above.
GourdFreeMan - Thursday, September 24, 2009 - link
Popularity and quality are strong arguments for World of Warcraft, Left 4 Dead, Crysis, Far Cry, the newly released Batman game... and *maybe* Resident Evil (though it is has far greater popularity among console gamers). However, HAWX? Battleforge? I would never have even heard of these games had I not looked them up on Wikipedia. In retrospect I can see you using Battleforge due to it being the only DirectX 11 title, but I still don't find your list of games compelling or comprehensive.To me *PC* gaming needs to offer something more than simple action to justify its cost of entry. In the past this included open worlds, multiplayer, greater graphical realism and attempts at balancing realistic simulation with entertaining game play. Console gaming has since offered the first two, but the latter are still lacking.
It's games like Crysis, Stalker and ArmA II along with the potential of modding that attract people to PC gaming in the first place...
dvijaydev46 - Wednesday, September 23, 2009 - link
Good review but it would be good if you could also add Steam and Cuda benchmarks. Now you have a common software Mediashow Espresso right?