Half Life 2 GPU Roundup Part 1 - DirectX 9 Shootout
by Anand Lal Shimpi on November 17, 2004 11:22 AM EST- Posted in
- GPUs
The Slowest Level in the Game
For our fifth and final demo we turn to one of the last levels in the game – d3_c17_12. This city level takes place mostly outdoors and gave us the lowest average frame rates out of any level we played in during our testing of Half Life 2. A combination of all of the fire shaders as well as the explosions and weapon fire and the outdoor lighting make for one very stressful test.
Our player fires upon one of the mammoth mechanical spiders using a handful of weapons, including the RPG which in itself ends up being decently stressful on the GPU.
Although this demo yielded the lowest average frame rates out of our entire suite of demos, the level itself does not appear to be overly GPU bound. Instead it would seem that even at 1280 x 1024, we’re mostly CPU bound here which is definitely not what we expected.
79 Comments
View All Comments
Anand Lal Shimpi - Wednesday, November 17, 2004 - link
Thanks for all of the comments guys. Just so you know, I started on Part 2 the minute the first article was done. I'm hoping to be done with testing by sometime tomorrow and then I've just got to write the article. Here's a list of the new cards being tested:9600XT, 9550, 9700, X300, GF 6200, GF 5900XT, GF4 Ti 4600, GF4 MX440
I'm doing both DX9 and DX8 comparisons, including image quality.
After Part 2 I think I'll go ahead and do the CPU comparison, although I've been thinking about doing a more investigative type of article into Half Life 2 performance in trying to figure out where its performance limitations exist, so things may get shuffled around a bit.
We used the PCI Express 6600GT for our tests, but the AGP version should perform quite similarly.
The one issue I'm struggling with right now is the fact that the X700 XT is still not available in retail, while the X700 Pro (256MB) is. If I have the time I may go back and run some X700 Pro numbers to make this a more realistic present-day comparison.
Any other requests?
Take care,
Anand
Cybercat - Wednesday, November 17, 2004 - link
You guys made my day comparing the X700XT, 6800, and 6600GT together. One question though (and I apologize if this was mentioned in the article and I missed it), did you guys use the PCIe or AGP version of the 6600GT?Houdani - Wednesday, November 17, 2004 - link
18: Many users rely on hardware review sites to get a feel for what technology is worth upgrading and when.Most of us have financial contraints which preclude us from upgrading to the best hardware, therefore we are more interested in knowing how the mainstream hardware performs.
You are correct that it would not be an efficient use of resources to have AT repeat the tests on hardware that is two or three generations old ... but sampling the previous generation seems appropriate. Fortunately, that's where part 2 will come in handy.
I expect that part 2 will be sufficient in showing whether or not the previous generation's hardware will be a bottleneck. The results will be invaluable for helping me establish my minimum level of satisfaction for today's applications.
stelleg151 - Wednesday, November 17, 2004 - link
forget what i said in 34.....pio!pio! - Wednesday, November 17, 2004 - link
So how do you softmod a 6800NU to a 6800GT???or unlock the extra stuff....
stelleg151 - Wednesday, November 17, 2004 - link
What drivers were being used here, 4.12 + 67.02??Akira1224 - Wednesday, November 17, 2004 - link
Jedilol I should have seen that one coming!
nastyemu25 - Wednesday, November 17, 2004 - link
i bought a 9600XT because it came boxed with a free coupon for HL2. and now i can't even see how it matches up :(coldpower27 - Wednesday, November 17, 2004 - link
These benchmarks are more in line with what I was predicting, the x800 Pro should be equal to 6800 GT due to similar Pixel Shader fillrate while the X800 XT should have an advantage at higher resolutions due to it's having a higher fillrate being clocked higher.Unlike DriverATIheaven:P.
This is great I am happy knowing Nvidia's current generation of hardware is very competitive in performance in all aspects when at equal amounts of fillrate.
Da3dalus - Wednesday, November 17, 2004 - link
In the 67.02 Forceware driver there's a new option called "Negative LOD bias", if I understand what I've read correctly it's supposed to reduce shimmering.What was that option set to in the tests? And how did it affect performance, image quality and shimmering?