Follow Slashdot blog updates by subscribing to our blog RSS feed

 



Forgot your password?
typodupeerror
×
PC Games (Games) Hardware

Benchmarking the Benchmarks 126

apoppin writes "HardOCP put video card benchmarking on trial and comes back with some pretty incredible verdicts. They show one video returning benchmark scores much better than another compared to what you get when you actually play the game. Lies, damn lies, and benchmarks."
This discussion has been archived. No new comments can be posted.

Benchmarking the Benchmarks

Comments Filter:
  • by majorme ( 515104 ) on Monday February 11, 2008 @11:49AM (#22379634) Journal
    damn i hate benchmarks
  • Re:OSS (Score:2, Insightful)

    by joaommp ( 685612 ) on Monday February 11, 2008 @11:55AM (#22379708) Homepage Journal
    aren't you being just a little bit... oh, I dunno... offtopic?

    Either I misunderstood you, or I don't see how the license can be a metric of performance or accuracy.
  • Benchmarks (Score:5, Insightful)

    by Anonymous Coward on Monday February 11, 2008 @12:13PM (#22379878)
    Duh, a benchmark is a controlled test performed "on a bench" - meaning, in a controlled environment with specific, well-described procedures.

    You must perform the same exact test on all video cards, disclose any variables, and you must not "pick a subset of completed tests to publish". You must not compare tests performed using different procedures, no matter how slight the deviation of the procedures are.

    One cannot draw conclusions about "real world" performance from a benchmark. The benchmark is merely an indicator. A "real world" test that uses the strong, formalized procedures of a benchmark IS a benchmark - and suddenly, the benchmark is not "real world" - because the "real world" doesn't have formal procedures for gameplay.

    Haphazard "non-blind" gameplay on a random machine is NOT a benchmark, and it can not provide useful, comparable numbers.

    A good benchmark is one where (1) most experts agree that it has validity, and (2) one where the tester cannot change the rules of the game.

    The numbers of a benchmark are meaningless, except in terms of being compared to one another using the same exact procedure.

  • by jonnythan ( 79727 ) on Monday February 11, 2008 @12:33PM (#22380080)
    Um, they come up with what is probably the most useful data of all:

    The highest playable settings for given hardware.

    They then change the video card and find the highest playable settings for that hardware.

    I'd much rather compare the highest playable settings for two different cards than the timedemo benchmark numbers for two different cards.
  • by jonnythan ( 79727 ) on Monday February 11, 2008 @12:35PM (#22380098)
    "As long as you don't run two 30 inch monitors, any name brand video card for about 200 bucks will give you great playable rates at 1680 x 1050."

    Not in Crysis, Call of Duty 4, UT3, etc.

    When I go to plunk down $200 - $300 on a video card, and one of them performs comfortably at my LCD's native resolution and the other one doesn't, that matters. Saying all cards in a given price range are roughly equivalent is saying that you are completely, 100% blind to the reality of video cards today.
  • Not the same card (Score:3, Insightful)

    by jandrese ( 485 ) <kensama@vt.edu> on Monday February 11, 2008 @12:47PM (#22380238) Homepage Journal
    One thing that's bothering me is that HardOCP said "Anandtech benchmarked this card vs. an 8800GTS and said it came out faster, then we benchmarked it against an 8800GTX and it game out faster, then people complained that our results didn't match". Isn't that expected? The GTX is a faster card than the GTS last time I looked. Why is it such a shock that the ATI card came in between them in performance?

    It is a bit of a shock that ATI's latest and greatest can't seem to consistently beat nVidia's over a year old GTX cards I guess.
  • by Dracolytch ( 714699 ) on Monday February 11, 2008 @12:58PM (#22380382) Homepage
    You know that's totally intractable, right?

    For example: 1620x1050 with no AA may be considered unplayable (jaggies) for some, but others it's perfectly fine...

    Or, maybe you can turn on the AA, but deactivate shadows, changing your whole "playable" demographic again.

    It's like asking someone to benchmark coffee at different resturants to grade whether it is palletable or not.

    ~D
  • by MobileTatsu-NJG ( 946591 ) on Monday February 11, 2008 @12:58PM (#22380384)

    It's not the benchmark-scores that count. Sure, you need a specific minimum to enjoy the game, but it's the actual gameplay that makes the game fun, no matter the hardware.

    I'm pretty sure these benchmarks are invented by men.
    These benchmark scores are important when trying to determine a balance of cost vs. performance. So yes, these benchmarks were invented by men. This is because the old standard of picking the one whose color matches their shoes also resulted with the invention of the credit card.
  • by donscarletti ( 569232 ) on Monday February 11, 2008 @01:11PM (#22380564)

    There is indeed a bare minimum hardware performance required to play but sadly many new games, especially Crysis, that bare minimum is scarily close to the market's maximum. Benchmarks are supposed to be a way to isolate this and objectively measure it so that a good purchasing decision can be made by the consumer and when the game is played hopefully the subjective experience of enjoyment will follow. A framerate above human perception is needed for fun (as jerky frames lead to nausia and frustration), high detail is needed for the beauty of a game which is probably just as important (it's been the basis for visual art, music and poetry for millennia).

    The reason we've got so far and now can have computers, electricity, aeroplanes, cars, etc. is because of the willingness of scientifically inclined individuals to isolate, experiment and measure. Technology is one of the things in life that can be measured and I think it is a good idea to continue to do it, provided we can do it right. Experimentation and science is what got us out of caves no?

    As for Hardocp, what have they proven? Apparently traditional time demos run a fairly linear amount faster than realtime demos, even though it has been acknowledged that realtime demos render more including weapons, characters and effects that the canned demo does not. This would be interesting if the question was "how fast can Crysis run on different cards" but that's not what people want to know. What I'd want to know is which card should I buy to allow me to continue to play cutting edge games for as long as possible while enjoying their whole beauty but not getting a framerate low enough to make me uncomfortable. It just so happens that the card with the best timedemo benchmark has the best actual playthrough benchmark and by roughly the same factor. The only difference is that the traditional timedemo depends on only the graphics hardware whereas the playthrough benchmark depends on efficiency elsewhere in the engine (AI physics), where the player spent most time and if reviewing subjectively, the reviewers current mindset and biases.

    Somebody please think of the science!

  • by cHiphead ( 17854 ) on Monday February 11, 2008 @02:11PM (#22381244)
    Some of us make purchasing decisions based on the piece of shit game we are thinking of buying. Crysis is a joke with such high requirements for a playable experience. I base my game purchases on what will run on my old pos single core p4 2.8ghz box. Any game that can't impress with such insanely fast hardware as we have these days even on the 'budget' boxes is not a game worth investing in.

    I must be getting old, I haven't upgraded my box in almost 2 years.

    Cheers.
  • by Sancho ( 17056 ) on Monday February 11, 2008 @03:39PM (#22382328) Homepage
    You're conflating benchmarking games vs. benchmarking graphics cards. If you're looking for raw power for an arbitrary amount of money, you'd want to get the graphics card which has the maximum frame rate at that price. If you're looking to play a specific game, you'd look for a graphics card which most people (quite subjectively, obviously) say plays the game well.

    The point is that you can't use a standard game (plus FPS meter) played by a human player to judge a graphics card's raw capabilities. To reduce subjectivity and error, you need a consistency in what is being rendered.

Beware of Programmers who carry screwdrivers. -- Leonard Brandwein

Working...