WinRAR 3.62
Many servers and workstations have to compress a lot of data. WinRAR is one of the most popular compression applications and now features a multi-threaded benchmark.
Compression algorithms work on large streams of data, so fast memory access is important. The WinRAR Benchmark has a rather high margin of error, but it is still is interesting to look at the scaling numbers.
The picture gets clearer as you compare the gains from extra cores in percentages.
The algorithm does scale somewhat but it is another example of how hard it is to scale well as more cores get added. NUMA architectures like AMD's Opteron have the potential to extract more memory performance, but there's still the problem of properly coding an algorithm to work with NUMA.
Many servers and workstations have to compress a lot of data. WinRAR is one of the most popular compression applications and now features a multi-threaded benchmark.
WinRAR 3.62 | ||
Multi | Single | |
Dual Xeon E5345 2.33 | 1501 | 522 |
Dual Opteron 2224 SE | 1259 | 529 |
Dual Xeon 5160 3.0 | 1236 | 549 |
Dual Opteron 2222 | 1219 | 471 |
Dual Opteron 8218HE 2.6 | 1172 | 426 |
Xeon E5345 2.33 | 1169 | 522 |
Xeon 5160 3 | 923 | 549 |
Compression algorithms work on large streams of data, so fast memory access is important. The WinRAR Benchmark has a rather high margin of error, but it is still is interesting to look at the scaling numbers.
WinRAR Scaling | ||||
Single | Dual | Quad | Octal | |
Xeon 5345 2.3 GHz | 522 | 901 | 1169 | 1501 |
Xeon 5160 3 GHz | 549 | 923 | 1236 | N/A |
Opteron 2224 SE 3.2 GHz | 529 | 957 | 1259 | N/A |
The picture gets clearer as you compare the gains from extra cores in percentages.
WinRAR Scaling - Percentages | |||
Dual vs. Single | Quad vs. Dual | Octal vs. Quad | |
Xeon 5345 2.3 GHz | 73% | 30% | 28% |
Xeon 5160 3 GHz | 68% | 34% | N/A |
Opteron 2224 SE 3.2 GHz | 81% | 32% | N/A |
The algorithm does scale somewhat but it is another example of how hard it is to scale well as more cores get added. NUMA architectures like AMD's Opteron have the potential to extract more memory performance, but there's still the problem of properly coding an algorithm to work with NUMA.
30 Comments
View All Comments
piroroadkill - Tuesday, August 7, 2007 - link
it is a car analogyGul Westfale - Monday, August 6, 2007 - link
good analogy there, except that mustangs (and various other cars) use pickup truck engines for cost reasons. large trucks use larger engines (often diesels) because they offer considerably more torque at much lower RPM than a smaller gasoline engine; and thus provide more pulling power.Gul Westfale - Monday, August 6, 2007 - link
these are not regular consumer cpus, but intended for use in commercial servers and workstations. they and their motherboards cost more because they support features such as multiple sockets (so in addition to having multiple cores on one chip you can also have multiple chips on one motherboard).yyrkoon - Monday, August 6, 2007 - link
they win 1 of 2 tests, and it is clear they are the winner ? Why ? Because they won the software rendering also ? Anyone interrested enough in rendering, and HAVING to have this sort of hardware for it is NOT going to bother with software . . .
This means your conclusion on this point is incorrect, and in which case, it boils down to which application the rendering machine is going to do.
Man you guys come to the wierdest conclusions based on your own data, and I am not even the first to notice/mention this sort of thing . . .
JohanAnandtech - Monday, August 6, 2007 - link
The Quadcore wins all high resolution rendering tests. Where do you see the DC opterons win against the Quadcore Intel in high resolution rendering? Show me a rendering engine where a 3 GHz K8 DC core is faster in high resolution renderering than a 2.33 GHz Quadcore. All decent and used in the realworld rendering engines will more or less show the same picture.In fact, the "rendering performance" situation will get worse for the K8 as SSE-2 tuning will get more common. All Intel CPUs since core and all AMD CPUs since Barcelona will show (or are already showing) high performance boost from using better SSE-2 code.
yyrkoon - Monday, August 6, 2007 - link
Ok, I see now with the graphs 'lower is better' on 3ds max, I missed that with the tables, which is actually what I meant this morning 'table obfustication'. I personally do not mind tables, but when the data is not in a uniform spot, it confuses/makes it harder to read at a glance.Anyhow, I was tired when I posted this morning, cranky, and was overly harsh I think. However it *is* much easier for me personaly to read the graphs at a glance (I cannot speak for everyone though).
yyrkoon - Monday, August 6, 2007 - link
Oh, and while on the subject, you guys here at anandtech have lately mastered the art of graph obfustication. Is it really THAT hard leaving items in the same rows / columns for different tests ? Are we trying to confuse the results, or is there some other reason this happens, and has gone completely over my head ?JohanAnandtech - Monday, August 6, 2007 - link
The only reason is that until very recently I didn't master the graphing engine. I got some weird error messages and gave up. But I have found the error, and you should see some nice graphs which don't obfusticate...Spoelie - Monday, August 6, 2007 - link
the gif on page 2 is non-looping, so after a very quick jump from 1ghz -> 2.8ghz (why??) -> 3.2ghz , it stays put on the 3.2ghz image. If reading the article, by the time the reader sees the image, it's already 5 minutes on the last image and staying there, making it for all intents and purposes a static image instead of an animated one:)
JohanAnandtech - Monday, August 6, 2007 - link
Thanks, fixed that. The reason to show 2.8 GHz is that for example Specjbb and other applications sometimes don't completely stress the CPU and then the cpu dynamically goes back to 2.8 GHz. It are simply the 3 stages I saw the most, and found the most interesting to show.