Message boards :
Number crunching :
Congratulations. We're the 171st fastest supercomputer
Message board moderation
Previous · 1 · 2
Author | Message |
---|---|
tbret Send message Joined: 28 May 99 Posts: 3380 Credit: 296,162,071 RAC: 40 |
Let's pick at this a little bit. The Linpack benchmark used for estimating the Top 500's "real" power is pretty specific. How would any machine we might own fair running it? But, of course, the more basic question is, "Does Linpack represent a benchmark that has relevance to what we do?" I read that Linpack is a double-precision benchmark. We don't use double-precision. Is it possible that Eric's number is correct using Linpack as it's yardstick, but that in single-precision calculations we would double, triple, or more our performance? (considering the huge difference in single and double precision times on an NVIDIA GPU) Is it possible that the difference in AP and MB task "credits" we see are real based on what Linpack would say IF Linpack were running? It would be very interesting to know how Eric derived his number. My question is really this: Is it possible that the type of calculations being done at other projects really are better suited for the hardware we run and they really are getting multiple times the FLOPS we get? Take for instance GPUGRID -> I've long considered those "credits" to be fictional. BUT... is that my failure? When I ran GPUGRID I found it almost impossible to keep my cards cool. They were consuming a lot of power, producing a lot of heat, and racking-up a lot of credits. If my card shows a 95% GPU busy-state at Einstein, and a 95% busy-state at GPUGRID, but the card is 20C hotter at GPUGRID, additional electrons must be swirling around in there somewhere. Additional work is being done. More FLOPS per unit "busy-ness" of the card, right? ...or wrong? EDIT - Maybe what's called-for is to look at identical computers across the projects and see what their RAC is. The highest scoring machines at Einstein are all using AMD cards; notsomuch here. The difference is partially due to which hardware is better-suited to the tasks. Einstein's still running CUDA 3.2. If I ran my hardware here on CUDA 3.2 my TFLOPS would be much reduced. I'd really like to know how Eric estimates our collective capacity. |
Richard Haselgrove Send message Joined: 4 Jul 99 Posts: 14653 Credit: 200,643,578 RAC: 874 |
I'd really like to know how Eric estimates our collective capacity. (I'm not an expert on benchmarking, so I'll leave that part) If you haven't already, have a read of CreditNew, to remind yourself how we got into this situation. The bit which is relevant to your question is: The second credit system Although this project is now running the Third (i.e. New) Credit system, the mechanics of the second system are still in place. Looking at the most recent task I reported from this computer, I see that stderr_txt contains Flopcounter: 38994473207145.844000 What's more, the actual report back to the project (I'm looking at the sched_request file) also contains <fpops_cumulative>111104300000000.000000</fpops_cumulative> A second task contains Flopcounter: 17692153721950.531000 <fpops_cumulative>50314560000000.000000</fpops_cumulative> Current workunit headers still contain <credit_rate>2.8499999</credit_rate>, aka the old "Credit Multiplier: 2.85" that old-timers like me will remember - and that's very close to the ratio of 'counted' and 'cumulative' floating point operations. So, all the components are there for a true estimate of FPOPs to be saved in the SETI database, and - since, obviously, we record time too - we can calculate true FLOPs directly. My guess (and it can only be a guess, until Eric gets back from the conference) is that Eric has access to the raw FLOPs data and had a quick look at it while preparing his speech. |
kittyman Send message Joined: 9 Jul 00 Posts: 51468 Credit: 1,018,363,574 RAC: 1,004 |
Richard........ You have to perpetuate the credit screwed scenario again? I thought it was calming down. I truly don't give the back end of a rat about it anymore. I just don't. In the true meaning of the project, it is simply just a non-issue. Not that I don't luv my creds. But slicing and dicing about why they could not be higher or better? Not my issue anymore. I do not care. Every other user here is issued credits on the same level and basis as I am. So it's all good, as far as I am concerned. I know, Let's have a spelling contest. That would work. "Freedom is just Chaos, with better lighting." Alan Dean Foster |
tbret Send message Joined: 28 May 99 Posts: 3380 Credit: 296,162,071 RAC: 40 |
I hadn't put that two and two together, Richard. Thanks. I suspect you are right, then. I also suspect you are correct about the other projects. Maybe Eric can tell us if that's how he got his number. I'd like to know. Still, it does leave the heat issue unresolved. Maybe Jason will drop-by to tell us if heat is an indication of FLOPing or if heat is a result of some sort of hardware electrical effect caused by fewer memory accesses or something. (Mark - this isn't about credits, it's about calculating FLOPSs. It's possible to talk about this without whining about the credit numbers. We just did.) So, if our total RAC is ________, and our actual speed is 237 TFLOPS, I should be able to calculate my own speed within a reasonable margin of error for such things. Or did I miss something in your message? I've read it three times, but that doesn't mean I couldn't have misunderstood. |
Raistmer Send message Joined: 16 Jun 01 Posts: 6325 Credit: 106,370,077 RAC: 121 |
Bigger temp most probably indicates fewer arithmetic unit stalls on data awaiting. FLOP conseption doesn't reflect the need of algorithm in memory. Some algorithms more memory-demanding, some not. Lets consider 2 simplified samples. One computes Y[i]=X[i]+C Another - Y[i]=C*sin(X[i]) Both need to read constant, read one variable and save(write) one result per array element, but amount of calculations will be quite different. First algorithm will be constrained by speed of memory access on all modern hardware CPU or GPU. Second one has some chances to hide at least part of memory access time by sine computations. SETI apps news We're not gonna fight them. We're gonna transcend them. |
tbret Send message Joined: 28 May 99 Posts: 3380 Credit: 296,162,071 RAC: 40 |
So, if I understand the implication of what you are saying, then the hotter card IS doing more FLOPS because it waits less per second. If that is true, then there is a real difference in the FLOPS performance of the hardware and the hotter project is making more complete use of the resources. Said differently; a real count of the floating point operations WILL be higher per second. That means that if "credit" were granted on actual FLOPS, the code that produces the fewest waits will get the largest "credit." Interesting... |
Richard Haselgrove Send message Joined: 4 Jul 99 Posts: 14653 Credit: 200,643,578 RAC: 874 |
First algorithm will be constrained by speed of memory access on all modern hardware CPU or GPU. Second one has some chances to hide at least part of memory access time by sine computations. Yes, that's a valid way of looking at it: one of the tricks of optimisation is to make sure that the information you're going to need in a moment is exactly where you're going to need it, with the fastest possible route from there to the computational engine that's going to process it. The same applies to CPUs too: it's nothing special for GPUs (although the effects may be more dramatic, because of the longer route the data has to travel over the PCI bus from main memory). It's one of the reasons why the old AK_v8 applications for CPU (now withdrawn) were so good: because Alex Kan put a lot of time and cleverness into getting the memory access right for that particular generation of CPUs. Or so I'm told, anyway. |
Raistmer Send message Joined: 16 Jun 01 Posts: 6325 Credit: 106,370,077 RAC: 121 |
Slightly wrong. Cause AKv8 not withdrown first of all. It's still our base for MB7 opt CPU apps. Only ICC/IPP version abandoned for now. Hand-made optimizations still in work. Or replaced by even better ones. SETI apps news We're not gonna fight them. We're gonna transcend them. |
tbret Send message Joined: 28 May 99 Posts: 3380 Credit: 296,162,071 RAC: 40 |
Just a general comment, then: I'm more enthusiastic about NVIDIA's Maxwell line and the rumored inclusion of a Tegra processor on the card. IF I understand what I think I understand about what Jason said in an old thread somewhere, that will allow a programmer to put the entirety of the data in VRAM on the card, then program the Tegra to feed the GPU's "processors" when needed. Suddenly even antiquated PCIe bus speeds wouldn't affect total processing time very much. We ought to be able to screeeeeeam through some data if we can keep the cards cool enough and the Tegra can handle whatever CPU tasks the work we do requires (beyond RAM read/writes). I hope it is as large of an order of magnitude leap as it sounds. If so, we might either; A)challenge the bandwidth even at the COLO, B)risk running-out of data to process. The latter would be a good thing. The former might get us booted out of the COLO. |
kittyman Send message Joined: 9 Jul 00 Posts: 51468 Credit: 1,018,363,574 RAC: 1,004 |
I highly suspect that if the bandwidth usage ever became an issue with the colo facilities, they would simply throttle the project's bandwidth, not boot the project. "Freedom is just Chaos, with better lighting." Alan Dean Foster |
ExchangeMan Send message Joined: 9 Jan 00 Posts: 115 Credit: 157,719,104 RAC: 0 |
I don't think that 5 GPU cards or 4 explain the gap between your RAC and mine. We both have 8 GPUs, that's the bottom line. I have 1 Titan, 1 680 and 6 690s. You have 8 690s. The likely explanation is the higher performance of the Titan. |
Claggy Send message Joined: 5 Jul 99 Posts: 4654 Credit: 47,537,079 RAC: 4 |
There are app and parameter differences between your hosts, you both run the Cuda50 app, Batterup as Stock, ExchangeMan as anonymous platform, Batterup uses: mbcuda.cfg, Global pfblockspersm key being used for this device pulsefind: blocks per SM 8 mbcuda.cfg, Global pfperiodsperlaunch key being used for this device pulsefind: periods per launch 200 Priority of process set to HIGH successfully Priority of worker thread set successfully ExchangeMan uses: mbcuda.cfg, Global pfblockspersm key being used for this device pulsefind: blocks per SM 16 mbcuda.cfg, Global pfperiodsperlaunch key being used for this device pulsefind: periods per launch 200 Priority of process set to ABOVE_NORMAL successfully Priority of worker thread set successfully Batterup runs the Stock r1316 AP app with the default parameters, but with the -hp switch: DATA_CHUNK_UNROLL at default:2 Priority of worker thread raised successfully Priority of process adjusted successfully, high priority class used ExchangeMan runs the r2058 AP app with the following parameters: DATA_CHUNK_UNROLL set to:16 FFA thread block override value:16384 FFA thread fetchblock override value:8192 Sleep() & wait for event loops will be used in some places Priority of worker thread raised successfully Priority of process adjusted successfully, high priority class used ExchangeMan is also doing AP on the CPU with r2137 x64, that will give him a good Boost to his RAC too. Claggy |
juan BFP Send message Joined: 16 Mar 07 Posts: 9786 Credit: 572,710,851 RAC: 3,799 |
And not forget Titan+680 produces a lot more than a 690, actualy the 690 is equivalent to a little less than 2x670 aproximately. |
Batter Up Send message Joined: 5 May 99 Posts: 1946 Credit: 24,860,347 RAC: 0 |
And not forget Titan+680 produces a lot more than a 690, actualy the 690 is equivalent to a little less than 2x670 aproximately. Point of order; the 690 has two 680 detuned chips. My RAC just took a hit of 10,000 cobblestones because of two CPU shut downs, about 16 hours lost. I had to cut back the 4.25 overclock because of the warmer weather. The top cruncher is tuned a bit better than mine and most likely has an up time of 99 and 44/100%, no crashes. When I made changes to the code or hardware the time it took caused me to fall behind so I kept it to a minimum. Summer is coming so there has to be less crunching but that will give me a chance to tune so, I'll be back. |
juan BFP Send message Joined: 16 Mar 07 Posts: 9786 Credit: 572,710,851 RAC: 3,799 |
Point of order; the 690 has two 680 detuned chips. You are right but they are so detuned so they actualy produces about the same daily production of about 2x670 running (+/- 50K/day), who is not to bad anyway. |
©2024 University of California
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.