Message boards :
Number crunching :
Unannounced Server-Side Change?
Message board moderation
Author | Message |
---|---|
Cruncher-American Send message Joined: 25 Mar 02 Posts: 1513 Credit: 370,893,186 RAC: 340 |
I just noticed that when I do an "Update" on my machines (one 6.10.xx, the other 7.0.25) no matter how many WUs I am reporting, only 64 get reported; if I have > 64 to report, I have to Update again and again to get all WUs reported - 64 at a time. This change occurred well after the recent outage. This is new, is it not? WHY? Also, most times when I report, I get stuff like this: 5/22/2012 9:29:21 PM | SETI@home | Sending scheduler request: Requested by user. 5/22/2012 9:29:21 PM | SETI@home | Reporting 79 completed tasks, requesting new tasks for CPU 5/22/2012 9:29:32 PM | SETI@home | Scheduler request completed: got 15 new tasks 5/22/2012 9:29:32 PM | SETI@home | Resent lost task 09ja11aa.1031.24607.11.10.132_0 5/22/2012 9:29:32 PM | SETI@home | Resent lost task 09ja11aa.1031.24607.11.10.135_0 5/22/2012 9:29:32 PM | SETI@home | Resent lost task 09ja11aa.1031.24607.11.10.141_0 5/22/2012 9:29:32 PM | SETI@home | Resent lost task 09ja11aa.1031.24607.11.10.140_0 5/22/2012 9:29:32 PM | SETI@home | Resent lost task 09ja11aa.1031.24607.11.10.147_1 5/22/2012 9:29:32 PM | SETI@home | Resent lost task 09ja11aa.1291.24607.12.10.148_1 5/22/2012 9:29:32 PM | SETI@home | Resent lost task 09ja11aa.1291.24607.12.10.151_1 5/22/2012 9:29:32 PM | SETI@home | Resent lost task 09ja11aa.1291.24607.12.10.152_0 5/22/2012 9:29:32 PM | SETI@home | Resent lost task 09ja11aa.1291.24607.12.10.169_0 5/22/2012 9:29:32 PM | SETI@home | Resent lost task 09ja11aa.1031.24607.11.10.182_0 5/22/2012 9:29:32 PM | SETI@home | Resent lost task 09ja11aa.1291.24607.12.10.175_0 5/22/2012 9:29:32 PM | SETI@home | Resent lost task 09ja11aa.1291.24607.12.10.180_0 5/22/2012 9:29:32 PM | SETI@home | Resent lost task 09ja11aa.1291.24607.12.10.183_0 5/22/2012 9:29:32 PM | SETI@home | Resent lost task 09ja11aa.1291.24607.12.10.181_0 5/22/2012 9:29:32 PM | SETI@home | Resent lost task 09ja11aa.1291.24607.12.10.184_0 5/22/2012 9:29:32 PM | SETI@home | [error] Already have task 09ja11aa.1031.24607.11.10.132_0 5/22/2012 9:29:32 PM | SETI@home | [error] Already have task 09ja11aa.1031.24607.11.10.135_0 5/22/2012 9:29:32 PM | SETI@home | [error] Already have task 09ja11aa.1031.24607.11.10.141_0 5/22/2012 9:29:32 PM | SETI@home | [error] Already have task 09ja11aa.1031.24607.11.10.140_0 5/22/2012 9:29:32 PM | SETI@home | [error] Already have task 09ja11aa.1031.24607.11.10.147_1 5/22/2012 9:29:32 PM | SETI@home | [error] Already have task 09ja11aa.1291.24607.12.10.148_1 5/22/2012 9:29:32 PM | SETI@home | [error] Already have task 09ja11aa.1291.24607.12.10.151_1 5/22/2012 9:29:32 PM | SETI@home | [error] Already have task 09ja11aa.1291.24607.12.10.152_0 5/22/2012 9:29:32 PM | SETI@home | [error] Already have task 09ja11aa.1291.24607.12.10.169_0 5/22/2012 9:29:32 PM | SETI@home | [error] Already have task 09ja11aa.1031.24607.11.10.182_0 5/22/2012 9:29:32 PM | SETI@home | [error] Already have task 09ja11aa.1291.24607.12.10.175_0 5/22/2012 9:29:32 PM | SETI@home | [error] Already have task 09ja11aa.1291.24607.12.10.180_0 5/22/2012 9:29:32 PM | SETI@home | [error] Already have task 09ja11aa.1291.24607.12.10.183_0 5/22/2012 9:29:32 PM | SETI@home | [error] Already have task 09ja11aa.1291.24607.12.10.181_0 5/22/2012 9:29:32 PM | SETI@home | [error] Already have task 09ja11aa.1291.24607.12.10.184_0 5/22/2012 9:29:49 PM | SETI@home | update requested by user 5/22/2012 9:29:53 PM | SETI@home | Sending scheduler request: Requested by user. 5/22/2012 9:29:53 PM | SETI@home | Reporting 15 completed tasks, requesting new tasks for CPU 5/22/2012 9:30:02 PM | SETI@home | Scheduler request completed: got 0 new tasks It keeps resending stuff I already have. What's up with that? |
BilBg Send message Joined: 27 May 07 Posts: 3720 Credit: 9,385,827 RAC: 0 |
Maybe some testing by Eric Korpela?: http://setiathome.berkeley.edu/forum_thread.php?id=68127&nowrap=true#1234568 Â - ALF - "Find out what you don't do well ..... then don't do it!" :) Â |
Slavac Send message Joined: 27 Apr 11 Posts: 1932 Credit: 17,952,639 RAC: 0 |
Interesting: 5/22/2012 10:07:44 PM SETI@home Message from server: Resent lost task 01dc10ae.32316.127760.8.10.148_4 5/22/2012 10:07:44 PM SETI@home Message from server: Resent lost task 16au10aa.27061.25042.15.10.220_2 5/22/2012 10:07:44 PM SETI@home Message from server: Resent lost task 15my11aa.26769.408007.13.10.212_2 5/22/2012 10:07:44 PM SETI@home Message from server: Resent lost task 16au10aa.27061.25042.15.10.204_2 5/22/2012 10:07:44 PM SETI@home Message from server: Resent lost task 15my11aa.26115.408007.14.10.228_2 5/22/2012 10:07:44 PM SETI@home Message from server: Resent lost task 15my11aa.26769.408007.13.10.229_2 5/22/2012 10:07:44 PM SETI@home Message from server: Resent lost task 01ap10aa.3212.2525.16.10.145_2 5/22/2012 10:07:44 PM SETI@home Message from server: Resent lost task 01ap10aa.3212.2525.16.10.151_2 5/22/2012 10:07:44 PM SETI@home Message from server: Resent lost task 18mr10af.19525.8836.6.10.229_2 5/22/2012 10:07:44 PM SETI@home Message from server: Resent lost task 15my11aa.26115.408007.14.10.232_2 5/22/2012 10:07:44 PM SETI@home Message from server: Resent lost task 15my11aa.26115.408007.14.10.234_2 5/22/2012 10:07:44 PM SETI@home Message from server: Resent lost task 11dc10ad.27977.3104.7.10.207_2 5/22/2012 10:07:44 PM SETI@home Message from server: Resent lost task 15my11aa.26115.408007.14.10.216_2 5/22/2012 10:07:44 PM SETI@home Message from server: Resent lost task 20au10af.5841.125317.14.10.25_3 5/22/2012 10:07:44 PM SETI@home Message from server: Resent lost task 14my11ac.12673.220918.10.10.214_3 5/22/2012 10:07:44 PM SETI@home Message from server: Resent lost task 03jn11ae.9659.135763.4.10.149_2 5/22/2012 10:07:44 PM SETI@home Message from server: Resent lost task 09ja11ac.11724.9345.9.10.187_2 5/22/2012 10:07:44 PM SETI@home Message from server: Resent lost task 22my11aa.10097.24198.12.10.116_3 5/22/2012 10:07:44 PM SETI@home Message from server: Resent lost task 20se10ac.18925.59534.16.10.187_2 5/22/2012 10:07:44 PM SETI@home Message from server: Resent lost task 03jn11ae.9659.135763.4.10.163_2 5/22/2012 10:07:44 PM SETI@home [error] Already have task 01dc10ae.32316.127760.8.10.148_4 5/22/2012 10:07:44 PM SETI@home [error] Already have task 16au10aa.27061.25042.15.10.220_2 5/22/2012 10:07:44 PM SETI@home [error] Already have task 15my11aa.26769.408007.13.10.212_2 5/22/2012 10:07:44 PM SETI@home [error] Already have task 16au10aa.27061.25042.15.10.204_2 5/22/2012 10:07:44 PM SETI@home [error] Already have task 15my11aa.26115.408007.14.10.228_2 5/22/2012 10:07:44 PM SETI@home [error] Already have task 15my11aa.26769.408007.13.10.229_2 5/22/2012 10:07:44 PM SETI@home [error] Already have task 01ap10aa.3212.2525.16.10.145_2 5/22/2012 10:07:44 PM SETI@home [error] Already have task 01ap10aa.3212.2525.16.10.151_2 5/22/2012 10:07:44 PM SETI@home [error] Already have task 18mr10af.19525.8836.6.10.229_2 5/22/2012 10:07:44 PM SETI@home [error] Already have task 15my11aa.26115.408007.14.10.232_2 5/22/2012 10:07:44 PM SETI@home [error] Already have task 15my11aa.26115.408007.14.10.234_2 5/22/2012 10:07:44 PM SETI@home [error] Already have task 11dc10ad.27977.3104.7.10.207_2 5/22/2012 10:07:44 PM SETI@home [error] Already have task 15my11aa.26115.408007.14.10.216_2 5/22/2012 10:07:44 PM SETI@home [error] Already have task 20au10af.5841.125317.14.10.25_3 5/22/2012 10:07:44 PM SETI@home [error] Already have task 14my11ac.12673.220918.10.10.214_3 5/22/2012 10:07:44 PM SETI@home [error] Already have task 03jn11ae.9659.135763.4.10.149_2 5/22/2012 10:07:44 PM SETI@home [error] Already have task 09ja11ac.11724.9345.9.10.187_2 5/22/2012 10:07:44 PM SETI@home [error] Already have task 22my11aa.10097.24198.12.10.116_3 5/22/2012 10:07:44 PM SETI@home [error] Already have task 20se10ac.18925.59534.16.10.187_2 5/22/2012 10:07:44 PM SETI@home [error] Already have task 03jn11ae.9659.135763.4.10.163_2 Executive Director GPU Users Group Inc. - brad@gpuug.org |
MikeN Send message Joined: 24 Jan 11 Posts: 319 Credit: 64,719,409 RAC: 85 |
I also noticed that after yesterdays outage, my main cruncher (the only 1 with >60 results to report) would only report 64 at a time. I just manually kept pressing update ignoring the 5 min count down until all 200 of them had gone. Not a problem with 200 tasks to report, but I would not like to have to do it with 2000! The implication is that if you leave a system to report automatically, you can only report 64 tasks every 5 minutes or 768 tasks per hour. This could cause problems for the really big crunchers, not so much from a standard weekly outage, but certainly after a 2.5 day power cut. |
kittyman Send message Joined: 9 Jul 00 Posts: 51468 Credit: 1,018,363,574 RAC: 1,004 |
Hmmm... I know Eric said he would look into the situation, but this is not the resolution I would think correct. At least not long term. Maybe he has kludged in a temporary fix until the root of the problem can be resolved. "Freedom is just Chaos, with better lighting." Alan Dean Foster |
David S Send message Joined: 4 Oct 99 Posts: 18352 Credit: 27,761,924 RAC: 12 |
Hmmm... This is probably a good solution, except that he should have picked a higher number than 64. IIRC, people only had trouble if they tried to report more than 200 at a time. David Sitting on my butt while others boldly go, Waiting for a message from a small furry creature from Alpha Centauri. |
Richard Haselgrove Send message Joined: 4 Jul 99 Posts: 14650 Credit: 200,643,578 RAC: 874 |
Well, the number I suggested when I put this idea to Eric and David on Monday was 500 - after all, the suggested value for <max_tasks_reported> in client configuration is 1000. It was David who chose to cap it at 64. I've reproduced the 'Resent lost task' problem that jravin opened this thread with, and passed the evidence on to Eric and David. |
kittyman Send message Joined: 9 Jul 00 Posts: 51468 Credit: 1,018,363,574 RAC: 1,004 |
Well, the number I suggested when I put this idea to Eric and David on Monday was 500 - after all, the suggested value for <max_tasks_reported> in client configuration is 1000. It was David who chose to cap it at 64. In the past, this was a server side settings problem that I believe Matt found and fixed rather quickly. "Freedom is just Chaos, with better lighting." Alan Dean Foster |
Richard Haselgrove Send message Joined: 4 Jul 99 Posts: 14650 Credit: 200,643,578 RAC: 874 |
Well, the number I suggested when I put this idea to Eric and David on Monday was 500 - after all, the suggested value for <max_tasks_reported> in client configuration is 1000. It was David who chose to cap it at 64. No, this will have been David writing code in a hurry, and forgetting there's important stuff like <other_results> after the last <result> he processes. He almost made the same mistake 18 months ago, until Joe prompted him to add [trac]changeset:22500[/trac]. |
skildude Send message Joined: 4 Oct 00 Posts: 9541 Credit: 50,759,529 RAC: 60 |
I had a problem with 1100+ errored WU's that wouldn't go home. Mark told me to use the config to sent a max 100 at a time. I was able to sent 100 at a time using this In a rich man's house there is no place to spit but his face. Diogenes Of Sinope |
Cruncher-American Send message Joined: 25 Mar 02 Posts: 1513 Credit: 370,893,186 RAC: 340 |
Further info: One of my rigs went 24 hours between (self-generated) updates, then reported 64 of 290 WUs. I waited > 5 minutes, but it did not self-report any more, so I hit update ("reporting" 228 - 4 more WU had finished in the interim -, 164, 100, 36) rapidly, to empty the queue. It did NOT say "too soon" until after the last one; for all but the last, I did get more "phantom" resends (i.e., already had them). |
©2024 University of California
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.