Message boards :
Number crunching :
Panic Mode On (77) Server Problems?
Message board moderation
Previous · 1 . . . 10 · 11 · 12 · 13 · 14 · 15 · 16 . . . 22 · Next
Author | Message |
---|---|
Speedy Send message Joined: 26 Jun 04 Posts: 1639 Credit: 12,921,799 RAC: 89 |
From Tech news 2nd Oct. However one sudden crisis at the end of the day today: the air conditioning in the building seems to have gone kaput. Our server closet is just fine (phew!) but we do have several servers not in the closet and they are burning up. We are shutting a few of the less necessary ones off for the evening. Hopefully the a/c will be fixed before too long. I'd say this is the reason for Splitters running slow. As I type Current result creation rate is 31.2241/sec As of* 6m & to send buffer is 0 |
JohnDK Send message Joined: 28 May 00 Posts: 1222 Credit: 451,243,443 RAC: 1,127 |
Now if they could just crank the splitters up a couple of notches... Since the cricket is maxed as it is now, I think they should leave things as they are now. |
Wiggo Send message Joined: 24 Jan 00 Posts: 34744 Credit: 261,360,520 RAC: 489 |
Now if they could just crank the splitters up a couple of notches... The splitters could probably produce work except that the file "24my12ad" has been stuck at 14 for nearly 24hrs now. Cheers. |
Grant (SSSF) Send message Joined: 19 Aug 99 Posts: 13720 Credit: 208,696,464 RAC: 304 |
Now if they could just crank the splitters up a couple of notches... The problem is that people's caches aren't being refilled. More work split will mean they will be refilled, even with the heavier network traffic (and congestion). Grant Darwin NT |
Grant (SSSF) Send message Joined: 19 Aug 99 Posts: 13720 Credit: 208,696,464 RAC: 304 |
From Tech news 2nd Oct. I noticed that, and while those machines were shut down down none of them were used for splitting work, feeding or Scheduling. This is some other problem. And even with the reduced availability of work, i'm still getting Scheduler timeouts. Grant Darwin NT |
Link Send message Joined: 18 Sep 03 Posts: 834 Credit: 1,807,369 RAC: 0 |
Now if they could just crank the splitters up a couple of notches... And how does it help them to get tasks assigned, which they can't download? I have to agree here with JohnDK, right now they are sending out exactly as much as they can, more will only make the things worse. You won't get more thru a network connection when you push it harder, you'll get less. |
Grant (SSSF) Send message Joined: 19 Aug 99 Posts: 13720 Credit: 208,696,464 RAC: 304 |
And how does it help them to get tasks assigned, which they can't download? I have to agree here with JohnDK, right now they are sending out exactly as much as they can, more will only make the things worse. You won't get more thru a network connection when you push it harder, you'll get less. That's generally the case. The interesting thing is that when there is more work available, i'm able to get it & eventually download it. That helps reduce the load on the Scheduler because it doesn't have to deaal with nearly as many requests. At the moment, my caches are re-filling, but at the present rate it will take a couple of weeks- if there are no outages or hiccups between now & then. Normally, inspite of the load, even after an extended outage it usually only take 8-12 hours for my caches to be filled. I haven't had a full cache for over 3 weeks now. Grant Darwin NT |
Morten Ross Send message Joined: 30 Apr 01 Posts: 183 Credit: 385,664,915 RAC: 0 |
Something is definitely changing in regards to work distribution as I broke the 100 mark today for number of tasks assigned after a request: 06/10/2012 11:10:37 | SETI@home | Sending scheduler request: To fetch work. 06/10/2012 11:10:37 | SETI@home | Reporting 31 completed tasks, requesting new tasks for CPU and NVIDIA 06/10/2012 11:10:40 | SETI@home | Computation for task 12mr10ab.30517.20926.140733193388047.10.170_0 finished 06/10/2012 11:10:40 | SETI@home | Starting task 12mr10ab.30517.20926.140733193388047.10.158_0 using setiathome_enhanced version 610 (cuda_fermi) in slot 31 06/10/2012 11:10:43 | SETI@home | Finished upload of 12mr10ab.30517.20926.140733193388047.10.152_0_0 06/10/2012 11:10:43 | SETI@home | Started upload of 12mr10ab.30517.20926.140733193388047.10.170_0_0 06/10/2012 11:10:51 | SETI@home | Finished upload of 12mr10ab.30517.20926.140733193388047.10.170_0_0 06/10/2012 11:11:27 | SETI@home | Scheduler request completed: got 113 new tasks Morten Ross |
Link Send message Joined: 18 Sep 03 Posts: 834 Credit: 1,807,369 RAC: 0 |
And how does it help them to get tasks assigned, which they can't download? I have to agree here with JohnDK, right now they are sending out exactly as much as they can, more will only make the things worse. You won't get more thru a network connection when you push it harder, you'll get less. Not if the scheduler replies get lost due to overloaded network connection like it happened recently quite often. Than the resend lost tasks thingy has to work a lot, let many VLAR tasks time out, new replacement results have to be created and send to someone else (so the scheduler has to send out the same tasks more than once). I might be wrong, but that does not sound like less load to me, it's more and not only for the scheduler. |
Grant (SSSF) Send message Joined: 19 Aug 99 Posts: 13720 Credit: 208,696,464 RAC: 304 |
The interesting thing is that when there is more work available, i'm able to get it & eventually download it. The Scheduler timeouts are a major problem, and i too expect they are the major cause of Ghost WUs & resends. However they don't appear to be related to network traffic load. I've been getting a lot of Scheduler "timeout reached" messages for the last 3 weeks, when we were having upload issues, as well as the present lack of work produced by the splitters. Yet in the past, after multi-day outages when the ready to send buffer was 200,000+ WUs & the download speeds were lucky to be 2kB/s, Sheduler "timeout reached messages" were few & far between. Usually it was "couldn't contact Scheduler" (or similar). Whatever the present issue is with the Scheduler timeouts, it's not due to the download traffic, and it's not due to the change in download server software as the problems were occuring before that was implemented. Grant Darwin NT |
Wiggo Send message Joined: 24 Jan 00 Posts: 34744 Credit: 261,360,520 RAC: 489 |
Well I can't complain now as all 3 of my rigs have their caches back up to scratch. Cheers. |
Mike Send message Joined: 17 Feb 01 Posts: 34253 Credit: 79,922,639 RAC: 80 |
And how does it help them to get tasks assigned, which they can't download? I have to agree here with JohnDK, right now they are sending out exactly as much as they can, more will only make the things worse. You won't get more thru a network connection when you push it harder, you'll get less. I totally agree on that. With each crime and every kindness we birth our future. |
.clair. Send message Joined: 4 Nov 04 Posts: 1300 Credit: 55,390,408 RAC: 69 |
Something is definitely changing in regards to work distribution as I broke the 100 mark today for number of tasks assigned after a request: I saw a got 95 yesterday, that is the most i have ever seen in one go i did not think it was posible to get more than 100 at a time, unless the que has been enlarged !! |
Morten Ross Send message Joined: 30 Apr 01 Posts: 183 Credit: 385,664,915 RAC: 0 |
Something is definitely changing in regards to work distribution as I broke the 100 mark today for number of tasks assigned after a request: I've just maxed out at 131: 06/10/2012 12:58:34 | SETI@home | Sending scheduler request: To fetch work. 06/10/2012 12:58:34 | SETI@home | Requesting new tasks for CPU 06/10/2012 13:00:02 | SETI@home | Scheduler request completed: got 131 new tasks Morten Ross |
Link Send message Joined: 18 Sep 03 Posts: 834 Credit: 1,807,369 RAC: 0 |
The Scheduler timeouts are a major problem, and i too expect they are the major cause of Ghost WUs & resends. Well, from here we can only guess, what the current reason is, but the general rule of thumb is not to push the entire system harder than it's slowest part (here the 100Mbit connection to the outside world) can take, maybe even a bit less than that. There it usually will work best. |
fscheel Send message Joined: 13 Apr 12 Posts: 73 Credit: 11,135,641 RAC: 0 |
Is there a utility or something that will easily show me how many and what type tasks are presently on my machine? |
Fred E. Send message Joined: 22 Jul 99 Posts: 768 Credit: 24,140,697 RAC: 0 |
Is there a utility or something that will easily show me how many and what type tasks are presently on my machine? BOINC Tasks will do that and more - you can monitor all your computers from one. It doesn't replace BOINC Manager, just provides a better user interface with task counts, sum of estimated completion times, etc. Another Fred Support SETI@home when you search the Web with GoodSearch or shop online with GoodShop. |
fscheel Send message Joined: 13 Apr 12 Posts: 73 Credit: 11,135,641 RAC: 0 |
Is there a utility or something that will easily show me how many and what type tasks are presently on my machine? Thanks.. Will give that a try. |
Fred E. Send message Joined: 22 Jul 99 Posts: 768 Credit: 24,140,697 RAC: 0 |
Is there a utility or something that will easily show me how many and what type tasks are presently on my machine? This page shows how to set it up for multiple computers and has some more screenshots. Another Fred Support SETI@home when you search the Web with GoodSearch or shop online with GoodShop. |
Tim Send message Joined: 19 May 99 Posts: 211 Credit: 278,575,259 RAC: 0 |
I finally did it. I manage to finish a daily quota... :-) 6/10/2012 6:32:54 μμ SETI@home Scheduler request completed: got 0 new tasks 6/10/2012 6:32:54 μμ SETI@home Message from server: No tasks sent 6/10/2012 6:32:54 μμ SETI@home Message from server: No tasks are available for SETI@home Enhanced 6/10/2012 6:32:54 μμ SETI@home Message from server: This computer has finished a daily quota of 1 tasks |
©2024 University of California
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.