Message boards :
Number crunching :
Panic Mode On (63) Server problems?
Message board moderation
Previous · 1 . . . 5 · 6 · 7 · 8 · 9 · Next
Author | Message |
---|---|
kittyman ![]() ![]() ![]() ![]() Send message Joined: 9 Jul 00 Posts: 51525 Credit: 1,018,363,574 RAC: 1,004 ![]() ![]() |
That's why I have been stumping for weeks now for the Admins and Devs to address the Boinc code problems, get them behind us, and get the dang limits lifted. "Time is simply the mechanism that keeps everything from happening all at once." ![]() |
kittyman ![]() ![]() ![]() ![]() Send message Joined: 9 Jul 00 Posts: 51525 Credit: 1,018,363,574 RAC: 1,004 ![]() ![]() |
Shouldn't we be in the middle of the usual Tuesday outage by now? Maybe they'll skip the outage this time, because staff is on leave during Christmas/New Year? That could be the case, but I am not certain. If they are on hiatus for the whole week, they may let things limp along as they are other than possibly what could be addressed by remote. Or, possibly an outage later in the week. Just dunno for sure. "Time is simply the mechanism that keeps everything from happening all at once." ![]() |
Cosmic_Ocean ![]() Send message Joined: 23 Dec 00 Posts: 3027 Credit: 13,516,867 RAC: 13 ![]() ![]() |
What is so odd about that? All things being equal, the quad core is going to do 4 times the amount of work than the single core would. So would have to get and successfully download 4 times the amount of work just to stay even, much less build it's cache. So when up/download and work requests are not flowing well, it's going to be the first one to feel the pain. I do agree, however the part that I forgot was that the single-core machine would get at least one task about 95% of the time it asked for work. The quad-core machine would have about a 10% success rate. Slow machine would get its ~50 MBs in less than 10 requests, but the quad would have to ask for work 50+ times to get maybe 75. Something else I'm pondering is if there is any way to speed up the refill rate for the feeder. I've heard that it fills up every two seconds. I wonder if that can be dropped to 1 second if it's even possible? That may alleviate a lot of those "project has no tasks available" messages when server status shows 200,000+ waiting to be assigned. Linux laptop: record uptime: 1511d 20h 19m (ended due to the power brick giving-up) |
kittyman ![]() ![]() ![]() ![]() Send message Joined: 9 Jul 00 Posts: 51525 Credit: 1,018,363,574 RAC: 1,004 ![]() ![]() |
What is so odd about that? All things being equal, the quad core is going to do 4 times the amount of work than the single core would. So would have to get and successfully download 4 times the amount of work just to stay even, much less build it's cache. So when up/download and work requests are not flowing well, it's going to be the first one to feel the pain. I think optimizing the scheduler is a moot point until such time as there is bandwidth available to support it. My view is that has to happen first, then scheduler or other server based bottlenecks can be addressed as they are identified. You can schedule all the work you want, but if the hosts cannot get it downloaded, it cannot be processed. "Time is simply the mechanism that keeps everything from happening all at once." ![]() |
Cosmic_Ocean ![]() Send message Joined: 23 Dec 00 Posts: 3027 Credit: 13,516,867 RAC: 13 ![]() ![]() |
I think optimizing the scheduler is a moot point until such time as there is bandwidth available to support it. My view is that has to happen first, then scheduler or other server based bottlenecks can be addressed as they are identified. You can schedule all the work you want, but if the hosts cannot get it downloaded, it cannot be processed. That is true. And I've stated a few times that if we can get more bandwidth, it may create a whole new pile of problems all by itself by allowing more successful contacts to the database. It's one of those things that we'll just have to wait and see what happens and have some contingency plans lined up for some of the possible scenarios. However, the good news is that with all of the enterprise-class networking equipment that is in place, we can get an actual gigabit link, but still rate-limit it to 100mbit, or 150mbit, whatever seems to allow the smoothest data transfer while keeping the database from getting DDoS'ed. Linux laptop: record uptime: 1511d 20h 19m (ended due to the power brick giving-up) |
kittyman ![]() ![]() ![]() ![]() Send message Joined: 9 Jul 00 Posts: 51525 Credit: 1,018,363,574 RAC: 1,004 ![]() ![]() |
I think optimizing the scheduler is a moot point until such time as there is bandwidth available to support it. My view is that has to happen first, then scheduler or other server based bottlenecks can be addressed as they are identified. You can schedule all the work you want, but if the hosts cannot get it downloaded, it cannot be processed. Well, if you peruse the information in the GPUUG fundraising thread, you will see than many hardware upgrades are well on their way to being completed. With more to come. As far as I know, we still do not have a real path in place for upgrading the bandwidth, other than having the project's pleas fall on the deaf ears of the Berk IT admins. "Time is simply the mechanism that keeps everything from happening all at once." ![]() |
Josef W. Segur Send message Joined: 30 Oct 99 Posts: 4504 Credit: 1,414,761 RAC: 0 ![]() |
Shouldn't we be in the middle of the usual Tuesday outage by now? Maybe they'll skip the outage this time, because staff is on leave during Christmas/New Year? Hmm, the UC Berkeley Academic Calendar shows Monday, Tuesday, Thursday, and Friday as "Academic and Administrative Holiday". Joe |
kittyman ![]() ![]() ![]() ![]() Send message Joined: 9 Jul 00 Posts: 51525 Credit: 1,018,363,574 RAC: 1,004 ![]() ![]() |
Shouldn't we be in the middle of the usual Tuesday outage by now? Maybe they'll skip the outage this time, because staff is on leave during Christmas/New Year? Ahhh.... So it looks like some of our indentured servants may be in the lab tomorrow for an outage party. "Time is simply the mechanism that keeps everything from happening all at once." ![]() |
Richard1949 Send message Joined: 20 Oct 99 Posts: 18 Credit: 232,635 RAC: 0 ![]() |
"I do agree, however the part that I forgot was that the single-core machine would get at least one task about 95% of the time it asked for work." --------------------------------------------------- I can't even get anything for my single core machine. |
Richard1949 Send message Joined: 20 Oct 99 Posts: 18 Credit: 232,635 RAC: 0 ![]() |
"Something else I'm pondering is if there is any way to speed up the refill rate for the feeder. I've heard that it fills up every two seconds. I wonder if that can be dropped to 1 second if it's even possible? That may alleviate a lot of those "project has no tasks available" messages when server status shows 200,000+ waiting to be assigned." ---------------------------------------------- I keep getting "not requesting any tasks." |
Grant (SSSF) Send message Joined: 19 Aug 99 Posts: 13904 Credit: 208,696,464 RAC: 304 ![]() ![]() |
15min to download 1 WU is a bit of a PITA when it takes less than 3min to do 2. Grant Darwin NT |
![]() ![]() Send message Joined: 24 Jan 00 Posts: 37613 Credit: 261,360,520 RAC: 489 ![]() ![]() |
Personally I still put the current problems on the connection itself between the USA side of our undersea cable and HE as using a proxy here quickly clears any backlogs that occur. Cheers. |
![]() Send message Joined: 19 Sep 00 Posts: 3184 Credit: 446,358 RAC: 0 ![]() |
I keep getting "not requesting any tasks." And why would that be a server problem when your client doesn't ask for work? Gruß, Gundolf |
Cosmic_Ocean ![]() Send message Joined: 23 Dec 00 Posts: 3027 Credit: 13,516,867 RAC: 13 ![]() ![]() |
Yeah, that certainly would appear to be that under-sea cable. I noticed in my messages tab last night that between "starting download" and "finished download" for an AP, 19 seconds elapsed (~430KB/sec). Of course it was a B3_P1 WU, so it took 24 seconds to error out once processing started. Go figure. Linux laptop: record uptime: 1511d 20h 19m (ended due to the power brick giving-up) |
Dave Send message Joined: 29 Mar 02 Posts: 778 Credit: 25,001,396 RAC: 0 ![]() |
So are we in an outage? Why is there no work? |
rob smith ![]() ![]() ![]() Send message Joined: 7 Mar 03 Posts: 22723 Credit: 416,307,556 RAC: 380 ![]() ![]() |
No signs of an outage, and I've been getting a fairly steady stream of work Bob Smith Member of Seti PIPPS (Pluto is a Planet Protest Society) Somewhere in the (un)known Universe? |
Richard1949 Send message Joined: 20 Oct 99 Posts: 18 Credit: 232,635 RAC: 0 ![]() |
I keep getting "not requesting any tasks." I don't know whats going on. I have uninstalled and reinstalled. I have reset the project. I have tried different BOINC versions. Nothing works. All I get when I request tasks is "not requesting tasks" message. Every other BOINC project works fine. It's only SETI I can't get anything from even though it shows it has 100's of thousands of tasks ready to send out. |
Kevin Olley Send message Joined: 3 Aug 99 Posts: 906 Credit: 261,085,289 RAC: 572 ![]() ![]() |
No signs of an outage, and I've been getting a fairly steady stream of work Bouncing off the limits here. Pity its all sitting in my download queue:-( Dropped back to 1 WU per card and with max button pushing only getting enough to keep on average 2 out of 3 cards running. Shorties and AP's, never a good combination. Kevin ![]() ![]() ![]() |
Claggy Send message Joined: 5 Jul 99 Posts: 4654 Credit: 47,537,079 RAC: 4 ![]() |
I keep getting "not requesting any tasks." If your host isn't asking for work from Seti then eithier you're got No New Tasks set, you're got the Seti Project Suspended, you're got one or more Seti Tasks Suspended, or your Cache is Full already. Claggy |
Dave Send message Joined: 29 Mar 02 Posts: 778 Credit: 25,001,396 RAC: 0 ![]() |
Shorties and AP's, never a good combination. We're after the Goldilocks zone. |
©2025 University of California
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.