Panic Mode On (55) Server problems?

Message boards : Number crunching : Panic Mode On (55) Server problems?
Message board moderation

To post messages, you must log in.

Previous · 1 · 2 · 3 · 4 · 5 · 6 . . . 10 · Next

AuthorMessage
Profile James Sotherden
Avatar

Send message
Joined: 16 May 99
Posts: 10436
Credit: 110,373,059
RAC: 54
United States
Message 1154864 - Posted: 22 Sep 2011, 12:10:55 UTC

Just checked the servers and and see no MB work lots of AP . Tried to get me some but no luck, got two lousy 4 minute GPUs that took one 5 minutes of button abuse to get and the other one refused to download.
[/quote]

Old James
ID: 1154864 · Report as offensive
Profile Mad Fritz
Avatar

Send message
Joined: 20 Jul 01
Posts: 87
Credit: 11,334,904
RAC: 0
Switzerland
Message 1154866 - Posted: 22 Sep 2011, 12:19:27 UTC - in response to Message 1154858.  

...
Keep it to yourself though, dont want everybody overloading it!

:)

Regards,
Andy



Too late ;-)
Many thanks, this proxy does ATM work for me as well.
ID: 1154866 · Report as offensive
kittyman Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 9 Jul 00
Posts: 51468
Credit: 1,018,363,574
RAC: 1,004
United States
Message 1154921 - Posted: 22 Sep 2011, 16:07:21 UTC - in response to Message 1154850.  
Last modified: 22 Sep 2011, 16:08:02 UTC

Well, I don't know where that leaves my top rig.......
4 GPUs (2 GTX295s), and it has 1335 GPU tasks left, 863 of which are VHAR, which it tosses back at the rate of 4 about every 30 seconds.

You're way below 1600 and if you keep getting limit message it indicates that the number of GPU cores is not in the mix. Let us see what'll happen when it gets below 800... Maybe it is 400 per card?

Dunno.......
Lesse....400 per card, 2 GPUs per card, 2 minutes per WU.....
That would be a cache of what, 6.67 hours?


I don't think it's even going to be as good as 400 per card...

My triple 460 machine has been pegging against the 400 limit for the past couple hours. Report 10, get 10. Report 20, get 20. Rinse and repeat. It looks like the limits in place this time are per processor type (CPU/GPU), per host. 50 for CPU, 400 per GPU. Unlike the old limits that were in place during the days of the 3 day outages, which IIRC, were 40 per CPU core, and 320 per GPU core.

With the current shorty storm, my triple machine's 400 WU cache is just under 4 hours of work.

Ugh.

Apparently not 400 per card........
Last night I had 1335 WUs for the GPUs on my top rig, of which 863 were VHAR.
This morning, I have 677 left, of which 328 are VHAR.

And still banging it's head against the 'in process' limit.

Meowwwarrggghhhhhhhhhhhhhhhhhhhhh..........
"Freedom is just Chaos, with better lighting." Alan Dean Foster

ID: 1154921 · Report as offensive
kittyman Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 9 Jul 00
Posts: 51468
Credit: 1,018,363,574
RAC: 1,004
United States
Message 1154954 - Posted: 22 Sep 2011, 17:26:49 UTC

And it appears the splitters are slowing down even more...
I suspect we shall be observing some chinks in the Cricket graph shortly.

Need some more assimilator power online to deal with the shorty influx.
"Freedom is just Chaos, with better lighting." Alan Dean Foster

ID: 1154954 · Report as offensive
Profile zoom3+1=4
Volunteer tester
Avatar

Send message
Joined: 30 Nov 03
Posts: 65746
Credit: 55,293,173
RAC: 49
United States
Message 1154958 - Posted: 22 Sep 2011, 17:31:47 UTC - in response to Message 1154954.  

And it appears the splitters are slowing down even more...
I suspect we shall be observing some chinks in the Cricket graph shortly.

Need some more assimilator power online to deal with the shorty influx.

Yeah and this 4 wu limit is getting to be a real drag on My 4 gpu pc(2 GTX295 cards, It's capable of holding 3)...

I keep getting pestered to do AP v505 too...
The T1 Trust, PRR T1 Class 4-4-4-4 #5550, 1 of America's First HST's
ID: 1154958 · Report as offensive
kittyman Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 9 Jul 00
Posts: 51468
Credit: 1,018,363,574
RAC: 1,004
United States
Message 1154960 - Posted: 22 Sep 2011, 17:36:32 UTC - in response to Message 1154958.  
Last modified: 22 Sep 2011, 17:37:15 UTC

And it appears the splitters are slowing down even more...
I suspect we shall be observing some chinks in the Cricket graph shortly.

Need some more assimilator power online to deal with the shorty influx.

Yeah and this 4 wu limit is getting to be a real drag on My 4 gpu pc(2 GTX295 cards, It's capable of holding 3)...

I keep getting pestered to do AP v505 too...

It's not a 4 WU limit....
It appears to be 400 per host with a GPU (for the GPU cache).
Your DCF has probably gone so low that you are only requesting 1 task per GPU.

See the 'cannot get any cache built up' thread.
"Freedom is just Chaos, with better lighting." Alan Dean Foster

ID: 1154960 · Report as offensive
Profile zoom3+1=4
Volunteer tester
Avatar

Send message
Joined: 30 Nov 03
Posts: 65746
Credit: 55,293,173
RAC: 49
United States
Message 1154962 - Posted: 22 Sep 2011, 17:42:18 UTC - in response to Message 1154960.  

And it appears the splitters are slowing down even more...
I suspect we shall be observing some chinks in the Cricket graph shortly.

Need some more assimilator power online to deal with the shorty influx.

Yeah and this 4 wu limit is getting to be a real drag on My 4 gpu pc(2 GTX295 cards, It's capable of holding 3)...

I keep getting pestered to do AP v505 too...

It's not a 4 WU limit....
It appears to be 400 per host with a GPU (for the GPU cache).
Your DCF has probably gone so low that you are only requesting 1 task per GPU.

See the 'cannot get any cache built up' thread.

Ok, maybe It has, Oh wise non-machine liter(pun intended). So I'm essentially in limp along mode, figures. :(
The T1 Trust, PRR T1 Class 4-4-4-4 #5550, 1 of America's First HST's
ID: 1154962 · Report as offensive
kittyman Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 9 Jul 00
Posts: 51468
Credit: 1,018,363,574
RAC: 1,004
United States
Message 1154965 - Posted: 22 Sep 2011, 17:51:18 UTC - in response to Message 1154962.  

And it appears the splitters are slowing down even more...
I suspect we shall be observing some chinks in the Cricket graph shortly.

Need some more assimilator power online to deal with the shorty influx.

Yeah and this 4 wu limit is getting to be a real drag on My 4 gpu pc(2 GTX295 cards, It's capable of holding 3)...

I keep getting pestered to do AP v505 too...

It's not a 4 WU limit....
It appears to be 400 per host with a GPU (for the GPU cache).
Your DCF has probably gone so low that you are only requesting 1 task per GPU.

See the 'cannot get any cache built up' thread.

Ok, maybe It has, Oh wise non-machine liter(pun intended). So I'm essentially in limp along mode, figures. :(

You might try my suggestion of slowing your GPUs way down for a bit and see if the work requests don't pick up again.
"Freedom is just Chaos, with better lighting." Alan Dean Foster

ID: 1154965 · Report as offensive
tbret
Volunteer tester
Avatar

Send message
Joined: 28 May 99
Posts: 3380
Credit: 296,162,071
RAC: 40
United States
Message 1154988 - Posted: 22 Sep 2011, 18:53:59 UTC - in response to Message 1154798.  

OK....
Just WHAT is the limit on tasks in progress right now?
My top rig is getting mighty low on cache.

Meowgrrrrrrrrrrrrrrr.


I haven't changed anything on my three speedier computers with GPUs running Lunatics since this whole thing started. Two 460s and a 560Ti (2 WUs simultaneously on each) ran completely out of GPU work.

Now all three of them are maintaining a cache of work between 230-235 WUs, total. It might not mean anything, but since the caches are grouped so tightly together, it might. (we may have accidentally stumbled into one of those "math tricks" where you add a magic number to any number and divide by another given number... and the result is you'll always get your birthday)

These three computers are "about" equally quick with a GPU work unit.

The fourth machine (P4 with GT 240) -- which is substantially slower with a GPU work unit -- had a nice big cache, but that cache is now 330-ish and falling. It will be interesting to see if it, too, gets into that 230 range and begins to hold.

I hope I don't get the chance to observe that, but if all four fall to that 230 WU cache level, then I would be comfortable guessing that the formula (intentional or not) is resulting in a max cache of about 230 WUs. If it falls through the 230-cache level, I'll feel comfortable guessing that the formula is yielding a really short "time restriction."

It will take another couple of days for me to think I can glean anything from what I'm seeing at this altitude (not getting into each computer's individual DCFs, etc). But I continue to leave my computers alone so that there is a chance I can observe something. I hesitate to get into trying to straighten this out by hand for fear that I'll defeat(or fight)the servers' attempts to compensate and just make long-term matters worse. I just don't understand the problem well enough to think I can fix it.



ID: 1154988 · Report as offensive
Profile S@NL Etienne Dokkum
Volunteer tester
Avatar

Send message
Joined: 11 Jun 99
Posts: 212
Credit: 43,822,095
RAC: 0
Netherlands
Message 1154994 - Posted: 22 Sep 2011, 19:21:03 UTC
Last modified: 22 Sep 2011, 19:22:09 UTC

well, as a non gpu user I might be not have a high tasks in progress limit, but when on both machines I get a max. of 25 tasks per core and they are all shorties which take about 35 to 40 minutes to complete I wonder how this will help te network problems as reported by Jeff yesterday.

Every time my machines contact berkeley ( every 5 min. to fetch work ) I report 1 and get 1 WU. This couldn't be helpful to the load on the servers...

Well, all to do is wait and see what comes...
ID: 1154994 · Report as offensive
Profile kepan

Send message
Joined: 17 Sep 99
Posts: 7
Credit: 27,442,770
RAC: 0
Sweden
Message 1155004 - Posted: 22 Sep 2011, 19:52:13 UTC

I'm still having problems to upload/report/download WU's. The problems are like in august.

/Per, Sweden
ID: 1155004 · Report as offensive
Profile Dimly Lit Lightbulb 😀
Volunteer tester
Avatar

Send message
Joined: 30 Aug 08
Posts: 15399
Credit: 7,423,413
RAC: 1
United Kingdom
Message 1155014 - Posted: 22 Sep 2011, 20:43:05 UTC

Lil' ol' mini cruncher Sparky has reached a limit on tasks in progress, sayeth the messages/event log. Humbug :).
ID: 1155014 · Report as offensive
Bill Beeman

Send message
Joined: 15 May 99
Posts: 11
Credit: 7,722,342
RAC: 0
United States
Message 1155017 - Posted: 22 Sep 2011, 20:46:42 UTC

Looks like the router needs reloading again. I managed to get some uploaded by using Hotspot Shield, but it is blind to all my machines without that.

Same symptoms as before.
ID: 1155017 · Report as offensive
Profile Mad Fritz
Avatar

Send message
Joined: 20 Jul 01
Posts: 87
Credit: 11,334,904
RAC: 0
Switzerland
Message 1155032 - Posted: 22 Sep 2011, 22:18:11 UTC - in response to Message 1155014.  

Lil' ol' mini cruncher Sparky has reached a limit on tasks in progress, sayeth the messages/event log. Humbug :).


At least you get messages - for some others and me the HE-router is acting like a firewall ^^
ID: 1155032 · Report as offensive
Profile Dave Barstow

Send message
Joined: 14 May 99
Posts: 76
Credit: 15,064,044
RAC: 0
Philippines
Message 1155036 - Posted: 22 Sep 2011, 22:41:36 UTC - in response to Message 1155032.  

Just a note from the other side of the world... for some anyway.

I am now down to only one task running (GPU GRID) and can't report, upload or download from S@H.

This IS getting tiresome...............
ID: 1155036 · Report as offensive
Profile Akio
Avatar

Send message
Joined: 18 May 11
Posts: 375
Credit: 32,129,242
RAC: 0
United States
Message 1155050 - Posted: 22 Sep 2011, 23:23:22 UTC - in response to Message 1155036.  

I had to "no more tasks" my Seti...I had a plethora of shorties. As far as the uploads everything is going smooth for me. I saw a FB post from Seti@Home that there were issues, but everything seems to be working fine on my end. Are there still on-going issues with people have trouble with downloads and uploads?
ID: 1155050 · Report as offensive
Profile Mad Fritz
Avatar

Send message
Joined: 20 Jul 01
Posts: 87
Credit: 11,334,904
RAC: 0
Switzerland
Message 1155061 - Posted: 22 Sep 2011, 23:42:32 UTC - in response to Message 1155050.  
Last modified: 23 Sep 2011, 0:15:45 UTC

Yes.
_______________________________

PS C:\> tracert -h 15 208.68.240.13

Routenverfolgung zu boinc2.ssl.berkeley.edu [208.68.240.13] über maximal 15 Abschnitte:

1 2 ms 1 ms 1 ms 192.168.62.1
2 11 ms 12 ms 9 ms 217-162-191-1.dynamic.hispeed.ch [217.162.191.1]
3 10 ms 9 ms 9 ms 217-168-54-61.static.cablecom.ch [217.168.54.61]
4 11 ms 11 ms 11 ms 172.31.208.69
5 122 ms 120 ms 119 ms 84-116-130-49.aorta.net [84.116.130.49]
6 120 ms 119 ms 150 ms us-was03a-rd1-xe-0-3-0.aorta.net [84.116.130.66]
7 122 ms 121 ms 121 ms us-nyc01c-rd1-ge-15-0-0.aorta.net [84.116.130.161]
8 123 ms 121 ms 121 ms us-nyc01b-ri1-xe-4-1-0.aorta.net [213.46.190.98]
9 122 ms 124 ms 124 ms core1.nyc4.he.net [198.32.118.57]
10 192 ms 191 ms 194 ms 10gigabitethernet10-2.core1.sjc2.he.net [184.105.213.197]
11 193 ms 200 ms 199 ms 10gigabitethernet3-2.core1.pao1.he.net [72.52.92.69]
12 * * * Zeitüberschreitung der Anforderung.

PS C:\> tracert -h 15 208.68.240.16

Routenverfolgung zu setiboincdata.ssl.berkeley.edu [208.68.240.16] über maximal 15 Abschnitte:

1 2 ms 1 ms 1 ms 192.168.62.1
2 9 ms 9 ms 9 ms 217-162-191-1.dynamic.hispeed.ch [217.162.191.1]
3 9 ms 10 ms 8 ms 217-168-54-61.static.cablecom.ch [217.168.54.61]
4 11 ms 11 ms 14 ms 172.31.208.69
5 120 ms 122 ms 123 ms 84-116-130-49.aorta.net [84.116.130.49]
6 120 ms 120 ms 119 ms us-was03a-rd1-xe-0-3-0.aorta.net [84.116.130.66]
7 122 ms 120 ms 120 ms us-nyc01c-rd1-ge-15-0-0.aorta.net [84.116.130.161]
8 122 ms 124 ms 120 ms us-nyc01b-ri1-xe-4-1-0.aorta.net [213.46.190.98]
9 124 ms 121 ms 122 ms core1.nyc4.he.net [198.32.118.57]
10 197 ms 199 ms 199 ms 10gigabitethernet10-1.core1.sjc2.he.net [184.105.213.173]
11 192 ms 191 ms 195 ms 10gigabitethernet3-2.core1.pao1.he.net [72.52.92.69]
12 * * * Zeitüberschreitung der Anforderung.

PS C:\> tracert -h 15 208.68.240.18

Routenverfolgung zu boinc2.ssl.berkeley.edu [208.68.240.18] über maximal 15 Abschnitte:

1 2 ms 1 ms 1 ms 192.168.62.1
2 9 ms 9 ms 10 ms 217-162-191-1.dynamic.hispeed.ch [217.162.191.1]
3 8 ms 9 ms 9 ms 217-168-54-61.static.cablecom.ch [217.168.54.61]
4 25 ms 10 ms 11 ms 172.31.208.69
5 138 ms 122 ms 119 ms 84.116.134.25
6 120 ms 119 ms 119 ms us-was03a-rd1-xe-1-3-0.aorta.net [84.116.130.70]
7 120 ms 122 ms 119 ms us-nyc01c-rd1-ge-15-0-0.aorta.net [84.116.130.161]
8 120 ms 119 ms 119 ms us-nyc01b-ri1-xe-4-1-0.aorta.net [213.46.190.98]
9 127 ms 123 ms 125 ms core1.nyc4.he.net [198.32.118.57]
10 190 ms 189 ms 199 ms 10gigabitethernet10-2.core1.sjc2.he.net [184.105.213.197]
11 203 ms 200 ms 199 ms 10gigabitethernet3-2.core1.pao1.he.net [72.52.92.69]
12 * * * Zeitüberschreitung der Anforderung.

PS C:\> tracert -h 15 208.68.240.20

Routenverfolgung zu setiboinc.ssl.berkeley.edu [208.68.240.20] über maximal 15 Abschnitte:

1 2 ms 1 ms 1 ms 192.168.62.1
2 9 ms 9 ms 9 ms 217-162-191-1.dynamic.hispeed.ch [217.162.191.1]
3 9 ms 9 ms 9 ms 217-168-54-61.static.cablecom.ch [217.168.54.61]
4 10 ms 11 ms 12 ms 172.31.208.69
5 118 ms 118 ms 118 ms 84-116-130-53.aorta.net [84.116.130.53]
6 120 ms 119 ms 119 ms us-was03a-rd1-xe-1-3-0.aorta.net [84.116.130.70]
7 120 ms 119 ms 119 ms us-nyc01c-rd1-ge-15-0-0.aorta.net [84.116.130.161]
8 121 ms 119 ms 119 ms us-nyc01b-ri1-xe-4-1-0.aorta.net [213.46.190.98]
9 122 ms 124 ms 125 ms core1.nyc4.he.net [198.32.118.57]
10 191 ms 190 ms 199 ms 10gigabitethernet10-2.core1.sjc2.he.net [184.105.213.197]
11 190 ms 191 ms 195 ms 10gigabitethernet3-2.core1.pao1.he.net [72.52.92.69]
12 * * * Zeitüberschreitung der Anforderung.
ID: 1155061 · Report as offensive
Rick
Avatar

Send message
Joined: 3 Dec 99
Posts: 79
Credit: 11,486,227
RAC: 0
United States
Message 1155065 - Posted: 23 Sep 2011, 0:10:04 UTC
Last modified: 23 Sep 2011, 0:13:33 UTC

Seem to be getting tasks now but downloads are still a bit iffy. I just got a set of 18 AP tasks that are trying to download. Hadn't noticed it before but those are 8MB each. If things were flowing normally I would be getting transfers at between 30 and 40KBps but now I'm lucky to see 10MBps on most of my downloads.

Also see that the estimates for these tasks is about 31 hours.
ID: 1155065 · Report as offensive
Profile soft^spirit
Avatar

Send message
Joined: 18 May 99
Posts: 6497
Credit: 34,134,168
RAC: 0
United States
Message 1155092 - Posted: 23 Sep 2011, 1:39:34 UTC

It seems I have finally hit the current server side cap of 450 units.

Janice
ID: 1155092 · Report as offensive
Profile zoom3+1=4
Volunteer tester
Avatar

Send message
Joined: 30 Nov 03
Posts: 65746
Credit: 55,293,173
RAC: 49
United States
Message 1155100 - Posted: 23 Sep 2011, 2:04:09 UTC - in response to Message 1155092.  

It seems I have finally hit the current server side cap of 450 units.

Mine seems to be at 631 right now. :D All cause of the Flops setting in My seti xml file.

Of course I put Boinc back to its normal summer schedule of 6-8 hours a night, as I'm not crazy about the room being 82F on humid cloudy days.
The T1 Trust, PRR T1 Class 4-4-4-4 #5550, 1 of America's First HST's
ID: 1155100 · Report as offensive
Previous · 1 · 2 · 3 · 4 · 5 · 6 . . . 10 · Next

Message boards : Number crunching : Panic Mode On (55) Server problems?


 
©2024 University of California
 
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.