Panic Mode On (86) Server Problems?

Message boards : Number crunching : Panic Mode On (86) Server Problems?
Message board moderation

To post messages, you must log in.

Previous · 1 . . . 20 · 21 · 22 · 23 · 24 · Next

AuthorMessage
Profile Wiggo
Avatar

Send message
Joined: 24 Jan 00
Posts: 34744
Credit: 261,360,520
RAC: 489
Australia
Message 1481550 - Posted: 25 Feb 2014, 6:37:43 UTC

Well my uploads just cleared again and I'm back up on the limits.

Cheers.
ID: 1481550 · Report as offensive
Grant (SSSF)
Volunteer tester

Send message
Joined: 19 Aug 99
Posts: 13736
Credit: 208,696,464
RAC: 304
Australia
Message 1481559 - Posted: 25 Feb 2014, 6:47:51 UTC - in response to Message 1481550.  
Last modified: 25 Feb 2014, 6:50:23 UTC

Well my uploads just cleared again and I'm back up on the limits.

Cheers.

A few retries & the uploads cleared, but the most common response to a request for work is "No tasks sent". Every now & then some get allocated, but probably 75% of the requests at the moment result in "No tasks sent" responses.


EDIT- just looked at the network graphs & there's a big surge in outbound & inbound traffic. Things might be coming back to life...
Uploads appear to be going back under their own steam, although it is taking 10-20 seconds for them to actually start transferring data once the clock starts.
Grant
Darwin NT
ID: 1481559 · Report as offensive
Speedy
Volunteer tester
Avatar

Send message
Joined: 26 Jun 04
Posts: 1643
Credit: 12,921,799
RAC: 89
New Zealand
Message 1481562 - Posted: 25 Feb 2014, 6:52:12 UTC

As I write this 90,000 results have been returned and incoming traffic is up to just over 27 MB per second. My upload queue cleared an hour or so ago.

Out of interest how many results are stored in a row in the database? I am just wondering how many rows are been taken up by result waiting for DB purging there are currently 1,606,679 results waiting to be deleted and the work unit waiting purging is 779,696
ID: 1481562 · Report as offensive
Profile Wiggo
Avatar

Send message
Joined: 24 Jan 00
Posts: 34744
Credit: 261,360,520
RAC: 489
Australia
Message 1481564 - Posted: 25 Feb 2014, 6:56:20 UTC - in response to Message 1481559.  

Well my uploads just cleared again and I'm back up on the limits.

Cheers.

A few retries & the uploads cleared, but the most common response to a request for work is "No tasks sent". Every now & then some get allocated, but probably 75% of the requests at the moment result in "No tasks sent" responses.


EDIT- just looked at the network graphs & there's a big surge in outbound & inbound traffic. Things might be coming back to life...
Uploads appear to be going back under their own steam, although it is taking 10-20 seconds for them to actually start transferring data once the clock starts.

Yeah we'll just have to wait and see how it goes, though Einstein has been helpful to keep my main rigs' GPU's active the last few days while trying to get to get my SETI cache back up.

Cheers.
ID: 1481564 · Report as offensive
juan BFP Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 16 Mar 07
Posts: 9786
Credit: 572,710,851
RAC: 3,799
Panama
Message 1481590 - Posted: 25 Feb 2014, 9:28:07 UTC
Last modified: 25 Feb 2014, 9:36:01 UTC

Aparently UL are working now (still no AP work avaiable). As usual we will never know what happenig. Sometimes i think, takes only few seconds to write a note about the outages, maybe just to keep the curious updated.
ID: 1481590 · Report as offensive
Profile HAL9000
Volunteer tester
Avatar

Send message
Joined: 11 Sep 99
Posts: 6534
Credit: 196,805,888
RAC: 57
United States
Message 1481630 - Posted: 25 Feb 2014, 13:39:05 UTC - in response to Message 1481562.  

As I write this 90,000 results have been returned and incoming traffic is up to just over 27 MB per second. My upload queue cleared an hour or so ago.

Out of interest how many results are stored in a row in the database? I am just wondering how many rows are been taken up by result waiting for DB purging there are currently 1,606,679 results waiting to be deleted and the work unit waiting purging is 779,696

As I recall it is one entry per row. Which was the reason limits were needed. It was taking to long to scan the table with "Results out in the field" I believe.
SETI@home classic workunits: 93,865 CPU time: 863,447 hours
Join the [url=http://tinyurl.com/8y46zvu]BP6/VP6 User Group[
ID: 1481630 · Report as offensive
Ulrich Metzner
Volunteer tester
Avatar

Send message
Joined: 3 Jul 02
Posts: 1256
Credit: 13,565,513
RAC: 13
Germany
Message 1481634 - Posted: 25 Feb 2014, 14:15:08 UTC
Last modified: 25 Feb 2014, 14:15:46 UTC

Things appear normal again from BOINC clients view, no stuck up- or donwloads.
But the SSP looks ridiculous to me, as if 15 MB and 15 AP splitters are running momentary... :?
Aloha, Uli

ID: 1481634 · Report as offensive
Richard Haselgrove Project Donor
Volunteer tester

Send message
Joined: 4 Jul 99
Posts: 14650
Credit: 200,643,578
RAC: 874
United Kingdom
Message 1481637 - Posted: 25 Feb 2014, 14:30:42 UTC - in response to Message 1481630.  

As I write this 90,000 results have been returned and incoming traffic is up to just over 27 MB per second. My upload queue cleared an hour or so ago.

Out of interest how many results are stored in a row in the database? I am just wondering how many rows are been taken up by result waiting for DB purging there are currently 1,606,679 results waiting to be deleted and the work unit waiting purging is 779,696

As I recall it is one entry per row. Which was the reason limits were needed. It was taking to long to scan the table with "Results out in the field" I believe.

One entry per row is standard for any good database design. 'Result' and 'Row' are pretty much synonyms here.

Scanning won't take much time, because the key fields are indexed - and a database index search is *very* fast (that's what they're optimised for). Retrieving the rest of the record is what takes the time - especially if you have to go to a physical disk for it. I think the staff have said that it makes a huge difference when the whole database is small enough to be held completely in RAM, both tables and indexes.
ID: 1481637 · Report as offensive
Profile HAL9000
Volunteer tester
Avatar

Send message
Joined: 11 Sep 99
Posts: 6534
Credit: 196,805,888
RAC: 57
United States
Message 1481641 - Posted: 25 Feb 2014, 15:00:18 UTC - in response to Message 1481637.  

As I write this 90,000 results have been returned and incoming traffic is up to just over 27 MB per second. My upload queue cleared an hour or so ago.

Out of interest how many results are stored in a row in the database? I am just wondering how many rows are been taken up by result waiting for DB purging there are currently 1,606,679 results waiting to be deleted and the work unit waiting purging is 779,696

As I recall it is one entry per row. Which was the reason limits were needed. It was taking to long to scan the table with "Results out in the field" I believe.

One entry per row is standard for any good database design. 'Result' and 'Row' are pretty much synonyms here.

Scanning won't take much time, because the key fields are indexed - and a database index search is *very* fast (that's what they're optimised for). Retrieving the rest of the record is what takes the time - especially if you have to go to a physical disk for it. I think the staff have said that it makes a huge difference when the whole database is small enough to be held completely in RAM, both tables and indexes.

Having seen many poor db implementations over the years I tend to take caution when assuming things.
In fact I have to work with a system where someone thought it would be a good idea to make a table for groups of users and then stuff all the users into each line with the group as the key. Instead of just adding a group field to the user table.

Perhaps the limit I was thinking about was just once we exceeded the RAM for the db. I did not follow this issue closely after they asked if there were db experts with idea for optimization.
I tend to just know enough to muck things up and then get them back to where they were before I touched it.
SETI@home classic workunits: 93,865 CPU time: 863,447 hours
Join the [url=http://tinyurl.com/8y46zvu]BP6/VP6 User Group[
ID: 1481641 · Report as offensive
Richard Haselgrove Project Donor
Volunteer tester

Send message
Joined: 4 Jul 99
Posts: 14650
Credit: 200,643,578
RAC: 874
United Kingdom
Message 1481648 - Posted: 25 Feb 2014, 15:25:39 UTC - in response to Message 1481641.  

As I write this 90,000 results have been returned and incoming traffic is up to just over 27 MB per second. My upload queue cleared an hour or so ago.

Out of interest how many results are stored in a row in the database? I am just wondering how many rows are been taken up by result waiting for DB purging there are currently 1,606,679 results waiting to be deleted and the work unit waiting purging is 779,696

As I recall it is one entry per row. Which was the reason limits were needed. It was taking to long to scan the table with "Results out in the field" I believe.

One entry per row is standard for any good database design. 'Result' and 'Row' are pretty much synonyms here.

Scanning won't take much time, because the key fields are indexed - and a database index search is *very* fast (that's what they're optimised for). Retrieving the rest of the record is what takes the time - especially if you have to go to a physical disk for it. I think the staff have said that it makes a huge difference when the whole database is small enough to be held completely in RAM, both tables and indexes.

Having seen many poor db implementations over the years I tend to take caution when assuming things.
In fact I have to work with a system where someone thought it would be a good idea to make a table for groups of users and then stuff all the users into each line with the group as the key. Instead of just adding a group field to the user table.

Perhaps the limit I was thinking about was just once we exceeded the RAM for the db. I did not follow this issue closely after they asked if there were db experts with idea for optimization.
I tend to just know enough to muck things up and then get them back to where they were before I touched it.

The basic design is viewable in http://boinc.berkeley.edu/trac/browser/boinc-v2/db/schema.sql, and seems to be reasonably normalised for tables like 'workunit' and 'result'. The BOINC developers do have a tendency to over-use blobs, especially for xml replacements for proper fields, but that doesn't affect our question here.
ID: 1481648 · Report as offensive
Speedy
Volunteer tester
Avatar

Send message
Joined: 26 Jun 04
Posts: 1643
Credit: 12,921,799
RAC: 89
New Zealand
Message 1481672 - Posted: 25 Feb 2014, 21:05:38 UTC - in response to Message 1481648.  

As I write this 90,000 results have been returned and incoming traffic is up to just over 27 MB per second. My upload queue cleared an hour or so ago.

Out of interest how many results are stored in a row in the database? I am just wondering how many rows are been taken up by result waiting for DB purging there are currently 1,606,679 results waiting to be deleted and the work unit waiting purging is 779,696

As I recall it is one entry per row. Which was the reason limits were needed. It was taking to long to scan the table with "Results out in the field" I believe.

One entry per row is standard for any good database design. 'Result' and 'Row' are pretty much synonyms here.

Scanning won't take much time, because the key fields are indexed - and a database index search is *very* fast (that's what they're optimised for). Retrieving the rest of the record is what takes the time - especially if you have to go to a physical disk for it. I think the staff have said that it makes a huge difference when the whole database is small enough to be held completely in RAM, both tables and indexes.

Having seen many poor db implementations over the years I tend to take caution when assuming things.
In fact I have to work with a system where someone thought it would be a good idea to make a table for groups of users and then stuff all the users into each line with the group as the key. Instead of just adding a group field to the user table.

Perhaps the limit I was thinking about was just once we exceeded the RAM for the db. I did not follow this issue closely after they asked if there were db experts with idea for optimization.
I tend to just know enough to muck things up and then get them back to where they were before I touched it.

The basic design is viewable in http://boinc.berkeley.edu/trac/browser/boinc-v2/db/schema.sql, and seems to be reasonably normalised for tables like 'workunit' and 'result'. The BOINC developers do have a tendency to over-use blobs, especially for xml replacements for proper fields, but that doesn't affect our question here.

Thanks for all the answers.
ID: 1481672 · Report as offensive
TBar
Volunteer tester

Send message
Joined: 22 May 99
Posts: 5204
Credit: 840,779,836
RAC: 2,768
United States
Message 1481695 - Posted: 25 Feb 2014, 21:41:02 UTC
Last modified: 25 Feb 2014, 21:41:36 UTC

Weekly Maintenance is over and the AstroPulse Splitters went from Not Running to Disabled. I suppose that may be an improvement, however, my Mac can't tell any difference. Is there any hope of resurrecting the AP Splitters anytime soon?
ID: 1481695 · Report as offensive
Profile Michael W.F. Miles
Avatar

Send message
Joined: 24 Mar 07
Posts: 268
Credit: 34,410,870
RAC: 0
Canada
Message 1481745 - Posted: 26 Feb 2014, 0:40:33 UTC - in response to Message 1481695.  

I sure hope AP comes on line
Fully expected it back up after the outage
This is a waste of volunteers time

Michael
ID: 1481745 · Report as offensive
Profile Sutaru Tsureku
Volunteer tester

Send message
Joined: 6 Apr 07
Posts: 7105
Credit: 147,663,825
RAC: 5
Germany
Message 1481766 - Posted: 26 Feb 2014, 3:53:41 UTC

The AstroPulse splitters comes again online when they checked why they made so much errors.
I was in contact with Eric and he said it is/was not normal.

SETI@home never said "We will feed your PC with tasks 24/7". Or they did?

http://setiathome.berkeley.edu
.. on the left hand:
Keep your computer busy when SETI@home has no work - participate in other BOINC-based projects.


TBar, you have "good luck" that the MAC BOINC don't support the "limit of tasks in progress", so your PC can fill up more tasks than other members with other OS. And you complain nevertheless?
If you are a hardcore SETIzen, make a 2nd OS (Linux, Windows) to your HDD and use it if no AstroPulse tasks are available. Just an idea.
ID: 1481766 · Report as offensive
Profile Geek@Play
Volunteer tester
Avatar

Send message
Joined: 31 Jul 01
Posts: 2467
Credit: 86,146,931
RAC: 0
United States
Message 1481767 - Posted: 26 Feb 2014, 3:54:14 UTC
Last modified: 26 Feb 2014, 4:02:53 UTC

I have been checking my AP work that finally got validated today because somebody finally returned their work. It's easy to see who has been hoarding work as ALL AP work should have been returned by now from a 200 work unit limit. Any outstanding AP work has been hoarded by using various techniques discussed here. Some 56K AP work still out in the field as of now.

Much work is being re-issued because my wingman was late returning work. Great waste of resources just to build up a gigantic cache.

Take a look and take names! Perhaps Berkely will put some meaningful limits that can be enforced by them when work continues. We can always hope!
Boinc....Boinc....Boinc....Boinc....
ID: 1481767 · Report as offensive
Dave Stegner
Volunteer tester
Avatar

Send message
Joined: 20 Oct 04
Posts: 540
Credit: 65,583,328
RAC: 27
United States
Message 1481768 - Posted: 26 Feb 2014, 4:02:02 UTC - in response to Message 1481767.  

I have been checking my AP work that finally got validated today because somebody finally returned their work. It's easy to see who has been hoarding work as ALL AP work should have been returned by now from a 200 work unit limit. Any outstanding AP work has been hoarded by using various techniques discussed here. Some 56K AP work still out in the field as of now.

Much work is being re-issued because my wingman was late returning work. Great waste of resources just to build up a gigantic cache.

Take a look and take names! Perhaps Berkely will put some meaningful limits that can be enforced by them when work continues. We can always hope!



NOT entirely true. NO hoarding going on here. Actually I resent the implication.

I run mostly AP with a 10 day cache. That amounts to about 20-25 tasks per machine. AP has not been down that long, my machines have about 5-6 days left.
Dave

ID: 1481768 · Report as offensive
Profile Michael W.F. Miles
Avatar

Send message
Joined: 24 Mar 07
Posts: 268
Credit: 34,410,870
RAC: 0
Canada
Message 1481771 - Posted: 26 Feb 2014, 4:26:42 UTC

The real pisser is seeing a lot of channels loaded and not knowing why work is not being sent.
We were not told until now of why this is going on
As to hoarding AP work my 200 limit is 5 minutes away from my last WU completing then the well runs dry.

I have no interest in other projects at all.
Thanks for at least telling what is wrong......
ID: 1481771 · Report as offensive
Profile RottenMutt
Avatar

Send message
Joined: 15 Mar 01
Posts: 1011
Credit: 230,314,058
RAC: 0
United States
Message 1481772 - Posted: 26 Feb 2014, 4:27:03 UTC - in response to Message 1481767.  
Last modified: 26 Feb 2014, 4:27:53 UTC

I have been checking my AP work that finally got validated today because somebody finally returned their work. It's easy to see who has been hoarding work as ALL AP work should have been returned by now from a 200 work unit limit. Any outstanding AP work has been hoarded by using various techniques discussed here. Some 56K AP work still out in the field as of now....


i blame the 56k on the slow suns of a female dog! any computer in the top 500 rigs should be excluded from the limits as they stand.
ID: 1481772 · Report as offensive
Profile Geek@Play
Volunteer tester
Avatar

Send message
Joined: 31 Jul 01
Posts: 2467
Credit: 86,146,931
RAC: 0
United States
Message 1481773 - Posted: 26 Feb 2014, 4:34:10 UTC
Last modified: 26 Feb 2014, 4:36:12 UTC

Now I look and see many wu with March deadlines so obviously I was wrong. I failed to realize that many are still content to work 24+ hours on one AP work unit.

And excluding the top 500 computers from the limits would just be wrong!
Boinc....Boinc....Boinc....Boinc....
ID: 1481773 · Report as offensive
TBar
Volunteer tester

Send message
Joined: 22 May 99
Posts: 5204
Credit: 840,779,836
RAC: 2,768
United States
Message 1481774 - Posted: 26 Feb 2014, 4:41:52 UTC - in response to Message 1481768.  
Last modified: 26 Feb 2014, 5:07:45 UTC

I have been checking my AP work that finally got validated today because somebody finally returned their work. It's easy to see who has been hoarding work as ALL AP work should have been returned by now from a 200 work unit limit. Any outstanding AP work has been hoarded by using various techniques discussed here. Some 56K AP work still out in the field as of now.

Much work is being re-issued because my wingman was late returning work. Great waste of resources just to build up a gigantic cache.

Take a look and take names! Perhaps Berkely will put some meaningful limits that can be enforced by them when work continues. We can always hope!



NOT entirely true. NO hoarding going on here. Actually I resent the implication.

I run mostly AP with a 10 day cache. That amounts to about 20-25 tasks per machine. AP has not been down that long, my machines have about 5-6 days left.

Well, some of the people here are not the swiftest. Obviously they don't bother to check a thing, or even give it a thought. My Mac has been working two hosts since Jan 26th, supposedly working 2 groups of 200 is somehow much better than 1 group of 400. I guess I should have made it 3 groups of 200. It is STILL working a BETA APP in hopes a Stock Mac App will be released soon. I suppose you geniuses out there would rather see a Machine working a Beta App just quit? Heck, why not, no one needs to test any new Apps around here, they just pop outta thin air. It has quit BTW, going on a couple days now. No problem for those complainers though, if it doesn't help them then it doesn't matter. Are any of you complainers running any tests on new software out there?

There are plenty of other machines out there running a 10 day cache. I have one still working off a 10 day cache on the CPU, they aren't due until the 7th of Mar. I suppose some people can't think that far ahead though. Oh, my machine working APs on the CPU is Still running the New CUDA App on the GPUs which will eventually be released as x42. I don't remember a whole lotta people offering to test that one either...
ID: 1481774 · Report as offensive
Previous · 1 . . . 20 · 21 · 22 · 23 · 24 · Next

Message boards : Number crunching : Panic Mode On (86) Server Problems?


 
©2024 University of California
 
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.