Mono (Feb 25 2009)


log in

Advanced search

Message boards : Technical News : Mono (Feb 25 2009)

Author Message
Profile Matt Lebofsky
Volunteer moderator
Project administrator
Project developer
Project scientist
Avatar
Send message
Joined: 1 Mar 99
Posts: 1389
Credit: 74,079
RAC: 0
United States
Message 869467 - Posted: 25 Feb 2009, 22:48:42 UTC

It looked like we got beyond the current deluge without too much intervention. Good. Then our bandwidth spiked again. Bad. But then it recovered once more. Good. Oh well, whatever. We're still just in "wait and see if it gets better on its own" mode around here - if we hit our bandwidth limits (and we understand why) there's not much else we can do.

Spent a chunk of the day tracking down current donation processing issues. What a pain. I really need to document the whole crazy donation system so other people around here can fix these problems when they arise. Maybe I'll do that later today. Other than that, just some data pipeline/sysadmin type stuff.

A note about the server status page: Every 10 minutes a BOINC script runs which does several things including: 1. start/restart servers that aren't running but should be, and 2. run a bunch of "task" scripts, like the one that generates the server status page. Since this status page script runs once every ten minutes, it is only a snapshot in time - not a continuum. It also could take several minutes to run its course, as it is scanning many heavily loaded servers. So the data towards the top of the page is representative of a minute or two earlier than the data towards the bottom. And server processes, like ap_validator, hiccup from time to time and get restarted every 10 minutes, then maybe process a few hundred workunits, but fail again a second before the status page checks its status. So even though it was running the past couple of minutes it shows up as "Not Running." In short, don't trust anything on that page at first glance.

- Matt

____________
-- BOINC/SETI@home network/web/science/development person
-- "Any idiot can have a good idea. What is hard is to do it." - Jeanne-Claude

gomeyer
Volunteer tester
Send message
Joined: 21 May 99
Posts: 488
Credit: 50,157,953
RAC: 0
United States
Message 869474 - Posted: 25 Feb 2009, 23:03:21 UTC

I imagine this is pretty low on your priority list, but if you find a minute (ha) could you make sure the stats xml's are running? It looks as if they have not done so since the 23rd.

And as always, thanks for taking time to keep us informed.

Profile Mike O
Avatar
Send message
Joined: 1 Sep 07
Posts: 428
Credit: 6,670,998
RAC: 0
United States
Message 869586 - Posted: 26 Feb 2009, 3:11:07 UTC

Hello Matt...
I'm still stuck on this issue about bandwidth... my bad.

Here's a crazy idea.
What about compressing the WU's after being created and sending then in that state. When the clients get the data, they could decompress before crunching. Afterwards, compress the results for the trip home.
I know.. more work recoding the apps but... even a 1/10 ratio effectively multiplies the bandwidth by 10. Another issue is, it would put even more burdon on the closets computers. With all the different version of the core clients, some kind of bootstrap (if thats the right term) would need to be created to sit between BOINC and the core client (if thats even possible).


Thanks for making my computers busy again. HOORAH!

Wolverine
Avatar
Send message
Joined: 9 Jan 00
Posts: 35
Credit: 7,349,242
RAC: 1,273
Canada
Message 869597 - Posted: 26 Feb 2009, 4:22:05 UTC

Mike O

Refer to this post on that subject.. (I beat ya to it)... 8-)


http://setiathome.berkeley.edu/forum_thread.php?id=52213


- Wol
____________

John McLeod VII
Volunteer developer
Volunteer tester
Avatar
Send message
Joined: 15 Jul 99
Posts: 24384
Credit: 519,750
RAC: 37
United States
Message 869598 - Posted: 26 Feb 2009, 4:26:09 UTC - in response to Message 869586.

Hello Matt...
I'm still stuck on this issue about bandwidth... my bad.

Here's a crazy idea.
What about compressing the WU's after being created and sending then in that state. When the clients get the data, they could decompress before crunching. Afterwards, compress the results for the trip home.
I know.. more work recoding the apps but... even a 1/10 ratio effectively multiplies the bandwidth by 10. Another issue is, it would put even more burdon on the closets computers. With all the different version of the core clients, some kind of bootstrap (if thats the right term) would need to be created to sit between BOINC and the core client (if thats even possible).


Thanks for making my computers busy again. HOORAH!

Are you certain that this is not already being done? There is a facility built into BOINC to do precisely this.
____________


BOINC WIKI

Message boards : Technical News : Mono (Feb 25 2009)

Copyright © 2014 University of California