Tuesday Twinge (Sep 02 2008)

Message boards : Technical News : Tuesday Twinge (Sep 02 2008)
Message board moderation

To post messages, you must log in.

AuthorMessage
Profile Matt Lebofsky
Volunteer moderator
Project administrator
Project developer
Project scientist
Avatar

Send message
Joined: 1 Mar 99
Posts: 1444
Credit: 957,058
RAC: 0
United States
Message 804356 - Posted: 2 Sep 2008, 22:16:36 UTC

Currently as I write this we're recovering from the weekly outage (during which we take care of database backups and other sundry server details). It may take a while...

This past Friday we overloaded our science database trying to create a new index. A database engine restart solved the problem, but not after choking the whole local network. As mentioned in many posts past, we're strangely sensitive to heavy network bandwidth (I think due to linux's imperfect handling of NFS dropouts), and such periods cause random unexpected events. This time, for example, the bottleneck from the primary science database server ultimately caused the BOINC/mysql replica server to disconnect from the master. So the replica fell behind all weekend. Sigh. Instead of actually letting it catch up we're just re-mirroring it from the master as we just backed it up this morning.

Meanwhile, we're out of space again on the workunit server, and with no fast/easy way to add space. Eric's playing with the splitter mix to reduce the number of Astropulse workunits being generated (they are much larger than SETI@home workunits). Maybe that will help, but not immediately. This is what's mostly causing our headaches today as we can't create enough work to keep up with demand.

- Matt

-- BOINC/SETI@home network/web/science/development person
-- "Any idiot can have a good idea. What is hard is to do it." - Jeanne-Claude
ID: 804356 · Report as offensive
Profile Dr. C.E.T.I.
Avatar

Send message
Joined: 29 Feb 00
Posts: 16019
Credit: 794,685
RAC: 0
United States
Message 804358 - Posted: 2 Sep 2008, 22:24:53 UTC


. . . Thanks - as usual Matt - for your Updates - iT is certainly appreciated Sir!

> Eric - Best of Luck w/ the splitter mix - looks like another Servers required for the Astropulse Units eh . . .

< Accolades to Each of You @ Berkeley


BOINC Wiki . . .

Science Status Page . . .
ID: 804358 · Report as offensive
Profile JDWhale
Volunteer tester
Avatar

Send message
Joined: 6 Apr 99
Posts: 921
Credit: 21,935,817
RAC: 3
United States
Message 804366 - Posted: 2 Sep 2008, 22:38:03 UTC - in response to Message 804356.  

Meanwhile, we're out of space again on the workunit server, and with no fast/easy way to add space. Eric's playing with the splitter mix to reduce the number of Astropulse workunits being generated (they are much larger than SETI@home workunits). Maybe that will help, but not immediately. This is what's mostly causing our headaches today as we can't create enough work to keep up with demand.

- Matt


Maybe if Astropulse paid more, there'd be more demand for its WUs... I'm about to finish up an AP WU sent to me 3 weeks ago, but it's worth less when it was issued to me than now... This will not thrill my new wingman who is claiming more credit than me. In the economy this is call deflation :-( WUs should be worth more over time rather than less... There is no investment value if the claim doesn't increase over time with new wingmen!

How can the same WU be worth more credit just because it had to be reissued?

Maybe all users should cancel their AP WUs so the WUs will increase in value!
ID: 804366 · Report as offensive
OzzFan Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 9 Apr 02
Posts: 15691
Credit: 84,761,841
RAC: 28
United States
Message 804391 - Posted: 3 Sep 2008, 0:13:35 UTC - in response to Message 804366.  

Meanwhile, we're out of space again on the workunit server, and with no fast/easy way to add space. Eric's playing with the splitter mix to reduce the number of Astropulse workunits being generated (they are much larger than SETI@home workunits). Maybe that will help, but not immediately. This is what's mostly causing our headaches today as we can't create enough work to keep up with demand.

- Matt


Maybe if Astropulse paid more, there'd be more demand for its WUs...


I don't think that's what Matt meant, JD. I think he meant that they can't create enough work in general (both SETI & AstroPulse) to keep up with demand. There were a couple reports over in Q&A of users being unable to get more SETI work. The limiting factor is that both SETI and AP are using the same disk (or set of disks) and the drives are filling up fast making it impossible to create new work.
ID: 804391 · Report as offensive
DJStarfox

Send message
Joined: 23 May 01
Posts: 1066
Credit: 1,226,053
RAC: 2
United States
Message 804405 - Posted: 3 Sep 2008, 0:56:12 UTC - in response to Message 804356.  

This time, for example, the bottleneck from the primary science database server ultimately caused the BOINC/mysql replica server to disconnect from the master. So the replica fell behind all weekend.


I'm not sure if anyone has mentioned this.... But did you try installing separate NIC cards in the database and replica servers? You could either 1) bundle them to make one faster NIC, or 2) dedicate the secondary NICs as DB replica traffic only (using private subnet).
ID: 804405 · Report as offensive
Profile Henk Haneveld
Volunteer tester

Send message
Joined: 16 May 99
Posts: 154
Credit: 1,577,293
RAC: 1
Netherlands
Message 804490 - Posted: 3 Sep 2008, 9:45:51 UTC - in response to Message 804356.  


Meanwhile, we're out of space again on the workunit server, and with no fast/easy way to add space. Eric's playing with the splitter mix to reduce the number of Astropulse workunits being generated (they are much larger than SETI@home workunits). Maybe that will help, but not immediately. This is what's mostly causing our headaches today as we can't create enough work to keep up with demand.


I think this is the wrong way around.

1 AP result file is about the size of 22 MB result files.
But to process 1 AP result takes longer then 22 MB results.

To reduce the need for diskspace you need the get more AP results to the clients not less.
ID: 804490 · Report as offensive
Profile ML1
Volunteer moderator
Volunteer tester

Send message
Joined: 25 Nov 01
Posts: 21788
Credit: 7,508,002
RAC: 20
United Kingdom
Message 804547 - Posted: 3 Sep 2008, 14:15:51 UTC - in response to Message 804356.  
Last modified: 3 Sep 2008, 14:17:15 UTC

... we overloaded our science database trying to create a new index. A database engine restart solved the problem, but not after choking the whole local network. As mentioned in many posts past, we're strangely sensitive to heavy network bandwidth (I think due to linux's imperfect handling of NFS dropouts), and such periods cause random unexpected events. ... server to disconnect from the master...

Is the real overload due to packet fragmentation, or lost packets due to congestion on one of your switches or server NICs?... NFS doesn't handle packet loss very well. Eventually, things time out but not before descending into a slow death...

Excessive fragmentation due to an MTU=1500? Can you use jumbo packets for that network? Or at least increase the MTU for a subnet of machines?

Too small kernel buffers for the network packet assembly?

Are the switches getting their routing tables poisoned from noise traffic from elsewhere?

Note: Optimizing NFS Performance

5.3. Overflow of Fragmented Packets

Using an rsize or wsize larger than your network's MTU (often set to 1500, in many networks) will cause IP packet fragmentation...


There's a few good comments in there.


Hope that helps,

Regards,
Martin
See new freedom: Mageia Linux
Take a look for yourself: Linux Format
The Future is what We all make IT (GPLv3)
ID: 804547 · Report as offensive
Profile KWSN THE Holy Hand Grenade!
Volunteer tester
Avatar

Send message
Joined: 20 Dec 05
Posts: 3187
Credit: 57,163,290
RAC: 0
United States
Message 804598 - Posted: 3 Sep 2008, 18:59:27 UTC - in response to Message 804490.  
Last modified: 3 Sep 2008, 19:01:46 UTC


Meanwhile, we're out of space again on the workunit server, and with no fast/easy way to add space. Eric's playing with the splitter mix to reduce the number of Astropulse workunits being generated (they are much larger than SETI@home workunits). Maybe that will help, but not immediately. This is what's mostly causing our headaches today as we can't create enough work to keep up with demand.


I think this is the wrong way around.

1 AP result file is about the size of 22 MB result files.
But to process 1 AP result takes longer then 22 MB results.

To reduce the need for diskspace you need the get more AP results to the clients not less.



Your typical result file for MB is 23-35Kb, not Mb, your typical AP result is about 45 Kb... but a MB WU is 350Kb, and an AP WU is 8Mb - and they have to store the WU, so that if someone misses deadline, they can send it out again.

MB = Multi-Beam
AP = Astro Pulse
.

Hello, from Albany, CA!...
ID: 804598 · Report as offensive
Profile Henk Haneveld
Volunteer tester

Send message
Joined: 16 May 99
Posts: 154
Credit: 1,577,293
RAC: 1
Netherlands
Message 804616 - Posted: 3 Sep 2008, 20:43:40 UTC - in response to Message 804598.  


Meanwhile, we're out of space again on the workunit server, and with no fast/easy way to add space. Eric's playing with the splitter mix to reduce the number of Astropulse workunits being generated (they are much larger than SETI@home workunits). Maybe that will help, but not immediately. This is what's mostly causing our headaches today as we can't create enough work to keep up with demand.


I think this is the wrong way around.

1 AP result file is about the size of 22 MB result files.
But to process 1 AP result takes longer then 22 MB results.

To reduce the need for diskspace you need the get more AP results to the clients not less.



Your typical result file for MB is 23-35Kb, not Mb, your typical AP result is about 45 Kb... but a MB WU is 350Kb, and an AP WU is 8Mb - and they have to store the WU, so that if someone misses deadline, they can send it out again.

MB = Multi-Beam
AP = Astro Pulse


I may have not be clear enough. I was talking about the 350 Kb and 8 Mb files.
8 Mb is about 22 * 350 Kb.

So my point still stands that it is beter to send out more AP.
ID: 804616 · Report as offensive
Profile Matt Lebofsky
Volunteer moderator
Project administrator
Project developer
Project scientist
Avatar

Send message
Joined: 1 Mar 99
Posts: 1444
Credit: 957,058
RAC: 0
United States
Message 804638 - Posted: 3 Sep 2008, 21:59:59 UTC

Looks like more erratic behavior thanks to the bottlenecks du jour - the lack of the workunit storage and general NFS malaise. This will all clear up sooner or later. The NFS problems will always be there until we get about 50% more servers and shut everything down for a week to reorganize all the services and storage in such a manner that each system is only doing one thing. Until then server A is mounting server B for reason X, and server C is mounting server A for reason Y, etc. Anyway, don't see that getting better any time soon.

- Matt

-- BOINC/SETI@home network/web/science/development person
-- "Any idiot can have a good idea. What is hard is to do it." - Jeanne-Claude
ID: 804638 · Report as offensive
Profile squishymaster
Avatar

Send message
Joined: 21 Oct 03
Posts: 34
Credit: 784,496
RAC: 0
United States
Message 804727 - Posted: 4 Sep 2008, 3:46:15 UTC - in response to Message 804638.  

Are you talking about having more actual servers or you just need more storage space? How much would something like that cost?
ID: 804727 · Report as offensive
Profile ML1
Volunteer moderator
Volunteer tester

Send message
Joined: 25 Nov 01
Posts: 21788
Credit: 7,508,002
RAC: 20
United Kingdom
Message 804816 - Posted: 4 Sep 2008, 11:28:43 UTC - in response to Message 804727.  
Last modified: 4 Sep 2008, 11:29:51 UTC

Are you talking about having more actual servers or you just need more storage space? How much would something like that cost?

I would expect both, and then also the time needed to plan and set it all up...

A short term help-fix could be to just install additional Gbit NICs in whichever servers and then wire direct point-point ethernet connections for exclusive use for the NFS between each pair of servers. Set fixed IPs and set the routing tables and viola! (You can also set a jumbo MTU=9660 to avoid all fragmentation. BIG performance boost for 8k NFS read/writes!)


Good luck,

Regards,
Martin

(You don't even need cross-over cables, the Gbit NICs will correctly autonegotiate for a normal patch lead.)
See new freedom: Mageia Linux
Take a look for yourself: Linux Format
The Future is what We all make IT (GPLv3)
ID: 804816 · Report as offensive
Profile WimTea
Volunteer tester

Send message
Joined: 15 Feb 02
Posts: 34
Credit: 909,865
RAC: 0
Netherlands
Message 804860 - Posted: 4 Sep 2008, 13:44:03 UTC

A short term help-fix could be to just install additional Gbit NICs in whichever servers and then wire direct point-point ethernet connections for exclusive use for the NFS between each pair of servers. Set fixed IPs and set the routing tables and viola! (You can also set a jumbo MTU=9660 to avoid all fragmentation. BIG performance boost for 8k NFS read/writes!)


You could even expand this idea by throwing in a small Gb switch and a few more NICS and slowly migrate all NFS links to connect over this separate network. Sort of a distributed NAS, with bonusses for offloading the existing network which handles the client traffic and slightly reducing the overall server load by reducing the serverbased network handling (less fragmentation and less resends).
ID: 804860 · Report as offensive
Profile ML1
Volunteer moderator
Volunteer tester

Send message
Joined: 25 Nov 01
Posts: 21788
Credit: 7,508,002
RAC: 20
United Kingdom
Message 804888 - Posted: 4 Sep 2008, 16:32:17 UTC - in response to Message 804860.  

A short term help-fix could be to just install additional Gbit NICs in whichever servers and then wire direct point-point ethernet connections for exclusive use for the NFS between each pair of servers. Set fixed IPs and set the routing tables and viola! (You can also set a jumbo MTU=9660 to avoid all fragmentation. BIG performance boost for 8k NFS read/writes!)


You could even expand this idea by throwing in a small Gb switch and a few more NICS and slowly migrate all NFS links to connect over this separate network. Sort of a distributed NAS, with bonusses for offloading the existing network which handles the client traffic and slightly reducing the overall server load by reducing the serverbased network handling (less fragmentation and less resends).

Nope. Poor idea. That gives little advantage over their existing Gbit switch. Note that the internet traffic on that switch maxes out taking about 8% capacity for bandwidth. However, it might be running out of routing/persistency table space...

The whole idea of a point-to-point link (NO intermediate switch) is to guarantee no packet loss due to switch overload or due to congestion on any one NIC.

The more immediate question is whether there are any spare IO slots for additional NICs... And whether Matt is game for more cuts and bruises in giving it a try...

Regards,
Martin

See new freedom: Mageia Linux
Take a look for yourself: Linux Format
The Future is what We all make IT (GPLv3)
ID: 804888 · Report as offensive
Profile Matt Lebofsky
Volunteer moderator
Project administrator
Project developer
Project scientist
Avatar

Send message
Joined: 1 Mar 99
Posts: 1444
Credit: 957,058
RAC: 0
United States
Message 804889 - Posted: 4 Sep 2008, 16:34:29 UTC - in response to Message 804816.  


A short term help-fix could be to just install additional Gbit NICs in whichever servers and then wire direct point-point ethernet connections for exclusive use for the NFS between each pair of servers. Set fixed IPs and set the routing tables and viola! (You can also set a jumbo MTU=9660 to avoid all fragmentation. BIG performance boost for 8k NFS read/writes!)


The NICs would help, but the greater issue is that each server is doing multiple things, and hosting all kinds of files to all other servers. Basically every machine is mounting each other's disks, so if one system goes down all kinds of random hell breaks loose. Things like informix esql libraries, data sets, upload directories, download directories, home accounts, pieces of web sites...

Of course we'd like to clean this up, but that would require a lot more servers doing one specific thing (which means more servers *and* space *and* time to set all this up *and* move things safely to their new locations *and* make sure services haven't broken in the meantime).

- Matt
-- BOINC/SETI@home network/web/science/development person
-- "Any idiot can have a good idea. What is hard is to do it." - Jeanne-Claude
ID: 804889 · Report as offensive
Profile WimTea
Volunteer tester

Send message
Joined: 15 Feb 02
Posts: 34
Credit: 909,865
RAC: 0
Netherlands
Message 804934 - Posted: 4 Sep 2008, 19:18:19 UTC

Note that the internet traffic on that switch maxes out taking about 8% capacity for bandwidth.

OK, this low load makes my suggestion a bad one.

... *and* ... *and* ... *and* ... *and* ...

Sounds like money (also as in 'time = ...') is more of an issue than anything else. That's science and the shoestring budgets that so often seem to come with it I guess...
ID: 804934 · Report as offensive

Message boards : Technical News : Tuesday Twinge (Sep 02 2008)


 
©2025 University of California
 
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.