Work buffer cache settings don't seem to work

Questions and Answers : Preferences : Work buffer cache settings don't seem to work
Message board moderation

To post messages, you must log in.

AuthorMessage
Jon Rook

Send message
Joined: 9 Jan 02
Posts: 34
Credit: 6,679,310
RAC: 0
United States
Message 1328076 - Posted: 16 Jan 2013, 21:11:37 UTC

Hi,

I've been running S@H for a while now and can't figure something out. I recently changed both my online preferences and my local computing preferences to specify a minimum work buffer of 10 days. I had previously had this setting at .1 day which seems to be the default. I also specify 10 days for the Max additional work buffer. This second option has always been set to 10 days for me.

I made the change because the number of work units that were being cached for me was insufficient for more than about a 2 day downtime of the SETI servers. The problem is that since I made the change, I see no difference in the number of work units in my local work buffer. My machine is regularly asking for more work units, but I get messages that say "The computer has reached a limit on tasks in progress."

I ran out of CPU work units again during the most recent outage on 1/14-1/15. Thus, it doesn't seem that this option works. Does anyone else see this? Is there anything I can do to make it work, or is this a S@H restriction?

Thanks in advance,
Jon



ID: 1328076 · Report as offensive
rob smith Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Volunteer moderator
Volunteer tester

Send message
Joined: 7 Mar 03
Posts: 22160
Credit: 416,307,556
RAC: 380
United Kingdom
Message 1328080 - Posted: 16 Jan 2013, 21:27:09 UTC - in response to Message 1328076.  

Just about everyone with a even a moderately fast processor ran out of work during the most recent outage.
There are currently limits set on the number of work units you can have "in progress" - 100 per CPU (not cpu core) and 100 per GPU. You are bouncing off that limit.

So there is little point in "playing" with your cache settings. It would be better to reduce them to 3/0.1 otherwise you find you are starved of work on a very regular basis due to some changes in the way the cache works.
Bob Smith
Member of Seti PIPPS (Pluto is a Planet Protest Society)
Somewhere in the (un)known Universe?
ID: 1328080 · Report as offensive
Profile Jord
Volunteer tester
Avatar

Send message
Joined: 9 Jun 99
Posts: 15184
Credit: 4,362,181
RAC: 3
Netherlands
Message 1328082 - Posted: 16 Jan 2013, 21:31:15 UTC - in response to Message 1328076.  
Last modified: 16 Jan 2013, 21:31:40 UTC

Well, it's been in the News:
Continued server problems.
We're continuing to have issues due to a database problem early last week and a botched attempt to fix it.

The problem is that the result and host tables in the database have grown large enough, and hosts have gotten fast enough that the lookup of result in process for a host and the enumeration of new results to send don't finish before the web connection times out either on the server or the client side. This resulted in hosts being assigned large number or results to compute without the transaction that tells them about these results being completed. The host. think it received no results would then contact the server for more results, which it would again not receive.

This isn't a hardware problem. The database currently fits in memory and the processors are fast. We've just crossed a threshold where each host computes fast enough that host queues and the result table have become large enough to cause this problem. To solve it, we've put per host limits on results in process back in place. But hosts that are having this problem will probably continue to have it until the average number of results per host has fallen to a workable level. That could take weeks.

For a more permanent fix, we plan do more work in each result by quadrupling the size of the workunits. But that fix will probably take months to implement and test.

The limits are 100 tasks per CPU (not per core, but per actual CPU) and I think 200 per GPU.
ID: 1328082 · Report as offensive
rob smith Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Volunteer moderator
Volunteer tester

Send message
Joined: 7 Mar 03
Posts: 22160
Credit: 416,307,556
RAC: 380
United Kingdom
Message 1328088 - Posted: 16 Jan 2013, 22:11:10 UTC

I can vouch for the 100 per CPU - I've been bouncing off that limit for weeks, and am still doing so.
Bob Smith
Member of Seti PIPPS (Pluto is a Planet Protest Society)
Somewhere in the (un)known Universe?
ID: 1328088 · Report as offensive
Jon Rook

Send message
Joined: 9 Jan 02
Posts: 34
Credit: 6,679,310
RAC: 0
United States
Message 1328276 - Posted: 17 Jan 2013, 15:40:36 UTC

Thanks!

I do remember seeing that news post, but I didn't associate it with this problem. Looks like we have to wait for the permanent fix.

Regards,
J



ID: 1328276 · Report as offensive

Questions and Answers : Preferences : Work buffer cache settings don't seem to work


 
©2024 University of California
 
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.