留言板 :
Number crunching :
The Server Issues / Outages Thread - Panic Mode On! (119)
留言板合理
前 · 1 . . . 105 · 106 · 107 · 108 · 109 · 110 · 后
| 作者 | 消息 |
|---|---|
|
Ville Saari 发送消息 已加入:30 Nov 00 贴子:1123 积分:49,177,052 近期平均积分:82,530
|
While that's been affirmed, I'm not seeing any improvement on the number of items waiting to be satisfied.I don't think any real improvement will happen before whatever problem is preventing the assimilator from assimilating has been found and fixed. |
|
AllgoodGuy 发送消息 已加入:29 May 01 贴子:293 积分:16,348,499 近期平均积分:266
|
I can't help wondering if the splitters are being deliberately throttled in an attempt to reduce the amount of work sitting around in the various queues. After all work not being split will have that effectYes they are, it was stated by Eric that this is being done to try and keep the system within it's RAM limits, I just can't remember where that post was made and whether Eric actually made it or it was passed along ATM. While that's been affirmed, I'm not seeing any improvement on the number of items waiting to be satisfied. |
Wiggo "Democratic Socialist" 发送消息 已加入:24 Jan 00 贴子:18713 积分:261,360,520 近期平均积分:489
|
Thanks Mr Kevvy, I was suddenly very pressed for time just as I started to make that post. :-)...it was stated by Eric that this is being done to try and keep the system within it's RAM limits, I just can't remember where that post was made and whether Eric actually made it or it was passed along ATM.It's actually right in the News forum so visible in the BOINC client. (Edit: of course it was there yesterday and now that I post this it's disappeared lol.) Cheers. |
Mr. Kevvy ![]() 发送消息 已加入:15 May 99 贴子:3202 积分:1,114,826,392 近期平均积分:3,319
|
...it was stated by Eric that this is being done to try and keep the system within it's RAM limits, I just can't remember where that post was made and whether Eric actually made it or it was passed along ATM. It's actually right in the News forum so visible in the BOINC client. (Edit: of course it was there yesterday and now that I post this it's disappeared lol.)
|
Wiggo "Democratic Socialist" 发送消息 已加入:24 Jan 00 贴子:18713 积分:261,360,520 近期平均积分:489
|
I can't help wondering if the splitters are being deliberately throttled in an attempt to reduce the amount of work sitting around in the various queues. After all work not being split will have that effectYes they are, it was stated by Eric that this is being done to try and keep the system within it's RAM limits, I just can't remember where that post was made and whether Eric actually made it or it was passed along ATM. Cheers. |
rob smith ![]() 发送消息 已加入:7 Mar 03 贴子:18752 积分:416,307,556 近期平均积分:380
|
I can't help wondering if the splitters are being deliberately throttled in an attempt to reduce the amount of work sitting around in the various queues. After all work not being split will have that effect Bob Smith Member of Seti PIPPS (Pluto is a Planet Protest Society) Somewhere in the (un)known Universe? |
Jimbocous ![]() 发送消息 已加入:1 Apr 13 贴子:1849 积分:268,616,081 近期平均积分:1,349
|
That's my experience, too. I now have two machines in the 'high RAC' category (top 100): they were both completely dry yesterday morning. I did a little Einstein backup work while the servers were sorting themselves out, but once work started flowing, I ramped them up gently by requesting an hour of work at a time (0.05 days) and increasing the cache a step at a time as they filled up. Reached full cache by evening, with just a little tweak any time I happened to be passing.I'm still convinced that somehow, whether it be intent or just net result, the higher your RAC is the lower you are in the priority stack in terms of actually getting work during a recovery from outage. This is entirely too consistent to be the luck of the draw.Maybe related, but I think it is more to do with how much work the host requests. Sounds like a reality, not "an illusion". Main cruncher cache here is around
|
|
AllgoodGuy 发送消息 已加入:29 May 01 贴子:293 积分:16,348,499 近期平均积分:266
|
Validation Pending still steadily growing, looks like around 23 million objects waiting to be satisfied. Still getting work though, despite the RTS showing a pretty steady 0. I even fell asleep in the wrong configuration night before last to decrease my Pending column below normal average, but I'm well over that again. This poor system needs a break. |
Richard Haselgrove ![]() 发送消息 已加入:4 Jul 99 贴子:14141 积分:200,643,578 近期平均积分:874
|
That's my experience, too. I now have two machines in the 'high RAC' category (top 100): they were both completely dry yesterday morning. I did a little Einstein backup work while the servers were sorting themselves out, but once work started flowing, I ramped them up gently by requesting an hour of work at a time (0.05 days) and increasing the cache a step at a time as they filled up. Reached full cache by evening, with just a little tweak any time I happened to be passing.I'm still convinced that somehow, whether it be intent or just net result, the higher your RAC is the lower you are in the priority stack in terms of actually getting work during a recovery from outage. This is entirely too consistent to be the luck of the draw.Maybe related, but I think it is more to do with how much work the host requests. |
W-K 666 ![]() 发送消息 已加入:18 May 99 贴子:13873 积分:40,757,560 近期平均积分:67
|
I'm still convinced that somehow, whether it be intent or just net result, the higher your RAC is the lower you are in the priority stack in terms of actually getting work during a recovery from outage. This is entirely too consistent to be the luck of the draw. Maybe related, but I think it is more to do with how much work the host requests. When I get up Wednesday mornings, UTC times rule in the UK winter, if the computer hasn't started receiving work, I set the cache to a very low level. I find that usually works after a few attempts, and as I receive work, I increase the cache in steps up to 0.6 days which, unless the servers give me oddles of AP, fills the GPU cache to 150 tasks. Also, here are some numbers on tasks downloaded and validated in the 24 hrs since 08:06:31 26th Feb, ~24hours ago. After 12 hours at ~20:00 26th Downloaded - 345; In Progress 150; Valid 86 Processed = 345 - 150 = 195 Percentage of tasks downloaded and Validated in 12 hours = 100 * 86 / 195 = 44.1% After 4 hours at ~08:00 27th Downloaded - 523; In Progress 150; Valid 253 Processed = 523 - 150 = 373 Percentage of tasks downloaded and Validated in 24 hours = 100 * 253 / 373 = 67.8% I only crunch on the GPU so it is fairly simple just to scroll through the pages and count each page then add up the page numbers. edit] Prior to 08:06 yesterday the Seti cache was empty. |
|
Ville Saari 发送消息 已加入:30 Nov 00 贴子:1123 积分:49,177,052 近期平均积分:82,530
|
Could it be that the assimilation process is the problem.There has clearly been some problem in assimilation for the last several weeks, but the problem can be in many different places. It could be the throughput of the boinc database that somehow hits the assimilator harder than the other processes. Or it could be a problem in the assimilator program itself. Or it can be the throughput of the science databases. Or the throughput of the upload filesystem where the result files the assimilator needs to read are. |
|
Ville Saari 发送消息 已加入:30 Nov 00 贴子:1123 积分:49,177,052 近期平均积分:82,530
|
I'm still convinced that somehow, whether it be intent or just net result, the higher your RAC is the lower you are in the priority stack in terms of actually getting work during a recovery from outage. This is entirely too consistent to be the luck of the draw.It is an illusion. Everyone has the same priority but higher your RAC, the more successful scheduler request you need to keep your cache not depleting. If every 12th request wins the lottery and gets some work, then you get some work once every hour and this may be all that a slow host needs to refill its cache to the brim but nowhere near the one hour production of a fast host. |
W-K 666 ![]() 发送消息 已加入:18 May 99 贴子:13873 积分:40,757,560 近期平均积分:67
|
Could it be that the assimilation process is the problem. How difficult is it to translate the data we produce and all the other details necessary and put onto the science database. this is what the Server Status page says; sah_assimilator/ap_assimilator : Takes scientific data from validated results and puts them in the SETI@home (or Astropulse) database for later analysis. |
petri33 发送消息 已加入:6 Jun 02 贴子:1668 积分:623,086,772 近期平均积分:156
|
I'm still convinced that somehow, whether it be intent or just net result, the higher your RAC is the lower you are in the priority stack in terms of actually getting work during a recovery from outage. This is entirely too consistent to be the luck of the draw. +1 To overcome Heisenbergs: "You can't always get what you want / but if you try sometimes you just might find / you get what you need." -- Rolling Stones |
Jimbocous ![]() 发送消息 已加入:1 Apr 13 贴子:1849 积分:268,616,081 近期平均积分:1,349
|
I'm still convinced that somehow, whether it be intent or just net result, the higher your RAC is the lower you are in the priority stack in terms of actually getting work during a recovery from outage. This is entirely too consistent to be the luck of the draw.
|
|
Boiler Paul 发送消息 已加入:4 May 00 贴子:232 积分:4,965,771 近期平均积分:64
|
and, of course, after I post, I receive work! |
|
Boiler Paul 发送消息 已加入:4 May 00 贴子:232 积分:4,965,771 近期平均积分:64
|
work can be hard to come by. all I've gotten over the past few hours is the Project has no tasks available in the log. Just need to be patient |
|
TBar 发送消息 已加入:22 May 99 贴子:5204 积分:840,779,836 近期平均积分:2,768
|
That's what you get for assuming, Since making that post the machine that had been run out of work now has 400 tasks instead of Zero. I had absolutely nothing to do with it. |
W-K 666 ![]() 发送消息 已加入:18 May 99 贴子:13873 积分:40,757,560 近期平均积分:67
|
Now the 2nd out of 3 machines has run Out of Work, https://setiathome.berkeley.edu/results.php?hostid=6796479 I can only assume it is your problem. I've had very few problems since 08:00 26th UTC. |
|
TBar 发送消息 已加入:22 May 99 贴子:5204 积分:840,779,836 近期平均积分:2,768
|
Now the 2nd out of 3 machines has run Out of Work, https://setiathome.berkeley.edu/results.php?hostid=6796479 That leaves 1 machine still working. I suppose when that one runs Out of Work I'll just shut every thing down and brag about how much money I'm saving on electricity. |
©2020 University of California
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.