Message boards :
Number crunching :
SETI loves me.
Message board moderation
Previous · 1 · 2
Author | Message |
---|---|
soft^spirit Send message Joined: 18 May 99 Posts: 6497 Credit: 34,134,168 RAC: 0 |
CPU tasks: 158 (158 VLAR, 0 VHAR) It is obvious other blocks have been put in place. fully loaded on 3 days: I had enough work for approx 2 days (over if you count other projects and CPU in the math, CPU does far less than low end GPU). I raised it to 6 days. The amount of work I had on hand DID NOT CHANGE.. and it still refused to request more. Including after reboot. so basically, we have badly written limits(combining CPU and GPU) .. suffered a work around ADDITIONAL badly written limit(20 units only) all weekend.. Compounded with a replica data base crash.. and now we are back to the original badly written limit. The chance of finding Extra terrestrial intelligence seems dismal in light of our efforts to find terrestrial inteligence. Janice |
perryjay Send message Joined: 20 Aug 02 Posts: 3377 Credit: 20,676,751 RAC: 0 |
Soft^spirit, The separate DCFs were supposed to set the right timing on each different type of WU. By the time they found out it wasn't working they didn't have time to get it fixed so they went back to the old way with just one DCF for all. This also isn't working but.... Hopefully something can be worked out to get the separate DCFs to work right and get the times down to what they should be. Once that is done, we should all get the correct amount of work and things will go much smoother. PROUD MEMBER OF Team Starfire World BOINC |
soft^spirit Send message Joined: 18 May 99 Posts: 6497 Credit: 34,134,168 RAC: 0 |
well perry, I will believe it when I see it. But at least before Jeff's thread got hijacked he had at least read and seemed to understand.. I am hoping he understands it was a problem in the previous limits as well. So I am not without hope. Still a bit dismayed at this raw test bed we seem to be used for... And of course being a former technician.... I am a natural enemy of anyone from engineering. ;) Janice |
SciManStev Send message Joined: 20 Jun 99 Posts: 6652 Credit: 121,090,076 RAC: 0 |
well perry, I will believe it when I see it. But at least before Jeff's thread got hijacked he had at least read and seemed to understand.. I am hoping he understands it was a problem in the previous limits as well. Uh Oh! I'm an engineer! Back when I was a technician, I could really relate to that. I found things that were next to impossible to repair or even wire up. Now as a software engineer I use what I learn, and listen to those I am programming for. When I get requests I take them very seriously, and do my best to implement them as long as they do not compromise the end goal. :) Steve Warning, addicted to SETI crunching! Crunching as a member of GPU Users Group. GPUUG Website |
soft^spirit Send message Joined: 18 May 99 Posts: 6497 Credit: 34,134,168 RAC: 0 |
I have seen hundreds of thousands of dollars in engineered hardware.. that could have been done for several thousand... if they had talked to a tech first. And it would have worked better. And massive upgrades that would have paid off 10 times.. if anyone had listened to the "cheap seats"... But they saved $10 on something like a cable. So instead it took 6 months longer to get going... and never got fully utilized. Could I have designed those systems? Nope. But I could walk up to them and in 5 minutes ask a few questions that could have saved thousands... if they had not already ordered everything. Which always made the engineers very angry. Janice |
SciManStev Send message Joined: 20 Jun 99 Posts: 6652 Credit: 121,090,076 RAC: 0 |
I totally understand what you are saying. I am trying to get approval to automate a temperature chamber. It will cost about $100,000. I have already developed the root structure in a massive database, and automated test stands, and we would be able to test 12 units at one time, rather than one at a time. I have been trying to do this for almost 8 years, and I'm finally getting closer. My system has already had a major impact on customer satisfaction and unifying test methods. I have only been an engineer for a year, and a technician for 30 years at least. My company makes force feedback servo inclinometers and accelerometers, as well as other products. Edit: I have always worked closely with our engineers, as close communication is how things get done. Steve Warning, addicted to SETI crunching! Crunching as a member of GPU Users Group. GPUUG Website |
hiamps Send message Joined: 23 May 99 Posts: 4292 Credit: 72,971,319 RAC: 0 |
CPU tasks: 158 (158 VLAR, 0 VHAR) Your DCF probably screwed you, I had to set mine to .01 this morning a few times but I got lots of work. Official Abuser of Boinc Buttons... And no good credit hound! |
soft^spirit Send message Joined: 18 May 99 Posts: 6497 Credit: 34,134,168 RAC: 0 |
I totally understand what you are saying. I am trying to get approval to automate a temperature chamber. It will cost about $100,000. I have already developed the root structure in a massive database, and automated test stands, and we would be able to test 12 units at one time, rather than one at a time. I have been trying to do this for almost 8 years, and I'm finally getting closer. My system has already had a major impact on customer satisfaction and unifying test methods. I have only been an engineer for a year, and a technician for 30 years at least. My company makes force feedback servo inclinometers and accelerometers, as well as other products. Unfortunately the corporation decided meetings went much more smoothly without technician input. And eventually dropped having us sign off jobs as completed as well. Much easier to have someone that had no clue do it. Janice |
JohnDK Send message Joined: 28 May 00 Posts: 1222 Credit: 451,243,443 RAC: 1,127 |
CPU tasks: 158 (158 VLAR, 0 VHAR) Mine is 0.018013 and the completion times are correct on GPU/CPU/AP. I did try lowering DCF but didn't help. |
SciManStev Send message Joined: 20 Jun 99 Posts: 6652 Credit: 121,090,076 RAC: 0 |
I've seen that before too. There is the right way, the wrong way, and the bosses way. You kind of have to go along with the one who signs the paychecks. :) Steve Warning, addicted to SETI crunching! Crunching as a member of GPU Users Group. GPUUG Website |
soft^spirit Send message Joined: 18 May 99 Posts: 6497 Credit: 34,134,168 RAC: 0 |
I fixed that. Maybe not a perfect fix. But no longer an issue. Janice |
SciManStev Send message Joined: 20 Jun 99 Posts: 6652 Credit: 121,090,076 RAC: 0 |
I fixed that. I can't wait until I'm 396 and I can finally stop working. I'm counting the seconds. :) Steve Warning, addicted to SETI crunching! Crunching as a member of GPU Users Group. GPUUG Website |
Ellis Hardin Send message Joined: 15 Mar 01 Posts: 33 Credit: 26,603,764 RAC: 0 |
I noticed a glitch a few weeks ago. Something messed up and all the estimated completion times were way off for some reason. Some work units that had been estimated to complete in an hour, were now estimated to complete in a minute or two, etc. One of the results of the glitch was that my client then downloaded about 30 days of work (I had it set to 10 days at the time). They also need to get a some real software engineers to look at the BOINC code. Is a database driven system the best system to control a project like this? Or is it just all they know? |
John McLeod VII Send message Joined: 15 Jul 99 Posts: 24806 Credit: 790,712 RAC: 0 |
Bill Walker wrote: While it is true that sending work and reporting work does not involve the science database, sending new work relies on the splitters which do require the science database. BOINC WIKI |
perryjay Send message Joined: 20 Aug 02 Posts: 3377 Credit: 20,676,751 RAC: 0 |
Anyone that wants can look at the BOINC code. Everything is open source. PROUD MEMBER OF Team Starfire World BOINC |
soft^spirit Send message Joined: 18 May 99 Posts: 6497 Credit: 34,134,168 RAC: 0 |
I fixed that. I had enough.. there was of course a last straw.. But I finally decided to just lower my expectations. Janice |
SciManStev Send message Joined: 20 Jun 99 Posts: 6652 Credit: 121,090,076 RAC: 0 |
I noticed a glitch a few weeks ago. Something messed up and all the estimated completion times were way off for some reason. Some work units that had been estimated to complete in an hour, were now estimated to complete in a minute or two, etc. One of the results of the glitch was that my client then downloaded about 30 days of work (I had it set to 10 days at the time). SETI is a work in progress. Both on the SETI side, and the user side, there are some very sharp people working on this in order to solve many different issues. I am still learning several computer languages, but I can say for certain, based on their posts, that the developers on both sides are definitly the ones you want in those positions. Even on the user side, look how calm and precise the developers are. They know what they are talking about, and don't lose their cool. BOINC is a collection of systems. You have the BOINC infrastructure, with the individual project apps that process some sort of raw data depending on which project you crunch for. And you have us the crunchers working without some of the knowledge the developers have. Each of those different projects has their own specialists, that know the ins and outs of everything. I think that things will work out in the long run, and the issues we face today will be history. Steve Warning, addicted to SETI crunching! Crunching as a member of GPU Users Group. GPUUG Website |
Josef W. Segur Send message Joined: 30 Oct 99 Posts: 4504 Credit: 1,414,761 RAC: 0 |
Bill Walker wrote: Very true, as I said. But it is creating new work which hits the science database, after that there's often a significant delay in the "Results ready to send" queue before sending. The splitters can be disabled or "not running" and still many tasks can be sent. I won't claim that's practical, there's ample evidence they don't want to stop a splitter in the middle of a channel so it would be a fuzzy shutdown as channels finish. Joe |
Josef W. Segur Send message Joined: 30 Oct 99 Posts: 4504 Credit: 1,414,761 RAC: 0 |
I noticed a glitch a few weeks ago. Something messed up and all the estimated completion times were way off for some reason. Some work units that had been estimated to complete in an hour, were now estimated to complete in a minute or two, etc. One of the results of the glitch was that my client then downloaded about 30 days of work (I had it set to 10 days at the time). That glitch was a transition to a new system of estimating run times (and granting credit). It has temporarily been turned off for those of us using anonymous platform, you can expect a similar glitch whenever it is tried again. But it ought to be just that, a temporary glitch which declines fairly quickly. There could be additional bugs, and because BOINC must be prepared to handle projects with many different characteristics it seldom achieves performance ideal for any one project. They also need to get a some real software engineers to look at the BOINC code. Is a database driven system the best system to control a project like this? Or is it just all they know? You might want to look at the wikipedia entry for Dr. Anderson and follow some of the external links. One paper he, Eric Korpela, and Rom Walton authored specifically discusses some of the design decisions. Joe |
Odan Send message Joined: 8 May 03 Posts: 91 Credit: 15,331,177 RAC: 0 |
Wasn't that the idea? Hi Jord, I feel your frustration. I did reduce my cache BTW but I also fiddled with some DCFs to get what I expect to just about tide me over until the servers come back on line. All in all, If I had just left my cache at 8 days i would have ended up with about the same 4-ish days of work. I do look forward to the time when we have more accurate cache filling. I know this is being attempted & I wait patiently for the improvement & realise that I am effectively taking part in RC testing after Beta testing was unfortunately too small to be conclusive. One thing that occurred to me while I was waiting for the 20 WU limit to be removed was that it was somewhat counter productive. During the approx 3.5 days of restricted issue the number of MB in progress gradually fell. I believe that this is at least partly due to the draining of remaining larger caches for the fastest crunchers among us who can actually crunch more than 20 units at a time(!) as well as the slightly more usual among us who get through 20 WU quite quickly. AP exhibited an interesting ratcheting up pattern that seems to be related to AP units being available for something like 1 hour in every 4 or so much more freely than at other times (don't shoot me here I'm really not quite sure what happened but it looks this way to me :) ) Anyway, my point is that the 20 WU limit was a very good idea for the initial restart; it allowed almost everyone to get some form of cache & to get crunching straight away without us being able to max out the downloads. Unfortunately I think it was too severe for too long if S@h still wants to be available for people to crunch continually (I'm not sure that is what the team wants or not, I haven't seen any statement either way. I know that nobody guarantees 24/7 availability I just wonder if Berkley want us all to continue crunching at about the same rate or if they want a reduction) Right, back to the point :) If you look at the Cricket graphs for the week you can see that for those 3+ days we were only downloading WUs at something like 45% of practical capacity for most of that time. Remember that during that time the reservoir of work in progress was falling steadily; individuals were seeing their previously still large caches draining steadily and panicked, increasing their cache sizes further in the hope of snagging WUs when the flood gates opened. For these people we effectively had a 6 day cache drain so that when the limit was removed for 26 hours or so there was a lot of cache to fill up. If some of the 55% of available bandwidth could have been freed up sooner this would have reduced the cache drain and given longer for the caches to fill up. It would also have helped a bit to calm the urge in some people to increase cache sizes "just in case". Of course, if the splitters could not have kept up it wouldn't have mattered but we did start this latest outage with a reservoir of WUs on the servers that could not be sent; I also noticed not all the splitters were running all the time so it appeared that there was some capacity not used. Anyway, I hope that we all have a calm outage :) and that everything comes up smoothly on Friday. I hope there is a bit freer availability of WUs. Happy crunching all! |
©2024 University of California
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.