Monitoring inconclusive GBT validations and harvesting data for testing

Message boards : Number crunching : Monitoring inconclusive GBT validations and harvesting data for testing
Message board moderation

To post messages, you must log in.

Previous · 1 . . . 9 · 10 · 11 · 12 · 13 · 14 · 15 . . . 36 · Next

AuthorMessage
Profile Wiggo
Avatar

Send message
Joined: 24 Jan 00
Posts: 34748
Credit: 261,360,520
RAC: 489
Australia
Message 1815354 - Posted: 6 Sep 2016, 9:40:56 UTC - in response to Message 1815315.  

Things seem to improve. Look further down and look at my consecutive valid tasks with Petris latest revision.

http://setiathome.berkeley.edu/host_app_versions.php?hostid=8053171

*Thumbs up*

You may give it a thumbs up ATM, but if you have a thorough look, the app still isn't giving a straight out conclusive result against a standard CPU app running on an error free rig, it still needs a 3rd to validate.

It's just something to have a closer look for those interested. ;-)

Cheers.
ID: 1815354 · Report as offensive
Profile petri33
Volunteer tester

Send message
Joined: 6 Jun 02
Posts: 1668
Credit: 623,086,772
RAC: 156
Finland
Message 1815407 - Posted: 6 Sep 2016, 20:51:28 UTC
Last modified: 6 Sep 2016, 20:58:47 UTC

I've seen some false positive signals i.e. an extra pulse found or a case when the exaclty same pulse (frequency, time and chirp) is found but a different power for the pulse is reported.

a) all setiathome signals so far are false positives. All and every one. Even the WOW signal.

b) A pulse that is at the same frequency, time and chirp is the same pulse by another application. The power of the pulse may be 1/100000 or 1/1e6 different but all signals that are suspected valid are to be rechecked byt seti institution anyway by not with our public software. They'll run a double double precision SW then. The difference in power comes from the different hw architecture, different implementation of calculating the average (that is an estimate at best since the average need a sum of thousands of floating point approximations of a number and can be done in sequential, pairwise, recurring or whatever way and still be 'right') and the different order of doing a+b+c+d vs (a+b)+(c+d) or any other combination of 32 bit floating point arithmetic with rounding errors.

The found pulses should be matched for frequency, time and chirp a priori and with a tolerance for error in the power. (That is what they do now and that is why some results need a third oppinion). An occasional false positive is not a bad thing. They happen with low power signals that are most probably due to noise.


The MAC problem with a certain OS version is not my problem. It affects the OpenCL version too.
To overcome Heisenbergs:
"You can't always get what you want / but if you try sometimes you just might find / you get what you need." -- Rolling Stones
ID: 1815407 · Report as offensive
Profile Raistmer
Volunteer developer
Volunteer tester
Avatar

Send message
Joined: 16 Jun 01
Posts: 6325
Credit: 106,370,077
RAC: 121
Russia
Message 1815411 - Posted: 6 Sep 2016, 21:00:28 UTC - in response to Message 1815407.  
Last modified: 6 Sep 2016, 21:06:08 UTC

I've seen some false positive signals i.e. an extra pulse found or a case when the exaclty same pulse (time and chirp) is found but a different power for the pulse is reported.

So they are not quite false positive, that just have incorrect power.


a) all setiathome signals so far are false positives. All and every one. Even the WOW signal.

Not correct. One need to clearly distinguish false positive in sense of validator passing and in scientific sense. You used "scientific sense" here but w/o real ground. SETI@home doesn'r look into single signals at all. PErcistency is requirement for anything of near to threshold power. That is, to say if we detected signal or not postprocessing required. And until it will be performed we (no one) can't say that we catch only false positives so far.


b) A pulse that is at the same time and chirp is the same pulse by another application. The power of the pulse may be 1/100000 or 1/1e6 different but all signals that are suspected valid are to be rechecked byt seti institution anyway by not with our public software. They'll run a double double precision SW then. The difference in power comes from the different hw architecture, different implementation of calculating the average (that is an estimate at best since the average need a sum of thousands of floating point approximations of a number and can be done in sequential, pairwise, recurring or whatever way and still be 'right') and the different order of doing a+b+c+d vs (a+b)+(c+d) or any other combination of 32 bit floating point arithmetic with rounding errors.

Our aim on this distributed phase is just make results to pass validator with as low number of resends as possible. No need to mix this task with further stages of signal post-processing.
EDIT: and of course near-to-threshold inconsistency unavoidable in all apps.


The MAC problem with a certain version is not my problem. It affects the OpenCL version too.

Did you see some reports of the same issue for CUDA too?
SETI apps news
We're not gonna fight them. We're gonna transcend them.
ID: 1815411 · Report as offensive
Profile petri33
Volunteer tester

Send message
Joined: 6 Jun 02
Posts: 1668
Credit: 623,086,772
RAC: 156
Finland
Message 1815414 - Posted: 6 Sep 2016, 21:06:07 UTC - in response to Message 1815411.  




The MAC problem with a certain version is not my problem. It affects the OpenCL version too.

Did you see some reports of the same issue for CUDA too?


I think TBar reported just that at this thread or in the 'built some osx apps' thread.
To overcome Heisenbergs:
"You can't always get what you want / but if you try sometimes you just might find / you get what you need." -- Rolling Stones
ID: 1815414 · Report as offensive
Profile petri33
Volunteer tester

Send message
Joined: 6 Jun 02
Posts: 1668
Credit: 623,086,772
RAC: 156
Finland
Message 1815417 - Posted: 6 Sep 2016, 21:08:27 UTC - in response to Message 1815411.  
Last modified: 6 Sep 2016, 21:10:30 UTC


Our aim on this distributed phase is just make results to pass validator with as low number of resends as possible. No need to mix this task with further stages of signal post-processing.
EDIT: and of course near-to-threshold inconsistency unavoidable in all apps.


And to lower the number of resends 'they' might lower the bar on 'power' just because of the unavoidable inconsistency.
To overcome Heisenbergs:
"You can't always get what you want / but if you try sometimes you just might find / you get what you need." -- Rolling Stones
ID: 1815417 · Report as offensive
Profile Raistmer
Volunteer developer
Volunteer tester
Avatar

Send message
Joined: 16 Jun 01
Posts: 6325
Credit: 106,370,077
RAC: 121
Russia
Message 1815418 - Posted: 6 Sep 2016, 21:08:41 UTC - in response to Message 1815414.  




The MAC problem with a certain version is not my problem. It affects the OpenCL version too.

Did you see some reports of the same issue for CUDA too?


I think TBar reported just that at this thread or in the 'built some osx apps' thread.

Ah, maybe... I thought the same but looked into app names in posted bench saw only vs OpenCL comparison. Of course if OpenCL computes incorrectly any comparison with it just void and has apriory no sense.

Would be good if soem comparison vs known to be OK CPU would made. Cause so far I did not saw definitive reports of CUDA issue for particular OS X versions (TBar, could you clarify this a little?? )
SETI apps news
We're not gonna fight them. We're gonna transcend them.
ID: 1815418 · Report as offensive
Profile Raistmer
Volunteer developer
Volunteer tester
Avatar

Send message
Joined: 16 Jun 01
Posts: 6325
Credit: 106,370,077
RAC: 121
Russia
Message 1815421 - Posted: 6 Sep 2016, 21:11:29 UTC - in response to Message 1815417.  
Last modified: 6 Sep 2016, 21:11:48 UTC


Our aim on this distributed phase is just make results to pass validator with as low number of resends as possible. No need to mix this task with further stages of signal post-processing.
EDIT: and of course near-to-threshold inconsistency unavoidable in all apps.


And to lower the number of resends 'they' might lower the bar on 'power' just because of the unavoidable inconsistency.

Well, so far tolerance range was OK enough (actually, we in 99+% similarity for all already released apps on working HW/soft combos). This doesn't exclude occasional inconsistencies of course. But to tell if something too bad happens with particular build or just "too near to threshold" case some additional observations required.
SETI apps news
We're not gonna fight them. We're gonna transcend them.
ID: 1815421 · Report as offensive
Profile petri33
Volunteer tester

Send message
Joined: 6 Jun 02
Posts: 1668
Credit: 623,086,772
RAC: 156
Finland
Message 1815426 - Posted: 6 Sep 2016, 21:18:43 UTC - in response to Message 1815421.  


Our aim on this distributed phase is just make results to pass validator with as low number of resends as possible. No need to mix this task with further stages of signal post-processing.
EDIT: and of course near-to-threshold inconsistency unavoidable in all apps.


And to lower the number of resends 'they' might lower the bar on 'power' just because of the unavoidable inconsistency.

Well, so far tolerance range was OK enough (actually, we in 99+% similarity for all already released apps on working HW/soft combos). This doesn't exclude occasional inconsistencies of course. But to tell if something too bad happens with particular build or just "too near to threshold" case some additional observations required.


My latest version (and the previous one too) zi3f-zi3g give 99%+ similarity on tests. One guppi vlar finds an extra signal (low power, low average) and my inconclusives is falling rapidly. The old ones from earlier version do persist a month or so in the inconclusives though until they are validated by a third computer.
To overcome Heisenbergs:
"You can't always get what you want / but if you try sometimes you just might find / you get what you need." -- Rolling Stones
ID: 1815426 · Report as offensive
Profile Raistmer
Volunteer developer
Volunteer tester
Avatar

Send message
Joined: 16 Jun 01
Posts: 6325
Credit: 106,370,077
RAC: 121
Russia
Message 1815428 - Posted: 6 Sep 2016, 21:27:59 UTC - in response to Message 1815426.  
Last modified: 6 Sep 2016, 21:28:45 UTC

When you will feel ready to wider testing your build should be deployed on SETI beta site for wider testing (both for compatible HW/SW combos designation and for validness). Some preliminary beta testing can be made too but adequate validness checking almost impossible on SETI main (this thread clearly shows what difficulties encountered on main). So, some time shold be spend on beta site too (under anonymous platform or as new stock distribution).
On beta results persist much longer + 3-way validation better shows any issues if any.
SETI apps news
We're not gonna fight them. We're gonna transcend them.
ID: 1815428 · Report as offensive
Profile petri33
Volunteer tester

Send message
Joined: 6 Jun 02
Posts: 1668
Credit: 623,086,772
RAC: 156
Finland
Message 1815429 - Posted: 6 Sep 2016, 21:29:29 UTC - in response to Message 1815428.  
Last modified: 6 Sep 2016, 21:30:28 UTC

When you will feel ready to wider testing your build should be deployed on SETI beta site for wider testing (both for compatible HW/SW combos designation and for validness). Some preliminary beta testing can be made too but adequate validness checking almost impossible on SETI main (this thread clearly shows what difficulties encountered on main). So, some time shold be spend on beta site too (under anonymous platform or as new stock distribution).
On beta results persist much longer + 3-way validation better shows any issues if any.


I could set mine to run 5% beta. Would that help?

EDIT: When I set it to beta how do I set it up to run my version instead of the downloaded app?
To overcome Heisenbergs:
"You can't always get what you want / but if you try sometimes you just might find / you get what you need." -- Rolling Stones
ID: 1815429 · Report as offensive
Profile Raistmer
Volunteer developer
Volunteer tester
Avatar

Send message
Joined: 16 Jun 01
Posts: 6325
Credit: 106,370,077
RAC: 121
Russia
Message 1815432 - Posted: 6 Sep 2016, 21:33:03 UTC - in response to Message 1815429.  
Last modified: 6 Sep 2016, 21:34:16 UTC

When you will feel ready to wider testing your build should be deployed on SETI beta site for wider testing (both for compatible HW/SW combos designation and for validness). Some preliminary beta testing can be made too but adequate validness checking almost impossible on SETI main (this thread clearly shows what difficulties encountered on main). So, some time shold be spend on beta site too (under anonymous platform or as new stock distribution).
On beta results persist much longer + 3-way validation better shows any issues if any.


I could set mine to run 5% beta. Would that help?

Well, this will slowly create some set of tasks completed by your app that persists long enough to be analysed. So definitely it would be good thing even if running only on single host.
Would be good if your current alpha-testers will do the same. And if no hazards will be discovered, semi-open or open beta will be the next stage IMHO.
If your app much faster would be good if it will be distributed as wide as possible (after adequate testing of course).
SETI apps news
We're not gonna fight them. We're gonna transcend them.
ID: 1815432 · Report as offensive
Profile Raistmer
Volunteer developer
Volunteer tester
Avatar

Send message
Joined: 16 Jun 01
Posts: 6325
Credit: 106,370,077
RAC: 121
Russia
Message 1815433 - Posted: 6 Sep 2016, 21:33:42 UTC - in response to Message 1815429.  


EDIT: When I set it to beta how do I set it up to run my version instead of the downloaded app?

Just the same way as for main - via app_info.xml and anonymous platform setup.
SETI apps news
We're not gonna fight them. We're gonna transcend them.
ID: 1815433 · Report as offensive
Profile -= Vyper =-
Volunteer tester
Avatar

Send message
Joined: 5 Sep 99
Posts: 1652
Credit: 1,065,191,981
RAC: 2,537
Sweden
Message 1815436 - Posted: 6 Sep 2016, 21:41:44 UTC

If we're about to take the original S@H code today and compare it against all platforms would all variants produce 100% instead of 99+%?

We're talking a bunch of WU's and a whole lot of variants:

Debian, Ubuntu , Arch , Mint etc x32, x64.
Apple variants
Windows Vista,7,8,8.1,10,Server 2008,Server 2012 x32 and x64
Arm Linux, Android etc etc.

There are alot to compare through in a chart to see if the cpu portion of the code works as intended on all platforms with 100% similarity.
If not then i Think s@h themselves need to have a plan to solve this because if the cpu portion of the executables isn't a 100% match how on Earth would we ever get as Close to 100% on GPUs when the problems would be multiplied within the code.

Just a thought and my 2 cents of thinking outside the box now.

_________________________________________________________________________
Addicted to SETI crunching!
Founder of GPU Users Group
ID: 1815436 · Report as offensive
Profile petri33
Volunteer tester

Send message
Joined: 6 Jun 02
Posts: 1668
Credit: 623,086,772
RAC: 156
Finland
Message 1815437 - Posted: 6 Sep 2016, 21:44:52 UTC - in response to Message 1815433.  
Last modified: 6 Sep 2016, 21:47:16 UTC


EDIT: When I set it to beta how do I set it up to run my version instead of the downloaded app?

Just the same way as for main - via app_info.xml and anonymous platform setup.



Thanks, I'm running now some blc7_guppi_57449 v 8.10 tasks on beta.
EDIT: http://setiweb.ssl.berkeley.edu/beta/hosts_user.php?userid=17240
To overcome Heisenbergs:
"You can't always get what you want / but if you try sometimes you just might find / you get what you need." -- Rolling Stones
ID: 1815437 · Report as offensive
Profile Raistmer
Volunteer developer
Volunteer tester
Avatar

Send message
Joined: 16 Jun 01
Posts: 6325
Credit: 106,370,077
RAC: 121
Russia
Message 1815438 - Posted: 6 Sep 2016, 21:46:09 UTC - in response to Message 1815436.  
Last modified: 6 Sep 2016, 21:47:30 UTC

If we're about to take the original S@H code today and compare it against all platforms would all variants produce 100% instead of 99+%?

Of course they will be 99+%. You just listed reasons why and nobody expects 100% similarity and never expected. That's why validator has some validation tolerance range (!) (and, btw, it quite wider than our Lunatics 99+ standart).
But if validator marks result as inconclusive, especially for new quite re-designed build, we need to establish the reason why and how often this happens.
Nothing new here, all apps passed and still passing through the same testing.
SETI apps news
We're not gonna fight them. We're gonna transcend them.
ID: 1815438 · Report as offensive
Profile Raistmer
Volunteer developer
Volunteer tester
Avatar

Send message
Joined: 16 Jun 01
Posts: 6325
Credit: 106,370,077
RAC: 121
Russia
Message 1815440 - Posted: 6 Sep 2016, 21:48:42 UTC - in response to Message 1815437.  


EDIT: When I set it to beta how do I set it up to run my version instead of the downloaded app?

Just the same way as for main - via app_info.xml and anonymous platform setup.



Thanks, I'm running now some blc7_guppi_57449 v 8.10 tasks on beta.
EDIT: http://setiweb.ssl.berkeley.edu/beta/hosts_user.php?userid=17240

Thanks for link, this will simplify testing analyse for those who interested.
SETI apps news
We're not gonna fight them. We're gonna transcend them.
ID: 1815440 · Report as offensive
Profile -= Vyper =-
Volunteer tester
Avatar

Send message
Joined: 5 Sep 99
Posts: 1652
Credit: 1,065,191,981
RAC: 2,537
Sweden
Message 1815441 - Posted: 6 Sep 2016, 21:50:00 UTC
Last modified: 6 Sep 2016, 21:51:33 UTC

Has anyone the knowledge exactly how the validator works? I mean how much is the tolerance, does it compare spikes, tripplets, gaussians and pulses with exactly the same tolerance in percentage?
How much is the +/- percentage even what it consider a mismatch and why does it later on even validate the result when it first marked it with a invalid.
Whats the difference there first and second time?

EDIT: Could perhaps look at the server portion of the code if its public available

_________________________________________________________________________
Addicted to SETI crunching!
Founder of GPU Users Group
ID: 1815441 · Report as offensive
Profile Raistmer
Volunteer developer
Volunteer tester
Avatar

Send message
Joined: 16 Jun 01
Posts: 6325
Credit: 106,370,077
RAC: 121
Russia
Message 1815442 - Posted: 6 Sep 2016, 21:53:12 UTC - in response to Message 1815441.  
Last modified: 6 Sep 2016, 21:55:15 UTC


EDIT: Could perhaps look at the server portion of the code if its public available

yep, it's publicly available and in the same repository as all SETI code.

Why validated later, simple example:

1 and 3 - differs by 2.
But {1,2,3} 2 differs only by 1 to 3 and to 1, 1 differs only by 1 to 2 the same for 3 - all passed validation.
It's quite rough example of course.
SETI apps news
We're not gonna fight them. We're gonna transcend them.
ID: 1815442 · Report as offensive
Kiska
Volunteer tester

Send message
Joined: 31 Mar 12
Posts: 302
Credit: 3,067,762
RAC: 0
Australia
Message 1815443 - Posted: 6 Sep 2016, 21:53:21 UTC - in response to Message 1815441.  
Last modified: 6 Sep 2016, 21:54:43 UTC

Look the the rescmpv on lunatics, they include the source code that is used for the server.

EDIT: also on the repo for all seti code
ID: 1815443 · Report as offensive
TBar
Volunteer tester

Send message
Joined: 22 May 99
Posts: 5204
Credit: 840,779,836
RAC: 2,768
United States
Message 1815464 - Posted: 6 Sep 2016, 23:20:37 UTC - in response to Message 1815418.  
Last modified: 7 Sep 2016, 0:00:56 UTC




The MAC problem with a certain version is not my problem. It affects the OpenCL version too.

Did you see some reports of the same issue for CUDA too?


I think TBar reported just that at this thread or in the 'built some osx apps' thread.

Ah, maybe... I thought the same but looked into app names in posted bench saw only vs OpenCL comparison. Of course if OpenCL computes incorrectly any comparison with it just void and has apriory no sense.

Would be good if soem comparison vs known to be OK CPU would made. Cause so far I did not saw definitive reports of CUDA issue for particular OS X versions (TBar, could you clarify this a little?? )

For Petri's Apps the only problems I've noted are also present in Linux. But it Appears those problems may have been Solved with the latest version x41p_zi3g. So far, running zi3g in Darwin 14.5 with CUDA driver 7.5.27 has been trouble free. Items that Appear to have been fixed are;
1) Runs with the 7.5.x driver without causing the 750Ti to Stall
2) Occasional reports of ...exited with zero status but no 'finished' file, Gone
3) Extreme Autocorrelation Peaks Non-existent

So Far, So Good.

The other problem with the Baseline Apps only affect the Laptops Occasionally.
That is a problem with the Baseline code, which I think Jason controls.
Petri's App doesn't work on those Laptops.


Hmmm, just noticed an Invalid overflow with x41p_zi3g, http://setiathome.berkeley.edu/result.php?resultid=5139355961
Looking at the time, Received: 6 Sep 2016, 5:31:32 UTC, that must have been one of the first tasks run with x41p_zi3g, and it was on the 750Ti. We'll have to see if anymore appear.
I suppose the next thing to do would be to build a Linux version. This should be fun, since the last Linux build I've installed a clean system.

Not good, another Overflow on the 750Ti that the WingPerson doesn't agree with, http://setiathome.berkeley.edu/workunit.php?wuid=2257472392
Seems there is still no love for the 750Ti with driver 7.5.x, although it has Not Stalled.
ID: 1815464 · Report as offensive
Previous · 1 . . . 9 · 10 · 11 · 12 · 13 · 14 · 15 . . . 36 · Next

Message boards : Number crunching : Monitoring inconclusive GBT validations and harvesting data for testing


 
©2024 University of California
 
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.