Message boards :
SETI@home Science :
Question about the Graphical FFT
Message board moderation
Author | Message |
---|---|
Finnen78 Send message Joined: 15 Oct 09 Posts: 4 Credit: 307,964 RAC: 0 |
Hello all Boinc crunchers! i read all that nice info on the page about how the screensaver works and what all that nonsence ;) in those windows mean... although... i got a question about that graphical thing... how come sometimes those power bars are wide through the frequency axis and slow and sometimes very thin and fast through the time axis, does that say anything about the signal at that given time? |
jason_gee Send message Joined: 24 Nov 06 Posts: 7489 Credit: 91,093,184 RAC: 0 |
...how come sometimes those power bars are wide through the frequency axis and slow and sometimes very thin and fast through the time axis, does that say anything about the signal at that given time? Hi there, the number of frequency 'bins' (fftlength) varies from 8 through 128k bins, from the dataset of 1*1024*1024 complex single precision floating point dataset. Since we have a fixed data set size, the # of FFT frequency bins will form one side of a rectangular 'new 2d data set' of Frequency (power at each) x time When the FFT Length is short, CPUs can calculate these component FFTs (Time offset is the other dimension of the 'rectangle') much faster due to the way CPU cache works, and how FFT algorithms exchange data internally. Longer FFTs need to swap/combine data much further apart in memory, so cache & memory subsystem comes under more pressure, as well as longer FFTs requiring more processing, making longer FFTs Actually slower..(Opposite to your observation :-O )... However! The other dimension of the graph (Time) is primarily concerned with looking for pulses & triplets, and by far dominates the time taken as the time axis 'extends'. That is because there are many more possible offsets / periods to look for indications of a signal, and that becomes memory & CPU intensive ... overtaking the shorter FFT requirement by a significant proportion (In processing time) Fast Fourier transforms have been improved over decades, and heaps of papers have been written about them, yet pulsefinding, apart from some few key papers is quite difficult to find detailed work on. So for the time being,short searches in the FFT axis, but long in the Time axis will continue to dominate the time taken ... much to the chagrin of us GPU crunchers also, but would like to hope that eventually we can get the timexfrequecy 'area' fairly constant. Hope that points you toward some of the answers you're seeking. Jason "Living by the wisdom of computer science doesn't sound so bad after all. And unlike most advice, it's backed up by proofs." -- Algorithms to live by: The computer science of human decisions. |
Finnen78 Send message Joined: 15 Oct 09 Posts: 4 Credit: 307,964 RAC: 0 |
Hahahaha!!! Thank you Jason! well :) ... i kind of understod about 70% of the computer stuff but u totaly lost me with the other stuff. so... the longer the "bar" on the FFT graph the more time the computer needed to work with that specific prequency at that given time of the package could u stupify that other stuff a little? ;D i mean... im crunching 24/7 with at least one of my computers and sometimes even i have to access it for stuff i only save at that computer, sometimes i have to wait 2 minutes for IE/wordpad/HDD whatever to respond & sometimes the responce is instantanious, and yes... i have allocated 100% of my resources to crunshing. |
jason_gee Send message Joined: 24 Nov 06 Posts: 7489 Credit: 91,093,184 RAC: 0 |
Hahahaha!!! Sure! , yep you have it pretty much right anyway :). Short time axis with fewer points, means 'easy' pulsefinds, which can run fast and smooth. Long time axis means longer pulsefinds, which can 'choke'. Somewhere on the screensaver you should see a task parameter called "AR' or 'Angle Range', and the low ones less than about 0.13 tend to have many long pulsefinds. These would show the worst lagging behaviour you describe. Since you are running CPU multibeam, and experiencing lag problems, once you get sick of the screen saver you might want to look at the possibility of using optimised applications for about 2x speedup or so. I wouldn't recommend that if you prefer 'set & forget' type operation, don't want to have to check for updates regularly, or a particularly attached to the screensaver. Jason "Living by the wisdom of computer science doesn't sound so bad after all. And unlike most advice, it's backed up by proofs." -- Algorithms to live by: The computer science of human decisions. |
ML1 Send message Joined: 25 Nov 01 Posts: 20283 Credit: 7,508,002 RAC: 20 |
... Since we have a fixed data set size... Thanks for a good description including the hardware problems. (Nicely on the intelligible side of technobabble! :-) ) Are there no parallel segmentation algorithms for the pulse finding that would be blindingly fast on GPUs? Keep searchin', Martin See new freedom: Mageia Linux Take a look for yourself: Linux Format The Future is what We all make IT (GPLv3) |
jason_gee Send message Joined: 24 Nov 06 Posts: 7489 Credit: 91,093,184 RAC: 0 |
Are there no parallel segmentation algorithms for the pulse finding that would be blindingly fast on GPUs? Sure, there're ways to do things faster (& more reliably) than current GPU apps, and in the scheme of things we're all learning how to use these new languages & tools, that are based on supercomuter architectures from prior decades, with little prior work to draw on other than the odd SDK samples and stock code. Aside from that, one of the major constraints on technique, is having to achieve results that will validate within tolerances to stock (mostly) serial CPU code, most importantly warts and all. On the high level side of things, there's *relatively* little published work on pulsefinding (serial OR parallel) in comparison to FFT, which has huge volumes of science & mathematics papers spanning decades (about 6 in current serial split Radix form, IIRC) to wring out every flop for different applications. That could perhaps be surprising, seeing as the existing serial algorithms are each of similar algorithmic nature & complexity in many ways. As well as still maturing tools, drivers & techniques, there isn't the same decades of experience to draw on as for [the 'black art' of] CPU micro-architectural optimisation either. Parallel algorithms that have heavy publication tend to have been focussed on the supercomputing platforms of those decades, and the (relatively specialised & obscure) tasks they were expected to perform ... which didn't have to do things like watch videos/screensavers or even run operating systems ;). So all and all, IMO, while there is some experience amongst us with the predecessor architectures & principles, and there are quite a few notable similarities, the current & maturing platforms are turning out to be a different beast than, e.g. The CSIRO GIS Lab 'Hitachi super torus' I worked on back in University. We're definitely 'getting there' though. Jason "Living by the wisdom of computer science doesn't sound so bad after all. And unlike most advice, it's backed up by proofs." -- Algorithms to live by: The computer science of human decisions. |
ML1 Send message Joined: 25 Nov 01 Posts: 20283 Credit: 7,508,002 RAC: 20 |
We're definitely 'getting there' though. Oooer... Quite a jump from the old Seymour Cray vector pipelines... ;-o Segmentation is important in image processing. Surprised that hasn't been pushed for 'fast' optimised processing. Thanks and good luck for pushing the computations. Keep searchin', Martin See new freedom: Mageia Linux Take a look for yourself: Linux Format The Future is what We all make IT (GPLv3) |
©2024 University of California
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.