Message boards :
Number crunching :
Are some gpu tasks longer now?
Message board moderation
Previous · 1 · 2 · 3
Author | Message |
---|---|
TBar Send message Joined: 22 May 99 Posts: 5204 Credit: 840,779,836 RAC: 2,768 |
Hopefully things will work out. BTW, I just downloaded another copy of the sah_v7_opt folder and I'm still getting the same error with the PetriR_raw2 files; Undefined symbols for architecture x86_64: "cudaAcc_GetAutoCorrelation(float*, int, int)", referenced from: seti_analyze(ANALYSIS_STATE&) in seti_cuda-analyzeFuncs.o "cudaAcc_FindAutoCorrelations(int, int)", referenced from: seti_analyze(ANALYSIS_STATE&) in seti_cuda-analyzeFuncs.o ld: symbol(s) not found for architecture x86_64 clang: error: linker command failed with exit code 1 |
jason_gee Send message Joined: 24 Nov 06 Posts: 7489 Credit: 91,093,184 RAC: 0 |
Hopefully things will work out. Probably one of the first things I'll end up looking at, because the autocorrelation streamlining is one of the safest areas, and should give a near constant improvement at all angle ranges. "Living by the wisdom of computer science doesn't sound so bad after all. And unlike most advice, it's backed up by proofs." -- Algorithms to live by: The computer science of human decisions. |
petri33 Send message Joined: 6 Jun 02 Posts: 1668 Credit: 623,086,772 RAC: 156 |
Hopefully things will work out. @TBar: analyzeFuncs.cpp in main folder does not see the cuda/cudaAcc_autocorr.cu GetAutoCorrelations-function correctly defined. The linker expects to find the function, having as a parameter a pointer to a float (or an array of floats), and integer and an integer. I'd try 'make clean' and then 'make' to build all from the scratch. If that doesn't work I'd try to find a duplicate definition for the function (in a *.h file). p.s. I've been a week away from my home (and shut down my computer for the time) and haven't had a time to look at any new (old to me) source that has been published. I've been busy making my rig to recognize a gtx1080 properly under Linux and then getting any 'acceptable' results from my modified code. I had to revert back to pr_zi. That is the one most of You are running if You are experimenting. One good thing. It still works - with the new hardware too. p.p.s I'm sorry that I can't give You an acceptable working executable/source for the NVIDIA-cuda platform at the moment. But, I'm an optimistic. There is a whole summer (in the northern hemisphere) time to make things work all OK. p.p.p.s And having JasonG there backing/leading up I'm sure a Superb Cuda Build will emerge. To overcome Heisenbergs: "You can't always get what you want / but if you try sometimes you just might find / you get what you need." -- Rolling Stones |
Grant (SSSF) Send message Joined: 19 Aug 99 Posts: 13736 Credit: 208,696,464 RAC: 304 |
A nice point Richard. But I run only one at a time. May or may not be of interest/use, or just a red herring. I'm presently running the CUDA50 application on my 2*GTX 750Tis, Win10, 353.82 driver, using the -poll option with 1 CPU core reserved for each GPU WU. I had been running 2WUs at a time till last night where I decided to give 3WUs at a time a go. Even with the loss of 2 more CPU cores output, the increase in WUs per hour from the GPU has offset the CPU loss. Previously (before Guppies) running 3WUs at a time gave better output than 2 with mid range WUs, and much better output with longer running WUs. However the effect with shorties was a massive increase in processing time. So severe was the increase in shortie processing time that it completely offset any gains from the other WUs and resulted in less work per hour being done, even if there were only the occasional shortie in the mix. Now with barely any shortie to be seen, and the Guppies long running processing time like the earlier longer running (but non VLAR) WUs, 3 WUs at a time gives the most work per hour. One thing I have noticed with the Guppies, and is even more noticeable with 3WUs running- when monitoring the GPUs with GPU-z the power used by the GPU varies depending on the Memory Controller load, not the GPU load as it did before the introduction of the Guppies. The GPU load remains around 98-99% (94% the lowest; no WUs finishing or starting in that period). Power consumption (as a % of TDP) varies between 54% (80% memory controller load) down to 31% (30% memory controller load). My personal wild arse guess is that the Memory Controller load isn't what's causing the increased power consumption, but that it's just an indicator of the work the GPU is doing around that time when the data is being moved. Grant Darwin NT |
©2024 University of California
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.