Question about dedicating CPU cores to GPU support w/Lunatics apps.

Message boards : Number crunching : Question about dedicating CPU cores to GPU support w/Lunatics apps.
Message board moderation

To post messages, you must log in.

Previous · 1 · 2 · 3 · Next

AuthorMessage
Profile Mike Special Project $75 donor
Volunteer tester
Avatar

Send message
Joined: 17 Feb 01
Posts: 34258
Credit: 79,922,639
RAC: 80
Germany
Message 1467339 - Posted: 22 Jan 2014, 16:48:09 UTC - in response to Message 1467285.  

Mike

I use -use_sleep -unroll 18 -ffa_block 8192 -ffa_block_fetch 2048 on my 780, as sugested by someone else in other thread, and i see you sugest a different setting for the 780. As i´m sure you have a lot more experience than me, could you tell me what is the best setting? If not to much work, you know the setting for the 670 or the 690 too? Thanks in advance.

As a sugestion somebody could make a table with the optimal setting for each card, i´m sure that will be very wellcomed by the comunity.


First of all Helli is using the older app without -use sleep function.
For that app my settings is the best to go with the 780.
Steve is using it for month on his 480`s like i suggested it to him.

Evenso it depends what your goal is.
Reducing CPU usage or getting best run times.

What i can see for sure is that your times are to slow for a 780.


With each crime and every kindness we birth our future.
ID: 1467339 · Report as offensive
Profile Mike Special Project $75 donor
Volunteer tester
Avatar

Send message
Joined: 17 Feb 01
Posts: 34258
Credit: 79,922,639
RAC: 80
Germany
Message 1467340 - Posted: 22 Jan 2014, 16:53:25 UTC - in response to Message 1467288.  

Also add this to the comandline or ap_cmdline_win_x86_SSE2_OpenCL_NV.txt file

-unroll 12 -ffa_block 12288 -ffa_block_fetch 6144

This should speed your 780 up.

Mike, any reason you didn't include the -hp switch to bump the priority up above the default of "below normal"? I usually recommend that for the AP/Nvidia apps.


HP switch is slowing processing down.


With each crime and every kindness we birth our future.
ID: 1467340 · Report as offensive
juan BFP Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 16 Mar 07
Posts: 9786
Credit: 572,710,851
RAC: 3,799
Panama
Message 1467358 - Posted: 22 Jan 2014, 17:35:17 UTC - in response to Message 1467339.  

What i can see for sure is that your times are to slow for a 780.

Even if i run 3 WU at a time and my CPU is a slow I5?

Thanks for your help, switching the file now.
ID: 1467358 · Report as offensive
Profile Mike Special Project $75 donor
Volunteer tester
Avatar

Send message
Joined: 17 Feb 01
Posts: 34258
Credit: 79,922,639
RAC: 80
Germany
Message 1467359 - Posted: 22 Jan 2014, 17:39:14 UTC - in response to Message 1467358.  

What i can see for sure is that your times are to slow for a 780.

Even if i run 3 WU at a time and my CPU is a slow I5?

Thanks for your help, switching the file now.


Which file ?

Just adjust the settings.


With each crime and every kindness we birth our future.
ID: 1467359 · Report as offensive
juan BFP Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 16 Mar 07
Posts: 9786
Credit: 572,710,851
RAC: 3,799
Panama
Message 1467362 - Posted: 22 Jan 2014, 17:45:23 UTC - in response to Message 1467359.  

Soryy was a translation error, i just change the settings.
ID: 1467362 · Report as offensive
kittyman Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 9 Jul 00
Posts: 51468
Credit: 1,018,363,574
RAC: 1,004
United States
Message 1467369 - Posted: 22 Jan 2014, 17:56:56 UTC - in response to Message 1467339.  
Last modified: 22 Jan 2014, 17:58:21 UTC



Evenso it depends what your goal is.
Reducing CPU usage or getting best run times.

Could you possibly explain how the settings affect CPU usage VS best run times?
What would you change to achieve one over the other?

I would prefer best run times, as on my crunch-only rigs CPU usage is not really my concern.

Meow?
"Freedom is just Chaos, with better lighting." Alan Dean Foster

ID: 1467369 · Report as offensive
Profile Mike Special Project $75 donor
Volunteer tester
Avatar

Send message
Joined: 17 Feb 01
Posts: 34258
Credit: 79,922,639
RAC: 80
Germany
Message 1467373 - Posted: 22 Jan 2014, 18:10:11 UTC - in response to Message 1467369.  
Last modified: 22 Jan 2014, 18:12:17 UTC



Evenso it depends what your goal is.
Reducing CPU usage or getting best run times.

Could you possibly explain how the settings affect CPU usage VS best run times?
What would you change to achieve one over the other?

I would prefer best run times, as on my crunch-only rigs CPU usage is not really my concern.

Meow?


Higher app values increasing the processing speed.

Since r_2058 -sleep switch is available to trigger CPU usage on Nvidia hosts.

For better run times just increase -unroll and ffa_block/fetch like i suggested earlier.

BTW
To high unrolls have no affect in speed just causing crashes.
Especially on multi GPU hosts.


With each crime and every kindness we birth our future.
ID: 1467373 · Report as offensive
kittyman Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 9 Jul 00
Posts: 51468
Credit: 1,018,363,574
RAC: 1,004
United States
Message 1467399 - Posted: 22 Jan 2014, 19:03:39 UTC - in response to Message 1467373.  



Evenso it depends what your goal is.
Reducing CPU usage or getting best run times.

Could you possibly explain how the settings affect CPU usage VS best run times?
What would you change to achieve one over the other?

I would prefer best run times, as on my crunch-only rigs CPU usage is not really my concern.

Meow?


Higher app values increasing the processing speed.

Since r_2058 -sleep switch is available to trigger CPU usage on Nvidia hosts.

For better run times just increase -unroll and ffa_block/fetch like i suggested earlier.

BTW
To high unrolls have no affect in speed just causing crashes.
Especially on multi GPU hosts.
Does ffa block fetch have to be increased in certain increments? And does that increase GPU memory usage?
"Freedom is just Chaos, with better lighting." Alan Dean Foster

ID: 1467399 · Report as offensive
Profile Mike Special Project $75 donor
Volunteer tester
Avatar

Send message
Joined: 17 Feb 01
Posts: 34258
Credit: 79,922,639
RAC: 80
Germany
Message 1467490 - Posted: 22 Jan 2014, 23:19:08 UTC - in response to Message 1467399.  



Evenso it depends what your goal is.
Reducing CPU usage or getting best run times.

Could you possibly explain how the settings affect CPU usage VS best run times?
What would you change to achieve one over the other?

I would prefer best run times, as on my crunch-only rigs CPU usage is not really my concern.

Meow?


Higher app values increasing the processing speed.

Since r_2058 -sleep switch is available to trigger CPU usage on Nvidia hosts.

For better run times just increase -unroll and ffa_block/fetch like i suggested earlier.

BTW
To high unrolls have no affect in speed just causing crashes.
Especially on multi GPU hosts.
Does ffa block fetch have to be increased in certain increments? And does that increase GPU memory usage?


Check the read me for details.

You can try -unroll 12 -ffa_block 12288 -ffa_block_fetch 6144.

Memory usage is a little higher but not much.
It just adjusts amount of data beeing processed in one kernel call.


With each crime and every kindness we birth our future.
ID: 1467490 · Report as offensive
S197Mach1

Send message
Joined: 16 Jan 11
Posts: 9
Credit: 20,120,713
RAC: 0
United States
Message 1467587 - Posted: 23 Jan 2014, 6:33:00 UTC

Not to hijack the thread but since its about dedicated CPU cores, I went ahead and modified the app info file and changed the cpu usage for v7 mbs to .25 and astropulse to 1.0. I came home today to find that on my 12 thread processor it was over provisioned to 12 1/2. So I had 10 cpu tasks going, one cpu paused(because a astropulse unit started), 2 astro pulse going(at 12 cores utilized) and 2 v7 MB going at .25 (12 1/2). Is this a bug in Boinc? I was just idling cores/threads prior to reading this thread and implementing in app info which seems more efficient to me, but I just trying to figure out the over provisioning part and why boinc didnt pause anymore cpu threads to keep it under 12 cpu/threads utilized for tasks.
ID: 1467587 · Report as offensive
Profile Helli_retiered
Volunteer tester
Avatar

Send message
Joined: 15 Dec 99
Posts: 707
Credit: 108,785,585
RAC: 0
Germany
Message 1480506 - Posted: 22 Feb 2014, 5:42:24 UTC - in response to Message 1467315.  
Last modified: 22 Feb 2014, 5:44:07 UTC

I think it was suggested that by modifying app_info, one could basically get Boinc to 'auto switch', and dedicate the CPU cores to the GPUs when needed, but release them for CPU MB or AP work when the GPUs are doing MB and don't need them.
If so, that would make better use of my CPU resources between AP splitting runs.

How would that work with multiple GPUs? And with 2 AP tasks per GPU? It the setting per GPU? Per running instance of the app?


Edit app_info.xml and change the "avg CPU" and "max CPU" lines as follows:

<app_version>
<app_name>astropulse_v6</app_name>
<version_num>604</version_num>
<avg_ncpus>1.0</avg_ncpus>
<max_ncpus>1.0</max_ncpus>

There are two app sections for 1843, so either pick the right plan class or just do both of them.

This will reserve 1 cpu core for each task running and release them when not needed. A value of .5 will reserve 1 for every two tasks running - you might want to try that if you install the Beta Juan mentioned. I use .5 for cuda 50 tasks (2 at a time) and do not reserve any cores in BOINC's computing preferences. I use app_config.xml for these settings but you can't do that with your version of BOINC.




Thank you, I have been waiting for this. :-)

Helli


@Helli
On your system i would suggest to set it like this.

<avg_ncpus>0.5</avg_ncpus>
<max_ncpus>0.5</max_ncpus>

Also add this to the comandline or ap_cmdline_win_x86_SSE2_OpenCL_NV.txt file

-unroll 12 -ffa_block 12288 -ffa_block_fetch 6144

This should speed your 780 up.



Thanks Mike, i will try your advise immediately.

......




Mike, i've added a (second) GTX780 to my Rig - do i have to change
something for optimal workload?

Perhaps - i'm running two Workunits per GPU.

Helli
ID: 1480506 · Report as offensive
Profile Mike Special Project $75 donor
Volunteer tester
Avatar

Send message
Joined: 17 Feb 01
Posts: 34258
Credit: 79,922,639
RAC: 80
Germany
Message 1480558 - Posted: 22 Feb 2014, 8:44:50 UTC - in response to Message 1480506.  

I think it was suggested that by modifying app_info, one could basically get Boinc to 'auto switch', and dedicate the CPU cores to the GPUs when needed, but release them for CPU MB or AP work when the GPUs are doing MB and don't need them.
If so, that would make better use of my CPU resources between AP splitting runs.

How would that work with multiple GPUs? And with 2 AP tasks per GPU? It the setting per GPU? Per running instance of the app?


Edit app_info.xml and change the "avg CPU" and "max CPU" lines as follows:

<app_version>
<app_name>astropulse_v6</app_name>
<version_num>604</version_num>
<avg_ncpus>1.0</avg_ncpus>
<max_ncpus>1.0</max_ncpus>

There are two app sections for 1843, so either pick the right plan class or just do both of them.

This will reserve 1 cpu core for each task running and release them when not needed. A value of .5 will reserve 1 for every two tasks running - you might want to try that if you install the Beta Juan mentioned. I use .5 for cuda 50 tasks (2 at a time) and do not reserve any cores in BOINC's computing preferences. I use app_config.xml for these settings but you can't do that with your version of BOINC.




Thank you, I have been waiting for this. :-)

Helli


@Helli
On your system i would suggest to set it like this.

<avg_ncpus>0.5</avg_ncpus>
<max_ncpus>0.5</max_ncpus>

Also add this to the comandline or ap_cmdline_win_x86_SSE2_OpenCL_NV.txt file

-unroll 12 -ffa_block 12288 -ffa_block_fetch 6144

This should speed your 780 up.



Thanks Mike, i will try your advise immediately.

......




Mike, i've added a (second) GTX780 to my Rig - do i have to change
something for optimal workload?

Perhaps - i'm running two Workunits per GPU.

Helli


Not necessarily.
But i`d suggest to use a more recent app version if you want to reduce CPU usage as well.
You can download r_2058 from my website for example and add -sleep switch.

Like this.

-unroll 12 -ffa_block 12288 -ffa_block_fetch 6144 -sleep.

But then you need to addjust max_ncpus as well.


With each crime and every kindness we birth our future.
ID: 1480558 · Report as offensive
Profile Helli_retiered
Volunteer tester
Avatar

Send message
Joined: 15 Dec 99
Posts: 707
Credit: 108,785,585
RAC: 0
Germany
Message 1480592 - Posted: 22 Feb 2014, 10:40:50 UTC - in response to Message 1480558.  

Thanks for your generous Advice, Mike.
I will try it at my Best...

Helli
ID: 1480592 · Report as offensive
Profile Mike Special Project $75 donor
Volunteer tester
Avatar

Send message
Joined: 17 Feb 01
Posts: 34258
Credit: 79,922,639
RAC: 80
Germany
Message 1480593 - Posted: 22 Feb 2014, 10:47:25 UTC

You are welcome Helli.

I made a typo.

It has to be like this with the new app.

-unroll 12 -ffa_block 12288 -ffa_block_fetch 6144 -use_sleep.


With each crime and every kindness we birth our future.
ID: 1480593 · Report as offensive
Kieron Walsh

Send message
Joined: 2 Mar 00
Posts: 74
Credit: 43,502,325
RAC: 112
United Kingdom
Message 1481596 - Posted: 25 Feb 2014, 10:41:25 UTC

Hi,
I'm a long-time SETI participant but have so far stuck to just running the Lunatics apps with its default settings. I'd like to squeeze a bit more performance out of my PC if possible so to run 2 GPU tasks simultaneously I should be leaving 1 CPU core free - or is it 2?

When I look at my app_info file I see multiple entries for <avg_ncpus> and </max_ncpus>; do I need to change them all or just some of them?

Also what do I change to run 2 simultaneous tasks on the GPU?

My system: Quad Core Q6600 @2.40GHz, 8GB RAM, 64-bit Win8.1, GTX460 1GB GPU.

Apologies for posting in its entirety but I'm not certain how much is required/relevant so my current app_info.xml is:

<app_info>
<app>
<name>setiathome_v7</name>
</app>
<file_info>
<name>AKv8c_Bb_r1846_winx86_SSSE3x.exe</name>
<executable/>
</file_info>
<file_info>
<name>libfftw3f-3.dll</name>
<executable/>
</file_info>
<file_info>
<name>AKv8c_Bb_r1846_winx86_SSSE3x.txt</name>
</file_info>
<app_version>
<app_name>setiathome_v7</app_name>
<version_num>700</version_num>
<platform>windows_intelx86</platform>
<file_ref>
<file_name>AKv8c_Bb_r1846_winx86_SSSE3x.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>libfftw3f-3.dll</file_name>
</file_ref>
<file_ref>
<file_name>AKv8c_Bb_r1846_winx86_SSSE3x.txt</file_name>
<open_name>mb_cmdline.txt</open_name>
</file_ref>
</app_version>
<app_version>
<app_name>setiathome_v7</app_name>
<version_num>700</version_num>
<platform>windows_x86_64</platform>
<file_ref>
<file_name>AKv8c_Bb_r1846_winx86_SSSE3x.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>libfftw3f-3.dll</file_name>
</file_ref>
<file_ref>
<file_name>AKv8c_Bb_r1846_winx86_SSSE3x.txt</file_name>
<open_name>mb_cmdline.txt</open_name>
</file_ref>
</app_version>
<app>
<name>setiathome_enhanced</name>
</app>
<file_info>
<name>AKv8c_Bb_r1846_winx86_SSSE3x.exe</name>
<executable/>
</file_info>
<file_info>
<name>libfftw3f-3.dll</name>
<executable/>
</file_info>
<file_info>
<name>AKv8c_Bb_r1846_winx86_SSSE3x.txt</name>
</file_info>
<app_version>
<app_name>setiathome_enhanced</app_name>
<version_num>603</version_num>
<platform>windows_intelx86</platform>
<file_ref>
<file_name>AKv8c_Bb_r1846_winx86_SSSE3x.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>libfftw3f-3.dll</file_name>
</file_ref>
<file_ref>
<file_name>AKv8c_Bb_r1846_winx86_SSSE3x.txt</file_name>
<open_name>mb_cmdline.txt</open_name>
</file_ref>
</app_version>
<app_version>
<app_name>setiathome_enhanced</app_name>
<version_num>603</version_num>
<platform>windows_x86_64</platform>
<file_ref>
<file_name>AKv8c_Bb_r1846_winx86_SSSE3x.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>libfftw3f-3.dll</file_name>
</file_ref>
<file_ref>
<file_name>AKv8c_Bb_r1846_winx86_SSSE3x.txt</file_name>
<open_name>mb_cmdline.txt</open_name>
</file_ref>
</app_version>
<app>
<name>astropulse_v6</name>
</app>
<file_info>
<name>AP6_win_x86_SSE2_OpenCL_NV_r1843.exe</name>
<executable/>
</file_info>
<file_info>
<name>libfftw3f-3.dll</name>
<executable/>
</file_info>
<file_info>
<name>ap_cmdline_win_x86_SSE2_OpenCL_NV.txt</name>
</file_info>
<file_info>
<name>AstroPulse_Kernels_r1843.cl</name>
</file_info>
<app_version>
<app_name>astropulse_v6</app_name>
<version_num>604</version_num>
<platform>windows_intelx86</platform>
<avg_ncpus>0.04</avg_ncpus>
<max_ncpus>0.2</max_ncpus>
<plan_class>cuda_opencl_100</plan_class>
<cmdline></cmdline>
<coproc>
<type>CUDA</type>
<count>1</count>
</coproc>
<file_ref>
<file_name>AP6_win_x86_SSE2_OpenCL_NV_r1843.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>libfftw3f-3.dll</file_name>
</file_ref>
<file_ref>
<file_name>ap_cmdline_win_x86_SSE2_OpenCL_NV.txt</file_name>
<open_name>ap_cmdline.txt</open_name>
</file_ref>
<file_ref>
<file_name>AstroPulse_Kernels_r1843.cl</file_name>
</file_ref>
</app_version>
<app_version>
<app_name>astropulse_v6</app_name>
<version_num>604</version_num>
<platform>windows_intelx86</platform>
<avg_ncpus>0.04</avg_ncpus>
<max_ncpus>0.2</max_ncpus>
<plan_class>opencl_nvidia_100</plan_class>
<cmdline></cmdline>
<coproc>
<type>CUDA</type>
<count>1</count>
</coproc>
<file_ref>
<file_name>AP6_win_x86_SSE2_OpenCL_NV_r1843.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>libfftw3f-3.dll</file_name>
</file_ref>
<file_ref>
<file_name>ap_cmdline_win_x86_SSE2_OpenCL_NV.txt</file_name>
<open_name>ap_cmdline.txt</open_name>
</file_ref>
<file_ref>
<file_name>AstroPulse_Kernels_r1843.cl</file_name>
</file_ref>
</app_version>
<app_version>
<app_name>astropulse_v6</app_name>
<version_num>604</version_num>
<platform>windows_x86_64</platform>
<avg_ncpus>0.04</avg_ncpus>
<max_ncpus>0.2</max_ncpus>
<plan_class>cuda_opencl_100</plan_class>
<cmdline></cmdline>
<coproc>
<type>CUDA</type>
<count>1</count>
</coproc>
<file_ref>
<file_name>AP6_win_x86_SSE2_OpenCL_NV_r1843.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>libfftw3f-3.dll</file_name>
</file_ref>
<file_ref>
<file_name>ap_cmdline_win_x86_SSE2_OpenCL_NV.txt</file_name>
<open_name>ap_cmdline.txt</open_name>
</file_ref>
<file_ref>
<file_name>AstroPulse_Kernels_r1843.cl</file_name>
</file_ref>
</app_version>
<app_version>
<app_name>astropulse_v6</app_name>
<version_num>604</version_num>
<platform>windows_x86_64</platform>
<avg_ncpus>0.04</avg_ncpus>
<max_ncpus>0.2</max_ncpus>
<plan_class>opencl_nvidia_100</plan_class>
<cmdline></cmdline>
<coproc>
<type>CUDA</type>
<count>1</count>
</coproc>
<file_ref>
<file_name>AP6_win_x86_SSE2_OpenCL_NV_r1843.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>libfftw3f-3.dll</file_name>
</file_ref>
<file_ref>
<file_name>ap_cmdline_win_x86_SSE2_OpenCL_NV.txt</file_name>
<open_name>ap_cmdline.txt</open_name>
</file_ref>
<file_ref>
<file_name>AstroPulse_Kernels_r1843.cl</file_name>
</file_ref>
</app_version>
<app>
<name>astropulse_v6</name>
</app>
<file_info>
<name>AP6_win_x86_SSE_CPU_r1797.exe</name>
<executable/>
</file_info>
<file_info>
<name>libfftw3f-3.dll</name>
<executable/>
</file_info>
<file_info>
<name>ap_cmdline_win_x86_SSE_CPU.txt</name>
</file_info>
<app_version>
<app_name>astropulse_v6</app_name>
<version_num>601</version_num>
<platform>windows_intelx86</platform>
<cmdline></cmdline>
<file_ref>
<file_name>AP6_win_x86_SSE_CPU_r1797.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>libfftw3f-3.dll</file_name>
</file_ref>
<file_ref>
<file_name>ap_cmdline_win_x86_SSE_CPU.txt</file_name>
<open_name>ap_cmdline.txt</open_name>
</file_ref>
</app_version>
<app_version>
<app_name>astropulse_v6</app_name>
<version_num>601</version_num>
<platform>windows_x86_64</platform>
<cmdline></cmdline>
<file_ref>
<file_name>AP6_win_x86_SSE_CPU_r1797.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>libfftw3f-3.dll</file_name>
</file_ref>
<file_ref>
<file_name>ap_cmdline_win_x86_SSE_CPU.txt</file_name>
<open_name>ap_cmdline.txt</open_name>
</file_ref>
</app_version>
<app>
<name>setiathome_v7</name>
</app>
<file_info>
<name>Lunatics_x41zc_win32_cuda32.exe</name>
<executable/>
</file_info>
<file_info>
<name>cudart32_32_16.dll</name>
<executable/>
</file_info>
<file_info>
<name>cufft32_32_16.dll</name>
<executable/>
</file_info>
<file_info>
<name>mbcuda.cfg</name>
</file_info>
<app_version>
<app_name>setiathome_v7</app_name>
<version_num>700</version_num>
<platform>windows_intelx86</platform>
<plan_class>cuda32</plan_class>
<avg_ncpus>0.040000</avg_ncpus>
<max_ncpus>0.040000</max_ncpus>
<coproc>
<type>CUDA</type>
<count>1</count>
</coproc>
<file_ref>
<file_name>Lunatics_x41zc_win32_cuda32.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>cudart32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>cufft32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>mbcuda.cfg</file_name>
</file_ref>
</app_version>
<app_version>
<app_name>setiathome_v7</app_name>
<version_num>700</version_num>
<platform>windows_intelx86</platform>
<plan_class>cuda22</plan_class>
<avg_ncpus>0.040000</avg_ncpus>
<max_ncpus>0.040000</max_ncpus>
<coproc>
<type>CUDA</type>
<count>1</count>
</coproc>
<file_ref>
<file_name>Lunatics_x41zc_win32_cuda32.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>cudart32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>cufft32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>mbcuda.cfg</file_name>
</file_ref>
</app_version>
<app_version>
<app_name>setiathome_v7</app_name>
<version_num>700</version_num>
<platform>windows_intelx86</platform>
<plan_class>cuda23</plan_class>
<avg_ncpus>0.040000</avg_ncpus>
<max_ncpus>0.040000</max_ncpus>
<coproc>
<type>CUDA</type>
<count>1</count>
</coproc>
<file_ref>
<file_name>Lunatics_x41zc_win32_cuda32.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>cudart32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>cufft32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>mbcuda.cfg</file_name>
</file_ref>
</app_version>
<app_version>
<app_name>setiathome_v7</app_name>
<version_num>700</version_num>
<platform>windows_intelx86</platform>
<plan_class>cuda42</plan_class>
<avg_ncpus>0.040000</avg_ncpus>
<max_ncpus>0.040000</max_ncpus>
<coproc>
<type>CUDA</type>
<count>1</count>
</coproc>
<file_ref>
<file_name>Lunatics_x41zc_win32_cuda32.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>cudart32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>cufft32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>mbcuda.cfg</file_name>
</file_ref>
</app_version>
<app_version>
<app_name>setiathome_v7</app_name>
<version_num>700</version_num>
<platform>windows_intelx86</platform>
<plan_class>cuda50</plan_class>
<avg_ncpus>0.040000</avg_ncpus>
<max_ncpus>0.040000</max_ncpus>
<coproc>
<type>CUDA</type>
<count>1</count>
</coproc>
<file_ref>
<file_name>Lunatics_x41zc_win32_cuda32.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>cudart32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>cufft32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>mbcuda.cfg</file_name>
</file_ref>
</app_version>
<app_version>
<app_name>setiathome_v7</app_name>
<version_num>700</version_num>
<platform>windows_x86_64</platform>
<plan_class>cuda32</plan_class>
<avg_ncpus>0.040000</avg_ncpus>
<max_ncpus>0.040000</max_ncpus>
<coproc>
<type>CUDA</type>
<count>1</count>
</coproc>
<file_ref>
<file_name>Lunatics_x41zc_win32_cuda32.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>cudart32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>cufft32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>mbcuda.cfg</file_name>
</file_ref>
</app_version>
<app_version>
<app_name>setiathome_v7</app_name>
<version_num>700</version_num>
<platform>windows_x86_64</platform>
<plan_class>cuda22</plan_class>
<avg_ncpus>0.040000</avg_ncpus>
<max_ncpus>0.040000</max_ncpus>
<coproc>
<type>CUDA</type>
<count>1</count>
</coproc>
<file_ref>
<file_name>Lunatics_x41zc_win32_cuda32.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>cudart32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>cufft32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>mbcuda.cfg</file_name>
</file_ref>
</app_version>
<app_version>
<app_name>setiathome_v7</app_name>
<version_num>700</version_num>
<platform>windows_x86_64</platform>
<plan_class>cuda23</plan_class>
<avg_ncpus>0.040000</avg_ncpus>
<max_ncpus>0.040000</max_ncpus>
<coproc>
<type>CUDA</type>
<count>1</count>
</coproc>
<file_ref>
<file_name>Lunatics_x41zc_win32_cuda32.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>cudart32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>cufft32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>mbcuda.cfg</file_name>
</file_ref>
</app_version>
<app_version>
<app_name>setiathome_v7</app_name>
<version_num>700</version_num>
<platform>windows_x86_64</platform>
<plan_class>cuda42</plan_class>
<avg_ncpus>0.040000</avg_ncpus>
<max_ncpus>0.040000</max_ncpus>
<coproc>
<type>CUDA</type>
<count>1</count>
</coproc>
<file_ref>
<file_name>Lunatics_x41zc_win32_cuda32.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>cudart32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>cufft32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>mbcuda.cfg</file_name>
</file_ref>
</app_version>
<app_version>
<app_name>setiathome_v7</app_name>
<version_num>700</version_num>
<platform>windows_x86_64</platform>
<plan_class>cuda50</plan_class>
<avg_ncpus>0.040000</avg_ncpus>
<max_ncpus>0.040000</max_ncpus>
<coproc>
<type>CUDA</type>
<count>1</count>
</coproc>
<file_ref>
<file_name>Lunatics_x41zc_win32_cuda32.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>cudart32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>cufft32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>mbcuda.cfg</file_name>
</file_ref>
</app_version>
<app>
<name>setiathome_enhanced</name>
</app>
<file_info>
<name>Lunatics_x41zc_win32_cuda32.exe</name>
<executable/>
</file_info>
<file_info>
<name>cudart32_32_16.dll</name>
<executable/>
</file_info>
<file_info>
<name>cufft32_32_16.dll</name>
<executable/>
</file_info>
<file_info>
<name>mbcuda.cfg</name>
</file_info>
<app_version>
<app_name>setiathome_enhanced</app_name>
<version_num>610</version_num>
<platform>windows_intelx86</platform>
<plan_class>cuda_fermi</plan_class>
<avg_ncpus>0.040000</avg_ncpus>
<max_ncpus>0.040000</max_ncpus>
<coproc>
<type>CUDA</type>
<count>1</count>
</coproc>
<file_ref>
<file_name>Lunatics_x41zc_win32_cuda32.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>cudart32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>cufft32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>mbcuda.cfg</file_name>
</file_ref>
</app_version>
<app_version>
<app_name>setiathome_enhanced</app_name>
<version_num>609</version_num>
<platform>windows_intelx86</platform>
<plan_class>cuda23</plan_class>
<avg_ncpus>0.040000</avg_ncpus>
<max_ncpus>0.040000</max_ncpus>
<coproc>
<type>CUDA</type>
<count>1</count>
</coproc>
<file_ref>
<file_name>Lunatics_x41zc_win32_cuda32.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>cudart32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>cufft32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>mbcuda.cfg</file_name>
</file_ref>
</app_version>
<app_version>
<app_name>setiathome_enhanced</app_name>
<version_num>608</version_num>
<platform>windows_intelx86</platform>
<plan_class>cuda</plan_class>
<avg_ncpus>0.040000</avg_ncpus>
<max_ncpus>0.040000</max_ncpus>
<coproc>
<type>CUDA</type>
<count>1</count>
</coproc>
<file_ref>
<file_name>Lunatics_x41zc_win32_cuda32.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>cudart32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>cufft32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>mbcuda.cfg</file_name>
</file_ref>
</app_version>
<app_version>
<app_name>setiathome_enhanced</app_name>
<version_num>610</version_num>
<platform>windows_x86_64</platform>
<plan_class>cuda_fermi</plan_class>
<avg_ncpus>0.040000</avg_ncpus>
<max_ncpus>0.040000</max_ncpus>
<coproc>
<type>CUDA</type>
<count>1</count>
</coproc>
<file_ref>
<file_name>Lunatics_x41zc_win32_cuda32.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>cudart32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>cufft32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>mbcuda.cfg</file_name>
</file_ref>
</app_version>
<app_version>
<app_name>setiathome_enhanced</app_name>
<version_num>609</version_num>
<platform>windows_x86_64</platform>
<plan_class>cuda23</plan_class>
<avg_ncpus>0.040000</avg_ncpus>
<max_ncpus>0.040000</max_ncpus>
<coproc>
<type>CUDA</type>
<count>1</count>
</coproc>
<file_ref>
<file_name>Lunatics_x41zc_win32_cuda32.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>cudart32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>cufft32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>mbcuda.cfg</file_name>
</file_ref>
</app_version>
<app_version>
<app_name>setiathome_enhanced</app_name>
<version_num>608</version_num>
<platform>windows_x86_64</platform>
<plan_class>cuda</plan_class>
<avg_ncpus>0.040000</avg_ncpus>
<max_ncpus>0.040000</max_ncpus>
<coproc>
<type>CUDA</type>
<count>1</count>
</coproc>
<file_ref>
<file_name>Lunatics_x41zc_win32_cuda32.exe</file_name>
<main_program/>
</file_ref>
<file_ref>
<file_name>cudart32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>cufft32_32_16.dll</file_name>
</file_ref>
<file_ref>
<file_name>mbcuda.cfg</file_name>
</file_ref>
</app_version>
</app_info>


Thanks for looking.

Cheers,
Kieron.
ID: 1481596 · Report as offensive
Claggy
Volunteer tester

Send message
Joined: 5 Jul 99
Posts: 4654
Credit: 47,537,079
RAC: 4
United Kingdom
Message 1481601 - Posted: 25 Feb 2014, 11:14:50 UTC - in response to Message 1481596.  
Last modified: 25 Feb 2014, 11:29:04 UTC

Hi,
I'm a long-time SETI participant but have so far stuck to just running the Lunatics apps with its default settings. I'd like to squeeze a bit more performance out of my PC if possible so to run 2 GPU tasks simultaneously I should be leaving 1 CPU core free - or is it 2?

When I look at my app_info file I see multiple entries for <avg_ncpus> and </max_ncpus>; do I need to change them all or just some of them?

Also what do I change to run 2 simultaneous tasks on the GPU?

My system: Quad Core Q6600 @2.40GHz, 8GB RAM, 64-bit Win8.1, GTX460 1GB GPU.

First of all, rerun the Installer, and choose the Cuda42 app, the Cuda32 app is the default choice, on a GTX460 the Cuda42 app will be fastest,

Rather than doing modifications to the app_info, and having the risk of making a typo and loosing your work, make an app_config.xml (with notepad) with the following instead:

<app_config>
<app>
<name>astropulse_v6</name>
<gpu_versions>
<gpu_usage>0.5</gpu_usage>
<cpu_usage>1.0</cpu_usage>
</gpu_versions>
</app>
<app>
<name>setiathome_v7</name>
<gpu_versions>
<gpu_usage>0.5</gpu_usage>
<cpu_usage>0.1</cpu_usage>
</gpu_versions>
</app_config>

Drop the file in your setiathome project directory (after making sure it's not named app_config.xml.txt), and do a read config file (It also means you can make changes without having to restart Boinc),
the above app_config.xml is setup to run two Seti v7 tasks on the GPU, and two AP tasks on the GPU, and while running AP tasks to reserve a CPU core.

To improve the x41zc app's load, in the mbcuda.cfg file, remove the ; from the processpriority line, and up the priority line from belownormal to normal or abovenormal, the next time the app starts the new priority will be used.

You can also supply some incresed parameters to NV AP app to improve it's performance,
In the ap_cmdline_win_x86_SSE2_OpenCL_NV.txt file add the following text: -unroll 10 -ffa_block 6144 -ffa_block_fetch 1536

Interesting that your GTX460 running 334.89 drivers and the rev 1843 AP app has very low CPU usage, perhaps Nvidia has listened to Raistmer and fixed the High CPU usage Bug already:

All AstroPulse v6 tasks for computer 7212588

Claggy
ID: 1481601 · Report as offensive
Kieron Walsh

Send message
Joined: 2 Mar 00
Posts: 74
Credit: 43,502,325
RAC: 112
United Kingdom
Message 1481613 - Posted: 25 Feb 2014, 12:10:19 UTC

Thanks Claggy, much appreciated.
I ran the Lunatics installer choosing Cuda42 which seems to have had a positive effect. Newly downloaded GPU tasks are badged as Cuda42

After creating the app_config.xml file however, the number/mix of tasks running hasn't changed; I'm still running 4xCPU "setiathome_v7 7.00" and 1xGPU "setiathome_v7 7.00 (Cuda32)". Against the single running GPU task it says "Running (0.04CPUs + 1NVIDIA GPU)".

The app_config.xml file I created is being 'Found' as shown in the Boinc Event Log but it doesn't seem to be having the desired effect. Anything else I need to do?

Cheers,
Kieron.
ID: 1481613 · Report as offensive
Kieron Walsh

Send message
Joined: 2 Mar 00
Posts: 74
Credit: 43,502,325
RAC: 112
United Kingdom
Message 1481615 - Posted: 25 Feb 2014, 12:17:07 UTC

Me again!
In the mbcuda.cfg file which 'processpriority' line should I be un-remarking?

;;; This configuration file is for optional control of Cuda Multibeam x41zc
;;; Currently, the available options are for
;;; application process priority control (without external tools), and
;;; per gpu priority control (useful for multiple Cuda GPU systems)
[mbcuda]
;;;;; Global applications settings, to apply to all Cuda devices
;;; You can uncomment the processpriority line below, by removing the ';', to engage machine global priority control of x41x
;;; possible options are 'belownormal' (which is the default), 'normal', 'abovenormal', or 'high'
;;; For dedicated crunching machines, 'abovenormal' is recommended
;;; raising global application priorities above the default
;;; may have system dependant usability effects, and can have positive or negative effects on overall throughput
;processpriority = abovenormal
;;; Pulsefinding: Advanced options for long pulsefinds (affect display usability & long kernel runs)
;;; defaults are conservative.
;;; WARNING: Excessive values may induce display lag, driver timeout & recovery, or errors.
;;; pulsefinding blocks per multiprocessor (1-16), default is 1 for Pre-Fermi, 4 for Fermi or newer GPUs
;pfblockspersm = 8
;;; pulsefinding maximum periods per kernel launch (1-1000), default is 100, as per 6.09
;pfperiodsperlaunch = 200

;[bus1slot0]
;;; Optional GPU specifc control (requires Cuda 3.2 or newer app), example
;processpriority = abovenormal
;pfblockspersm = 8
;pfperiodsperlaunch = 200


Thanks.
ID: 1481615 · Report as offensive
juan BFP Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 16 Mar 07
Posts: 9786
Credit: 572,710,851
RAC: 3,799
Panama
Message 1481616 - Posted: 25 Feb 2014, 12:18:34 UTC

The blue one.
ID: 1481616 · Report as offensive
Kieron Walsh

Send message
Joined: 2 Mar 00
Posts: 74
Credit: 43,502,325
RAC: 112
United Kingdom
Message 1481618 - Posted: 25 Feb 2014, 12:23:32 UTC - in response to Message 1481616.  

Thanks Juan.

Doh! As soon as I had sent that follow-up question I knew I was going to look stupid! The crucial word 'example' in the lower section kind of gives the answer doesn't it!
ID: 1481618 · Report as offensive
Previous · 1 · 2 · 3 · Next

Message boards : Number crunching : Question about dedicating CPU cores to GPU support w/Lunatics apps.


 
©2024 University of California
 
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.