HTTP internal server error for 4 days, can't phone home...

Message boards : Number crunching : HTTP internal server error for 4 days, can't phone home...
Message board moderation

To post messages, you must log in.

Previous · 1 · 2 · 3

AuthorMessage
Al Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Avatar

Send message
Joined: 3 Apr 99
Posts: 1682
Credit: 477,343,364
RAC: 482
United States
Message 1851995 - Posted: 28 Feb 2017, 4:06:24 UTC - in response to Message 1851868.  

The RAC is climbing again, and I am currently running 2 tasks per GPU, and at least when it was running CUDA, .5 core per GPU, but not sure if SOG defaults to 1-1 or if it has to be configed like CUDA. Regardless, it appears to be running ok right now, so I am going to leave it alone for a while to stabilize, and will look into tinkering with it once its been up and running for a while with none of those catistrophic GPU errors again. Crossing my fingers that that was a one time occurance...

ID: 1851995 · Report as offensive
Profile Keith Myers Special Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 29 Apr 01
Posts: 13164
Credit: 1,160,866,277
RAC: 1,873
United States
Message 1852010 - Posted: 28 Feb 2017, 6:00:06 UTC
Last modified: 28 Feb 2017, 6:06:02 UTC

Depends on whether you had the CUDA tasks running two up via a <count> entry in app_info directly or via an app_config. When you run the Lunatics installer it defaults to <count>1</count> entry in the MB section of the app_info. If you simply did a Replace of the r3557 app with the r3584 app, it would have left that as default. If you are running two up now with the SoG app, either you modified the count or have a MB app section defined to run two up in a app_config file.

I would let it run for a while so you can observe the trend. It should increase your RAC and that can be increased a great deal more with tuning parameters that the SoG app provides. There is quite a bit more tuning available compared to the CUDA50 app.

The SoG app will take whatever CPU core resource that it needs. All the .5 CPU resource does is give BOINC an idea of how many CPU tasks it should start. If you check with Task Manager or any of the other tools available to monitor BOINC, you will see that each GPU task is allocated almost a full CPU core. You can set that up explicitly in an app_config.

This is my app_config for example:

<app_config>
  <app_version>
    <app_name>setiathome_v8</app_name>
    <plan_class>opencl_nvidia_SoG</plan_class>
    <avg_ncpus>1</avg_ncpus>
    <ngpus>0.5</ngpus>
    <cmdline>-sbs 2048 -period_iterations_num 2 -tt 1500 -high_perf -spike_fft_thresh 4096 -tune 1 64 1 4 -oclfft_tune_gr 256 -oclfft_tune_lr 16 -oclfft_tune_wg 256 -oclfft_tune_ls 512 -oclfft_tune_bn 64 -oclfft_tune_cw 64 -high_prec_timer</cmdline>
  </app_version>
  <app_version>
    <app_name>astropulse_v7</app_name>
    <plan_class>opencl_nvidia_100</plan_class>
    <avg_ncpus>1</avg_ncpus>
    <ngpus>0.5</ngpus>
    <cmdline>-unroll 24 -oclFFT_plan 256 16 256 -ffa_block 16384 -ffa_block_fetch 8192 -tune 1 64 8 1 -tune 2 64 8 1</cmdline>
  </app_version>
<project_max_concurrent>7</project_max_concurrent>
</app_config>


You can see my callout for 1 CPU core in both the MB and AP sections.
Seti@Home classic workunits:20,676 CPU time:74,226 hours

A proud member of the OFA (Old Farts Association)
ID: 1852010 · Report as offensive
AMDave
Volunteer tester

Send message
Joined: 9 Mar 01
Posts: 234
Credit: 11,671,730
RAC: 0
United States
Message 1852912 - Posted: 4 Mar 2017, 17:32:22 UTC - in response to Message 1851156.  

EDIT- I usually exit and restart BOINC to make sure it gets read- not sure if Options, Read Config files works with that file.
Look just above "Logging flags" here.
ID: 1852912 · Report as offensive
Previous · 1 · 2 · 3

Message boards : Number crunching : HTTP internal server error for 4 days, can't phone home...


 
©2024 University of California
 
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.