LotzaCores and a GTX 1080 FTW

Message boards : Number crunching : LotzaCores and a GTX 1080 FTW
Message board moderation

To post messages, you must log in.

Previous · 1 . . . 4 · 5 · 6 · 7 · 8 · 9 · 10 . . . 11 · Next

AuthorMessage
Al Special Project $250 donor
Avatar

Send message
Joined: 3 Apr 99
Posts: 1618
Credit: 342,633,783
RAC: 274,409
United States
Message 1793834 - Posted: 6 Jun 2016, 2:52:32 UTC - in response to Message 1793832.  
Last modified: 6 Jun 2016, 3:22:41 UTC

Well, that was interesting. I copied and pasted your text above and created the file, then I restarted BOINC, it took maybe 20-30 seconds to start, and then said that my GTX 950 couldn't be ejected. I thought ok, that's good, nobody wants my video card ejected from the computer, that would be a little extreme. I tweeked it to .5 on both and now it suspended 2 of the 8 GPU tasks that were running, but I still have all 8 cores running. I renamed it to prevent it from running on the next startup, and now my 4th card is no longer recognized, I have 3 cards running 2 tasks each, and one task that used to be running, now saying waiting to run, for a GPU to open up it appears. So, apparently somehow it did eject my 950, at least in it's mind, my question is how do I get it back? Sheeesh, such drama...

*edit* Just for fun, I restarted my computer, just to start everything fresh, looked in Precision 16, and it lists only the 3 cards. Here is my logfile, now really not sure what is going on here but for whatever reason it seems like it's having a hardware issue:

6/5/2016 9:55:14 PM |  | Starting BOINC client version 7.6.22 for windows_x86_64
6/5/2016 9:55:14 PM |  | log flags: file_xfer, sched_ops, task
6/5/2016 9:55:14 PM |  | Libraries: libcurl/7.45.0 OpenSSL/1.0.2d zlib/1.2.8
6/5/2016 9:55:14 PM |  | Data directory: C:\ProgramData\BOINC
6/5/2016 9:55:14 PM |  | Running under account Flash
6/5/2016 9:55:14 PM |  | CUDA: NVIDIA GPU 0: GeForce GTX 950 (driver version 368.22, CUDA version 8.0, compute capability 5.2, 2048MB, 1940MB available, 2158 GFLOPS peak)
6/5/2016 9:55:14 PM |  | CUDA: NVIDIA GPU 1: GeForce GTX 670 (driver version 368.22, CUDA version 8.0, compute capability 3.0, 2048MB, 1959MB available, 2915 GFLOPS peak)
6/5/2016 9:55:14 PM |  | CUDA: NVIDIA GPU 2: GeForce GTX 670 (driver version 368.22, CUDA version 8.0, compute capability 3.0, 2048MB, 1951MB available, 2915 GFLOPS peak)
6/5/2016 9:55:14 PM |  | OpenCL: NVIDIA GPU 0: GeForce GTX 950 (driver version 368.22, device version OpenCL 1.2 CUDA, 2048MB, 1940MB available, 2158 GFLOPS peak)
6/5/2016 9:55:14 PM |  | OpenCL: NVIDIA GPU 1: GeForce GTX 670 (driver version 368.22, device version OpenCL 1.2 CUDA, 2048MB, 1959MB available, 2915 GFLOPS peak)
6/5/2016 9:55:14 PM |  | OpenCL: NVIDIA GPU 2: GeForce GTX 670 (driver version 368.22, device version OpenCL 1.2 CUDA, 2048MB, 1951MB available, 2915 GFLOPS peak)
6/5/2016 9:55:14 PM | SETI@home | Found app_info.xml; using anonymous platform
6/5/2016 9:55:14 PM |  | Host name: FlashFlyer
6/5/2016 9:55:14 PM |  | Processor: 8 GenuineIntel        Intel(R) Core(TM) i7-3770K CPU @ 3.50GHz [Family 6 Model 58 Stepping 9]
6/5/2016 9:55:14 PM |  | Processor features: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss htt tm pni ssse3 cx16 sse4_1 sse4_2 popcnt aes f16c rdrandsyscall nx lm avx vmx tm2 pbe fsgsbase smep
6/5/2016 9:55:14 PM |  | OS: Microsoft Windows 7: Ultimate x64 Edition, Service Pack 1, (06.01.7601.00)
6/5/2016 9:55:14 PM |  | Memory: 31.95 GB physical, 63.89 GB virtual
6/5/2016 9:55:14 PM |  | Disk: 223.47 GB total, 137.07 GB free
6/5/2016 9:55:14 PM |  | Local time is UTC -5 hours
6/5/2016 9:55:14 PM |  | Config: event log limit disabled
6/5/2016 9:55:14 PM |  | Config: use all coprocessors
6/5/2016 9:55:14 PM | SETI@home | URL http://setiathome.berkeley.edu/; Computer ID 8017700; resource share 100
6/5/2016 9:55:14 PM | SETI@home | General prefs: from SETI@home (last modified 03-Apr-2013 23:59:56)
6/5/2016 9:55:14 PM | SETI@home | Computer location: home
6/5/2016 9:55:14 PM | SETI@home | General prefs: no separate prefs for home; using your defaults
6/5/2016 9:55:14 PM |  | Reading preferences override file
6/5/2016 9:55:14 PM |  | Preferences:
6/5/2016 9:55:14 PM |  | max memory usage when active: 16357.26MB
6/5/2016 9:55:14 PM |  | max memory usage when idle: 31078.80MB
6/5/2016 9:55:14 PM |  | max disk usage: 100.00GB
6/5/2016 9:55:14 PM |  | (to change preferences, visit a project web site or select Preferences in the Manager)


ID: 1793834 · Report as offensive
Profile Zalster Special Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 27 May 99
Posts: 4433
Credit: 260,389,843
RAC: 11,499
United States
Message 1793838 - Posted: 6 Jun 2016, 3:23:22 UTC - in response to Message 1793834.  

Al,

You have a cc_config.xml also telling it to use all the GPUs?

Can't remember if you said if you did or not...
ID: 1793838 · Report as offensive
Grant (SSSF)
Volunteer tester

Send message
Joined: 19 Aug 99
Posts: 9893
Credit: 128,233,389
RAC: 77,965
Australia
Message 1793841 - Posted: 6 Jun 2016, 3:29:09 UTC - in response to Message 1793834.  

<app_config>
<app>
<name>setiathome_v8</name>
<gpu_versions>
<gpu_usage>0.50</gpu_usage>
<cpu_usage>1.00</cpu_usage>
</gpu_versions>
</app>
</app_config>

As long as it's in the /projects/setiathome.berkeley.edu folder and is named app_config.xml then BOINC should be able to read it & use it.


Easy way to check if it is being used or not- change <gpu_usage>0.50</gpu_usage> to <gpu_usage>1.00</gpu_usage> (or <gpu_usage>0.33</gpu_usage>) and then Options, Read config files. That will result in 1 WU per card (or 3).
If that works, then change it back to <gpu_usage>0.50</gpu_usage> & then make <cpu_usage>2.00</cpu_usage> and re-read the config file- that should result in 2 CPU cores no longer crunching CPU work units for every 1 GPU WU running. Change it back to <cpu_usage>0.50</cpu_usage> & re-read config files & you should have 1 CPU core reserved for every 2 GPU WUs running.


If nothing happens then it's the wrong name, wrong location or something in the file got mucked up.
Grant
Darwin NT
ID: 1793841 · Report as offensive
Grant (SSSF)
Volunteer tester

Send message
Joined: 19 Aug 99
Posts: 9893
Credit: 128,233,389
RAC: 77,965
Australia
Message 1793847 - Posted: 6 Jun 2016, 3:44:19 UTC - in response to Message 1793838.  
Last modified: 6 Jun 2016, 3:45:11 UTC

Al,

You have a cc_config.xml also telling it to use all the GPUs?

Can't remember if you said if you did or not...


For Al,

From the BOINC reference,
<use_all_gpus>0|1</use_all_gpus>If 1, use all GPUs (otherwise only the most capable ones are used).

So it would be

cc_config
  <options>
    <use_all_gpus>1</use_all_gpus>
  </options>
</cc_config


in a file named cc_config.xml, and this one is located in the
C:\Program data\BOINC
folder
Grant
Darwin NT
ID: 1793847 · Report as offensive
Al Special Project $250 donor
Avatar

Send message
Joined: 3 Apr 99
Posts: 1618
Credit: 342,633,783
RAC: 274,409
United States
Message 1793913 - Posted: 6 Jun 2016, 10:19:10 UTC - in response to Message 1793838.  

Zalster, yes, I have that line added, it is running all the GPUs, well, all but one, it appears that it may have been an unfortunate coincidence that the card flaked out just as I was testing that config file, as looking at the device manager, it is disabled in Windows, possibly due to a resource conflict. I am going to have to try digging into that one a little more, as the card does show up, it'a just disabled. I did do a search about ejecting cards, and found a post on Reddit about the Nvidia video driver issue from about 4 months or so ago, basically saying it was a bug in that version, and that they had added support for external Thunderbolt 3 GPUs in that update, and for some reason it shows it for all GPUs and not just GPUs connected through Thunderbolt. As there had been a couple releases since then, not sure what to make of it, but I am letting my cache run down on this machine, and may just end up reloading it, as there comes a point where it just doesn't make sense to futz with it chasing tails as such. It may fix it, possibly not, but if after a few tries (with things like driver reinstall/regression, etc.) I don't come up with a good solution, it's Fdisk time for this bad boy.

ID: 1793913 · Report as offensive
Al Special Project $250 donor
Avatar

Send message
Joined: 3 Apr 99
Posts: 1618
Credit: 342,633,783
RAC: 274,409
United States
Message 1793914 - Posted: 6 Jun 2016, 10:22:39 UTC - in response to Message 1793847.  

Thanks Grant, this is the cc_config I use on all of my machines, weather they have one or 4+ GPUs in them, as I don't believe it will harm anything having it in there with just one card?

<cc_config> 
 <options> 
  <use_all_gpus>1</use_all_gpus>
  <save_stats_days>10000</save_stats_days> 
  <max_event_log_lines>0</max_event_log_lines>
  <max_stdout_file_size>40</max_stdout_file_size>
 </options>
</cc_config>


ID: 1793914 · Report as offensive
Richard Haselgrove Project Donor
Volunteer tester

Send message
Joined: 4 Jul 99
Posts: 11991
Credit: 118,567,752
RAC: 41,012
United Kingdom
Message 1793916 - Posted: 6 Jun 2016, 10:25:11 UTC - in response to Message 1793914.  
Last modified: 6 Jun 2016, 10:29:35 UTC

... I don't believe it will harm anything having it in there with just one card?

Correct. No problem with that.

Edit - not sure <max_event_log_lines> 0 is such a good idea, though. The Event Log is your friend.
ID: 1793916 · Report as offensive
Al Special Project $250 donor
Avatar

Send message
Joined: 3 Apr 99
Posts: 1618
Credit: 342,633,783
RAC: 274,409
United States
Message 1793921 - Posted: 6 Jun 2016, 10:49:26 UTC - in response to Message 1793916.  

Richard, I was told that setting it to 0 disables the limit, it will continue to grow and not overwrite itself. I completely agree with you, it is def your friend, certainly wouldn't want it to not work! :-)

ID: 1793921 · Report as offensive
Richard Haselgrove Project Donor
Volunteer tester

Send message
Joined: 4 Jul 99
Posts: 11991
Credit: 118,567,752
RAC: 41,012
United Kingdom
Message 1793926 - Posted: 6 Jun 2016, 11:22:01 UTC - in response to Message 1793921.  

I think 'max lines' controls the number of current lines (since startup) sent by the client for live display in BOINC Manager - that can get sluggish if you go a long way beyond the 2,000 (IIRC) line default.

You might be more interested in enlarging the permanent disk archive:

<max_stdout_file_size>N</max_stdout_file_size>
Specify the maximum size of the standard out log file (stdoutdae.txt); default is 2 MB.

Two tips:

1) use the "Event Log options..." picker (Ctrl+Shift+F) once - and it creates a fully-populated cc_config.xml file, with all possible tags listed and current/default values.

2) Bookmark http://boinc.berkeley.edu/wiki/Client_configuration to find out what they all mean.

(yes, the default Event Log display limit is 2,000 lines, and 0 removes the limit - it just falls over when you run out of memory)
(the value of N in the suggestion above is bytes, not MB)
ID: 1793926 · Report as offensive
Al Special Project $250 donor
Avatar

Send message
Joined: 3 Apr 99
Posts: 1618
Credit: 342,633,783
RAC: 274,409
United States
Message 1793935 - Posted: 6 Jun 2016, 11:35:20 UTC - in response to Message 1793926.  


(yes, the default Event Log display limit is 2,000 lines, and 0 removes the limit - it just falls over when you run out of memory)
(the value of N in the suggestion above is bytes, not MB)

Oooh.. Didn't know either of those things. Hmm, well, I'll have to rethink some of those settings then, as I don't want it to tip over now.. ;-) And thanks for the heads up on the stdout file, I thought it was MB, not bytes. I swear I learn something new here every day...

ID: 1793935 · Report as offensive
Richard Haselgrove Project Donor
Volunteer tester

Send message
Joined: 4 Jul 99
Posts: 11991
Credit: 118,567,752
RAC: 41,012
United Kingdom
Message 1793971 - Posted: 6 Jun 2016, 14:19:27 UTC

Has anybody here got a GTX 1080 or 1070 active and running BOINC? If so, could you please post the GPU detection line from the startup log - including, specifically, the 'GFLOPs peak' value - I've been asked to check if that's accurate.

(and please don't just compare it with what SIV says - it's the author of SIV who's asked me, because both programs use the same formula - they'll either both be right or both be wrong.)
ID: 1793971 · Report as offensive
Al Special Project $250 donor
Avatar

Send message
Joined: 3 Apr 99
Posts: 1618
Credit: 342,633,783
RAC: 274,409
United States
Message 1794037 - Posted: 6 Jun 2016, 17:58:11 UTC - in response to Message 1793971.  

Well Richard, don't look to me anytime soon, I just got off the phone with EVGA about my new machines video weirdness, so I gave him the order # to check the shipping status, and he said that it would be shipping later in the month. Min a couple weeks, but almost for certain by the end of the month. I asked if there was a way to change the shipping method, because I had thought it would be shipping by the end of this week initially, so I chose 2 day and paid an extra $18 for it, but if it's coming that late, well, who cares about a couple 3 more days? Of course, the only way to change it is to cancel the pre-order, and then there is no chance of re-ordering it at this time. Yippee. So, as I said I'll make a note of it here when I get it installed and running, and maybe use it as a comparison of the Reference version vs. the FTW version, not sure how much difference there will be, but at least the data will be out there for all to see.

ID: 1794037 · Report as offensive
Profile tullio Project Donor
Volunteer moderator
Volunteer tester

Send message
Joined: 9 Apr 04
Posts: 6680
Credit: 1,966,057
RAC: 346
Italy
Message 1794052 - Posted: 6 Jun 2016, 18:32:57 UTC
Last modified: 6 Jun 2016, 18:35:16 UTC

I don't have a GTX 1080 but I have downloaded its nVidia driver, 365.19 and it works on my GTX 750 OC.
Tullio
ID: 1794052 · Report as offensive
Rasputin42
Volunteer tester

Send message
Joined: 25 Jul 08
Posts: 412
Credit: 5,834,453
RAC: 0
United States
Message 1794056 - Posted: 6 Jun 2016, 18:48:59 UTC - in response to Message 1794052.  

Meaning?
ID: 1794056 · Report as offensive
archae86

Send message
Joined: 31 Aug 99
Posts: 909
Credit: 1,582,816
RAC: 0
United States
Message 1794065 - Posted: 6 Jun 2016, 19:32:48 UTC - in response to Message 1794052.  
Last modified: 6 Jun 2016, 19:34:16 UTC

I don't have a GTX 1080 but I have downloaded its nVidia driver, 365.19 and it works on my GTX 750 OC.
Tullio

Curious, this Nvidia driver download site currently advises using 368.25 for a Windows 10 64-bit 1080 host, and 368.22 for the same host with a 750.
ID: 1794065 · Report as offensive
Al Special Project $250 donor
Avatar

Send message
Joined: 3 Apr 99
Posts: 1618
Credit: 342,633,783
RAC: 274,409
United States
Message 1794087 - Posted: 6 Jun 2016, 21:25:27 UTC - in response to Message 1794065.  
Last modified: 6 Jun 2016, 21:26:28 UTC

They are different, and apparently non-compatible drivers as I found out, because I didn't choose the correct OS (needed Win 7-64 bit, but picked the default one which was Win 10). I tried installing it, and it popped up saying Wrong OS, Dummy. lol It lists the Win 10 drivers separately from the Vista, 7 & 8 drivers, not sure what changed in 10 but looks like they had to tweak something for 10 and gave it a different version #, so there you go.

ID: 1794087 · Report as offensive
Profile jason_gee
Volunteer developer
Volunteer tester
Avatar

Send message
Joined: 24 Nov 06
Posts: 7489
Credit: 91,093,184
RAC: 0
Australia
Message 1794094 - Posted: 6 Jun 2016, 22:03:07 UTC

Comparing the information for 368.25 on nVidia's download page:
Supported Products: GeForce 10 Series:
GeForce GTX 1080

and regular 368.22:

Supported Products: GeForce 900 Series:
GeForce GTX TITAN X, GeForce GTX 980 Ti, GeForce GTX 980, GeForce GTX 970, GeForce GTX 960, GeForce GTX 950

GeForce 700 Series:
GeForce GTX TITAN Z, GeForce GTX TITAN Black, GeForce GTX TITAN, GeForce GTX 780 Ti, GeForce GTX 780, GeForce GTX 770, GeForce GTX 760, GeForce GTX 760 Ti (OEM), GeForce GTX 750 Ti, GeForce GTX 750, GeForce GTX 745, GeForce GT 740, GeForce GT 730, GeForce GT 720, GeForce GT 710, GeForce GT 705

GeForce 600 Series:
GeForce GTX 690, GeForce GTX 680, GeForce GTX 670, GeForce GTX 660 Ti, GeForce GTX 660, GeForce GTX 650 Ti BOOST, GeForce GTX 650 Ti, GeForce GTX 650, GeForce GTX 645, GeForce GT 645, GeForce GT 640, GeForce GT 635, GeForce GT 630, GeForce GT 620, GeForce GT 610, GeForce 605

GeForce 500 Series:
GeForce GTX 590, GeForce GTX 580, GeForce GTX 570, GeForce GTX 560 Ti, GeForce GTX 560 SE, GeForce GTX 560, GeForce GTX 555, GeForce GTX 550 Ti, GeForce GT 545, GeForce GT 530, GeForce GT 520, GeForce 510

GeForce 400 Series:
GeForce GTX 480, GeForce GTX 470, GeForce GTX 465, GeForce GTX 460 SE v2, GeForce GTX 460 SE, GeForce GTX 460, GeForce GTS 450, GeForce GT 440, GeForce GT 430, GeForce GT 420


Seems the Pascal branch isn't merged into the mainstream yet. Don't see a 1070 yet either. I wouldn't advise mixing GPU generations with a 1080 until they are (usually happens in another driver release, or a few, after initial)
"Living by the wisdom of computer science doesn't sound so bad after all. And unlike most advice, it's backed up by proofs." -- Algorithms to live by: The computer science of human decisions.
ID: 1794094 · Report as offensive
Al Special Project $250 donor
Avatar

Send message
Joined: 3 Apr 99
Posts: 1618
Credit: 342,633,783
RAC: 274,409
United States
Message 1794095 - Posted: 6 Jun 2016, 22:13:40 UTC - in response to Message 1794094.  

Well, then maybe my card being delayed for a few more weeks isn't necessarily a bad thing, it will allow them to work out some of the driver kinks? Although I hadn't planned initially to do a multi GPU setup with it, I am going to toss it into my 48 core machine and let it run for a month or so to see how well it works.

ID: 1794095 · Report as offensive
Profile jason_gee
Volunteer developer
Volunteer tester
Avatar

Send message
Joined: 24 Nov 06
Posts: 7489
Credit: 91,093,184
RAC: 0
Australia
Message 1794101 - Posted: 6 Jun 2016, 22:24:51 UTC - in response to Message 1794095.  

Well, then maybe my card being delayed for a few more weeks isn't necessarily a bad thing, it will allow them to work out some of the driver kinks? Although I hadn't planned initially to do a multi GPU setup with it, I am going to toss it into my 48 core machine and let it run for a month or so to see how well it works.


Typically that'll be a shifting target for 3-4 months, after which the drivers settle down. Could be a bit of extra jostling than normal, due to DirectX12, OpenCL(2.1?)/Vulkan, Cuda 8.0, all attempting to mature at the same time.

At least there are Raistmer's OpenCL apps, Familiar Cuda ones, and multiple new avenues being explored on all fronts for when we know more about this architecture. One of those times when having plenty of eggs in different baskets could pay off (even if initially there is a few unwanted surprises)
"Living by the wisdom of computer science doesn't sound so bad after all. And unlike most advice, it's backed up by proofs." -- Algorithms to live by: The computer science of human decisions.
ID: 1794101 · Report as offensive
archae86

Send message
Joined: 31 Aug 99
Posts: 909
Credit: 1,582,816
RAC: 0
United States
Message 1794248 - Posted: 7 Jun 2016, 13:26:58 UTC

A thread on the Nvidia forums regarding a widely reported complaint of abnormal fan speed variation on the initial 1080 cards asserts that a driver fix has been found, tested, and scheduled for release in a driver update on June 7, 2016

Checking the Nvidia driver download page, as of a couple of minutes ago, the currently recommended driver for the 1080 and 1070 is 368.39, and unlike the previous 1080 driver this one lists a very large set of supported products, including for example the 750, 980, and a fullish looking list back through the 400 series.
ID: 1794248 · Report as offensive
Previous · 1 . . . 4 · 5 · 6 · 7 · 8 · 9 · 10 . . . 11 · Next

Message boards : Number crunching : LotzaCores and a GTX 1080 FTW


 
©2018 University of California
 
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.